x64 always has MMX, MMXEXT, SSE and SSE2 and this means that some functions for MMX, MMXEXT, SSE and 3dnow are always overridden by other functions (unless one e.g. explicitly disables SSE2). So given that the only systems which benefit from the 8x8 MMX (overridden by MMXEXT) or the 16x16 MMXEXT (overridden by SSE2) are truely ancient 32bit x86s they are removed. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
		
			
				
	
	
		
			329 lines
		
	
	
		
			9.6 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
			
		
		
	
	
			329 lines
		
	
	
		
			9.6 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
| ;******************************************************************************
 | |
| ;* Pixel utilities SIMD
 | |
| ;*
 | |
| ;* Copyright (C) 2002-2004 Michael Niedermayer <michaelni@gmx.at>
 | |
| ;* Copyright (C) 2014 Clément Bœsch <u pkh me>
 | |
| ;*
 | |
| ;* This file is part of FFmpeg.
 | |
| ;*
 | |
| ;* FFmpeg is free software; you can redistribute it and/or
 | |
| ;* modify it under the terms of the GNU Lesser General Public
 | |
| ;* License as published by the Free Software Foundation; either
 | |
| ;* version 2.1 of the License, or (at your option) any later version.
 | |
| ;*
 | |
| ;* FFmpeg is distributed in the hope that it will be useful,
 | |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
| ;* Lesser General Public License for more details.
 | |
| ;*
 | |
| ;* You should have received a copy of the GNU Lesser General Public
 | |
| ;* License along with FFmpeg; if not, write to the Free Software
 | |
| ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
| ;******************************************************************************
 | |
| 
 | |
| %include "libavutil/x86/x86util.asm"
 | |
| 
 | |
| SECTION .text
 | |
| 
 | |
| ;-------------------------------------------------------------------------------
 | |
| ; int ff_pixelutils_sad_8x8_mmxext(const uint8_t *src1, ptrdiff_t stride1,
 | |
| ;                                  const uint8_t *src2, ptrdiff_t stride2);
 | |
| ;-------------------------------------------------------------------------------
 | |
| INIT_MMX mmxext
 | |
| cglobal pixelutils_sad_8x8, 4,4,0, src1, stride1, src2, stride2
 | |
|     pxor        m2, m2
 | |
| %rep 4
 | |
|     mova        m0, [src1q]
 | |
|     mova        m1, [src1q + stride1q]
 | |
|     psadbw      m0, [src2q]
 | |
|     psadbw      m1, [src2q + stride2q]
 | |
|     paddw       m2, m0
 | |
|     paddw       m2, m1
 | |
|     lea         src1q, [src1q + 2*stride1q]
 | |
|     lea         src2q, [src2q + 2*stride2q]
 | |
| %endrep
 | |
|     movd        eax, m2
 | |
|     RET
 | |
| 
 | |
| ;-------------------------------------------------------------------------------
 | |
| ; int ff_pixelutils_sad_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1,
 | |
| ;                                  const uint8_t *src2, ptrdiff_t stride2);
 | |
| ;-------------------------------------------------------------------------------
 | |
| INIT_XMM sse2
 | |
| cglobal pixelutils_sad_16x16, 4,4,5, src1, stride1, src2, stride2
 | |
|     movu        m4, [src1q]
 | |
|     movu        m2, [src2q]
 | |
|     movu        m1, [src1q + stride1q]
 | |
|     movu        m3, [src2q + stride2q]
 | |
|     psadbw      m4, m2
 | |
|     psadbw      m1, m3
 | |
|     paddw       m4, m1
 | |
| %rep 7
 | |
|     lea         src1q, [src1q + 2*stride1q]
 | |
|     lea         src2q, [src2q + 2*stride2q]
 | |
|     movu        m0, [src1q]
 | |
|     movu        m2, [src2q]
 | |
|     movu        m1, [src1q + stride1q]
 | |
|     movu        m3, [src2q + stride2q]
 | |
|     psadbw      m0, m2
 | |
|     psadbw      m1, m3
 | |
|     paddw       m4, m0
 | |
|     paddw       m4, m1
 | |
| %endrep
 | |
|     movhlps     m0, m4
 | |
|     paddw       m4, m0
 | |
|     movd        eax, m4
 | |
|     RET
 | |
| 
 | |
| ;-------------------------------------------------------------------------------
 | |
| ; int ff_pixelutils_sad_[au]_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1,
 | |
| ;                                       const uint8_t *src2, ptrdiff_t stride2);
 | |
| ;-------------------------------------------------------------------------------
 | |
| %macro SAD_XMM_16x16 1
 | |
| INIT_XMM sse2
 | |
| cglobal pixelutils_sad_%1_16x16, 4,4,3, src1, stride1, src2, stride2
 | |
|     mov%1       m2, [src2q]
 | |
|     psadbw      m2, [src1q]
 | |
|     mov%1       m1, [src2q + stride2q]
 | |
|     psadbw      m1, [src1q + stride1q]
 | |
|     paddw       m2, m1
 | |
| %rep 7
 | |
|     lea         src1q, [src1q + 2*stride1q]
 | |
|     lea         src2q, [src2q + 2*stride2q]
 | |
|     mov%1       m0, [src2q]
 | |
|     psadbw      m0, [src1q]
 | |
|     mov%1       m1, [src2q + stride2q]
 | |
|     psadbw      m1, [src1q + stride1q]
 | |
|     paddw       m2, m0
 | |
|     paddw       m2, m1
 | |
| %endrep
 | |
|     movhlps     m0, m2
 | |
|     paddw       m2, m0
 | |
|     movd        eax, m2
 | |
|     RET
 | |
| %endmacro
 | |
| 
 | |
| SAD_XMM_16x16 a
 | |
| SAD_XMM_16x16 u
 | |
| 
 | |
| 
 | |
| %macro PROCESS_SAD_32x4_U 0
 | |
|     movu    m1,  [r2]
 | |
|     movu    m2,  [r2 + 16]
 | |
|     movu    m3,  [r0]
 | |
|     movu    m4,  [r0 + 16]
 | |
|     psadbw  m1,  m3
 | |
|     psadbw  m2,  m4
 | |
|     paddd   m1,  m2
 | |
|     paddd   m0,  m1
 | |
|     lea     r2,  [r2 + r3]
 | |
|     lea     r0,  [r0 + r1]
 | |
| 
 | |
|     movu    m1,  [r2]
 | |
|     movu    m2,  [r2 + 16]
 | |
|     movu    m3,  [r0]
 | |
|     movu    m4,  [r0 + 16]
 | |
|     psadbw  m1,  m3
 | |
|     psadbw  m2,  m4
 | |
|     paddd   m1,  m2
 | |
|     paddd   m0,  m1
 | |
|     lea     r2,  [r2 + r3]
 | |
|     lea     r0,  [r0 + r1]
 | |
| 
 | |
|     movu    m1,  [r2]
 | |
|     movu    m2,  [r2 + 16]
 | |
|     movu    m3,  [r0]
 | |
|     movu    m4,  [r0 + 16]
 | |
|     psadbw  m1,  m3
 | |
|     psadbw  m2,  m4
 | |
|     paddd   m1,  m2
 | |
|     paddd   m0,  m1
 | |
|     lea     r2,  [r2 + r3]
 | |
|     lea     r0,  [r0 + r1]
 | |
| 
 | |
|     movu    m1,  [r2]
 | |
|     movu    m2,  [r2 + 16]
 | |
|     movu    m3,  [r0]
 | |
|     movu    m4,  [r0 + 16]
 | |
|     psadbw  m1,  m3
 | |
|     psadbw  m2,  m4
 | |
|     paddd   m1,  m2
 | |
|     paddd   m0,  m1
 | |
|     lea     r2,  [r2 + r3]
 | |
|     lea     r0,  [r0 + r1]
 | |
| %endmacro
 | |
| 
 | |
| %macro PROCESS_SAD_32x4 1
 | |
|     mov%1   m1,  [r2]
 | |
|     mov%1   m2,  [r2 + 16]
 | |
|     psadbw  m1,  [r0]
 | |
|     psadbw  m2,  [r0 + 16]
 | |
|     paddd   m1,  m2
 | |
|     paddd   m0,  m1
 | |
|     lea     r2,  [r2 + r3]
 | |
|     lea     r0,  [r0 + r1]
 | |
| 
 | |
|     mov%1   m1,  [r2]
 | |
|     mov%1   m2,  [r2 + 16]
 | |
|     psadbw  m1,  [r0]
 | |
|     psadbw  m2,  [r0 + 16]
 | |
|     paddd   m1,  m2
 | |
|     paddd   m0,  m1
 | |
|     lea     r2,  [r2 + r3]
 | |
|     lea     r0,  [r0 + r1]
 | |
| 
 | |
|     mov%1   m1,  [r2]
 | |
|     mov%1   m2,  [r2 + 16]
 | |
|     psadbw  m1,  [r0]
 | |
|     psadbw  m2,  [r0 + 16]
 | |
|     paddd   m1,  m2
 | |
|     paddd   m0,  m1
 | |
|     lea     r2,  [r2 + r3]
 | |
|     lea     r0,  [r0 + r1]
 | |
| 
 | |
|     mov%1   m1,  [r2]
 | |
|     mov%1   m2,  [r2 + 16]
 | |
|     psadbw  m1,  [r0]
 | |
|     psadbw  m2,  [r0 + 16]
 | |
|     paddd   m1,  m2
 | |
|     paddd   m0,  m1
 | |
|     lea     r2,  [r2 + r3]
 | |
|     lea     r0,  [r0 + r1]
 | |
| %endmacro
 | |
| 
 | |
| ;-----------------------------------------------------------------------------
 | |
| ; int ff_pixelutils_sad_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1,
 | |
| ;                                  const uint8_t *src2, ptrdiff_t stride2);
 | |
| ;-----------------------------------------------------------------------------
 | |
| INIT_XMM sse2
 | |
| cglobal pixelutils_sad_32x32, 4,5,5, src1, stride1, src2, stride2
 | |
|     pxor  m0,  m0
 | |
|     mov   r4d, 4
 | |
| .loop:
 | |
|     PROCESS_SAD_32x4_U
 | |
|     PROCESS_SAD_32x4_U
 | |
|     dec r4d
 | |
|     jnz .loop
 | |
| 
 | |
|     movhlps m1,  m0
 | |
|     paddd   m0,  m1
 | |
|     movd    eax, m0
 | |
|     RET
 | |
| 
 | |
| ;-------------------------------------------------------------------------------
 | |
| ; int ff_pixelutils_sad_[au]_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1,
 | |
| ;                                       const uint8_t *src2, ptrdiff_t stride2);
 | |
| ;-------------------------------------------------------------------------------
 | |
| %macro SAD_XMM_32x32 1
 | |
| INIT_XMM sse2
 | |
| cglobal pixelutils_sad_%1_32x32, 4,5,3, src1, stride1, src2, stride2
 | |
|     pxor  m0,  m0
 | |
|     mov   r4d, 4
 | |
| .loop:
 | |
|     PROCESS_SAD_32x4 %1
 | |
|     PROCESS_SAD_32x4 %1
 | |
|     dec r4d
 | |
|     jnz .loop
 | |
| 
 | |
|     movhlps m1,  m0
 | |
|     paddd   m0,  m1
 | |
|     movd    eax, m0
 | |
|     RET
 | |
| %endmacro
 | |
| 
 | |
| SAD_XMM_32x32 a
 | |
| SAD_XMM_32x32 u
 | |
| 
 | |
| %if HAVE_AVX2_EXTERNAL
 | |
| ;-------------------------------------------------------------------------------
 | |
| ; int ff_pixelutils_sad_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1,
 | |
| ;                                  const uint8_t *src2, ptrdiff_t stride2);
 | |
| ;-------------------------------------------------------------------------------
 | |
| INIT_YMM avx2
 | |
| cglobal pixelutils_sad_32x32, 4,7,5, src1, stride1, src2, stride2
 | |
|     pxor            m0, m0
 | |
|     mov             r4d, 32/4
 | |
|     lea             r5, [stride1q * 3]
 | |
|     lea             r6, [stride2q * 3]
 | |
| 
 | |
| .loop:
 | |
|     movu           m1, [src1q]               ; row 0 of pix0
 | |
|     movu           m2, [src2q]               ; row 0 of pix1
 | |
|     movu           m3, [src1q + stride1q]    ; row 1 of pix0
 | |
|     movu           m4, [src2q + stride2q]    ; row 1 of pix1
 | |
| 
 | |
|     psadbw         m1, m2
 | |
|     psadbw         m3, m4
 | |
|     paddd          m0, m1
 | |
|     paddd          m0, m3
 | |
| 
 | |
|     movu           m1, [src1q + 2 * stride1q] ; row 2 of pix0
 | |
|     movu           m2, [src2q + 2 * stride2q] ; row 2 of pix1
 | |
|     movu           m3, [src1q + r5]           ; row 3 of pix0
 | |
|     movu           m4, [src2q + r6]           ; row 3 of pix1
 | |
| 
 | |
|     psadbw         m1, m2
 | |
|     psadbw         m3, m4
 | |
|     paddd          m0, m1
 | |
|     paddd          m0, m3
 | |
| 
 | |
|     lea            src2q,     [src2q + 4 * stride2q]
 | |
|     lea            src1q,     [src1q + 4 * stride1q]
 | |
| 
 | |
|     dec            r4d
 | |
|     jnz           .loop
 | |
| 
 | |
|     vextracti128   xm1, m0, 1
 | |
|     paddd          xm0, xm1
 | |
|     pshufd         xm1, xm0, 2
 | |
|     paddd          xm0, xm1
 | |
|     movd           eax, xm0
 | |
|     RET
 | |
| 
 | |
| ;-------------------------------------------------------------------------------
 | |
| ; int ff_pixelutils_sad_[au]_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1,
 | |
| ;                                       const uint8_t *src2, ptrdiff_t stride2);
 | |
| ;-------------------------------------------------------------------------------
 | |
| %macro SAD_AVX2_32x32 1
 | |
| INIT_YMM avx2
 | |
| cglobal pixelutils_sad_%1_32x32, 4,7,3, src1, stride1, src2, stride2
 | |
|     pxor           m0, m0
 | |
|     mov            r4d, 32/4
 | |
|     lea            r5, [stride1q * 3]
 | |
|     lea            r6, [stride2q * 3]
 | |
| 
 | |
| .loop:
 | |
|     mov%1          m1, [src2q]                ; row 0 of pix1
 | |
|     psadbw         m1, [src1q]
 | |
|     mov%1          m2, [src2q + stride2q]     ; row 1 of pix1
 | |
|     psadbw         m2, [src1q + stride1q]
 | |
| 
 | |
|     paddd          m0, m1
 | |
|     paddd          m0, m2
 | |
| 
 | |
|     mov%1          m1, [src2q + 2 * stride2q] ; row 2 of pix1
 | |
|     psadbw         m1, [src1q + 2 * stride1q]
 | |
|     mov%1          m2, [src2q + r6]           ; row 3 of pix1
 | |
|     psadbw         m2, [src1q + r5]
 | |
| 
 | |
|     paddd          m0, m1
 | |
|     paddd          m0, m2
 | |
| 
 | |
|     lea            src2q,     [src2q + 4 * stride2q]
 | |
|     lea            src1q,     [src1q + 4 * stride1q]
 | |
| 
 | |
|     dec            r4d
 | |
|     jnz           .loop
 | |
| 
 | |
|     vextracti128   xm1, m0, 1
 | |
|     paddd          xm0, xm1
 | |
|     pshufd         xm1, xm0, 2
 | |
|     paddd          xm0, xm1
 | |
|     movd           eax, xm0
 | |
|     RET
 | |
| %endmacro
 | |
| 
 | |
| SAD_AVX2_32x32 a
 | |
| SAD_AVX2_32x32 u
 | |
| %endif
 |