The .text section is already 16-byte aligned by default on all supported platforms so `SECTION_TEXT` isn't any different from `SECTION .text`.
		
			
				
	
	
		
			1262 lines
		
	
	
		
			43 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
			
		
		
	
	
			1262 lines
		
	
	
		
			43 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
| ;******************************************************************************
 | |
| ;* x86 optimized Format Conversion Utils
 | |
| ;* Copyright (c) 2008 Loren Merritt
 | |
| ;* Copyright (c) 2012 Justin Ruggles <justin.ruggles@gmail.com>
 | |
| ;*
 | |
| ;* This file is part of FFmpeg.
 | |
| ;*
 | |
| ;* FFmpeg is free software; you can redistribute it and/or
 | |
| ;* modify it under the terms of the GNU Lesser General Public
 | |
| ;* License as published by the Free Software Foundation; either
 | |
| ;* version 2.1 of the License, or (at your option) any later version.
 | |
| ;*
 | |
| ;* FFmpeg is distributed in the hope that it will be useful,
 | |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
| ;* Lesser General Public License for more details.
 | |
| ;*
 | |
| ;* You should have received a copy of the GNU Lesser General Public
 | |
| ;* License along with FFmpeg; if not, write to the Free Software
 | |
| ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
| ;******************************************************************************
 | |
| 
 | |
| %include "libavutil/x86/x86util.asm"
 | |
| %include "util.asm"
 | |
| 
 | |
| SECTION_RODATA 32
 | |
| 
 | |
| pf_s32_inv_scale: times 8 dd 0x30000000
 | |
| pf_s32_scale:     times 8 dd 0x4f000000
 | |
| pf_s32_clip:      times 8 dd 0x4effffff
 | |
| pf_s16_inv_scale: times 4 dd 0x38000000
 | |
| pf_s16_scale:     times 4 dd 0x47000000
 | |
| pb_shuf_unpack_even:      db -1, -1,  0,  1, -1, -1,  2,  3, -1, -1,  8,  9, -1, -1, 10, 11
 | |
| pb_shuf_unpack_odd:       db -1, -1,  4,  5, -1, -1,  6,  7, -1, -1, 12, 13, -1, -1, 14, 15
 | |
| pb_interleave_words: SHUFFLE_MASK_W  0,  4,  1,  5,  2,  6,  3,  7
 | |
| pb_deinterleave_words: SHUFFLE_MASK_W  0,  2,  4,  6,  1,  3,  5,  7
 | |
| pw_zero_even:     times 4 dw 0x0000, 0xffff
 | |
| 
 | |
| SECTION .text
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16_to_s32(int32_t *dst, const int16_t *src, int len);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| INIT_XMM sse2
 | |
| cglobal conv_s16_to_s32, 3,3,3, dst, src, len
 | |
|     lea      lenq, [2*lend]
 | |
|     lea      dstq, [dstq+2*lenq]
 | |
|     add      srcq, lenq
 | |
|     neg      lenq
 | |
| .loop:
 | |
|     mova       m2, [srcq+lenq]
 | |
|     pxor       m0, m0
 | |
|     pxor       m1, m1
 | |
|     punpcklwd  m0, m2
 | |
|     punpckhwd  m1, m2
 | |
|     mova  [dstq+2*lenq       ], m0
 | |
|     mova  [dstq+2*lenq+mmsize], m1
 | |
|     add      lenq, mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16_to_flt(float *dst, const int16_t *src, int len);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S16_TO_FLT 0
 | |
| cglobal conv_s16_to_flt, 3,3,3, dst, src, len
 | |
|     lea      lenq, [2*lend]
 | |
|     add      srcq, lenq
 | |
|     lea      dstq, [dstq + 2*lenq]
 | |
|     neg      lenq
 | |
|     mova       m2, [pf_s16_inv_scale]
 | |
|     ALIGN 16
 | |
| .loop:
 | |
|     mova       m0, [srcq+lenq]
 | |
|     S16_TO_S32_SX 0, 1
 | |
|     cvtdq2ps   m0, m0
 | |
|     cvtdq2ps   m1, m1
 | |
|     mulps      m0, m2
 | |
|     mulps      m1, m2
 | |
|     mova  [dstq+2*lenq       ], m0
 | |
|     mova  [dstq+2*lenq+mmsize], m1
 | |
|     add      lenq, mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S16_TO_FLT
 | |
| INIT_XMM sse4
 | |
| CONV_S16_TO_FLT
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s32_to_s16(int16_t *dst, const int32_t *src, int len);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S32_TO_S16 0
 | |
| cglobal conv_s32_to_s16, 3,3,4, dst, src, len
 | |
|     lea     lenq, [2*lend]
 | |
|     lea     srcq, [srcq+2*lenq]
 | |
|     add     dstq, lenq
 | |
|     neg     lenq
 | |
| .loop:
 | |
|     mova      m0, [srcq+2*lenq         ]
 | |
|     mova      m1, [srcq+2*lenq+  mmsize]
 | |
|     mova      m2, [srcq+2*lenq+2*mmsize]
 | |
|     mova      m3, [srcq+2*lenq+3*mmsize]
 | |
|     psrad     m0, 16
 | |
|     psrad     m1, 16
 | |
|     psrad     m2, 16
 | |
|     psrad     m3, 16
 | |
|     packssdw  m0, m1
 | |
|     packssdw  m2, m3
 | |
|     mova  [dstq+lenq       ], m0
 | |
|     mova  [dstq+lenq+mmsize], m2
 | |
|     add     lenq, mmsize*2
 | |
|     jl .loop
 | |
| %if mmsize == 8
 | |
|     emms
 | |
|     RET
 | |
| %else
 | |
|     REP_RET
 | |
| %endif
 | |
| %endmacro
 | |
| 
 | |
| INIT_MMX mmx
 | |
| CONV_S32_TO_S16
 | |
| INIT_XMM sse2
 | |
| CONV_S32_TO_S16
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s32_to_flt(float *dst, const int32_t *src, int len);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S32_TO_FLT 0
 | |
| cglobal conv_s32_to_flt, 3,3,3, dst, src, len
 | |
|     lea     lenq, [4*lend]
 | |
|     add     srcq, lenq
 | |
|     add     dstq, lenq
 | |
|     neg     lenq
 | |
|     mova      m0, [pf_s32_inv_scale]
 | |
|     ALIGN 16
 | |
| .loop:
 | |
|     cvtdq2ps  m1, [srcq+lenq       ]
 | |
|     cvtdq2ps  m2, [srcq+lenq+mmsize]
 | |
|     mulps     m1, m1, m0
 | |
|     mulps     m2, m2, m0
 | |
|     mova  [dstq+lenq       ], m1
 | |
|     mova  [dstq+lenq+mmsize], m2
 | |
|     add     lenq, mmsize*2
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S32_TO_FLT
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_YMM avx
 | |
| CONV_S32_TO_FLT
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_flt_to_s16(int16_t *dst, const float *src, int len);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| INIT_XMM sse2
 | |
| cglobal conv_flt_to_s16, 3,3,5, dst, src, len
 | |
|     lea     lenq, [2*lend]
 | |
|     lea     srcq, [srcq+2*lenq]
 | |
|     add     dstq, lenq
 | |
|     neg     lenq
 | |
|     mova      m4, [pf_s16_scale]
 | |
| .loop:
 | |
|     mova      m0, [srcq+2*lenq         ]
 | |
|     mova      m1, [srcq+2*lenq+1*mmsize]
 | |
|     mova      m2, [srcq+2*lenq+2*mmsize]
 | |
|     mova      m3, [srcq+2*lenq+3*mmsize]
 | |
|     mulps     m0, m4
 | |
|     mulps     m1, m4
 | |
|     mulps     m2, m4
 | |
|     mulps     m3, m4
 | |
|     cvtps2dq  m0, m0
 | |
|     cvtps2dq  m1, m1
 | |
|     cvtps2dq  m2, m2
 | |
|     cvtps2dq  m3, m3
 | |
|     packssdw  m0, m1
 | |
|     packssdw  m2, m3
 | |
|     mova  [dstq+lenq       ], m0
 | |
|     mova  [dstq+lenq+mmsize], m2
 | |
|     add     lenq, mmsize*2
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_flt_to_s32(int32_t *dst, const float *src, int len);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_FLT_TO_S32 0
 | |
| cglobal conv_flt_to_s32, 3,3,6, dst, src, len
 | |
|     lea     lenq, [lend*4]
 | |
|     add     srcq, lenq
 | |
|     add     dstq, lenq
 | |
|     neg     lenq
 | |
|     mova      m4, [pf_s32_scale]
 | |
|     mova      m5, [pf_s32_clip]
 | |
| .loop:
 | |
|     mulps     m0, m4, [srcq+lenq         ]
 | |
|     mulps     m1, m4, [srcq+lenq+1*mmsize]
 | |
|     mulps     m2, m4, [srcq+lenq+2*mmsize]
 | |
|     mulps     m3, m4, [srcq+lenq+3*mmsize]
 | |
|     minps     m0, m0, m5
 | |
|     minps     m1, m1, m5
 | |
|     minps     m2, m2, m5
 | |
|     minps     m3, m3, m5
 | |
|     cvtps2dq  m0, m0
 | |
|     cvtps2dq  m1, m1
 | |
|     cvtps2dq  m2, m2
 | |
|     cvtps2dq  m3, m3
 | |
|     mova  [dstq+lenq         ], m0
 | |
|     mova  [dstq+lenq+1*mmsize], m1
 | |
|     mova  [dstq+lenq+2*mmsize], m2
 | |
|     mova  [dstq+lenq+3*mmsize], m3
 | |
|     add     lenq, mmsize*4
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_FLT_TO_S32
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_YMM avx
 | |
| CONV_FLT_TO_S32
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16p_to_s16_2ch(int16_t *dst, int16_t *const *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S16P_TO_S16_2CH 0
 | |
| cglobal conv_s16p_to_s16_2ch, 3,4,5, dst, src0, len, src1
 | |
|     mov       src1q, [src0q+gprsize]
 | |
|     mov       src0q, [src0q        ]
 | |
|     lea        lenq, [2*lend]
 | |
|     add       src0q, lenq
 | |
|     add       src1q, lenq
 | |
|     lea        dstq, [dstq+2*lenq]
 | |
|     neg        lenq
 | |
| .loop:
 | |
|     mova         m0, [src0q+lenq       ]
 | |
|     mova         m1, [src1q+lenq       ]
 | |
|     mova         m2, [src0q+lenq+mmsize]
 | |
|     mova         m3, [src1q+lenq+mmsize]
 | |
|     SBUTTERFLY2  wd, 0, 1, 4
 | |
|     SBUTTERFLY2  wd, 2, 3, 4
 | |
|     mova  [dstq+2*lenq+0*mmsize], m0
 | |
|     mova  [dstq+2*lenq+1*mmsize], m1
 | |
|     mova  [dstq+2*lenq+2*mmsize], m2
 | |
|     mova  [dstq+2*lenq+3*mmsize], m3
 | |
|     add        lenq, 2*mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S16P_TO_S16_2CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_S16P_TO_S16_2CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16p_to_s16_6ch(int16_t *dst, int16_t *const *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; NOTE: In the 6-channel functions, len could be used as an index on x86-64
 | |
| ;       instead of just a counter, which would avoid incrementing the
 | |
| ;       pointers, but the extra complexity and amount of code is not worth
 | |
| ;       the small gain. On x86-32 there are not enough registers to use len
 | |
| ;       as an index without keeping two of the pointers on the stack and
 | |
| ;       loading them in each iteration.
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S16P_TO_S16_6CH 0
 | |
| %if ARCH_X86_64
 | |
| cglobal conv_s16p_to_s16_6ch, 3,8,7, dst, src0, len, src1, src2, src3, src4, src5
 | |
| %else
 | |
| cglobal conv_s16p_to_s16_6ch, 2,7,7, dst, src0, src1, src2, src3, src4, src5
 | |
| %define lend dword r2m
 | |
| %endif
 | |
|     mov      src1q, [src0q+1*gprsize]
 | |
|     mov      src2q, [src0q+2*gprsize]
 | |
|     mov      src3q, [src0q+3*gprsize]
 | |
|     mov      src4q, [src0q+4*gprsize]
 | |
|     mov      src5q, [src0q+5*gprsize]
 | |
|     mov      src0q, [src0q]
 | |
|     sub      src1q, src0q
 | |
|     sub      src2q, src0q
 | |
|     sub      src3q, src0q
 | |
|     sub      src4q, src0q
 | |
|     sub      src5q, src0q
 | |
| .loop:
 | |
| %if cpuflag(sse2slow)
 | |
|     movq        m0, [src0q      ]   ; m0 =  0,  6, 12, 18,  x,  x,  x,  x
 | |
|     movq        m1, [src0q+src1q]   ; m1 =  1,  7, 13, 19,  x,  x,  x,  x
 | |
|     movq        m2, [src0q+src2q]   ; m2 =  2,  8, 14, 20,  x,  x,  x,  x
 | |
|     movq        m3, [src0q+src3q]   ; m3 =  3,  9, 15, 21,  x,  x,  x,  x
 | |
|     movq        m4, [src0q+src4q]   ; m4 =  4, 10, 16, 22,  x,  x,  x,  x
 | |
|     movq        m5, [src0q+src5q]   ; m5 =  5, 11, 17, 23,  x,  x,  x,  x
 | |
|                                     ; unpack words:
 | |
|     punpcklwd   m0, m1              ; m0 =  0,  1,  6,  7, 12, 13, 18, 19
 | |
|     punpcklwd   m2, m3              ; m2 =  4,  5, 10, 11, 16, 17, 22, 23
 | |
|     punpcklwd   m4, m5              ; m4 =  2,  3,  8,  9, 14, 15, 20, 21
 | |
|                                     ; blend dwords
 | |
|     shufps      m1, m0, m2, q2020   ; m1 =  0,  1, 12, 13,  2,  3, 14, 15
 | |
|     shufps      m0, m4, q2031       ; m0 =  6,  7, 18, 19,  4,  5, 16, 17
 | |
|     shufps      m2, m4, q3131       ; m2 =  8,  9, 20, 21, 10, 11, 22, 23
 | |
|                                     ; shuffle dwords
 | |
|     pshufd      m0, m0, q1302       ; m0 =  4,  5,  6,  7, 16, 17, 18, 19
 | |
|     pshufd      m1, m1, q3120       ; m1 =  0,  1,  2,  3, 12, 13, 14, 15
 | |
|     pshufd      m2, m2, q3120       ; m2 =  8,  9, 10, 11, 20, 21, 22, 23
 | |
|     movq   [dstq+0*mmsize/2], m1
 | |
|     movq   [dstq+1*mmsize/2], m0
 | |
|     movq   [dstq+2*mmsize/2], m2
 | |
|     movhps [dstq+3*mmsize/2], m1
 | |
|     movhps [dstq+4*mmsize/2], m0
 | |
|     movhps [dstq+5*mmsize/2], m2
 | |
|     add      src0q, mmsize/2
 | |
|     add       dstq, mmsize*3
 | |
|     sub       lend, mmsize/4
 | |
| %else
 | |
|     mova        m0, [src0q      ]   ; m0 =  0,  6, 12, 18, 24, 30, 36, 42
 | |
|     mova        m1, [src0q+src1q]   ; m1 =  1,  7, 13, 19, 25, 31, 37, 43
 | |
|     mova        m2, [src0q+src2q]   ; m2 =  2,  8, 14, 20, 26, 32, 38, 44
 | |
|     mova        m3, [src0q+src3q]   ; m3 =  3,  9, 15, 21, 27, 33, 39, 45
 | |
|     mova        m4, [src0q+src4q]   ; m4 =  4, 10, 16, 22, 28, 34, 40, 46
 | |
|     mova        m5, [src0q+src5q]   ; m5 =  5, 11, 17, 23, 29, 35, 41, 47
 | |
|                                     ; unpack words:
 | |
|     SBUTTERFLY2 wd, 0, 1, 6         ; m0 =  0,  1,  6,  7, 12, 13, 18, 19
 | |
|                                     ; m1 = 24, 25, 30, 31, 36, 37, 42, 43
 | |
|     SBUTTERFLY2 wd, 2, 3, 6         ; m2 =  2,  3,  8,  9, 14, 15, 20, 21
 | |
|                                     ; m3 = 26, 27, 32, 33, 38, 39, 44, 45
 | |
|     SBUTTERFLY2 wd, 4, 5, 6         ; m4 =  4,  5, 10, 11, 16, 17, 22, 23
 | |
|                                     ; m5 = 28, 29, 34, 35, 40, 41, 46, 47
 | |
|                                     ; blend dwords
 | |
|     shufps      m6, m0, m2, q2020   ; m6 =  0,  1, 12, 13,  2,  3, 14, 15
 | |
|     shufps      m0, m4, q2031       ; m0 =  6,  7, 18, 19,  4,  5, 16, 17
 | |
|     shufps      m2, m4, q3131       ; m2 =  8,  9, 20, 21, 10, 11, 22, 23
 | |
|     SWAP 4,6                        ; m4 =  0,  1, 12, 13,  2,  3, 14, 15
 | |
|     shufps      m6, m1, m3, q2020   ; m6 = 24, 25, 36, 37, 26, 27, 38, 39
 | |
|     shufps      m1, m5, q2031       ; m1 = 30, 31, 42, 43, 28, 29, 40, 41
 | |
|     shufps      m3, m5, q3131       ; m3 = 32, 33, 44, 45, 34, 35, 46, 47
 | |
|     SWAP 5,6                        ; m5 = 24, 25, 36, 37, 26, 27, 38, 39
 | |
|                                     ; shuffle dwords
 | |
|     pshufd      m0, m0, q1302       ; m0 =  4,  5,  6,  7, 16, 17, 18, 19
 | |
|     pshufd      m2, m2, q3120       ; m2 =  8,  9, 10, 11, 20, 21, 22, 23
 | |
|     pshufd      m4, m4, q3120       ; m4 =  0,  1,  2,  3, 12, 13, 14, 15
 | |
|     pshufd      m1, m1, q1302       ; m1 = 28, 29, 30, 31, 40, 41, 42, 43
 | |
|     pshufd      m3, m3, q3120       ; m3 = 32, 33, 34, 35, 44, 45, 46, 47
 | |
|     pshufd      m5, m5, q3120       ; m5 = 24, 25, 26, 27, 36, 37, 38, 39
 | |
|                                     ; shuffle qwords
 | |
|     punpcklqdq  m6, m4, m0          ; m6 =  0,  1,  2,  3,  4,  5,  6,  7
 | |
|     punpckhqdq  m0, m2              ; m0 = 16, 17, 18, 19, 20, 21, 22, 23
 | |
|     shufps      m2, m4, q3210       ; m2 =  8,  9, 10, 11, 12, 13, 14, 15
 | |
|     SWAP 4,6                        ; m4 =  0,  1,  2,  3,  4,  5,  6,  7
 | |
|     punpcklqdq  m6, m5, m1          ; m6 = 24, 25, 26, 27, 28, 29, 30, 31
 | |
|     punpckhqdq  m1, m3              ; m1 = 40, 41, 42, 43, 44, 45, 46, 47
 | |
|     shufps      m3, m5, q3210       ; m3 = 32, 33, 34, 35, 36, 37, 38, 39
 | |
|     SWAP 5,6                        ; m5 = 24, 25, 26, 27, 28, 29, 30, 31
 | |
|     mova   [dstq+0*mmsize], m4
 | |
|     mova   [dstq+1*mmsize], m2
 | |
|     mova   [dstq+2*mmsize], m0
 | |
|     mova   [dstq+3*mmsize], m5
 | |
|     mova   [dstq+4*mmsize], m3
 | |
|     mova   [dstq+5*mmsize], m1
 | |
|     add      src0q, mmsize
 | |
|     add       dstq, mmsize*6
 | |
|     sub       lend, mmsize/2
 | |
| %endif
 | |
|     jg .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S16P_TO_S16_6CH
 | |
| INIT_XMM sse2slow
 | |
| CONV_S16P_TO_S16_6CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_S16P_TO_S16_6CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16p_to_flt_2ch(float *dst, int16_t *const *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S16P_TO_FLT_2CH 0
 | |
| cglobal conv_s16p_to_flt_2ch, 3,4,6, dst, src0, len, src1
 | |
|     lea       lenq, [2*lend]
 | |
|     mov      src1q, [src0q+gprsize]
 | |
|     mov      src0q, [src0q        ]
 | |
|     lea       dstq, [dstq+4*lenq]
 | |
|     add      src0q, lenq
 | |
|     add      src1q, lenq
 | |
|     neg       lenq
 | |
|     mova        m5, [pf_s32_inv_scale]
 | |
| .loop:
 | |
|     mova        m2, [src0q+lenq]    ; m2 =  0,  2,  4,  6,  8, 10, 12, 14
 | |
|     mova        m4, [src1q+lenq]    ; m4 =  1,  3,  5,  7,  9, 11, 13, 15
 | |
|     SBUTTERFLY2 wd, 2, 4, 3         ; m2 =  0,  1,  2,  3,  4,  5,  6,  7
 | |
|                                     ; m4 =  8,  9, 10, 11, 12, 13, 14, 15
 | |
|     pxor        m3, m3
 | |
|     punpcklwd   m0, m3, m2          ; m0 =      0,      1,      2,      3
 | |
|     punpckhwd   m1, m3, m2          ; m1 =      4,      5,      6,      7
 | |
|     punpcklwd   m2, m3, m4          ; m2 =      8,      9,     10,     11
 | |
|     punpckhwd   m3, m4              ; m3 =     12,     13,     14,     15
 | |
|     cvtdq2ps    m0, m0
 | |
|     cvtdq2ps    m1, m1
 | |
|     cvtdq2ps    m2, m2
 | |
|     cvtdq2ps    m3, m3
 | |
|     mulps       m0, m5
 | |
|     mulps       m1, m5
 | |
|     mulps       m2, m5
 | |
|     mulps       m3, m5
 | |
|     mova  [dstq+4*lenq         ], m0
 | |
|     mova  [dstq+4*lenq+  mmsize], m1
 | |
|     mova  [dstq+4*lenq+2*mmsize], m2
 | |
|     mova  [dstq+4*lenq+3*mmsize], m3
 | |
|     add       lenq, mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S16P_TO_FLT_2CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_S16P_TO_FLT_2CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16p_to_flt_6ch(float *dst, int16_t *const *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S16P_TO_FLT_6CH 0
 | |
| %if ARCH_X86_64
 | |
| cglobal conv_s16p_to_flt_6ch, 3,8,8, dst, src, len, src1, src2, src3, src4, src5
 | |
| %else
 | |
| cglobal conv_s16p_to_flt_6ch, 2,7,8, dst, src, src1, src2, src3, src4, src5
 | |
| %define lend dword r2m
 | |
| %endif
 | |
|     mov     src1q, [srcq+1*gprsize]
 | |
|     mov     src2q, [srcq+2*gprsize]
 | |
|     mov     src3q, [srcq+3*gprsize]
 | |
|     mov     src4q, [srcq+4*gprsize]
 | |
|     mov     src5q, [srcq+5*gprsize]
 | |
|     mov      srcq, [srcq]
 | |
|     sub     src1q, srcq
 | |
|     sub     src2q, srcq
 | |
|     sub     src3q, srcq
 | |
|     sub     src4q, srcq
 | |
|     sub     src5q, srcq
 | |
|     mova       m7, [pf_s32_inv_scale]
 | |
| %if cpuflag(ssse3)
 | |
|     %define unpack_even m6
 | |
|     mova       m6, [pb_shuf_unpack_even]
 | |
| %if ARCH_X86_64
 | |
|     %define unpack_odd m8
 | |
|     mova       m8, [pb_shuf_unpack_odd]
 | |
| %else
 | |
|     %define unpack_odd [pb_shuf_unpack_odd]
 | |
| %endif
 | |
| %endif
 | |
| .loop:
 | |
|     movq       m0, [srcq      ]  ; m0 =  0,  6, 12, 18,  x,  x,  x,  x
 | |
|     movq       m1, [srcq+src1q]  ; m1 =  1,  7, 13, 19,  x,  x,  x,  x
 | |
|     movq       m2, [srcq+src2q]  ; m2 =  2,  8, 14, 20,  x,  x,  x,  x
 | |
|     movq       m3, [srcq+src3q]  ; m3 =  3,  9, 15, 21,  x,  x,  x,  x
 | |
|     movq       m4, [srcq+src4q]  ; m4 =  4, 10, 16, 22,  x,  x,  x,  x
 | |
|     movq       m5, [srcq+src5q]  ; m5 =  5, 11, 17, 23,  x,  x,  x,  x
 | |
|                                  ; unpack words:
 | |
|     punpcklwd  m0, m1            ; m0 =  0,  1,  6,  7, 12, 13, 18, 19
 | |
|     punpcklwd  m2, m3            ; m2 =  2,  3,  8,  9, 14, 15, 20, 21
 | |
|     punpcklwd  m4, m5            ; m4 =  4,  5, 10, 11, 16, 17, 22, 23
 | |
|                                  ; blend dwords
 | |
|     shufps     m1, m4, m0, q3120 ; m1 =  4,  5, 16, 17,  6,  7, 18, 19
 | |
|     shufps         m0, m2, q2020 ; m0 =  0,  1, 12, 13,  2,  3, 14, 15
 | |
|     shufps         m2, m4, q3131 ; m2 =  8,  9, 20, 21, 10, 11, 22, 23
 | |
| %if cpuflag(ssse3)
 | |
|     pshufb     m3, m0, unpack_odd   ; m3 =  12,     13,     14,     15
 | |
|     pshufb         m0, unpack_even  ; m0 =   0,      1,      2,      3
 | |
|     pshufb     m4, m1, unpack_odd   ; m4 =  16,     17,     18,     19
 | |
|     pshufb         m1, unpack_even  ; m1 =   4,      5,      6,      7
 | |
|     pshufb     m5, m2, unpack_odd   ; m5 =  20,     21,     22,     23
 | |
|     pshufb         m2, unpack_even  ; m2 =   8,      9,     10,     11
 | |
| %else
 | |
|                                  ; shuffle dwords
 | |
|     pshufd     m0, m0, q3120     ; m0 =  0,  1,  2,  3, 12, 13, 14, 15
 | |
|     pshufd     m1, m1, q3120     ; m1 =  4,  5,  6,  7, 16, 17, 18, 19
 | |
|     pshufd     m2, m2, q3120     ; m2 =  8,  9, 10, 11, 20, 21, 22, 23
 | |
|     pxor       m6, m6            ; convert s16 in m0-m2 to s32 in m0-m5
 | |
|     punpcklwd  m3, m6, m0        ; m3 =      0,      1,      2,      3
 | |
|     punpckhwd  m4, m6, m0        ; m4 =     12,     13,     14,     15
 | |
|     punpcklwd  m0, m6, m1        ; m0 =      4,      5,      6,      7
 | |
|     punpckhwd  m5, m6, m1        ; m5 =     16,     17,     18,     19
 | |
|     punpcklwd  m1, m6, m2        ; m1 =      8,      9,     10,     11
 | |
|     punpckhwd      m6, m2        ; m6 =     20,     21,     22,     23
 | |
|     SWAP 6,2,1,0,3,4,5           ; swap registers 3,0,1,4,5,6 to 0,1,2,3,4,5
 | |
| %endif
 | |
|     cvtdq2ps   m0, m0            ; convert s32 to float
 | |
|     cvtdq2ps   m1, m1
 | |
|     cvtdq2ps   m2, m2
 | |
|     cvtdq2ps   m3, m3
 | |
|     cvtdq2ps   m4, m4
 | |
|     cvtdq2ps   m5, m5
 | |
|     mulps      m0, m7            ; scale float from s32 range to [-1.0,1.0]
 | |
|     mulps      m1, m7
 | |
|     mulps      m2, m7
 | |
|     mulps      m3, m7
 | |
|     mulps      m4, m7
 | |
|     mulps      m5, m7
 | |
|     mova  [dstq         ], m0
 | |
|     mova  [dstq+  mmsize], m1
 | |
|     mova  [dstq+2*mmsize], m2
 | |
|     mova  [dstq+3*mmsize], m3
 | |
|     mova  [dstq+4*mmsize], m4
 | |
|     mova  [dstq+5*mmsize], m5
 | |
|     add      srcq, mmsize/2
 | |
|     add      dstq, mmsize*6
 | |
|     sub      lend, mmsize/4
 | |
|     jg .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S16P_TO_FLT_6CH
 | |
| INIT_XMM ssse3
 | |
| CONV_S16P_TO_FLT_6CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_S16P_TO_FLT_6CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_fltp_to_s16_2ch(int16_t *dst, float *const *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_FLTP_TO_S16_2CH 0
 | |
| cglobal conv_fltp_to_s16_2ch, 3,4,3, dst, src0, len, src1
 | |
|     lea      lenq, [4*lend]
 | |
|     mov     src1q, [src0q+gprsize]
 | |
|     mov     src0q, [src0q        ]
 | |
|     add      dstq, lenq
 | |
|     add     src0q, lenq
 | |
|     add     src1q, lenq
 | |
|     neg      lenq
 | |
|     mova       m2, [pf_s16_scale]
 | |
| %if cpuflag(ssse3)
 | |
|     mova       m3, [pb_interleave_words]
 | |
| %endif
 | |
| .loop:
 | |
|     mulps      m0, m2, [src0q+lenq] ; m0 =    0,    2,    4,    6
 | |
|     mulps      m1, m2, [src1q+lenq] ; m1 =    1,    3,    5,    7
 | |
|     cvtps2dq   m0, m0
 | |
|     cvtps2dq   m1, m1
 | |
| %if cpuflag(ssse3)
 | |
|     packssdw   m0, m1               ; m0 = 0, 2, 4, 6, 1, 3, 5, 7
 | |
|     pshufb     m0, m3               ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
 | |
| %else
 | |
|     packssdw   m0, m0               ; m0 = 0, 2, 4, 6, x, x, x, x
 | |
|     packssdw   m1, m1               ; m1 = 1, 3, 5, 7, x, x, x, x
 | |
|     punpcklwd  m0, m1               ; m0 = 0, 1, 2, 3, 4, 5, 6, 7
 | |
| %endif
 | |
|     mova  [dstq+lenq], m0
 | |
|     add      lenq, mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_FLTP_TO_S16_2CH
 | |
| INIT_XMM ssse3
 | |
| CONV_FLTP_TO_S16_2CH
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_fltp_to_s16_6ch(int16_t *dst, float *const *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_FLTP_TO_S16_6CH 0
 | |
| %if ARCH_X86_64
 | |
| cglobal conv_fltp_to_s16_6ch, 3,8,7, dst, src, len, src1, src2, src3, src4, src5
 | |
| %else
 | |
| cglobal conv_fltp_to_s16_6ch, 2,7,7, dst, src, src1, src2, src3, src4, src5
 | |
| %define lend dword r2m
 | |
| %endif
 | |
|     mov        src1q, [srcq+1*gprsize]
 | |
|     mov        src2q, [srcq+2*gprsize]
 | |
|     mov        src3q, [srcq+3*gprsize]
 | |
|     mov        src4q, [srcq+4*gprsize]
 | |
|     mov        src5q, [srcq+5*gprsize]
 | |
|     mov         srcq, [srcq]
 | |
|     sub        src1q, srcq
 | |
|     sub        src2q, srcq
 | |
|     sub        src3q, srcq
 | |
|     sub        src4q, srcq
 | |
|     sub        src5q, srcq
 | |
|     movaps      xmm6, [pf_s16_scale]
 | |
| .loop:
 | |
| %if cpuflag(sse2)
 | |
|     mulps         m0, m6, [srcq      ]
 | |
|     mulps         m1, m6, [srcq+src1q]
 | |
|     mulps         m2, m6, [srcq+src2q]
 | |
|     mulps         m3, m6, [srcq+src3q]
 | |
|     mulps         m4, m6, [srcq+src4q]
 | |
|     mulps         m5, m6, [srcq+src5q]
 | |
|     cvtps2dq      m0, m0
 | |
|     cvtps2dq      m1, m1
 | |
|     cvtps2dq      m2, m2
 | |
|     cvtps2dq      m3, m3
 | |
|     cvtps2dq      m4, m4
 | |
|     cvtps2dq      m5, m5
 | |
|     packssdw      m0, m3            ; m0 =  0,  6, 12, 18,  3,  9, 15, 21
 | |
|     packssdw      m1, m4            ; m1 =  1,  7, 13, 19,  4, 10, 16, 22
 | |
|     packssdw      m2, m5            ; m2 =  2,  8, 14, 20,  5, 11, 17, 23
 | |
|                                     ; unpack words:
 | |
|     movhlps       m3, m0            ; m3 =  3,  9, 15, 21,  x,  x,  x,  x
 | |
|     punpcklwd     m0, m1            ; m0 =  0,  1,  6,  7, 12, 13, 18, 19
 | |
|     punpckhwd     m1, m2            ; m1 =  4,  5, 10, 11, 16, 17, 22, 23
 | |
|     punpcklwd     m2, m3            ; m2 =  2,  3,  8,  9, 14, 15, 20, 21
 | |
|                                     ; blend dwords:
 | |
|     shufps        m3, m0, m2, q2020 ; m3 =  0,  1, 12, 13,  2,  3, 14, 15
 | |
|     shufps        m0, m1, q2031     ; m0 =  6,  7, 18, 19,  4,  5, 16, 17
 | |
|     shufps        m2, m1, q3131     ; m2 =  8,  9, 20, 21, 10, 11, 22, 23
 | |
|                                     ; shuffle dwords:
 | |
|     shufps        m1, m2, m3, q3120 ; m1 =  8,  9, 10, 11, 12, 13, 14, 15
 | |
|     shufps        m3, m0,     q0220 ; m3 =  0,  1,  2,  3,  4,  5,  6,  7
 | |
|     shufps        m0, m2,     q3113 ; m0 = 16, 17, 18, 19, 20, 21, 22, 23
 | |
|     mova  [dstq+0*mmsize], m3
 | |
|     mova  [dstq+1*mmsize], m1
 | |
|     mova  [dstq+2*mmsize], m0
 | |
| %else ; sse
 | |
|     movlps      xmm0, [srcq      ]
 | |
|     movlps      xmm1, [srcq+src1q]
 | |
|     movlps      xmm2, [srcq+src2q]
 | |
|     movlps      xmm3, [srcq+src3q]
 | |
|     movlps      xmm4, [srcq+src4q]
 | |
|     movlps      xmm5, [srcq+src5q]
 | |
|     mulps       xmm0, xmm6
 | |
|     mulps       xmm1, xmm6
 | |
|     mulps       xmm2, xmm6
 | |
|     mulps       xmm3, xmm6
 | |
|     mulps       xmm4, xmm6
 | |
|     mulps       xmm5, xmm6
 | |
|     cvtps2pi     mm0, xmm0
 | |
|     cvtps2pi     mm1, xmm1
 | |
|     cvtps2pi     mm2, xmm2
 | |
|     cvtps2pi     mm3, xmm3
 | |
|     cvtps2pi     mm4, xmm4
 | |
|     cvtps2pi     mm5, xmm5
 | |
|     packssdw     mm0, mm3           ; m0 =  0,  6,  3,  9
 | |
|     packssdw     mm1, mm4           ; m1 =  1,  7,  4, 10
 | |
|     packssdw     mm2, mm5           ; m2 =  2,  8,  5, 11
 | |
|                                     ; unpack words
 | |
|     pshufw       mm3, mm0, q1032    ; m3 =  3,  9,  0,  6
 | |
|     punpcklwd    mm0, mm1           ; m0 =  0,  1,  6,  7
 | |
|     punpckhwd    mm1, mm2           ; m1 =  4,  5, 10, 11
 | |
|     punpcklwd    mm2, mm3           ; m2 =  2,  3,  8,  9
 | |
|                                     ; unpack dwords
 | |
|     pshufw       mm3, mm0, q1032    ; m3 =  6,  7,  0,  1
 | |
|     punpckldq    mm0, mm2           ; m0 =  0,  1,  2,  3 (final)
 | |
|     punpckhdq    mm2, mm1           ; m2 =  8,  9, 10, 11 (final)
 | |
|     punpckldq    mm1, mm3           ; m1 =  4,  5,  6,  7 (final)
 | |
|     mova  [dstq+0*mmsize], mm0
 | |
|     mova  [dstq+1*mmsize], mm1
 | |
|     mova  [dstq+2*mmsize], mm2
 | |
| %endif
 | |
|     add       srcq, mmsize
 | |
|     add       dstq, mmsize*3
 | |
|     sub       lend, mmsize/4
 | |
|     jg .loop
 | |
| %if mmsize == 8
 | |
|     emms
 | |
|     RET
 | |
| %else
 | |
|     REP_RET
 | |
| %endif
 | |
| %endmacro
 | |
| 
 | |
| INIT_MMX sse
 | |
| CONV_FLTP_TO_S16_6CH
 | |
| INIT_XMM sse2
 | |
| CONV_FLTP_TO_S16_6CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_FLTP_TO_S16_6CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_fltp_to_flt_2ch(float *dst, float *const *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_FLTP_TO_FLT_2CH 0
 | |
| cglobal conv_fltp_to_flt_2ch, 3,4,5, dst, src0, len, src1
 | |
|     mov  src1q, [src0q+gprsize]
 | |
|     mov  src0q, [src0q]
 | |
|     lea   lenq, [4*lend]
 | |
|     add  src0q, lenq
 | |
|     add  src1q, lenq
 | |
|     lea   dstq, [dstq+2*lenq]
 | |
|     neg   lenq
 | |
| .loop:
 | |
|     mova    m0, [src0q+lenq       ]
 | |
|     mova    m1, [src1q+lenq       ]
 | |
|     mova    m2, [src0q+lenq+mmsize]
 | |
|     mova    m3, [src1q+lenq+mmsize]
 | |
|     SBUTTERFLYPS 0, 1, 4
 | |
|     SBUTTERFLYPS 2, 3, 4
 | |
|     mova  [dstq+2*lenq+0*mmsize], m0
 | |
|     mova  [dstq+2*lenq+1*mmsize], m1
 | |
|     mova  [dstq+2*lenq+2*mmsize], m2
 | |
|     mova  [dstq+2*lenq+3*mmsize], m3
 | |
|     add   lenq, 2*mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse
 | |
| CONV_FLTP_TO_FLT_2CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_FLTP_TO_FLT_2CH
 | |
| %endif
 | |
| 
 | |
| ;-----------------------------------------------------------------------------
 | |
| ; void ff_conv_fltp_to_flt_6ch(float *dst, float *const *src, int len,
 | |
| ;                              int channels);
 | |
| ;-----------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_FLTP_TO_FLT_6CH 0
 | |
| cglobal conv_fltp_to_flt_6ch, 2,8,7, dst, src, src1, src2, src3, src4, src5, len
 | |
| %if ARCH_X86_64
 | |
|     mov     lend, r2d
 | |
| %else
 | |
|     %define lend dword r2m
 | |
| %endif
 | |
|     mov    src1q, [srcq+1*gprsize]
 | |
|     mov    src2q, [srcq+2*gprsize]
 | |
|     mov    src3q, [srcq+3*gprsize]
 | |
|     mov    src4q, [srcq+4*gprsize]
 | |
|     mov    src5q, [srcq+5*gprsize]
 | |
|     mov     srcq, [srcq]
 | |
|     sub    src1q, srcq
 | |
|     sub    src2q, srcq
 | |
|     sub    src3q, srcq
 | |
|     sub    src4q, srcq
 | |
|     sub    src5q, srcq
 | |
| .loop:
 | |
|     mova      m0, [srcq      ]
 | |
|     mova      m1, [srcq+src1q]
 | |
|     mova      m2, [srcq+src2q]
 | |
|     mova      m3, [srcq+src3q]
 | |
|     mova      m4, [srcq+src4q]
 | |
|     mova      m5, [srcq+src5q]
 | |
| %if cpuflag(sse4)
 | |
|     SBUTTERFLYPS 0, 1, 6
 | |
|     SBUTTERFLYPS 2, 3, 6
 | |
|     SBUTTERFLYPS 4, 5, 6
 | |
| 
 | |
|     blendps   m6, m4, m0, 1100b
 | |
|     movlhps   m0, m2
 | |
|     movhlps   m4, m2
 | |
|     blendps   m2, m5, m1, 1100b
 | |
|     movlhps   m1, m3
 | |
|     movhlps   m5, m3
 | |
| 
 | |
|     movaps [dstq   ], m0
 | |
|     movaps [dstq+16], m6
 | |
|     movaps [dstq+32], m4
 | |
|     movaps [dstq+48], m1
 | |
|     movaps [dstq+64], m2
 | |
|     movaps [dstq+80], m5
 | |
| %else ; mmx
 | |
|     SBUTTERFLY dq, 0, 1, 6
 | |
|     SBUTTERFLY dq, 2, 3, 6
 | |
|     SBUTTERFLY dq, 4, 5, 6
 | |
| 
 | |
|     movq   [dstq   ], m0
 | |
|     movq   [dstq+ 8], m2
 | |
|     movq   [dstq+16], m4
 | |
|     movq   [dstq+24], m1
 | |
|     movq   [dstq+32], m3
 | |
|     movq   [dstq+40], m5
 | |
| %endif
 | |
|     add      srcq, mmsize
 | |
|     add      dstq, mmsize*6
 | |
|     sub      lend, mmsize/4
 | |
|     jg .loop
 | |
| %if mmsize == 8
 | |
|     emms
 | |
|     RET
 | |
| %else
 | |
|     REP_RET
 | |
| %endif
 | |
| %endmacro
 | |
| 
 | |
| INIT_MMX mmx
 | |
| CONV_FLTP_TO_FLT_6CH
 | |
| INIT_XMM sse4
 | |
| CONV_FLTP_TO_FLT_6CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_FLTP_TO_FLT_6CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16_to_s16p_2ch(int16_t *const *dst, int16_t *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S16_TO_S16P_2CH 0
 | |
| cglobal conv_s16_to_s16p_2ch, 3,4,4, dst0, src, len, dst1
 | |
|     lea       lenq, [2*lend]
 | |
|     mov      dst1q, [dst0q+gprsize]
 | |
|     mov      dst0q, [dst0q        ]
 | |
|     lea       srcq, [srcq+2*lenq]
 | |
|     add      dst0q, lenq
 | |
|     add      dst1q, lenq
 | |
|     neg       lenq
 | |
| %if cpuflag(ssse3)
 | |
|     mova        m3, [pb_deinterleave_words]
 | |
| %endif
 | |
| .loop:
 | |
|     mova        m0, [srcq+2*lenq       ]  ; m0 =  0,  1,  2,  3,  4,  5,  6,  7
 | |
|     mova        m1, [srcq+2*lenq+mmsize]  ; m1 =  8,  9, 10, 11, 12, 13, 14, 15
 | |
| %if cpuflag(ssse3)
 | |
|     pshufb      m0, m3                    ; m0 =  0,  2,  4,  6,  1,  3,  5,  7
 | |
|     pshufb      m1, m3                    ; m1 =  8, 10, 12, 14,  9, 11, 13, 15
 | |
|     SBUTTERFLY2 qdq, 0, 1, 2              ; m0 =  0,  2,  4,  6,  8, 10, 12, 14
 | |
|                                           ; m1 =  1,  3,  5,  7,  9, 11, 13, 15
 | |
| %else ; sse2
 | |
|     pshuflw     m0, m0, q3120             ; m0 =  0,  2,  1,  3,  4,  5,  6,  7
 | |
|     pshufhw     m0, m0, q3120             ; m0 =  0,  2,  1,  3,  4,  6,  5,  7
 | |
|     pshuflw     m1, m1, q3120             ; m1 =  8, 10,  9, 11, 12, 13, 14, 15
 | |
|     pshufhw     m1, m1, q3120             ; m1 =  8, 10,  9, 11, 12, 14, 13, 15
 | |
|     DEINT2_PS    0, 1, 2                  ; m0 =  0,  2,  4,  6,  8, 10, 12, 14
 | |
|                                           ; m1 =  1,  3,  5,  7,  9, 11, 13, 15
 | |
| %endif
 | |
|     mova  [dst0q+lenq], m0
 | |
|     mova  [dst1q+lenq], m1
 | |
|     add       lenq, mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S16_TO_S16P_2CH
 | |
| INIT_XMM ssse3
 | |
| CONV_S16_TO_S16P_2CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_S16_TO_S16P_2CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16_to_s16p_6ch(int16_t *const *dst, int16_t *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S16_TO_S16P_6CH 0
 | |
| %if ARCH_X86_64
 | |
| cglobal conv_s16_to_s16p_6ch, 3,8,5, dst, src, len, dst1, dst2, dst3, dst4, dst5
 | |
| %else
 | |
| cglobal conv_s16_to_s16p_6ch, 2,7,5, dst, src, dst1, dst2, dst3, dst4, dst5
 | |
| %define lend dword r2m
 | |
| %endif
 | |
|     mov     dst1q, [dstq+  gprsize]
 | |
|     mov     dst2q, [dstq+2*gprsize]
 | |
|     mov     dst3q, [dstq+3*gprsize]
 | |
|     mov     dst4q, [dstq+4*gprsize]
 | |
|     mov     dst5q, [dstq+5*gprsize]
 | |
|     mov      dstq, [dstq          ]
 | |
|     sub     dst1q, dstq
 | |
|     sub     dst2q, dstq
 | |
|     sub     dst3q, dstq
 | |
|     sub     dst4q, dstq
 | |
|     sub     dst5q, dstq
 | |
| .loop:
 | |
|     mova       m0, [srcq+0*mmsize]      ; m0 =  0,  1,  2,  3,  4,  5,  6,  7
 | |
|     mova       m3, [srcq+1*mmsize]      ; m3 =  8,  9, 10, 11, 12, 13, 14, 15
 | |
|     mova       m2, [srcq+2*mmsize]      ; m2 = 16, 17, 18, 19, 20, 21, 22, 23
 | |
|     PALIGNR    m1, m3, m0, 12, m4       ; m1 =  6,  7,  8,  9, 10, 11,  x,  x
 | |
|     shufps     m3, m2, q1032            ; m3 = 12, 13, 14, 15, 16, 17, 18, 19
 | |
|     psrldq     m2, 4                    ; m2 = 18, 19, 20, 21, 22, 23,  x,  x
 | |
|     SBUTTERFLY2 wd, 0, 1, 4             ; m0 =  0,  6,  1,  7,  2,  8,  3,  9
 | |
|                                         ; m1 =  4, 10,  5, 11,  x,  x,  x,  x
 | |
|     SBUTTERFLY2 wd, 3, 2, 4             ; m3 = 12, 18, 13, 19, 14, 20, 15, 21
 | |
|                                         ; m2 = 16, 22, 17, 23,  x,  x,  x,  x
 | |
|     SBUTTERFLY2 dq, 0, 3, 4             ; m0 =  0,  6, 12, 18,  1,  7, 13, 19
 | |
|                                         ; m3 =  2,  8, 14, 20,  3,  9, 15, 21
 | |
|     punpckldq  m1, m2                   ; m1 =  4, 10, 16, 22,  5, 11, 17, 23
 | |
|     movq    [dstq      ], m0
 | |
|     movhps  [dstq+dst1q], m0
 | |
|     movq    [dstq+dst2q], m3
 | |
|     movhps  [dstq+dst3q], m3
 | |
|     movq    [dstq+dst4q], m1
 | |
|     movhps  [dstq+dst5q], m1
 | |
|     add      srcq, mmsize*3
 | |
|     add      dstq, mmsize/2
 | |
|     sub      lend, mmsize/4
 | |
|     jg .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S16_TO_S16P_6CH
 | |
| INIT_XMM ssse3
 | |
| CONV_S16_TO_S16P_6CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_S16_TO_S16P_6CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16_to_fltp_2ch(float *const *dst, int16_t *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S16_TO_FLTP_2CH 0
 | |
| cglobal conv_s16_to_fltp_2ch, 3,4,5, dst0, src, len, dst1
 | |
|     lea       lenq, [4*lend]
 | |
|     mov      dst1q, [dst0q+gprsize]
 | |
|     mov      dst0q, [dst0q        ]
 | |
|     add       srcq, lenq
 | |
|     add      dst0q, lenq
 | |
|     add      dst1q, lenq
 | |
|     neg       lenq
 | |
|     mova        m3, [pf_s32_inv_scale]
 | |
|     mova        m4, [pw_zero_even]
 | |
| .loop:
 | |
|     mova        m1, [srcq+lenq]
 | |
|     pslld       m0, m1, 16
 | |
|     pand        m1, m4
 | |
|     cvtdq2ps    m0, m0
 | |
|     cvtdq2ps    m1, m1
 | |
|     mulps       m0, m0, m3
 | |
|     mulps       m1, m1, m3
 | |
|     mova  [dst0q+lenq], m0
 | |
|     mova  [dst1q+lenq], m1
 | |
|     add       lenq, mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S16_TO_FLTP_2CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_S16_TO_FLTP_2CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_s16_to_fltp_6ch(float *const *dst, int16_t *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_S16_TO_FLTP_6CH 0
 | |
| %if ARCH_X86_64
 | |
| cglobal conv_s16_to_fltp_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5
 | |
| %else
 | |
| cglobal conv_s16_to_fltp_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5
 | |
| %define lend dword r2m
 | |
| %endif
 | |
|     mov     dst1q, [dstq+  gprsize]
 | |
|     mov     dst2q, [dstq+2*gprsize]
 | |
|     mov     dst3q, [dstq+3*gprsize]
 | |
|     mov     dst4q, [dstq+4*gprsize]
 | |
|     mov     dst5q, [dstq+5*gprsize]
 | |
|     mov      dstq, [dstq          ]
 | |
|     sub     dst1q, dstq
 | |
|     sub     dst2q, dstq
 | |
|     sub     dst3q, dstq
 | |
|     sub     dst4q, dstq
 | |
|     sub     dst5q, dstq
 | |
|     mova       m6, [pf_s16_inv_scale]
 | |
| .loop:
 | |
|     mova       m0, [srcq+0*mmsize]  ; m0 =  0,  1,  2,  3,  4,  5,  6,  7
 | |
|     mova       m3, [srcq+1*mmsize]  ; m3 =  8,  9, 10, 11, 12, 13, 14, 15
 | |
|     mova       m2, [srcq+2*mmsize]  ; m2 = 16, 17, 18, 19, 20, 21, 22, 23
 | |
|     PALIGNR    m1, m3, m0, 12, m4   ; m1 =  6,  7,  8,  9, 10, 11,  x,  x
 | |
|     shufps     m3, m2, q1032        ; m3 = 12, 13, 14, 15, 16, 17, 18, 19
 | |
|     psrldq     m2, 4                ; m2 = 18, 19, 20, 21, 22, 23,  x,  x
 | |
|     SBUTTERFLY2 wd, 0, 1, 4         ; m0 =  0,  6,  1,  7,  2,  8,  3,  9
 | |
|                                     ; m1 =  4, 10,  5, 11,  x,  x,  x,  x
 | |
|     SBUTTERFLY2 wd, 3, 2, 4         ; m3 = 12, 18, 13, 19, 14, 20, 15, 21
 | |
|                                     ; m2 = 16, 22, 17, 23,  x,  x,  x,  x
 | |
|     SBUTTERFLY2 dq, 0, 3, 4         ; m0 =  0,  6, 12, 18,  1,  7, 13, 19
 | |
|                                     ; m3 =  2,  8, 14, 20,  3,  9, 15, 21
 | |
|     punpckldq  m1, m2               ; m1 =  4, 10, 16, 22,  5, 11, 17, 23
 | |
|     S16_TO_S32_SX 0, 2              ; m0 =      0,      6,     12,     18
 | |
|                                     ; m2 =      1,      7,     13,     19
 | |
|     S16_TO_S32_SX 3, 4              ; m3 =      2,      8,     14,     20
 | |
|                                     ; m4 =      3,      9,     15,     21
 | |
|     S16_TO_S32_SX 1, 5              ; m1 =      4,     10,     16,     22
 | |
|                                     ; m5 =      5,     11,     17,     23
 | |
|     SWAP 1,2,3,4
 | |
|     cvtdq2ps   m0, m0
 | |
|     cvtdq2ps   m1, m1
 | |
|     cvtdq2ps   m2, m2
 | |
|     cvtdq2ps   m3, m3
 | |
|     cvtdq2ps   m4, m4
 | |
|     cvtdq2ps   m5, m5
 | |
|     mulps      m0, m6
 | |
|     mulps      m1, m6
 | |
|     mulps      m2, m6
 | |
|     mulps      m3, m6
 | |
|     mulps      m4, m6
 | |
|     mulps      m5, m6
 | |
|     mova  [dstq      ], m0
 | |
|     mova  [dstq+dst1q], m1
 | |
|     mova  [dstq+dst2q], m2
 | |
|     mova  [dstq+dst3q], m3
 | |
|     mova  [dstq+dst4q], m4
 | |
|     mova  [dstq+dst5q], m5
 | |
|     add      srcq, mmsize*3
 | |
|     add      dstq, mmsize
 | |
|     sub      lend, mmsize/4
 | |
|     jg .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_S16_TO_FLTP_6CH
 | |
| INIT_XMM ssse3
 | |
| CONV_S16_TO_FLTP_6CH
 | |
| INIT_XMM sse4
 | |
| CONV_S16_TO_FLTP_6CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_S16_TO_FLTP_6CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_flt_to_s16p_2ch(int16_t *const *dst, float *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_FLT_TO_S16P_2CH 0
 | |
| cglobal conv_flt_to_s16p_2ch, 3,4,6, dst0, src, len, dst1
 | |
|     lea       lenq, [2*lend]
 | |
|     mov      dst1q, [dst0q+gprsize]
 | |
|     mov      dst0q, [dst0q        ]
 | |
|     lea       srcq, [srcq+4*lenq]
 | |
|     add      dst0q, lenq
 | |
|     add      dst1q, lenq
 | |
|     neg       lenq
 | |
|     mova        m5, [pf_s16_scale]
 | |
| .loop:
 | |
|     mova       m0, [srcq+4*lenq         ]
 | |
|     mova       m1, [srcq+4*lenq+  mmsize]
 | |
|     mova       m2, [srcq+4*lenq+2*mmsize]
 | |
|     mova       m3, [srcq+4*lenq+3*mmsize]
 | |
|     DEINT2_PS   0, 1, 4
 | |
|     DEINT2_PS   2, 3, 4
 | |
|     mulps      m0, m0, m5
 | |
|     mulps      m1, m1, m5
 | |
|     mulps      m2, m2, m5
 | |
|     mulps      m3, m3, m5
 | |
|     cvtps2dq   m0, m0
 | |
|     cvtps2dq   m1, m1
 | |
|     cvtps2dq   m2, m2
 | |
|     cvtps2dq   m3, m3
 | |
|     packssdw   m0, m2
 | |
|     packssdw   m1, m3
 | |
|     mova  [dst0q+lenq], m0
 | |
|     mova  [dst1q+lenq], m1
 | |
|     add      lenq, mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_FLT_TO_S16P_2CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_FLT_TO_S16P_2CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_flt_to_s16p_6ch(int16_t *const *dst, float *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_FLT_TO_S16P_6CH 0
 | |
| %if ARCH_X86_64
 | |
| cglobal conv_flt_to_s16p_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5
 | |
| %else
 | |
| cglobal conv_flt_to_s16p_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5
 | |
| %define lend dword r2m
 | |
| %endif
 | |
|     mov     dst1q, [dstq+  gprsize]
 | |
|     mov     dst2q, [dstq+2*gprsize]
 | |
|     mov     dst3q, [dstq+3*gprsize]
 | |
|     mov     dst4q, [dstq+4*gprsize]
 | |
|     mov     dst5q, [dstq+5*gprsize]
 | |
|     mov      dstq, [dstq          ]
 | |
|     sub     dst1q, dstq
 | |
|     sub     dst2q, dstq
 | |
|     sub     dst3q, dstq
 | |
|     sub     dst4q, dstq
 | |
|     sub     dst5q, dstq
 | |
|     mova       m6, [pf_s16_scale]
 | |
| .loop:
 | |
|     mulps      m0, m6, [srcq+0*mmsize]
 | |
|     mulps      m3, m6, [srcq+1*mmsize]
 | |
|     mulps      m1, m6, [srcq+2*mmsize]
 | |
|     mulps      m4, m6, [srcq+3*mmsize]
 | |
|     mulps      m2, m6, [srcq+4*mmsize]
 | |
|     mulps      m5, m6, [srcq+5*mmsize]
 | |
|     cvtps2dq   m0, m0
 | |
|     cvtps2dq   m1, m1
 | |
|     cvtps2dq   m2, m2
 | |
|     cvtps2dq   m3, m3
 | |
|     cvtps2dq   m4, m4
 | |
|     cvtps2dq   m5, m5
 | |
|     packssdw   m0, m3               ; m0 =  0,  1,  2,  3,  4,  5,  6,  7
 | |
|     packssdw   m1, m4               ; m1 =  8,  9, 10, 11, 12, 13, 14, 15
 | |
|     packssdw   m2, m5               ; m2 = 16, 17, 18, 19, 20, 21, 22, 23
 | |
|     PALIGNR    m3, m1, m0, 12, m4   ; m3 =  6,  7,  8,  9, 10, 11,  x,  x
 | |
|     shufps     m1, m2, q1032        ; m1 = 12, 13, 14, 15, 16, 17, 18, 19
 | |
|     psrldq     m2, 4                ; m2 = 18, 19, 20, 21, 22, 23,  x,  x
 | |
|     SBUTTERFLY2 wd, 0, 3, 4         ; m0 =  0,  6,  1,  7,  2,  8,  3,  9
 | |
|                                     ; m3 =  4, 10,  5, 11,  x,  x,  x,  x
 | |
|     SBUTTERFLY2 wd, 1, 2, 4         ; m1 = 12, 18, 13, 19, 14, 20, 15, 21
 | |
|                                     ; m2 = 16, 22, 17, 23,  x,  x,  x,  x
 | |
|     SBUTTERFLY2 dq, 0, 1, 4         ; m0 =  0,  6, 12, 18,  1,  7, 13, 19
 | |
|                                     ; m1 =  2,  8, 14, 20,  3,  9, 15, 21
 | |
|     punpckldq  m3, m2               ; m3 =  4, 10, 16, 22,  5, 11, 17, 23
 | |
|     movq    [dstq      ], m0
 | |
|     movhps  [dstq+dst1q], m0
 | |
|     movq    [dstq+dst2q], m1
 | |
|     movhps  [dstq+dst3q], m1
 | |
|     movq    [dstq+dst4q], m3
 | |
|     movhps  [dstq+dst5q], m3
 | |
|     add      srcq, mmsize*6
 | |
|     add      dstq, mmsize/2
 | |
|     sub      lend, mmsize/4
 | |
|     jg .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_FLT_TO_S16P_6CH
 | |
| INIT_XMM ssse3
 | |
| CONV_FLT_TO_S16P_6CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_FLT_TO_S16P_6CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_flt_to_fltp_2ch(float *const *dst, float *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_FLT_TO_FLTP_2CH 0
 | |
| cglobal conv_flt_to_fltp_2ch, 3,4,3, dst0, src, len, dst1
 | |
|     lea    lenq, [4*lend]
 | |
|     mov   dst1q, [dst0q+gprsize]
 | |
|     mov   dst0q, [dst0q        ]
 | |
|     lea    srcq, [srcq+2*lenq]
 | |
|     add   dst0q, lenq
 | |
|     add   dst1q, lenq
 | |
|     neg    lenq
 | |
| .loop:
 | |
|     mova     m0, [srcq+2*lenq       ]
 | |
|     mova     m1, [srcq+2*lenq+mmsize]
 | |
|     DEINT2_PS 0, 1, 2
 | |
|     mova  [dst0q+lenq], m0
 | |
|     mova  [dst1q+lenq], m1
 | |
|     add    lenq, mmsize
 | |
|     jl .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse
 | |
| CONV_FLT_TO_FLTP_2CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_FLT_TO_FLTP_2CH
 | |
| %endif
 | |
| 
 | |
| ;------------------------------------------------------------------------------
 | |
| ; void ff_conv_flt_to_fltp_6ch(float *const *dst, float *src, int len,
 | |
| ;                              int channels);
 | |
| ;------------------------------------------------------------------------------
 | |
| 
 | |
| %macro CONV_FLT_TO_FLTP_6CH 0
 | |
| %if ARCH_X86_64
 | |
| cglobal conv_flt_to_fltp_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5
 | |
| %else
 | |
| cglobal conv_flt_to_fltp_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5
 | |
| %define lend dword r2m
 | |
| %endif
 | |
|     mov     dst1q, [dstq+  gprsize]
 | |
|     mov     dst2q, [dstq+2*gprsize]
 | |
|     mov     dst3q, [dstq+3*gprsize]
 | |
|     mov     dst4q, [dstq+4*gprsize]
 | |
|     mov     dst5q, [dstq+5*gprsize]
 | |
|     mov      dstq, [dstq          ]
 | |
|     sub     dst1q, dstq
 | |
|     sub     dst2q, dstq
 | |
|     sub     dst3q, dstq
 | |
|     sub     dst4q, dstq
 | |
|     sub     dst5q, dstq
 | |
| .loop:
 | |
|     mova       m0, [srcq+0*mmsize]  ; m0 =  0,  1,  2,  3
 | |
|     mova       m1, [srcq+1*mmsize]  ; m1 =  4,  5,  6,  7
 | |
|     mova       m2, [srcq+2*mmsize]  ; m2 =  8,  9, 10, 11
 | |
|     mova       m3, [srcq+3*mmsize]  ; m3 = 12, 13, 14, 15
 | |
|     mova       m4, [srcq+4*mmsize]  ; m4 = 16, 17, 18, 19
 | |
|     mova       m5, [srcq+5*mmsize]  ; m5 = 20, 21, 22, 23
 | |
| 
 | |
|     SBUTTERFLY2 dq, 0, 3, 6         ; m0 =  0, 12,  1, 13
 | |
|                                     ; m3 =  2, 14,  3, 15
 | |
|     SBUTTERFLY2 dq, 1, 4, 6         ; m1 =  4, 16,  5, 17
 | |
|                                     ; m4 =  6, 18,  7, 19
 | |
|     SBUTTERFLY2 dq, 2, 5, 6         ; m2 =  8, 20,  9, 21
 | |
|                                     ; m5 = 10, 22, 11, 23
 | |
|     SBUTTERFLY2 dq, 0, 4, 6         ; m0 =  0,  6, 12, 18
 | |
|                                     ; m4 =  1,  7, 13, 19
 | |
|     SBUTTERFLY2 dq, 3, 2, 6         ; m3 =  2,  8, 14, 20
 | |
|                                     ; m2 =  3,  9, 15, 21
 | |
|     SBUTTERFLY2 dq, 1, 5, 6         ; m1 =  4, 10, 16, 22
 | |
|                                     ; m5 =  5, 11, 17, 23
 | |
|     mova [dstq      ], m0
 | |
|     mova [dstq+dst1q], m4
 | |
|     mova [dstq+dst2q], m3
 | |
|     mova [dstq+dst3q], m2
 | |
|     mova [dstq+dst4q], m1
 | |
|     mova [dstq+dst5q], m5
 | |
|     add      srcq, mmsize*6
 | |
|     add      dstq, mmsize
 | |
|     sub      lend, mmsize/4
 | |
|     jg .loop
 | |
|     REP_RET
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| CONV_FLT_TO_FLTP_6CH
 | |
| %if HAVE_AVX_EXTERNAL
 | |
| INIT_XMM avx
 | |
| CONV_FLT_TO_FLTP_6CH
 | |
| %endif
 |