429 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
			
		
		
	
	
			429 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			NASM
		
	
	
	
	
	
| ;******************************************************************************
 | |
| ;* x86-SIMD-optimized IDCT for prores
 | |
| ;* this is identical to "simple" IDCT except for the clip range
 | |
| ;*
 | |
| ;* Copyright (c) 2011 Ronald S. Bultje <rsbultje@gmail.com>
 | |
| ;*
 | |
| ;* This file is part of Libav.
 | |
| ;*
 | |
| ;* Libav is free software; you can redistribute it and/or
 | |
| ;* modify it under the terms of the GNU Lesser General Public
 | |
| ;* License as published by the Free Software Foundation; either
 | |
| ;* version 2.1 of the License, or (at your option) any later version.
 | |
| ;*
 | |
| ;* Libav is distributed in the hope that it will be useful,
 | |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
| ;* Lesser General Public License for more details.
 | |
| ;*
 | |
| ;* You should have received a copy of the GNU Lesser General Public
 | |
| ;* License along with Libav; if not, write to the Free Software
 | |
| ;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
| ;******************************************************************************
 | |
| 
 | |
| %include "libavutil/x86/x86util.asm"
 | |
| 
 | |
| %define W1sh2 22725 ; W1 = 90901 = 22725<<2 + 1
 | |
| %define W2sh2 21407 ; W2 = 85627 = 21407<<2 - 1
 | |
| %define W3sh2 19265 ; W3 = 77062 = 19265<<2 + 2
 | |
| %define W4sh2 16384 ; W4 = 65535 = 16384<<2 - 1
 | |
| %define W5sh2 12873 ; W5 = 51491 = 12873<<2 - 1
 | |
| %define W6sh2  8867 ; W6 = 35468 =  8867<<2
 | |
| %define W7sh2  4520 ; W7 = 18081 =  4520<<2 + 1
 | |
| 
 | |
| %if ARCH_X86_64
 | |
| 
 | |
| SECTION_RODATA
 | |
| 
 | |
| w4_plus_w2: times 4 dw W4sh2, +W2sh2
 | |
| w4_min_w2:  times 4 dw W4sh2, -W2sh2
 | |
| w4_plus_w6: times 4 dw W4sh2, +W6sh2
 | |
| w4_min_w6:  times 4 dw W4sh2, -W6sh2
 | |
| w1_plus_w3: times 4 dw W1sh2, +W3sh2
 | |
| w3_min_w1:  times 4 dw W3sh2, -W1sh2
 | |
| w7_plus_w3: times 4 dw W7sh2, +W3sh2
 | |
| w3_min_w7:  times 4 dw W3sh2, -W7sh2
 | |
| w1_plus_w5: times 4 dw W1sh2, +W5sh2
 | |
| w5_min_w1:  times 4 dw W5sh2, -W1sh2
 | |
| w5_plus_w7: times 4 dw W5sh2, +W7sh2
 | |
| w7_min_w5:  times 4 dw W7sh2, -W5sh2
 | |
| row_round:  times 8 dw (1<<14)
 | |
| 
 | |
| cextern pw_4
 | |
| cextern pw_8
 | |
| cextern pw_512
 | |
| cextern pw_1019
 | |
| 
 | |
| section .text align=16
 | |
| 
 | |
| ; interleave data while maintaining source
 | |
| ; %1=type, %2=dstlo, %3=dsthi, %4=src, %5=interleave
 | |
| %macro SBUTTERFLY3 5
 | |
|     punpckl%1   m%2, m%4, m%5
 | |
|     punpckh%1   m%3, m%4, m%5
 | |
| %endmacro
 | |
| 
 | |
| ; %1/%2=src1/dst1, %3/%4=dst2, %5/%6=src2, %7=shift
 | |
| ; action: %3/%4 = %1/%2 - %5/%6; %1/%2 += %5/%6
 | |
| ;         %1/%2/%3/%4 >>= %7; dword -> word (in %1/%3)
 | |
| %macro SUMSUB_SHPK 7
 | |
|     psubd       %3,  %1,  %5       ; { a0 - b0 }[0-3]
 | |
|     psubd       %4,  %2,  %6       ; { a0 - b0 }[4-7]
 | |
|     paddd       %1,  %5            ; { a0 + b0 }[0-3]
 | |
|     paddd       %2,  %6            ; { a0 + b0 }[4-7]
 | |
|     psrad       %1,  %7
 | |
|     psrad       %2,  %7
 | |
|     psrad       %3,  %7
 | |
|     psrad       %4,  %7
 | |
|     packssdw    %1,  %2            ; row[0]
 | |
|     packssdw    %3,  %4            ; row[7]
 | |
| %endmacro
 | |
| 
 | |
| ; %1 = row or col (for rounding variable)
 | |
| ; %2 = number of bits to shift at the end
 | |
| %macro IDCT_1D 2
 | |
|     ; a0 = (W4 * row[0]) + (1 << (15 - 1));
 | |
|     ; a1 = a0;
 | |
|     ; a2 = a0;
 | |
|     ; a3 = a0;
 | |
|     ; a0 += W2 * row[2];
 | |
|     ; a1 += W6 * row[2];
 | |
|     ; a2 -= W6 * row[2];
 | |
|     ; a3 -= W2 * row[2];
 | |
| %ifidn %1, col
 | |
|     paddw       m10,[pw_8]
 | |
| %endif
 | |
|     SBUTTERFLY3 wd,  0,  1, 10,  8 ; { row[0], row[2] }[0-3]/[4-7]
 | |
| %ifidn %1, row
 | |
|     psubw       m10,[row_round]
 | |
| %endif
 | |
|     SIGNEXTEND  m8,  m9,  m14      ; { row[2] }[0-3] / [4-7]
 | |
|     SIGNEXTEND  m10, m11, m14      ; { row[0] }[0-3] / [4-7]
 | |
|     pmaddwd     m2,  m0, [w4_plus_w6]
 | |
|     pmaddwd     m3,  m1, [w4_plus_w6]
 | |
|     pmaddwd     m4,  m0, [w4_min_w6]
 | |
|     pmaddwd     m5,  m1, [w4_min_w6]
 | |
|     pmaddwd     m6,  m0, [w4_min_w2]
 | |
|     pmaddwd     m7,  m1, [w4_min_w2]
 | |
|     pmaddwd     m0, [w4_plus_w2]
 | |
|     pmaddwd     m1, [w4_plus_w2]
 | |
|     pslld       m2,  2
 | |
|     pslld       m3,  2
 | |
|     pslld       m4,  2
 | |
|     pslld       m5,  2
 | |
|     pslld       m6,  2
 | |
|     pslld       m7,  2
 | |
|     pslld       m0,  2
 | |
|     pslld       m1,  2
 | |
| 
 | |
|     ; a0: -1*row[0]-1*row[2]
 | |
|     ; a1: -1*row[0]
 | |
|     ; a2: -1*row[0]
 | |
|     ; a3: -1*row[0]+1*row[2]
 | |
|     psubd       m2,  m10           ; a1[0-3]
 | |
|     psubd       m3,  m11           ; a1[4-7]
 | |
|     psubd       m4,  m10           ; a2[0-3]
 | |
|     psubd       m5,  m11           ; a2[4-7]
 | |
|     psubd       m0,  m10
 | |
|     psubd       m1,  m11
 | |
|     psubd       m6,  m10
 | |
|     psubd       m7,  m11
 | |
|     psubd       m0,  m8            ; a0[0-3]
 | |
|     psubd       m1,  m9            ; a0[4-7]
 | |
|     paddd       m6,  m8            ; a3[0-3]
 | |
|     paddd       m7,  m9            ; a3[4-7]
 | |
| 
 | |
|     ; a0 +=   W4*row[4] + W6*row[6]; i.e. -1*row[4]
 | |
|     ; a1 -=   W4*row[4] + W2*row[6]; i.e. -1*row[4]-1*row[6]
 | |
|     ; a2 -=   W4*row[4] - W2*row[6]; i.e. -1*row[4]+1*row[6]
 | |
|     ; a3 +=   W4*row[4] - W6*row[6]; i.e. -1*row[4]
 | |
|     SBUTTERFLY3 wd,  8,  9, 13, 12 ; { row[4], row[6] }[0-3]/[4-7]
 | |
|     SIGNEXTEND  m13, m14, m10      ; { row[4] }[0-3] / [4-7]
 | |
|     pmaddwd     m10, m8, [w4_plus_w6]
 | |
|     pmaddwd     m11, m9, [w4_plus_w6]
 | |
|     pslld       m10, 2
 | |
|     pslld       m11, 2
 | |
|     psubd       m10,  m13
 | |
|     psubd       m11,  m14
 | |
|     paddd       m0,  m10            ; a0[0-3]
 | |
|     paddd       m1,  m11            ; a0[4-7]
 | |
|     pmaddwd     m10, m8, [w4_min_w6]
 | |
|     pmaddwd     m11, m9, [w4_min_w6]
 | |
|     pslld       m10, 2
 | |
|     pslld       m11, 2
 | |
|     psubd       m10, m13
 | |
|     psubd       m11, m14
 | |
|     paddd       m6,  m10           ; a3[0-3]
 | |
|     paddd       m7,  m11           ; a3[4-7]
 | |
|     pmaddwd     m10, m8, [w4_min_w2]
 | |
|     pmaddwd     m11, m9, [w4_min_w2]
 | |
|     pmaddwd     m8, [w4_plus_w2]
 | |
|     pmaddwd     m9, [w4_plus_w2]
 | |
|     pslld       m10, 2
 | |
|     pslld       m11, 2
 | |
|     pslld       m8,  2
 | |
|     pslld       m9,  2
 | |
|     psubd       m10, m13
 | |
|     psubd       m11, m14
 | |
|     psubd       m8,  m13
 | |
|     psubd       m9,  m14
 | |
|     psubd       m4,  m10           ; a2[0-3] intermediate
 | |
|     psubd       m5,  m11           ; a2[4-7] intermediate
 | |
|     psubd       m2,  m8            ; a1[0-3] intermediate
 | |
|     psubd       m3,  m9            ; a1[4-7] intermediate
 | |
|     SIGNEXTEND  m12, m13, m10      ; { row[6] }[0-3] / [4-7]
 | |
|     psubd       m4,  m12           ; a2[0-3]
 | |
|     psubd       m5,  m13           ; a2[4-7]
 | |
|     paddd       m2,  m12           ; a1[0-3]
 | |
|     paddd       m3,  m13           ; a1[4-7]
 | |
| 
 | |
|     ; load/store
 | |
|     mova   [r2+  0], m0
 | |
|     mova   [r2+ 32], m2
 | |
|     mova   [r2+ 64], m4
 | |
|     mova   [r2+ 96], m6
 | |
|     mova        m10,[r2+ 16]       ; { row[1] }[0-7]
 | |
|     mova        m8, [r2+ 48]       ; { row[3] }[0-7]
 | |
|     mova        m13,[r2+ 80]       ; { row[5] }[0-7]
 | |
|     mova        m14,[r2+112]       ; { row[7] }[0-7]
 | |
|     mova   [r2+ 16], m1
 | |
|     mova   [r2+ 48], m3
 | |
|     mova   [r2+ 80], m5
 | |
|     mova   [r2+112], m7
 | |
| %ifidn %1, row
 | |
|     pmullw      m10,[r3+ 16]
 | |
|     pmullw      m8, [r3+ 48]
 | |
|     pmullw      m13,[r3+ 80]
 | |
|     pmullw      m14,[r3+112]
 | |
| %endif
 | |
| 
 | |
|     ; b0 = MUL(W1, row[1]);
 | |
|     ; MAC(b0, W3, row[3]);
 | |
|     ; b1 = MUL(W3, row[1]);
 | |
|     ; MAC(b1, -W7, row[3]);
 | |
|     ; b2 = MUL(W5, row[1]);
 | |
|     ; MAC(b2, -W1, row[3]);
 | |
|     ; b3 = MUL(W7, row[1]);
 | |
|     ; MAC(b3, -W5, row[3]);
 | |
|     SBUTTERFLY3 wd,  0,  1, 10, 8  ; { row[1], row[3] }[0-3]/[4-7]
 | |
|     SIGNEXTEND  m10, m11, m12      ; { row[1] }[0-3] / [4-7]
 | |
|     SIGNEXTEND  m8,  m9,  m12      ; { row[3] }[0-3] / [4-7]
 | |
|     pmaddwd     m2,  m0, [w3_min_w7]
 | |
|     pmaddwd     m3,  m1, [w3_min_w7]
 | |
|     pmaddwd     m4,  m0, [w5_min_w1]
 | |
|     pmaddwd     m5,  m1, [w5_min_w1]
 | |
|     pmaddwd     m6,  m0, [w7_min_w5]
 | |
|     pmaddwd     m7,  m1, [w7_min_w5]
 | |
|     pmaddwd     m0, [w1_plus_w3]
 | |
|     pmaddwd     m1, [w1_plus_w3]
 | |
|     pslld       m2,  2
 | |
|     pslld       m3,  2
 | |
|     pslld       m4,  2
 | |
|     pslld       m5,  2
 | |
|     pslld       m6,  2
 | |
|     pslld       m7,  2
 | |
|     pslld       m0,  2
 | |
|     pslld       m1,  2
 | |
| 
 | |
|     ; b0: +1*row[1]+2*row[3]
 | |
|     ; b1: +2*row[1]-1*row[3]
 | |
|     ; b2: -1*row[1]-1*row[3]
 | |
|     ; b3: +1*row[1]+1*row[3]
 | |
|     psubd       m2,  m8
 | |
|     psubd       m3,  m9
 | |
|     paddd       m0,  m8
 | |
|     paddd       m1,  m9
 | |
|     paddd       m8,  m10           ; { row[1] + row[3] }[0-3]
 | |
|     paddd       m9,  m11           ; { row[1] + row[3] }[4-7]
 | |
|     paddd       m10, m10
 | |
|     paddd       m11, m11
 | |
|     paddd       m0,  m8            ; b0[0-3]
 | |
|     paddd       m1,  m9            ; b0[4-7]
 | |
|     paddd       m2,  m10           ; b1[0-3]
 | |
|     paddd       m3,  m11           ; b2[4-7]
 | |
|     psubd       m4,  m8            ; b2[0-3]
 | |
|     psubd       m5,  m9            ; b2[4-7]
 | |
|     paddd       m6,  m8            ; b3[0-3]
 | |
|     paddd       m7,  m9            ; b3[4-7]
 | |
| 
 | |
|     ; MAC(b0,  W5, row[5]);
 | |
|     ; MAC(b0,  W7, row[7]);
 | |
|     ; MAC(b1, -W1, row[5]);
 | |
|     ; MAC(b1, -W5, row[7]);
 | |
|     ; MAC(b2,  W7, row[5]);
 | |
|     ; MAC(b2,  W3, row[7]);
 | |
|     ; MAC(b3,  W3, row[5]);
 | |
|     ; MAC(b3, -W1, row[7]);
 | |
|     SBUTTERFLY3 wd,  8,  9, 13, 14 ; { row[5], row[7] }[0-3]/[4-7]
 | |
|     SIGNEXTEND  m13, m12, m11      ; { row[5] }[0-3] / [4-7]
 | |
|     SIGNEXTEND  m14, m11, m10      ; { row[7] }[0-3] / [4-7]
 | |
| 
 | |
|     ; b0: -1*row[5]+1*row[7]
 | |
|     ; b1: -1*row[5]+1*row[7]
 | |
|     ; b2: +1*row[5]+2*row[7]
 | |
|     ; b3: +2*row[5]-1*row[7]
 | |
|     paddd       m4,  m13
 | |
|     paddd       m5,  m12
 | |
|     paddd       m6,  m13
 | |
|     paddd       m7,  m12
 | |
|     psubd       m13, m14           ; { row[5] - row[7] }[0-3]
 | |
|     psubd       m12, m11           ; { row[5] - row[7] }[4-7]
 | |
|     paddd       m14, m14
 | |
|     paddd       m11, m11
 | |
|     psubd       m0,  m13
 | |
|     psubd       m1,  m12
 | |
|     psubd       m2,  m13
 | |
|     psubd       m3,  m12
 | |
|     paddd       m4,  m14
 | |
|     paddd       m5,  m11
 | |
|     paddd       m6,  m13
 | |
|     paddd       m7,  m12
 | |
| 
 | |
|     pmaddwd     m10, m8, [w1_plus_w5]
 | |
|     pmaddwd     m11, m9, [w1_plus_w5]
 | |
|     pmaddwd     m12, m8, [w5_plus_w7]
 | |
|     pmaddwd     m13, m9, [w5_plus_w7]
 | |
|     pslld       m10, 2
 | |
|     pslld       m11, 2
 | |
|     pslld       m12,  2
 | |
|     pslld       m13,  2
 | |
|     psubd       m2,  m10           ; b1[0-3]
 | |
|     psubd       m3,  m11           ; b1[4-7]
 | |
|     paddd       m0,  m12            ; b0[0-3]
 | |
|     paddd       m1,  m13            ; b0[4-7]
 | |
|     pmaddwd     m12, m8, [w7_plus_w3]
 | |
|     pmaddwd     m13, m9, [w7_plus_w3]
 | |
|     pmaddwd     m8, [w3_min_w1]
 | |
|     pmaddwd     m9, [w3_min_w1]
 | |
|     pslld       m12, 2
 | |
|     pslld       m13, 2
 | |
|     pslld       m8,  2
 | |
|     pslld       m9,  2
 | |
|     paddd       m4,  m12           ; b2[0-3]
 | |
|     paddd       m5,  m13           ; b2[4-7]
 | |
|     paddd       m6,  m8            ; b3[0-3]
 | |
|     paddd       m7,  m9            ; b3[4-7]
 | |
| 
 | |
|     ; row[0] = (a0 + b0) >> 15;
 | |
|     ; row[7] = (a0 - b0) >> 15;
 | |
|     ; row[1] = (a1 + b1) >> 15;
 | |
|     ; row[6] = (a1 - b1) >> 15;
 | |
|     ; row[2] = (a2 + b2) >> 15;
 | |
|     ; row[5] = (a2 - b2) >> 15;
 | |
|     ; row[3] = (a3 + b3) >> 15;
 | |
|     ; row[4] = (a3 - b3) >> 15;
 | |
|     mova        m8, [r2+ 0]        ; a0[0-3]
 | |
|     mova        m9, [r2+16]        ; a0[4-7]
 | |
|     SUMSUB_SHPK m8,  m9,  m10, m11, m0,  m1,  %2
 | |
|     mova        m0, [r2+32]        ; a1[0-3]
 | |
|     mova        m1, [r2+48]        ; a1[4-7]
 | |
|     SUMSUB_SHPK m0,  m1,  m9,  m11, m2,  m3,  %2
 | |
|     mova        m1, [r2+64]        ; a2[0-3]
 | |
|     mova        m2, [r2+80]        ; a2[4-7]
 | |
|     SUMSUB_SHPK m1,  m2,  m11, m3,  m4,  m5,  %2
 | |
|     mova        m2, [r2+96]        ; a3[0-3]
 | |
|     mova        m3, [r2+112]       ; a3[4-7]
 | |
|     SUMSUB_SHPK m2,  m3,  m4,  m5,  m6,  m7,  %2
 | |
| %endmacro
 | |
| 
 | |
| ; void ff_prores_idct_put_10_<opt>(uint8_t *pixels, int stride,
 | |
| ;                                  int16_t *block, const int16_t *qmat);
 | |
| %macro idct_put_fn 1
 | |
| cglobal prores_idct_put_10, 4, 4, %1
 | |
|     movsxd      r1,  r1d
 | |
|     pxor        m15, m15           ; zero
 | |
| 
 | |
|     ; for (i = 0; i < 8; i++)
 | |
|     ;     idctRowCondDC(block + i*8);
 | |
|     mova        m10,[r2+ 0]        ; { row[0] }[0-7]
 | |
|     mova        m8, [r2+32]        ; { row[2] }[0-7]
 | |
|     mova        m13,[r2+64]        ; { row[4] }[0-7]
 | |
|     mova        m12,[r2+96]        ; { row[6] }[0-7]
 | |
| 
 | |
|     pmullw      m10,[r3+ 0]
 | |
|     pmullw      m8, [r3+32]
 | |
|     pmullw      m13,[r3+64]
 | |
|     pmullw      m12,[r3+96]
 | |
| 
 | |
|     IDCT_1D     row, 17
 | |
| 
 | |
|     ; transpose for second part of IDCT
 | |
|     TRANSPOSE8x8W 8, 0, 1, 2, 4, 11, 9, 10, 3
 | |
|     mova   [r2+ 16], m0
 | |
|     mova   [r2+ 48], m2
 | |
|     mova   [r2+ 80], m11
 | |
|     mova   [r2+112], m10
 | |
|     SWAP         8,  10
 | |
|     SWAP         1,   8
 | |
|     SWAP         4,  13
 | |
|     SWAP         9,  12
 | |
| 
 | |
|     ; for (i = 0; i < 8; i++)
 | |
|     ;     idctSparseColAdd(dest + i, line_size, block + i);
 | |
|     IDCT_1D     col, 20
 | |
| 
 | |
|     ; clip/store
 | |
|     mova        m6, [pw_512]
 | |
|     mova        m3, [pw_4]
 | |
|     mova        m5, [pw_1019]
 | |
|     paddw       m8,  m6
 | |
|     paddw       m0,  m6
 | |
|     paddw       m1,  m6
 | |
|     paddw       m2,  m6
 | |
|     paddw       m4,  m6
 | |
|     paddw       m11, m6
 | |
|     paddw       m9,  m6
 | |
|     paddw       m10, m6
 | |
|     pmaxsw      m8,  m3
 | |
|     pmaxsw      m0,  m3
 | |
|     pmaxsw      m1,  m3
 | |
|     pmaxsw      m2,  m3
 | |
|     pmaxsw      m4,  m3
 | |
|     pmaxsw      m11, m3
 | |
|     pmaxsw      m9,  m3
 | |
|     pmaxsw      m10, m3
 | |
|     pminsw      m8,  m5
 | |
|     pminsw      m0,  m5
 | |
|     pminsw      m1,  m5
 | |
|     pminsw      m2,  m5
 | |
|     pminsw      m4,  m5
 | |
|     pminsw      m11, m5
 | |
|     pminsw      m9,  m5
 | |
|     pminsw      m10, m5
 | |
| 
 | |
|     lea         r2, [r1*3]
 | |
|     mova  [r0     ], m8
 | |
|     mova  [r0+r1  ], m0
 | |
|     mova  [r0+r1*2], m1
 | |
|     mova  [r0+r2  ], m2
 | |
|     lea         r0, [r0+r1*4]
 | |
|     mova  [r0     ], m4
 | |
|     mova  [r0+r1  ], m11
 | |
|     mova  [r0+r1*2], m9
 | |
|     mova  [r0+r2  ], m10
 | |
|     RET
 | |
| %endmacro
 | |
| 
 | |
| %macro SIGNEXTEND 2-3
 | |
| %if cpuflag(sse4) ; dstlow, dsthigh
 | |
|     movhlps     %2,  %1
 | |
|     pmovsxwd    %1,  %1
 | |
|     pmovsxwd    %2,  %2
 | |
| %elif cpuflag(sse2) ; dstlow, dsthigh, tmp
 | |
|     pxor        %3,  %3
 | |
|     pcmpgtw     %3,  %1
 | |
|     mova        %2,  %1
 | |
|     punpcklwd   %1,  %3
 | |
|     punpckhwd   %2,  %3
 | |
| %endif
 | |
| %endmacro
 | |
| 
 | |
| INIT_XMM sse2
 | |
| idct_put_fn 16
 | |
| INIT_XMM sse4
 | |
| idct_put_fn 16
 | |
| INIT_XMM avx
 | |
| idct_put_fn 16
 | |
| 
 | |
| %endif
 |