411 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
			
		
		
	
	
			411 lines
		
	
	
		
			13 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
	
	
| /*
 | |
|  * ARM NEON optimised DSP functions
 | |
|  * Copyright (c) 2008 Mans Rullgard <mans@mansr.com>
 | |
|  *
 | |
|  * This file is part of FFmpeg.
 | |
|  *
 | |
|  * FFmpeg is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of the GNU Lesser General Public
 | |
|  * License as published by the Free Software Foundation; either
 | |
|  * version 2.1 of the License, or (at your option) any later version.
 | |
|  *
 | |
|  * FFmpeg is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
|  * Lesser General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU Lesser General Public
 | |
|  * License along with FFmpeg; if not, write to the Free Software
 | |
|  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
|  */
 | |
| 
 | |
| #include "libavutil/arm/asm.S"
 | |
| 
 | |
| .macro  pixels16        rnd=1, avg=0
 | |
|   .if \avg
 | |
|         mov             r12, r0
 | |
|   .endif
 | |
| 1:      vld1.8          {q0},     [r1], r2
 | |
|         vld1.8          {q1},     [r1], r2
 | |
|         vld1.8          {q2},     [r1], r2
 | |
|         pld             [r1, r2, lsl #2]
 | |
|         vld1.8          {q3},     [r1], r2
 | |
|         pld             [r1]
 | |
|         pld             [r1, r2]
 | |
|         pld             [r1, r2, lsl #1]
 | |
|   .if \avg
 | |
|         vld1.8          {q8},     [r12,:128], r2
 | |
|         vrhadd.u8       q0,  q0,  q8
 | |
|         vld1.8          {q9},     [r12,:128], r2
 | |
|         vrhadd.u8       q1,  q1,  q9
 | |
|         vld1.8          {q10},    [r12,:128], r2
 | |
|         vrhadd.u8       q2,  q2,  q10
 | |
|         vld1.8          {q11},    [r12,:128], r2
 | |
|         vrhadd.u8       q3,  q3,  q11
 | |
|   .endif
 | |
|         subs            r3,  r3,  #4
 | |
|         vst1.64         {q0},     [r0,:128], r2
 | |
|         vst1.64         {q1},     [r0,:128], r2
 | |
|         vst1.64         {q2},     [r0,:128], r2
 | |
|         vst1.64         {q3},     [r0,:128], r2
 | |
|         bne             1b
 | |
|         bx              lr
 | |
| .endm
 | |
| 
 | |
| .macro  pixels16_x2     rnd=1, avg=0
 | |
| 1:      vld1.8          {d0-d2},  [r1], r2
 | |
|         vld1.8          {d4-d6},  [r1], r2
 | |
|         pld             [r1]
 | |
|         pld             [r1, r2]
 | |
|         subs            r3,  r3,  #2
 | |
|         vext.8          q1,  q0,  q1,  #1
 | |
|         avg             q0,  q0,  q1
 | |
|         vext.8          q3,  q2,  q3,  #1
 | |
|         avg             q2,  q2,  q3
 | |
|   .if \avg
 | |
|         vld1.8          {q1},     [r0,:128], r2
 | |
|         vld1.8          {q3},     [r0,:128]
 | |
|         vrhadd.u8       q0,  q0,  q1
 | |
|         vrhadd.u8       q2,  q2,  q3
 | |
|         sub             r0,  r0,  r2
 | |
|   .endif
 | |
|         vst1.8          {q0},     [r0,:128], r2
 | |
|         vst1.8          {q2},     [r0,:128], r2
 | |
|         bne             1b
 | |
|         bx              lr
 | |
| .endm
 | |
| 
 | |
| .macro  pixels16_y2     rnd=1, avg=0
 | |
|         sub             r3,  r3,  #2
 | |
|         vld1.8          {q0},     [r1], r2
 | |
|         vld1.8          {q1},     [r1], r2
 | |
| 1:      subs            r3,  r3,  #2
 | |
|         avg             q2,  q0,  q1
 | |
|         vld1.8          {q0},     [r1], r2
 | |
|         avg             q3,  q0,  q1
 | |
|         vld1.8          {q1},     [r1], r2
 | |
|         pld             [r1]
 | |
|         pld             [r1, r2]
 | |
|   .if \avg
 | |
|         vld1.8          {q8},     [r0,:128], r2
 | |
|         vld1.8          {q9},     [r0,:128]
 | |
|         vrhadd.u8       q2,  q2,  q8
 | |
|         vrhadd.u8       q3,  q3,  q9
 | |
|         sub             r0,  r0,  r2
 | |
|   .endif
 | |
|         vst1.8          {q2},     [r0,:128], r2
 | |
|         vst1.8          {q3},     [r0,:128], r2
 | |
|         bne             1b
 | |
| 
 | |
|         avg             q2,  q0,  q1
 | |
|         vld1.8          {q0},     [r1], r2
 | |
|         avg             q3,  q0,  q1
 | |
|   .if \avg
 | |
|         vld1.8          {q8},     [r0,:128], r2
 | |
|         vld1.8          {q9},     [r0,:128]
 | |
|         vrhadd.u8       q2,  q2,  q8
 | |
|         vrhadd.u8       q3,  q3,  q9
 | |
|         sub             r0,  r0,  r2
 | |
|   .endif
 | |
|         vst1.8          {q2},     [r0,:128], r2
 | |
|         vst1.8          {q3},     [r0,:128], r2
 | |
| 
 | |
|         bx              lr
 | |
| .endm
 | |
| 
 | |
| .macro  pixels16_xy2    rnd=1, avg=0
 | |
|         sub             r3,  r3,  #2
 | |
|         vld1.8          {d0-d2},  [r1], r2
 | |
|         vld1.8          {d4-d6},  [r1], r2
 | |
| NRND    vmov.i16        q13, #1
 | |
|         pld             [r1]
 | |
|         pld             [r1, r2]
 | |
|         vext.8          q1,  q0,  q1,  #1
 | |
|         vext.8          q3,  q2,  q3,  #1
 | |
|         vaddl.u8        q8,  d0,  d2
 | |
|         vaddl.u8        q10, d1,  d3
 | |
|         vaddl.u8        q9,  d4,  d6
 | |
|         vaddl.u8        q11, d5,  d7
 | |
| 1:      subs            r3,  r3,  #2
 | |
|         vld1.8          {d0-d2},  [r1], r2
 | |
|         vadd.u16        q12, q8,  q9
 | |
|         pld             [r1]
 | |
| NRND    vadd.u16        q12, q12, q13
 | |
|         vext.8          q15, q0,  q1,  #1
 | |
|         vadd.u16        q1 , q10, q11
 | |
|         shrn            d28, q12, #2
 | |
| NRND    vadd.u16        q1,  q1,  q13
 | |
|         shrn            d29, q1,  #2
 | |
|   .if \avg
 | |
|         vld1.8          {q8},     [r0,:128]
 | |
|         vrhadd.u8       q14, q14, q8
 | |
|   .endif
 | |
|         vaddl.u8        q8,  d0,  d30
 | |
|         vld1.8          {d2-d4},  [r1], r2
 | |
|         vaddl.u8        q10, d1,  d31
 | |
|         vst1.8          {q14},    [r0,:128], r2
 | |
|         vadd.u16        q12, q8,  q9
 | |
|         pld             [r1, r2]
 | |
| NRND    vadd.u16        q12, q12, q13
 | |
|         vext.8          q2,  q1,  q2,  #1
 | |
|         vadd.u16        q0,  q10, q11
 | |
|         shrn            d30, q12, #2
 | |
| NRND    vadd.u16        q0,  q0,  q13
 | |
|         shrn            d31, q0,  #2
 | |
|   .if \avg
 | |
|         vld1.8          {q9},     [r0,:128]
 | |
|         vrhadd.u8       q15, q15, q9
 | |
|   .endif
 | |
|         vaddl.u8        q9,  d2,  d4
 | |
|         vaddl.u8        q11, d3,  d5
 | |
|         vst1.8          {q15},    [r0,:128], r2
 | |
|         bgt             1b
 | |
| 
 | |
|         vld1.8          {d0-d2},  [r1], r2
 | |
|         vadd.u16        q12, q8,  q9
 | |
| NRND    vadd.u16        q12, q12, q13
 | |
|         vext.8          q15, q0,  q1,  #1
 | |
|         vadd.u16        q1 , q10, q11
 | |
|         shrn            d28, q12, #2
 | |
| NRND    vadd.u16        q1,  q1,  q13
 | |
|         shrn            d29, q1,  #2
 | |
|   .if \avg
 | |
|         vld1.8          {q8},     [r0,:128]
 | |
|         vrhadd.u8       q14, q14, q8
 | |
|   .endif
 | |
|         vaddl.u8        q8,  d0,  d30
 | |
|         vaddl.u8        q10, d1,  d31
 | |
|         vst1.8          {q14},    [r0,:128], r2
 | |
|         vadd.u16        q12, q8,  q9
 | |
| NRND    vadd.u16        q12, q12, q13
 | |
|         vadd.u16        q0,  q10, q11
 | |
|         shrn            d30, q12, #2
 | |
| NRND    vadd.u16        q0,  q0,  q13
 | |
|         shrn            d31, q0,  #2
 | |
|   .if \avg
 | |
|         vld1.8          {q9},     [r0,:128]
 | |
|         vrhadd.u8       q15, q15, q9
 | |
|   .endif
 | |
|         vst1.8          {q15},    [r0,:128], r2
 | |
| 
 | |
|         bx              lr
 | |
| .endm
 | |
| 
 | |
| .macro  pixels8         rnd=1, avg=0
 | |
| 1:      vld1.8          {d0},     [r1], r2
 | |
|         vld1.8          {d1},     [r1], r2
 | |
|         vld1.8          {d2},     [r1], r2
 | |
|         pld             [r1, r2, lsl #2]
 | |
|         vld1.8          {d3},     [r1], r2
 | |
|         pld             [r1]
 | |
|         pld             [r1, r2]
 | |
|         pld             [r1, r2, lsl #1]
 | |
|   .if \avg
 | |
|         vld1.8          {d4},     [r0,:64], r2
 | |
|         vrhadd.u8       d0,  d0,  d4
 | |
|         vld1.8          {d5},     [r0,:64], r2
 | |
|         vrhadd.u8       d1,  d1,  d5
 | |
|         vld1.8          {d6},     [r0,:64], r2
 | |
|         vrhadd.u8       d2,  d2,  d6
 | |
|         vld1.8          {d7},     [r0,:64], r2
 | |
|         vrhadd.u8       d3,  d3,  d7
 | |
|         sub             r0,  r0,  r2,  lsl #2
 | |
|   .endif
 | |
|         subs            r3,  r3,  #4
 | |
|         vst1.8          {d0},     [r0,:64], r2
 | |
|         vst1.8          {d1},     [r0,:64], r2
 | |
|         vst1.8          {d2},     [r0,:64], r2
 | |
|         vst1.8          {d3},     [r0,:64], r2
 | |
|         bne             1b
 | |
|         bx              lr
 | |
| .endm
 | |
| 
 | |
| .macro  pixels8_x2      rnd=1, avg=0
 | |
| 1:      vld1.8          {q0},     [r1], r2
 | |
|         vext.8          d1,  d0,  d1,  #1
 | |
|         vld1.8          {q1},     [r1], r2
 | |
|         vext.8          d3,  d2,  d3,  #1
 | |
|         pld             [r1]
 | |
|         pld             [r1, r2]
 | |
|         subs            r3,  r3,  #2
 | |
|         vswp            d1,  d2
 | |
|         avg             q0,  q0,  q1
 | |
|   .if \avg
 | |
|         vld1.8          {d4},     [r0,:64], r2
 | |
|         vld1.8          {d5},     [r0,:64]
 | |
|         vrhadd.u8       q0,  q0,  q2
 | |
|         sub             r0,  r0,  r2
 | |
|   .endif
 | |
|         vst1.8          {d0},     [r0,:64], r2
 | |
|         vst1.8          {d1},     [r0,:64], r2
 | |
|         bne             1b
 | |
|         bx              lr
 | |
| .endm
 | |
| 
 | |
| .macro  pixels8_y2      rnd=1, avg=0
 | |
|         sub             r3,  r3,  #2
 | |
|         vld1.8          {d0},     [r1], r2
 | |
|         vld1.8          {d1},     [r1], r2
 | |
| 1:      subs            r3,  r3,  #2
 | |
|         avg             d4,  d0,  d1
 | |
|         vld1.8          {d0},     [r1], r2
 | |
|         avg             d5,  d0,  d1
 | |
|         vld1.8          {d1},     [r1], r2
 | |
|         pld             [r1]
 | |
|         pld             [r1, r2]
 | |
|   .if \avg
 | |
|         vld1.8          {d2},     [r0,:64], r2
 | |
|         vld1.8          {d3},     [r0,:64]
 | |
|         vrhadd.u8       q2,  q2,  q1
 | |
|         sub             r0,  r0,  r2
 | |
|   .endif
 | |
|         vst1.8          {d4},     [r0,:64], r2
 | |
|         vst1.8          {d5},     [r0,:64], r2
 | |
|         bne             1b
 | |
| 
 | |
|         avg             d4,  d0,  d1
 | |
|         vld1.8          {d0},     [r1], r2
 | |
|         avg             d5,  d0,  d1
 | |
|   .if \avg
 | |
|         vld1.8          {d2},     [r0,:64], r2
 | |
|         vld1.8          {d3},     [r0,:64]
 | |
|         vrhadd.u8       q2,  q2,  q1
 | |
|         sub             r0,  r0,  r2
 | |
|   .endif
 | |
|         vst1.8          {d4},     [r0,:64], r2
 | |
|         vst1.8          {d5},     [r0,:64], r2
 | |
| 
 | |
|         bx              lr
 | |
| .endm
 | |
| 
 | |
| .macro  pixels8_xy2     rnd=1, avg=0
 | |
|         sub             r3,  r3,  #2
 | |
|         vld1.8          {q0},     [r1], r2
 | |
|         vld1.8          {q1},     [r1], r2
 | |
| NRND    vmov.i16        q11, #1
 | |
|         pld             [r1]
 | |
|         pld             [r1, r2]
 | |
|         vext.8          d4,  d0,  d1,  #1
 | |
|         vext.8          d6,  d2,  d3,  #1
 | |
|         vaddl.u8        q8,  d0,  d4
 | |
|         vaddl.u8        q9,  d2,  d6
 | |
| 1:      subs            r3,  r3,  #2
 | |
|         vld1.8          {q0},     [r1], r2
 | |
|         pld             [r1]
 | |
|         vadd.u16        q10, q8,  q9
 | |
|         vext.8          d4,  d0,  d1,  #1
 | |
| NRND    vadd.u16        q10, q10, q11
 | |
|         vaddl.u8        q8,  d0,  d4
 | |
|         shrn            d5,  q10, #2
 | |
|         vld1.8          {q1},     [r1], r2
 | |
|         vadd.u16        q10, q8,  q9
 | |
|         pld             [r1, r2]
 | |
|   .if \avg
 | |
|         vld1.8          {d7},     [r0,:64]
 | |
|         vrhadd.u8       d5,  d5,  d7
 | |
|   .endif
 | |
| NRND    vadd.u16        q10, q10, q11
 | |
|         vst1.8          {d5},     [r0,:64], r2
 | |
|         shrn            d7,  q10, #2
 | |
|   .if \avg
 | |
|         vld1.8          {d5},     [r0,:64]
 | |
|         vrhadd.u8       d7,  d7,  d5
 | |
|   .endif
 | |
|         vext.8          d6,  d2,  d3,  #1
 | |
|         vaddl.u8        q9,  d2,  d6
 | |
|         vst1.8          {d7},     [r0,:64], r2
 | |
|         bgt             1b
 | |
| 
 | |
|         vld1.8          {q0},     [r1], r2
 | |
|         vadd.u16        q10, q8,  q9
 | |
|         vext.8          d4,  d0,  d1,  #1
 | |
| NRND    vadd.u16        q10, q10, q11
 | |
|         vaddl.u8        q8,  d0,  d4
 | |
|         shrn            d5,  q10, #2
 | |
|         vadd.u16        q10, q8,  q9
 | |
|   .if \avg
 | |
|         vld1.8          {d7},     [r0,:64]
 | |
|         vrhadd.u8       d5,  d5,  d7
 | |
|   .endif
 | |
| NRND    vadd.u16        q10, q10, q11
 | |
|         vst1.8          {d5},     [r0,:64], r2
 | |
|         shrn            d7,  q10, #2
 | |
|   .if \avg
 | |
|         vld1.8          {d5},     [r0,:64]
 | |
|         vrhadd.u8       d7,  d7,  d5
 | |
|   .endif
 | |
|         vst1.8          {d7},     [r0,:64], r2
 | |
| 
 | |
|         bx              lr
 | |
| .endm
 | |
| 
 | |
| .macro  pixfunc         pfx, name, suf, rnd=1, avg=0
 | |
|   .if \rnd
 | |
|     .macro avg  rd, rn, rm
 | |
|         vrhadd.u8       \rd, \rn, \rm
 | |
|     .endm
 | |
|     .macro shrn rd, rn, rm
 | |
|         vrshrn.u16      \rd, \rn, \rm
 | |
|     .endm
 | |
|     .macro NRND insn:vararg
 | |
|     .endm
 | |
|   .else
 | |
|     .macro avg  rd, rn, rm
 | |
|         vhadd.u8        \rd, \rn, \rm
 | |
|     .endm
 | |
|     .macro shrn rd, rn, rm
 | |
|         vshrn.u16       \rd, \rn, \rm
 | |
|     .endm
 | |
|     .macro NRND insn:vararg
 | |
|         \insn
 | |
|     .endm
 | |
|   .endif
 | |
| function ff_\pfx\name\suf\()_neon, export=1
 | |
|         \name           \rnd, \avg
 | |
| endfunc
 | |
|         .purgem         avg
 | |
|         .purgem         shrn
 | |
|         .purgem         NRND
 | |
| .endm
 | |
| 
 | |
| .macro  pixfunc2        pfx, name, avg=0
 | |
|         pixfunc         \pfx, \name,          rnd=1, avg=\avg
 | |
|         pixfunc         \pfx, \name, _no_rnd, rnd=0, avg=\avg
 | |
| .endm
 | |
| 
 | |
| function ff_put_h264_qpel16_mc00_neon, export=1
 | |
|         mov             r3,  #16
 | |
| endfunc
 | |
| 
 | |
|         pixfunc         put_, pixels16,     avg=0
 | |
|         pixfunc2        put_, pixels16_x2,  avg=0
 | |
|         pixfunc2        put_, pixels16_y2,  avg=0
 | |
|         pixfunc2        put_, pixels16_xy2, avg=0
 | |
| 
 | |
| function ff_avg_h264_qpel16_mc00_neon, export=1
 | |
|         mov             r3,  #16
 | |
| endfunc
 | |
| 
 | |
|         pixfunc         avg_, pixels16,     avg=1
 | |
|         pixfunc2        avg_, pixels16_x2,  avg=1
 | |
|         pixfunc2        avg_, pixels16_y2,  avg=1
 | |
|         pixfunc2        avg_, pixels16_xy2, avg=1
 | |
| 
 | |
| function ff_put_h264_qpel8_mc00_neon, export=1
 | |
|         mov             r3,  #8
 | |
| endfunc
 | |
| 
 | |
|         pixfunc         put_, pixels8,     avg=0
 | |
|         pixfunc2        put_, pixels8_x2,  avg=0
 | |
|         pixfunc2        put_, pixels8_y2,  avg=0
 | |
|         pixfunc2        put_, pixels8_xy2, avg=0
 | |
| 
 | |
| function ff_avg_h264_qpel8_mc00_neon, export=1
 | |
|         mov             r3,  #8
 | |
| endfunc
 | |
| 
 | |
|         pixfunc         avg_, pixels8,     avg=1
 | |
|         pixfunc         avg_, pixels8_x2,  avg=1
 | |
|         pixfunc         avg_, pixels8_y2,  avg=1
 | |
|         pixfunc         avg_, pixels8_xy2, avg=1
 |