huffyuvdsp: move functions only used by huffyuv from lossless_videodsp
Signed-off-by: James Almer <jamrial@gmail.com>
This commit is contained in:
		
							parent
							
								
									cf9ef83960
								
							
						
					
					
						commit
						47f212329e
					
				@ -297,8 +297,8 @@ static av_cold int decode_init(AVCodecContext *avctx)
 | 
				
			|||||||
    if (ret < 0)
 | 
					    if (ret < 0)
 | 
				
			||||||
        return ret;
 | 
					        return ret;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    ff_huffyuvdsp_init(&s->hdsp);
 | 
					    ff_huffyuvdsp_init(&s->hdsp, avctx);
 | 
				
			||||||
    ff_llviddsp_init(&s->llviddsp, avctx);
 | 
					    ff_llviddsp_init(&s->llviddsp);
 | 
				
			||||||
    memset(s->vlc, 0, 4 * sizeof(VLC));
 | 
					    memset(s->vlc, 0, 4 * sizeof(VLC));
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    s->interlaced = avctx->height > 288;
 | 
					    s->interlaced = avctx->height > 288;
 | 
				
			||||||
@ -891,7 +891,7 @@ static void add_bytes(HYuvContext *s, uint8_t *dst, uint8_t *src, int w)
 | 
				
			|||||||
    if (s->bps <= 8) {
 | 
					    if (s->bps <= 8) {
 | 
				
			||||||
        s->llviddsp.add_bytes(dst, src, w);
 | 
					        s->llviddsp.add_bytes(dst, src, w);
 | 
				
			||||||
    } else {
 | 
					    } else {
 | 
				
			||||||
        s->llviddsp.add_int16((uint16_t*)dst, (const uint16_t*)src, s->n - 1, w);
 | 
					        s->hdsp.add_int16((uint16_t*)dst, (const uint16_t*)src, s->n - 1, w);
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@ -900,7 +900,7 @@ static void add_median_prediction(HYuvContext *s, uint8_t *dst, const uint8_t *s
 | 
				
			|||||||
    if (s->bps <= 8) {
 | 
					    if (s->bps <= 8) {
 | 
				
			||||||
        s->llviddsp.add_median_pred(dst, src, diff, w, left, left_top);
 | 
					        s->llviddsp.add_median_pred(dst, src, diff, w, left, left_top);
 | 
				
			||||||
    } else {
 | 
					    } else {
 | 
				
			||||||
        s->llviddsp.add_hfyu_median_pred_int16((uint16_t *)dst, (const uint16_t *)src, (const uint16_t *)diff, s->n-1, w, left, left_top);
 | 
					        s->hdsp.add_hfyu_median_pred_int16((uint16_t *)dst, (const uint16_t *)src, (const uint16_t *)diff, s->n-1, w, left, left_top);
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame,
 | 
					static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame,
 | 
				
			||||||
 | 
				
			|||||||
@ -23,6 +23,36 @@
 | 
				
			|||||||
#include "mathops.h"
 | 
					#include "mathops.h"
 | 
				
			||||||
#include "huffyuvdsp.h"
 | 
					#include "huffyuvdsp.h"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static void add_int16_c(uint16_t *dst, const uint16_t *src, unsigned mask, int w){
 | 
				
			||||||
 | 
					    long i;
 | 
				
			||||||
 | 
					    unsigned long pw_lsb = (mask >> 1) * 0x0001000100010001ULL;
 | 
				
			||||||
 | 
					    unsigned long pw_msb = pw_lsb +  0x0001000100010001ULL;
 | 
				
			||||||
 | 
					    for (i = 0; i <= w - (int)sizeof(long)/2; i += sizeof(long)/2) {
 | 
				
			||||||
 | 
					        long a = *(long*)(src+i);
 | 
				
			||||||
 | 
					        long b = *(long*)(dst+i);
 | 
				
			||||||
 | 
					        *(long*)(dst+i) = ((a&pw_lsb) + (b&pw_lsb)) ^ ((a^b)&pw_msb);
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					    for(; i<w; i++)
 | 
				
			||||||
 | 
					        dst[i] = (dst[i] + src[i]) & mask;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					static void add_hfyu_median_pred_int16_c(uint16_t *dst, const uint16_t *src, const uint16_t *diff, unsigned mask, int w, int *left, int *left_top){
 | 
				
			||||||
 | 
					    int i;
 | 
				
			||||||
 | 
					    uint16_t l, lt;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    l  = *left;
 | 
				
			||||||
 | 
					    lt = *left_top;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    for(i=0; i<w; i++){
 | 
				
			||||||
 | 
					        l  = (mid_pred(l, src[i], (l + src[i] - lt) & mask) + diff[i]) & mask;
 | 
				
			||||||
 | 
					        lt = src[i];
 | 
				
			||||||
 | 
					        dst[i] = l;
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    *left     = l;
 | 
				
			||||||
 | 
					    *left_top = lt;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static void add_hfyu_left_pred_bgr32_c(uint8_t *dst, const uint8_t *src,
 | 
					static void add_hfyu_left_pred_bgr32_c(uint8_t *dst, const uint8_t *src,
 | 
				
			||||||
                                       intptr_t w, uint8_t *left)
 | 
					                                       intptr_t w, uint8_t *left)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
@ -47,10 +77,12 @@ static void add_hfyu_left_pred_bgr32_c(uint8_t *dst, const uint8_t *src,
 | 
				
			|||||||
    left[A] = a;
 | 
					    left[A] = a;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
av_cold void ff_huffyuvdsp_init(HuffYUVDSPContext *c)
 | 
					av_cold void ff_huffyuvdsp_init(HuffYUVDSPContext *c, AVCodecContext *avctx)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
 | 
					    c->add_int16 = add_int16_c;
 | 
				
			||||||
 | 
					    c->add_hfyu_median_pred_int16 = add_hfyu_median_pred_int16_c;
 | 
				
			||||||
    c->add_hfyu_left_pred_bgr32 = add_hfyu_left_pred_bgr32_c;
 | 
					    c->add_hfyu_left_pred_bgr32 = add_hfyu_left_pred_bgr32_c;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if (ARCH_X86)
 | 
					    if (ARCH_X86)
 | 
				
			||||||
        ff_huffyuvdsp_init_x86(c);
 | 
					        ff_huffyuvdsp_init_x86(c, avctx);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
				
			|||||||
@ -21,6 +21,7 @@
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
#include <stdint.h>
 | 
					#include <stdint.h>
 | 
				
			||||||
#include "config.h"
 | 
					#include "config.h"
 | 
				
			||||||
 | 
					#include "avcodec.h"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#if HAVE_BIGENDIAN
 | 
					#if HAVE_BIGENDIAN
 | 
				
			||||||
#define B 3
 | 
					#define B 3
 | 
				
			||||||
@ -35,12 +36,18 @@
 | 
				
			|||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
typedef struct HuffYUVDSPContext {
 | 
					typedef struct HuffYUVDSPContext {
 | 
				
			||||||
 | 
					    void (*add_int16)(uint16_t *dst/*align 16*/, const uint16_t *src/*align 16*/,
 | 
				
			||||||
 | 
					                      unsigned mask, int w);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    void (*add_hfyu_median_pred_int16)(uint16_t *dst, const uint16_t *top,
 | 
				
			||||||
 | 
					                                       const uint16_t *diff, unsigned mask,
 | 
				
			||||||
 | 
					                                       int w, int *left, int *left_top);
 | 
				
			||||||
    void (*add_hfyu_left_pred_bgr32)(uint8_t *dst, const uint8_t *src,
 | 
					    void (*add_hfyu_left_pred_bgr32)(uint8_t *dst, const uint8_t *src,
 | 
				
			||||||
                                     intptr_t w, uint8_t *left);
 | 
					                                     intptr_t w, uint8_t *left);
 | 
				
			||||||
} HuffYUVDSPContext;
 | 
					} HuffYUVDSPContext;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void ff_huffyuvdsp_init(HuffYUVDSPContext *c);
 | 
					void ff_huffyuvdsp_init(HuffYUVDSPContext *c, AVCodecContext *avctx);
 | 
				
			||||||
void ff_huffyuvdsp_init_ppc(HuffYUVDSPContext *c);
 | 
					void ff_huffyuvdsp_init_ppc(HuffYUVDSPContext *c, AVCodecContext *avctx);
 | 
				
			||||||
void ff_huffyuvdsp_init_x86(HuffYUVDSPContext *c);
 | 
					void ff_huffyuvdsp_init_x86(HuffYUVDSPContext *c, AVCodecContext *avctx);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#endif /* AVCODEC_HUFFYUVDSP_H */
 | 
					#endif /* AVCODEC_HUFFYUVDSP_H */
 | 
				
			||||||
 | 
				
			|||||||
@ -725,7 +725,7 @@ static av_cold int lag_decode_init(AVCodecContext *avctx)
 | 
				
			|||||||
    LagarithContext *l = avctx->priv_data;
 | 
					    LagarithContext *l = avctx->priv_data;
 | 
				
			||||||
    l->avctx = avctx;
 | 
					    l->avctx = avctx;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    ff_llviddsp_init(&l->llviddsp, avctx);
 | 
					    ff_llviddsp_init(&l->llviddsp);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    return 0;
 | 
					    return 0;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
				
			|||||||
@ -79,36 +79,6 @@ static int add_left_pred_c(uint8_t *dst, const uint8_t *src, intptr_t w,
 | 
				
			|||||||
    return acc;
 | 
					    return acc;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static void add_int16_c(uint16_t *dst, const uint16_t *src, unsigned mask, int w){
 | 
					 | 
				
			||||||
    long i;
 | 
					 | 
				
			||||||
    unsigned long pw_lsb = (mask >> 1) * 0x0001000100010001ULL;
 | 
					 | 
				
			||||||
    unsigned long pw_msb = pw_lsb +  0x0001000100010001ULL;
 | 
					 | 
				
			||||||
    for (i = 0; i <= w - (int)sizeof(long)/2; i += sizeof(long)/2) {
 | 
					 | 
				
			||||||
        long a = *(long*)(src+i);
 | 
					 | 
				
			||||||
        long b = *(long*)(dst+i);
 | 
					 | 
				
			||||||
        *(long*)(dst+i) = ((a&pw_lsb) + (b&pw_lsb)) ^ ((a^b)&pw_msb);
 | 
					 | 
				
			||||||
    }
 | 
					 | 
				
			||||||
    for(; i<w; i++)
 | 
					 | 
				
			||||||
        dst[i] = (dst[i] + src[i]) & mask;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static void add_hfyu_median_pred_int16_c(uint16_t *dst, const uint16_t *src, const uint16_t *diff, unsigned mask, int w, int *left, int *left_top){
 | 
					 | 
				
			||||||
    int i;
 | 
					 | 
				
			||||||
    uint16_t l, lt;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
    l  = *left;
 | 
					 | 
				
			||||||
    lt = *left_top;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
    for(i=0; i<w; i++){
 | 
					 | 
				
			||||||
        l  = (mid_pred(l, src[i], (l + src[i] - lt) & mask) + diff[i]) & mask;
 | 
					 | 
				
			||||||
        lt = src[i];
 | 
					 | 
				
			||||||
        dst[i] = l;
 | 
					 | 
				
			||||||
    }
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
    *left     = l;
 | 
					 | 
				
			||||||
    *left_top = lt;
 | 
					 | 
				
			||||||
}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
static int add_hfyu_left_pred_int16_c(uint16_t *dst, const uint16_t *src, unsigned mask, int w, unsigned acc){
 | 
					static int add_hfyu_left_pred_int16_c(uint16_t *dst, const uint16_t *src, unsigned mask, int w, unsigned acc){
 | 
				
			||||||
    int i;
 | 
					    int i;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@ -129,16 +99,14 @@ static int add_hfyu_left_pred_int16_c(uint16_t *dst, const uint16_t *src, unsign
 | 
				
			|||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void ff_llviddsp_init(LLVidDSPContext *c, AVCodecContext *avctx)
 | 
					void ff_llviddsp_init(LLVidDSPContext *c)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
    c->add_bytes                  = add_bytes_c;
 | 
					    c->add_bytes                  = add_bytes_c;
 | 
				
			||||||
    c->add_median_pred            = add_median_pred_c;
 | 
					    c->add_median_pred            = add_median_pred_c;
 | 
				
			||||||
    c->add_left_pred              = add_left_pred_c;
 | 
					    c->add_left_pred              = add_left_pred_c;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    c->add_int16 = add_int16_c;
 | 
					 | 
				
			||||||
    c->add_hfyu_left_pred_int16   = add_hfyu_left_pred_int16_c;
 | 
					    c->add_hfyu_left_pred_int16   = add_hfyu_left_pred_int16_c;
 | 
				
			||||||
    c->add_hfyu_median_pred_int16 = add_hfyu_median_pred_int16_c;
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if (ARCH_X86)
 | 
					    if (ARCH_X86)
 | 
				
			||||||
        ff_llviddsp_init_x86(c, avctx);
 | 
					        ff_llviddsp_init_x86(c);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
				
			|||||||
@ -34,14 +34,11 @@ typedef struct LLVidDSPContext {
 | 
				
			|||||||
    int (*add_left_pred)(uint8_t *dst, const uint8_t *src,
 | 
					    int (*add_left_pred)(uint8_t *dst, const uint8_t *src,
 | 
				
			||||||
                         intptr_t w, int left);
 | 
					                         intptr_t w, int left);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    void (*add_int16)(uint16_t *dst/*align 16*/, const uint16_t *src/*align 16*/, unsigned mask, int w);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
    void (*add_hfyu_median_pred_int16)(uint16_t *dst, const uint16_t *top, const uint16_t *diff, unsigned mask, int w, int *left, int *left_top);
 | 
					 | 
				
			||||||
    int  (*add_hfyu_left_pred_int16)(uint16_t *dst, const uint16_t *src, unsigned mask, int w, unsigned left);
 | 
					    int  (*add_hfyu_left_pred_int16)(uint16_t *dst, const uint16_t *src, unsigned mask, int w, unsigned left);
 | 
				
			||||||
} LLVidDSPContext;
 | 
					} LLVidDSPContext;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void ff_llviddsp_init(LLVidDSPContext *llviddsp, AVCodecContext *avctx);
 | 
					void ff_llviddsp_init(LLVidDSPContext *llviddsp);
 | 
				
			||||||
void ff_llviddsp_init_x86(LLVidDSPContext *llviddsp, AVCodecContext *avctx);
 | 
					void ff_llviddsp_init_x86(LLVidDSPContext *llviddsp);
 | 
				
			||||||
void ff_llviddsp_init_ppc(LLVidDSPContext *llviddsp, AVCodecContext *avctx);
 | 
					void ff_llviddsp_init_ppc(LLVidDSPContext *llviddsp);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#endif //AVCODEC_LOSSLESS_VIDEODSP_H
 | 
					#endif //AVCODEC_LOSSLESS_VIDEODSP_H
 | 
				
			||||||
 | 
				
			|||||||
@ -697,7 +697,7 @@ static int magy_init_thread_copy(AVCodecContext *avctx)
 | 
				
			|||||||
static av_cold int magy_decode_init(AVCodecContext *avctx)
 | 
					static av_cold int magy_decode_init(AVCodecContext *avctx)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
    MagicYUVContext *s = avctx->priv_data;
 | 
					    MagicYUVContext *s = avctx->priv_data;
 | 
				
			||||||
    ff_llviddsp_init(&s->llviddsp, avctx);
 | 
					    ff_llviddsp_init(&s->llviddsp);
 | 
				
			||||||
    return 0;
 | 
					    return 0;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
				
			|||||||
@ -51,7 +51,7 @@ static void add_bytes_altivec(uint8_t *dst, uint8_t *src, intptr_t w)
 | 
				
			|||||||
}
 | 
					}
 | 
				
			||||||
#endif /* HAVE_ALTIVEC */
 | 
					#endif /* HAVE_ALTIVEC */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
av_cold void ff_llviddsp_init_ppc(LLVidDSPContext *c, AVCodecContext *avctx)
 | 
					av_cold void ff_llviddsp_init_ppc(LLVidDSPContext *c)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
#if HAVE_ALTIVEC
 | 
					#if HAVE_ALTIVEC
 | 
				
			||||||
    if (!PPC_ALTIVEC(av_get_cpu_flags()))
 | 
					    if (!PPC_ALTIVEC(av_get_cpu_flags()))
 | 
				
			||||||
 | 
				
			|||||||
@ -827,7 +827,7 @@ static av_cold int decode_init(AVCodecContext *avctx)
 | 
				
			|||||||
    c->avctx = avctx;
 | 
					    c->avctx = avctx;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    ff_bswapdsp_init(&c->bdsp);
 | 
					    ff_bswapdsp_init(&c->bdsp);
 | 
				
			||||||
    ff_llviddsp_init(&c->llviddsp, avctx);
 | 
					    ff_llviddsp_init(&c->llviddsp);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if (avctx->extradata_size >= 16) {
 | 
					    if (avctx->extradata_size >= 16) {
 | 
				
			||||||
        av_log(avctx, AV_LOG_DEBUG, "Encoder version %d.%d.%d.%d\n",
 | 
					        av_log(avctx, AV_LOG_DEBUG, "Encoder version %d.%d.%d.%d\n",
 | 
				
			||||||
 | 
				
			|||||||
@ -185,7 +185,7 @@ static av_cold int vble_decode_init(AVCodecContext *avctx)
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
    /* Stash for later use */
 | 
					    /* Stash for later use */
 | 
				
			||||||
    ctx->avctx = avctx;
 | 
					    ctx->avctx = avctx;
 | 
				
			||||||
    ff_llviddsp_init(&ctx->llviddsp, avctx);
 | 
					    ff_llviddsp_init(&ctx->llviddsp);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    avctx->pix_fmt = AV_PIX_FMT_YUV420P;
 | 
					    avctx->pix_fmt = AV_PIX_FMT_YUV420P;
 | 
				
			||||||
    avctx->bits_per_raw_sample = 8;
 | 
					    avctx->bits_per_raw_sample = 8;
 | 
				
			||||||
 | 
				
			|||||||
@ -24,6 +24,78 @@
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
SECTION .text
 | 
					SECTION .text
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					%macro INT16_LOOP 2 ; %1 = a/u (aligned/unaligned), %2 = add/sub
 | 
				
			||||||
 | 
					    movd    m4, maskd
 | 
				
			||||||
 | 
					    SPLATW  m4, m4
 | 
				
			||||||
 | 
					    add     wd, wd
 | 
				
			||||||
 | 
					    test    wq, 2*mmsize - 1
 | 
				
			||||||
 | 
					    jz %%.tomainloop
 | 
				
			||||||
 | 
					    push  tmpq
 | 
				
			||||||
 | 
					%%.wordloop:
 | 
				
			||||||
 | 
					    sub     wq, 2
 | 
				
			||||||
 | 
					%ifidn %2, add
 | 
				
			||||||
 | 
					    mov   tmpw, [srcq+wq]
 | 
				
			||||||
 | 
					    add   tmpw, [dstq+wq]
 | 
				
			||||||
 | 
					%else
 | 
				
			||||||
 | 
					    mov   tmpw, [src1q+wq]
 | 
				
			||||||
 | 
					    sub   tmpw, [src2q+wq]
 | 
				
			||||||
 | 
					%endif
 | 
				
			||||||
 | 
					    and   tmpw, maskw
 | 
				
			||||||
 | 
					    mov     [dstq+wq], tmpw
 | 
				
			||||||
 | 
					    test    wq, 2*mmsize - 1
 | 
				
			||||||
 | 
					    jnz %%.wordloop
 | 
				
			||||||
 | 
					    pop   tmpq
 | 
				
			||||||
 | 
					%%.tomainloop:
 | 
				
			||||||
 | 
					%ifidn %2, add
 | 
				
			||||||
 | 
					    add     srcq, wq
 | 
				
			||||||
 | 
					%else
 | 
				
			||||||
 | 
					    add     src1q, wq
 | 
				
			||||||
 | 
					    add     src2q, wq
 | 
				
			||||||
 | 
					%endif
 | 
				
			||||||
 | 
					    add     dstq, wq
 | 
				
			||||||
 | 
					    neg     wq
 | 
				
			||||||
 | 
					    jz      %%.end
 | 
				
			||||||
 | 
					%%.loop:
 | 
				
			||||||
 | 
					%ifidn %2, add
 | 
				
			||||||
 | 
					    mov%1   m0, [srcq+wq]
 | 
				
			||||||
 | 
					    mov%1   m1, [dstq+wq]
 | 
				
			||||||
 | 
					    mov%1   m2, [srcq+wq+mmsize]
 | 
				
			||||||
 | 
					    mov%1   m3, [dstq+wq+mmsize]
 | 
				
			||||||
 | 
					%else
 | 
				
			||||||
 | 
					    mov%1   m0, [src1q+wq]
 | 
				
			||||||
 | 
					    mov%1   m1, [src2q+wq]
 | 
				
			||||||
 | 
					    mov%1   m2, [src1q+wq+mmsize]
 | 
				
			||||||
 | 
					    mov%1   m3, [src2q+wq+mmsize]
 | 
				
			||||||
 | 
					%endif
 | 
				
			||||||
 | 
					    p%2w    m0, m1
 | 
				
			||||||
 | 
					    p%2w    m2, m3
 | 
				
			||||||
 | 
					    pand    m0, m4
 | 
				
			||||||
 | 
					    pand    m2, m4
 | 
				
			||||||
 | 
					    mov%1   [dstq+wq]       , m0
 | 
				
			||||||
 | 
					    mov%1   [dstq+wq+mmsize], m2
 | 
				
			||||||
 | 
					    add     wq, 2*mmsize
 | 
				
			||||||
 | 
					    jl %%.loop
 | 
				
			||||||
 | 
					%%.end:
 | 
				
			||||||
 | 
					    RET
 | 
				
			||||||
 | 
					%endmacro
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					%if ARCH_X86_32
 | 
				
			||||||
 | 
					INIT_MMX mmx
 | 
				
			||||||
 | 
					cglobal add_int16, 4,4,5, dst, src, mask, w, tmp
 | 
				
			||||||
 | 
					    INT16_LOOP a, add
 | 
				
			||||||
 | 
					%endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					INIT_XMM sse2
 | 
				
			||||||
 | 
					cglobal add_int16, 4,4,5, dst, src, mask, w, tmp
 | 
				
			||||||
 | 
					    test srcq, mmsize-1
 | 
				
			||||||
 | 
					    jnz .unaligned
 | 
				
			||||||
 | 
					    test dstq, mmsize-1
 | 
				
			||||||
 | 
					    jnz .unaligned
 | 
				
			||||||
 | 
					    INT16_LOOP a, add
 | 
				
			||||||
 | 
					.unaligned:
 | 
				
			||||||
 | 
					    INT16_LOOP u, add
 | 
				
			||||||
 | 
					
 | 
				
			||||||
; void add_hfyu_left_pred_bgr32(uint8_t *dst, const uint8_t *src,
 | 
					; void add_hfyu_left_pred_bgr32(uint8_t *dst, const uint8_t *src,
 | 
				
			||||||
;                               intptr_t w, uint8_t *left)
 | 
					;                               intptr_t w, uint8_t *left)
 | 
				
			||||||
%macro LEFT_BGR32 0
 | 
					%macro LEFT_BGR32 0
 | 
				
			||||||
@ -63,3 +135,68 @@ LEFT_BGR32
 | 
				
			|||||||
%endif
 | 
					%endif
 | 
				
			||||||
INIT_XMM sse2
 | 
					INIT_XMM sse2
 | 
				
			||||||
LEFT_BGR32
 | 
					LEFT_BGR32
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					; void add_hfyu_median_prediction_mmxext(uint8_t *dst, const uint8_t *top, const uint8_t *diff, int mask, int w, int *left, int *left_top)
 | 
				
			||||||
 | 
					INIT_MMX mmxext
 | 
				
			||||||
 | 
					cglobal add_hfyu_median_pred_int16, 7,7,0, dst, top, diff, mask, w, left, left_top
 | 
				
			||||||
 | 
					    add      wd, wd
 | 
				
			||||||
 | 
					    movd    mm6, maskd
 | 
				
			||||||
 | 
					    SPLATW  mm6, mm6
 | 
				
			||||||
 | 
					    movq    mm0, [topq]
 | 
				
			||||||
 | 
					    movq    mm2, mm0
 | 
				
			||||||
 | 
					    movd    mm4, [left_topq]
 | 
				
			||||||
 | 
					    psllq   mm2, 16
 | 
				
			||||||
 | 
					    movq    mm1, mm0
 | 
				
			||||||
 | 
					    por     mm4, mm2
 | 
				
			||||||
 | 
					    movd    mm3, [leftq]
 | 
				
			||||||
 | 
					    psubw   mm0, mm4 ; t-tl
 | 
				
			||||||
 | 
					    add    dstq, wq
 | 
				
			||||||
 | 
					    add    topq, wq
 | 
				
			||||||
 | 
					    add   diffq, wq
 | 
				
			||||||
 | 
					    neg      wq
 | 
				
			||||||
 | 
					    jmp .skip
 | 
				
			||||||
 | 
					.loop:
 | 
				
			||||||
 | 
					    movq    mm4, [topq+wq]
 | 
				
			||||||
 | 
					    movq    mm0, mm4
 | 
				
			||||||
 | 
					    psllq   mm4, 16
 | 
				
			||||||
 | 
					    por     mm4, mm1
 | 
				
			||||||
 | 
					    movq    mm1, mm0 ; t
 | 
				
			||||||
 | 
					    psubw   mm0, mm4 ; t-tl
 | 
				
			||||||
 | 
					.skip:
 | 
				
			||||||
 | 
					    movq    mm2, [diffq+wq]
 | 
				
			||||||
 | 
					%assign i 0
 | 
				
			||||||
 | 
					%rep 4
 | 
				
			||||||
 | 
					    movq    mm4, mm0
 | 
				
			||||||
 | 
					    paddw   mm4, mm3 ; t-tl+l
 | 
				
			||||||
 | 
					    pand    mm4, mm6
 | 
				
			||||||
 | 
					    movq    mm5, mm3
 | 
				
			||||||
 | 
					    pmaxsw  mm3, mm1
 | 
				
			||||||
 | 
					    pminsw  mm5, mm1
 | 
				
			||||||
 | 
					    pminsw  mm3, mm4
 | 
				
			||||||
 | 
					    pmaxsw  mm3, mm5 ; median
 | 
				
			||||||
 | 
					    paddw   mm3, mm2 ; +residual
 | 
				
			||||||
 | 
					    pand    mm3, mm6
 | 
				
			||||||
 | 
					%if i==0
 | 
				
			||||||
 | 
					    movq    mm7, mm3
 | 
				
			||||||
 | 
					    psllq   mm7, 48
 | 
				
			||||||
 | 
					%else
 | 
				
			||||||
 | 
					    movq    mm4, mm3
 | 
				
			||||||
 | 
					    psrlq   mm7, 16
 | 
				
			||||||
 | 
					    psllq   mm4, 48
 | 
				
			||||||
 | 
					    por     mm7, mm4
 | 
				
			||||||
 | 
					%endif
 | 
				
			||||||
 | 
					%if i<3
 | 
				
			||||||
 | 
					    psrlq   mm0, 16
 | 
				
			||||||
 | 
					    psrlq   mm1, 16
 | 
				
			||||||
 | 
					    psrlq   mm2, 16
 | 
				
			||||||
 | 
					%endif
 | 
				
			||||||
 | 
					%assign i i+1
 | 
				
			||||||
 | 
					%endrep
 | 
				
			||||||
 | 
					    movq [dstq+wq], mm7
 | 
				
			||||||
 | 
					    add      wq, 8
 | 
				
			||||||
 | 
					    jl .loop
 | 
				
			||||||
 | 
					    movzx   r2d, word [dstq-2]
 | 
				
			||||||
 | 
					    mov [leftq], r2d
 | 
				
			||||||
 | 
					    movzx   r2d, word [topq-2]
 | 
				
			||||||
 | 
					    mov [left_topq], r2d
 | 
				
			||||||
 | 
					    RET
 | 
				
			||||||
 | 
				
			|||||||
@ -21,24 +21,35 @@
 | 
				
			|||||||
#include "config.h"
 | 
					#include "config.h"
 | 
				
			||||||
#include "libavutil/attributes.h"
 | 
					#include "libavutil/attributes.h"
 | 
				
			||||||
#include "libavutil/cpu.h"
 | 
					#include "libavutil/cpu.h"
 | 
				
			||||||
 | 
					#include "libavutil/pixdesc.h"
 | 
				
			||||||
#include "libavutil/x86/asm.h"
 | 
					#include "libavutil/x86/asm.h"
 | 
				
			||||||
#include "libavutil/x86/cpu.h"
 | 
					#include "libavutil/x86/cpu.h"
 | 
				
			||||||
#include "libavcodec/huffyuvdsp.h"
 | 
					#include "libavcodec/huffyuvdsp.h"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					void ff_add_int16_mmx(uint16_t *dst, const uint16_t *src, unsigned mask, int w);
 | 
				
			||||||
 | 
					void ff_add_int16_sse2(uint16_t *dst, const uint16_t *src, unsigned mask, int w);
 | 
				
			||||||
void ff_add_hfyu_left_pred_bgr32_mmx(uint8_t *dst, const uint8_t *src,
 | 
					void ff_add_hfyu_left_pred_bgr32_mmx(uint8_t *dst, const uint8_t *src,
 | 
				
			||||||
                                     intptr_t w, uint8_t *left);
 | 
					                                     intptr_t w, uint8_t *left);
 | 
				
			||||||
void ff_add_hfyu_left_pred_bgr32_sse2(uint8_t *dst, const uint8_t *src,
 | 
					void ff_add_hfyu_left_pred_bgr32_sse2(uint8_t *dst, const uint8_t *src,
 | 
				
			||||||
                                      intptr_t w, uint8_t *left);
 | 
					                                      intptr_t w, uint8_t *left);
 | 
				
			||||||
 | 
					void ff_add_hfyu_median_pred_int16_mmxext(uint16_t *dst, const uint16_t *top, const uint16_t *diff, unsigned mask, int w, int *left, int *left_top);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
av_cold void ff_huffyuvdsp_init_x86(HuffYUVDSPContext *c)
 | 
					av_cold void ff_huffyuvdsp_init_x86(HuffYUVDSPContext *c, AVCodecContext *avctx)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
    int cpu_flags = av_get_cpu_flags();
 | 
					    int cpu_flags = av_get_cpu_flags();
 | 
				
			||||||
 | 
					    const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(avctx->pix_fmt);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if (ARCH_X86_32 && EXTERNAL_MMX(cpu_flags)) {
 | 
					    if (ARCH_X86_32 && EXTERNAL_MMX(cpu_flags)) {
 | 
				
			||||||
        c->add_hfyu_left_pred_bgr32 = ff_add_hfyu_left_pred_bgr32_mmx;
 | 
					        c->add_hfyu_left_pred_bgr32 = ff_add_hfyu_left_pred_bgr32_mmx;
 | 
				
			||||||
 | 
					        c->add_int16 = ff_add_int16_mmx;
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    if (EXTERNAL_MMXEXT(cpu_flags) && pix_desc && pix_desc->comp[0].depth<16) {
 | 
				
			||||||
 | 
					        c->add_hfyu_median_pred_int16 = ff_add_hfyu_median_pred_int16_mmxext;
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if (EXTERNAL_SSE2(cpu_flags)) {
 | 
					    if (EXTERNAL_SSE2(cpu_flags)) {
 | 
				
			||||||
 | 
					        c->add_int16 = ff_add_int16_sse2;
 | 
				
			||||||
        c->add_hfyu_left_pred_bgr32 = ff_add_hfyu_left_pred_bgr32_sse2;
 | 
					        c->add_hfyu_left_pred_bgr32 = ff_add_hfyu_left_pred_bgr32_sse2;
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
				
			|||||||
@ -217,77 +217,6 @@ ADD_BYTES
 | 
				
			|||||||
INIT_XMM sse2
 | 
					INIT_XMM sse2
 | 
				
			||||||
ADD_BYTES
 | 
					ADD_BYTES
 | 
				
			||||||
 | 
					
 | 
				
			||||||
%macro INT16_LOOP 2 ; %1 = a/u (aligned/unaligned), %2 = add/sub
 | 
					 | 
				
			||||||
    movd    m4, maskd
 | 
					 | 
				
			||||||
    SPLATW  m4, m4
 | 
					 | 
				
			||||||
    add     wd, wd
 | 
					 | 
				
			||||||
    test    wq, 2*mmsize - 1
 | 
					 | 
				
			||||||
    jz %%.tomainloop
 | 
					 | 
				
			||||||
    push  tmpq
 | 
					 | 
				
			||||||
%%.wordloop:
 | 
					 | 
				
			||||||
    sub     wq, 2
 | 
					 | 
				
			||||||
%ifidn %2, add
 | 
					 | 
				
			||||||
    mov   tmpw, [srcq+wq]
 | 
					 | 
				
			||||||
    add   tmpw, [dstq+wq]
 | 
					 | 
				
			||||||
%else
 | 
					 | 
				
			||||||
    mov   tmpw, [src1q+wq]
 | 
					 | 
				
			||||||
    sub   tmpw, [src2q+wq]
 | 
					 | 
				
			||||||
%endif
 | 
					 | 
				
			||||||
    and   tmpw, maskw
 | 
					 | 
				
			||||||
    mov     [dstq+wq], tmpw
 | 
					 | 
				
			||||||
    test    wq, 2*mmsize - 1
 | 
					 | 
				
			||||||
    jnz %%.wordloop
 | 
					 | 
				
			||||||
    pop   tmpq
 | 
					 | 
				
			||||||
%%.tomainloop:
 | 
					 | 
				
			||||||
%ifidn %2, add
 | 
					 | 
				
			||||||
    add     srcq, wq
 | 
					 | 
				
			||||||
%else
 | 
					 | 
				
			||||||
    add     src1q, wq
 | 
					 | 
				
			||||||
    add     src2q, wq
 | 
					 | 
				
			||||||
%endif
 | 
					 | 
				
			||||||
    add     dstq, wq
 | 
					 | 
				
			||||||
    neg     wq
 | 
					 | 
				
			||||||
    jz      %%.end
 | 
					 | 
				
			||||||
%%.loop:
 | 
					 | 
				
			||||||
%ifidn %2, add
 | 
					 | 
				
			||||||
    mov%1   m0, [srcq+wq]
 | 
					 | 
				
			||||||
    mov%1   m1, [dstq+wq]
 | 
					 | 
				
			||||||
    mov%1   m2, [srcq+wq+mmsize]
 | 
					 | 
				
			||||||
    mov%1   m3, [dstq+wq+mmsize]
 | 
					 | 
				
			||||||
%else
 | 
					 | 
				
			||||||
    mov%1   m0, [src1q+wq]
 | 
					 | 
				
			||||||
    mov%1   m1, [src2q+wq]
 | 
					 | 
				
			||||||
    mov%1   m2, [src1q+wq+mmsize]
 | 
					 | 
				
			||||||
    mov%1   m3, [src2q+wq+mmsize]
 | 
					 | 
				
			||||||
%endif
 | 
					 | 
				
			||||||
    p%2w    m0, m1
 | 
					 | 
				
			||||||
    p%2w    m2, m3
 | 
					 | 
				
			||||||
    pand    m0, m4
 | 
					 | 
				
			||||||
    pand    m2, m4
 | 
					 | 
				
			||||||
    mov%1   [dstq+wq]       , m0
 | 
					 | 
				
			||||||
    mov%1   [dstq+wq+mmsize], m2
 | 
					 | 
				
			||||||
    add     wq, 2*mmsize
 | 
					 | 
				
			||||||
    jl %%.loop
 | 
					 | 
				
			||||||
%%.end:
 | 
					 | 
				
			||||||
    RET
 | 
					 | 
				
			||||||
%endmacro
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
%if ARCH_X86_32
 | 
					 | 
				
			||||||
INIT_MMX mmx
 | 
					 | 
				
			||||||
cglobal add_int16, 4,4,5, dst, src, mask, w, tmp
 | 
					 | 
				
			||||||
    INT16_LOOP a, add
 | 
					 | 
				
			||||||
%endif
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
INIT_XMM sse2
 | 
					 | 
				
			||||||
cglobal add_int16, 4,4,5, dst, src, mask, w, tmp
 | 
					 | 
				
			||||||
    test srcq, mmsize-1
 | 
					 | 
				
			||||||
    jnz .unaligned
 | 
					 | 
				
			||||||
    test dstq, mmsize-1
 | 
					 | 
				
			||||||
    jnz .unaligned
 | 
					 | 
				
			||||||
    INT16_LOOP a, add
 | 
					 | 
				
			||||||
.unaligned:
 | 
					 | 
				
			||||||
    INT16_LOOP u, add
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
%macro ADD_HFYU_LEFT_LOOP_INT16 2 ; %1 = dst alignment (a/u), %2 = src alignment (a/u)
 | 
					%macro ADD_HFYU_LEFT_LOOP_INT16 2 ; %1 = dst alignment (a/u), %2 = src alignment (a/u)
 | 
				
			||||||
    add     wd, wd
 | 
					    add     wd, wd
 | 
				
			||||||
    add     srcq, wq
 | 
					    add     srcq, wq
 | 
				
			||||||
@ -359,68 +288,3 @@ cglobal add_hfyu_left_pred_int16, 4,4,8, dst, src, mask, w, left
 | 
				
			|||||||
    ADD_HFYU_LEFT_LOOP_INT16 u, a
 | 
					    ADD_HFYU_LEFT_LOOP_INT16 u, a
 | 
				
			||||||
.src_unaligned:
 | 
					.src_unaligned:
 | 
				
			||||||
    ADD_HFYU_LEFT_LOOP_INT16 u, u
 | 
					    ADD_HFYU_LEFT_LOOP_INT16 u, u
 | 
				
			||||||
 | 
					 | 
				
			||||||
; void add_hfyu_median_prediction_mmxext(uint8_t *dst, const uint8_t *top, const uint8_t *diff, int mask, int w, int *left, int *left_top)
 | 
					 | 
				
			||||||
INIT_MMX mmxext
 | 
					 | 
				
			||||||
cglobal add_hfyu_median_pred_int16, 7,7,0, dst, top, diff, mask, w, left, left_top
 | 
					 | 
				
			||||||
    add      wd, wd
 | 
					 | 
				
			||||||
    movd    mm6, maskd
 | 
					 | 
				
			||||||
    SPLATW  mm6, mm6
 | 
					 | 
				
			||||||
    movq    mm0, [topq]
 | 
					 | 
				
			||||||
    movq    mm2, mm0
 | 
					 | 
				
			||||||
    movd    mm4, [left_topq]
 | 
					 | 
				
			||||||
    psllq   mm2, 16
 | 
					 | 
				
			||||||
    movq    mm1, mm0
 | 
					 | 
				
			||||||
    por     mm4, mm2
 | 
					 | 
				
			||||||
    movd    mm3, [leftq]
 | 
					 | 
				
			||||||
    psubw   mm0, mm4 ; t-tl
 | 
					 | 
				
			||||||
    add    dstq, wq
 | 
					 | 
				
			||||||
    add    topq, wq
 | 
					 | 
				
			||||||
    add   diffq, wq
 | 
					 | 
				
			||||||
    neg      wq
 | 
					 | 
				
			||||||
    jmp .skip
 | 
					 | 
				
			||||||
.loop:
 | 
					 | 
				
			||||||
    movq    mm4, [topq+wq]
 | 
					 | 
				
			||||||
    movq    mm0, mm4
 | 
					 | 
				
			||||||
    psllq   mm4, 16
 | 
					 | 
				
			||||||
    por     mm4, mm1
 | 
					 | 
				
			||||||
    movq    mm1, mm0 ; t
 | 
					 | 
				
			||||||
    psubw   mm0, mm4 ; t-tl
 | 
					 | 
				
			||||||
.skip:
 | 
					 | 
				
			||||||
    movq    mm2, [diffq+wq]
 | 
					 | 
				
			||||||
%assign i 0
 | 
					 | 
				
			||||||
%rep 4
 | 
					 | 
				
			||||||
    movq    mm4, mm0
 | 
					 | 
				
			||||||
    paddw   mm4, mm3 ; t-tl+l
 | 
					 | 
				
			||||||
    pand    mm4, mm6
 | 
					 | 
				
			||||||
    movq    mm5, mm3
 | 
					 | 
				
			||||||
    pmaxsw  mm3, mm1
 | 
					 | 
				
			||||||
    pminsw  mm5, mm1
 | 
					 | 
				
			||||||
    pminsw  mm3, mm4
 | 
					 | 
				
			||||||
    pmaxsw  mm3, mm5 ; median
 | 
					 | 
				
			||||||
    paddw   mm3, mm2 ; +residual
 | 
					 | 
				
			||||||
    pand    mm3, mm6
 | 
					 | 
				
			||||||
%if i==0
 | 
					 | 
				
			||||||
    movq    mm7, mm3
 | 
					 | 
				
			||||||
    psllq   mm7, 48
 | 
					 | 
				
			||||||
%else
 | 
					 | 
				
			||||||
    movq    mm4, mm3
 | 
					 | 
				
			||||||
    psrlq   mm7, 16
 | 
					 | 
				
			||||||
    psllq   mm4, 48
 | 
					 | 
				
			||||||
    por     mm7, mm4
 | 
					 | 
				
			||||||
%endif
 | 
					 | 
				
			||||||
%if i<3
 | 
					 | 
				
			||||||
    psrlq   mm0, 16
 | 
					 | 
				
			||||||
    psrlq   mm1, 16
 | 
					 | 
				
			||||||
    psrlq   mm2, 16
 | 
					 | 
				
			||||||
%endif
 | 
					 | 
				
			||||||
%assign i i+1
 | 
					 | 
				
			||||||
%endrep
 | 
					 | 
				
			||||||
    movq [dstq+wq], mm7
 | 
					 | 
				
			||||||
    add      wq, 8
 | 
					 | 
				
			||||||
    jl .loop
 | 
					 | 
				
			||||||
    movzx   r2d, word [dstq-2]
 | 
					 | 
				
			||||||
    mov [leftq], r2d
 | 
					 | 
				
			||||||
    movzx   r2d, word [topq-2]
 | 
					 | 
				
			||||||
    mov [left_topq], r2d
 | 
					 | 
				
			||||||
    RET
 | 
					 | 
				
			||||||
 | 
				
			|||||||
@ -21,7 +21,6 @@
 | 
				
			|||||||
#include "config.h"
 | 
					#include "config.h"
 | 
				
			||||||
#include "libavutil/x86/asm.h"
 | 
					#include "libavutil/x86/asm.h"
 | 
				
			||||||
#include "../lossless_videodsp.h"
 | 
					#include "../lossless_videodsp.h"
 | 
				
			||||||
#include "libavutil/pixdesc.h"
 | 
					 | 
				
			||||||
#include "libavutil/x86/cpu.h"
 | 
					#include "libavutil/x86/cpu.h"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void ff_add_bytes_mmx(uint8_t *dst, uint8_t *src, intptr_t w);
 | 
					void ff_add_bytes_mmx(uint8_t *dst, uint8_t *src, intptr_t w);
 | 
				
			||||||
@ -39,11 +38,8 @@ int  ff_add_left_pred_ssse3(uint8_t *dst, const uint8_t *src,
 | 
				
			|||||||
int  ff_add_left_pred_sse4(uint8_t *dst, const uint8_t *src,
 | 
					int  ff_add_left_pred_sse4(uint8_t *dst, const uint8_t *src,
 | 
				
			||||||
                            intptr_t w, int left);
 | 
					                            intptr_t w, int left);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void ff_add_int16_mmx(uint16_t *dst, const uint16_t *src, unsigned mask, int w);
 | 
					 | 
				
			||||||
void ff_add_int16_sse2(uint16_t *dst, const uint16_t *src, unsigned mask, int w);
 | 
					 | 
				
			||||||
int ff_add_hfyu_left_pred_int16_ssse3(uint16_t *dst, const uint16_t *src, unsigned mask, int w, unsigned acc);
 | 
					int ff_add_hfyu_left_pred_int16_ssse3(uint16_t *dst, const uint16_t *src, unsigned mask, int w, unsigned acc);
 | 
				
			||||||
int ff_add_hfyu_left_pred_int16_sse4(uint16_t *dst, const uint16_t *src, unsigned mask, int w, unsigned acc);
 | 
					int ff_add_hfyu_left_pred_int16_sse4(uint16_t *dst, const uint16_t *src, unsigned mask, int w, unsigned acc);
 | 
				
			||||||
void ff_add_hfyu_median_pred_int16_mmxext(uint16_t *dst, const uint16_t *top, const uint16_t *diff, unsigned mask, int w, int *left, int *left_top);
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
#if HAVE_INLINE_ASM && HAVE_7REGS && ARCH_X86_32
 | 
					#if HAVE_INLINE_ASM && HAVE_7REGS && ARCH_X86_32
 | 
				
			||||||
static void add_median_pred_cmov(uint8_t *dst, const uint8_t *top,
 | 
					static void add_median_pred_cmov(uint8_t *dst, const uint8_t *top,
 | 
				
			||||||
@ -83,10 +79,9 @@ static void add_median_pred_cmov(uint8_t *dst, const uint8_t *top,
 | 
				
			|||||||
}
 | 
					}
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void ff_llviddsp_init_x86(LLVidDSPContext *c, AVCodecContext *avctx)
 | 
					void ff_llviddsp_init_x86(LLVidDSPContext *c)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
    int cpu_flags = av_get_cpu_flags();
 | 
					    int cpu_flags = av_get_cpu_flags();
 | 
				
			||||||
    const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(avctx->pix_fmt);
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
#if HAVE_INLINE_ASM && HAVE_7REGS && ARCH_X86_32
 | 
					#if HAVE_INLINE_ASM && HAVE_7REGS && ARCH_X86_32
 | 
				
			||||||
    if (cpu_flags & AV_CPU_FLAG_CMOV)
 | 
					    if (cpu_flags & AV_CPU_FLAG_CMOV)
 | 
				
			||||||
@ -95,7 +90,6 @@ void ff_llviddsp_init_x86(LLVidDSPContext *c, AVCodecContext *avctx)
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
    if (ARCH_X86_32 && EXTERNAL_MMX(cpu_flags)) {
 | 
					    if (ARCH_X86_32 && EXTERNAL_MMX(cpu_flags)) {
 | 
				
			||||||
        c->add_bytes = ff_add_bytes_mmx;
 | 
					        c->add_bytes = ff_add_bytes_mmx;
 | 
				
			||||||
        c->add_int16 = ff_add_int16_mmx;
 | 
					 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if (ARCH_X86_32 && EXTERNAL_MMXEXT(cpu_flags)) {
 | 
					    if (ARCH_X86_32 && EXTERNAL_MMXEXT(cpu_flags)) {
 | 
				
			||||||
@ -104,15 +98,9 @@ void ff_llviddsp_init_x86(LLVidDSPContext *c, AVCodecContext *avctx)
 | 
				
			|||||||
            c->add_median_pred = ff_add_median_pred_mmxext;
 | 
					            c->add_median_pred = ff_add_median_pred_mmxext;
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if (EXTERNAL_MMXEXT(cpu_flags) && pix_desc && pix_desc->comp[0].depth<16) {
 | 
					 | 
				
			||||||
        c->add_hfyu_median_pred_int16 = ff_add_hfyu_median_pred_int16_mmxext;
 | 
					 | 
				
			||||||
    }
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
    if (EXTERNAL_SSE2(cpu_flags)) {
 | 
					    if (EXTERNAL_SSE2(cpu_flags)) {
 | 
				
			||||||
        c->add_bytes       = ff_add_bytes_sse2;
 | 
					        c->add_bytes       = ff_add_bytes_sse2;
 | 
				
			||||||
        c->add_median_pred = ff_add_median_pred_sse2;
 | 
					        c->add_median_pred = ff_add_median_pred_sse2;
 | 
				
			||||||
 | 
					 | 
				
			||||||
        c->add_int16 = ff_add_int16_sse2;
 | 
					 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if (EXTERNAL_SSSE3(cpu_flags)) {
 | 
					    if (EXTERNAL_SSSE3(cpu_flags)) {
 | 
				
			||||||
 | 
				
			|||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user