The majority of frame-threaded decoders (mainly the intra-only) need exactly one part of ThreadFrame: The AVFrame. They don't need the owners nor the progress, yet they had to use it because ff_thread_(get|release)_buffer() requires it. This commit changes this and makes these functions work with ordinary AVFrames; the decoders that need the extra fields for progress use ff_thread_(get|release)_ext_buffer() which work exactly as ff_thread_(get|release)_buffer() used to do. This also avoids some unnecessary allocations of progress AVBuffers, namely for H.264 and HEVC film grain frames: These frames are not used for synchronization and therefore don't need a ThreadFrame. Also move the ThreadFrame structure as well as ff_thread_ref_frame() to threadframe.h, the header for frame-threaded decoders with inter-frame dependencies. Reviewed-by: Anton Khirnov <anton@khirnov.net> Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
		
			
				
	
	
		
			229 lines
		
	
	
		
			6.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			229 lines
		
	
	
		
			6.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * V210 decoder
 | |
|  *
 | |
|  * Copyright (C) 2009 Michael Niedermayer <michaelni@gmx.at>
 | |
|  * Copyright (c) 2009 Baptiste Coudurier <baptiste dot coudurier at gmail dot com>
 | |
|  *
 | |
|  * This file is part of FFmpeg.
 | |
|  *
 | |
|  * FFmpeg is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of the GNU Lesser General Public
 | |
|  * License as published by the Free Software Foundation; either
 | |
|  * version 2.1 of the License, or (at your option) any later version.
 | |
|  *
 | |
|  * FFmpeg is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
|  * Lesser General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU Lesser General Public
 | |
|  * License along with FFmpeg; if not, write to the Free Software
 | |
|  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
|  */
 | |
| 
 | |
| #include "avcodec.h"
 | |
| #include "internal.h"
 | |
| #include "v210dec.h"
 | |
| #include "libavutil/bswap.h"
 | |
| #include "libavutil/internal.h"
 | |
| #include "libavutil/intreadwrite.h"
 | |
| #include "thread.h"
 | |
| 
 | |
| #define READ_PIXELS(a, b, c)         \
 | |
|     do {                             \
 | |
|         val  = av_le2ne32(*src++);   \
 | |
|         *a++ =  val & 0x3FF;         \
 | |
|         *b++ = (val >> 10) & 0x3FF;  \
 | |
|         *c++ = (val >> 20) & 0x3FF;  \
 | |
|     } while (0)
 | |
| 
 | |
| typedef struct ThreadData {
 | |
|     AVFrame *frame;
 | |
|     uint8_t *buf;
 | |
|     int stride;
 | |
| } ThreadData;
 | |
| 
 | |
| static void v210_planar_unpack_c(const uint32_t *src, uint16_t *y, uint16_t *u, uint16_t *v, int width)
 | |
| {
 | |
|     uint32_t val;
 | |
|     int i;
 | |
| 
 | |
|     for( i = 0; i < width-5; i += 6 ){
 | |
|         READ_PIXELS(u, y, v);
 | |
|         READ_PIXELS(y, u, y);
 | |
|         READ_PIXELS(v, y, u);
 | |
|         READ_PIXELS(y, v, y);
 | |
|     }
 | |
| }
 | |
| 
 | |
| av_cold void ff_v210dec_init(V210DecContext *s)
 | |
| {
 | |
|     s->unpack_frame = v210_planar_unpack_c;
 | |
|     if (ARCH_X86)
 | |
|         ff_v210_x86_init(s);
 | |
| }
 | |
| 
 | |
| static av_cold int decode_init(AVCodecContext *avctx)
 | |
| {
 | |
|     V210DecContext *s = avctx->priv_data;
 | |
| 
 | |
|     avctx->pix_fmt             = AV_PIX_FMT_YUV422P10;
 | |
|     avctx->bits_per_raw_sample = 10;
 | |
| 
 | |
|     s->thread_count  = av_clip(avctx->thread_count, 1, avctx->height/4);
 | |
|     s->aligned_input = 0;
 | |
|     ff_v210dec_init(s);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int v210_decode_slice(AVCodecContext *avctx, void *arg, int jobnr, int threadnr)
 | |
| {
 | |
|     V210DecContext *s = avctx->priv_data;
 | |
|     int h, w;
 | |
|     ThreadData *td = arg;
 | |
|     AVFrame *frame = td->frame;
 | |
|     int stride = td->stride;
 | |
|     int slice_start = (avctx->height *  jobnr) / s->thread_count;
 | |
|     int slice_end = (avctx->height * (jobnr+1)) / s->thread_count;
 | |
|     uint8_t *psrc = td->buf + stride * slice_start;
 | |
|     uint16_t *y, *u, *v;
 | |
| 
 | |
|     y = (uint16_t*)frame->data[0] + slice_start * frame->linesize[0] / 2;
 | |
|     u = (uint16_t*)frame->data[1] + slice_start * frame->linesize[1] / 2;
 | |
|     v = (uint16_t*)frame->data[2] + slice_start * frame->linesize[2] / 2;
 | |
|     for (h = slice_start; h < slice_end; h++) {
 | |
|         const uint32_t *src = (const uint32_t*)psrc;
 | |
|         uint32_t val;
 | |
| 
 | |
|         w = (avctx->width / 12) * 12;
 | |
|         s->unpack_frame(src, y, u, v, w);
 | |
| 
 | |
|         y += w;
 | |
|         u += w >> 1;
 | |
|         v += w >> 1;
 | |
|         src += (w << 1) / 3;
 | |
| 
 | |
|         if (w < avctx->width - 5) {
 | |
|             READ_PIXELS(u, y, v);
 | |
|             READ_PIXELS(y, u, y);
 | |
|             READ_PIXELS(v, y, u);
 | |
|             READ_PIXELS(y, v, y);
 | |
|             w += 6;
 | |
|         }
 | |
| 
 | |
|         if (w < avctx->width - 1) {
 | |
|             READ_PIXELS(u, y, v);
 | |
| 
 | |
|             val  = av_le2ne32(*src++);
 | |
|             *y++ =  val & 0x3FF;
 | |
|             if (w < avctx->width - 3) {
 | |
|                 *u++ = (val >> 10) & 0x3FF;
 | |
|                 *y++ = (val >> 20) & 0x3FF;
 | |
| 
 | |
|                 val  = av_le2ne32(*src++);
 | |
|                 *v++ =  val & 0x3FF;
 | |
|                 *y++ = (val >> 10) & 0x3FF;
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         psrc += stride;
 | |
|         y += frame->linesize[0] / 2 - avctx->width + (avctx->width & 1);
 | |
|         u += frame->linesize[1] / 2 - avctx->width / 2;
 | |
|         v += frame->linesize[2] / 2 - avctx->width / 2;
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame,
 | |
|                         AVPacket *avpkt)
 | |
| {
 | |
|     V210DecContext *s = avctx->priv_data;
 | |
|     ThreadData td;
 | |
|     int ret, stride, aligned_input;
 | |
|     AVFrame *pic = data;
 | |
|     const uint8_t *psrc = avpkt->data;
 | |
| 
 | |
|     if (s->custom_stride )
 | |
|         stride = s->custom_stride;
 | |
|     else {
 | |
|         int aligned_width = ((avctx->width + 47) / 48) * 48;
 | |
|         stride = aligned_width * 8 / 3;
 | |
|     }
 | |
| 
 | |
|     if (avpkt->size < stride * avctx->height) {
 | |
|         if ((((avctx->width + 23) / 24) * 24 * 8) / 3 * avctx->height == avpkt->size) {
 | |
|             stride = avpkt->size / avctx->height;
 | |
|             if (!s->stride_warning_shown)
 | |
|                 av_log(avctx, AV_LOG_WARNING, "Broken v210 with too small padding (64 byte) detected\n");
 | |
|             s->stride_warning_shown = 1;
 | |
|         } else {
 | |
|             av_log(avctx, AV_LOG_ERROR, "packet too small\n");
 | |
|             return AVERROR_INVALIDDATA;
 | |
|         }
 | |
|     }
 | |
|     td.stride = stride;
 | |
|     if (   avctx->codec_tag == MKTAG('C', '2', '1', '0')
 | |
|         && avpkt->size > 64
 | |
|         && AV_RN32(psrc) == AV_RN32("INFO")
 | |
|         && avpkt->size - 64 >= stride * avctx->height)
 | |
|         psrc += 64;
 | |
| 
 | |
|     aligned_input = !((uintptr_t)psrc & 0x1f) && !(stride & 0x1f);
 | |
|     if (aligned_input != s->aligned_input) {
 | |
|         s->aligned_input = aligned_input;
 | |
|         ff_v210dec_init(s);
 | |
|     }
 | |
| 
 | |
|     if ((ret = ff_thread_get_buffer(avctx, pic, 0)) < 0)
 | |
|         return ret;
 | |
| 
 | |
|     pic->pict_type = AV_PICTURE_TYPE_I;
 | |
|     pic->key_frame = 1;
 | |
| 
 | |
|     td.buf = (uint8_t*)psrc;
 | |
|     td.frame = pic;
 | |
|     avctx->execute2(avctx, v210_decode_slice, &td, NULL, s->thread_count);
 | |
| 
 | |
|     if (avctx->field_order > AV_FIELD_PROGRESSIVE) {
 | |
|         /* we have interlaced material flagged in container */
 | |
|         pic->interlaced_frame = 1;
 | |
|         if (avctx->field_order == AV_FIELD_TT || avctx->field_order == AV_FIELD_TB)
 | |
|             pic->top_field_first = 1;
 | |
|     }
 | |
| 
 | |
|     *got_frame      = 1;
 | |
| 
 | |
|     return avpkt->size;
 | |
| }
 | |
| 
 | |
| #define V210DEC_FLAGS AV_OPT_FLAG_DECODING_PARAM | AV_OPT_FLAG_VIDEO_PARAM
 | |
| static const AVOption v210dec_options[] = {
 | |
|     {"custom_stride", "Custom V210 stride", offsetof(V210DecContext, custom_stride), AV_OPT_TYPE_INT,
 | |
|      {.i64 = 0}, INT_MIN, INT_MAX, V210DEC_FLAGS},
 | |
|     {NULL}
 | |
| };
 | |
| 
 | |
| static const AVClass v210dec_class = {
 | |
|     .class_name = "V210 Decoder",
 | |
|     .item_name  = av_default_item_name,
 | |
|     .option     = v210dec_options,
 | |
|     .version    = LIBAVUTIL_VERSION_INT,
 | |
| };
 | |
| 
 | |
| const AVCodec ff_v210_decoder = {
 | |
|     .name           = "v210",
 | |
|     .long_name      = NULL_IF_CONFIG_SMALL("Uncompressed 4:2:2 10-bit"),
 | |
|     .type           = AVMEDIA_TYPE_VIDEO,
 | |
|     .id             = AV_CODEC_ID_V210,
 | |
|     .priv_data_size = sizeof(V210DecContext),
 | |
|     .init           = decode_init,
 | |
|     .decode         = decode_frame,
 | |
|     .capabilities   = AV_CODEC_CAP_DR1 |
 | |
|                       AV_CODEC_CAP_SLICE_THREADS |
 | |
|                       AV_CODEC_CAP_FRAME_THREADS,
 | |
|     .priv_class     = &v210dec_class,
 | |
|     .caps_internal  = FF_CODEC_CAP_INIT_THREADSAFE,
 | |
| };
 |