The encoders such as libx264 support different QPs offset for different MBs, it makes possible for ROI-based encoding. It makes sense to add support within ffmpeg to generate/accept ROI infos and pass into encoders. Typical usage: After AVFrame is decoded, a ffmpeg filter or user's code generates ROI info for that frame, and the encoder finally does the ROI-based encoding. The ROI info is maintained as side data of AVFrame. Signed-off-by: Guo, Yejun <yejun.guo@intel.com> Signed-off-by: Derek Buitenhuis <derek.buitenhuis@gmail.com>
		
			
				
	
	
		
			948 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			948 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * This file is part of FFmpeg.
 | 
						|
 *
 | 
						|
 * FFmpeg is free software; you can redistribute it and/or
 | 
						|
 * modify it under the terms of the GNU Lesser General Public
 | 
						|
 * License as published by the Free Software Foundation; either
 | 
						|
 * version 2.1 of the License, or (at your option) any later version.
 | 
						|
 *
 | 
						|
 * FFmpeg is distributed in the hope that it will be useful,
 | 
						|
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | 
						|
 * Lesser General Public License for more details.
 | 
						|
 *
 | 
						|
 * You should have received a copy of the GNU Lesser General Public
 | 
						|
 * License along with FFmpeg; if not, write to the Free Software
 | 
						|
 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | 
						|
 */
 | 
						|
 | 
						|
#include "channel_layout.h"
 | 
						|
#include "avassert.h"
 | 
						|
#include "buffer.h"
 | 
						|
#include "common.h"
 | 
						|
#include "dict.h"
 | 
						|
#include "frame.h"
 | 
						|
#include "imgutils.h"
 | 
						|
#include "mem.h"
 | 
						|
#include "samplefmt.h"
 | 
						|
 | 
						|
#if FF_API_FRAME_GET_SET
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, int64_t, best_effort_timestamp)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, int64_t, pkt_duration)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, int64_t, pkt_pos)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, int64_t, channel_layout)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, int,     channels)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, int,     sample_rate)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, AVDictionary *, metadata)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, int,     decode_error_flags)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, int,     pkt_size)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, enum AVColorSpace, colorspace)
 | 
						|
MAKE_ACCESSORS(AVFrame, frame, enum AVColorRange, color_range)
 | 
						|
#endif
 | 
						|
 | 
						|
#define CHECK_CHANNELS_CONSISTENCY(frame) \
 | 
						|
    av_assert2(!(frame)->channel_layout || \
 | 
						|
               (frame)->channels == \
 | 
						|
               av_get_channel_layout_nb_channels((frame)->channel_layout))
 | 
						|
 | 
						|
#if FF_API_FRAME_QP
 | 
						|
struct qp_properties {
 | 
						|
    int stride;
 | 
						|
    int type;
 | 
						|
};
 | 
						|
 | 
						|
int av_frame_set_qp_table(AVFrame *f, AVBufferRef *buf, int stride, int qp_type)
 | 
						|
{
 | 
						|
    struct qp_properties *p;
 | 
						|
    AVFrameSideData *sd;
 | 
						|
    AVBufferRef *ref;
 | 
						|
 | 
						|
FF_DISABLE_DEPRECATION_WARNINGS
 | 
						|
    av_buffer_unref(&f->qp_table_buf);
 | 
						|
 | 
						|
    f->qp_table_buf = buf;
 | 
						|
    f->qscale_table = buf->data;
 | 
						|
    f->qstride      = stride;
 | 
						|
    f->qscale_type  = qp_type;
 | 
						|
FF_ENABLE_DEPRECATION_WARNINGS
 | 
						|
 | 
						|
    av_frame_remove_side_data(f, AV_FRAME_DATA_QP_TABLE_PROPERTIES);
 | 
						|
    av_frame_remove_side_data(f, AV_FRAME_DATA_QP_TABLE_DATA);
 | 
						|
 | 
						|
    ref = av_buffer_ref(buf);
 | 
						|
    if (!av_frame_new_side_data_from_buf(f, AV_FRAME_DATA_QP_TABLE_DATA, ref)) {
 | 
						|
        av_buffer_unref(&ref);
 | 
						|
        return AVERROR(ENOMEM);
 | 
						|
    }
 | 
						|
 | 
						|
    sd = av_frame_new_side_data(f, AV_FRAME_DATA_QP_TABLE_PROPERTIES,
 | 
						|
                                sizeof(struct qp_properties));
 | 
						|
    if (!sd)
 | 
						|
        return AVERROR(ENOMEM);
 | 
						|
 | 
						|
    p = (struct qp_properties *)sd->data;
 | 
						|
    p->stride = stride;
 | 
						|
    p->type = qp_type;
 | 
						|
 | 
						|
    return 0;
 | 
						|
}
 | 
						|
 | 
						|
int8_t *av_frame_get_qp_table(AVFrame *f, int *stride, int *type)
 | 
						|
{
 | 
						|
    AVBufferRef *buf = NULL;
 | 
						|
 | 
						|
    *stride = 0;
 | 
						|
    *type   = 0;
 | 
						|
 | 
						|
FF_DISABLE_DEPRECATION_WARNINGS
 | 
						|
    if (f->qp_table_buf) {
 | 
						|
        *stride = f->qstride;
 | 
						|
        *type   = f->qscale_type;
 | 
						|
        buf     = f->qp_table_buf;
 | 
						|
FF_ENABLE_DEPRECATION_WARNINGS
 | 
						|
    } else {
 | 
						|
        AVFrameSideData *sd;
 | 
						|
        struct qp_properties *p;
 | 
						|
        sd = av_frame_get_side_data(f, AV_FRAME_DATA_QP_TABLE_PROPERTIES);
 | 
						|
        if (!sd)
 | 
						|
            return NULL;
 | 
						|
        p = (struct qp_properties *)sd->data;
 | 
						|
        sd = av_frame_get_side_data(f, AV_FRAME_DATA_QP_TABLE_DATA);
 | 
						|
        if (!sd)
 | 
						|
            return NULL;
 | 
						|
        *stride = p->stride;
 | 
						|
        *type   = p->type;
 | 
						|
        buf     = sd->buf;
 | 
						|
    }
 | 
						|
 | 
						|
    return buf ? buf->data : NULL;
 | 
						|
}
 | 
						|
#endif
 | 
						|
 | 
						|
const char *av_get_colorspace_name(enum AVColorSpace val)
 | 
						|
{
 | 
						|
    static const char * const name[] = {
 | 
						|
        [AVCOL_SPC_RGB]       = "GBR",
 | 
						|
        [AVCOL_SPC_BT709]     = "bt709",
 | 
						|
        [AVCOL_SPC_FCC]       = "fcc",
 | 
						|
        [AVCOL_SPC_BT470BG]   = "bt470bg",
 | 
						|
        [AVCOL_SPC_SMPTE170M] = "smpte170m",
 | 
						|
        [AVCOL_SPC_SMPTE240M] = "smpte240m",
 | 
						|
        [AVCOL_SPC_YCOCG]     = "YCgCo",
 | 
						|
    };
 | 
						|
    if ((unsigned)val >= FF_ARRAY_ELEMS(name))
 | 
						|
        return NULL;
 | 
						|
    return name[val];
 | 
						|
}
 | 
						|
 | 
						|
static void get_frame_defaults(AVFrame *frame)
 | 
						|
{
 | 
						|
    if (frame->extended_data != frame->data)
 | 
						|
        av_freep(&frame->extended_data);
 | 
						|
 | 
						|
    memset(frame, 0, sizeof(*frame));
 | 
						|
 | 
						|
    frame->pts                   =
 | 
						|
    frame->pkt_dts               = AV_NOPTS_VALUE;
 | 
						|
#if FF_API_PKT_PTS
 | 
						|
FF_DISABLE_DEPRECATION_WARNINGS
 | 
						|
    frame->pkt_pts               = AV_NOPTS_VALUE;
 | 
						|
FF_ENABLE_DEPRECATION_WARNINGS
 | 
						|
#endif
 | 
						|
    frame->best_effort_timestamp = AV_NOPTS_VALUE;
 | 
						|
    frame->pkt_duration        = 0;
 | 
						|
    frame->pkt_pos             = -1;
 | 
						|
    frame->pkt_size            = -1;
 | 
						|
    frame->key_frame           = 1;
 | 
						|
    frame->sample_aspect_ratio = (AVRational){ 0, 1 };
 | 
						|
    frame->format              = -1; /* unknown */
 | 
						|
    frame->extended_data       = frame->data;
 | 
						|
    frame->color_primaries     = AVCOL_PRI_UNSPECIFIED;
 | 
						|
    frame->color_trc           = AVCOL_TRC_UNSPECIFIED;
 | 
						|
    frame->colorspace          = AVCOL_SPC_UNSPECIFIED;
 | 
						|
    frame->color_range         = AVCOL_RANGE_UNSPECIFIED;
 | 
						|
    frame->chroma_location     = AVCHROMA_LOC_UNSPECIFIED;
 | 
						|
    frame->flags               = 0;
 | 
						|
}
 | 
						|
 | 
						|
static void free_side_data(AVFrameSideData **ptr_sd)
 | 
						|
{
 | 
						|
    AVFrameSideData *sd = *ptr_sd;
 | 
						|
 | 
						|
    av_buffer_unref(&sd->buf);
 | 
						|
    av_dict_free(&sd->metadata);
 | 
						|
    av_freep(ptr_sd);
 | 
						|
}
 | 
						|
 | 
						|
static void wipe_side_data(AVFrame *frame)
 | 
						|
{
 | 
						|
    int i;
 | 
						|
 | 
						|
    for (i = 0; i < frame->nb_side_data; i++) {
 | 
						|
        free_side_data(&frame->side_data[i]);
 | 
						|
    }
 | 
						|
    frame->nb_side_data = 0;
 | 
						|
 | 
						|
    av_freep(&frame->side_data);
 | 
						|
}
 | 
						|
 | 
						|
AVFrame *av_frame_alloc(void)
 | 
						|
{
 | 
						|
    AVFrame *frame = av_mallocz(sizeof(*frame));
 | 
						|
 | 
						|
    if (!frame)
 | 
						|
        return NULL;
 | 
						|
 | 
						|
    frame->extended_data = NULL;
 | 
						|
    get_frame_defaults(frame);
 | 
						|
 | 
						|
    return frame;
 | 
						|
}
 | 
						|
 | 
						|
void av_frame_free(AVFrame **frame)
 | 
						|
{
 | 
						|
    if (!frame || !*frame)
 | 
						|
        return;
 | 
						|
 | 
						|
    av_frame_unref(*frame);
 | 
						|
    av_freep(frame);
 | 
						|
}
 | 
						|
 | 
						|
static int get_video_buffer(AVFrame *frame, int align)
 | 
						|
{
 | 
						|
    const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(frame->format);
 | 
						|
    int ret, i, padded_height;
 | 
						|
    int plane_padding = FFMAX(16 + 16/*STRIDE_ALIGN*/, align);
 | 
						|
 | 
						|
    if (!desc)
 | 
						|
        return AVERROR(EINVAL);
 | 
						|
 | 
						|
    if ((ret = av_image_check_size(frame->width, frame->height, 0, NULL)) < 0)
 | 
						|
        return ret;
 | 
						|
 | 
						|
    if (!frame->linesize[0]) {
 | 
						|
        if (align <= 0)
 | 
						|
            align = 32; /* STRIDE_ALIGN. Should be av_cpu_max_align() */
 | 
						|
 | 
						|
        for(i=1; i<=align; i+=i) {
 | 
						|
            ret = av_image_fill_linesizes(frame->linesize, frame->format,
 | 
						|
                                          FFALIGN(frame->width, i));
 | 
						|
            if (ret < 0)
 | 
						|
                return ret;
 | 
						|
            if (!(frame->linesize[0] & (align-1)))
 | 
						|
                break;
 | 
						|
        }
 | 
						|
 | 
						|
        for (i = 0; i < 4 && frame->linesize[i]; i++)
 | 
						|
            frame->linesize[i] = FFALIGN(frame->linesize[i], align);
 | 
						|
    }
 | 
						|
 | 
						|
    padded_height = FFALIGN(frame->height, 32);
 | 
						|
    if ((ret = av_image_fill_pointers(frame->data, frame->format, padded_height,
 | 
						|
                                      NULL, frame->linesize)) < 0)
 | 
						|
        return ret;
 | 
						|
 | 
						|
    frame->buf[0] = av_buffer_alloc(ret + 4*plane_padding);
 | 
						|
    if (!frame->buf[0]) {
 | 
						|
        ret = AVERROR(ENOMEM);
 | 
						|
        goto fail;
 | 
						|
    }
 | 
						|
 | 
						|
    if ((ret = av_image_fill_pointers(frame->data, frame->format, padded_height,
 | 
						|
                                      frame->buf[0]->data, frame->linesize)) < 0)
 | 
						|
        goto fail;
 | 
						|
 | 
						|
    for (i = 1; i < 4; i++) {
 | 
						|
        if (frame->data[i])
 | 
						|
            frame->data[i] += i * plane_padding;
 | 
						|
    }
 | 
						|
 | 
						|
    frame->extended_data = frame->data;
 | 
						|
 | 
						|
    return 0;
 | 
						|
fail:
 | 
						|
    av_frame_unref(frame);
 | 
						|
    return ret;
 | 
						|
}
 | 
						|
 | 
						|
static int get_audio_buffer(AVFrame *frame, int align)
 | 
						|
{
 | 
						|
    int channels;
 | 
						|
    int planar   = av_sample_fmt_is_planar(frame->format);
 | 
						|
    int planes;
 | 
						|
    int ret, i;
 | 
						|
 | 
						|
    if (!frame->channels)
 | 
						|
        frame->channels = av_get_channel_layout_nb_channels(frame->channel_layout);
 | 
						|
 | 
						|
    channels = frame->channels;
 | 
						|
    planes = planar ? channels : 1;
 | 
						|
 | 
						|
    CHECK_CHANNELS_CONSISTENCY(frame);
 | 
						|
    if (!frame->linesize[0]) {
 | 
						|
        ret = av_samples_get_buffer_size(&frame->linesize[0], channels,
 | 
						|
                                         frame->nb_samples, frame->format,
 | 
						|
                                         align);
 | 
						|
        if (ret < 0)
 | 
						|
            return ret;
 | 
						|
    }
 | 
						|
 | 
						|
    if (planes > AV_NUM_DATA_POINTERS) {
 | 
						|
        frame->extended_data = av_mallocz_array(planes,
 | 
						|
                                          sizeof(*frame->extended_data));
 | 
						|
        frame->extended_buf  = av_mallocz_array((planes - AV_NUM_DATA_POINTERS),
 | 
						|
                                          sizeof(*frame->extended_buf));
 | 
						|
        if (!frame->extended_data || !frame->extended_buf) {
 | 
						|
            av_freep(&frame->extended_data);
 | 
						|
            av_freep(&frame->extended_buf);
 | 
						|
            return AVERROR(ENOMEM);
 | 
						|
        }
 | 
						|
        frame->nb_extended_buf = planes - AV_NUM_DATA_POINTERS;
 | 
						|
    } else
 | 
						|
        frame->extended_data = frame->data;
 | 
						|
 | 
						|
    for (i = 0; i < FFMIN(planes, AV_NUM_DATA_POINTERS); i++) {
 | 
						|
        frame->buf[i] = av_buffer_alloc(frame->linesize[0]);
 | 
						|
        if (!frame->buf[i]) {
 | 
						|
            av_frame_unref(frame);
 | 
						|
            return AVERROR(ENOMEM);
 | 
						|
        }
 | 
						|
        frame->extended_data[i] = frame->data[i] = frame->buf[i]->data;
 | 
						|
    }
 | 
						|
    for (i = 0; i < planes - AV_NUM_DATA_POINTERS; i++) {
 | 
						|
        frame->extended_buf[i] = av_buffer_alloc(frame->linesize[0]);
 | 
						|
        if (!frame->extended_buf[i]) {
 | 
						|
            av_frame_unref(frame);
 | 
						|
            return AVERROR(ENOMEM);
 | 
						|
        }
 | 
						|
        frame->extended_data[i + AV_NUM_DATA_POINTERS] = frame->extended_buf[i]->data;
 | 
						|
    }
 | 
						|
    return 0;
 | 
						|
 | 
						|
}
 | 
						|
 | 
						|
int av_frame_get_buffer(AVFrame *frame, int align)
 | 
						|
{
 | 
						|
    if (frame->format < 0)
 | 
						|
        return AVERROR(EINVAL);
 | 
						|
 | 
						|
    if (frame->width > 0 && frame->height > 0)
 | 
						|
        return get_video_buffer(frame, align);
 | 
						|
    else if (frame->nb_samples > 0 && (frame->channel_layout || frame->channels > 0))
 | 
						|
        return get_audio_buffer(frame, align);
 | 
						|
 | 
						|
    return AVERROR(EINVAL);
 | 
						|
}
 | 
						|
 | 
						|
static int frame_copy_props(AVFrame *dst, const AVFrame *src, int force_copy)
 | 
						|
{
 | 
						|
    int i;
 | 
						|
 | 
						|
    dst->key_frame              = src->key_frame;
 | 
						|
    dst->pict_type              = src->pict_type;
 | 
						|
    dst->sample_aspect_ratio    = src->sample_aspect_ratio;
 | 
						|
    dst->crop_top               = src->crop_top;
 | 
						|
    dst->crop_bottom            = src->crop_bottom;
 | 
						|
    dst->crop_left              = src->crop_left;
 | 
						|
    dst->crop_right             = src->crop_right;
 | 
						|
    dst->pts                    = src->pts;
 | 
						|
    dst->repeat_pict            = src->repeat_pict;
 | 
						|
    dst->interlaced_frame       = src->interlaced_frame;
 | 
						|
    dst->top_field_first        = src->top_field_first;
 | 
						|
    dst->palette_has_changed    = src->palette_has_changed;
 | 
						|
    dst->sample_rate            = src->sample_rate;
 | 
						|
    dst->opaque                 = src->opaque;
 | 
						|
#if FF_API_PKT_PTS
 | 
						|
FF_DISABLE_DEPRECATION_WARNINGS
 | 
						|
    dst->pkt_pts                = src->pkt_pts;
 | 
						|
FF_ENABLE_DEPRECATION_WARNINGS
 | 
						|
#endif
 | 
						|
    dst->pkt_dts                = src->pkt_dts;
 | 
						|
    dst->pkt_pos                = src->pkt_pos;
 | 
						|
    dst->pkt_size               = src->pkt_size;
 | 
						|
    dst->pkt_duration           = src->pkt_duration;
 | 
						|
    dst->reordered_opaque       = src->reordered_opaque;
 | 
						|
    dst->quality                = src->quality;
 | 
						|
    dst->best_effort_timestamp  = src->best_effort_timestamp;
 | 
						|
    dst->coded_picture_number   = src->coded_picture_number;
 | 
						|
    dst->display_picture_number = src->display_picture_number;
 | 
						|
    dst->flags                  = src->flags;
 | 
						|
    dst->decode_error_flags     = src->decode_error_flags;
 | 
						|
    dst->color_primaries        = src->color_primaries;
 | 
						|
    dst->color_trc              = src->color_trc;
 | 
						|
    dst->colorspace             = src->colorspace;
 | 
						|
    dst->color_range            = src->color_range;
 | 
						|
    dst->chroma_location        = src->chroma_location;
 | 
						|
 | 
						|
    av_dict_copy(&dst->metadata, src->metadata, 0);
 | 
						|
 | 
						|
#if FF_API_ERROR_FRAME
 | 
						|
FF_DISABLE_DEPRECATION_WARNINGS
 | 
						|
    memcpy(dst->error, src->error, sizeof(dst->error));
 | 
						|
FF_ENABLE_DEPRECATION_WARNINGS
 | 
						|
#endif
 | 
						|
 | 
						|
    for (i = 0; i < src->nb_side_data; i++) {
 | 
						|
        const AVFrameSideData *sd_src = src->side_data[i];
 | 
						|
        AVFrameSideData *sd_dst;
 | 
						|
        if (   sd_src->type == AV_FRAME_DATA_PANSCAN
 | 
						|
            && (src->width != dst->width || src->height != dst->height))
 | 
						|
            continue;
 | 
						|
        if (force_copy) {
 | 
						|
            sd_dst = av_frame_new_side_data(dst, sd_src->type,
 | 
						|
                                            sd_src->size);
 | 
						|
            if (!sd_dst) {
 | 
						|
                wipe_side_data(dst);
 | 
						|
                return AVERROR(ENOMEM);
 | 
						|
            }
 | 
						|
            memcpy(sd_dst->data, sd_src->data, sd_src->size);
 | 
						|
        } else {
 | 
						|
            AVBufferRef *ref = av_buffer_ref(sd_src->buf);
 | 
						|
            sd_dst = av_frame_new_side_data_from_buf(dst, sd_src->type, ref);
 | 
						|
            if (!sd_dst) {
 | 
						|
                av_buffer_unref(&ref);
 | 
						|
                wipe_side_data(dst);
 | 
						|
                return AVERROR(ENOMEM);
 | 
						|
            }
 | 
						|
        }
 | 
						|
        av_dict_copy(&sd_dst->metadata, sd_src->metadata, 0);
 | 
						|
    }
 | 
						|
 | 
						|
#if FF_API_FRAME_QP
 | 
						|
FF_DISABLE_DEPRECATION_WARNINGS
 | 
						|
    dst->qscale_table = NULL;
 | 
						|
    dst->qstride      = 0;
 | 
						|
    dst->qscale_type  = 0;
 | 
						|
    av_buffer_unref(&dst->qp_table_buf);
 | 
						|
    if (src->qp_table_buf) {
 | 
						|
        dst->qp_table_buf = av_buffer_ref(src->qp_table_buf);
 | 
						|
        if (dst->qp_table_buf) {
 | 
						|
            dst->qscale_table = dst->qp_table_buf->data;
 | 
						|
            dst->qstride      = src->qstride;
 | 
						|
            dst->qscale_type  = src->qscale_type;
 | 
						|
        }
 | 
						|
    }
 | 
						|
FF_ENABLE_DEPRECATION_WARNINGS
 | 
						|
#endif
 | 
						|
 | 
						|
    av_buffer_unref(&dst->opaque_ref);
 | 
						|
    av_buffer_unref(&dst->private_ref);
 | 
						|
    if (src->opaque_ref) {
 | 
						|
        dst->opaque_ref = av_buffer_ref(src->opaque_ref);
 | 
						|
        if (!dst->opaque_ref)
 | 
						|
            return AVERROR(ENOMEM);
 | 
						|
    }
 | 
						|
    if (src->private_ref) {
 | 
						|
        dst->private_ref = av_buffer_ref(src->private_ref);
 | 
						|
        if (!dst->private_ref)
 | 
						|
            return AVERROR(ENOMEM);
 | 
						|
    }
 | 
						|
    return 0;
 | 
						|
}
 | 
						|
 | 
						|
int av_frame_ref(AVFrame *dst, const AVFrame *src)
 | 
						|
{
 | 
						|
    int i, ret = 0;
 | 
						|
 | 
						|
    av_assert1(dst->width == 0 && dst->height == 0);
 | 
						|
    av_assert1(dst->channels == 0);
 | 
						|
 | 
						|
    dst->format         = src->format;
 | 
						|
    dst->width          = src->width;
 | 
						|
    dst->height         = src->height;
 | 
						|
    dst->channels       = src->channels;
 | 
						|
    dst->channel_layout = src->channel_layout;
 | 
						|
    dst->nb_samples     = src->nb_samples;
 | 
						|
 | 
						|
    ret = frame_copy_props(dst, src, 0);
 | 
						|
    if (ret < 0)
 | 
						|
        return ret;
 | 
						|
 | 
						|
    /* duplicate the frame data if it's not refcounted */
 | 
						|
    if (!src->buf[0]) {
 | 
						|
        ret = av_frame_get_buffer(dst, 32);
 | 
						|
        if (ret < 0)
 | 
						|
            return ret;
 | 
						|
 | 
						|
        ret = av_frame_copy(dst, src);
 | 
						|
        if (ret < 0)
 | 
						|
            av_frame_unref(dst);
 | 
						|
 | 
						|
        return ret;
 | 
						|
    }
 | 
						|
 | 
						|
    /* ref the buffers */
 | 
						|
    for (i = 0; i < FF_ARRAY_ELEMS(src->buf); i++) {
 | 
						|
        if (!src->buf[i])
 | 
						|
            continue;
 | 
						|
        dst->buf[i] = av_buffer_ref(src->buf[i]);
 | 
						|
        if (!dst->buf[i]) {
 | 
						|
            ret = AVERROR(ENOMEM);
 | 
						|
            goto fail;
 | 
						|
        }
 | 
						|
    }
 | 
						|
 | 
						|
    if (src->extended_buf) {
 | 
						|
        dst->extended_buf = av_mallocz_array(sizeof(*dst->extended_buf),
 | 
						|
                                       src->nb_extended_buf);
 | 
						|
        if (!dst->extended_buf) {
 | 
						|
            ret = AVERROR(ENOMEM);
 | 
						|
            goto fail;
 | 
						|
        }
 | 
						|
        dst->nb_extended_buf = src->nb_extended_buf;
 | 
						|
 | 
						|
        for (i = 0; i < src->nb_extended_buf; i++) {
 | 
						|
            dst->extended_buf[i] = av_buffer_ref(src->extended_buf[i]);
 | 
						|
            if (!dst->extended_buf[i]) {
 | 
						|
                ret = AVERROR(ENOMEM);
 | 
						|
                goto fail;
 | 
						|
            }
 | 
						|
        }
 | 
						|
    }
 | 
						|
 | 
						|
    if (src->hw_frames_ctx) {
 | 
						|
        dst->hw_frames_ctx = av_buffer_ref(src->hw_frames_ctx);
 | 
						|
        if (!dst->hw_frames_ctx) {
 | 
						|
            ret = AVERROR(ENOMEM);
 | 
						|
            goto fail;
 | 
						|
        }
 | 
						|
    }
 | 
						|
 | 
						|
    /* duplicate extended data */
 | 
						|
    if (src->extended_data != src->data) {
 | 
						|
        int ch = src->channels;
 | 
						|
 | 
						|
        if (!ch) {
 | 
						|
            ret = AVERROR(EINVAL);
 | 
						|
            goto fail;
 | 
						|
        }
 | 
						|
        CHECK_CHANNELS_CONSISTENCY(src);
 | 
						|
 | 
						|
        dst->extended_data = av_malloc_array(sizeof(*dst->extended_data), ch);
 | 
						|
        if (!dst->extended_data) {
 | 
						|
            ret = AVERROR(ENOMEM);
 | 
						|
            goto fail;
 | 
						|
        }
 | 
						|
        memcpy(dst->extended_data, src->extended_data, sizeof(*src->extended_data) * ch);
 | 
						|
    } else
 | 
						|
        dst->extended_data = dst->data;
 | 
						|
 | 
						|
    memcpy(dst->data,     src->data,     sizeof(src->data));
 | 
						|
    memcpy(dst->linesize, src->linesize, sizeof(src->linesize));
 | 
						|
 | 
						|
    return 0;
 | 
						|
 | 
						|
fail:
 | 
						|
    av_frame_unref(dst);
 | 
						|
    return ret;
 | 
						|
}
 | 
						|
 | 
						|
AVFrame *av_frame_clone(const AVFrame *src)
 | 
						|
{
 | 
						|
    AVFrame *ret = av_frame_alloc();
 | 
						|
 | 
						|
    if (!ret)
 | 
						|
        return NULL;
 | 
						|
 | 
						|
    if (av_frame_ref(ret, src) < 0)
 | 
						|
        av_frame_free(&ret);
 | 
						|
 | 
						|
    return ret;
 | 
						|
}
 | 
						|
 | 
						|
void av_frame_unref(AVFrame *frame)
 | 
						|
{
 | 
						|
    int i;
 | 
						|
 | 
						|
    if (!frame)
 | 
						|
        return;
 | 
						|
 | 
						|
    wipe_side_data(frame);
 | 
						|
 | 
						|
    for (i = 0; i < FF_ARRAY_ELEMS(frame->buf); i++)
 | 
						|
        av_buffer_unref(&frame->buf[i]);
 | 
						|
    for (i = 0; i < frame->nb_extended_buf; i++)
 | 
						|
        av_buffer_unref(&frame->extended_buf[i]);
 | 
						|
    av_freep(&frame->extended_buf);
 | 
						|
    av_dict_free(&frame->metadata);
 | 
						|
#if FF_API_FRAME_QP
 | 
						|
FF_DISABLE_DEPRECATION_WARNINGS
 | 
						|
    av_buffer_unref(&frame->qp_table_buf);
 | 
						|
FF_ENABLE_DEPRECATION_WARNINGS
 | 
						|
#endif
 | 
						|
 | 
						|
    av_buffer_unref(&frame->hw_frames_ctx);
 | 
						|
 | 
						|
    av_buffer_unref(&frame->opaque_ref);
 | 
						|
    av_buffer_unref(&frame->private_ref);
 | 
						|
 | 
						|
    get_frame_defaults(frame);
 | 
						|
}
 | 
						|
 | 
						|
void av_frame_move_ref(AVFrame *dst, AVFrame *src)
 | 
						|
{
 | 
						|
    av_assert1(dst->width == 0 && dst->height == 0);
 | 
						|
    av_assert1(dst->channels == 0);
 | 
						|
 | 
						|
    *dst = *src;
 | 
						|
    if (src->extended_data == src->data)
 | 
						|
        dst->extended_data = dst->data;
 | 
						|
    memset(src, 0, sizeof(*src));
 | 
						|
    get_frame_defaults(src);
 | 
						|
}
 | 
						|
 | 
						|
int av_frame_is_writable(AVFrame *frame)
 | 
						|
{
 | 
						|
    int i, ret = 1;
 | 
						|
 | 
						|
    /* assume non-refcounted frames are not writable */
 | 
						|
    if (!frame->buf[0])
 | 
						|
        return 0;
 | 
						|
 | 
						|
    for (i = 0; i < FF_ARRAY_ELEMS(frame->buf); i++)
 | 
						|
        if (frame->buf[i])
 | 
						|
            ret &= !!av_buffer_is_writable(frame->buf[i]);
 | 
						|
    for (i = 0; i < frame->nb_extended_buf; i++)
 | 
						|
        ret &= !!av_buffer_is_writable(frame->extended_buf[i]);
 | 
						|
 | 
						|
    return ret;
 | 
						|
}
 | 
						|
 | 
						|
int av_frame_make_writable(AVFrame *frame)
 | 
						|
{
 | 
						|
    AVFrame tmp;
 | 
						|
    int ret;
 | 
						|
 | 
						|
    if (!frame->buf[0])
 | 
						|
        return AVERROR(EINVAL);
 | 
						|
 | 
						|
    if (av_frame_is_writable(frame))
 | 
						|
        return 0;
 | 
						|
 | 
						|
    memset(&tmp, 0, sizeof(tmp));
 | 
						|
    tmp.format         = frame->format;
 | 
						|
    tmp.width          = frame->width;
 | 
						|
    tmp.height         = frame->height;
 | 
						|
    tmp.channels       = frame->channels;
 | 
						|
    tmp.channel_layout = frame->channel_layout;
 | 
						|
    tmp.nb_samples     = frame->nb_samples;
 | 
						|
    ret = av_frame_get_buffer(&tmp, 32);
 | 
						|
    if (ret < 0)
 | 
						|
        return ret;
 | 
						|
 | 
						|
    ret = av_frame_copy(&tmp, frame);
 | 
						|
    if (ret < 0) {
 | 
						|
        av_frame_unref(&tmp);
 | 
						|
        return ret;
 | 
						|
    }
 | 
						|
 | 
						|
    ret = av_frame_copy_props(&tmp, frame);
 | 
						|
    if (ret < 0) {
 | 
						|
        av_frame_unref(&tmp);
 | 
						|
        return ret;
 | 
						|
    }
 | 
						|
 | 
						|
    av_frame_unref(frame);
 | 
						|
 | 
						|
    *frame = tmp;
 | 
						|
    if (tmp.data == tmp.extended_data)
 | 
						|
        frame->extended_data = frame->data;
 | 
						|
 | 
						|
    return 0;
 | 
						|
}
 | 
						|
 | 
						|
int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
 | 
						|
{
 | 
						|
    return frame_copy_props(dst, src, 1);
 | 
						|
}
 | 
						|
 | 
						|
AVBufferRef *av_frame_get_plane_buffer(AVFrame *frame, int plane)
 | 
						|
{
 | 
						|
    uint8_t *data;
 | 
						|
    int planes, i;
 | 
						|
 | 
						|
    if (frame->nb_samples) {
 | 
						|
        int channels = frame->channels;
 | 
						|
        if (!channels)
 | 
						|
            return NULL;
 | 
						|
        CHECK_CHANNELS_CONSISTENCY(frame);
 | 
						|
        planes = av_sample_fmt_is_planar(frame->format) ? channels : 1;
 | 
						|
    } else
 | 
						|
        planes = 4;
 | 
						|
 | 
						|
    if (plane < 0 || plane >= planes || !frame->extended_data[plane])
 | 
						|
        return NULL;
 | 
						|
    data = frame->extended_data[plane];
 | 
						|
 | 
						|
    for (i = 0; i < FF_ARRAY_ELEMS(frame->buf) && frame->buf[i]; i++) {
 | 
						|
        AVBufferRef *buf = frame->buf[i];
 | 
						|
        if (data >= buf->data && data < buf->data + buf->size)
 | 
						|
            return buf;
 | 
						|
    }
 | 
						|
    for (i = 0; i < frame->nb_extended_buf; i++) {
 | 
						|
        AVBufferRef *buf = frame->extended_buf[i];
 | 
						|
        if (data >= buf->data && data < buf->data + buf->size)
 | 
						|
            return buf;
 | 
						|
    }
 | 
						|
    return NULL;
 | 
						|
}
 | 
						|
 | 
						|
AVFrameSideData *av_frame_new_side_data_from_buf(AVFrame *frame,
 | 
						|
                                                 enum AVFrameSideDataType type,
 | 
						|
                                                 AVBufferRef *buf)
 | 
						|
{
 | 
						|
    AVFrameSideData *ret, **tmp;
 | 
						|
 | 
						|
    if (!buf)
 | 
						|
        return NULL;
 | 
						|
 | 
						|
    if (frame->nb_side_data > INT_MAX / sizeof(*frame->side_data) - 1)
 | 
						|
        return NULL;
 | 
						|
 | 
						|
    tmp = av_realloc(frame->side_data,
 | 
						|
                     (frame->nb_side_data + 1) * sizeof(*frame->side_data));
 | 
						|
    if (!tmp)
 | 
						|
        return NULL;
 | 
						|
    frame->side_data = tmp;
 | 
						|
 | 
						|
    ret = av_mallocz(sizeof(*ret));
 | 
						|
    if (!ret)
 | 
						|
        return NULL;
 | 
						|
 | 
						|
    ret->buf = buf;
 | 
						|
    ret->data = ret->buf->data;
 | 
						|
    ret->size = buf->size;
 | 
						|
    ret->type = type;
 | 
						|
 | 
						|
    frame->side_data[frame->nb_side_data++] = ret;
 | 
						|
 | 
						|
    return ret;
 | 
						|
}
 | 
						|
 | 
						|
AVFrameSideData *av_frame_new_side_data(AVFrame *frame,
 | 
						|
                                        enum AVFrameSideDataType type,
 | 
						|
                                        int size)
 | 
						|
{
 | 
						|
    AVFrameSideData *ret;
 | 
						|
    AVBufferRef *buf = av_buffer_alloc(size);
 | 
						|
    ret = av_frame_new_side_data_from_buf(frame, type, buf);
 | 
						|
    if (!ret)
 | 
						|
        av_buffer_unref(&buf);
 | 
						|
    return ret;
 | 
						|
}
 | 
						|
 | 
						|
AVFrameSideData *av_frame_get_side_data(const AVFrame *frame,
 | 
						|
                                        enum AVFrameSideDataType type)
 | 
						|
{
 | 
						|
    int i;
 | 
						|
 | 
						|
    for (i = 0; i < frame->nb_side_data; i++) {
 | 
						|
        if (frame->side_data[i]->type == type)
 | 
						|
            return frame->side_data[i];
 | 
						|
    }
 | 
						|
    return NULL;
 | 
						|
}
 | 
						|
 | 
						|
static int frame_copy_video(AVFrame *dst, const AVFrame *src)
 | 
						|
{
 | 
						|
    const uint8_t *src_data[4];
 | 
						|
    int i, planes;
 | 
						|
 | 
						|
    if (dst->width  < src->width ||
 | 
						|
        dst->height < src->height)
 | 
						|
        return AVERROR(EINVAL);
 | 
						|
 | 
						|
    planes = av_pix_fmt_count_planes(dst->format);
 | 
						|
    for (i = 0; i < planes; i++)
 | 
						|
        if (!dst->data[i] || !src->data[i])
 | 
						|
            return AVERROR(EINVAL);
 | 
						|
 | 
						|
    memcpy(src_data, src->data, sizeof(src_data));
 | 
						|
    av_image_copy(dst->data, dst->linesize,
 | 
						|
                  src_data, src->linesize,
 | 
						|
                  dst->format, src->width, src->height);
 | 
						|
 | 
						|
    return 0;
 | 
						|
}
 | 
						|
 | 
						|
static int frame_copy_audio(AVFrame *dst, const AVFrame *src)
 | 
						|
{
 | 
						|
    int planar   = av_sample_fmt_is_planar(dst->format);
 | 
						|
    int channels = dst->channels;
 | 
						|
    int planes   = planar ? channels : 1;
 | 
						|
    int i;
 | 
						|
 | 
						|
    if (dst->nb_samples     != src->nb_samples ||
 | 
						|
        dst->channels       != src->channels ||
 | 
						|
        dst->channel_layout != src->channel_layout)
 | 
						|
        return AVERROR(EINVAL);
 | 
						|
 | 
						|
    CHECK_CHANNELS_CONSISTENCY(src);
 | 
						|
 | 
						|
    for (i = 0; i < planes; i++)
 | 
						|
        if (!dst->extended_data[i] || !src->extended_data[i])
 | 
						|
            return AVERROR(EINVAL);
 | 
						|
 | 
						|
    av_samples_copy(dst->extended_data, src->extended_data, 0, 0,
 | 
						|
                    dst->nb_samples, channels, dst->format);
 | 
						|
 | 
						|
    return 0;
 | 
						|
}
 | 
						|
 | 
						|
int av_frame_copy(AVFrame *dst, const AVFrame *src)
 | 
						|
{
 | 
						|
    if (dst->format != src->format || dst->format < 0)
 | 
						|
        return AVERROR(EINVAL);
 | 
						|
 | 
						|
    if (dst->width > 0 && dst->height > 0)
 | 
						|
        return frame_copy_video(dst, src);
 | 
						|
    else if (dst->nb_samples > 0 && dst->channels > 0)
 | 
						|
        return frame_copy_audio(dst, src);
 | 
						|
 | 
						|
    return AVERROR(EINVAL);
 | 
						|
}
 | 
						|
 | 
						|
void av_frame_remove_side_data(AVFrame *frame, enum AVFrameSideDataType type)
 | 
						|
{
 | 
						|
    int i;
 | 
						|
 | 
						|
    for (i = 0; i < frame->nb_side_data; i++) {
 | 
						|
        AVFrameSideData *sd = frame->side_data[i];
 | 
						|
        if (sd->type == type) {
 | 
						|
            free_side_data(&frame->side_data[i]);
 | 
						|
            frame->side_data[i] = frame->side_data[frame->nb_side_data - 1];
 | 
						|
            frame->nb_side_data--;
 | 
						|
        }
 | 
						|
    }
 | 
						|
}
 | 
						|
 | 
						|
const char *av_frame_side_data_name(enum AVFrameSideDataType type)
 | 
						|
{
 | 
						|
    switch(type) {
 | 
						|
    case AV_FRAME_DATA_PANSCAN:         return "AVPanScan";
 | 
						|
    case AV_FRAME_DATA_A53_CC:          return "ATSC A53 Part 4 Closed Captions";
 | 
						|
    case AV_FRAME_DATA_STEREO3D:        return "Stereo 3D";
 | 
						|
    case AV_FRAME_DATA_MATRIXENCODING:  return "AVMatrixEncoding";
 | 
						|
    case AV_FRAME_DATA_DOWNMIX_INFO:    return "Metadata relevant to a downmix procedure";
 | 
						|
    case AV_FRAME_DATA_REPLAYGAIN:      return "AVReplayGain";
 | 
						|
    case AV_FRAME_DATA_DISPLAYMATRIX:   return "3x3 displaymatrix";
 | 
						|
    case AV_FRAME_DATA_AFD:             return "Active format description";
 | 
						|
    case AV_FRAME_DATA_MOTION_VECTORS:  return "Motion vectors";
 | 
						|
    case AV_FRAME_DATA_SKIP_SAMPLES:    return "Skip samples";
 | 
						|
    case AV_FRAME_DATA_AUDIO_SERVICE_TYPE:          return "Audio service type";
 | 
						|
    case AV_FRAME_DATA_MASTERING_DISPLAY_METADATA:  return "Mastering display metadata";
 | 
						|
    case AV_FRAME_DATA_CONTENT_LIGHT_LEVEL:         return "Content light level metadata";
 | 
						|
    case AV_FRAME_DATA_GOP_TIMECODE:                return "GOP timecode";
 | 
						|
    case AV_FRAME_DATA_S12M_TIMECODE:               return "SMPTE 12-1 timecode";
 | 
						|
    case AV_FRAME_DATA_SPHERICAL:                   return "Spherical Mapping";
 | 
						|
    case AV_FRAME_DATA_ICC_PROFILE:                 return "ICC profile";
 | 
						|
#if FF_API_FRAME_QP
 | 
						|
    case AV_FRAME_DATA_QP_TABLE_PROPERTIES:         return "QP table properties";
 | 
						|
    case AV_FRAME_DATA_QP_TABLE_DATA:               return "QP table data";
 | 
						|
#endif
 | 
						|
    case AV_FRAME_DATA_DYNAMIC_HDR_PLUS: return "HDR Dynamic Metadata SMPTE2094-40 (HDR10+)";
 | 
						|
    case AV_FRAME_DATA_REGIONS_OF_INTEREST: return "Regions Of Interest";
 | 
						|
    }
 | 
						|
    return NULL;
 | 
						|
}
 | 
						|
 | 
						|
static int calc_cropping_offsets(size_t offsets[4], const AVFrame *frame,
 | 
						|
                                 const AVPixFmtDescriptor *desc)
 | 
						|
{
 | 
						|
    int i, j;
 | 
						|
 | 
						|
    for (i = 0; frame->data[i]; i++) {
 | 
						|
        const AVComponentDescriptor *comp = NULL;
 | 
						|
        int shift_x = (i == 1 || i == 2) ? desc->log2_chroma_w : 0;
 | 
						|
        int shift_y = (i == 1 || i == 2) ? desc->log2_chroma_h : 0;
 | 
						|
 | 
						|
        if (desc->flags & (AV_PIX_FMT_FLAG_PAL | FF_PSEUDOPAL) && i == 1) {
 | 
						|
            offsets[i] = 0;
 | 
						|
            break;
 | 
						|
        }
 | 
						|
 | 
						|
        /* find any component descriptor for this plane */
 | 
						|
        for (j = 0; j < desc->nb_components; j++) {
 | 
						|
            if (desc->comp[j].plane == i) {
 | 
						|
                comp = &desc->comp[j];
 | 
						|
                break;
 | 
						|
            }
 | 
						|
        }
 | 
						|
        if (!comp)
 | 
						|
            return AVERROR_BUG;
 | 
						|
 | 
						|
        offsets[i] = (frame->crop_top  >> shift_y) * frame->linesize[i] +
 | 
						|
                     (frame->crop_left >> shift_x) * comp->step;
 | 
						|
    }
 | 
						|
 | 
						|
    return 0;
 | 
						|
}
 | 
						|
 | 
						|
int av_frame_apply_cropping(AVFrame *frame, int flags)
 | 
						|
{
 | 
						|
    const AVPixFmtDescriptor *desc;
 | 
						|
    size_t offsets[4];
 | 
						|
    int i;
 | 
						|
 | 
						|
    if (!(frame->width > 0 && frame->height > 0))
 | 
						|
        return AVERROR(EINVAL);
 | 
						|
 | 
						|
    if (frame->crop_left >= INT_MAX - frame->crop_right        ||
 | 
						|
        frame->crop_top  >= INT_MAX - frame->crop_bottom       ||
 | 
						|
        (frame->crop_left + frame->crop_right) >= frame->width ||
 | 
						|
        (frame->crop_top + frame->crop_bottom) >= frame->height)
 | 
						|
        return AVERROR(ERANGE);
 | 
						|
 | 
						|
    desc = av_pix_fmt_desc_get(frame->format);
 | 
						|
    if (!desc)
 | 
						|
        return AVERROR_BUG;
 | 
						|
 | 
						|
    /* Apply just the right/bottom cropping for hwaccel formats. Bitstream
 | 
						|
     * formats cannot be easily handled here either (and corresponding decoders
 | 
						|
     * should not export any cropping anyway), so do the same for those as well.
 | 
						|
     * */
 | 
						|
    if (desc->flags & (AV_PIX_FMT_FLAG_BITSTREAM | AV_PIX_FMT_FLAG_HWACCEL)) {
 | 
						|
        frame->width      -= frame->crop_right;
 | 
						|
        frame->height     -= frame->crop_bottom;
 | 
						|
        frame->crop_right  = 0;
 | 
						|
        frame->crop_bottom = 0;
 | 
						|
        return 0;
 | 
						|
    }
 | 
						|
 | 
						|
    /* calculate the offsets for each plane */
 | 
						|
    calc_cropping_offsets(offsets, frame, desc);
 | 
						|
 | 
						|
    /* adjust the offsets to avoid breaking alignment */
 | 
						|
    if (!(flags & AV_FRAME_CROP_UNALIGNED)) {
 | 
						|
        int log2_crop_align = frame->crop_left ? ff_ctz(frame->crop_left) : INT_MAX;
 | 
						|
        int min_log2_align = INT_MAX;
 | 
						|
 | 
						|
        for (i = 0; frame->data[i]; i++) {
 | 
						|
            int log2_align = offsets[i] ? ff_ctz(offsets[i]) : INT_MAX;
 | 
						|
            min_log2_align = FFMIN(log2_align, min_log2_align);
 | 
						|
        }
 | 
						|
 | 
						|
        /* we assume, and it should always be true, that the data alignment is
 | 
						|
         * related to the cropping alignment by a constant power-of-2 factor */
 | 
						|
        if (log2_crop_align < min_log2_align)
 | 
						|
            return AVERROR_BUG;
 | 
						|
 | 
						|
        if (min_log2_align < 5) {
 | 
						|
            frame->crop_left &= ~((1 << (5 + log2_crop_align - min_log2_align)) - 1);
 | 
						|
            calc_cropping_offsets(offsets, frame, desc);
 | 
						|
        }
 | 
						|
    }
 | 
						|
 | 
						|
    for (i = 0; frame->data[i]; i++)
 | 
						|
        frame->data[i] += offsets[i];
 | 
						|
 | 
						|
    frame->width      -= (frame->crop_left + frame->crop_right);
 | 
						|
    frame->height     -= (frame->crop_top  + frame->crop_bottom);
 | 
						|
    frame->crop_left   = 0;
 | 
						|
    frame->crop_right  = 0;
 | 
						|
    frame->crop_top    = 0;
 | 
						|
    frame->crop_bottom = 0;
 | 
						|
 | 
						|
    return 0;
 | 
						|
}
 |