1257 lines
		
	
	
		
			42 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1257 lines
		
	
	
		
			42 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						||
 * Videotoolbox hardware acceleration
 | 
						||
 *
 | 
						||
 * copyright (c) 2012 Sebastien Zwickert
 | 
						||
 *
 | 
						||
 * This file is part of FFmpeg.
 | 
						||
 *
 | 
						||
 * FFmpeg is free software; you can redistribute it and/or
 | 
						||
 * modify it under the terms of the GNU Lesser General Public
 | 
						||
 * License as published by the Free Software Foundation; either
 | 
						||
 * version 2.1 of the License, or (at your option) any later version.
 | 
						||
 *
 | 
						||
 * FFmpeg is distributed in the hope that it will be useful,
 | 
						||
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						||
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
 | 
						||
 * Lesser General Public License for more details.
 | 
						||
 *
 | 
						||
 * You should have received a copy of the GNU Lesser General Public
 | 
						||
 * License along with FFmpeg; if not, write to the Free Software
 | 
						||
 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | 
						||
 */
 | 
						||
 | 
						||
#include "config.h"
 | 
						||
#include "videotoolbox.h"
 | 
						||
#include "libavutil/hwcontext_videotoolbox.h"
 | 
						||
#include "vt_internal.h"
 | 
						||
#include "libavutil/avutil.h"
 | 
						||
#include "libavutil/hwcontext.h"
 | 
						||
#include "libavutil/pixdesc.h"
 | 
						||
#include "bytestream.h"
 | 
						||
#include "decode.h"
 | 
						||
#include "h264dec.h"
 | 
						||
#include "hevcdec.h"
 | 
						||
#include "mpegvideo.h"
 | 
						||
#include <TargetConditionals.h>
 | 
						||
 | 
						||
#ifndef kVTVideoDecoderSpecification_RequireHardwareAcceleratedVideoDecoder
 | 
						||
#  define kVTVideoDecoderSpecification_RequireHardwareAcceleratedVideoDecoder CFSTR("RequireHardwareAcceleratedVideoDecoder")
 | 
						||
#endif
 | 
						||
#ifndef kVTVideoDecoderSpecification_EnableHardwareAcceleratedVideoDecoder
 | 
						||
#  define kVTVideoDecoderSpecification_EnableHardwareAcceleratedVideoDecoder CFSTR("EnableHardwareAcceleratedVideoDecoder")
 | 
						||
#endif
 | 
						||
 | 
						||
#if !HAVE_KCMVIDEOCODECTYPE_HEVC
 | 
						||
enum { kCMVideoCodecType_HEVC = 'hvc1' };
 | 
						||
#endif
 | 
						||
 | 
						||
#define VIDEOTOOLBOX_ESDS_EXTRADATA_PADDING  12
 | 
						||
 | 
						||
typedef struct VTHWFrame {
 | 
						||
    CVPixelBufferRef pixbuf;
 | 
						||
    AVBufferRef *hw_frames_ctx;
 | 
						||
} VTHWFrame;
 | 
						||
 | 
						||
static void videotoolbox_buffer_release(void *opaque, uint8_t *data)
 | 
						||
{
 | 
						||
    VTHWFrame *ref = (VTHWFrame *)data;
 | 
						||
    av_buffer_unref(&ref->hw_frames_ctx);
 | 
						||
    CVPixelBufferRelease(ref->pixbuf);
 | 
						||
 | 
						||
    av_free(data);
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_buffer_copy(VTContext *vtctx,
 | 
						||
                                    const uint8_t *buffer,
 | 
						||
                                    uint32_t size)
 | 
						||
{
 | 
						||
    void *tmp;
 | 
						||
 | 
						||
    tmp = av_fast_realloc(vtctx->bitstream,
 | 
						||
                         &vtctx->allocated_size,
 | 
						||
                         size);
 | 
						||
 | 
						||
    if (!tmp)
 | 
						||
        return AVERROR(ENOMEM);
 | 
						||
 | 
						||
    vtctx->bitstream = tmp;
 | 
						||
    memcpy(vtctx->bitstream, buffer, size);
 | 
						||
    vtctx->bitstream_size = size;
 | 
						||
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_postproc_frame(void *avctx, AVFrame *frame)
 | 
						||
{
 | 
						||
    VTHWFrame *ref = (VTHWFrame *)frame->buf[0]->data;
 | 
						||
 | 
						||
    if (!ref->pixbuf) {
 | 
						||
        av_log(avctx, AV_LOG_ERROR, "No frame decoded?\n");
 | 
						||
        av_frame_unref(frame);
 | 
						||
        return AVERROR_EXTERNAL;
 | 
						||
    }
 | 
						||
 | 
						||
    frame->crop_right = 0;
 | 
						||
    frame->crop_left = 0;
 | 
						||
    frame->crop_top = 0;
 | 
						||
    frame->crop_bottom = 0;
 | 
						||
 | 
						||
    frame->data[3] = (uint8_t*)ref->pixbuf;
 | 
						||
 | 
						||
    if (ref->hw_frames_ctx) {
 | 
						||
        av_buffer_unref(&frame->hw_frames_ctx);
 | 
						||
        frame->hw_frames_ctx = av_buffer_ref(ref->hw_frames_ctx);
 | 
						||
        if (!frame->hw_frames_ctx)
 | 
						||
            return AVERROR(ENOMEM);
 | 
						||
    }
 | 
						||
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
int ff_videotoolbox_alloc_frame(AVCodecContext *avctx, AVFrame *frame)
 | 
						||
{
 | 
						||
    size_t      size = sizeof(VTHWFrame);
 | 
						||
    uint8_t    *data = NULL;
 | 
						||
    AVBufferRef *buf = NULL;
 | 
						||
    int ret = ff_attach_decode_data(frame);
 | 
						||
    FrameDecodeData *fdd;
 | 
						||
    if (ret < 0)
 | 
						||
        return ret;
 | 
						||
 | 
						||
    data = av_mallocz(size);
 | 
						||
    if (!data)
 | 
						||
        return AVERROR(ENOMEM);
 | 
						||
    buf = av_buffer_create(data, size, videotoolbox_buffer_release, NULL, 0);
 | 
						||
    if (!buf) {
 | 
						||
        av_freep(&data);
 | 
						||
        return AVERROR(ENOMEM);
 | 
						||
    }
 | 
						||
    frame->buf[0] = buf;
 | 
						||
 | 
						||
    fdd = (FrameDecodeData*)frame->private_ref->data;
 | 
						||
    fdd->post_process = videotoolbox_postproc_frame;
 | 
						||
 | 
						||
    frame->width  = avctx->width;
 | 
						||
    frame->height = avctx->height;
 | 
						||
    frame->format = avctx->pix_fmt;
 | 
						||
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
#define AV_W8(p, v) *(p) = (v)
 | 
						||
 | 
						||
CFDataRef ff_videotoolbox_avcc_extradata_create(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
    H264Context *h = avctx->priv_data;
 | 
						||
    CFDataRef data = NULL;
 | 
						||
    uint8_t *p;
 | 
						||
    int vt_extradata_size = 6 + 2 + h->ps.sps->data_size + 3 + h->ps.pps->data_size;
 | 
						||
    uint8_t *vt_extradata = av_malloc(vt_extradata_size);
 | 
						||
    if (!vt_extradata)
 | 
						||
        return NULL;
 | 
						||
 | 
						||
    p = vt_extradata;
 | 
						||
 | 
						||
    AV_W8(p + 0, 1); /* version */
 | 
						||
    AV_W8(p + 1, h->ps.sps->data[1]); /* profile */
 | 
						||
    AV_W8(p + 2, h->ps.sps->data[2]); /* profile compat */
 | 
						||
    AV_W8(p + 3, h->ps.sps->data[3]); /* level */
 | 
						||
    AV_W8(p + 4, 0xff); /* 6 bits reserved (111111) + 2 bits nal size length - 3 (11) */
 | 
						||
    AV_W8(p + 5, 0xe1); /* 3 bits reserved (111) + 5 bits number of sps (00001) */
 | 
						||
    AV_WB16(p + 6, h->ps.sps->data_size);
 | 
						||
    memcpy(p + 8, h->ps.sps->data, h->ps.sps->data_size);
 | 
						||
    p += 8 + h->ps.sps->data_size;
 | 
						||
    AV_W8(p + 0, 1); /* number of pps */
 | 
						||
    AV_WB16(p + 1, h->ps.pps->data_size);
 | 
						||
    memcpy(p + 3, h->ps.pps->data, h->ps.pps->data_size);
 | 
						||
 | 
						||
    p += 3 + h->ps.pps->data_size;
 | 
						||
    av_assert0(p - vt_extradata == vt_extradata_size);
 | 
						||
 | 
						||
    // save sps header (profile/level) used to create decoder session,
 | 
						||
    // so we can detect changes and recreate it.
 | 
						||
    if (vtctx)
 | 
						||
        memcpy(vtctx->sps, h->ps.sps->data + 1, 3);
 | 
						||
 | 
						||
    data = CFDataCreate(kCFAllocatorDefault, vt_extradata, vt_extradata_size);
 | 
						||
    av_free(vt_extradata);
 | 
						||
    return data;
 | 
						||
}
 | 
						||
 | 
						||
CFDataRef ff_videotoolbox_hvcc_extradata_create(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    HEVCContext *h = avctx->priv_data;
 | 
						||
    int i, num_vps = 0, num_sps = 0, num_pps = 0;
 | 
						||
    const HEVCVPS *vps = h->ps.vps;
 | 
						||
    const HEVCSPS *sps = h->ps.sps;
 | 
						||
    const HEVCPPS *pps = h->ps.pps;
 | 
						||
    PTLCommon ptlc = vps->ptl.general_ptl;
 | 
						||
    VUI vui = sps->vui;
 | 
						||
    uint8_t parallelismType;
 | 
						||
    CFDataRef data = NULL;
 | 
						||
    uint8_t *p;
 | 
						||
    int vt_extradata_size = 23 + 3 + 3 + 3;
 | 
						||
    uint8_t *vt_extradata;
 | 
						||
 | 
						||
#define COUNT_SIZE_PS(T, t) \
 | 
						||
    for (i = 0; i < HEVC_MAX_##T##PS_COUNT; i++) { \
 | 
						||
        if (h->ps.t##ps_list[i]) { \
 | 
						||
            const HEVC##T##PS *lps = (const HEVC##T##PS *)h->ps.t##ps_list[i]->data; \
 | 
						||
            vt_extradata_size += 2 + lps->data_size; \
 | 
						||
            num_##t##ps++; \
 | 
						||
        } \
 | 
						||
    }
 | 
						||
 | 
						||
    COUNT_SIZE_PS(V, v)
 | 
						||
    COUNT_SIZE_PS(S, s)
 | 
						||
    COUNT_SIZE_PS(P, p)
 | 
						||
 | 
						||
    vt_extradata = av_malloc(vt_extradata_size);
 | 
						||
    if (!vt_extradata)
 | 
						||
        return NULL;
 | 
						||
    p = vt_extradata;
 | 
						||
 | 
						||
    /* unsigned int(8) configurationVersion = 1; */
 | 
						||
    AV_W8(p + 0, 1);
 | 
						||
 | 
						||
    /*
 | 
						||
     * unsigned int(2) general_profile_space;
 | 
						||
     * unsigned int(1) general_tier_flag;
 | 
						||
     * unsigned int(5) general_profile_idc;
 | 
						||
     */
 | 
						||
    AV_W8(p + 1, ptlc.profile_space << 6 |
 | 
						||
                 ptlc.tier_flag     << 5 |
 | 
						||
                 ptlc.profile_idc);
 | 
						||
 | 
						||
    /* unsigned int(32) general_profile_compatibility_flags; */
 | 
						||
    memcpy(p + 2, ptlc.profile_compatibility_flag, 4);
 | 
						||
 | 
						||
    /* unsigned int(48) general_constraint_indicator_flags; */
 | 
						||
    AV_W8(p + 6, ptlc.progressive_source_flag    << 7 |
 | 
						||
                 ptlc.interlaced_source_flag     << 6 |
 | 
						||
                 ptlc.non_packed_constraint_flag << 5 |
 | 
						||
                 ptlc.frame_only_constraint_flag << 4);
 | 
						||
    AV_W8(p + 7, 0);
 | 
						||
    AV_WN32(p + 8, 0);
 | 
						||
 | 
						||
    /* unsigned int(8) general_level_idc; */
 | 
						||
    AV_W8(p + 12, ptlc.level_idc);
 | 
						||
 | 
						||
    /*
 | 
						||
     * bit(4) reserved = ‘1111’b;
 | 
						||
     * unsigned int(12) min_spatial_segmentation_idc;
 | 
						||
     */
 | 
						||
    AV_W8(p + 13, 0xf0 | (vui.min_spatial_segmentation_idc >> 4));
 | 
						||
    AV_W8(p + 14, vui.min_spatial_segmentation_idc & 0xff);
 | 
						||
 | 
						||
    /*
 | 
						||
     * bit(6) reserved = ‘111111’b;
 | 
						||
     * unsigned int(2) parallelismType;
 | 
						||
     */
 | 
						||
    if (!vui.min_spatial_segmentation_idc)
 | 
						||
        parallelismType = 0;
 | 
						||
    else if (pps->entropy_coding_sync_enabled_flag && pps->tiles_enabled_flag)
 | 
						||
        parallelismType = 0;
 | 
						||
    else if (pps->entropy_coding_sync_enabled_flag)
 | 
						||
        parallelismType = 3;
 | 
						||
    else if (pps->tiles_enabled_flag)
 | 
						||
        parallelismType = 2;
 | 
						||
    else
 | 
						||
        parallelismType = 1;
 | 
						||
    AV_W8(p + 15, 0xfc | parallelismType);
 | 
						||
 | 
						||
    /*
 | 
						||
     * bit(6) reserved = ‘111111’b;
 | 
						||
     * unsigned int(2) chromaFormat;
 | 
						||
     */
 | 
						||
    AV_W8(p + 16, sps->chroma_format_idc | 0xfc);
 | 
						||
 | 
						||
    /*
 | 
						||
     * bit(5) reserved = ‘11111’b;
 | 
						||
     * unsigned int(3) bitDepthLumaMinus8;
 | 
						||
     */
 | 
						||
    AV_W8(p + 17, (sps->bit_depth - 8) | 0xfc);
 | 
						||
 | 
						||
    /*
 | 
						||
     * bit(5) reserved = ‘11111’b;
 | 
						||
     * unsigned int(3) bitDepthChromaMinus8;
 | 
						||
     */
 | 
						||
    AV_W8(p + 18, (sps->bit_depth_chroma - 8) | 0xfc);
 | 
						||
 | 
						||
    /* bit(16) avgFrameRate; */
 | 
						||
    AV_WB16(p + 19, 0);
 | 
						||
 | 
						||
    /*
 | 
						||
     * bit(2) constantFrameRate;
 | 
						||
     * bit(3) numTemporalLayers;
 | 
						||
     * bit(1) temporalIdNested;
 | 
						||
     * unsigned int(2) lengthSizeMinusOne;
 | 
						||
     */
 | 
						||
    AV_W8(p + 21, 0                             << 6 |
 | 
						||
                  sps->max_sub_layers           << 3 |
 | 
						||
                  sps->temporal_id_nesting_flag << 2 |
 | 
						||
                  3);
 | 
						||
 | 
						||
    /* unsigned int(8) numOfArrays; */
 | 
						||
    AV_W8(p + 22, 3);
 | 
						||
 | 
						||
    p += 23;
 | 
						||
 | 
						||
#define APPEND_PS(T, t) \
 | 
						||
    /* \
 | 
						||
     * bit(1) array_completeness; \
 | 
						||
     * unsigned int(1) reserved = 0; \
 | 
						||
     * unsigned int(6) NAL_unit_type; \
 | 
						||
     */ \
 | 
						||
    AV_W8(p, 1 << 7 | \
 | 
						||
             HEVC_NAL_##T##PS & 0x3f); \
 | 
						||
    /* unsigned int(16) numNalus; */ \
 | 
						||
    AV_WB16(p + 1, num_##t##ps); \
 | 
						||
    p += 3; \
 | 
						||
    for (i = 0; i < HEVC_MAX_##T##PS_COUNT; i++) { \
 | 
						||
        if (h->ps.t##ps_list[i]) { \
 | 
						||
            const HEVC##T##PS *lps = (const HEVC##T##PS *)h->ps.t##ps_list[i]->data; \
 | 
						||
            /* unsigned int(16) nalUnitLength; */ \
 | 
						||
            AV_WB16(p, lps->data_size); \
 | 
						||
            /* bit(8*nalUnitLength) nalUnit; */ \
 | 
						||
            memcpy(p + 2, lps->data, lps->data_size); \
 | 
						||
            p += 2 + lps->data_size; \
 | 
						||
        } \
 | 
						||
    }
 | 
						||
 | 
						||
    APPEND_PS(V, v)
 | 
						||
    APPEND_PS(S, s)
 | 
						||
    APPEND_PS(P, p)
 | 
						||
 | 
						||
    av_assert0(p - vt_extradata == vt_extradata_size);
 | 
						||
 | 
						||
    data = CFDataCreate(kCFAllocatorDefault, vt_extradata, vt_extradata_size);
 | 
						||
    av_free(vt_extradata);
 | 
						||
    return data;
 | 
						||
}
 | 
						||
 | 
						||
int ff_videotoolbox_h264_start_frame(AVCodecContext *avctx,
 | 
						||
                                     const uint8_t *buffer,
 | 
						||
                                     uint32_t size)
 | 
						||
{
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
    H264Context *h = avctx->priv_data;
 | 
						||
 | 
						||
    if (h->is_avc == 1) {
 | 
						||
        return videotoolbox_buffer_copy(vtctx, buffer, size);
 | 
						||
    }
 | 
						||
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_h264_decode_params(AVCodecContext *avctx,
 | 
						||
                                           int type,
 | 
						||
                                           const uint8_t *buffer,
 | 
						||
                                           uint32_t size)
 | 
						||
{
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
    H264Context *h = avctx->priv_data;
 | 
						||
 | 
						||
    // save sps header (profile/level) used to create decoder session
 | 
						||
    if (!vtctx->sps[0])
 | 
						||
        memcpy(vtctx->sps, h->ps.sps->data + 1, 3);
 | 
						||
 | 
						||
    if (type == H264_NAL_SPS) {
 | 
						||
        if (size > 4 && memcmp(vtctx->sps, buffer + 1, 3) != 0) {
 | 
						||
            vtctx->reconfig_needed = true;
 | 
						||
            memcpy(vtctx->sps, buffer + 1, 3);
 | 
						||
        }
 | 
						||
    }
 | 
						||
 | 
						||
    // pass-through SPS/PPS changes to the decoder
 | 
						||
    return ff_videotoolbox_h264_decode_slice(avctx, buffer, size);
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_common_decode_slice(AVCodecContext *avctx,
 | 
						||
                                            const uint8_t *buffer,
 | 
						||
                                            uint32_t size)
 | 
						||
{
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
    void *tmp;
 | 
						||
 | 
						||
    tmp = av_fast_realloc(vtctx->bitstream,
 | 
						||
                          &vtctx->allocated_size,
 | 
						||
                          vtctx->bitstream_size+size+4);
 | 
						||
    if (!tmp)
 | 
						||
        return AVERROR(ENOMEM);
 | 
						||
 | 
						||
    vtctx->bitstream = tmp;
 | 
						||
 | 
						||
    AV_WB32(vtctx->bitstream + vtctx->bitstream_size, size);
 | 
						||
    memcpy(vtctx->bitstream + vtctx->bitstream_size + 4, buffer, size);
 | 
						||
 | 
						||
    vtctx->bitstream_size += size + 4;
 | 
						||
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
int ff_videotoolbox_h264_decode_slice(AVCodecContext *avctx,
 | 
						||
                                      const uint8_t *buffer,
 | 
						||
                                      uint32_t size)
 | 
						||
{
 | 
						||
    H264Context *h = avctx->priv_data;
 | 
						||
 | 
						||
    if (h->is_avc == 1)
 | 
						||
        return 0;
 | 
						||
 | 
						||
    return videotoolbox_common_decode_slice(avctx, buffer, size);
 | 
						||
}
 | 
						||
 | 
						||
int ff_videotoolbox_uninit(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
    if (vtctx) {
 | 
						||
        av_freep(&vtctx->bitstream);
 | 
						||
        if (vtctx->frame)
 | 
						||
            CVPixelBufferRelease(vtctx->frame);
 | 
						||
    }
 | 
						||
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
#if CONFIG_VIDEOTOOLBOX
 | 
						||
// Return the AVVideotoolboxContext that matters currently. Where it comes from
 | 
						||
// depends on the API used.
 | 
						||
static AVVideotoolboxContext *videotoolbox_get_context(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    // Somewhat tricky because the user can call av_videotoolbox_default_free()
 | 
						||
    // at any time, even when the codec is closed.
 | 
						||
    if (avctx->internal && avctx->internal->hwaccel_priv_data) {
 | 
						||
        VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
        if (vtctx->vt_ctx)
 | 
						||
            return vtctx->vt_ctx;
 | 
						||
    }
 | 
						||
    return avctx->hwaccel_context;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_buffer_create(AVCodecContext *avctx, AVFrame *frame)
 | 
						||
{
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
    CVPixelBufferRef pixbuf = (CVPixelBufferRef)vtctx->frame;
 | 
						||
    OSType pixel_format = CVPixelBufferGetPixelFormatType(pixbuf);
 | 
						||
    enum AVPixelFormat sw_format = av_map_videotoolbox_format_to_pixfmt(pixel_format);
 | 
						||
    int width = CVPixelBufferGetWidth(pixbuf);
 | 
						||
    int height = CVPixelBufferGetHeight(pixbuf);
 | 
						||
    AVHWFramesContext *cached_frames;
 | 
						||
    VTHWFrame *ref;
 | 
						||
    int ret;
 | 
						||
 | 
						||
    if (!frame->buf[0] || frame->data[3]) {
 | 
						||
        av_log(avctx, AV_LOG_ERROR, "videotoolbox: invalid state\n");
 | 
						||
        av_frame_unref(frame);
 | 
						||
        return AVERROR_EXTERNAL;
 | 
						||
    }
 | 
						||
 | 
						||
    ref = (VTHWFrame *)frame->buf[0]->data;
 | 
						||
 | 
						||
    if (ref->pixbuf)
 | 
						||
        CVPixelBufferRelease(ref->pixbuf);
 | 
						||
    ref->pixbuf = vtctx->frame;
 | 
						||
    vtctx->frame = NULL;
 | 
						||
 | 
						||
    // Old API code path.
 | 
						||
    if (!vtctx->cached_hw_frames_ctx)
 | 
						||
        return 0;
 | 
						||
 | 
						||
    cached_frames = (AVHWFramesContext*)vtctx->cached_hw_frames_ctx->data;
 | 
						||
 | 
						||
    if (cached_frames->sw_format != sw_format ||
 | 
						||
        cached_frames->width != width ||
 | 
						||
        cached_frames->height != height) {
 | 
						||
        AVBufferRef *hw_frames_ctx = av_hwframe_ctx_alloc(cached_frames->device_ref);
 | 
						||
        AVHWFramesContext *hw_frames;
 | 
						||
        if (!hw_frames_ctx)
 | 
						||
            return AVERROR(ENOMEM);
 | 
						||
 | 
						||
        hw_frames = (AVHWFramesContext*)hw_frames_ctx->data;
 | 
						||
        hw_frames->format = cached_frames->format;
 | 
						||
        hw_frames->sw_format = sw_format;
 | 
						||
        hw_frames->width = width;
 | 
						||
        hw_frames->height = height;
 | 
						||
 | 
						||
        ret = av_hwframe_ctx_init(hw_frames_ctx);
 | 
						||
        if (ret < 0) {
 | 
						||
            av_buffer_unref(&hw_frames_ctx);
 | 
						||
            return ret;
 | 
						||
        }
 | 
						||
 | 
						||
        av_buffer_unref(&vtctx->cached_hw_frames_ctx);
 | 
						||
        vtctx->cached_hw_frames_ctx = hw_frames_ctx;
 | 
						||
    }
 | 
						||
 | 
						||
    av_buffer_unref(&ref->hw_frames_ctx);
 | 
						||
    ref->hw_frames_ctx = av_buffer_ref(vtctx->cached_hw_frames_ctx);
 | 
						||
    if (!ref->hw_frames_ctx)
 | 
						||
        return AVERROR(ENOMEM);
 | 
						||
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
static void videotoolbox_write_mp4_descr_length(PutByteContext *pb, int length)
 | 
						||
{
 | 
						||
    int i;
 | 
						||
    uint8_t b;
 | 
						||
 | 
						||
    for (i = 3; i >= 0; i--) {
 | 
						||
        b = (length >> (i * 7)) & 0x7F;
 | 
						||
        if (i != 0)
 | 
						||
            b |= 0x80;
 | 
						||
 | 
						||
        bytestream2_put_byteu(pb, b);
 | 
						||
    }
 | 
						||
}
 | 
						||
 | 
						||
static CFDataRef videotoolbox_esds_extradata_create(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    CFDataRef data;
 | 
						||
    uint8_t *rw_extradata;
 | 
						||
    PutByteContext pb;
 | 
						||
    int full_size = 3 + 5 + 13 + 5 + avctx->extradata_size + 3;
 | 
						||
    // ES_DescrTag data + DecoderConfigDescrTag + data + DecSpecificInfoTag + size + SLConfigDescriptor
 | 
						||
    int config_size = 13 + 5 + avctx->extradata_size;
 | 
						||
    int s;
 | 
						||
 | 
						||
    if (!(rw_extradata = av_mallocz(full_size + VIDEOTOOLBOX_ESDS_EXTRADATA_PADDING)))
 | 
						||
        return NULL;
 | 
						||
 | 
						||
    bytestream2_init_writer(&pb, rw_extradata, full_size + VIDEOTOOLBOX_ESDS_EXTRADATA_PADDING);
 | 
						||
    bytestream2_put_byteu(&pb, 0);        // version
 | 
						||
    bytestream2_put_ne24(&pb, 0);         // flags
 | 
						||
 | 
						||
    // elementary stream descriptor
 | 
						||
    bytestream2_put_byteu(&pb, 0x03);     // ES_DescrTag
 | 
						||
    videotoolbox_write_mp4_descr_length(&pb, full_size);
 | 
						||
    bytestream2_put_ne16(&pb, 0);         // esid
 | 
						||
    bytestream2_put_byteu(&pb, 0);        // stream priority (0-32)
 | 
						||
 | 
						||
    // decoder configuration descriptor
 | 
						||
    bytestream2_put_byteu(&pb, 0x04);     // DecoderConfigDescrTag
 | 
						||
    videotoolbox_write_mp4_descr_length(&pb, config_size);
 | 
						||
    bytestream2_put_byteu(&pb, 32);       // object type indication. 32 = AV_CODEC_ID_MPEG4
 | 
						||
    bytestream2_put_byteu(&pb, 0x11);     // stream type
 | 
						||
    bytestream2_put_ne24(&pb, 0);         // buffer size
 | 
						||
    bytestream2_put_ne32(&pb, 0);         // max bitrate
 | 
						||
    bytestream2_put_ne32(&pb, 0);         // avg bitrate
 | 
						||
 | 
						||
    // decoder specific descriptor
 | 
						||
    bytestream2_put_byteu(&pb, 0x05);     ///< DecSpecificInfoTag
 | 
						||
    videotoolbox_write_mp4_descr_length(&pb, avctx->extradata_size);
 | 
						||
 | 
						||
    bytestream2_put_buffer(&pb, avctx->extradata, avctx->extradata_size);
 | 
						||
 | 
						||
    // SLConfigDescriptor
 | 
						||
    bytestream2_put_byteu(&pb, 0x06);     // SLConfigDescrTag
 | 
						||
    bytestream2_put_byteu(&pb, 0x01);     // length
 | 
						||
    bytestream2_put_byteu(&pb, 0x02);     //
 | 
						||
 | 
						||
    s = bytestream2_size_p(&pb);
 | 
						||
 | 
						||
    data = CFDataCreate(kCFAllocatorDefault, rw_extradata, s);
 | 
						||
 | 
						||
    av_freep(&rw_extradata);
 | 
						||
    return data;
 | 
						||
}
 | 
						||
 | 
						||
static CMSampleBufferRef videotoolbox_sample_buffer_create(CMFormatDescriptionRef fmt_desc,
 | 
						||
                                                           void *buffer,
 | 
						||
                                                           int size)
 | 
						||
{
 | 
						||
    OSStatus status;
 | 
						||
    CMBlockBufferRef  block_buf;
 | 
						||
    CMSampleBufferRef sample_buf;
 | 
						||
 | 
						||
    block_buf  = NULL;
 | 
						||
    sample_buf = NULL;
 | 
						||
 | 
						||
    status = CMBlockBufferCreateWithMemoryBlock(kCFAllocatorDefault,// structureAllocator
 | 
						||
                                                buffer,             // memoryBlock
 | 
						||
                                                size,               // blockLength
 | 
						||
                                                kCFAllocatorNull,   // blockAllocator
 | 
						||
                                                NULL,               // customBlockSource
 | 
						||
                                                0,                  // offsetToData
 | 
						||
                                                size,               // dataLength
 | 
						||
                                                0,                  // flags
 | 
						||
                                                &block_buf);
 | 
						||
 | 
						||
    if (!status) {
 | 
						||
        status = CMSampleBufferCreate(kCFAllocatorDefault,  // allocator
 | 
						||
                                      block_buf,            // dataBuffer
 | 
						||
                                      TRUE,                 // dataReady
 | 
						||
                                      0,                    // makeDataReadyCallback
 | 
						||
                                      0,                    // makeDataReadyRefcon
 | 
						||
                                      fmt_desc,             // formatDescription
 | 
						||
                                      1,                    // numSamples
 | 
						||
                                      0,                    // numSampleTimingEntries
 | 
						||
                                      NULL,                 // sampleTimingArray
 | 
						||
                                      0,                    // numSampleSizeEntries
 | 
						||
                                      NULL,                 // sampleSizeArray
 | 
						||
                                      &sample_buf);
 | 
						||
    }
 | 
						||
 | 
						||
    if (block_buf)
 | 
						||
        CFRelease(block_buf);
 | 
						||
 | 
						||
    return sample_buf;
 | 
						||
}
 | 
						||
 | 
						||
static void videotoolbox_decoder_callback(void *opaque,
 | 
						||
                                          void *sourceFrameRefCon,
 | 
						||
                                          OSStatus status,
 | 
						||
                                          VTDecodeInfoFlags flags,
 | 
						||
                                          CVImageBufferRef image_buffer,
 | 
						||
                                          CMTime pts,
 | 
						||
                                          CMTime duration)
 | 
						||
{
 | 
						||
    AVCodecContext *avctx = opaque;
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
 | 
						||
    if (vtctx->frame) {
 | 
						||
        CVPixelBufferRelease(vtctx->frame);
 | 
						||
        vtctx->frame = NULL;
 | 
						||
    }
 | 
						||
 | 
						||
    if (!image_buffer) {
 | 
						||
        av_log(avctx, AV_LOG_DEBUG, "vt decoder cb: output image buffer is null\n");
 | 
						||
        return;
 | 
						||
    }
 | 
						||
 | 
						||
    vtctx->frame = CVPixelBufferRetain(image_buffer);
 | 
						||
}
 | 
						||
 | 
						||
static OSStatus videotoolbox_session_decode_frame(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    OSStatus status;
 | 
						||
    CMSampleBufferRef sample_buf;
 | 
						||
    AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx);
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
 | 
						||
    sample_buf = videotoolbox_sample_buffer_create(videotoolbox->cm_fmt_desc,
 | 
						||
                                                   vtctx->bitstream,
 | 
						||
                                                   vtctx->bitstream_size);
 | 
						||
 | 
						||
    if (!sample_buf)
 | 
						||
        return -1;
 | 
						||
 | 
						||
    status = VTDecompressionSessionDecodeFrame(videotoolbox->session,
 | 
						||
                                               sample_buf,
 | 
						||
                                               0,       // decodeFlags
 | 
						||
                                               NULL,    // sourceFrameRefCon
 | 
						||
                                               0);      // infoFlagsOut
 | 
						||
    if (status == noErr)
 | 
						||
        status = VTDecompressionSessionWaitForAsynchronousFrames(videotoolbox->session);
 | 
						||
 | 
						||
    CFRelease(sample_buf);
 | 
						||
 | 
						||
    return status;
 | 
						||
}
 | 
						||
 | 
						||
static CMVideoFormatDescriptionRef videotoolbox_format_desc_create(CMVideoCodecType codec_type,
 | 
						||
                                                                   CFDictionaryRef decoder_spec,
 | 
						||
                                                                   int width,
 | 
						||
                                                                   int height)
 | 
						||
{
 | 
						||
    CMFormatDescriptionRef cm_fmt_desc;
 | 
						||
    OSStatus status;
 | 
						||
 | 
						||
    status = CMVideoFormatDescriptionCreate(kCFAllocatorDefault,
 | 
						||
                                            codec_type,
 | 
						||
                                            width,
 | 
						||
                                            height,
 | 
						||
                                            decoder_spec, // Dictionary of extension
 | 
						||
                                            &cm_fmt_desc);
 | 
						||
 | 
						||
    if (status)
 | 
						||
        return NULL;
 | 
						||
 | 
						||
    return cm_fmt_desc;
 | 
						||
}
 | 
						||
 | 
						||
static CFDictionaryRef videotoolbox_buffer_attributes_create(int width,
 | 
						||
                                                             int height,
 | 
						||
                                                             OSType pix_fmt)
 | 
						||
{
 | 
						||
    CFMutableDictionaryRef buffer_attributes;
 | 
						||
    CFMutableDictionaryRef io_surface_properties;
 | 
						||
    CFNumberRef cv_pix_fmt;
 | 
						||
    CFNumberRef w;
 | 
						||
    CFNumberRef h;
 | 
						||
 | 
						||
    w = CFNumberCreate(kCFAllocatorDefault, kCFNumberSInt32Type, &width);
 | 
						||
    h = CFNumberCreate(kCFAllocatorDefault, kCFNumberSInt32Type, &height);
 | 
						||
    cv_pix_fmt = CFNumberCreate(kCFAllocatorDefault, kCFNumberSInt32Type, &pix_fmt);
 | 
						||
 | 
						||
    buffer_attributes = CFDictionaryCreateMutable(kCFAllocatorDefault,
 | 
						||
                                                  4,
 | 
						||
                                                  &kCFTypeDictionaryKeyCallBacks,
 | 
						||
                                                  &kCFTypeDictionaryValueCallBacks);
 | 
						||
    io_surface_properties = CFDictionaryCreateMutable(kCFAllocatorDefault,
 | 
						||
                                                      0,
 | 
						||
                                                      &kCFTypeDictionaryKeyCallBacks,
 | 
						||
                                                      &kCFTypeDictionaryValueCallBacks);
 | 
						||
 | 
						||
    if (pix_fmt)
 | 
						||
        CFDictionarySetValue(buffer_attributes, kCVPixelBufferPixelFormatTypeKey, cv_pix_fmt);
 | 
						||
    CFDictionarySetValue(buffer_attributes, kCVPixelBufferIOSurfacePropertiesKey, io_surface_properties);
 | 
						||
    CFDictionarySetValue(buffer_attributes, kCVPixelBufferWidthKey, w);
 | 
						||
    CFDictionarySetValue(buffer_attributes, kCVPixelBufferHeightKey, h);
 | 
						||
#if TARGET_OS_IPHONE
 | 
						||
    CFDictionarySetValue(buffer_attributes, kCVPixelBufferOpenGLESCompatibilityKey, kCFBooleanTrue);
 | 
						||
#else
 | 
						||
    CFDictionarySetValue(buffer_attributes, kCVPixelBufferIOSurfaceOpenGLTextureCompatibilityKey, kCFBooleanTrue);
 | 
						||
#endif
 | 
						||
 | 
						||
    CFRelease(io_surface_properties);
 | 
						||
    CFRelease(cv_pix_fmt);
 | 
						||
    CFRelease(w);
 | 
						||
    CFRelease(h);
 | 
						||
 | 
						||
    return buffer_attributes;
 | 
						||
}
 | 
						||
 | 
						||
static CFDictionaryRef videotoolbox_decoder_config_create(CMVideoCodecType codec_type,
 | 
						||
                                                          AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    CFMutableDictionaryRef config_info = CFDictionaryCreateMutable(kCFAllocatorDefault,
 | 
						||
                                                                   0,
 | 
						||
                                                                   &kCFTypeDictionaryKeyCallBacks,
 | 
						||
                                                                   &kCFTypeDictionaryValueCallBacks);
 | 
						||
 | 
						||
    CFDictionarySetValue(config_info,
 | 
						||
                         codec_type == kCMVideoCodecType_HEVC ?
 | 
						||
                            kVTVideoDecoderSpecification_EnableHardwareAcceleratedVideoDecoder :
 | 
						||
                            kVTVideoDecoderSpecification_RequireHardwareAcceleratedVideoDecoder,
 | 
						||
                         kCFBooleanTrue);
 | 
						||
 | 
						||
    CFMutableDictionaryRef avc_info;
 | 
						||
    CFDataRef data = NULL;
 | 
						||
 | 
						||
    avc_info = CFDictionaryCreateMutable(kCFAllocatorDefault,
 | 
						||
                                         1,
 | 
						||
                                         &kCFTypeDictionaryKeyCallBacks,
 | 
						||
                                         &kCFTypeDictionaryValueCallBacks);
 | 
						||
 | 
						||
    switch (codec_type) {
 | 
						||
    case kCMVideoCodecType_MPEG4Video :
 | 
						||
        if (avctx->extradata_size)
 | 
						||
            data = videotoolbox_esds_extradata_create(avctx);
 | 
						||
        if (data)
 | 
						||
            CFDictionarySetValue(avc_info, CFSTR("esds"), data);
 | 
						||
        break;
 | 
						||
    case kCMVideoCodecType_H264 :
 | 
						||
        data = ff_videotoolbox_avcc_extradata_create(avctx);
 | 
						||
        if (data)
 | 
						||
            CFDictionarySetValue(avc_info, CFSTR("avcC"), data);
 | 
						||
        break;
 | 
						||
    case kCMVideoCodecType_HEVC :
 | 
						||
        data = ff_videotoolbox_hvcc_extradata_create(avctx);
 | 
						||
        if (data)
 | 
						||
            CFDictionarySetValue(avc_info, CFSTR("hvcC"), data);
 | 
						||
        break;
 | 
						||
    default:
 | 
						||
        break;
 | 
						||
    }
 | 
						||
 | 
						||
    CFDictionarySetValue(config_info,
 | 
						||
            kCMFormatDescriptionExtension_SampleDescriptionExtensionAtoms,
 | 
						||
            avc_info);
 | 
						||
 | 
						||
    if (data)
 | 
						||
        CFRelease(data);
 | 
						||
 | 
						||
    CFRelease(avc_info);
 | 
						||
    return config_info;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_start(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx);
 | 
						||
    OSStatus status;
 | 
						||
    VTDecompressionOutputCallbackRecord decoder_cb;
 | 
						||
    CFDictionaryRef decoder_spec;
 | 
						||
    CFDictionaryRef buf_attr;
 | 
						||
 | 
						||
    if (!videotoolbox) {
 | 
						||
        av_log(avctx, AV_LOG_ERROR, "hwaccel context is not set\n");
 | 
						||
        return -1;
 | 
						||
    }
 | 
						||
 | 
						||
    switch( avctx->codec_id ) {
 | 
						||
    case AV_CODEC_ID_H263 :
 | 
						||
        videotoolbox->cm_codec_type = kCMVideoCodecType_H263;
 | 
						||
        break;
 | 
						||
    case AV_CODEC_ID_H264 :
 | 
						||
        videotoolbox->cm_codec_type = kCMVideoCodecType_H264;
 | 
						||
        break;
 | 
						||
    case AV_CODEC_ID_HEVC :
 | 
						||
        videotoolbox->cm_codec_type = kCMVideoCodecType_HEVC;
 | 
						||
        break;
 | 
						||
    case AV_CODEC_ID_MPEG1VIDEO :
 | 
						||
        videotoolbox->cm_codec_type = kCMVideoCodecType_MPEG1Video;
 | 
						||
        break;
 | 
						||
    case AV_CODEC_ID_MPEG2VIDEO :
 | 
						||
        videotoolbox->cm_codec_type = kCMVideoCodecType_MPEG2Video;
 | 
						||
        break;
 | 
						||
    case AV_CODEC_ID_MPEG4 :
 | 
						||
        videotoolbox->cm_codec_type = kCMVideoCodecType_MPEG4Video;
 | 
						||
        break;
 | 
						||
    default :
 | 
						||
        break;
 | 
						||
    }
 | 
						||
 | 
						||
    decoder_spec = videotoolbox_decoder_config_create(videotoolbox->cm_codec_type, avctx);
 | 
						||
 | 
						||
    if (!decoder_spec) {
 | 
						||
        av_log(avctx, AV_LOG_ERROR, "decoder specification creation failed\n");
 | 
						||
        return -1;
 | 
						||
    }
 | 
						||
 | 
						||
    videotoolbox->cm_fmt_desc = videotoolbox_format_desc_create(videotoolbox->cm_codec_type,
 | 
						||
                                                                decoder_spec,
 | 
						||
                                                                avctx->width,
 | 
						||
                                                                avctx->height);
 | 
						||
    if (!videotoolbox->cm_fmt_desc) {
 | 
						||
        if (decoder_spec)
 | 
						||
            CFRelease(decoder_spec);
 | 
						||
 | 
						||
        av_log(avctx, AV_LOG_ERROR, "format description creation failed\n");
 | 
						||
        return -1;
 | 
						||
    }
 | 
						||
 | 
						||
    buf_attr = videotoolbox_buffer_attributes_create(avctx->width,
 | 
						||
                                                     avctx->height,
 | 
						||
                                                     videotoolbox->cv_pix_fmt_type);
 | 
						||
 | 
						||
    decoder_cb.decompressionOutputCallback = videotoolbox_decoder_callback;
 | 
						||
    decoder_cb.decompressionOutputRefCon   = avctx;
 | 
						||
 | 
						||
    status = VTDecompressionSessionCreate(NULL,                      // allocator
 | 
						||
                                          videotoolbox->cm_fmt_desc, // videoFormatDescription
 | 
						||
                                          decoder_spec,              // videoDecoderSpecification
 | 
						||
                                          buf_attr,                  // destinationImageBufferAttributes
 | 
						||
                                          &decoder_cb,               // outputCallback
 | 
						||
                                          &videotoolbox->session);   // decompressionSessionOut
 | 
						||
 | 
						||
    if (decoder_spec)
 | 
						||
        CFRelease(decoder_spec);
 | 
						||
    if (buf_attr)
 | 
						||
        CFRelease(buf_attr);
 | 
						||
 | 
						||
    switch (status) {
 | 
						||
    case kVTVideoDecoderNotAvailableNowErr:
 | 
						||
        av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox session not available.\n");
 | 
						||
        return AVERROR(ENOSYS);
 | 
						||
    case kVTVideoDecoderUnsupportedDataFormatErr:
 | 
						||
        av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox does not support this format.\n");
 | 
						||
        return AVERROR(ENOSYS);
 | 
						||
    case kVTCouldNotFindVideoDecoderErr:
 | 
						||
        av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox decoder for this format not found.\n");
 | 
						||
        return AVERROR(ENOSYS);
 | 
						||
    case kVTVideoDecoderMalfunctionErr:
 | 
						||
        av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox malfunction.\n");
 | 
						||
        return AVERROR(EINVAL);
 | 
						||
    case kVTVideoDecoderBadDataErr:
 | 
						||
        av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox reported invalid data.\n");
 | 
						||
        return AVERROR_INVALIDDATA;
 | 
						||
    case 0:
 | 
						||
        return 0;
 | 
						||
    default:
 | 
						||
        av_log(avctx, AV_LOG_VERBOSE, "Unknown VideoToolbox session creation error %d\n", (int)status);
 | 
						||
        return AVERROR_UNKNOWN;
 | 
						||
    }
 | 
						||
}
 | 
						||
 | 
						||
static void videotoolbox_stop(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx);
 | 
						||
    if (!videotoolbox)
 | 
						||
        return;
 | 
						||
 | 
						||
    if (videotoolbox->cm_fmt_desc) {
 | 
						||
        CFRelease(videotoolbox->cm_fmt_desc);
 | 
						||
        videotoolbox->cm_fmt_desc = NULL;
 | 
						||
    }
 | 
						||
 | 
						||
    if (videotoolbox->session) {
 | 
						||
        VTDecompressionSessionInvalidate(videotoolbox->session);
 | 
						||
        CFRelease(videotoolbox->session);
 | 
						||
        videotoolbox->session = NULL;
 | 
						||
    }
 | 
						||
}
 | 
						||
 | 
						||
static const char *videotoolbox_error_string(OSStatus status)
 | 
						||
{
 | 
						||
    switch (status) {
 | 
						||
        case kVTVideoDecoderBadDataErr:
 | 
						||
            return "bad data";
 | 
						||
        case kVTVideoDecoderMalfunctionErr:
 | 
						||
            return "decoder malfunction";
 | 
						||
        case kVTInvalidSessionErr:
 | 
						||
            return "invalid session";
 | 
						||
    }
 | 
						||
    return "unknown";
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_common_end_frame(AVCodecContext *avctx, AVFrame *frame)
 | 
						||
{
 | 
						||
    OSStatus status;
 | 
						||
    AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx);
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
 | 
						||
    if (vtctx->reconfig_needed == true) {
 | 
						||
        vtctx->reconfig_needed = false;
 | 
						||
        av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox decoder needs reconfig, restarting..\n");
 | 
						||
        videotoolbox_stop(avctx);
 | 
						||
        if (videotoolbox_start(avctx) != 0) {
 | 
						||
            return AVERROR_EXTERNAL;
 | 
						||
        }
 | 
						||
    }
 | 
						||
 | 
						||
    if (!videotoolbox->session || !vtctx->bitstream || !vtctx->bitstream_size)
 | 
						||
        return AVERROR_INVALIDDATA;
 | 
						||
 | 
						||
    status = videotoolbox_session_decode_frame(avctx);
 | 
						||
    if (status != noErr) {
 | 
						||
        if (status == kVTVideoDecoderMalfunctionErr || status == kVTInvalidSessionErr)
 | 
						||
            vtctx->reconfig_needed = true;
 | 
						||
        av_log(avctx, AV_LOG_ERROR, "Failed to decode frame (%s, %d)\n", videotoolbox_error_string(status), (int)status);
 | 
						||
        return AVERROR_UNKNOWN;
 | 
						||
    }
 | 
						||
 | 
						||
    if (!vtctx->frame) {
 | 
						||
        vtctx->reconfig_needed = true;
 | 
						||
        return AVERROR_UNKNOWN;
 | 
						||
    }
 | 
						||
 | 
						||
    return videotoolbox_buffer_create(avctx, frame);
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_h264_end_frame(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    H264Context *h = avctx->priv_data;
 | 
						||
    AVFrame *frame = h->cur_pic_ptr->f;
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
    int ret = videotoolbox_common_end_frame(avctx, frame);
 | 
						||
    vtctx->bitstream_size = 0;
 | 
						||
    return ret;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_hevc_start_frame(AVCodecContext *avctx,
 | 
						||
                                         const uint8_t *buffer,
 | 
						||
                                         uint32_t size)
 | 
						||
{
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_hevc_decode_slice(AVCodecContext *avctx,
 | 
						||
                                          const uint8_t *buffer,
 | 
						||
                                          uint32_t size)
 | 
						||
{
 | 
						||
    return videotoolbox_common_decode_slice(avctx, buffer, size);
 | 
						||
}
 | 
						||
 | 
						||
 | 
						||
static int videotoolbox_hevc_decode_params(AVCodecContext *avctx,
 | 
						||
                                           int type,
 | 
						||
                                           const uint8_t *buffer,
 | 
						||
                                           uint32_t size)
 | 
						||
{
 | 
						||
    return videotoolbox_common_decode_slice(avctx, buffer, size);
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_hevc_end_frame(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    HEVCContext *h = avctx->priv_data;
 | 
						||
    AVFrame *frame = h->ref->frame;
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
 | 
						||
    h->output_frame->crop_right = 0;
 | 
						||
    h->output_frame->crop_left = 0;
 | 
						||
    h->output_frame->crop_top = 0;
 | 
						||
    h->output_frame->crop_bottom = 0;
 | 
						||
 | 
						||
    int ret = videotoolbox_common_end_frame(avctx, frame);
 | 
						||
    vtctx->bitstream_size = 0;
 | 
						||
    return ret;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_mpeg_start_frame(AVCodecContext *avctx,
 | 
						||
                                         const uint8_t *buffer,
 | 
						||
                                         uint32_t size)
 | 
						||
{
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
 | 
						||
    return videotoolbox_buffer_copy(vtctx, buffer, size);
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_mpeg_decode_slice(AVCodecContext *avctx,
 | 
						||
                                          const uint8_t *buffer,
 | 
						||
                                          uint32_t size)
 | 
						||
{
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_mpeg_end_frame(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    MpegEncContext *s = avctx->priv_data;
 | 
						||
    AVFrame *frame = s->current_picture_ptr->f;
 | 
						||
 | 
						||
    return videotoolbox_common_end_frame(avctx, frame);
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_uninit(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
    if (!vtctx)
 | 
						||
        return 0;
 | 
						||
 | 
						||
    ff_videotoolbox_uninit(avctx);
 | 
						||
 | 
						||
    if (vtctx->vt_ctx)
 | 
						||
        videotoolbox_stop(avctx);
 | 
						||
 | 
						||
    av_buffer_unref(&vtctx->cached_hw_frames_ctx);
 | 
						||
    av_freep(&vtctx->vt_ctx);
 | 
						||
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
static enum AVPixelFormat videotoolbox_best_pixel_format(AVCodecContext *avctx) {
 | 
						||
    const AVPixFmtDescriptor *descriptor = av_pix_fmt_desc_get(avctx->pix_fmt);
 | 
						||
    if (!descriptor)
 | 
						||
        return AV_PIX_FMT_NV12; // same as av_videotoolbox_alloc_context()
 | 
						||
 | 
						||
    int depth = descriptor->comp[0].depth;
 | 
						||
    if (depth > 8) {
 | 
						||
        return AV_PIX_FMT_P010;
 | 
						||
    }
 | 
						||
 | 
						||
    return AV_PIX_FMT_NV12;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_common_init(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    VTContext *vtctx = avctx->internal->hwaccel_priv_data;
 | 
						||
    AVHWFramesContext *hw_frames;
 | 
						||
    int err;
 | 
						||
 | 
						||
    // Old API - do nothing.
 | 
						||
    if (avctx->hwaccel_context)
 | 
						||
        return 0;
 | 
						||
 | 
						||
    if (!avctx->hw_frames_ctx && !avctx->hw_device_ctx) {
 | 
						||
        av_log(avctx, AV_LOG_ERROR,
 | 
						||
               "Either hw_frames_ctx or hw_device_ctx must be set.\n");
 | 
						||
        return AVERROR(EINVAL);
 | 
						||
    }
 | 
						||
 | 
						||
    vtctx->vt_ctx = av_videotoolbox_alloc_context();
 | 
						||
    if (!vtctx->vt_ctx) {
 | 
						||
        err = AVERROR(ENOMEM);
 | 
						||
        goto fail;
 | 
						||
    }
 | 
						||
 | 
						||
    if (avctx->hw_frames_ctx) {
 | 
						||
        hw_frames = (AVHWFramesContext*)avctx->hw_frames_ctx->data;
 | 
						||
    } else {
 | 
						||
        avctx->hw_frames_ctx = av_hwframe_ctx_alloc(avctx->hw_device_ctx);
 | 
						||
        if (!avctx->hw_frames_ctx) {
 | 
						||
            err = AVERROR(ENOMEM);
 | 
						||
            goto fail;
 | 
						||
        }
 | 
						||
 | 
						||
        hw_frames = (AVHWFramesContext*)avctx->hw_frames_ctx->data;
 | 
						||
        hw_frames->format = AV_PIX_FMT_VIDEOTOOLBOX;
 | 
						||
        hw_frames->sw_format = videotoolbox_best_pixel_format(avctx);
 | 
						||
        hw_frames->width = avctx->width;
 | 
						||
        hw_frames->height = avctx->height;
 | 
						||
 | 
						||
        err = av_hwframe_ctx_init(avctx->hw_frames_ctx);
 | 
						||
        if (err < 0) {
 | 
						||
            av_buffer_unref(&avctx->hw_frames_ctx);
 | 
						||
            goto fail;
 | 
						||
        }
 | 
						||
    }
 | 
						||
 | 
						||
    vtctx->cached_hw_frames_ctx = av_buffer_ref(avctx->hw_frames_ctx);
 | 
						||
    if (!vtctx->cached_hw_frames_ctx) {
 | 
						||
        err = AVERROR(ENOMEM);
 | 
						||
        goto fail;
 | 
						||
    }
 | 
						||
 | 
						||
    bool full_range = avctx->color_range == AVCOL_RANGE_JPEG;
 | 
						||
    vtctx->vt_ctx->cv_pix_fmt_type =
 | 
						||
        av_map_videotoolbox_format_from_pixfmt2(hw_frames->sw_format, full_range);
 | 
						||
    if (!vtctx->vt_ctx->cv_pix_fmt_type) {
 | 
						||
        av_log(avctx, AV_LOG_ERROR, "Unknown sw_format.\n");
 | 
						||
        err = AVERROR(EINVAL);
 | 
						||
        goto fail;
 | 
						||
    }
 | 
						||
 | 
						||
    err = videotoolbox_start(avctx);
 | 
						||
    if (err < 0)
 | 
						||
        goto fail;
 | 
						||
 | 
						||
    return 0;
 | 
						||
 | 
						||
fail:
 | 
						||
    videotoolbox_uninit(avctx);
 | 
						||
    return err;
 | 
						||
}
 | 
						||
 | 
						||
static int videotoolbox_frame_params(AVCodecContext *avctx,
 | 
						||
                                     AVBufferRef *hw_frames_ctx)
 | 
						||
{
 | 
						||
    AVHWFramesContext *frames_ctx = (AVHWFramesContext*)hw_frames_ctx->data;
 | 
						||
 | 
						||
    frames_ctx->format            = AV_PIX_FMT_VIDEOTOOLBOX;
 | 
						||
    frames_ctx->width             = avctx->coded_width;
 | 
						||
    frames_ctx->height            = avctx->coded_height;
 | 
						||
    frames_ctx->sw_format         = videotoolbox_best_pixel_format(avctx);
 | 
						||
 | 
						||
    return 0;
 | 
						||
}
 | 
						||
 | 
						||
const AVHWAccel ff_h263_videotoolbox_hwaccel = {
 | 
						||
    .name           = "h263_videotoolbox",
 | 
						||
    .type           = AVMEDIA_TYPE_VIDEO,
 | 
						||
    .id             = AV_CODEC_ID_H263,
 | 
						||
    .pix_fmt        = AV_PIX_FMT_VIDEOTOOLBOX,
 | 
						||
    .alloc_frame    = ff_videotoolbox_alloc_frame,
 | 
						||
    .start_frame    = videotoolbox_mpeg_start_frame,
 | 
						||
    .decode_slice   = videotoolbox_mpeg_decode_slice,
 | 
						||
    .end_frame      = videotoolbox_mpeg_end_frame,
 | 
						||
    .frame_params   = videotoolbox_frame_params,
 | 
						||
    .init           = videotoolbox_common_init,
 | 
						||
    .uninit         = videotoolbox_uninit,
 | 
						||
    .priv_data_size = sizeof(VTContext),
 | 
						||
};
 | 
						||
 | 
						||
const AVHWAccel ff_hevc_videotoolbox_hwaccel = {
 | 
						||
    .name           = "hevc_videotoolbox",
 | 
						||
    .type           = AVMEDIA_TYPE_VIDEO,
 | 
						||
    .id             = AV_CODEC_ID_HEVC,
 | 
						||
    .pix_fmt        = AV_PIX_FMT_VIDEOTOOLBOX,
 | 
						||
    .alloc_frame    = ff_videotoolbox_alloc_frame,
 | 
						||
    .start_frame    = videotoolbox_hevc_start_frame,
 | 
						||
    .decode_slice   = videotoolbox_hevc_decode_slice,
 | 
						||
    .decode_params  = videotoolbox_hevc_decode_params,
 | 
						||
    .end_frame      = videotoolbox_hevc_end_frame,
 | 
						||
    .frame_params   = videotoolbox_frame_params,
 | 
						||
    .init           = videotoolbox_common_init,
 | 
						||
    .uninit         = videotoolbox_uninit,
 | 
						||
    .priv_data_size = sizeof(VTContext),
 | 
						||
};
 | 
						||
 | 
						||
const AVHWAccel ff_h264_videotoolbox_hwaccel = {
 | 
						||
    .name           = "h264_videotoolbox",
 | 
						||
    .type           = AVMEDIA_TYPE_VIDEO,
 | 
						||
    .id             = AV_CODEC_ID_H264,
 | 
						||
    .pix_fmt        = AV_PIX_FMT_VIDEOTOOLBOX,
 | 
						||
    .alloc_frame    = ff_videotoolbox_alloc_frame,
 | 
						||
    .start_frame    = ff_videotoolbox_h264_start_frame,
 | 
						||
    .decode_slice   = ff_videotoolbox_h264_decode_slice,
 | 
						||
    .decode_params  = videotoolbox_h264_decode_params,
 | 
						||
    .end_frame      = videotoolbox_h264_end_frame,
 | 
						||
    .frame_params   = videotoolbox_frame_params,
 | 
						||
    .init           = videotoolbox_common_init,
 | 
						||
    .uninit         = videotoolbox_uninit,
 | 
						||
    .priv_data_size = sizeof(VTContext),
 | 
						||
};
 | 
						||
 | 
						||
const AVHWAccel ff_mpeg1_videotoolbox_hwaccel = {
 | 
						||
    .name           = "mpeg1_videotoolbox",
 | 
						||
    .type           = AVMEDIA_TYPE_VIDEO,
 | 
						||
    .id             = AV_CODEC_ID_MPEG1VIDEO,
 | 
						||
    .pix_fmt        = AV_PIX_FMT_VIDEOTOOLBOX,
 | 
						||
    .alloc_frame    = ff_videotoolbox_alloc_frame,
 | 
						||
    .start_frame    = videotoolbox_mpeg_start_frame,
 | 
						||
    .decode_slice   = videotoolbox_mpeg_decode_slice,
 | 
						||
    .end_frame      = videotoolbox_mpeg_end_frame,
 | 
						||
    .frame_params   = videotoolbox_frame_params,
 | 
						||
    .init           = videotoolbox_common_init,
 | 
						||
    .uninit         = videotoolbox_uninit,
 | 
						||
    .priv_data_size = sizeof(VTContext),
 | 
						||
};
 | 
						||
 | 
						||
const AVHWAccel ff_mpeg2_videotoolbox_hwaccel = {
 | 
						||
    .name           = "mpeg2_videotoolbox",
 | 
						||
    .type           = AVMEDIA_TYPE_VIDEO,
 | 
						||
    .id             = AV_CODEC_ID_MPEG2VIDEO,
 | 
						||
    .pix_fmt        = AV_PIX_FMT_VIDEOTOOLBOX,
 | 
						||
    .alloc_frame    = ff_videotoolbox_alloc_frame,
 | 
						||
    .start_frame    = videotoolbox_mpeg_start_frame,
 | 
						||
    .decode_slice   = videotoolbox_mpeg_decode_slice,
 | 
						||
    .end_frame      = videotoolbox_mpeg_end_frame,
 | 
						||
    .frame_params   = videotoolbox_frame_params,
 | 
						||
    .init           = videotoolbox_common_init,
 | 
						||
    .uninit         = videotoolbox_uninit,
 | 
						||
    .priv_data_size = sizeof(VTContext),
 | 
						||
};
 | 
						||
 | 
						||
const AVHWAccel ff_mpeg4_videotoolbox_hwaccel = {
 | 
						||
    .name           = "mpeg4_videotoolbox",
 | 
						||
    .type           = AVMEDIA_TYPE_VIDEO,
 | 
						||
    .id             = AV_CODEC_ID_MPEG4,
 | 
						||
    .pix_fmt        = AV_PIX_FMT_VIDEOTOOLBOX,
 | 
						||
    .alloc_frame    = ff_videotoolbox_alloc_frame,
 | 
						||
    .start_frame    = videotoolbox_mpeg_start_frame,
 | 
						||
    .decode_slice   = videotoolbox_mpeg_decode_slice,
 | 
						||
    .end_frame      = videotoolbox_mpeg_end_frame,
 | 
						||
    .frame_params   = videotoolbox_frame_params,
 | 
						||
    .init           = videotoolbox_common_init,
 | 
						||
    .uninit         = videotoolbox_uninit,
 | 
						||
    .priv_data_size = sizeof(VTContext),
 | 
						||
};
 | 
						||
 | 
						||
static AVVideotoolboxContext *av_videotoolbox_alloc_context_with_pix_fmt(enum AVPixelFormat pix_fmt,
 | 
						||
                                                                         bool full_range)
 | 
						||
{
 | 
						||
    AVVideotoolboxContext *ret = av_mallocz(sizeof(*ret));
 | 
						||
 | 
						||
    if (ret) {
 | 
						||
        ret->output_callback = videotoolbox_decoder_callback;
 | 
						||
 | 
						||
        OSType cv_pix_fmt_type = av_map_videotoolbox_format_from_pixfmt2(pix_fmt, full_range);
 | 
						||
        if (cv_pix_fmt_type == 0) {
 | 
						||
            cv_pix_fmt_type = kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange;
 | 
						||
        }
 | 
						||
        ret->cv_pix_fmt_type = cv_pix_fmt_type;
 | 
						||
    }
 | 
						||
 | 
						||
    return ret;
 | 
						||
}
 | 
						||
 | 
						||
AVVideotoolboxContext *av_videotoolbox_alloc_context(void)
 | 
						||
{
 | 
						||
    return av_videotoolbox_alloc_context_with_pix_fmt(AV_PIX_FMT_NONE, false);
 | 
						||
}
 | 
						||
 | 
						||
int av_videotoolbox_default_init(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
    return av_videotoolbox_default_init2(avctx, NULL);
 | 
						||
}
 | 
						||
 | 
						||
int av_videotoolbox_default_init2(AVCodecContext *avctx, AVVideotoolboxContext *vtctx)
 | 
						||
{
 | 
						||
    enum AVPixelFormat pix_fmt = videotoolbox_best_pixel_format(avctx);
 | 
						||
    bool full_range = avctx->color_range == AVCOL_RANGE_JPEG;
 | 
						||
    avctx->hwaccel_context = vtctx ?: av_videotoolbox_alloc_context_with_pix_fmt(pix_fmt, full_range);
 | 
						||
    if (!avctx->hwaccel_context)
 | 
						||
        return AVERROR(ENOMEM);
 | 
						||
    return videotoolbox_start(avctx);
 | 
						||
}
 | 
						||
 | 
						||
void av_videotoolbox_default_free(AVCodecContext *avctx)
 | 
						||
{
 | 
						||
 | 
						||
    videotoolbox_stop(avctx);
 | 
						||
    av_freep(&avctx->hwaccel_context);
 | 
						||
}
 | 
						||
#endif /* CONFIG_VIDEOTOOLBOX */
 |