Skip to content
Snippets Groups Projects
frame.c 27.9 KiB
Newer Older
  • Learn to ignore specific revisions
  •  * This file is part of FFmpeg.
    
     * FFmpeg is free software; you can redistribute it and/or
    
     * modify it under the terms of the GNU Lesser General Public
     * License as published by the Free Software Foundation; either
     * version 2.1 of the License, or (at your option) any later version.
     *
    
     * FFmpeg is distributed in the hope that it will be useful,
    
     * but WITHOUT ANY WARRANTY; without even the implied warranty of
     * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
     * Lesser General Public License for more details.
     *
     * You should have received a copy of the GNU Lesser General Public
    
     * License along with FFmpeg; if not, write to the Free Software
    
     * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
     */
    
    
    #include "avassert.h"
    
    #include "buffer.h"
    #include "common.h"
    
    Anton Khirnov's avatar
    Anton Khirnov committed
    #include "dict.h"
    
    #include "frame.h"
    #include "imgutils.h"
    #include "mem.h"
    #include "samplefmt.h"
    
    
    MAKE_ACCESSORS(AVFrame, frame, int64_t, best_effort_timestamp)
    MAKE_ACCESSORS(AVFrame, frame, int64_t, pkt_duration)
    MAKE_ACCESSORS(AVFrame, frame, int64_t, pkt_pos)
    MAKE_ACCESSORS(AVFrame, frame, int64_t, channel_layout)
    MAKE_ACCESSORS(AVFrame, frame, int,     channels)
    MAKE_ACCESSORS(AVFrame, frame, int,     sample_rate)
    MAKE_ACCESSORS(AVFrame, frame, AVDictionary *, metadata)
    MAKE_ACCESSORS(AVFrame, frame, int,     decode_error_flags)
    MAKE_ACCESSORS(AVFrame, frame, int,     pkt_size)
    
    MAKE_ACCESSORS(AVFrame, frame, enum AVColorSpace, colorspace)
    MAKE_ACCESSORS(AVFrame, frame, enum AVColorRange, color_range)
    
    #define CHECK_CHANNELS_CONSISTENCY(frame) \
        av_assert2(!(frame)->channel_layout || \
                   (frame)->channels == \
                   av_get_channel_layout_nb_channels((frame)->channel_layout))
    
    
    wm4's avatar
    wm4 committed
    struct qp_properties {
        int stride;
        int type;
    };
    
    
    int av_frame_set_qp_table(AVFrame *f, AVBufferRef *buf, int stride, int qp_type)
    {
    
    wm4's avatar
    wm4 committed
        struct qp_properties *p;
        AVFrameSideData *sd;
        AVBufferRef *ref;
    
    
    FF_DISABLE_DEPRECATION_WARNINGS
    
        av_buffer_unref(&f->qp_table_buf);
    
        f->qp_table_buf = buf;
        f->qscale_table = buf->data;
        f->qstride      = stride;
        f->qscale_type  = qp_type;
    
    FF_ENABLE_DEPRECATION_WARNINGS
    
    wm4's avatar
    wm4 committed
        av_frame_remove_side_data(f, AV_FRAME_DATA_QP_TABLE_PROPERTIES);
        av_frame_remove_side_data(f, AV_FRAME_DATA_QP_TABLE_DATA);
    
        ref = av_buffer_ref(buf);
        if (!av_frame_new_side_data_from_buf(f, AV_FRAME_DATA_QP_TABLE_DATA, ref)) {
            av_buffer_unref(&ref);
            return AVERROR(ENOMEM);
        }
    
        sd = av_frame_new_side_data(f, AV_FRAME_DATA_QP_TABLE_PROPERTIES,
                                    sizeof(struct qp_properties));
        if (!sd)
            return AVERROR(ENOMEM);
    
        p = (struct qp_properties *)sd->data;
        p->stride = stride;
        p->type = qp_type;
    
    
        return 0;
    }
    
    int8_t *av_frame_get_qp_table(AVFrame *f, int *stride, int *type)
    {
    
    wm4's avatar
    wm4 committed
        AVBufferRef *buf = NULL;
    
    wm4's avatar
    wm4 committed
        *stride = 0;
        *type   = 0;
    
    wm4's avatar
    wm4 committed
    FF_DISABLE_DEPRECATION_WARNINGS
        if (f->qp_table_buf) {
            *stride = f->qstride;
            *type   = f->qscale_type;
            buf     = f->qp_table_buf;
    
    FF_ENABLE_DEPRECATION_WARNINGS
    
    wm4's avatar
    wm4 committed
        } else {
            AVFrameSideData *sd;
            struct qp_properties *p;
            sd = av_frame_get_side_data(f, AV_FRAME_DATA_QP_TABLE_PROPERTIES);
            if (!sd)
                return NULL;
            p = (struct qp_properties *)sd->data;
            sd = av_frame_get_side_data(f, AV_FRAME_DATA_QP_TABLE_DATA);
            if (!sd)
                return NULL;
            *stride = p->stride;
            *type   = p->type;
            buf     = sd->buf;
        }
    
        return buf ? buf->data : NULL;
    
    const char *av_get_colorspace_name(enum AVColorSpace val)
    {
    
        static const char * const name[] = {
    
            [AVCOL_SPC_RGB]       = "GBR",
            [AVCOL_SPC_BT709]     = "bt709",
            [AVCOL_SPC_FCC]       = "fcc",
            [AVCOL_SPC_BT470BG]   = "bt470bg",
            [AVCOL_SPC_SMPTE170M] = "smpte170m",
            [AVCOL_SPC_SMPTE240M] = "smpte240m",
            [AVCOL_SPC_YCOCG]     = "YCgCo",
        };
    
            return NULL;
        return name[val];
    }
    
    
    static void get_frame_defaults(AVFrame *frame)
    {
        if (frame->extended_data != frame->data)
            av_freep(&frame->extended_data);
    
        memset(frame, 0, sizeof(*frame));
    
    
        frame->pkt_dts               = AV_NOPTS_VALUE;
    #if FF_API_PKT_PTS
    FF_DISABLE_DEPRECATION_WARNINGS
    
        frame->pkt_pts               = AV_NOPTS_VALUE;
    
    FF_ENABLE_DEPRECATION_WARNINGS
    #endif
    
        frame->best_effort_timestamp = AV_NOPTS_VALUE;
        frame->pkt_duration        = 0;
        frame->pkt_pos             = -1;
        frame->pkt_size            = -1;
    
        frame->key_frame           = 1;
        frame->sample_aspect_ratio = (AVRational){ 0, 1 };
        frame->format              = -1; /* unknown */
        frame->extended_data       = frame->data;
    
        frame->color_primaries     = AVCOL_PRI_UNSPECIFIED;
        frame->color_trc           = AVCOL_TRC_UNSPECIFIED;
        frame->colorspace          = AVCOL_SPC_UNSPECIFIED;
        frame->color_range         = AVCOL_RANGE_UNSPECIFIED;
        frame->chroma_location     = AVCHROMA_LOC_UNSPECIFIED;
    
    static void free_side_data(AVFrameSideData **ptr_sd)
    {
        AVFrameSideData *sd = *ptr_sd;
    
    
        av_buffer_unref(&sd->buf);
    
        av_dict_free(&sd->metadata);
        av_freep(ptr_sd);
    }
    
    
    static void wipe_side_data(AVFrame *frame)
    {
        int i;
    
        for (i = 0; i < frame->nb_side_data; i++) {
            free_side_data(&frame->side_data[i]);
        }
        frame->nb_side_data = 0;
    
        av_freep(&frame->side_data);
    }
    
    
    AVFrame *av_frame_alloc(void)
    {
        AVFrame *frame = av_mallocz(sizeof(*frame));
    
        if (!frame)
            return NULL;
    
    
        frame->extended_data = NULL;
    
        get_frame_defaults(frame);
    
        return frame;
    }
    
    void av_frame_free(AVFrame **frame)
    {
        if (!frame || !*frame)
            return;
    
        av_frame_unref(*frame);
        av_freep(frame);
    }
    
    static int get_video_buffer(AVFrame *frame, int align)
    {
        const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(frame->format);
    
        int ret, i, padded_height;
        int plane_padding = FFMAX(16 + 16/*STRIDE_ALIGN*/, align);
    
    
        if (!desc)
            return AVERROR(EINVAL);
    
        if ((ret = av_image_check_size(frame->width, frame->height, 0, NULL)) < 0)
            return ret;
    
        if (!frame->linesize[0]) {
    
                align = 32; /* STRIDE_ALIGN. Should be av_cpu_max_align() */
    
    
            for(i=1; i<=align; i+=i) {
                ret = av_image_fill_linesizes(frame->linesize, frame->format,
                                              FFALIGN(frame->width, i));
                if (ret < 0)
                    return ret;
                if (!(frame->linesize[0] & (align-1)))
                    break;
            }
    
    
            for (i = 0; i < 4 && frame->linesize[i]; i++)
                frame->linesize[i] = FFALIGN(frame->linesize[i], align);
        }
    
    
        padded_height = FFALIGN(frame->height, 32);
        if ((ret = av_image_fill_pointers(frame->data, frame->format, padded_height,
    
                                          NULL, frame->linesize)) < 0)
            return ret;
    
        frame->buf[0] = av_buffer_alloc(ret + 4*plane_padding);
    
        if (!frame->buf[0])
            goto fail;
    
        if (av_image_fill_pointers(frame->data, frame->format, padded_height,
    
                                   frame->buf[0]->data, frame->linesize) < 0)
            goto fail;
    
        for (i = 1; i < 4; i++) {
            if (frame->data[i])
                frame->data[i] += i * plane_padding;
    
        }
    
        frame->extended_data = frame->data;
    
        return 0;
    fail:
        av_frame_unref(frame);
        return AVERROR(ENOMEM);
    }
    
    static int get_audio_buffer(AVFrame *frame, int align)
    {
    
        int planar   = av_sample_fmt_is_planar(frame->format);
    
        int ret, i;
    
    
        if (!frame->channels)
            frame->channels = av_get_channel_layout_nb_channels(frame->channel_layout);
    
        channels = frame->channels;
        planes = planar ? channels : 1;
    
    
        CHECK_CHANNELS_CONSISTENCY(frame);
    
        if (!frame->linesize[0]) {
            ret = av_samples_get_buffer_size(&frame->linesize[0], channels,
                                             frame->nb_samples, frame->format,
                                             align);
            if (ret < 0)
                return ret;
        }
    
        if (planes > AV_NUM_DATA_POINTERS) {
    
            frame->extended_data = av_mallocz_array(planes,
    
                                              sizeof(*frame->extended_data));
    
            frame->extended_buf  = av_mallocz_array((planes - AV_NUM_DATA_POINTERS),
    
                                              sizeof(*frame->extended_buf));
            if (!frame->extended_data || !frame->extended_buf) {
                av_freep(&frame->extended_data);
                av_freep(&frame->extended_buf);
                return AVERROR(ENOMEM);
            }
            frame->nb_extended_buf = planes - AV_NUM_DATA_POINTERS;
        } else
            frame->extended_data = frame->data;
    
        for (i = 0; i < FFMIN(planes, AV_NUM_DATA_POINTERS); i++) {
            frame->buf[i] = av_buffer_alloc(frame->linesize[0]);
            if (!frame->buf[i]) {
                av_frame_unref(frame);
                return AVERROR(ENOMEM);
            }
            frame->extended_data[i] = frame->data[i] = frame->buf[i]->data;
        }
        for (i = 0; i < planes - AV_NUM_DATA_POINTERS; i++) {
            frame->extended_buf[i] = av_buffer_alloc(frame->linesize[0]);
            if (!frame->extended_buf[i]) {
                av_frame_unref(frame);
                return AVERROR(ENOMEM);
            }
            frame->extended_data[i + AV_NUM_DATA_POINTERS] = frame->extended_buf[i]->data;
        }
        return 0;
    
    }
    
    int av_frame_get_buffer(AVFrame *frame, int align)
    {
        if (frame->format < 0)
            return AVERROR(EINVAL);
    
        if (frame->width > 0 && frame->height > 0)
            return get_video_buffer(frame, align);
    
        else if (frame->nb_samples > 0 && (frame->channel_layout || frame->channels > 0))
    
            return get_audio_buffer(frame, align);
    
        return AVERROR(EINVAL);
    }
    
    
    static int frame_copy_props(AVFrame *dst, const AVFrame *src, int force_copy)
    
    {
        int i;
    
        dst->key_frame              = src->key_frame;
        dst->pict_type              = src->pict_type;
        dst->sample_aspect_ratio    = src->sample_aspect_ratio;
    
        dst->crop_top               = src->crop_top;
        dst->crop_bottom            = src->crop_bottom;
        dst->crop_left              = src->crop_left;
        dst->crop_right             = src->crop_right;
    
        dst->pts                    = src->pts;
        dst->repeat_pict            = src->repeat_pict;
        dst->interlaced_frame       = src->interlaced_frame;
        dst->top_field_first        = src->top_field_first;
        dst->palette_has_changed    = src->palette_has_changed;
        dst->sample_rate            = src->sample_rate;
        dst->opaque                 = src->opaque;
    
    #if FF_API_PKT_PTS
    FF_DISABLE_DEPRECATION_WARNINGS
    
        dst->pkt_pts                = src->pkt_pts;
    
    FF_ENABLE_DEPRECATION_WARNINGS
    #endif
    
        dst->pkt_dts                = src->pkt_dts;
        dst->pkt_pos                = src->pkt_pos;
        dst->pkt_size               = src->pkt_size;
        dst->pkt_duration           = src->pkt_duration;
        dst->reordered_opaque       = src->reordered_opaque;
        dst->quality                = src->quality;
        dst->best_effort_timestamp  = src->best_effort_timestamp;
        dst->coded_picture_number   = src->coded_picture_number;
        dst->display_picture_number = src->display_picture_number;
        dst->flags                  = src->flags;
        dst->decode_error_flags     = src->decode_error_flags;
        dst->color_primaries        = src->color_primaries;
        dst->color_trc              = src->color_trc;
        dst->colorspace             = src->colorspace;
        dst->color_range            = src->color_range;
        dst->chroma_location        = src->chroma_location;
    
        av_dict_copy(&dst->metadata, src->metadata, 0);
    
    
    #if FF_API_ERROR_FRAME
    FF_DISABLE_DEPRECATION_WARNINGS
    
        memcpy(dst->error, src->error, sizeof(dst->error));
    
    FF_ENABLE_DEPRECATION_WARNINGS
    #endif
    
    
        for (i = 0; i < src->nb_side_data; i++) {
            const AVFrameSideData *sd_src = src->side_data[i];
            AVFrameSideData *sd_dst;
            if (   sd_src->type == AV_FRAME_DATA_PANSCAN
                && (src->width != dst->width || src->height != dst->height))
                continue;
    
            if (force_copy) {
                sd_dst = av_frame_new_side_data(dst, sd_src->type,
                                                sd_src->size);
                if (!sd_dst) {
                    wipe_side_data(dst);
                    return AVERROR(ENOMEM);
                }
                memcpy(sd_dst->data, sd_src->data, sd_src->size);
            } else {
    
                AVBufferRef *ref = av_buffer_ref(sd_src->buf);
                sd_dst = av_frame_new_side_data_from_buf(dst, sd_src->type, ref);
    
                    wipe_side_data(dst);
                    return AVERROR(ENOMEM);
                }
    
            }
            av_dict_copy(&sd_dst->metadata, sd_src->metadata, 0);
        }
    
    
    #if FF_API_FRAME_QP
    FF_DISABLE_DEPRECATION_WARNINGS
    
        dst->qscale_table = NULL;
        dst->qstride      = 0;
        dst->qscale_type  = 0;
    
        av_buffer_unref(&dst->qp_table_buf);
    
        if (src->qp_table_buf) {
            dst->qp_table_buf = av_buffer_ref(src->qp_table_buf);
            if (dst->qp_table_buf) {
                dst->qscale_table = dst->qp_table_buf->data;
                dst->qstride      = src->qstride;
                dst->qscale_type  = src->qscale_type;
            }
        }
    
    FF_ENABLE_DEPRECATION_WARNINGS
    #endif
    
    wm4's avatar
    wm4 committed
        av_buffer_unref(&dst->opaque_ref);
    
        av_buffer_unref(&dst->private_ref);
    
    wm4's avatar
    wm4 committed
        if (src->opaque_ref) {
            dst->opaque_ref = av_buffer_ref(src->opaque_ref);
            if (!dst->opaque_ref)
                return AVERROR(ENOMEM);
        }
    
        if (src->private_ref) {
            dst->private_ref = av_buffer_ref(src->private_ref);
            if (!dst->private_ref)
                return AVERROR(ENOMEM);
        }
    
    int av_frame_ref(AVFrame *dst, const AVFrame *src)
    
    {
        int i, ret = 0;
    
    
        av_assert1(dst->width == 0 && dst->height == 0);
        av_assert1(dst->channels == 0);
    
    
        dst->format         = src->format;
        dst->width          = src->width;
        dst->height         = src->height;
    
        dst->channels       = src->channels;
    
        dst->channel_layout = src->channel_layout;
        dst->nb_samples     = src->nb_samples;
    
    
        ret = frame_copy_props(dst, src, 0);
    
        if (ret < 0)
            return ret;
    
        /* duplicate the frame data if it's not refcounted */
        if (!src->buf[0]) {
            ret = av_frame_get_buffer(dst, 32);
            if (ret < 0)
                return ret;
    
    
            ret = av_frame_copy(dst, src);
            if (ret < 0)
                av_frame_unref(dst);
    
            return ret;
    
        }
    
        /* ref the buffers */
    
        for (i = 0; i < FF_ARRAY_ELEMS(src->buf); i++) {
            if (!src->buf[i])
                continue;
    
            dst->buf[i] = av_buffer_ref(src->buf[i]);
            if (!dst->buf[i]) {
                ret = AVERROR(ENOMEM);
                goto fail;
            }
        }
    
        if (src->extended_buf) {
    
            dst->extended_buf = av_mallocz_array(sizeof(*dst->extended_buf),
    
                                           src->nb_extended_buf);
            if (!dst->extended_buf) {
                ret = AVERROR(ENOMEM);
                goto fail;
            }
            dst->nb_extended_buf = src->nb_extended_buf;
    
            for (i = 0; i < src->nb_extended_buf; i++) {
                dst->extended_buf[i] = av_buffer_ref(src->extended_buf[i]);
                if (!dst->extended_buf[i]) {
                    ret = AVERROR(ENOMEM);
                    goto fail;
                }
            }
        }
    
    
        if (src->hw_frames_ctx) {
            dst->hw_frames_ctx = av_buffer_ref(src->hw_frames_ctx);
            if (!dst->hw_frames_ctx) {
                ret = AVERROR(ENOMEM);
                goto fail;
            }
        }
    
    
        /* duplicate extended data */
        if (src->extended_data != src->data) {
    
            int ch = src->channels;
    
    
            if (!ch) {
                ret = AVERROR(EINVAL);
                goto fail;
            }
    
            CHECK_CHANNELS_CONSISTENCY(src);
    
            dst->extended_data = av_malloc_array(sizeof(*dst->extended_data), ch);
    
            if (!dst->extended_data) {
                ret = AVERROR(ENOMEM);
                goto fail;
            }
            memcpy(dst->extended_data, src->extended_data, sizeof(*src->extended_data) * ch);
        } else
            dst->extended_data = dst->data;
    
        memcpy(dst->data,     src->data,     sizeof(src->data));
        memcpy(dst->linesize, src->linesize, sizeof(src->linesize));
    
        return 0;
    
    fail:
        av_frame_unref(dst);
        return ret;
    }
    
    
    AVFrame *av_frame_clone(const AVFrame *src)
    
    {
        AVFrame *ret = av_frame_alloc();
    
        if (!ret)
            return NULL;
    
        if (av_frame_ref(ret, src) < 0)
            av_frame_free(&ret);
    
        return ret;
    }
    
    void av_frame_unref(AVFrame *frame)
    {
        int i;
    
    
        wipe_side_data(frame);
    
        for (i = 0; i < FF_ARRAY_ELEMS(frame->buf); i++)
            av_buffer_unref(&frame->buf[i]);
        for (i = 0; i < frame->nb_extended_buf; i++)
            av_buffer_unref(&frame->extended_buf[i]);
        av_freep(&frame->extended_buf);
    
        av_dict_free(&frame->metadata);
    
    FF_DISABLE_DEPRECATION_WARNINGS
    
        av_buffer_unref(&frame->qp_table_buf);
    
    FF_ENABLE_DEPRECATION_WARNINGS
    
        av_buffer_unref(&frame->hw_frames_ctx);
    
    
    wm4's avatar
    wm4 committed
        av_buffer_unref(&frame->opaque_ref);
    
        av_buffer_unref(&frame->private_ref);
    
        get_frame_defaults(frame);
    }
    
    void av_frame_move_ref(AVFrame *dst, AVFrame *src)
    {
    
        av_assert1(dst->width == 0 && dst->height == 0);
        av_assert1(dst->channels == 0);
    
    
        *dst = *src;
        if (src->extended_data == src->data)
            dst->extended_data = dst->data;
        memset(src, 0, sizeof(*src));
        get_frame_defaults(src);
    }
    
    int av_frame_is_writable(AVFrame *frame)
    {
        int i, ret = 1;
    
        /* assume non-refcounted frames are not writable */
        if (!frame->buf[0])
            return 0;
    
    
        for (i = 0; i < FF_ARRAY_ELEMS(frame->buf); i++)
            if (frame->buf[i])
                ret &= !!av_buffer_is_writable(frame->buf[i]);
    
        for (i = 0; i < frame->nb_extended_buf; i++)
            ret &= !!av_buffer_is_writable(frame->extended_buf[i]);
    
        return ret;
    }
    
    int av_frame_make_writable(AVFrame *frame)
    {
        AVFrame tmp;
        int ret;
    
        if (!frame->buf[0])
            return AVERROR(EINVAL);
    
        if (av_frame_is_writable(frame))
            return 0;
    
        memset(&tmp, 0, sizeof(tmp));
        tmp.format         = frame->format;
        tmp.width          = frame->width;
        tmp.height         = frame->height;
    
        tmp.channel_layout = frame->channel_layout;
        tmp.nb_samples     = frame->nb_samples;
        ret = av_frame_get_buffer(&tmp, 32);
        if (ret < 0)
            return ret;
    
    
        ret = av_frame_copy(&tmp, frame);
        if (ret < 0) {
            av_frame_unref(&tmp);
            return ret;
    
        }
    
        ret = av_frame_copy_props(&tmp, frame);
        if (ret < 0) {
            av_frame_unref(&tmp);
            return ret;
        }
    
        av_frame_unref(frame);
    
        *frame = tmp;
        if (tmp.data == tmp.extended_data)
            frame->extended_data = frame->data;
    
        return 0;
    }
    
    
    int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
    {
        return frame_copy_props(dst, src, 1);
    }
    
    
    AVBufferRef *av_frame_get_plane_buffer(AVFrame *frame, int plane)
    {
        uint8_t *data;
        int planes, i;
    
        if (frame->nb_samples) {
    
            int channels = frame->channels;
    
            if (!channels)
                return NULL;
    
            CHECK_CHANNELS_CONSISTENCY(frame);
    
            planes = av_sample_fmt_is_planar(frame->format) ? channels : 1;
        } else
            planes = 4;
    
        if (plane < 0 || plane >= planes || !frame->extended_data[plane])
            return NULL;
        data = frame->extended_data[plane];
    
        for (i = 0; i < FF_ARRAY_ELEMS(frame->buf) && frame->buf[i]; i++) {
            AVBufferRef *buf = frame->buf[i];
            if (data >= buf->data && data < buf->data + buf->size)
                return buf;
        }
        for (i = 0; i < frame->nb_extended_buf; i++) {
            AVBufferRef *buf = frame->extended_buf[i];
            if (data >= buf->data && data < buf->data + buf->size)
                return buf;
        }
        return NULL;
    }
    
    AVFrameSideData *av_frame_new_side_data_from_buf(AVFrame *frame,
                                                     enum AVFrameSideDataType type,
                                                     AVBufferRef *buf)
    
    Anton Khirnov's avatar
    Anton Khirnov committed
    {
        AVFrameSideData *ret, **tmp;
    
    
    Anton Khirnov's avatar
    Anton Khirnov committed
            return NULL;
    
    
        if (frame->nb_side_data > INT_MAX / sizeof(*frame->side_data) - 1)
    
    Anton Khirnov's avatar
    Anton Khirnov committed
        tmp = av_realloc(frame->side_data,
                         (frame->nb_side_data + 1) * sizeof(*frame->side_data));
        if (!tmp)
    
    Anton Khirnov's avatar
    Anton Khirnov committed
        frame->side_data = tmp;
    
        ret = av_mallocz(sizeof(*ret));
        if (!ret)
    
        ret->buf = buf;
        ret->data = ret->buf->data;
        ret->size = buf->size;
    
    Anton Khirnov's avatar
    Anton Khirnov committed
        ret->type = type;
    
        frame->side_data[frame->nb_side_data++] = ret;
    
        return ret;
    
    }
    
    AVFrameSideData *av_frame_new_side_data(AVFrame *frame,
                                            enum AVFrameSideDataType type,
                                            int size)
    {
    
        AVFrameSideData *ret;
        AVBufferRef *buf = av_buffer_alloc(size);
        ret = av_frame_new_side_data_from_buf(frame, type, buf);
        if (!ret)
            av_buffer_unref(&buf);
        return ret;
    
    AVFrameSideData *av_frame_get_side_data(const AVFrame *frame,
    
    Anton Khirnov's avatar
    Anton Khirnov committed
                                            enum AVFrameSideDataType type)
    {
        int i;
    
        for (i = 0; i < frame->nb_side_data; i++) {
            if (frame->side_data[i]->type == type)
                return frame->side_data[i];
        }
        return NULL;
    }
    
    
    static int frame_copy_video(AVFrame *dst, const AVFrame *src)
    {
        const uint8_t *src_data[4];
        int i, planes;
    
    
        if (dst->width  < src->width ||
            dst->height < src->height)
    
            return AVERROR(EINVAL);
    
        planes = av_pix_fmt_count_planes(dst->format);
        for (i = 0; i < planes; i++)
            if (!dst->data[i] || !src->data[i])
                return AVERROR(EINVAL);
    
        memcpy(src_data, src->data, sizeof(src_data));
        av_image_copy(dst->data, dst->linesize,
                      src_data, src->linesize,
    
    
        return 0;
    }
    
    static int frame_copy_audio(AVFrame *dst, const AVFrame *src)
    {
        int planar   = av_sample_fmt_is_planar(dst->format);
    
        int channels = dst->channels;
    
        int planes   = planar ? channels : 1;
        int i;
    
        if (dst->nb_samples     != src->nb_samples ||
    
            dst->channels       != src->channels ||
    
            dst->channel_layout != src->channel_layout)
            return AVERROR(EINVAL);
    
    
        CHECK_CHANNELS_CONSISTENCY(src);
    
    
        for (i = 0; i < planes; i++)
            if (!dst->extended_data[i] || !src->extended_data[i])
                return AVERROR(EINVAL);
    
        av_samples_copy(dst->extended_data, src->extended_data, 0, 0,
                        dst->nb_samples, channels, dst->format);
    
        return 0;
    }
    
    int av_frame_copy(AVFrame *dst, const AVFrame *src)
    {
        if (dst->format != src->format || dst->format < 0)
            return AVERROR(EINVAL);
    
        if (dst->width > 0 && dst->height > 0)
            return frame_copy_video(dst, src);
    
        else if (dst->nb_samples > 0 && dst->channels > 0)
    
            return frame_copy_audio(dst, src);
    
        return AVERROR(EINVAL);
    }
    
    
    void av_frame_remove_side_data(AVFrame *frame, enum AVFrameSideDataType type)
    {
        int i;
    
        for (i = 0; i < frame->nb_side_data; i++) {
            AVFrameSideData *sd = frame->side_data[i];
            if (sd->type == type) {
    
                free_side_data(&frame->side_data[i]);
    
                frame->side_data[i] = frame->side_data[frame->nb_side_data - 1];
                frame->nb_side_data--;
            }
        }
    }
    
    
    const char *av_frame_side_data_name(enum AVFrameSideDataType type)
    {
        switch(type) {
        case AV_FRAME_DATA_PANSCAN:         return "AVPanScan";
        case AV_FRAME_DATA_A53_CC:          return "ATSC A53 Part 4 Closed Captions";
    
        case AV_FRAME_DATA_STEREO3D:        return "Stereo 3D";
    
        case AV_FRAME_DATA_MATRIXENCODING:  return "AVMatrixEncoding";
        case AV_FRAME_DATA_DOWNMIX_INFO:    return "Metadata relevant to a downmix procedure";
        case AV_FRAME_DATA_REPLAYGAIN:      return "AVReplayGain";
        case AV_FRAME_DATA_DISPLAYMATRIX:   return "3x3 displaymatrix";
    
        case AV_FRAME_DATA_AFD:             return "Active format description";
    
        case AV_FRAME_DATA_MOTION_VECTORS:  return "Motion vectors";
    
        case AV_FRAME_DATA_SKIP_SAMPLES:    return "Skip samples";
        case AV_FRAME_DATA_AUDIO_SERVICE_TYPE:          return "Audio service type";
        case AV_FRAME_DATA_MASTERING_DISPLAY_METADATA:  return "Mastering display metadata";
    
        case AV_FRAME_DATA_CONTENT_LIGHT_LEVEL:         return "Content light level metadata";
    
        case AV_FRAME_DATA_GOP_TIMECODE:                return "GOP timecode";
    
        case AV_FRAME_DATA_SPHERICAL:                   return "Spherical Mapping";
    
        case AV_FRAME_DATA_ICC_PROFILE:                 return "ICC profile";
    
    wm4's avatar
    wm4 committed
        case AV_FRAME_DATA_QP_TABLE_PROPERTIES:         return "QP table properties";
        case AV_FRAME_DATA_QP_TABLE_DATA:               return "QP table data";
    
    
    static int calc_cropping_offsets(size_t offsets[4], const AVFrame *frame,
                                     const AVPixFmtDescriptor *desc)
    {
        int i, j;
    
        for (i = 0; frame->data[i]; i++) {
            const AVComponentDescriptor *comp = NULL;
            int shift_x = (i == 1 || i == 2) ? desc->log2_chroma_w : 0;
            int shift_y = (i == 1 || i == 2) ? desc->log2_chroma_h : 0;
    
    
            if (desc->flags & (AV_PIX_FMT_FLAG_PAL | FF_PSEUDOPAL) && i == 1) {
    
                offsets[i] = 0;
                break;
            }
    
            /* find any component descriptor for this plane */
            for (j = 0; j < desc->nb_components; j++) {
                if (desc->comp[j].plane == i) {
                    comp = &desc->comp[j];
                    break;
                }
            }
            if (!comp)
                return AVERROR_BUG;
    
            offsets[i] = (frame->crop_top  >> shift_y) * frame->linesize[i] +
                         (frame->crop_left >> shift_x) * comp->step;
        }
    
        return 0;
    }
    
    int av_frame_apply_cropping(AVFrame *frame, int flags)
    {
        const AVPixFmtDescriptor *desc;
        size_t offsets[4];
        int i;
    
        if (!(frame->width > 0 && frame->height > 0))
            return AVERROR(EINVAL);
    
        if (frame->crop_left >= INT_MAX - frame->crop_right        ||
            frame->crop_top  >= INT_MAX - frame->crop_bottom       ||
            (frame->crop_left + frame->crop_right) >= frame->width ||
            (frame->crop_top + frame->crop_bottom) >= frame->height)
            return AVERROR(ERANGE);
    
        desc = av_pix_fmt_desc_get(frame->format);
        if (!desc)
            return AVERROR_BUG;
    
        /* Apply just the right/bottom cropping for hwaccel formats. Bitstream
         * formats cannot be easily handled here either (and corresponding decoders
         * should not export any cropping anyway), so do the same for those as well.
         * */
        if (desc->flags & (AV_PIX_FMT_FLAG_BITSTREAM | AV_PIX_FMT_FLAG_HWACCEL)) {
            frame->width      -= frame->crop_right;
            frame->height     -= frame->crop_bottom;
            frame->crop_right  = 0;
            frame->crop_bottom = 0;
            return 0;
        }
    
        /* calculate the offsets for each plane */
        calc_cropping_offsets(offsets, frame, desc);
    
        /* adjust the offsets to avoid breaking alignment */
        if (!(flags & AV_FRAME_CROP_UNALIGNED)) {
            int log2_crop_align = frame->crop_left ? ff_ctz(frame->crop_left) : INT_MAX;
            int min_log2_align = INT_MAX;
    
            for (i = 0; frame->data[i]; i++) {
                int log2_align = offsets[i] ? ff_ctz(offsets[i]) : INT_MAX;
                min_log2_align = FFMIN(log2_align, min_log2_align);
            }
    
            /* we assume, and it should always be true, that the data alignment is
             * related to the cropping alignment by a constant power-of-2 factor */
            if (log2_crop_align < min_log2_align)
                return AVERROR_BUG;
    
            if (min_log2_align < 5) {
                frame->crop_left &= ~((1 << (5 + log2_crop_align - min_log2_align)) - 1);
                calc_cropping_offsets(offsets, frame, desc);
            }
        }
    
        for (i = 0; frame->data[i]; i++)
            frame->data[i] += offsets[i];
    
        frame->width      -= (frame->crop_left + frame->crop_right);
        frame->height     -= (frame->crop_top  + frame->crop_bottom);
        frame->crop_left   = 0;
        frame->crop_right  = 0;
        frame->crop_top    = 0;
        frame->crop_bottom = 0;
    
        return 0;
    }