diff options
author | Anton Khirnov <anton@khirnov.net> | 2020-10-29 16:00:23 +0100 |
---|---|---|
committer | Anton Khirnov <anton@khirnov.net> | 2020-11-20 15:20:24 +0100 |
commit | 29f33c107682162f9a9413d0d6760fc6974e860d (patch) | |
tree | 92ec069e8a8745c8cb996d2d64ea351d2012e594 /doc/examples/transcoding.c | |
parent | 800feae5d65ef07910857e0417b63bd6e1d15288 (diff) | |
download | ffmpeg-29f33c107682162f9a9413d0d6760fc6974e860d.tar.gz |
doc/examples/transcoding: stop constantly allocating AVFrames
Allocate just one and reuse it.
Diffstat (limited to 'doc/examples/transcoding.c')
-rw-r--r-- | doc/examples/transcoding.c | 60 |
1 files changed, 31 insertions, 29 deletions
diff --git a/doc/examples/transcoding.c b/doc/examples/transcoding.c index e48837cbd2..4a45717bea 100644 --- a/doc/examples/transcoding.c +++ b/doc/examples/transcoding.c @@ -41,12 +41,16 @@ typedef struct FilteringContext { AVFilterContext *buffersink_ctx; AVFilterContext *buffersrc_ctx; AVFilterGraph *filter_graph; + + AVFrame *filtered_frame; } FilteringContext; static FilteringContext *filter_ctx; typedef struct StreamContext { AVCodecContext *dec_ctx; AVCodecContext *enc_ctx; + + AVFrame *dec_frame; } StreamContext; static StreamContext *stream_ctx; @@ -102,6 +106,10 @@ static int open_input_file(const char *filename) } } stream_ctx[i].dec_ctx = codec_ctx; + + stream_ctx[i].dec_frame = av_frame_alloc(); + if (!stream_ctx[i].dec_frame) + return AVERROR(ENOMEM); } av_dump_format(ifmt_ctx, 0, filename, 0); @@ -398,6 +406,10 @@ static int init_filters(void) stream_ctx[i].enc_ctx, filter_spec); if (ret) return ret; + + filter_ctx[i].filtered_frame = av_frame_alloc(); + if (!filter_ctx[i].filtered_frame) + return AVERROR(ENOMEM); } return 0; } @@ -420,7 +432,6 @@ static int encode_write_frame(AVFrame *filt_frame, unsigned int stream_index, in av_init_packet(&enc_pkt); ret = enc_func(stream_ctx[stream_index].enc_ctx, &enc_pkt, filt_frame, got_frame); - av_frame_free(&filt_frame); if (ret < 0) return ret; if (!(*got_frame)) @@ -440,12 +451,12 @@ static int encode_write_frame(AVFrame *filt_frame, unsigned int stream_index, in static int filter_encode_write_frame(AVFrame *frame, unsigned int stream_index) { + FilteringContext *filter = &filter_ctx[stream_index]; int ret; - AVFrame *filt_frame; av_log(NULL, AV_LOG_INFO, "Pushing decoded frame to filters\n"); /* push the decoded frame into the filtergraph */ - ret = av_buffersrc_add_frame_flags(filter_ctx[stream_index].buffersrc_ctx, + ret = av_buffersrc_add_frame_flags(filter->buffersrc_ctx, frame, 0); if (ret < 0) { av_log(NULL, AV_LOG_ERROR, "Error while feeding the filtergraph\n"); @@ -454,14 +465,9 @@ static int filter_encode_write_frame(AVFrame *frame, unsigned int stream_index) /* pull filtered frames from the filtergraph */ while (1) { - filt_frame = av_frame_alloc(); - if (!filt_frame) { - ret = AVERROR(ENOMEM); - break; - } av_log(NULL, AV_LOG_INFO, "Pulling filtered frame from filters\n"); - ret = av_buffersink_get_frame(filter_ctx[stream_index].buffersink_ctx, - filt_frame); + ret = av_buffersink_get_frame(filter->buffersink_ctx, + filter->filtered_frame); if (ret < 0) { /* if no more frames for output - returns AVERROR(EAGAIN) * if flushed and no more frames for output - returns AVERROR_EOF @@ -469,12 +475,12 @@ static int filter_encode_write_frame(AVFrame *frame, unsigned int stream_index) */ if (ret == AVERROR(EAGAIN) || ret == AVERROR_EOF) ret = 0; - av_frame_free(&filt_frame); break; } - filt_frame->pict_type = AV_PICTURE_TYPE_NONE; - ret = encode_write_frame(filt_frame, stream_index, NULL); + filter->filtered_frame->pict_type = AV_PICTURE_TYPE_NONE; + ret = encode_write_frame(filter->filtered_frame, stream_index, NULL); + av_frame_unref(filter->filtered_frame); if (ret < 0) break; } @@ -506,7 +512,6 @@ int main(int argc, char **argv) { int ret; AVPacket packet = { .data = NULL, .size = 0 }; - AVFrame *frame = NULL; enum AVMediaType type; unsigned int stream_index; unsigned int i; @@ -535,33 +540,27 @@ int main(int argc, char **argv) stream_index); if (filter_ctx[stream_index].filter_graph) { + StreamContext *stream = &stream_ctx[stream_index]; + av_log(NULL, AV_LOG_DEBUG, "Going to reencode&filter the frame\n"); - frame = av_frame_alloc(); - if (!frame) { - ret = AVERROR(ENOMEM); - break; - } + av_packet_rescale_ts(&packet, ifmt_ctx->streams[stream_index]->time_base, - stream_ctx[stream_index].dec_ctx->time_base); + stream->dec_ctx->time_base); dec_func = (type == AVMEDIA_TYPE_VIDEO) ? avcodec_decode_video2 : avcodec_decode_audio4; - ret = dec_func(stream_ctx[stream_index].dec_ctx, frame, + ret = dec_func(stream->dec_ctx, stream->dec_frame, &got_frame, &packet); if (ret < 0) { - av_frame_free(&frame); av_log(NULL, AV_LOG_ERROR, "Decoding failed\n"); break; } if (got_frame) { - frame->pts = frame->best_effort_timestamp; - ret = filter_encode_write_frame(frame, stream_index); - av_frame_free(&frame); + stream->dec_frame->pts = stream->dec_frame->best_effort_timestamp; + ret = filter_encode_write_frame(stream->dec_frame, stream_index); if (ret < 0) goto end; - } else { - av_frame_free(&frame); } } else { /* remux this frame without reencoding */ @@ -598,13 +597,16 @@ int main(int argc, char **argv) av_write_trailer(ofmt_ctx); end: av_packet_unref(&packet); - av_frame_free(&frame); for (i = 0; i < ifmt_ctx->nb_streams; i++) { avcodec_free_context(&stream_ctx[i].dec_ctx); if (ofmt_ctx && ofmt_ctx->nb_streams > i && ofmt_ctx->streams[i] && stream_ctx[i].enc_ctx) avcodec_free_context(&stream_ctx[i].enc_ctx); - if (filter_ctx && filter_ctx[i].filter_graph) + if (filter_ctx && filter_ctx[i].filter_graph) { avfilter_graph_free(&filter_ctx[i].filter_graph); + av_frame_free(&filter_ctx[i].filtered_frame); + } + + av_frame_free(&stream_ctx[i].dec_frame); } av_free(filter_ctx); av_free(stream_ctx); |