aboutsummaryrefslogtreecommitdiffstats
path: root/libavcodec
diff options
context:
space:
mode:
authorZhao Zhili <zhilizhao@tencent.com>2022-12-07 17:31:20 +0800
committerZhao Zhili <zhilizhao@tencent.com>2022-12-17 04:30:36 +0800
commit7da1adfc026a57ab3fb42516354f8e62f3f7d35d (patch)
treeb4e935932da448257defa760c6a672552a047e4a /libavcodec
parente3f2d01d709f35e6e9544d542825eee4ef1d13b5 (diff)
downloadffmpeg-7da1adfc026a57ab3fb42516354f8e62f3f7d35d.tar.gz
avcodec/mediacodecenc: remove the strategy to create DTS
Use input PTS as DTS has multiple problems: 1. If there is no reordering, it's better to just use the output PTS as DTS, since encoder may change the timestamp value (do it on purpose or rounding error). 2. If there is reordering, input PTS should be shift a few frames as DTS to satisfy the requirement of PTS >= DTS. I can't find a reliable way to determine how many frames to be shift. For example, we don't known if the encoder use hierarchical B frames. The max_num_reorder_frames can be get from VUI, but VUI is optional. 3. Encoder dropping frames makes the case worse. Android has an BITRATE_MODE_CBR_FD option to allow it explicitly. Signed-off-by: Zhao Zhili <zhilizhao@tencent.com>
Diffstat (limited to 'libavcodec')
-rw-r--r--libavcodec/mediacodecenc.c28
1 files changed, 1 insertions, 27 deletions
diff --git a/libavcodec/mediacodecenc.c b/libavcodec/mediacodecenc.c
index 4e8716e3a5..96bc9c9c28 100644
--- a/libavcodec/mediacodecenc.c
+++ b/libavcodec/mediacodecenc.c
@@ -64,18 +64,6 @@ typedef struct MediaCodecEncContext {
uint8_t *extradata;
int extradata_size;
-
- // Since MediaCodec doesn't output DTS, use a timestamp queue to save pts
- // of AVFrame and generate DTS for AVPacket.
- //
- // This doesn't work when use Surface as input, in that case frames can be
- // sent to encoder without our notice. One exception is frames come from
- // our MediaCodec decoder wrapper, since we can control it's render by
- // av_mediacodec_release_buffer.
- int64_t timestamps[32];
- int ts_head;
- int ts_tail;
-
int eof_sent;
AVFrame *frame;
@@ -379,11 +367,6 @@ static int mediacodec_receive(AVCodecContext *avctx,
}
memcpy(pkt->data + extradata_size, out_buf + out_info.offset, out_info.size);
pkt->pts = av_rescale_q(out_info.presentationTimeUs, AV_TIME_BASE_Q, avctx->time_base);
- if (s->ts_tail != s->ts_head) {
- pkt->dts = s->timestamps[s->ts_tail];
- s->ts_tail = (s->ts_tail + 1) % FF_ARRAY_ELEMS(s->timestamps);
- }
-
if (out_info.flags & ff_AMediaCodec_getBufferFlagKeyFrame(codec))
pkt->flags |= AV_PKT_FLAG_KEY;
ret = 0;
@@ -448,14 +431,8 @@ static int mediacodec_send(AVCodecContext *avctx,
return ff_AMediaCodec_signalEndOfInputStream(codec);
}
-
- if (frame->data[3]) {
- pts = av_rescale_q(frame->pts, avctx->time_base, AV_TIME_BASE_Q);
- s->timestamps[s->ts_head] = frame->pts;
- s->ts_head = (s->ts_head + 1) % FF_ARRAY_ELEMS(s->timestamps);
-
+ if (frame->data[3])
av_mediacodec_release_buffer((AVMediaCodecBuffer *)frame->data[3], 1);
- }
return 0;
}
@@ -474,9 +451,6 @@ static int mediacodec_send(AVCodecContext *avctx,
copy_frame_to_buffer(avctx, frame, input_buf, input_size);
pts = av_rescale_q(frame->pts, avctx->time_base, AV_TIME_BASE_Q);
-
- s->timestamps[s->ts_head] = frame->pts;
- s->ts_head = (s->ts_head + 1) % FF_ARRAY_ELEMS(s->timestamps);
} else {
flags |= ff_AMediaCodec_getBufferFlagEndOfStream(codec);
s->eof_sent = 1;