aboutsummaryrefslogtreecommitdiffstats
path: root/libavcodec
diff options
context:
space:
mode:
authorRick Kern <kernrj@gmail.com>2016-06-02 02:25:21 -0400
committerRick Kern <kernrj@gmail.com>2016-06-16 21:02:41 -0400
commit143685a42bbc8861b626457ce4cb8b1ce4b0c436 (patch)
tree707b7631d2899aefcb70d2020a27e7eb1a4f1a73 /libavcodec
parentebda8216dcbc5251491ba2a723deb55ccaef4c30 (diff)
downloadffmpeg-143685a42bbc8861b626457ce4cb8b1ce4b0c436.tar.gz
lavc/audiotoolboxenc: fix dropped frames on iOS
AudioConverterFillComplexBuffer() doesn't always call its callback. A frame queue is used to prevent skipped audio samples. Signed-off-by: Rick Kern <kernrj@gmail.com>
Diffstat (limited to 'libavcodec')
-rw-r--r--libavcodec/audiotoolboxenc.c77
1 files changed, 50 insertions, 27 deletions
diff --git a/libavcodec/audiotoolboxenc.c b/libavcodec/audiotoolboxenc.c
index 855df0cf04..c47fbd1b2d 100644
--- a/libavcodec/audiotoolboxenc.c
+++ b/libavcodec/audiotoolboxenc.c
@@ -22,6 +22,9 @@
#include <AudioToolbox/AudioToolbox.h>
+#define FF_BUFQUEUE_SIZE 256
+#include "libavfilter/bufferqueue.h"
+
#include "config.h"
#include "audio_frame_queue.h"
#include "avcodec.h"
@@ -38,8 +41,8 @@ typedef struct ATDecodeContext {
int quality;
AudioConverterRef converter;
- AVFrame in_frame;
- AVFrame new_in_frame;
+ struct FFBufQueue frame_queue;
+ struct FFBufQueue used_frame_queue;
unsigned pkt_size;
AudioFrameQueue afq;
@@ -449,28 +452,30 @@ static OSStatus ffat_encode_callback(AudioConverterRef converter, UInt32 *nb_pac
{
AVCodecContext *avctx = inctx;
ATDecodeContext *at = avctx->priv_data;
+ AVFrame *frame;
- if (at->eof) {
- *nb_packets = 0;
- return 0;
+ if (!at->frame_queue.available) {
+ if (at->eof) {
+ *nb_packets = 0;
+ return 0;
+ } else {
+ *nb_packets = 0;
+ return 1;
+ }
}
- av_frame_unref(&at->in_frame);
- av_frame_move_ref(&at->in_frame, &at->new_in_frame);
-
- if (!at->in_frame.data[0]) {
- *nb_packets = 0;
- return 1;
- }
+ frame = ff_bufqueue_get(&at->frame_queue);
data->mNumberBuffers = 1;
data->mBuffers[0].mNumberChannels = avctx->channels;
- data->mBuffers[0].mDataByteSize = at->in_frame.nb_samples *
+ data->mBuffers[0].mDataByteSize = frame->nb_samples *
av_get_bytes_per_sample(avctx->sample_fmt) *
avctx->channels;
- data->mBuffers[0].mData = at->in_frame.data[0];
- if (*nb_packets > at->in_frame.nb_samples)
- *nb_packets = at->in_frame.nb_samples;
+ data->mBuffers[0].mData = frame->data[0];
+ if (*nb_packets > frame->nb_samples)
+ *nb_packets = frame->nb_samples;
+
+ ff_bufqueue_add(avctx, &at->used_frame_queue, frame);
return 0;
}
@@ -492,20 +497,35 @@ static int ffat_encode(AVCodecContext *avctx, AVPacket *avpkt,
};
AudioStreamPacketDescription out_pkt_desc = {0};
- if ((ret = ff_alloc_packet2(avctx, avpkt, at->pkt_size, 0)) < 0)
- return ret;
-
- av_frame_unref(&at->new_in_frame);
-
if (frame) {
+ AVFrame *in_frame;
+
+ if (ff_bufqueue_is_full(&at->frame_queue)) {
+ /*
+ * The frame queue is significantly larger than needed in practice,
+ * but no clear way to determine the minimum number of samples to
+ * get output from AudioConverterFillComplexBuffer().
+ */
+ av_log(avctx, AV_LOG_ERROR, "Bug: frame queue is too small.\n");
+ return AVERROR_BUG;
+ }
+
if ((ret = ff_af_queue_add(&at->afq, frame)) < 0)
return ret;
- if ((ret = av_frame_ref(&at->new_in_frame, frame)) < 0)
- return ret;
+
+ in_frame = av_frame_clone(frame);
+ if (!in_frame)
+ return AVERROR(ENOMEM);
+
+ ff_bufqueue_add(avctx, &at->frame_queue, in_frame);
} else {
at->eof = 1;
}
+ if ((ret = ff_alloc_packet2(avctx, avpkt, at->pkt_size, 0)) < 0)
+ return ret;
+
+
out_buffers.mBuffers[0].mData = avpkt->data;
*got_packet_ptr = avctx->frame_size / at->frame_size;
@@ -513,6 +533,9 @@ static int ffat_encode(AVCodecContext *avctx, AVPacket *avpkt,
ret = AudioConverterFillComplexBuffer(at->converter, ffat_encode_callback, avctx,
got_packet_ptr, &out_buffers,
(avctx->frame_size > at->frame_size) ? NULL : &out_pkt_desc);
+
+ ff_bufqueue_discard_all(&at->used_frame_queue);
+
if ((!ret || ret == 1) && *got_packet_ptr) {
avpkt->size = out_buffers.mBuffers[0].mDataByteSize;
ff_af_queue_remove(&at->afq, out_pkt_desc.mVariableFramesInPacket ?
@@ -531,16 +554,16 @@ static av_cold void ffat_encode_flush(AVCodecContext *avctx)
{
ATDecodeContext *at = avctx->priv_data;
AudioConverterReset(at->converter);
- av_frame_unref(&at->new_in_frame);
- av_frame_unref(&at->in_frame);
+ ff_bufqueue_discard_all(&at->frame_queue);
+ ff_bufqueue_discard_all(&at->used_frame_queue);
}
static av_cold int ffat_close_encoder(AVCodecContext *avctx)
{
ATDecodeContext *at = avctx->priv_data;
AudioConverterDispose(at->converter);
- av_frame_unref(&at->new_in_frame);
- av_frame_unref(&at->in_frame);
+ ff_bufqueue_discard_all(&at->frame_queue);
+ ff_bufqueue_discard_all(&at->used_frame_queue);
ff_af_queue_close(&at->afq);
return 0;
}