aboutsummaryrefslogtreecommitdiffstats
path: root/libavformat/gxfenc.c
diff options
context:
space:
mode:
authorBaptiste Coudurier <baptiste.coudurier@gmail.com>2009-03-09 01:10:28 +0000
committerBaptiste Coudurier <baptiste.coudurier@gmail.com>2009-03-09 01:10:28 +0000
commita96b1a903a9f2f7dbec4079415daf5bb4fe2befa (patch)
treea60c00724a4c22b9dc295d7d4fa0f74b3098fdc1 /libavformat/gxfenc.c
parent9885e9a63a9842ed94544d65cb89a2a91d4aecf6 (diff)
downloadffmpeg-a96b1a903a9f2f7dbec4079415daf5bb4fe2befa.tar.gz
correctly interleave audio and video according to specs
Originally committed as revision 17892 to svn://svn.ffmpeg.org/ffmpeg/trunk
Diffstat (limited to 'libavformat/gxfenc.c')
-rw-r--r--libavformat/gxfenc.c51
1 files changed, 42 insertions, 9 deletions
diff --git a/libavformat/gxfenc.c b/libavformat/gxfenc.c
index 710b1dfa6f..b422ac1a1d 100644
--- a/libavformat/gxfenc.c
+++ b/libavformat/gxfenc.c
@@ -43,6 +43,7 @@ typedef struct GXFStreamContext {
int p_per_gop;
int b_per_i_or_p; ///< number of B frames per I frame or P frame
int first_gop_closed;
+ unsigned order; ///< interleaving order
} GXFStreamContext;
typedef struct GXFContext {
@@ -56,7 +57,7 @@ typedef struct GXFContext {
uint32_t umf_length;
uint16_t umf_track_size;
uint16_t umf_media_size;
- int sample_rate;
+ AVRational time_base;
int flags;
GXFStreamContext timecode_track;
} GXFContext;
@@ -370,13 +371,14 @@ static int gxf_write_umf_material_description(AVFormatContext *s)
{
GXFContext *gxf = s->priv_data;
ByteIOContext *pb = s->pb;
+ int timecode_base = gxf->time_base.den == 60000 ? 60 : 50;
// XXX drop frame
uint32_t timecode =
- gxf->nb_fields / (gxf->sample_rate * 3600) % 24 << 24 | // hours
- gxf->nb_fields / (gxf->sample_rate * 60) % 60 << 16 | // minutes
- gxf->nb_fields / gxf->sample_rate % 60 << 8 | // seconds
- gxf->nb_fields % gxf->sample_rate; // fields
+ gxf->nb_fields / (timecode_base * 3600) % 24 << 24 | // hours
+ gxf->nb_fields / (timecode_base * 60) % 60 << 16 | // minutes
+ gxf->nb_fields / timecode_base % 60 << 8 | // seconds
+ gxf->nb_fields % timecode_base; // fields
put_le32(pb, gxf->flags);
put_le32(pb, gxf->nb_fields); /* length of the longest track */
@@ -647,19 +649,24 @@ static int gxf_write_header(AVFormatContext *s)
gxf->flags |= 0x04000000; /* audio is 16 bit pcm */
media_info = 'A';
} else if (st->codec->codec_type == CODEC_TYPE_VIDEO) {
+ if (i != 0) {
+ av_log(s, AV_LOG_ERROR, "video stream must be the first track\n");
+ return -1;
+ }
/* FIXME check from time_base ? */
if (st->codec->height == 480 || st->codec->height == 512) { /* NTSC or NTSC+VBI */
sc->frame_rate_index = 5;
sc->sample_rate = 60;
gxf->flags |= 0x00000080;
+ gxf->time_base = (AVRational){ 1001, 60000 };
} else { /* assume PAL */
sc->frame_rate_index = 6;
sc->media_type++;
sc->sample_rate = 50;
gxf->flags |= 0x00000040;
+ gxf->time_base = (AVRational){ 1, 50 };
}
- gxf->sample_rate = sc->sample_rate;
- av_set_pts_info(st, 64, 1, st->codec->time_base.den);
+ av_set_pts_info(st, 64, gxf->time_base.num, gxf->time_base.den);
if (gxf_find_lines_index(st) < 0)
sc->lines_index = -1;
sc->sample_size = st->codec->bit_rate;
@@ -704,6 +711,7 @@ static int gxf_write_header(AVFormatContext *s)
}
/* FIXME first 10 audio tracks are 0 to 9 next 22 are A to V */
sc->media_info = media_info<<8 | ('0'+tracks[media_info]++);
+ sc->order = s->nb_streams - st->index;
}
if (ff_audio_interleave_init(s, GXF_samples_per_frame, (AVRational){ 1, 48000 }) < 0)
@@ -770,7 +778,8 @@ static int gxf_write_media_preamble(AVFormatContext *s, AVPacket *pkt, int size)
if (st->codec->codec_type == CODEC_TYPE_VIDEO) {
field_nb = gxf->nb_fields;
} else {
- field_nb = av_rescale_rnd(pkt->dts, gxf->sample_rate, st->codec->time_base.den, AV_ROUND_UP);
+ field_nb = av_rescale_rnd(pkt->dts, gxf->time_base.den,
+ (int64_t)48000*gxf->time_base.num, AV_ROUND_UP);
}
put_byte(pb, sc->media_type);
@@ -828,10 +837,34 @@ static int gxf_write_packet(AVFormatContext *s, AVPacket *pkt)
return updatePacketSize(pb, pos);
}
+static int gxf_compare_field_nb(AVFormatContext *s, AVPacket *next, AVPacket *cur)
+{
+ GXFContext *gxf = s->priv_data;
+ AVPacket *pkt[2] = { cur, next };
+ int i, field_nb[2];
+ GXFStreamContext *sc[2];
+
+ for (i = 0; i < 2; i++) {
+ AVStream *st = s->streams[pkt[i]->stream_index];
+ sc[i] = st->priv_data;
+ if (st->codec->codec_type == CODEC_TYPE_AUDIO) {
+ field_nb[i] = av_rescale_rnd(pkt[i]->dts, gxf->time_base.den,
+ (int64_t)48000*gxf->time_base.num, AV_ROUND_UP);
+ field_nb[i] &= ~1; // compare against even field number because audio must be before video
+ } else
+ field_nb[i] = pkt[i]->dts; // dts are field based
+ }
+
+ return field_nb[1] > field_nb[0] ||
+ (field_nb[1] == field_nb[0] && sc[1]->order > sc[0]->order);
+}
+
static int gxf_interleave_packet(AVFormatContext *s, AVPacket *out, AVPacket *pkt, int flush)
{
+ if (pkt && s->streams[pkt->stream_index]->codec->codec_type == CODEC_TYPE_VIDEO)
+ pkt->duration = 2; // enforce 2 fields
return ff_audio_rechunk_interleave(s, out, pkt, flush,
- av_interleave_packet_per_dts, ff_interleave_compare_dts);
+ av_interleave_packet_per_dts, gxf_compare_field_nb);
}
AVOutputFormat gxf_muxer = {