aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorMichael Niedermayer <michaelni@gmx.at>2012-04-14 21:49:01 +0200
committerMichael Niedermayer <michaelni@gmx.at>2012-04-14 22:37:43 +0200
commit7432bcfe5a36331502c49fe03aefe39b9936247a (patch)
tree398b1874abbd4bcf7c103738ac4548e830af0a43
parent01bf2ad7351fdaa2e21b6bdf963d22d6ffccb920 (diff)
parent7bf9e3391fa21d90ff283fc03a12287fe73db9e8 (diff)
downloadffmpeg-7432bcfe5a36331502c49fe03aefe39b9936247a.tar.gz
Merge remote-tracking branch 'qatar/master'
* qatar/master: vsrc_buffer: fix check from 7ae7c41. libxvid: Reorder functions to avoid forward declarations; make functions static. libxvid: drop some pointless dead code wmal: vertical alignment cosmetics wmal: Warn about missing bitstream splicing feature and ask for sample. wmal: Skip seekable_frame_in_packet. wmal: Drop unused variable num_possible_block_size. avfiltergraph: make the AVFilterInOut alloc/free API public graphparser: allow specifying sws flags in the graph description. graphparser: fix the order of connecting unlabeled links. graphparser: add avfilter_graph_parse2(). vsrc_buffer: allow using a NULL buffer to signal EOF. swscale: handle last pixel if lines have an odd width. qdm2: fix a dubious pointer cast WMAL: Do not try to read rawpcm coefficients if bits is invalid mov: Fix detecting there is no sync sample. tiffdec: K&R cosmetics avf: has_duration does not check the global one dsputil: fix optimized emu_edge function on Win64. Conflicts: doc/APIchanges libavcodec/libxvid_rc.c libavcodec/libxvidff.c libavcodec/tiff.c libavcodec/wmalosslessdec.c libavfilter/avfiltergraph.h libavfilter/graphparser.c libavfilter/version.h libavfilter/vsrc_buffer.c libswscale/output.c Merged-by: Michael Niedermayer <michaelni@gmx.at>
-rw-r--r--doc/APIchanges3
-rw-r--r--doc/filters.texi8
-rw-r--r--libavcodec/libxvid_rc.c5
-rw-r--r--libavcodec/libxvidff.c541
-rw-r--r--libavcodec/qdm2.c11
-rw-r--r--libavcodec/tiff.c253
-rw-r--r--libavcodec/wmalosslessdec.c19
-rw-r--r--libavcodec/x86/dsputil_yasm.asm2
-rw-r--r--libavfilter/avfiltergraph.h52
-rw-r--r--libavfilter/buffersrc.h1
-rw-r--r--libavfilter/graphparser.c204
-rw-r--r--libavfilter/version.h4
-rw-r--r--libavfilter/vsrc_buffer.c32
-rw-r--r--libavformat/isom.h1
-rw-r--r--libavformat/mov.c5
-rw-r--r--libavformat/utils.c2
-rw-r--r--libswscale/output.c24
-rwxr-xr-xtests/lavfi-regression.sh2
18 files changed, 730 insertions, 439 deletions
diff --git a/doc/APIchanges b/doc/APIchanges
index 9a58117cff..8a0e5e6ee4 100644
--- a/doc/APIchanges
+++ b/doc/APIchanges
@@ -19,6 +19,9 @@ API changes, most recent first:
2012-03-26 - a67d9cf - lavfi 2.66.100
Add avfilter_fill_frame_from_{audio_,}buffer_ref() functions.
+2012-xx-xx - xxxxxxx - lavfi 2.16.0 - avfiltergraph.h
+ Add avfilter_graph_parse2()
+
2012-xx-xx - xxxxxxx - lavu 51.27.0 - samplefmt.h
Add av_get_packed_sample_fmt() and av_get_planar_sample_fmt()
diff --git a/doc/filters.texi b/doc/filters.texi
index 38312a5237..08ed11cc73 100644
--- a/doc/filters.texi
+++ b/doc/filters.texi
@@ -76,6 +76,12 @@ In a complete filterchain all the unlabelled filter input and output
pads must be connected. A filtergraph is considered valid if all the
filter input and output pads of all the filterchains are connected.
+Libavfilter will automatically insert scale filters where format
+conversion is required. It is possible to specify swscale flags
+for those automatically inserted scalers by prepending
+@code{sws_flags=@var{flags};}
+to the filtergraph description.
+
Follows a BNF description for the filtergraph syntax:
@example
@var{NAME} ::= sequence of alphanumeric characters and '_'
@@ -84,7 +90,7 @@ Follows a BNF description for the filtergraph syntax:
@var{FILTER_ARGUMENTS} ::= sequence of chars (eventually quoted)
@var{FILTER} ::= [@var{LINKNAMES}] @var{NAME} ["=" @var{ARGUMENTS}] [@var{LINKNAMES}]
@var{FILTERCHAIN} ::= @var{FILTER} [,@var{FILTERCHAIN}]
-@var{FILTERGRAPH} ::= @var{FILTERCHAIN} [;@var{FILTERGRAPH}]
+@var{FILTERGRAPH} ::= [sws_flags=@var{flags};] @var{FILTERCHAIN} [;@var{FILTERGRAPH}]
@end example
@c man end FILTERGRAPH DESCRIPTION
diff --git a/libavcodec/libxvid_rc.c b/libavcodec/libxvid_rc.c
index a9a1ee971d..71579ebe7a 100644
--- a/libavcodec/libxvid_rc.c
+++ b/libavcodec/libxvid_rc.c
@@ -31,16 +31,12 @@
#undef NDEBUG
#include <assert.h>
-extern unsigned int xvid_debug;
-
int ff_xvid_rate_control_init(MpegEncContext *s){
char *tmp_name;
int fd, i;
xvid_plg_create_t xvid_plg_create = { 0 };
xvid_plugin_2pass2_t xvid_2pass2 = { 0 };
-//xvid_debug=-1;
-
fd=av_tempfile("xvidrc.", &tmp_name, 0, s->avctx);
if (fd == -1) {
av_log(NULL, AV_LOG_ERROR, "Can't create temporary pass2 file.\n");
@@ -58,7 +54,6 @@ int ff_xvid_rate_control_init(MpegEncContext *s){
frame_types[rce->pict_type], (int)lrintf(rce->qscale / FF_QP2LAMBDA), rce->i_count, s->mb_num - rce->i_count - rce->skip_count,
rce->skip_count, (rce->i_tex_bits + rce->p_tex_bits + rce->misc_bits+7)/8, (rce->header_bits+rce->mv_bits+7)/8);
-//av_log(NULL, AV_LOG_ERROR, "%s\n", tmp);
if (write(fd, tmp, strlen(tmp)) < 0) {
av_log(NULL, AV_LOG_ERROR, "Error %s writing 2pass logfile\n", strerror(errno));
return AVERROR(errno);
diff --git a/libavcodec/libxvidff.c b/libavcodec/libxvidff.c
index 811efc3793..0426e9de14 100644
--- a/libavcodec/libxvidff.c
+++ b/libavcodec/libxvidff.c
@@ -72,11 +72,276 @@ struct xvid_ff_pass1 {
struct xvid_context *context; /**< Pointer to private context */
};
-/* Prototypes - See function implementation for details */
-int xvid_strip_vol_header(AVCodecContext *avctx, AVPacket *pkt, unsigned int header_len, unsigned int frame_len);
-int xvid_ff_2pass(void *ref, int opt, void *p1, void *p2);
-void xvid_correct_framerate(AVCodecContext *avctx);
+/*
+ * Xvid 2-Pass Kludge Section
+ *
+ * Xvid's default 2-pass doesn't allow us to create data as we need to, so
+ * this section spends time replacing the first pass plugin so we can write
+ * statistic information as libavcodec requests in. We have another kludge
+ * that allows us to pass data to the second pass in Xvid without a custom
+ * rate-control plugin.
+ */
+
+
+/**
+ * Initialize the two-pass plugin and context.
+ *
+ * @param param Input construction parameter structure
+ * @param handle Private context handle
+ * @return Returns XVID_ERR_xxxx on failure, or 0 on success.
+ */
+static int xvid_ff_2pass_create(xvid_plg_create_t * param,
+ void ** handle) {
+ struct xvid_ff_pass1 *x = (struct xvid_ff_pass1 *)param->param;
+ char *log = x->context->twopassbuffer;
+
+ /* Do a quick bounds check */
+ if( log == NULL )
+ return XVID_ERR_FAIL;
+
+ /* We use snprintf() */
+ /* This is because we can safely prevent a buffer overflow */
+ log[0] = 0;
+ snprintf(log, BUFFER_REMAINING(log),
+ "# ffmpeg 2-pass log file, using xvid codec\n");
+ snprintf(BUFFER_CAT(log), BUFFER_REMAINING(log),
+ "# Do not modify. libxvidcore version: %d.%d.%d\n\n",
+ XVID_VERSION_MAJOR(XVID_VERSION),
+ XVID_VERSION_MINOR(XVID_VERSION),
+ XVID_VERSION_PATCH(XVID_VERSION));
+
+ *handle = x->context;
+ return 0;
+}
+
+/**
+ * Destroy the two-pass plugin context.
+ *
+ * @param ref Context pointer for the plugin
+ * @param param Destrooy context
+ * @return Returns 0, success guaranteed
+ */
+static int xvid_ff_2pass_destroy(struct xvid_context *ref,
+ xvid_plg_destroy_t *param) {
+ /* Currently cannot think of anything to do on destruction */
+ /* Still, the framework should be here for reference/use */
+ if( ref->twopassbuffer != NULL )
+ ref->twopassbuffer[0] = 0;
+ return 0;
+}
+
+/**
+ * Enable fast encode mode during the first pass.
+ *
+ * @param ref Context pointer for the plugin
+ * @param param Frame data
+ * @return Returns 0, success guaranteed
+ */
+static int xvid_ff_2pass_before(struct xvid_context *ref,
+ xvid_plg_data_t *param) {
+ int motion_remove;
+ int motion_replacements;
+ int vop_remove;
+
+ /* Nothing to do here, result is changed too much */
+ if( param->zone && param->zone->mode == XVID_ZONE_QUANT )
+ return 0;
+
+ /* We can implement a 'turbo' first pass mode here */
+ param->quant = 2;
+
+ /* Init values */
+ motion_remove = ~XVID_ME_CHROMA_PVOP &
+ ~XVID_ME_CHROMA_BVOP &
+ ~XVID_ME_EXTSEARCH16 &
+ ~XVID_ME_ADVANCEDDIAMOND16;
+ motion_replacements = XVID_ME_FAST_MODEINTERPOLATE |
+ XVID_ME_SKIP_DELTASEARCH |
+ XVID_ME_FASTREFINE16 |
+ XVID_ME_BFRAME_EARLYSTOP;
+ vop_remove = ~XVID_VOP_MODEDECISION_RD &
+ ~XVID_VOP_FAST_MODEDECISION_RD &
+ ~XVID_VOP_TRELLISQUANT &
+ ~XVID_VOP_INTER4V &
+ ~XVID_VOP_HQACPRED;
+
+ param->vol_flags &= ~XVID_VOL_GMC;
+ param->vop_flags &= vop_remove;
+ param->motion_flags &= motion_remove;
+ param->motion_flags |= motion_replacements;
+
+ return 0;
+}
+
+/**
+ * Capture statistic data and write it during first pass.
+ *
+ * @param ref Context pointer for the plugin
+ * @param param Statistic data
+ * @return Returns XVID_ERR_xxxx on failure, or 0 on success
+ */
+static int xvid_ff_2pass_after(struct xvid_context *ref,
+ xvid_plg_data_t *param) {
+ char *log = ref->twopassbuffer;
+ const char *frame_types = " ipbs";
+ char frame_type;
+
+ /* Quick bounds check */
+ if( log == NULL )
+ return XVID_ERR_FAIL;
+
+ /* Convert the type given to us into a character */
+ if( param->type < 5 && param->type > 0 ) {
+ frame_type = frame_types[param->type];
+ } else {
+ return XVID_ERR_FAIL;
+ }
+
+ snprintf(BUFFER_CAT(log), BUFFER_REMAINING(log),
+ "%c %d %d %d %d %d %d\n",
+ frame_type, param->stats.quant, param->stats.kblks, param->stats.mblks,
+ param->stats.ublks, param->stats.length, param->stats.hlength);
+
+ return 0;
+}
+
+/**
+ * Dispatch function for our custom plugin.
+ * This handles the dispatch for the Xvid plugin. It passes data
+ * on to other functions for actual processing.
+ *
+ * @param ref Context pointer for the plugin
+ * @param cmd The task given for us to complete
+ * @param p1 First parameter (varies)
+ * @param p2 Second parameter (varies)
+ * @return Returns XVID_ERR_xxxx on failure, or 0 on success
+ */
+static int xvid_ff_2pass(void *ref, int cmd, void *p1, void *p2)
+{
+ switch( cmd ) {
+ case XVID_PLG_INFO:
+ case XVID_PLG_FRAME:
+ return 0;
+
+ case XVID_PLG_BEFORE:
+ return xvid_ff_2pass_before(ref, p1);
+
+ case XVID_PLG_CREATE:
+ return xvid_ff_2pass_create(p1, p2);
+ case XVID_PLG_AFTER:
+ return xvid_ff_2pass_after(ref, p1);
+
+ case XVID_PLG_DESTROY:
+ return xvid_ff_2pass_destroy(ref, p1);
+
+ default:
+ return XVID_ERR_FAIL;
+ }
+}
+
+/**
+ * Routine to create a global VO/VOL header for MP4 container.
+ * What we do here is extract the header from the Xvid bitstream
+ * as it is encoded. We also strip the repeated headers from the
+ * bitstream when a global header is requested for MPEG-4 ISO
+ * compliance.
+ *
+ * @param avctx AVCodecContext pointer to context
+ * @param frame Pointer to encoded frame data
+ * @param header_len Length of header to search
+ * @param frame_len Length of encoded frame data
+ * @return Returns new length of frame data
+ */
+static int xvid_strip_vol_header(AVCodecContext *avctx,
+ AVPacket *pkt,
+ unsigned int header_len,
+ unsigned int frame_len) {
+ int vo_len = 0, i;
+
+ for( i = 0; i < header_len - 3; i++ ) {
+ if( pkt->data[i] == 0x00 &&
+ pkt->data[i+1] == 0x00 &&
+ pkt->data[i+2] == 0x01 &&
+ pkt->data[i+3] == 0xB6 ) {
+ vo_len = i;
+ break;
+ }
+ }
+
+ if( vo_len > 0 ) {
+ /* We need to store the header, so extract it */
+ if( avctx->extradata == NULL ) {
+ avctx->extradata = av_malloc(vo_len);
+ memcpy(avctx->extradata, pkt->data, vo_len);
+ avctx->extradata_size = vo_len;
+ }
+ /* Less dangerous now, memmove properly copies the two
+ chunks of overlapping data */
+ memmove(pkt->data, &pkt->data[vo_len], frame_len - vo_len);
+ pkt->size = frame_len - vo_len;
+ }
+ return 0;
+}
+
+/**
+ * Routine to correct a possibly erroneous framerate being fed to us.
+ * Xvid currently chokes on framerates where the ticks per frame is
+ * extremely large. This function works to correct problems in this area
+ * by estimating a new framerate and taking the simpler fraction of
+ * the two presented.
+ *
+ * @param avctx Context that contains the framerate to correct.
+ */
+static void xvid_correct_framerate(AVCodecContext *avctx)
+{
+ int frate, fbase;
+ int est_frate, est_fbase;
+ int gcd;
+ float est_fps, fps;
+
+ frate = avctx->time_base.den;
+ fbase = avctx->time_base.num;
+
+ gcd = av_gcd(frate, fbase);
+ if( gcd > 1 ) {
+ frate /= gcd;
+ fbase /= gcd;
+ }
+
+ if( frate <= 65000 && fbase <= 65000 ) {
+ avctx->time_base.den = frate;
+ avctx->time_base.num = fbase;
+ return;
+ }
+
+ fps = (float)frate / (float)fbase;
+ est_fps = roundf(fps * 1000.0) / 1000.0;
+
+ est_frate = (int)est_fps;
+ if( est_fps > (int)est_fps ) {
+ est_frate = (est_frate + 1) * 1000;
+ est_fbase = (int)roundf((float)est_frate / est_fps);
+ } else
+ est_fbase = 1;
+
+ gcd = av_gcd(est_frate, est_fbase);
+ if( gcd > 1 ) {
+ est_frate /= gcd;
+ est_fbase /= gcd;
+ }
+
+ if( fbase > est_fbase ) {
+ avctx->time_base.den = est_frate;
+ avctx->time_base.num = est_fbase;
+ av_log(avctx, AV_LOG_DEBUG,
+ "Xvid: framerate re-estimated: %.2f, %.3f%% correction\n",
+ est_fps, (((est_fps - fps)/fps) * 100.0));
+ } else {
+ avctx->time_base.den = frate;
+ avctx->time_base.num = fbase;
+ }
+}
/**
* Create the private context for the encoder.
@@ -509,274 +774,6 @@ static av_cold int xvid_encode_close(AVCodecContext *avctx) {
}
/**
- * Routine to create a global VO/VOL header for MP4 container.
- * What we do here is extract the header from the Xvid bitstream
- * as it is encoded. We also strip the repeated headers from the
- * bitstream when a global header is requested for MPEG-4 ISO
- * compliance.
- *
- * @param avctx AVCodecContext pointer to context
- * @param frame Pointer to encoded frame data
- * @param header_len Length of header to search
- * @param frame_len Length of encoded frame data
- * @return Returns new length of frame data
- */
-int xvid_strip_vol_header(AVCodecContext *avctx,
- AVPacket *pkt,
- unsigned int header_len,
- unsigned int frame_len) {
- int vo_len = 0, i;
-
- for( i = 0; i < header_len - 3; i++ ) {
- if( pkt->data[i] == 0x00 &&
- pkt->data[i+1] == 0x00 &&
- pkt->data[i+2] == 0x01 &&
- pkt->data[i+3] == 0xB6 ) {
- vo_len = i;
- break;
- }
- }
-
- if( vo_len > 0 ) {
- /* We need to store the header, so extract it */
- if( avctx->extradata == NULL ) {
- avctx->extradata = av_malloc(vo_len);
- memcpy(avctx->extradata, pkt->data, vo_len);
- avctx->extradata_size = vo_len;
- }
- /* Less dangerous now, memmove properly copies the two
- chunks of overlapping data */
- memmove(pkt->data, &pkt->data[vo_len], frame_len - vo_len);
- pkt->size = frame_len - vo_len;
- }
- return 0;
-}
-
-/**
- * Routine to correct a possibly erroneous framerate being fed to us.
- * Xvid currently chokes on framerates where the ticks per frame is
- * extremely large. This function works to correct problems in this area
- * by estimating a new framerate and taking the simpler fraction of
- * the two presented.
- *
- * @param avctx Context that contains the framerate to correct.
- */
-void xvid_correct_framerate(AVCodecContext *avctx) {
- int frate, fbase;
- int est_frate, est_fbase;
- int gcd;
- float est_fps, fps;
-
- frate = avctx->time_base.den;
- fbase = avctx->time_base.num;
-
- gcd = av_gcd(frate, fbase);
- if( gcd > 1 ) {
- frate /= gcd;
- fbase /= gcd;
- }
-
- if( frate <= 65000 && fbase <= 65000 ) {
- avctx->time_base.den = frate;
- avctx->time_base.num = fbase;
- return;
- }
-
- fps = (float)frate / (float)fbase;
- est_fps = roundf(fps * 1000.0) / 1000.0;
-
- est_frate = (int)est_fps;
- if( est_fps > (int)est_fps ) {
- est_frate = (est_frate + 1) * 1000;
- est_fbase = (int)roundf((float)est_frate / est_fps);
- } else
- est_fbase = 1;
-
- gcd = av_gcd(est_frate, est_fbase);
- if( gcd > 1 ) {
- est_frate /= gcd;
- est_fbase /= gcd;
- }
-
- if( fbase > est_fbase ) {
- avctx->time_base.den = est_frate;
- avctx->time_base.num = est_fbase;
- av_log(avctx, AV_LOG_DEBUG,
- "Xvid: framerate re-estimated: %.2f, %.3f%% correction\n",
- est_fps, (((est_fps - fps)/fps) * 100.0));
- } else {
- avctx->time_base.den = frate;
- avctx->time_base.num = fbase;
- }
-}
-
-/*
- * Xvid 2-Pass Kludge Section
- *
- * Xvid's default 2-pass doesn't allow us to create data as we need to, so
- * this section spends time replacing the first pass plugin so we can write
- * statistic information as libavcodec requests in. We have another kludge
- * that allows us to pass data to the second pass in Xvid without a custom
- * rate-control plugin.
- */
-
-/**
- * Initialize the two-pass plugin and context.
- *
- * @param param Input construction parameter structure
- * @param handle Private context handle
- * @return Returns XVID_ERR_xxxx on failure, or 0 on success.
- */
-static int xvid_ff_2pass_create(xvid_plg_create_t * param,
- void ** handle) {
- struct xvid_ff_pass1 *x = (struct xvid_ff_pass1 *)param->param;
- char *log = x->context->twopassbuffer;
-
- /* Do a quick bounds check */
- if( log == NULL )
- return XVID_ERR_FAIL;
-
- /* We use snprintf() */
- /* This is because we can safely prevent a buffer overflow */
- log[0] = 0;
- snprintf(log, BUFFER_REMAINING(log),
- "# ffmpeg 2-pass log file, using xvid codec\n");
- snprintf(BUFFER_CAT(log), BUFFER_REMAINING(log),
- "# Do not modify. libxvidcore version: %d.%d.%d\n\n",
- XVID_VERSION_MAJOR(XVID_VERSION),
- XVID_VERSION_MINOR(XVID_VERSION),
- XVID_VERSION_PATCH(XVID_VERSION));
-
- *handle = x->context;
- return 0;
-}
-
-/**
- * Destroy the two-pass plugin context.
- *
- * @param ref Context pointer for the plugin
- * @param param Destrooy context
- * @return Returns 0, success guaranteed
- */
-static int xvid_ff_2pass_destroy(struct xvid_context *ref,
- xvid_plg_destroy_t *param) {
- /* Currently cannot think of anything to do on destruction */
- /* Still, the framework should be here for reference/use */
- if( ref->twopassbuffer != NULL )
- ref->twopassbuffer[0] = 0;
- return 0;
-}
-
-/**
- * Enable fast encode mode during the first pass.
- *
- * @param ref Context pointer for the plugin
- * @param param Frame data
- * @return Returns 0, success guaranteed
- */
-static int xvid_ff_2pass_before(struct xvid_context *ref,
- xvid_plg_data_t *param) {
- int motion_remove;
- int motion_replacements;
- int vop_remove;
-
- /* Nothing to do here, result is changed too much */
- if( param->zone && param->zone->mode == XVID_ZONE_QUANT )
- return 0;
-
- /* We can implement a 'turbo' first pass mode here */
- param->quant = 2;
-
- /* Init values */
- motion_remove = ~XVID_ME_CHROMA_PVOP &
- ~XVID_ME_CHROMA_BVOP &
- ~XVID_ME_EXTSEARCH16 &
- ~XVID_ME_ADVANCEDDIAMOND16;
- motion_replacements = XVID_ME_FAST_MODEINTERPOLATE |
- XVID_ME_SKIP_DELTASEARCH |
- XVID_ME_FASTREFINE16 |
- XVID_ME_BFRAME_EARLYSTOP;
- vop_remove = ~XVID_VOP_MODEDECISION_RD &
- ~XVID_VOP_FAST_MODEDECISION_RD &
- ~XVID_VOP_TRELLISQUANT &
- ~XVID_VOP_INTER4V &
- ~XVID_VOP_HQACPRED;
-
- param->vol_flags &= ~XVID_VOL_GMC;
- param->vop_flags &= vop_remove;
- param->motion_flags &= motion_remove;
- param->motion_flags |= motion_replacements;
-
- return 0;
-}
-
-/**
- * Capture statistic data and write it during first pass.
- *
- * @param ref Context pointer for the plugin
- * @param param Statistic data
- * @return Returns XVID_ERR_xxxx on failure, or 0 on success
- */
-static int xvid_ff_2pass_after(struct xvid_context *ref,
- xvid_plg_data_t *param) {
- char *log = ref->twopassbuffer;
- const char *frame_types = " ipbs";
- char frame_type;
-
- /* Quick bounds check */
- if( log == NULL )
- return XVID_ERR_FAIL;
-
- /* Convert the type given to us into a character */
- if( param->type < 5 && param->type > 0 ) {
- frame_type = frame_types[param->type];
- } else {
- return XVID_ERR_FAIL;
- }
-
- snprintf(BUFFER_CAT(log), BUFFER_REMAINING(log),
- "%c %d %d %d %d %d %d\n",
- frame_type, param->stats.quant, param->stats.kblks, param->stats.mblks,
- param->stats.ublks, param->stats.length, param->stats.hlength);
-
- return 0;
-}
-
-/**
- * Dispatch function for our custom plugin.
- * This handles the dispatch for the Xvid plugin. It passes data
- * on to other functions for actual processing.
- *
- * @param ref Context pointer for the plugin
- * @param cmd The task given for us to complete
- * @param p1 First parameter (varies)
- * @param p2 Second parameter (varies)
- * @return Returns XVID_ERR_xxxx on failure, or 0 on success
- */
-int xvid_ff_2pass(void *ref, int cmd, void *p1, void *p2) {
- switch( cmd ) {
- case XVID_PLG_INFO:
- case XVID_PLG_FRAME:
- return 0;
-
- case XVID_PLG_BEFORE:
- return xvid_ff_2pass_before(ref, p1);
-
- case XVID_PLG_CREATE:
- return xvid_ff_2pass_create(p1, p2);
-
- case XVID_PLG_AFTER:
- return xvid_ff_2pass_after(ref, p1);
-
- case XVID_PLG_DESTROY:
- return xvid_ff_2pass_destroy(ref, p1);
-
- default:
- return XVID_ERR_FAIL;
- }
-}
-
-/**
* Xvid codec definition for libavcodec.
*/
AVCodec ff_libxvid_encoder = {
diff --git a/libavcodec/qdm2.c b/libavcodec/qdm2.c
index bfac583973..74a138bc70 100644
--- a/libavcodec/qdm2.c
+++ b/libavcodec/qdm2.c
@@ -140,7 +140,6 @@ typedef struct {
/// Parameters built from header parameters, do not change during playback
int group_order; ///< order of frame group
int fft_order; ///< order of FFT (actually fftorder+1)
- int fft_frame_size; ///< size of fft frame, in components (1 comples = re + im)
int frame_size; ///< size of data frame
int frequency_range;
int sub_sampling; ///< subsampling: 0=25%, 1=50%, 2=100% */
@@ -1607,13 +1606,17 @@ static void qdm2_fft_tone_synthesizer (QDM2Context *q, int sub_packet)
static void qdm2_calculate_fft (QDM2Context *q, int channel, int sub_packet)
{
const float gain = (q->channels == 1 && q->nb_channels == 2) ? 0.5f : 1.0f;
+ float *out = q->output_buffer + channel;
int i;
q->fft.complex[channel][0].re *= 2.0f;
q->fft.complex[channel][0].im = 0.0f;
q->rdft_ctx.rdft_calc(&q->rdft_ctx, (FFTSample *)q->fft.complex[channel]);
/* add samples to output buffer */
- for (i = 0; i < ((q->fft_frame_size + 15) & ~15); i++)
- q->output_buffer[q->channels * i + channel] += ((float *) q->fft.complex[channel])[i] * gain;
+ for (i = 0; i < FFALIGN(q->fft_size, 8); i++) {
+ out[0] += q->fft.complex[channel][i].re * gain;
+ out[q->channels] += q->fft.complex[channel][i].im * gain;
+ out += 2 * q->channels;
+ }
}
@@ -1688,7 +1691,6 @@ static void dump_context(QDM2Context *q)
PRINT("checksum_size",q->checksum_size);
PRINT("channels",q->channels);
PRINT("nb_channels",q->nb_channels);
- PRINT("fft_frame_size",q->fft_frame_size);
PRINT("fft_size",q->fft_size);
PRINT("sub_sampling",q->sub_sampling);
PRINT("fft_order",q->fft_order);
@@ -1843,7 +1845,6 @@ static av_cold int qdm2_decode_init(AVCodecContext *avctx)
}
s->fft_order = av_log2(s->fft_size) + 1;
- s->fft_frame_size = 2 * s->fft_size; // complex has two floats
// something like max decodable tones
s->group_order = av_log2(s->group_size) + 1;
diff --git a/libavcodec/tiff.c b/libavcodec/tiff.c
index 0d79f6192b..8de1ebd1bf 100644
--- a/libavcodec/tiff.c
+++ b/libavcodec/tiff.c
@@ -52,35 +52,39 @@ typedef struct TiffContext {
int strips, rps, sstype;
int sot;
- const uint8_t* stripdata;
- const uint8_t* stripsizes;
+ const uint8_t *stripdata;
+ const uint8_t *stripsizes;
int stripsize, stripoff;
LZWState *lzw;
} TiffContext;
-static unsigned tget_short(const uint8_t **p, int le) {
+static unsigned tget_short(const uint8_t **p, int le)
+{
unsigned v = le ? AV_RL16(*p) : AV_RB16(*p);
*p += 2;
return v;
}
-static unsigned tget_long(const uint8_t **p, int le) {
+static unsigned tget_long(const uint8_t **p, int le)
+{
unsigned v = le ? AV_RL32(*p) : AV_RB32(*p);
*p += 4;
return v;
}
-static unsigned tget(const uint8_t **p, int type, int le) {
- switch(type){
+static unsigned tget(const uint8_t **p, int type, int le)
+{
+ switch (type) {
case TIFF_BYTE : return *(*p)++;
case TIFF_SHORT: return tget_short(p, le);
- case TIFF_LONG : return tget_long (p, le);
+ case TIFF_LONG : return tget_long(p, le);
default : return UINT_MAX;
}
}
#if CONFIG_ZLIB
-static int tiff_uncompress(uint8_t *dst, unsigned long *len, const uint8_t *src, int size)
+static int tiff_uncompress(uint8_t *dst, unsigned long *len, const uint8_t *src,
+ int size)
{
z_stream zstream = { 0 };
int zret;
@@ -141,7 +145,9 @@ static void av_always_inline horizontal_fill(unsigned int bpp, uint8_t* dst,
}
}
-static int tiff_unpack_strip(TiffContext *s, uint8_t* dst, int stride, const uint8_t *src, int size, int lines){
+static int tiff_unpack_strip(TiffContext *s, uint8_t *dst, int stride,
+ const uint8_t *src, int size, int lines)
+{
int c, line, pixels, code;
const uint8_t *ssrc = src;
int width = ((s->width * s->bpp) + 7) >> 3;
@@ -150,21 +156,24 @@ static int tiff_unpack_strip(TiffContext *s, uint8_t* dst, int stride, const uin
return AVERROR_INVALIDDATA;
#if CONFIG_ZLIB
- if(s->compr == TIFF_DEFLATE || s->compr == TIFF_ADOBE_DEFLATE){
- uint8_t *zbuf; unsigned long outlen;
+ if (s->compr == TIFF_DEFLATE || s->compr == TIFF_ADOBE_DEFLATE) {
+ uint8_t *zbuf;
+ unsigned long outlen;
int ret;
outlen = width * lines;
zbuf = av_malloc(outlen);
if (!zbuf)
return AVERROR(ENOMEM);
ret = tiff_uncompress(zbuf, &outlen, src, size);
- if(ret != Z_OK){
- av_log(s->avctx, AV_LOG_ERROR, "Uncompressing failed (%lu of %lu) with error %d\n", outlen, (unsigned long)width * lines, ret);
+ if (ret != Z_OK) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Uncompressing failed (%lu of %lu) with error %d\n", outlen,
+ (unsigned long)width * lines, ret);
av_free(zbuf);
return -1;
}
src = zbuf;
- for(line = 0; line < lines; line++){
+ for (line = 0; line < lines; line++) {
if(s->bpp < 8 && s->avctx->pix_fmt == PIX_FMT_PAL8){
horizontal_fill(s->bpp, dst, 1, src, 0, width, 0);
}else{
@@ -177,37 +186,42 @@ static int tiff_unpack_strip(TiffContext *s, uint8_t* dst, int stride, const uin
return 0;
}
#endif
- if(s->compr == TIFF_LZW){
- if(ff_lzw_decode_init(s->lzw, 8, src, size, FF_LZW_TIFF) < 0){
+ if (s->compr == TIFF_LZW) {
+ if (ff_lzw_decode_init(s->lzw, 8, src, size, FF_LZW_TIFF) < 0) {
av_log(s->avctx, AV_LOG_ERROR, "Error initializing LZW decoder\n");
return -1;
}
}
- if(s->compr == TIFF_CCITT_RLE || s->compr == TIFF_G3 || s->compr == TIFF_G4){
+ if (s->compr == TIFF_CCITT_RLE || s->compr == TIFF_G3
+ || s->compr == TIFF_G4) {
int i, ret = 0;
- uint8_t *src2 = av_malloc((unsigned)size + FF_INPUT_BUFFER_PADDING_SIZE);
+ uint8_t *src2 = av_malloc((unsigned)size +
+ FF_INPUT_BUFFER_PADDING_SIZE);
if (!src2) {
- av_log(s->avctx, AV_LOG_ERROR, "Error allocating temporary buffer\n");
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Error allocating temporary buffer\n");
return AVERROR(ENOMEM);
}
- if(s->fax_opts & 2){
- av_log(s->avctx, AV_LOG_ERROR, "Uncompressed fax mode is not supported (yet)\n");
+ if (s->fax_opts & 2) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Uncompressed fax mode is not supported (yet)\n");
av_free(src2);
return -1;
}
- if(!s->fill_order){
+ if (!s->fill_order) {
memcpy(src2, src, size);
- }else{
- for(i = 0; i < size; i++)
+ } else {
+ for (i = 0; i < size; i++)
src2[i] = av_reverse[src[i]];
}
- memset(src2+size, 0, FF_INPUT_BUFFER_PADDING_SIZE);
- switch(s->compr){
+ memset(src2 + size, 0, FF_INPUT_BUFFER_PADDING_SIZE);
+ switch (s->compr) {
case TIFF_CCITT_RLE:
case TIFF_G3:
case TIFF_G4:
- ret = ff_ccitt_unpack(s->avctx, src2, size, dst, lines, stride, s->compr, s->fax_opts);
+ ret = ff_ccitt_unpack(s->avctx, src2, size, dst, lines, stride,
+ s->compr, s->fax_opts);
break;
}
if (s->bpp < 8 && s->avctx->pix_fmt == PIX_FMT_PAL8)
@@ -218,12 +232,12 @@ static int tiff_unpack_strip(TiffContext *s, uint8_t* dst, int stride, const uin
av_free(src2);
return ret;
}
- for(line = 0; line < lines; line++){
- if(src - ssrc > size){
+ for (line = 0; line < lines; line++) {
+ if (src - ssrc > size) {
av_log(s->avctx, AV_LOG_ERROR, "Source data overread\n");
return -1;
}
- switch(s->compr){
+ switch (s->compr) {
case TIFF_RAW:
if (ssrc + size - src < width)
return AVERROR_INVALIDDATA;
@@ -238,22 +252,24 @@ static int tiff_unpack_strip(TiffContext *s, uint8_t* dst, int stride, const uin
src += width;
break;
case TIFF_PACKBITS:
- for(pixels = 0; pixels < width;){
- code = (int8_t)*src++;
- if(code >= 0){
+ for (pixels = 0; pixels < width;) {
+ code = (int8_t) * src++;
+ if (code >= 0) {
code++;
- if(pixels + code > width){
- av_log(s->avctx, AV_LOG_ERROR, "Copy went out of bounds\n");
+ if (pixels + code > width) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Copy went out of bounds\n");
return -1;
}
horizontal_fill(s->bpp * (s->avctx->pix_fmt == PIX_FMT_PAL8),
dst, 1, src, 0, code, pixels);
src += code;
pixels += code;
- }else if(code != -128){ // -127..-1
+ } else if (code != -128) { // -127..-1
code = (-code) + 1;
- if(pixels + code > width){
- av_log(s->avctx, AV_LOG_ERROR, "Run went out of bounds\n");
+ if (pixels + code > width) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Run went out of bounds\n");
return -1;
}
c = *src++;
@@ -265,8 +281,9 @@ static int tiff_unpack_strip(TiffContext *s, uint8_t* dst, int stride, const uin
break;
case TIFF_LZW:
pixels = ff_lzw_decode(s->lzw, dst, width);
- if(pixels < width){
- av_log(s->avctx, AV_LOG_ERROR, "Decoded only %i bytes of %i\n", pixels, width);
+ if (pixels < width) {
+ av_log(s->avctx, AV_LOG_ERROR, "Decoded only %i bytes of %i\n",
+ pixels, width);
return -1;
}
if (s->bpp < 8 && s->avctx->pix_fmt == PIX_FMT_PAL8)
@@ -342,7 +359,8 @@ static int init_image(TiffContext *s)
return 0;
}
-static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *buf, const uint8_t *end_buf)
+static int tiff_decode_tag(TiffContext *s, const uint8_t *start,
+ const uint8_t *buf, const uint8_t *end_buf)
{
unsigned tag, type, count, off, value = 0;
int i, j;
@@ -357,12 +375,13 @@ static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *
off = tget_long(&buf, s->le);
if (type == 0 || type >= FF_ARRAY_ELEMS(type_sizes)) {
- av_log(s->avctx, AV_LOG_DEBUG, "Unknown tiff type (%u) encountered\n", type);
+ av_log(s->avctx, AV_LOG_DEBUG, "Unknown tiff type (%u) encountered\n",
+ type);
return 0;
}
- if(count == 1){
- switch(type){
+ if (count == 1) {
+ switch (type) {
case TIFF_BYTE:
case TIFF_SHORT:
buf -= 4;
@@ -374,7 +393,7 @@ static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *
buf = NULL;
break;
case TIFF_STRING:
- if(count <= 4){
+ if (count <= 4) {
buf -= 4;
break;
}
@@ -390,12 +409,13 @@ static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *
}
}
- if(buf && (buf < start || buf > end_buf)){
- av_log(s->avctx, AV_LOG_ERROR, "Tag referencing position outside the image\n");
+ if (buf && (buf < start || buf > end_buf)) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Tag referencing position outside the image\n");
return -1;
}
- switch(tag){
+ switch (tag) {
case TIFF_WIDTH:
s->width = value;
break;
@@ -404,20 +424,25 @@ static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *
break;
case TIFF_BPP:
s->bppcount = count;
- if(count > 4){
- av_log(s->avctx, AV_LOG_ERROR, "This format is not supported (bpp=%d, %d components)\n", s->bpp, count);
+ if (count > 4) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "This format is not supported (bpp=%d, %d components)\n",
+ s->bpp, count);
return -1;
}
- if(count == 1) s->bpp = value;
- else{
- switch(type){
+ if (count == 1)
+ s->bpp = value;
+ else {
+ switch (type) {
case TIFF_BYTE:
- s->bpp = (off & 0xFF) + ((off >> 8) & 0xFF) + ((off >> 16) & 0xFF) + ((off >> 24) & 0xFF);
+ s->bpp = (off & 0xFF) + ((off >> 8) & 0xFF) +
+ ((off >> 16) & 0xFF) + ((off >> 24) & 0xFF);
break;
case TIFF_SHORT:
case TIFF_LONG:
s->bpp = 0;
- for(i = 0; i < count && buf < end_buf; i++) s->bpp += tget(&buf, type, s->le);
+ for (i = 0; i < count && buf < end_buf; i++)
+ s->bpp += tget(&buf, type, s->le);
break;
default:
s->bpp = -1;
@@ -437,7 +462,7 @@ static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *
case TIFF_COMPR:
s->compr = value;
s->predictor = 0;
- switch(s->compr){
+ switch (s->compr) {
case TIFF_RAW:
case TIFF_PACKBITS:
case TIFF_LZW:
@@ -457,48 +482,54 @@ static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *
#endif
case TIFF_JPEG:
case TIFF_NEWJPEG:
- av_log(s->avctx, AV_LOG_ERROR, "JPEG compression is not supported\n");
+ av_log(s->avctx, AV_LOG_ERROR,
+ "JPEG compression is not supported\n");
return -1;
default:
- av_log(s->avctx, AV_LOG_ERROR, "Unknown compression method %i\n", s->compr);
+ av_log(s->avctx, AV_LOG_ERROR, "Unknown compression method %i\n",
+ s->compr);
return -1;
}
break;
case TIFF_ROWSPERSTRIP:
if (type == TIFF_LONG && value == UINT_MAX)
value = s->avctx->height;
- if(value < 1){
- av_log(s->avctx, AV_LOG_ERROR, "Incorrect value of rows per strip\n");
+ if (value < 1) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Incorrect value of rows per strip\n");
return -1;
}
s->rps = value;
break;
case TIFF_STRIP_OFFS:
- if(count == 1){
+ if (count == 1) {
s->stripdata = NULL;
s->stripoff = value;
- }else
+ } else
s->stripdata = start + off;
s->strips = count;
- if(s->strips == 1) s->rps = s->height;
+ if (s->strips == 1)
+ s->rps = s->height;
s->sot = type;
- if(s->stripdata > end_buf){
- av_log(s->avctx, AV_LOG_ERROR, "Tag referencing position outside the image\n");
+ if (s->stripdata > end_buf) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Tag referencing position outside the image\n");
return -1;
}
break;
case TIFF_STRIP_SIZE:
- if(count == 1){
+ if (count == 1) {
s->stripsizes = NULL;
s->stripsize = value;
s->strips = 1;
- }else{
+ } else {
s->stripsizes = start + off;
}
s->strips = count;
s->sstype = type;
- if(s->stripsizes > end_buf){
- av_log(s->avctx, AV_LOG_ERROR, "Tag referencing position outside the image\n");
+ if (s->stripsizes > end_buf) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Tag referencing position outside the image\n");
return -1;
}
break;
@@ -513,7 +544,7 @@ static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *
s->predictor = value;
break;
case TIFF_INVERT:
- switch(value){
+ switch (value) {
case 0:
s->invert = 1;
break;
@@ -524,13 +555,15 @@ static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *
case 3:
break;
default:
- av_log(s->avctx, AV_LOG_ERROR, "Color mode %d is not supported\n", value);
+ av_log(s->avctx, AV_LOG_ERROR, "Color mode %d is not supported\n",
+ value);
return -1;
}
break;
case TIFF_FILL_ORDER:
- if(value < 1 || value > 2){
- av_log(s->avctx, AV_LOG_ERROR, "Unknown FillOrder value %d, trying default one\n", value);
+ if (value < 1 || value > 2) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Unknown FillOrder value %d, trying default one\n", value);
value = 1;
}
s->fill_order = value - 1;
@@ -544,44 +577,44 @@ static int tiff_decode_tag(TiffContext *s, const uint8_t *start, const uint8_t *
gp = buf + count / 3 * off;
bp = buf + count / 3 * off * 2;
off = (type_sizes[type] - 1) << 3;
- for(i = 0; i < count / 3; i++){
+ for (i = 0; i < count / 3; i++) {
j = 0xff << 24;
j |= (tget(&rp, type, s->le) >> off) << 16;
j |= (tget(&gp, type, s->le) >> off) << 8;
- j |= tget(&bp, type, s->le) >> off;
+ j |= tget(&bp, type, s->le) >> off;
pal[i] = j;
}
s->palette_is_set = 1;
break;
case TIFF_PLANAR:
- if(value == 2){
+ if (value == 2) {
av_log(s->avctx, AV_LOG_ERROR, "Planar format is not supported\n");
return -1;
}
break;
case TIFF_T4OPTIONS:
- if(s->compr == TIFF_G3)
+ if (s->compr == TIFF_G3)
s->fax_opts = value;
break;
case TIFF_T6OPTIONS:
- if(s->compr == TIFF_G4)
+ if (s->compr == TIFF_G4)
s->fax_opts = value;
break;
default:
- av_log(s->avctx, AV_LOG_DEBUG, "Unknown or unsupported tag %d/0X%0X\n", tag, tag);
+ av_log(s->avctx, AV_LOG_DEBUG, "Unknown or unsupported tag %d/0X%0X\n",
+ tag, tag);
}
return 0;
}
static int decode_frame(AVCodecContext *avctx,
- void *data, int *data_size,
- AVPacket *avpkt)
+ void *data, int *data_size, AVPacket *avpkt)
{
const uint8_t *buf = avpkt->data;
int buf_size = avpkt->size;
- TiffContext * const s = avctx->priv_data;
+ TiffContext *const s = avctx->priv_data;
AVFrame *picture = data;
- AVFrame * const p = &s->picture;
+ AVFrame *const p = &s->picture;
const uint8_t *orig_buf = buf, *end_buf = buf + buf_size;
unsigned off;
int id, le, ret;
@@ -593,10 +626,13 @@ static int decode_frame(AVCodecContext *avctx,
//parse image header
if (end_buf - buf < 8)
return AVERROR_INVALIDDATA;
- id = AV_RL16(buf); buf += 2;
- if(id == 0x4949) le = 1;
- else if(id == 0x4D4D) le = 0;
- else{
+ id = AV_RL16(buf);
+ buf += 2;
+ if (id == 0x4949)
+ le = 1;
+ else if (id == 0x4D4D)
+ le = 0;
+ else {
av_log(avctx, AV_LOG_ERROR, "TIFF header not found\n");
return -1;
}
@@ -606,8 +642,9 @@ static int decode_frame(AVCodecContext *avctx,
s->fill_order = 0;
// As TIFF 6.0 specification puts it "An arbitrary but carefully chosen number
// that further identifies the file as a TIFF file"
- if(tget_short(&buf, le) != 42){
- av_log(avctx, AV_LOG_ERROR, "The answer to life, universe and everything is not correct!\n");
+ if (tget_short(&buf, le) != 42) {
+ av_log(avctx, AV_LOG_ERROR,
+ "The answer to life, universe and everything is not correct!\n");
return -1;
}
// Reset these pointers so we can tell if they were set this frame
@@ -620,12 +657,12 @@ static int decode_frame(AVCodecContext *avctx,
}
buf = orig_buf + off;
entries = tget_short(&buf, le);
- for(i = 0; i < entries; i++){
- if(tiff_decode_tag(s, orig_buf, buf, end_buf) < 0)
+ for (i = 0; i < entries; i++) {
+ if (tiff_decode_tag(s, orig_buf, buf, end_buf) < 0)
return -1;
buf += 12;
}
- if(!s->stripdata && !s->stripoff){
+ if (!s->stripdata && !s->stripoff) {
av_log(avctx, AV_LOG_ERROR, "Image data is missing\n");
return -1;
}
@@ -633,36 +670,37 @@ static int decode_frame(AVCodecContext *avctx,
if ((ret = init_image(s)) < 0)
return ret;
- if(s->strips == 1 && !s->stripsize){
+ if (s->strips == 1 && !s->stripsize) {
av_log(avctx, AV_LOG_WARNING, "Image data size missing\n");
s->stripsize = buf_size - s->stripoff;
}
stride = p->linesize[0];
dst = p->data[0];
- for(i = 0; i < s->height; i += s->rps){
- if(s->stripsizes) {
+ for (i = 0; i < s->height; i += s->rps) {
+ if (s->stripsizes) {
if (s->stripsizes >= end_buf)
return AVERROR_INVALIDDATA;
ssize = tget(&s->stripsizes, s->sstype, s->le);
} else
ssize = s->stripsize;
- if(s->stripdata){
+ if (s->stripdata) {
if (s->stripdata >= end_buf)
return AVERROR_INVALIDDATA;
soff = tget(&s->stripdata, s->sot, s->le);
- }else
+ } else
soff = s->stripoff;
if (soff > buf_size || ssize > buf_size - soff) {
av_log(avctx, AV_LOG_ERROR, "Invalid strip size/offset\n");
return -1;
}
- if(tiff_unpack_strip(s, dst, stride, orig_buf + soff, ssize, FFMIN(s->rps, s->height - i)) < 0)
+ if (tiff_unpack_strip(s, dst, stride, orig_buf + soff, ssize,
+ FFMIN(s->rps, s->height - i)) < 0)
break;
dst += s->rps * stride;
}
- if(s->predictor == 2){
+ if (s->predictor == 2) {
dst = p->data[0];
soff = s->bpp >> 3;
ssize = s->width * soff;
@@ -681,18 +719,18 @@ static int decode_frame(AVCodecContext *avctx,
dst += stride;
}
} else {
- for(i = 0; i < s->height; i++) {
- for(j = soff; j < ssize; j++)
+ for (i = 0; i < s->height; i++) {
+ for (j = soff; j < ssize; j++)
dst[j] += dst[j - soff];
dst += stride;
}
}
}
- if(s->invert){
+ if (s->invert) {
dst = s->picture.data[0];
- for(i = 0; i < s->height; i++){
- for(j = 0; j < s->picture.linesize[0]; j++)
+ for (i = 0; i < s->height; i++) {
+ for (j = 0; j < s->picture.linesize[0]; j++)
dst[j] = (s->avctx->pix_fmt == PIX_FMT_PAL8 ? (1<<s->bpp) - 1 : 255) - dst[j];
dst += s->picture.linesize[0];
}
@@ -703,7 +741,8 @@ static int decode_frame(AVCodecContext *avctx,
return buf_size;
}
-static av_cold int tiff_init(AVCodecContext *avctx){
+static av_cold int tiff_init(AVCodecContext *avctx)
+{
TiffContext *s = avctx->priv_data;
s->width = 0;
@@ -719,10 +758,10 @@ static av_cold int tiff_init(AVCodecContext *avctx){
static av_cold int tiff_end(AVCodecContext *avctx)
{
- TiffContext * const s = avctx->priv_data;
+ TiffContext *const s = avctx->priv_data;
ff_lzw_decode_close(&s->lzw);
- if(s->picture.data[0])
+ if (s->picture.data[0])
avctx->release_buffer(avctx, &s->picture);
return 0;
}
diff --git a/libavcodec/wmalosslessdec.c b/libavcodec/wmalosslessdec.c
index 50c4725045..021ce496ff 100644
--- a/libavcodec/wmalosslessdec.c
+++ b/libavcodec/wmalosslessdec.c
@@ -229,7 +229,6 @@ static av_cold int decode_init(AVCodecContext *avctx)
s->max_subframe_len_bit = 0;
s->subframe_len_bits = av_log2(log2_max_num_subframes) + 1;
- num_possible_block_sizes = log2_max_num_subframes + 1;
s->min_samples_per_subframe = s->samples_per_frame / s->max_num_subframes;
s->dynamic_range_compression = s->decode_flags & 0x80;
s->bV3RTM = s->decode_flags & 0x100;
@@ -940,9 +939,10 @@ static int decode_subframe(WmallDecodeCtx *s)
if (rawpcm_tile) {
int bits = s->bits_per_sample - padding_zeroes;
- if (bits <= 0 ) {
- av_log(s->avctx, AV_LOG_ERROR, "rawpcm_tile bits invalid\n");
- return -1;
+ if (bits <= 0) {
+ av_log(s->avctx, AV_LOG_ERROR,
+ "Invalid number of padding bits in raw PCM tile\n");
+ return AVERROR_INVALIDDATA;
}
av_dlog(s->avctx, "RAWPCM %d bits per sample. "
"total %d bits, remain=%d\n", bits,
@@ -1169,8 +1169,7 @@ static int decode_packet(AVCodecContext *avctx, void *data, int *got_frame_ptr,
GetBitContext* gb = &s->pgb;
const uint8_t* buf = avpkt->data;
int buf_size = avpkt->size;
- int num_bits_prev_frame, packet_sequence_number,
- seekable_frame_in_packet, spliced_packet;
+ int num_bits_prev_frame, packet_sequence_number, spliced_packet;
if (s->packet_done || s->packet_loss) {
s->packet_done = 0;
@@ -1185,9 +1184,11 @@ static int decode_packet(AVCodecContext *avctx, void *data, int *got_frame_ptr,
/* parse packet header */
init_get_bits(gb, buf, s->buf_bit_size);
- packet_sequence_number = get_bits(gb, 4);
- seekable_frame_in_packet = get_bits1(gb);
- spliced_packet = get_bits1(gb);
+ packet_sequence_number = get_bits(gb, 4);
+ skip_bits(gb, 1); // Skip seekable_frame_in_packet, currently ununused
+ spliced_packet = get_bits1(gb);
+ if (spliced_packet)
+ av_log_missing_feature(avctx, "Bitstream splicing", 1);
/* get number of bits that need to be added to the previous frame */
num_bits_prev_frame = get_bits(gb, s->log2_frame_size);
diff --git a/libavcodec/x86/dsputil_yasm.asm b/libavcodec/x86/dsputil_yasm.asm
index 568ad29769..769582c060 100644
--- a/libavcodec/x86/dsputil_yasm.asm
+++ b/libavcodec/x86/dsputil_yasm.asm
@@ -594,7 +594,7 @@ cglobal emu_edge_core, 2, 7, 0
%define valw2 r7w
%define valw3 r3w
%if WIN64
-%define valw4 r4w
+%define valw4 r7w
%else ; unix64
%define valw4 r3w
%endif
diff --git a/libavfilter/avfiltergraph.h b/libavfilter/avfiltergraph.h
index cca9488f87..2c612b481d 100644
--- a/libavfilter/avfiltergraph.h
+++ b/libavfilter/avfiltergraph.h
@@ -91,11 +91,11 @@ void avfilter_graph_free(AVFilterGraph **graph);
/**
* A linked-list of the inputs/outputs of the filter chain.
*
- * This is mainly useful for avfilter_graph_parse(), since this
- * function may accept a description of a graph with not connected
- * input/output pads. This struct specifies, per each not connected
- * pad contained in the graph, the filter context and the pad index
- * required for establishing a link.
+ * This is mainly useful for avfilter_graph_parse() / avfilter_graph_parse2(),
+ * where it is used to communicate open (unlinked) inputs and outputs from and
+ * to the caller.
+ * This struct specifies, per each not connected pad contained in the graph, the
+ * filter context and the pad index required for establishing a link.
*/
typedef struct AVFilterInOut {
/** unique name for this input/output in the list */
@@ -112,13 +112,14 @@ typedef struct AVFilterInOut {
} AVFilterInOut;
/**
- * Create an AVFilterInOut.
- * Must be free with avfilter_inout_free().
+ * Allocate a single AVFilterInOut entry.
+ * Must be freed with avfilter_inout_free().
+ * @return allocated AVFilterInOut on success, NULL on failure.
*/
AVFilterInOut *avfilter_inout_alloc(void);
/**
- * Free the AVFilterInOut in *inout, and set its pointer to NULL.
+ * Free the supplied list of AVFilterInOut and set *inout to NULL.
* If *inout is NULL, do nothing.
*/
void avfilter_inout_free(AVFilterInOut **inout);
@@ -141,6 +142,41 @@ int avfilter_graph_parse(AVFilterGraph *graph, const char *filters,
void *log_ctx);
/**
+ * Add a graph described by a string to a graph.
+ *
+ * @param[in] graph the filter graph where to link the parsed graph context
+ * @param[in] filters string to be parsed
+ * @param[out] inputs a linked list of all free (unlinked) inputs of the
+ * parsed graph will be returned here. It is to be freed
+ * by the caller using avfilter_inout_free().
+ * @param[out] outputs a linked list of all free (unlinked) outputs of the
+ * parsed graph will be returned here. It is to be freed by the
+ * caller using avfilter_inout_free().
+ * @return zero on success, a negative AVERROR code on error
+ *
+ * @note the difference between avfilter_graph_parse2() and
+ * avfilter_graph_parse() is that in avfilter_graph_parse(), the caller provides
+ * the lists of inputs and outputs, which therefore must be known before calling
+ * the function. On the other hand, avfilter_graph_parse2() \em returns the
+ * inputs and outputs that are left unlinked after parsing the graph and the
+ * caller then deals with them. Another difference is that in
+ * avfilter_graph_parse(), the inputs parameter describes inputs of the
+ * <em>already existing</em> part of the graph; i.e. from the point of view of
+ * the newly created part, they are outputs. Similarly the outputs parameter
+ * describes outputs of the already existing filters, which are provided as
+ * inputs to the parsed filters.
+ * avfilter_graph_parse2() takes the opposite approach -- it makes no reference
+ * whatsoever to already existing parts of the graph and the inputs parameter
+ * will on return contain inputs of the newly parsed part of the graph.
+ * Analogously the outputs parameter will contain outputs of the newly created
+ * filters.
+ */
+int avfilter_graph_parse2(AVFilterGraph *graph, const char *filters,
+ AVFilterInOut **inputs,
+ AVFilterInOut **outputs);
+
+
+/**
* Send a command to one or more filter instances.
*
* @param graph the filter graph
diff --git a/libavfilter/buffersrc.h b/libavfilter/buffersrc.h
index bd82c065e4..918a54faee 100644
--- a/libavfilter/buffersrc.h
+++ b/libavfilter/buffersrc.h
@@ -32,6 +32,7 @@
*
* @param buf buffer containing frame data to be passed down the filtergraph.
* This function will take ownership of buf, the user must not free it.
+ * A NULL buf signals EOF -- i.e. no more frames will be sent to this filter.
*/
int av_buffersrc_buffer(AVFilterContext *s, AVFilterBufferRef *buf);
diff --git a/libavfilter/graphparser.c b/libavfilter/graphparser.c
index bd3022e035..af4b566f1b 100644
--- a/libavfilter/graphparser.c
+++ b/libavfilter/graphparser.c
@@ -189,13 +189,15 @@ static AVFilterInOut *extract_inout(const char *label, AVFilterInOut **links)
{
AVFilterInOut *ret;
- while (*links && strcmp((*links)->name, label))
+ while (*links && (!(*links)->name || strcmp((*links)->name, label)))
links = &((*links)->next);
ret = *links;
- if (ret)
+ if (ret) {
*links = ret->next;
+ ret->next = NULL;
+ }
return ret;
}
@@ -206,22 +208,31 @@ static void insert_inout(AVFilterInOut **inouts, AVFilterInOut *element)
*inouts = element;
}
+static void append_inout(AVFilterInOut **inouts, AVFilterInOut **element)
+{
+ while (*inouts && (*inouts)->next)
+ inouts = &((*inouts)->next);
+
+ if (!*inouts)
+ *inouts = *element;
+ else
+ (*inouts)->next = *element;
+ *element = NULL;
+}
+
static int link_filter_inouts(AVFilterContext *filt_ctx,
AVFilterInOut **curr_inputs,
AVFilterInOut **open_inputs, void *log_ctx)
{
- int pad = filt_ctx->input_count, ret;
+ int pad, ret;
- while (pad--) {
+ for (pad = 0; pad < filt_ctx->input_count; pad++) {
AVFilterInOut *p = *curr_inputs;
- if (!p) {
- av_log(log_ctx, AV_LOG_ERROR,
- "Not enough inputs specified for the \"%s\" filter.\n",
- filt_ctx->filter->name);
- return AVERROR(EINVAL);
- }
- *curr_inputs = (*curr_inputs)->next;
+ if (p)
+ *curr_inputs = (*curr_inputs)->next;
+ else if (!(p = av_mallocz(sizeof(*p))))
+ return AVERROR(ENOMEM);
if (p->filter_ctx) {
if ((ret = link_filter(p->filter_ctx, p->pad_idx, filt_ctx, pad, log_ctx)) < 0)
@@ -258,6 +269,7 @@ static int link_filter_inouts(AVFilterContext *filt_ctx,
static int parse_inputs(const char **buf, AVFilterInOut **curr_inputs,
AVFilterInOut **open_outputs, void *log_ctx)
{
+ AVFilterInOut *parsed_inputs = NULL;
int pad = 0;
while (**buf == '[') {
@@ -280,12 +292,15 @@ static int parse_inputs(const char **buf, AVFilterInOut **curr_inputs,
match->pad_idx = pad;
}
- insert_inout(curr_inputs, match);
+ append_inout(&parsed_inputs, &match);
*buf += strspn(*buf, WHITESPACES);
pad++;
}
+ append_inout(&parsed_inputs, curr_inputs);
+ *curr_inputs = parsed_inputs;
+
return pad;
}
@@ -334,10 +349,173 @@ static int parse_outputs(const char **buf, AVFilterInOut **curr_inputs,
return pad;
}
+#if FF_API_GRAPH_AVCLASS
+#define log_ctx graph
+#else
+#define log_ctx NULL
+#endif
+
+static int parse_sws_flags(const char **buf, AVFilterGraph *graph)
+{
+ char *p = strchr(*buf, ';');
+
+ if (strncmp(*buf, "sws_flags=", 10))
+ return 0;
+
+ if (!p) {
+ av_log(log_ctx, AV_LOG_ERROR, "sws_flags not terminated with ';'.\n");
+ return AVERROR(EINVAL);
+ }
+
+ *buf += 4; // keep the 'flags=' part
+
+ av_freep(&graph->scale_sws_opts);
+ if (!(graph->scale_sws_opts = av_mallocz(p - *buf + 1)))
+ return AVERROR(ENOMEM);
+ av_strlcpy(graph->scale_sws_opts, *buf, p - *buf + 1);
+
+ *buf = p + 1;
+ return 0;
+}
+
+int avfilter_graph_parse2(AVFilterGraph *graph, const char *filters,
+ AVFilterInOut **inputs,
+ AVFilterInOut **outputs)
+{
+ int index = 0, ret = 0;
+ char chr = 0;
+
+ AVFilterInOut *curr_inputs = NULL, *open_inputs = NULL, *open_outputs = NULL;
+
+ filters += strspn(filters, WHITESPACES);
+
+ if ((ret = parse_sws_flags(&filters, graph)) < 0)
+ goto fail;
+
+ do {
+ AVFilterContext *filter;
+ filters += strspn(filters, WHITESPACES);
+
+ if ((ret = parse_inputs(&filters, &curr_inputs, &open_outputs, log_ctx)) < 0)
+ goto end;
+
+ if ((ret = parse_filter(&filter, &filters, graph, index, log_ctx)) < 0)
+ goto end;
+
+ if ((ret = link_filter_inouts(filter, &curr_inputs, &open_inputs, log_ctx)) < 0)
+ goto end;
+
+ if ((ret = parse_outputs(&filters, &curr_inputs, &open_inputs, &open_outputs,
+ log_ctx)) < 0)
+ goto end;
+
+ filters += strspn(filters, WHITESPACES);
+ chr = *filters++;
+
+ if (chr == ';' && curr_inputs)
+ append_inout(&open_outputs, &curr_inputs);
+ index++;
+ } while (chr == ',' || chr == ';');
+
+ if (chr) {
+ av_log(log_ctx, AV_LOG_ERROR,
+ "Unable to parse graph description substring: \"%s\"\n",
+ filters - 1);
+ ret = AVERROR(EINVAL);
+ goto end;
+ }
+
+ append_inout(&open_outputs, &curr_inputs);
+
+
+ *inputs = open_inputs;
+ *outputs = open_outputs;
+ return 0;
+
+ fail:end:
+ for (; graph->filter_count > 0; graph->filter_count--)
+ avfilter_free(graph->filters[graph->filter_count - 1]);
+ av_freep(&graph->filters);
+ avfilter_inout_free(&open_inputs);
+ avfilter_inout_free(&open_outputs);
+ avfilter_inout_free(&curr_inputs);
+
+ *inputs = NULL;
+ *outputs = NULL;
+
+ return ret;
+}
+#undef log_ctx
+
int avfilter_graph_parse(AVFilterGraph *graph, const char *filters,
AVFilterInOut **open_inputs_ptr, AVFilterInOut **open_outputs_ptr,
void *log_ctx)
{
+#if 0
+ int ret;
+ AVFilterInOut *open_inputs = open_inputs_ptr ? *open_inputs_ptr : NULL;
+ AVFilterInOut *open_outputs = open_outputs_ptr ? *open_outputs_ptr : NULL;
+ AVFilterInOut *cur, *match, *inputs = NULL, *outputs = NULL;
+
+ if ((ret = avfilter_graph_parse2(graph, filters, &inputs, &outputs)) < 0)
+ goto fail;
+
+ /* First input can be omitted if it is "[in]" */
+ if (inputs && !inputs->name)
+ inputs->name = av_strdup("in");
+ for (cur = inputs; cur; cur = cur->next) {
+ if (!cur->name) {
+ av_log(log_ctx, AV_LOG_ERROR,
+ "Not enough inputs specified for the \"%s\" filter.\n",
+ cur->filter_ctx->filter->name);
+ ret = AVERROR(EINVAL);
+ goto fail;
+ }
+ if (!(match = extract_inout(cur->name, &open_outputs)))
+ continue;
+ ret = avfilter_link(match->filter_ctx, match->pad_idx,
+ cur->filter_ctx, cur->pad_idx);
+ avfilter_inout_free(&match);
+ if (ret < 0)
+ goto fail;
+ }
+
+ /* Last output can be omitted if it is "[out]" */
+ if (outputs && !outputs->name)
+ outputs->name = av_strdup("out");
+ for (cur = outputs; cur; cur = cur->next) {
+ if (!cur->name) {
+ av_log(log_ctx, AV_LOG_ERROR,
+ "Invalid filterchain containing an unlabelled output pad: \"%s\"\n",
+ filters);
+ ret = AVERROR(EINVAL);
+ goto fail;
+ }
+ if (!(match = extract_inout(cur->name, &open_inputs)))
+ continue;
+ ret = avfilter_link(cur->filter_ctx, cur->pad_idx,
+ match->filter_ctx, match->pad_idx);
+ avfilter_inout_free(&match);
+ if (ret < 0)
+ goto fail;
+ }
+
+ fail:
+ if (ret < 0) {
+ for (; graph->filter_count > 0; graph->filter_count--)
+ avfilter_free(graph->filters[graph->filter_count - 1]);
+ av_freep(&graph->filters);
+ }
+ avfilter_inout_free(&inputs);
+ avfilter_inout_free(&outputs);
+ /* clear open_in/outputs only if not passed as parameters */
+ if (open_inputs_ptr) *open_inputs_ptr = open_inputs;
+ else avfilter_inout_free(&open_inputs);
+ if (open_outputs_ptr) *open_outputs_ptr = open_outputs;
+ else avfilter_inout_free(&open_outputs);
+ return ret;
+}
+#else
int index = 0, ret = 0;
char chr = 0;
@@ -414,3 +592,5 @@ end:
}
return ret;
}
+
+#endif
diff --git a/libavfilter/version.h b/libavfilter/version.h
index 7df43a3915..e8e9f27566 100644
--- a/libavfilter/version.h
+++ b/libavfilter/version.h
@@ -29,8 +29,8 @@
#include "libavutil/avutil.h"
#define LIBAVFILTER_VERSION_MAJOR 2
-#define LIBAVFILTER_VERSION_MINOR 69
-#define LIBAVFILTER_VERSION_MICRO 101
+#define LIBAVFILTER_VERSION_MINOR 70
+#define LIBAVFILTER_VERSION_MICRO 100
#define LIBAVFILTER_VERSION_INT AV_VERSION_INT(LIBAVFILTER_VERSION_MAJOR, \
LIBAVFILTER_VERSION_MINOR, \
diff --git a/libavfilter/vsrc_buffer.c b/libavfilter/vsrc_buffer.c
index 73866a8c4a..d003add88b 100644
--- a/libavfilter/vsrc_buffer.c
+++ b/libavfilter/vsrc_buffer.c
@@ -39,6 +39,7 @@ typedef struct {
AVRational time_base; ///< time_base to set in the output link
AVRational sample_aspect_ratio;
char sws_param[256];
+ int eof;
} BufferSourceContext;
#define CHECK_PARAM_CHANGE(s, c, width, height, format)\
@@ -55,6 +56,12 @@ int av_vsrc_buffer_add_video_buffer_ref(AVFilterContext *buffer_filter,
AVFilterBufferRef *buf;
int ret;
+ if (!picref) {
+ c->eof = 1;
+ return 0;
+ } else if (c->eof)
+ return AVERROR(EINVAL);
+
if (!av_fifo_space(c->fifo) &&
(ret = av_fifo_realloc2(c->fifo, av_fifo_size(c->fifo) +
sizeof(buf))) < 0)
@@ -125,6 +132,12 @@ int av_buffersrc_buffer(AVFilterContext *s, AVFilterBufferRef *buf)
BufferSourceContext *c = s->priv;
int ret;
+ if (!buf) {
+ c->eof = 1;
+ return 0;
+ } else if (c->eof)
+ return AVERROR(EINVAL);
+
if (!av_fifo_space(c->fifo) &&
(ret = av_fifo_realloc2(c->fifo, av_fifo_size(c->fifo) +
sizeof(buf))) < 0)
@@ -144,9 +157,17 @@ int av_buffersrc_buffer(AVFilterContext *s, AVFilterBufferRef *buf)
int av_vsrc_buffer_add_frame(AVFilterContext *buffer_src,
const AVFrame *frame, int flags)
{
+ BufferSourceContext *c = buffer_src->priv;
+ AVFilterBufferRef *picref;
int ret;
- AVFilterBufferRef *picref =
- avfilter_get_video_buffer_ref_from_frame(frame, AV_PERM_WRITE);
+
+ if (!frame) {
+ c->eof = 1;
+ return 0;
+ } else if (c->eof)
+ return AVERROR(EINVAL);
+
+ picref = avfilter_get_video_buffer_ref_from_frame(frame, AV_PERM_WRITE);
if (!picref)
return AVERROR(ENOMEM);
ret = av_vsrc_buffer_add_video_buffer_ref(buffer_src, picref, flags);
@@ -226,6 +247,8 @@ static int request_frame(AVFilterLink *link)
AVFilterBufferRef *buf;
if (!av_fifo_size(c->fifo)) {
+ if (c->eof)
+ return AVERROR_EOF;
av_log(link->src, AV_LOG_WARNING,
"request_frame() called with no available frame!\n");
return AVERROR(EINVAL);
@@ -243,7 +266,10 @@ static int request_frame(AVFilterLink *link)
static int poll_frame(AVFilterLink *link)
{
BufferSourceContext *c = link->src->priv;
- return !!av_fifo_size(c->fifo);
+ int size = av_fifo_size(c->fifo);
+ if (!size && c->eof)
+ return AVERROR_EOF;
+ return size/sizeof(AVFilterBufferRef*);
}
AVFilter avfilter_vsrc_buffer = {
diff --git a/libavformat/isom.h b/libavformat/isom.h
index f5d03b20eb..542b76bf30 100644
--- a/libavformat/isom.h
+++ b/libavformat/isom.h
@@ -107,6 +107,7 @@ typedef struct MOVStreamContext {
unsigned int alt_sample_size; ///< always contains sample size from stsz atom
unsigned int sample_count;
int *sample_sizes;
+ int keyframe_absent;
unsigned int keyframe_count;
int *keyframes;
int time_scale;
diff --git a/libavformat/mov.c b/libavformat/mov.c
index 234c1f5ce1..f7ae19ab9f 100644
--- a/libavformat/mov.c
+++ b/libavformat/mov.c
@@ -1618,7 +1618,10 @@ static int mov_read_stss(MOVContext *c, AVIOContext *pb, MOVAtom atom)
av_dlog(c->fc, "keyframe_count = %d\n", entries);
if (!entries)
+ {
+ sc->keyframe_absent = 1;
return 0;
+ }
if (entries >= UINT_MAX / sizeof(int))
return AVERROR_INVALIDDATA;
sc->keyframes = av_malloc(entries * sizeof(int));
@@ -1873,7 +1876,7 @@ static void mov_build_index(MOVContext *mov, AVStream *st)
return;
}
- if (!sc->keyframe_count || current_sample+key_off == sc->keyframes[stss_index]) {
+ if (!sc->keyframe_absent && (!sc->keyframe_count || current_sample+key_off == sc->keyframes[stss_index])) {
keyframe = 1;
if (stss_index + 1 < sc->keyframe_count)
stss_index++;
diff --git a/libavformat/utils.c b/libavformat/utils.c
index 875d0d84a8..3c3ced75a0 100644
--- a/libavformat/utils.c
+++ b/libavformat/utils.c
@@ -1975,6 +1975,8 @@ static int has_duration(AVFormatContext *ic)
if (st->duration != AV_NOPTS_VALUE)
return 1;
}
+ if (ic->duration)
+ return 1;
return 0;
}
diff --git a/libswscale/output.c b/libswscale/output.c
index 6b17c922ba..0e5b3f8802 100644
--- a/libswscale/output.c
+++ b/libswscale/output.c
@@ -469,7 +469,7 @@ yuv2422_X_c_template(SwsContext *c, const int16_t *lumFilter,
{
int i;
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int j;
int Y1 = 1 << 18;
int Y2 = 1 << 18;
@@ -512,7 +512,7 @@ yuv2422_2_c_template(SwsContext *c, const int16_t *buf[2],
int uvalpha1 = 4095 - uvalpha;
int i;
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int Y1 = (buf0[i * 2] * yalpha1 + buf1[i * 2] * yalpha) >> 19;
int Y2 = (buf0[i * 2 + 1] * yalpha1 + buf1[i * 2 + 1] * yalpha) >> 19;
int U = (ubuf0[i] * uvalpha1 + ubuf1[i] * uvalpha) >> 19;
@@ -539,7 +539,7 @@ yuv2422_1_c_template(SwsContext *c, const int16_t *buf0,
int i;
if (uvalpha < 2048) {
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int Y1 = (buf0[i * 2 ]+64) >> 7;
int Y2 = (buf0[i * 2 + 1]+64) >> 7;
int U = (ubuf0[i] +64) >> 7;
@@ -561,7 +561,7 @@ yuv2422_1_c_template(SwsContext *c, const int16_t *buf0,
}
} else {
const int16_t *ubuf1 = ubuf[1], *vbuf1 = vbuf[1];
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int Y1 = (buf0[i * 2 ] + 64) >> 7;
int Y2 = (buf0[i * 2 + 1] + 64) >> 7;
int U = (ubuf0[i] + ubuf1[i]+128) >> 8;
@@ -608,7 +608,7 @@ yuv2rgb48_X_c_template(SwsContext *c, const int16_t *lumFilter,
{
int i;
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int j;
int Y1 = -0x40000000;
int Y2 = -0x40000000;
@@ -671,7 +671,7 @@ yuv2rgb48_2_c_template(SwsContext *c, const int32_t *buf[2],
int uvalpha1 = 4095 - uvalpha;
int i;
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int Y1 = (buf0[i * 2] * yalpha1 + buf1[i * 2] * yalpha) >> 14;
int Y2 = (buf0[i * 2 + 1] * yalpha1 + buf1[i * 2 + 1] * yalpha) >> 14;
int U = (ubuf0[i] * uvalpha1 + ubuf1[i] * uvalpha + (-128 << 23)) >> 14;
@@ -709,7 +709,7 @@ yuv2rgb48_1_c_template(SwsContext *c, const int32_t *buf0,
int i;
if (uvalpha < 2048) {
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int Y1 = (buf0[i * 2] ) >> 2;
int Y2 = (buf0[i * 2 + 1]) >> 2;
int U = (ubuf0[i] + (-128 << 11)) >> 2;
@@ -737,7 +737,7 @@ yuv2rgb48_1_c_template(SwsContext *c, const int32_t *buf0,
}
} else {
const int32_t *ubuf1 = ubuf[1], *vbuf1 = vbuf[1];
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int Y1 = (buf0[i * 2] ) >> 2;
int Y2 = (buf0[i * 2 + 1]) >> 2;
int U = (ubuf0[i] + ubuf1[i] + (-128 << 12)) >> 3;
@@ -952,7 +952,7 @@ yuv2rgb_X_c_template(SwsContext *c, const int16_t *lumFilter,
{
int i;
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int j, A1, A2;
int Y1 = 1 << 18;
int Y2 = 1 << 18;
@@ -1012,7 +1012,7 @@ yuv2rgb_2_c_template(SwsContext *c, const int16_t *buf[2],
int uvalpha1 = 4095 - uvalpha;
int i;
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int Y1 = (buf0[i * 2] * yalpha1 + buf1[i * 2] * yalpha) >> 19;
int Y2 = (buf0[i * 2 + 1] * yalpha1 + buf1[i * 2 + 1] * yalpha) >> 19;
int U = (ubuf0[i] * uvalpha1 + ubuf1[i] * uvalpha) >> 19;
@@ -1050,7 +1050,7 @@ yuv2rgb_1_c_template(SwsContext *c, const int16_t *buf0,
int i;
if (uvalpha < 2048) {
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int Y1 = (buf0[i * 2 ] + 64) >> 7;
int Y2 = (buf0[i * 2 + 1] + 64) >> 7;
int U = (ubuf0[i] + 64) >> 7;
@@ -1077,7 +1077,7 @@ yuv2rgb_1_c_template(SwsContext *c, const int16_t *buf0,
}
} else {
const int16_t *ubuf1 = ubuf[1], *vbuf1 = vbuf[1];
- for (i = 0; i < (dstW >> 1); i++) {
+ for (i = 0; i < ((dstW + 1) >> 1); i++) {
int Y1 = (buf0[i * 2 ] + 64) >> 7;
int Y2 = (buf0[i * 2 + 1] + 64) >> 7;
int U = (ubuf0[i] + ubuf1[i] + 128) >> 8;
diff --git a/tests/lavfi-regression.sh b/tests/lavfi-regression.sh
index 4857839fda..4d669f2d3c 100755
--- a/tests/lavfi-regression.sh
+++ b/tests/lavfi-regression.sh
@@ -40,7 +40,7 @@ do_lavfi "crop_vflip" "crop=iw-100:ih-100:100:100,vflip"
do_lavfi "drawbox" "drawbox=224:24:88:72:#FF8010@0.5"
do_lavfi "fade" "fade=in:5:15,fade=out:30:15"
do_lavfi "null" "null"
-do_lavfi "overlay" "split[m],scale=88:72,pad=96:80:4:4[o2];[m]fifo,[o2]overlay=240:16"
+do_lavfi "overlay" "split[m],scale=88:72,pad=96:80:4:4[o2];[m]fifo[o1],[o1][o2]overlay=240:16"
do_lavfi "pad" "pad=iw*1.5:ih*1.5:iw*0.3:ih*0.2"
do_lavfi "pp" "mp=pp=be/de/tn/l5/al"
do_lavfi "pp2" "mp=pp=be/fq:16/fa/lb"