aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorPaul B Mahol <onemda@gmail.com>2018-04-24 13:45:13 +0200
committerPaul B Mahol <onemda@gmail.com>2018-04-26 16:44:15 +0200
commit161e006cc02e991606ac1ef246afdbc1c3293867 (patch)
tree74f1fab579801d055e4112cb365c5560dbca7882
parent28503c5aea5f5ecaab95e9d0c56064132e4796d3 (diff)
downloadffmpeg-161e006cc02e991606ac1ef246afdbc1c3293867.tar.gz
avfilter: add tmix filter
Signed-off-by: Paul B Mahol <onemda@gmail.com>
-rw-r--r--Changelog1
-rw-r--r--doc/filters.texi15
-rw-r--r--libavfilter/Makefile1
-rw-r--r--libavfilter/allfilters.c1
-rw-r--r--libavfilter/version.h2
-rw-r--r--libavfilter/vf_mix.c164
6 files changed, 144 insertions, 40 deletions
diff --git a/Changelog b/Changelog
index 3ed4874b09..fe6a60a73c 100644
--- a/Changelog
+++ b/Changelog
@@ -3,6 +3,7 @@ releases are sorted from youngest to oldest.
version <next>:
- deblock filter
+- tmix filter
version 4.0:
diff --git a/doc/filters.texi b/doc/filters.texi
index c4318c5b6f..23d6063ef4 100644
--- a/doc/filters.texi
+++ b/doc/filters.texi
@@ -15561,6 +15561,21 @@ Vertical low-pass filtering can only be enabled for @option{mode}
@end table
+@section tmix
+
+Mix successive video frames.
+
+A description of the accepted options follows.
+
+@table @option
+@item frames
+The number of successive frames to mix. If unspecified, it defaults to 3.
+
+@item weights
+Specify weight of each input video frame.
+Each weight is separated by space.
+@end table
+
@section tonemap
Tone map colors from different dynamic ranges.
diff --git a/libavfilter/Makefile b/libavfilter/Makefile
index 8f41805527..7f8317c2f2 100644
--- a/libavfilter/Makefile
+++ b/libavfilter/Makefile
@@ -345,6 +345,7 @@ OBJS-$(CONFIG_THUMBNAIL_CUDA_FILTER) += vf_thumbnail_cuda.o vf_thumbnail
OBJS-$(CONFIG_TILE_FILTER) += vf_tile.o
OBJS-$(CONFIG_TINTERLACE_FILTER) += vf_tinterlace.o
OBJS-$(CONFIG_TLUT2_FILTER) += vf_lut2.o framesync.o
+OBJS-$(CONFIG_TMIX_FILTER) += vf_mix.o framesync.o
OBJS-$(CONFIG_TONEMAP_FILTER) += vf_tonemap.o
OBJS-$(CONFIG_TRANSPOSE_FILTER) += vf_transpose.o
OBJS-$(CONFIG_TRIM_FILTER) += trim.o
diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c
index 643eec287d..35e12d595f 100644
--- a/libavfilter/allfilters.c
+++ b/libavfilter/allfilters.c
@@ -336,6 +336,7 @@ extern AVFilter ff_vf_thumbnail_cuda;
extern AVFilter ff_vf_tile;
extern AVFilter ff_vf_tinterlace;
extern AVFilter ff_vf_tlut2;
+extern AVFilter ff_vf_tmix;
extern AVFilter ff_vf_tonemap;
extern AVFilter ff_vf_transpose;
extern AVFilter ff_vf_trim;
diff --git a/libavfilter/version.h b/libavfilter/version.h
index 625e95dd4c..2461a6f2e3 100644
--- a/libavfilter/version.h
+++ b/libavfilter/version.h
@@ -30,7 +30,7 @@
#include "libavutil/version.h"
#define LIBAVFILTER_VERSION_MAJOR 7
-#define LIBAVFILTER_VERSION_MINOR 19
+#define LIBAVFILTER_VERSION_MINOR 20
#define LIBAVFILTER_VERSION_MICRO 100
#define LIBAVFILTER_VERSION_INT AV_VERSION_INT(LIBAVFILTER_VERSION_MAJOR, \
diff --git a/libavfilter/vf_mix.c b/libavfilter/vf_mix.c
index a548f2663f..55307a4bb5 100644
--- a/libavfilter/vf_mix.c
+++ b/libavfilter/vf_mix.c
@@ -39,6 +39,9 @@ typedef struct MixContext {
float *weights;
float wfactor;
+ int tmix;
+ int nb_frames;
+
int depth;
int nb_planes;
int linesize[4];
@@ -71,6 +74,8 @@ static av_cold int init(AVFilterContext *ctx)
char *p, *arg, *saveptr = NULL;
int i, ret;
+ s->tmix = !strcmp(ctx->filter->name, "tmix");
+
s->frames = av_calloc(s->nb_inputs, sizeof(*s->frames));
if (!s->frames)
return AVERROR(ENOMEM);
@@ -79,17 +84,19 @@ static av_cold int init(AVFilterContext *ctx)
if (!s->weights)
return AVERROR(ENOMEM);
- for (i = 0; i < s->nb_inputs; i++) {
- AVFilterPad pad = { 0 };
+ if (!s->tmix) {
+ for (i = 0; i < s->nb_inputs; i++) {
+ AVFilterPad pad = { 0 };
- pad.type = AVMEDIA_TYPE_VIDEO;
- pad.name = av_asprintf("input%d", i);
- if (!pad.name)
- return AVERROR(ENOMEM);
+ pad.type = AVMEDIA_TYPE_VIDEO;
+ pad.name = av_asprintf("input%d", i);
+ if (!pad.name)
+ return AVERROR(ENOMEM);
- if ((ret = ff_insert_inpad(ctx, i, &pad)) < 0) {
- av_freep(&pad.name);
- return ret;
+ if ((ret = ff_insert_inpad(ctx, i, &pad)) < 0) {
+ av_freep(&pad.name);
+ return ret;
+ }
}
}
@@ -107,24 +114,9 @@ static av_cold int init(AVFilterContext *ctx)
return 0;
}
-static int process_frame(FFFrameSync *fs)
+static void mix_frames(MixContext *s, AVFrame **in, AVFrame *out)
{
- AVFilterContext *ctx = fs->parent;
- AVFilterLink *outlink = ctx->outputs[0];
- MixContext *s = fs->opaque;
- AVFrame **in = s->frames;
- AVFrame *out;
- int i, p, ret, x, y;
-
- for (i = 0; i < s->nb_inputs; i++) {
- if ((ret = ff_framesync_get_frame(&s->fs, i, &in[i], 0)) < 0)
- return ret;
- }
-
- out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
- if (!out)
- return AVERROR(ENOMEM);
- out->pts = av_rescale_q(s->fs.pts, s->fs.time_base, outlink->time_base);
+ int i, p, x, y;
if (s->depth <= 8) {
for (p = 0; p < s->nb_planes; p++) {
@@ -167,6 +159,28 @@ static int process_frame(FFFrameSync *fs)
}
}
}
+}
+
+static int process_frame(FFFrameSync *fs)
+{
+ AVFilterContext *ctx = fs->parent;
+ AVFilterLink *outlink = ctx->outputs[0];
+ MixContext *s = fs->opaque;
+ AVFrame **in = s->frames;
+ AVFrame *out;
+ int i, ret;
+
+ for (i = 0; i < s->nb_inputs; i++) {
+ if ((ret = ff_framesync_get_frame(&s->fs, i, &in[i], 0)) < 0)
+ return ret;
+ }
+
+ out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
+ if (!out)
+ return AVERROR(ENOMEM);
+ out->pts = av_rescale_q(s->fs.pts, s->fs.time_base, outlink->time_base);
+
+ mix_frames(s, in, out);
return ff_filter_frame(outlink, out);
}
@@ -183,10 +197,12 @@ static int config_output(AVFilterLink *outlink)
FFFrameSyncIn *in;
int i, ret;
- for (i = 1; i < s->nb_inputs; i++) {
- if (ctx->inputs[i]->h != height || ctx->inputs[i]->w != width) {
- av_log(ctx, AV_LOG_ERROR, "Input %d size (%dx%d) does not match input %d size (%dx%d).\n", i, ctx->inputs[i]->w, ctx->inputs[i]->h, 0, width, height);
- return AVERROR(EINVAL);
+ if (!s->tmix) {
+ for (i = 1; i < s->nb_inputs; i++) {
+ if (ctx->inputs[i]->h != height || ctx->inputs[i]->w != width) {
+ av_log(ctx, AV_LOG_ERROR, "Input %d size (%dx%d) does not match input %d size (%dx%d).\n", i, ctx->inputs[i]->w, ctx->inputs[i]->h, 0, width, height);
+ return AVERROR(EINVAL);
+ }
}
}
@@ -196,6 +212,15 @@ static int config_output(AVFilterLink *outlink)
s->nb_planes = av_pix_fmt_count_planes(outlink->format);
s->depth = s->desc->comp[0].depth;
+ if ((ret = av_image_fill_linesizes(s->linesize, inlink->format, inlink->w)) < 0)
+ return ret;
+
+ s->height[1] = s->height[2] = AV_CEIL_RSHIFT(inlink->h, s->desc->log2_chroma_h);
+ s->height[0] = s->height[3] = inlink->h;
+
+ if (s->tmix)
+ return 0;
+
outlink->w = width;
outlink->h = height;
outlink->time_base = time_base;
@@ -208,12 +233,6 @@ static int config_output(AVFilterLink *outlink)
s->fs.opaque = s;
s->fs.on_event = process_frame;
- if ((ret = av_image_fill_linesizes(s->linesize, inlink->format, inlink->w)) < 0)
- return ret;
-
- s->height[1] = s->height[2] = AV_CEIL_RSHIFT(inlink->h, s->desc->log2_chroma_h);
- s->height[0] = s->height[3] = inlink->h;
-
for (i = 0; i < s->nb_inputs; i++) {
AVFilterLink *inlink = ctx->inputs[i];
@@ -232,11 +251,16 @@ static av_cold void uninit(AVFilterContext *ctx)
int i;
ff_framesync_uninit(&s->fs);
- av_freep(&s->frames);
av_freep(&s->weights);
- for (i = 0; i < ctx->nb_inputs; i++)
- av_freep(&ctx->input_pads[i].name);
+ if (!s->tmix) {
+ for (i = 0; i < ctx->nb_inputs; i++)
+ av_freep(&ctx->input_pads[i].name);
+ } else {
+ for (i = 0; i < s->nb_frames; i++)
+ av_frame_free(&s->frames[i]);
+ }
+ av_freep(&s->frames);
}
static int activate(AVFilterContext *ctx)
@@ -267,6 +291,7 @@ static const AVFilterPad outputs[] = {
{ NULL }
};
+#if CONFIG_MIX_FILTER
AVFILTER_DEFINE_CLASS(mix);
AVFilter ff_vf_mix = {
@@ -281,3 +306,64 @@ AVFilter ff_vf_mix = {
.activate = activate,
.flags = AVFILTER_FLAG_DYNAMIC_INPUTS,
};
+
+#endif /* CONFIG_MIX_FILTER */
+
+#if CONFIG_TMIX_FILTER
+static int tmix_filter_frame(AVFilterLink *inlink, AVFrame *in)
+{
+ AVFilterContext *ctx = inlink->dst;
+ AVFilterLink *outlink = ctx->outputs[0];
+ MixContext *s = ctx->priv;
+ AVFrame *out;
+
+ if (s->nb_frames < s->nb_inputs) {
+ s->frames[s->nb_frames] = in;
+ s->nb_frames++;
+ return 0;
+ } else {
+ av_frame_free(&s->frames[0]);
+ memmove(&s->frames[0], &s->frames[1], sizeof(*s->frames) * (s->nb_inputs - 1));
+ s->frames[s->nb_inputs - 1] = in;
+ }
+
+ out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
+ if (!out)
+ return AVERROR(ENOMEM);
+ out->pts = s->frames[0]->pts;
+
+ mix_frames(s, s->frames, out);
+
+ return ff_filter_frame(outlink, out);
+}
+
+static const AVOption tmix_options[] = {
+ { "frames", "set number of successive frames to mix", OFFSET(nb_inputs), AV_OPT_TYPE_INT, {.i64=3}, 2, 128, .flags = FLAGS },
+ { "weights", "set weight for each frame", OFFSET(weights_str), AV_OPT_TYPE_STRING, {.str="1 1 1"}, 0, 0, .flags = FLAGS },
+ { NULL },
+};
+
+static const AVFilterPad inputs[] = {
+ {
+ .name = "default",
+ .type = AVMEDIA_TYPE_VIDEO,
+ .filter_frame = tmix_filter_frame,
+ },
+ { NULL }
+};
+
+AVFILTER_DEFINE_CLASS(tmix);
+
+AVFilter ff_vf_tmix = {
+ .name = "tmix",
+ .description = NULL_IF_CONFIG_SMALL("Mix successive video frames."),
+ .priv_size = sizeof(MixContext),
+ .priv_class = &tmix_class,
+ .query_formats = query_formats,
+ .outputs = outputs,
+ .inputs = inputs,
+ .init = init,
+ .uninit = uninit,
+};
+
+#endif /* CONFIG_TMIX_FILTER */