\input texinfo @c -*- texinfo -*- @settitle Video Hook Documentation @titlepage @sp 7 @center @titlefont{Video Hook Documentation} @sp 3 @end titlepage @chapter Introduction @var{Please be aware that vhook is deprecated, and hence its development is frozen (bug fixes are still accepted). The substitute will be the result of our 'Video Filter API' Google Summer of Code project. You may monitor its progress by subscribing to the ffmpeg-soc mailing list at @url{http://lists.mplayerhq.hu/mailman/listinfo/ffmpeg-soc}.} The video hook functionality is designed (mostly) for live video. It allows the video to be modified or examined between the decoder and the encoder. Any number of hook modules can be placed inline, and they are run in the order that they were specified on the ffmpeg command line. The video hook modules are provided for use as a base for your own modules, and are described below. Modules are loaded using the -vhook option to ffmpeg. The value of this parameter is a space separated list of arguments. The first is the module name, and the rest are passed as arguments to the Configure function of the module. The modules are dynamic libraries: They have different suffixes (.so, .dll, .dylib) depending on your platform. And your platform dictates if they need to be somewhere in your PATH, or in your LD_LIBRARY_PATH. Otherwise you will need to specify the full path of the vhook file that you are using. @section null.c This does nothing. Actually it converts the input image to RGB24 and then converts it back again. This is meant as a sample that you can use to test your setup. @section fish.c This implements a 'fish detector'. Essentially it converts the image into HSV space and tests whether more than a certain percentage of the pixels fall into a specific HSV cuboid. If so, then the image is saved into a file for processing by other bits of code. Why use HSV? It turns out that HSV cuboids represent a more compact range of colors than would an RGB cuboid. @section imlib2.c This module implements a text overlay for a video image. Currently it supports a fixed overlay or reading the text from a file. The string is passed through strftime() so that it is easy to imprint the date and time onto the image. This module depends on the external library imlib2, available on Sourceforge, among other places, if it is not already installed on your system. You may also overlay an image (even semi-transparent) like TV stations do. You may move either the text or the image around your video to create scrolling credits, for example. The font file used is looked for in a FONTPATH environment variable, and prepended to the point size as a command line option and can be specified with the full path to the font file, as in: @example -F /usr/X11R6/lib/X11/fonts/TTF/VeraBd.ttf/20 @end example where 20 is the point size. You can specify the filename to read RGB color names from. If none are specified, these defaults are used: @file{/usr/share/X11/rgb.txt} and @file{/usr/lib/X11/rgb.txt} Options: @multitable @columnfractions .2 .8 @item @option{-C <rgb.txt>} @tab The filename to read RGB color names from @item @option{-c <color>} @tab The color of the text @item @option{-F <fontname>} @tab The font face and size @item @option{-t <text>} @tab The text @item @option{-f <filename>} @tab The filename to read text from @item @option{-x <expression>}@tab x coordinate of text or image @item @option{-y <expression>}@tab y coordinate of text or image @item @option{-i <filename>} @tab The filename to read a image from @end multitable Expressions are functions of these variables: @multitable @columnfractions .2 .8 @item @var{N} @tab frame number (starting at zero) @item @var{H} @tab frame height @item @var{W} @tab frame width @item @var{h} @tab image height @item @var{w} @tab image width @item @var{X} @tab previous x coordinate of text or image @item @var{Y} @tab previous y coordinate of text or image @end multitable You may also use the constants @var{PI}, @var{E}, and the math functions available at the FFmpeg formula evaluator at (@url{ffmpeg-doc.html#SEC13}), except @var{bits2qp(bits)} and @var{qp2bits(qp)}. Usage examples: @example # Remember to set the path to your fonts FONTPATH="/cygdrive/c/WINDOWS/Fonts/" FONTPATH="$FONTPATH:/usr/share/imlib2/data/fonts/" FONTPATH="$FONTPATH:/usr/X11R6/lib/X11/fonts/TTF/" export FONTPATH # Bulb dancing in a Lissajous pattern ffmpeg -i input.avi -vhook \ 'vhook/imlib2.dll -x W*(0.5+0.25*sin(N/47*PI))-w/2 -y H*(0.5+0.50*cos(N/97*PI))-h/2 -i /usr/share/imlib2/data/images/bulb.png' \ -acodec copy -sameq output.avi # Text scrolling ffmpeg -i input.avi -vhook \ 'vhook/imlib2.dll -c red -F Vera.ttf/20 -x 150+0.5*N -y 70+0.25*N -t Hello' \ -acodec copy -sameq output.avi # Date and time stamp, security-camera style: ffmpeg -r 29.97 -s 320x256 -f video4linux -i /dev/video0 \ -vhook 'vhook/imlib2.so -x 0 -y 0 -i black-260x20.png' \ -vhook 'vhook/imlib2.so -c white -F VeraBd.ttf/12 -x 0 -y 0 -t %A-%D-%T' \ output.avi In this example the video is captured from the first video capture card as a 320x256 AVI, and a black 260 by 20 pixel PNG image is placed in the upper left corner, with the day, date and time overlaid on it in Vera Bold 12 point font. A simple black PNG file 260 pixels wide and 20 pixels tall was created in the GIMP for this purpose. # Scrolling credits from a text file ffmpeg -i input.avi -vhook \ 'vhook/imlib2.so -c white -F VeraBd.ttf/16 -x 100 -y -1.0*N -f credits.txt' \ -sameq output.avi In this example, the text is stored in a file, and is positioned 100 pixels from the left hand edge of the video. The text is scrolled from the bottom up. Making the y factor positive will scroll from the top down. Increasing the magnitude of the y factor makes the text scroll faster, decreasing it makes it scroll slower. Hint: Blank lines containing only a newline are treated as end-of-file. To create blank lines, use lines that consist of space characters only. # Scrolling credits with custom color from a text file ffmpeg -i input.avi -vhook \ 'vhook/imlib2.so -C rgb.txt -c CustomColor1 -F VeraBd.ttf/16 -x 100 -y -1.0*N -f credits.txt' \ -sameq output.avi This example does the same as the one above, but specifies an rgb.txt file to be used, which has a custom made color in it. # Variable colors ffmpeg -i input.avi -vhook \ 'vhook/imlib2.so -t Hello -R abs(255*sin(N/47*PI)) -G abs(255*sin(N/47*PI)) -B abs(255*sin(N/47*PI))' \ -sameq output.avi In this example, the color for the text goes up and down from black to white. # Text fade-out ffmpeg -i input.avi -vhook \ 'vhook/imlib2.so -t Hello -A max(0,255-exp(N/47))' \ -sameq output.avi In this example, the text fades out in about 10 seconds for a 25 fps input video file. # scrolling credits from a graphics file ffmpeg -sameq -i input.avi \ -vhook 'vhook/imlib2.so -x 0 -y -1.0*N -i credits.png' output.avi In this example, a transparent PNG file the same width as the video (e.g. 320 pixels), but very long, (e.g. 3000 pixels), was created, and text, graphics, brushstrokes, etc, were added to the image. The image is then scrolled up, from the bottom of the frame. @end example @section ppm.c It's basically a launch point for a PPM pipe, so you can use any executable (or script) which consumes a PPM on stdin and produces a PPM on stdout (and flushes each frame). The Netpbm utilities are a series of such programs. A list of them is here: @url{http://netpbm.sourceforge.net/doc/directory.html} Usage example: @example ffmpeg -i input -vhook "/path/to/ppm.so some-ppm-filter args" output @end example @section drawtext.c This module implements a text overlay for a video image. Currently it supports a fixed overlay or reading the text from a file. The string is passed through strftime() so that it is easy to imprint the date and time onto the image. Features: @itemize @minus @item TrueType, Type1 and others via the FreeType2 library @item Font kerning (better output) @item Line Wrap (put the text that doesn't fit one line on the next line) @item Background box (currently in development) @item Outline @end itemize Options: @multitable @columnfractions .2 .8 @item @option{-c <color>} @tab Foreground color of the text ('internet' way) <#RRGGBB> [default #FFFFFF] @item @option{-C <color>} @tab Background color of the text ('internet' way) <#RRGGBB> [default #000000] @item @option{-f <font-filename>} @tab font file to use @item @option{-t <text>} @tab text to display @item @option{-T <filename>} @tab file to read text from @item @option{-x <pos>} @tab x coordinate of the start of the text @item @option{-y <pos>} @tab y coordinate of the start of the text @end multitable Text fonts are being looked for in a FONTPATH environment variable. If the FONTPATH environment variable is not available, or is not checked by your target (i.e. Cygwin), then specify the full path to the font file as in: @example -f /usr/X11R6/lib/X11/fonts/TTF/VeraBd.ttf @end example Usage Example: @example # Remember to set the path to your fonts FONTPATH="/cygdrive/c/WINDOWS/Fonts/" FONTPATH="$FONTPATH:/usr/share/imlib2/data/fonts/" FONTPATH="$FONTPATH:/usr/X11R6/lib/X11/fonts/TTF/" export FONTPATH # Time and date display ffmpeg -f video4linux2 -i /dev/video0 \ -vhook 'vhook/drawtext.so -f VeraBd.ttf -t %A-%D-%T' movie.mpg This example grabs video from the first capture card and outputs it to an MPEG video, and places "Weekday-dd/mm/yy-hh:mm:ss" at the top left of the frame, updated every second, using the Vera Bold TrueType Font, which should exist in: /usr/X11R6/lib/X11/fonts/TTF/ @end example Check the man page for strftime() for all the various ways you can format the date and time. @section watermark.c Command Line options: @multitable @columnfractions .2 .8 @item @option{-m [0|1]} @tab Mode (default: 0, see below) @item @option{-t 000000 - FFFFFF} @tab Threshold, six digit hex number @item @option{-f <filename>} @tab Watermark image filename, must be specified! @end multitable MODE 0: The watermark picture works like this (assuming color intensities 0..0xFF): Per color do this: If mask color is 0x80, no change to the original frame. If mask color is < 0x80 the absolute difference is subtracted from the frame. If result < 0, result = 0. If mask color is > 0x80 the absolute difference is added to the frame. If result > 0xFF, result = 0xFF. You can override the 0x80 level with the -t flag. E.g. if threshold is 000000 the color value of watermark is added to the destination. This way a mask that is visible both in light and dark pictures can be made (e.g. by using a picture generated by the Gimp and the bump map tool). An example watermark file is at: @url{http://engene.se/ffmpeg_watermark.gif} MODE 1: Per color do this: If mask color > threshold color then the watermark pixel is used. Example usage: @example ffmpeg -i infile -vhook '/path/watermark.so -f wm.gif' -an out.mov ffmpeg -i infile -vhook '/path/watermark.so -f wm.gif -m 1 -t 222222' -an out.mov @end example @bye