diff options
author | Anton Khirnov <anton@khirnov.net> | 2015-02-12 00:07:29 +0100 |
---|---|---|
committer | Anton Khirnov <anton@khirnov.net> | 2015-02-28 21:51:24 +0100 |
commit | 990e4a6639d0714198583c2812b1074b5751898f (patch) | |
tree | 5cd38ccb8d1832831a4ccc70537aa1b261e38d70 | |
parent | ea1d0b7ece1881c2f1360f8c1e2116791105af21 (diff) | |
download | ffmpeg-990e4a6639d0714198583c2812b1074b5751898f.tar.gz |
Add a QSV decoding example.
-rwxr-xr-x | configure | 7 | ||||
-rw-r--r-- | doc/Makefile | 1 | ||||
-rw-r--r-- | doc/examples/qsvdec.c | 484 |
3 files changed, 492 insertions, 0 deletions
@@ -1137,6 +1137,7 @@ EXAMPLE_LIST=" filter_audio_example metadata_example output_example + qsvdec_example transcode_aac_example " @@ -1550,6 +1551,7 @@ HAVE_LIST=" sdl section_data_rel_ro threads + vaapi_x11 vdpau_x11 xlib " @@ -2187,6 +2189,7 @@ avcodec_example_deps="avcodec avutil" filter_audio_example_deps="avfilter avutil" metadata_example_deps="avformat avutil" output_example_deps="avcodec avformat avutil swscale" +qsvdec_example_deps="avcodec avutil libmfx h264_qsv_decoder vaapi_x11" transcode_aac_example_deps="avcodec avformat avresample" # libraries, in linking order @@ -4330,6 +4333,10 @@ if enabled x11grab; then require Xfixes X11/extensions/Xfixes.h XFixesGetCursorImage -lXfixes fi +enabled vaapi && enabled xlib && + check_lib2 "va/va.h va/va_x11.h" vaGetDisplay -lva -lva-x11 && + enable vaapi_x11 + enabled vdpau && check_cpp_condition vdpau/vdpau.h "defined VDP_DECODER_PROFILE_MPEG4_PART2_ASP" || disable vdpau diff --git a/doc/Makefile b/doc/Makefile index 1c205a5236..2f6a5fb0c1 100644 --- a/doc/Makefile +++ b/doc/Makefile @@ -19,6 +19,7 @@ DOC_EXAMPLES-$(CONFIG_AVCODEC_EXAMPLE) += avcodec DOC_EXAMPLES-$(CONFIG_FILTER_AUDIO_EXAMPLE) += filter_audio DOC_EXAMPLES-$(CONFIG_METADATA_EXAMPLE) += metadata DOC_EXAMPLES-$(CONFIG_OUTPUT_EXAMPLE) += output +DOC_EXAMPLES-$(CONFIG_QSVDEC_EXAMPLE) += qsvdec DOC_EXAMPLES-$(CONFIG_TRANSCODE_AAC_EXAMPLE) += transcode_aac ALL_DOC_EXAMPLES = avcodec filter_audio metadata output transcode_aac diff --git a/doc/examples/qsvdec.c b/doc/examples/qsvdec.c new file mode 100644 index 0000000000..6dbb21031d --- /dev/null +++ b/doc/examples/qsvdec.c @@ -0,0 +1,484 @@ +/* + * Copyright (c) 2015 Anton Khirnov + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to deal + * in the Software without restriction, including without limitation the rights + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell + * copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in + * all copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN + * THE SOFTWARE. + */ + +/** + * @file + * Intel QSV-accelerated H.264 decoding example. + * + * @example qsvdec.c + * This example shows how to do QSV-accelerated H.264 decoding with output + * frames in the VA-API video surfaces. + */ + +#include "config.h" + +#include <stdio.h> + +#include <mfx/mfxvideo.h> + +#include <va/va.h> +#include <va/va_x11.h> +#include <X11/Xlib.h> + +#include "libavformat/avformat.h" +#include "libavformat/avio.h" + +#include "libavcodec/avcodec.h" +#include "libavcodec/qsv.h" + +#include "libavutil/error.h" +#include "libavutil/mem.h" + +typedef struct DecodeContext { + mfxSession mfx_session; + VADisplay va_dpy; + + VASurfaceID *surfaces; + mfxMemId *surface_ids; + int *surface_used; + int nb_surfaces; + + mfxFrameInfo frame_info; +} DecodeContext; + +static mfxStatus frame_alloc(mfxHDL pthis, mfxFrameAllocRequest *req, + mfxFrameAllocResponse *resp) +{ + DecodeContext *decode = pthis; + int err, i; + + if (decode->surfaces) { + fprintf(stderr, "Multiple allocation requests.\n"); + return MFX_ERR_MEMORY_ALLOC; + } + if (!(req->Type & MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET)) { + fprintf(stderr, "Unsupported surface type: %d\n", req->Type); + return MFX_ERR_UNSUPPORTED; + } + if (req->Info.BitDepthLuma != 8 || req->Info.BitDepthChroma != 8 || + req->Info.Shift || req->Info.FourCC != MFX_FOURCC_NV12 || + req->Info.ChromaFormat != MFX_CHROMAFORMAT_YUV420) { + fprintf(stderr, "Unsupported surface properties.\n"); + return MFX_ERR_UNSUPPORTED; + } + + decode->surfaces = av_malloc_array (req->NumFrameSuggested, sizeof(*decode->surfaces)); + decode->surface_ids = av_malloc_array (req->NumFrameSuggested, sizeof(*decode->surface_ids)); + decode->surface_used = av_mallocz_array(req->NumFrameSuggested, sizeof(*decode->surface_used)); + if (!decode->surfaces || !decode->surface_ids || !decode->surface_used) + goto fail; + + err = vaCreateSurfaces(decode->va_dpy, VA_RT_FORMAT_YUV420, + req->Info.Width, req->Info.Height, + decode->surfaces, req->NumFrameSuggested, + NULL, 0); + if (err != VA_STATUS_SUCCESS) { + fprintf(stderr, "Error allocating VA surfaces\n"); + goto fail; + } + decode->nb_surfaces = req->NumFrameSuggested; + + for (i = 0; i < decode->nb_surfaces; i++) + decode->surface_ids[i] = &decode->surfaces[i]; + + resp->mids = decode->surface_ids; + resp->NumFrameActual = decode->nb_surfaces; + + decode->frame_info = req->Info; + + return MFX_ERR_NONE; +fail: + av_freep(&decode->surfaces); + av_freep(&decode->surface_ids); + av_freep(&decode->surface_used); + + return MFX_ERR_MEMORY_ALLOC; +} + +static mfxStatus frame_free(mfxHDL pthis, mfxFrameAllocResponse *resp) +{ + DecodeContext *decode = pthis; + + if (decode->surfaces) + vaDestroySurfaces(decode->va_dpy, decode->surfaces, decode->nb_surfaces); + av_freep(&decode->surfaces); + av_freep(&decode->surface_ids); + av_freep(&decode->surface_used); + decode->nb_surfaces = 0; + + return MFX_ERR_NONE; +} + +static mfxStatus frame_lock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr) +{ + return MFX_ERR_UNSUPPORTED; +} + +static mfxStatus frame_unlock(mfxHDL pthis, mfxMemId mid, mfxFrameData *ptr) +{ + return MFX_ERR_UNSUPPORTED; +} + +static mfxStatus frame_get_hdl(mfxHDL pthis, mfxMemId mid, mfxHDL *hdl) +{ + *hdl = mid; + return MFX_ERR_NONE; +} + +static void free_buffer(void *opaque, uint8_t *data) +{ + int *used = opaque; + *used = 0; + av_freep(&data); +} + +static int get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags) +{ + DecodeContext *decode = avctx->opaque; + + mfxFrameSurface1 *surf; + AVBufferRef *surf_buf; + int idx; + + for (idx = 0; idx < decode->nb_surfaces; idx++) { + if (!decode->surface_used[idx]) + break; + } + if (idx == decode->nb_surfaces) { + fprintf(stderr, "No free surfaces\n"); + return AVERROR(ENOMEM); + } + + surf = av_mallocz(sizeof(*surf)); + if (!surf) + return AVERROR(ENOMEM); + surf_buf = av_buffer_create((uint8_t*)surf, sizeof(*surf), free_buffer, + &decode->surface_used[idx], AV_BUFFER_FLAG_READONLY); + if (!surf_buf) { + av_freep(&surf); + return AVERROR(ENOMEM); + } + + surf->Info = decode->frame_info; + surf->Data.MemId = &decode->surfaces[idx]; + + frame->buf[0] = surf_buf; + frame->data[3] = (uint8_t*)surf; + + decode->surface_used[idx] = 1; + + return 0; +} + +static int get_format(AVCodecContext *avctx, const enum AVPixelFormat *pix_fmts) +{ + while (*pix_fmts != AV_PIX_FMT_NONE) { + if (*pix_fmts == AV_PIX_FMT_QSV) { + if (!avctx->hwaccel_context) { + DecodeContext *decode = avctx->opaque; + AVQSVContext *qsv = av_qsv_alloc_context(); + if (!qsv) + return AV_PIX_FMT_NONE; + + qsv->session = decode->mfx_session; + qsv->iopattern = MFX_IOPATTERN_OUT_VIDEO_MEMORY; + + avctx->hwaccel_context = qsv; + } + + return AV_PIX_FMT_QSV; + } + + pix_fmts++; + } + + fprintf(stderr, "The QSV pixel format not offered in get_format()\n"); + + return AV_PIX_FMT_NONE; +} + +static int decode_packet(DecodeContext *decode, AVCodecContext *decoder_ctx, + AVFrame *frame, AVPacket *pkt, + AVIOContext *output_ctx) +{ + int ret = 0; + int got_frame = 1; + + while (pkt->size > 0 || (!pkt->data && got_frame)) { + ret = avcodec_decode_video2(decoder_ctx, frame, &got_frame, pkt); + if (ret < 0) { + fprintf(stderr, "Error during decoding\n"); + return ret; + } + + pkt->data += ret; + pkt->size -= ret; + + /* A real program would do something useful with the decoded frame here. + * We just retrieve the raw data and write it to a file, which is rather + * useless but pedagogic. */ + if (got_frame) { + mfxFrameSurface1 *surf = (mfxFrameSurface1*)frame->data[3]; + VASurfaceID surface = *(VASurfaceID*)surf->Data.MemId; + + VAImageFormat img_fmt = { + .fourcc = VA_FOURCC_NV12, + .byte_order = VA_LSB_FIRST, + .bits_per_pixel = 8, + .depth = 8, + }; + + VAImage img; + + VAStatus err; + uint8_t *data; + int i, j; + + img.buf = VA_INVALID_ID; + img.image_id = VA_INVALID_ID; + + err = vaCreateImage(decode->va_dpy, &img_fmt, + frame->width, frame->height, &img); + if (err != VA_STATUS_SUCCESS) { + fprintf(stderr, "Error creating an image: %s\n", + vaErrorStr(err)); + ret = AVERROR_UNKNOWN; + goto fail; + } + + err = vaGetImage(decode->va_dpy, surface, 0, 0, + frame->width, frame->height, + img.image_id); + if (err != VA_STATUS_SUCCESS) { + fprintf(stderr, "Error getting an image: %s\n", + vaErrorStr(err)); + ret = AVERROR_UNKNOWN; + goto fail; + } + + err = vaMapBuffer(decode->va_dpy, img.buf, (void**)&data); + if (err != VA_STATUS_SUCCESS) { + fprintf(stderr, "Error mapping the image buffer: %s\n", + vaErrorStr(err)); + ret = AVERROR_UNKNOWN; + goto fail; + } + + for (i = 0; i < img.num_planes; i++) + for (j = 0; j < (img.height >> (i > 0)); j++) + avio_write(output_ctx, data + img.offsets[i] + j * img.pitches[i], img.width); + +fail: + if (img.buf != VA_INVALID_ID) + vaUnmapBuffer(decode->va_dpy, img.buf); + if (img.image_id != VA_INVALID_ID) + vaDestroyImage(decode->va_dpy, img.image_id); + av_frame_unref(frame); + + if (ret < 0) + return ret; + } + } + + return 0; +} + +int main(int argc, char **argv) +{ + AVFormatContext *input_ctx = NULL; + AVStream *video_st = NULL; + AVCodecContext *decoder_ctx = NULL; + const AVCodec *decoder; + + AVPacket pkt = { 0 }; + AVFrame *frame = NULL; + + DecodeContext decode = { NULL }; + + Display *dpy = NULL; + int va_ver_major, va_ver_minor; + + mfxIMPL mfx_impl = MFX_IMPL_AUTO_ANY; + mfxVersion mfx_ver = { { 1, 1 } }; + + mfxFrameAllocator frame_allocator = { + .pthis = &decode, + .Alloc = frame_alloc, + .Lock = frame_lock, + .Unlock = frame_unlock, + .GetHDL = frame_get_hdl, + .Free = frame_free, + }; + + AVIOContext *output_ctx = NULL; + + int ret, i, err; + + av_register_all(); + + if (argc < 3) { + fprintf(stderr, "Usage: %s <input file> <output file>\n", argv[0]); + return 1; + } + + /* open the input file */ + ret = avformat_open_input(&input_ctx, argv[1], NULL, NULL); + if (ret < 0) { + fprintf(stderr, "Cannot open input file '%s': ", argv[1]); + goto finish; + } + + /* find the first H.264 video stream */ + for (i = 0; i < input_ctx->nb_streams; i++) { + AVStream *st = input_ctx->streams[i]; + + if (st->codec->codec_id == AV_CODEC_ID_H264 && !video_st) + video_st = st; + else + st->discard = AVDISCARD_ALL; + } + if (!video_st) { + fprintf(stderr, "No H.264 video stream in the input file\n"); + goto finish; + } + + /* initialize VA-API */ + dpy = XOpenDisplay(NULL); + if (!dpy) { + fprintf(stderr, "Cannot open the X display\n"); + goto finish; + } + decode.va_dpy = vaGetDisplay(dpy); + if (!decode.va_dpy) { + fprintf(stderr, "Cannot open the VA display\n"); + goto finish; + } + + err = vaInitialize(decode.va_dpy, &va_ver_major, &va_ver_minor); + if (err != VA_STATUS_SUCCESS) { + fprintf(stderr, "Cannot initialize VA: %s\n", vaErrorStr(err)); + goto finish; + } + fprintf(stderr, "Initialized VA v%d.%d\n", va_ver_major, va_ver_minor); + + /* initialize an MFX session */ + err = MFXInit(mfx_impl, &mfx_ver, &decode.mfx_session); + if (err != MFX_ERR_NONE) { + fprintf(stderr, "Error initializing an MFX session\n"); + goto finish; + } + + MFXVideoCORE_SetHandle(decode.mfx_session, MFX_HANDLE_VA_DISPLAY, decode.va_dpy); + MFXVideoCORE_SetFrameAllocator(decode.mfx_session, &frame_allocator); + + /* initialize the decoder */ + decoder = avcodec_find_decoder_by_name("h264_qsv"); + if (!decoder) { + fprintf(stderr, "The QSV decoder is not present in libavcodec\n"); + goto finish; + } + + decoder_ctx = avcodec_alloc_context3(decoder); + if (!decoder_ctx) { + ret = AVERROR(ENOMEM); + goto finish; + } + decoder_ctx->codec_id = AV_CODEC_ID_H264; + if (video_st->codec->extradata_size) { + decoder_ctx->extradata = av_mallocz(video_st->codec->extradata_size + + FF_INPUT_BUFFER_PADDING_SIZE); + if (!decoder_ctx->extradata) { + ret = AVERROR(ENOMEM); + goto finish; + } + memcpy(decoder_ctx->extradata, video_st->codec->extradata, + video_st->codec->extradata_size); + decoder_ctx->extradata_size = video_st->codec->extradata_size; + } + decoder_ctx->refcounted_frames = 1; + + decoder_ctx->opaque = &decode; + decoder_ctx->get_buffer2 = get_buffer; + decoder_ctx->get_format = get_format; + + ret = avcodec_open2(decoder_ctx, NULL, NULL); + if (ret < 0) { + fprintf(stderr, "Error opening the decoder: "); + goto finish; + } + + /* open the output stream */ + ret = avio_open(&output_ctx, argv[2], AVIO_FLAG_WRITE); + if (ret < 0) { + fprintf(stderr, "Error opening the output context: "); + goto finish; + } + + frame = av_frame_alloc(); + if (!frame) { + ret = AVERROR(ENOMEM); + goto finish; + } + + /* actual decoding */ + while (ret >= 0) { + ret = av_read_frame(input_ctx, &pkt); + if (ret < 0) + break; + + if (pkt.stream_index == video_st->index) + ret = decode_packet(&decode, decoder_ctx, frame, &pkt, output_ctx); + + av_packet_unref(&pkt); + } + + /* flush the decoder */ + pkt.data = NULL; + pkt.size = 0; + ret = decode_packet(&decode, decoder_ctx, frame, &pkt, output_ctx); + +finish: + if (ret < 0) { + char buf[1024]; + av_strerror(ret, buf, sizeof(buf)); + fprintf(stderr, "%s\n", buf); + } + + avformat_close_input(&input_ctx); + + av_frame_free(&frame); + + if (decode.mfx_session) + MFXClose(decode.mfx_session); + if (decode.va_dpy) + vaTerminate(decode.va_dpy); + if (dpy) + XCloseDisplay(dpy); + + if (decoder_ctx) + av_freep(&decoder_ctx->hwaccel_context); + avcodec_free_context(&decoder_ctx); + + avio_close(output_ctx); + + return ret; +} |