[FFmpeg-devel] [PATCH v3] avfilter: Add tonemap vaapi filter for H2S
Vittorio Giovara
vittorio.giovara at gmail.com
Mon Dec 2 20:27:51 EET 2019
On Mon, Dec 2, 2019 at 2:19 AM Xinpeng Sun <xinpeng.sun at intel.com> wrote:
> It performs HDR(High Dynamic Range) to SDR(Standard Dynamic Range)
> conversion
> with tone-mapping. It only supports HDR10 as input temporarily.
>
> An example command to use this filter with vaapi codecs:
> FFMPEG -hwaccel vaapi -vaapi_device /dev/dri/renderD128
> -hwaccel_output_format vaapi \
> -i INPUT -vf 'tonemap_vaapi=format=p010' -c:v hevc_vaapi -profile 2 OUTPUT
>
> Signed-off-by: Xinpeng Sun <xinpeng.sun at intel.com>
> Signed-off-by: Zachary Zhou <zachary.zhou at intel.com>
> ---
> configure | 2 +
> doc/filters.texi | 81 +++++++
> libavfilter/Makefile | 1 +
> libavfilter/allfilters.c | 1 +
> libavfilter/vf_tonemap_vaapi.c | 420 +++++++++++++++++++++++++++++++++
> 5 files changed, 505 insertions(+)
> create mode 100644 libavfilter/vf_tonemap_vaapi.c
>
> diff --git a/configure b/configure
> index ca7137f341..5272fb2a57 100755
> --- a/configure
> +++ b/configure
> @@ -3576,6 +3576,7 @@ tinterlace_filter_deps="gpl"
> tinterlace_merge_test_deps="tinterlace_filter"
> tinterlace_pad_test_deps="tinterlace_filter"
> tonemap_filter_deps="const_nan"
> +tonemap_vaapi_filter_deps="vaapi
> VAProcPipelineParameterBuffer_output_hdr_metadata"
> tonemap_opencl_filter_deps="opencl const_nan"
> transpose_opencl_filter_deps="opencl"
> transpose_vaapi_filter_deps="vaapi VAProcPipelineCaps_rotation_flags"
> @@ -6576,6 +6577,7 @@ if enabled vaapi; then
>
> check_type "va/va.h va/va_dec_hevc.h" "VAPictureParameterBufferHEVC"
> check_struct "va/va.h" "VADecPictureParameterBufferVP9" bit_depth
> + check_struct "va/va.h va/va_vpp.h" "VAProcPipelineParameterBuffer"
> output_hdr_metadata
> check_struct "va/va.h va/va_vpp.h" "VAProcPipelineCaps" rotation_flags
> check_type "va/va.h va/va_enc_hevc.h"
> "VAEncPictureParameterBufferHEVC"
> check_type "va/va.h va/va_enc_jpeg.h"
> "VAEncPictureParameterBufferJPEG"
> diff --git a/doc/filters.texi b/doc/filters.texi
> index 5fdec6f015..7223ab89a3 100644
> --- a/doc/filters.texi
> +++ b/doc/filters.texi
> @@ -20972,6 +20972,87 @@ Apply a strong blur of both luma and chroma
> parameters:
>
> @c man end OPENCL VIDEO FILTERS
>
> + at chapter VAAPI Video Filters
> + at c man begin VAAPI VIDEO FILTERS
> +
> +VAAPI Video filters are usually used with VAAPI decoder and VAAPI
> encoder. Below is a description of VAAPI video filters.
> +
> +To enable compilation of these filters you need to configure FFmpeg with
> + at code{--enable-vaapi}.
> +
> +Running VAAPI filters requires you to initialize a hardware device and to
> pass that device to all filters in any filter graph.
> + at table @option
> +
> + at item -hwaccel vaapi
> +Specify the hardware accelerator as @var{vaapi}.
> +
> + at item -vaapi_device @var{driver_path}
> +Specify the vaapi driver path with @var{driver_path}.
> +
> + at item -hwaccel_output_format @var{vaapi}
> +Specify the output format of hardware accelerator as @var{vaapi}. All
> VAAPI hardware surfaces in ffmpeg are represented by the @var{vaapi} pixfmt.
> +
> + at end table
> +
> + at itemize
> + at item
> +Example of running tonemap_vaapi filter with default parameters on it.
> + at example
> +-hwaccel vaapi -vaapi_device /dev/dri/renderD128 -hwaccel_output_format
> vaapi -i INPUT -vf "tonemap_vaapi, hwdownload" OUTPUT
> + at end example
> + at end itemize
> +
> +Since VAAPI filters are not able to access frame data in arbitrary
> memory, so if you use a decoder other than VAAPI decoder before VAAPI
> filters, all frame data needs to be uploaded(@ref{hwupload}) to hardware
> surfaces connected to the appropriate device before being used. Also if you
> add a encoder other than VAAPI encoder after VAAPI filters, the hardware
> surfaces should be downloaded(@ref{hwdownload}) back to normal memory. Note
> that @ref{hwupload} will upload to a surface with the same layout as the
> software frame, so it may be necessary to add a @ref{format} filter
> immediately before to get the input into the right format and
> @ref{hwdownload} does not support all formats on the output - it may be
> necessary to insert an additional @ref{format} filter immediately following
> in the graph to get the output in a supported format.
> +
> + at section tonemap_vappi
> +
> +Perform HDR(High Dynamic Range) to SDR(Standard Dynamic Range) conversion
> with tone-mapping.
> +It maps the dynamic range of HDR10 content to the SDR content.
> +It only accepts HDR10 as input temporarily.
> +
> +It accepts the following parameters:
> +
> + at table @option
> + at item format
> +Specify the output pixel format.
> +
> +Currently supported formats are:
> + at table @var
> + at item p010
> + at item nv12
> + at end table
> +
> +Default is nv12.
> +
> + at item primaries, p
> +Set the output color primaries.
> +
> +Default is same as input.
> +
> + at item transfer, t
> +Set the output transfer characteristics.
> +
> +Default is bt709.
> +
> + at item matrix, m
> +Set the output colorspace matrix.
> +
> +Default is same as input.
> +
> + at end table
> +
> + at subsection Example
> +
> + at itemize
> + at item
> +Convert HDR(HDR10) video to bt2020-transfer-characteristic p010 format
> + at example
> +-i INPUT -vf "hwupload,tonemap_vaapi=format=p010:t=bt2020-10,hwdownload"
> OUTPUT
> + at end example
> + at end itemize
> +
> + at c man end VAAPI VIDEO FILTERS
> +
> @chapter Video Sources
> @c man begin VIDEO SOURCES
>
> diff --git a/libavfilter/Makefile b/libavfilter/Makefile
> index 46e3eecf9a..c568454af5 100644
> --- a/libavfilter/Makefile
> +++ b/libavfilter/Makefile
> @@ -411,6 +411,7 @@ OBJS-$(CONFIG_TMIX_FILTER) +=
> vf_mix.o framesync.o
> OBJS-$(CONFIG_TONEMAP_FILTER) += vf_tonemap.o colorspace.o
> OBJS-$(CONFIG_TONEMAP_OPENCL_FILTER) += vf_tonemap_opencl.o
> colorspace.o opencl.o \
> opencl/tonemap.o
> opencl/colorspace_common.o
> +OBJS-$(CONFIG_TONEMAP_VAAPI_FILTER) += vf_tonemap_vaapi.o
> vaapi_vpp.o
> OBJS-$(CONFIG_TPAD_FILTER) += vf_tpad.o
> OBJS-$(CONFIG_TRANSPOSE_FILTER) += vf_transpose.o
> OBJS-$(CONFIG_TRANSPOSE_NPP_FILTER) += vf_transpose_npp.o
> diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c
> index 2a69227476..c203b87dd9 100644
> --- a/libavfilter/allfilters.c
> +++ b/libavfilter/allfilters.c
> @@ -391,6 +391,7 @@ extern AVFilter ff_vf_tlut2;
> extern AVFilter ff_vf_tmix;
> extern AVFilter ff_vf_tonemap;
> extern AVFilter ff_vf_tonemap_opencl;
> +extern AVFilter ff_vf_tonemap_vaapi;
> extern AVFilter ff_vf_tpad;
> extern AVFilter ff_vf_transpose;
> extern AVFilter ff_vf_transpose_npp;
> diff --git a/libavfilter/vf_tonemap_vaapi.c
> b/libavfilter/vf_tonemap_vaapi.c
> new file mode 100644
> index 0000000000..c0c54662f5
> --- /dev/null
> +++ b/libavfilter/vf_tonemap_vaapi.c
> @@ -0,0 +1,420 @@
> +/*
> + * This file is part of FFmpeg.
> + *
> + * FFmpeg is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU Lesser General Public
> + * License as published by the Free Software Foundation; either
> + * version 2.1 of the License, or (at your option) any later version.
> + *
> + * FFmpeg is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
> + * Lesser General Public License for more details.
> + *
> + * You should have received a copy of the GNU Lesser General Public
> + * License along with FFmpeg; if not, write to the Free Software
> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
> 02110-1301 USA
> + */
> +#include <string.h>
> +
> +#include "libavutil/avassert.h"
> +#include "libavutil/mem.h"
> +#include "libavutil/opt.h"
> +#include "libavutil/pixdesc.h"
> +#include "libavutil/mastering_display_metadata.h"
> +
> +#include "avfilter.h"
> +#include "formats.h"
> +#include "internal.h"
> +#include "vaapi_vpp.h"
> +
> +typedef struct HDRVAAPIContext {
> + VAAPIVPPContext vpp_ctx; // must be the first field
> +
> + char *output_format_string;
> +
> + char *color_primaries_string;
> + char *color_transfer_string;
> + char *color_matrix_string;
> +
> + enum AVColorPrimaries color_primaries;
> + enum AVColorTransferCharacteristic color_transfer;
> + enum AVColorSpace color_matrix;
> +
> + VAHdrMetaDataHDR10 in_metadata;
> +
> + AVFrameSideData *src_display;
> + AVFrameSideData *src_light;
> +} HDRVAAPIContext;
> +
> +static int tonemap_vaapi_save_metadata(AVFilterContext *avctx, AVFrame
> *input_frame)
> +{
> + HDRVAAPIContext *ctx = avctx->priv;
> + AVMasteringDisplayMetadata *hdr_meta;
> + AVContentLightMetadata *light_meta;
> +
> + if (input_frame->color_trc != AVCOL_TRC_SMPTE2084) {
> + av_log(avctx, AV_LOG_WARNING, "Only support HDR10 as input for
> vaapi tone-mapping\n");
> + input_frame->color_trc = AVCOL_TRC_SMPTE2084;
> + }
> +
> + ctx->src_display = av_frame_get_side_data(input_frame,
> +
> AV_FRAME_DATA_MASTERING_DISPLAY_METADATA);
> + if (ctx->src_display) {
> + hdr_meta = (AVMasteringDisplayMetadata *)ctx->src_display->data;
> + if (!hdr_meta) {
> + av_log(avctx, AV_LOG_ERROR, "No mastering display data\n");
> + return AVERROR(EINVAL);
> + }
> +
> + if (hdr_meta->has_luminance) {
> + const int luma_den = 10000;
> + ctx->in_metadata.max_display_mastering_luminance =
> + lrint(luma_den * av_q2d(hdr_meta->max_luminance));
> + ctx->in_metadata.min_display_mastering_luminance =
> + FFMIN(lrint(luma_den * av_q2d(hdr_meta->min_luminance)),
> + ctx->in_metadata.max_display_mastering_luminance);
> +
> + av_log(avctx, AV_LOG_DEBUG,
> + "Mastering Display Metadata(in luminance):\n");
> + av_log(avctx, AV_LOG_DEBUG,
> + "min_luminance=%u, max_luminance=%u\n",
> + ctx->in_metadata.min_display_mastering_luminance,
> + ctx->in_metadata.max_display_mastering_luminance);
> + }
> +
> + if (hdr_meta->has_primaries) {
> + int i;
> + const int mapping[3] = {1, 2, 0}; //green, blue, red
> + const int chroma_den = 50000;
> +
> + for (i = 0; i < 3; i++) {
> + const int j = mapping[i];
> + ctx->in_metadata.display_primaries_x[i] =
> + FFMIN(lrint(chroma_den *
> +
> av_q2d(hdr_meta->display_primaries[j][0])),
> + chroma_den);
> + ctx->in_metadata.display_primaries_y[i] =
> + FFMIN(lrint(chroma_den *
> +
> av_q2d(hdr_meta->display_primaries[j][1])),
> + chroma_den);
> + }
> +
> + ctx->in_metadata.white_point_x =
> + FFMIN(lrint(chroma_den *
> av_q2d(hdr_meta->white_point[0])),
> + chroma_den);
> + ctx->in_metadata.white_point_y =
> + FFMIN(lrint(chroma_den *
> av_q2d(hdr_meta->white_point[1])),
> + chroma_den);
> +
> + av_log(avctx, AV_LOG_DEBUG,
> + "Mastering Display Metadata(in primaries):\n");
> + av_log(avctx, AV_LOG_DEBUG,
> + "G(%u,%u) B(%u,%u) R(%u,%u) WP(%u,%u)\n",
> + ctx->in_metadata.display_primaries_x[0],
> + ctx->in_metadata.display_primaries_y[0],
> + ctx->in_metadata.display_primaries_x[1],
> + ctx->in_metadata.display_primaries_y[1],
> + ctx->in_metadata.display_primaries_x[2],
> + ctx->in_metadata.display_primaries_y[2],
> + ctx->in_metadata.white_point_x,
> + ctx->in_metadata.white_point_y);
> + }
> + } else {
> + av_log(avctx, AV_LOG_ERROR, "No mastering display data from
> input\n");
> + return AVERROR(EINVAL);
> + }
> +
> + ctx->src_light = av_frame_get_side_data(input_frame,
> +
> AV_FRAME_DATA_CONTENT_LIGHT_LEVEL);
> + if (ctx->src_light) {
> + light_meta = (AVContentLightMetadata *)ctx->src_light->data;
> + if (!light_meta) {
> + av_log(avctx, AV_LOG_ERROR, "No light meta data\n");
>
nit: metadata
(no need to send this patch again, whoever pushes this should take care of
it)
+ return AVERROR(EINVAL);
> + }
> +
> + ctx->in_metadata.max_content_light_level = light_meta->MaxCLL;
> + ctx->in_metadata.max_pic_average_light_level =
> light_meta->MaxFALL;
> +
> + av_log(avctx, AV_LOG_DEBUG,
> + "Mastering Content Light Level (in):\n");
> + av_log(avctx, AV_LOG_DEBUG,
> + "MaxCLL(%u) MaxFALL(%u)\n",
> + ctx->in_metadata.max_content_light_level,
> + ctx->in_metadata.max_pic_average_light_level);
> + } else {
> + av_log(avctx, AV_LOG_DEBUG, "No content light level from
> input\n");
> + }
> + return 0;
> +}
> +
> +static int tonemap_vaapi_set_filter_params(AVFilterContext *avctx,
> AVFrame *input_frame)
> +{
> + VAAPIVPPContext *vpp_ctx = avctx->priv;
> + HDRVAAPIContext *ctx = avctx->priv;
> + VAStatus vas;
> + VAProcFilterParameterBufferHDRToneMapping *hdrtm_param;
> +
> + vas = vaMapBuffer(vpp_ctx->hwctx->display, vpp_ctx->filter_buffers[0],
> + (void**)&hdrtm_param);
> + if (vas != VA_STATUS_SUCCESS) {
> + av_log(avctx, AV_LOG_ERROR, "Failed to map "
> + "buffer (%d): %d (%s).\n",
> + vpp_ctx->filter_buffers[0], vas, vaErrorStr(vas));
> + return AVERROR(EIO);
> + }
> +
> + memcpy(hdrtm_param->data.metadata, &ctx->in_metadata,
> sizeof(VAHdrMetaDataHDR10));
> +
> + vas = vaUnmapBuffer(vpp_ctx->hwctx->display,
> vpp_ctx->filter_buffers[0]);
> + if (vas != VA_STATUS_SUCCESS) {
> + av_log(avctx, AV_LOG_ERROR, "Failed to unmap output buffers: "
> + "%d (%s).\n", vas, vaErrorStr(vas));
> + return AVERROR(EIO);
> + }
> +
> + return 0;
> +}
> +
> +static int tonemap_vaapi_build_filter_params(AVFilterContext *avctx)
> +{
> + VAAPIVPPContext *vpp_ctx = avctx->priv;
> + HDRVAAPIContext *ctx = avctx->priv;
> + VAStatus vas;
> + VAProcFilterParameterBufferHDRToneMapping hdrtm_param;
> + VAProcFilterCapHighDynamicRange
> hdr_cap[VAProcHighDynamicRangeMetadataTypeCount];
> + int num_query_caps;
> + int i;
> +
> + memset(&hdrtm_param, 0, sizeof(hdrtm_param));
> + memset(&ctx->in_metadata, 0, sizeof(ctx->in_metadata));
> +
> + num_query_caps = VAProcHighDynamicRangeMetadataTypeCount;
> + vas = vaQueryVideoProcFilterCaps(vpp_ctx->hwctx->display,
> + vpp_ctx->va_context,
> +
> VAProcFilterHighDynamicRangeToneMapping,
> + &hdr_cap, &num_query_caps);
> + if (vas != VA_STATUS_SUCCESS) {
> + av_log(avctx, AV_LOG_ERROR, "Failed to query HDR caps "
> + "context: %d (%s).\n", vas, vaErrorStr(vas));
> + return AVERROR(EIO);
> + }
> +
> + for (i = 0; i < num_query_caps; i++) {
> + if (hdr_cap[i].metadata_type !=
> VAProcHighDynamicRangeMetadataNone)
> + break;
> + }
> +
> + if (i >= num_query_caps) {
> + av_log(avctx, AV_LOG_ERROR, "VAAPI driver doesn't support HDR\n");
> + return AVERROR(EINVAL);
> + }
> +
> + for (i = 0; i < num_query_caps; i++) {
> + if (VA_TONE_MAPPING_HDR_TO_SDR & hdr_cap[i].caps_flag)
> + break;
> + }
> +
> + if (i >= num_query_caps) {
> + av_log(avctx, AV_LOG_ERROR,
> + "VAAPI driver doesn't support HDR to SDR\n");
> + return AVERROR(EINVAL);
> + }
> +
> + hdrtm_param.type = VAProcFilterHighDynamicRangeToneMapping;
> + hdrtm_param.data.metadata_type = VAProcHighDynamicRangeMetadataHDR10;
> + hdrtm_param.data.metadata = &ctx->in_metadata;
> + hdrtm_param.data.metadata_size = sizeof(VAHdrMetaDataHDR10);
> +
> + return ff_vaapi_vpp_make_param_buffers(avctx,
> +
> VAProcFilterParameterBufferType,
> + &hdrtm_param,
> sizeof(hdrtm_param), 1);
> +}
> +
> +static int tonemap_vaapi_filter_frame(AVFilterLink *inlink, AVFrame
> *input_frame)
> +{
> + AVFilterContext *avctx = inlink->dst;
> + AVFilterLink *outlink = avctx->outputs[0];
> + VAAPIVPPContext *vpp_ctx = avctx->priv;
> + HDRVAAPIContext *ctx = avctx->priv;
> + AVFrame *output_frame = NULL;
> + VASurfaceID input_surface, output_surface;
> +
> + VAProcPipelineParameterBuffer params;
> + int err;
> +
> + av_log(avctx, AV_LOG_DEBUG, "Filter input: %s, %ux%u (%"PRId64").\n",
> + av_get_pix_fmt_name(input_frame->format),
> + input_frame->width, input_frame->height, input_frame->pts);
> +
> + if (vpp_ctx->va_context == VA_INVALID_ID){
> + av_frame_free(&input_frame);
> + return AVERROR(EINVAL);
> + }
> +
> + err = tonemap_vaapi_save_metadata(avctx, input_frame);
> + if (err < 0)
> + goto fail;
> +
> + err = tonemap_vaapi_set_filter_params(avctx, input_frame);
> + if (err < 0)
> + goto fail;
> +
> + input_surface = (VASurfaceID)(uintptr_t)input_frame->data[3];
> + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for tonemap vpp
> input.\n",
> + input_surface);
> +
> + output_frame = ff_get_video_buffer(outlink, vpp_ctx->output_width,
> + vpp_ctx->output_height);
> + if (!output_frame) {
> + err = AVERROR(ENOMEM);
> + goto fail;
> + }
> +
> + output_surface = (VASurfaceID)(uintptr_t)output_frame->data[3];
> + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for tonemap vpp
> output.\n",
> + output_surface);
> + memset(¶ms, 0, sizeof(params));
> +
> + err = av_frame_copy_props(output_frame, input_frame);
> + if (err < 0)
> + return err;
> +
> + if (ctx->color_primaries != AVCOL_PRI_UNSPECIFIED)
> + output_frame->color_primaries = ctx->color_primaries;
> +
> + if (ctx->color_transfer != AVCOL_TRC_UNSPECIFIED)
> + output_frame->color_trc = ctx->color_transfer;
> + else
> + output_frame->color_trc = AVCOL_TRC_BT709
>
why does only this setting get special treatment?
> + if (ctx->color_matrix != AVCOL_SPC_UNSPECIFIED)
> + output_frame->colorspace = ctx->color_matrix;
> +
> + err = ff_vaapi_vpp_init_params(avctx, ¶ms,
> + input_frame, output_frame);
> + if (err < 0)
> + goto fail;
> +
> + err = ff_vaapi_vpp_render_picture(avctx, ¶ms, output_frame);
> + if (err < 0)
> + goto fail;
> +
> + av_frame_free(&input_frame);
> +
> + av_log(avctx, AV_LOG_DEBUG, "Filter output: %s, %ux%u (%"PRId64").\n",
> + av_get_pix_fmt_name(output_frame->format),
> + output_frame->width, output_frame->height, output_frame->pts);
> +
> + return ff_filter_frame(outlink, output_frame);
> +
> +fail:
> + av_frame_free(&input_frame);
> + av_frame_free(&output_frame);
> + return err;
> +}
> +
> +static av_cold int tonemap_vaapi_init(AVFilterContext *avctx)
> +{
> + VAAPIVPPContext *vpp_ctx = avctx->priv;
> + HDRVAAPIContext *ctx = avctx->priv;
> +
> + ff_vaapi_vpp_ctx_init(avctx);
> + vpp_ctx->build_filter_params = tonemap_vaapi_build_filter_params;
> + vpp_ctx->pipeline_uninit = ff_vaapi_vpp_pipeline_uninit;
> +
> + if (ctx->output_format_string) {
> + vpp_ctx->output_format =
> av_get_pix_fmt(ctx->output_format_string);
> + switch (vpp_ctx->output_format) {
> + case AV_PIX_FMT_NV12:
> + case AV_PIX_FMT_P010:
> + break;
> + default:
> + av_log(avctx, AV_LOG_ERROR, "Invalid output format.\n");
> + return AVERROR(EINVAL);
> + }
> + } else {
> + vpp_ctx->output_format = AV_PIX_FMT_NV12;
> + av_log(avctx, AV_LOG_WARNING, "Output format not set, use default
> format NV12\n");
> + }
> +
> +#define STRING_OPTION(var_name, func_name, default_value) do { \
> + if (ctx->var_name ## _string) { \
> + int var = av_ ## func_name ## _from_name(ctx->var_name ##
> _string); \
> + if (var < 0) { \
> + av_log(avctx, AV_LOG_ERROR, "Invalid %s.\n", #var_name); \
> + return AVERROR(EINVAL); \
> + } \
> + ctx->var_name = var; \
> + } else { \
> + ctx->var_name = default_value; \
> + } \
> + } while (0)
> +
> + STRING_OPTION(color_primaries, color_primaries,
> AVCOL_PRI_UNSPECIFIED);
> + STRING_OPTION(color_transfer, color_transfer,
> AVCOL_TRC_UNSPECIFIED);
> + STRING_OPTION(color_matrix, color_space,
> AVCOL_SPC_UNSPECIFIED);
> +
> + return 0;
> +}
> +
> +#define OFFSET(x) offsetof(HDRVAAPIContext, x)
> +#define FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_FILTERING_PARAM)
> +static const AVOption tonemap_vaapi_options[] = {
> + { "format", "Output pixel format set", OFFSET(output_format_string),
> AV_OPT_TYPE_STRING, .flags = FLAGS, "format" },
> + { "matrix", "Output color matrix coefficient set",
> + OFFSET(color_matrix_string), AV_OPT_TYPE_STRING,
> + { .str = NULL }, .flags = FLAGS, "matrix" },
> + { "m", "Output color matrix coefficient set",
> + OFFSET(color_matrix_string), AV_OPT_TYPE_STRING,
> + { .str = NULL }, .flags = FLAGS, "matrix" },
> + { "primaries", "Output color primaries set",
> + OFFSET(color_primaries_string), AV_OPT_TYPE_STRING,
> + { .str = NULL }, .flags = FLAGS, "primaries" },
> + { "p", "Output color primaries set",
> + OFFSET(color_primaries_string), AV_OPT_TYPE_STRING,
> + { .str = NULL }, .flags = FLAGS, "primaries" },
> + { "transfer", "Output color transfer characteristics set",
> + OFFSET(color_transfer_string), AV_OPT_TYPE_STRING,
> + { .str = NULL }, .flags = FLAGS, "transfer" },
> + { "t", "Output color transfer characteristics set",
> + OFFSET(color_transfer_string), AV_OPT_TYPE_STRING,
> + { .str = NULL }, .flags = FLAGS, "transfer" },
> + { NULL }
> +};
> +
> +
> +AVFILTER_DEFINE_CLASS(tonemap_vaapi);
> +
> +static const AVFilterPad tonemap_vaapi_inputs[] = {
> + {
> + .name = "default",
> + .type = AVMEDIA_TYPE_VIDEO,
> + .filter_frame = &tonemap_vaapi_filter_frame,
> + .config_props = &ff_vaapi_vpp_config_input,
> + },
> + { NULL }
> +};
> +
> +static const AVFilterPad tonemap_vaapi_outputs[] = {
> + {
> + .name = "default",
> + .type = AVMEDIA_TYPE_VIDEO,
> + .config_props = &ff_vaapi_vpp_config_output,
> + },
> + { NULL }
> +};
> +
> +AVFilter ff_vf_tonemap_vaapi = {
> + .name = "tonemap_vaapi",
> + .description = NULL_IF_CONFIG_SMALL("VAAPI VPP for tone-mapping"),
> + .priv_size = sizeof(HDRVAAPIContext),
> + .init = &tonemap_vaapi_init,
> + .uninit = &ff_vaapi_vpp_ctx_uninit,
> + .query_formats = &ff_vaapi_vpp_query_formats,
> + .inputs = tonemap_vaapi_inputs,
> + .outputs = tonemap_vaapi_outputs,
> + .priv_class = &tonemap_vaapi_class,
> + .flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE,
> +};
>
Overall this lgtm, I'd push it but I don't have a platform to test it on.
--
Vittorio
More information about the ffmpeg-devel
mailing list