[FFmpeg-devel] [PATCH V2 3/3] avfilter/vf_dnn_processing.c: add frame size change support for planar yuv format

mypopy at gmail.com mypopy at gmail.com
Fri Mar 6 04:49:10 EET 2020


On Tue, Feb 25, 2020 at 5:24 PM Guo, Yejun <yejun.guo at intel.com> wrote:
>
> The Y channel is handled by dnn, and also resized by dnn. The UV channels
> are resized with swscale.
For me, this is a little weird to resize Y with dnn backend but resize
UV channel with FFmpeg swscale, is it used the same scale algorithm ?

> The command to use espcn.pb (see vf_sr) looks like:
> ./ffmpeg -i 480p.jpg -vf format=yuv420p,dnn_processing=dnn_backend=tensorflow:model=espcn.pb:input=x:output=y -y tmp.espcn.jpg
>
> Signed-off-by: Guo, Yejun <yejun.guo at intel.com>
> ---
>  doc/filters.texi                |  9 +++++++++
>  libavfilter/vf_dnn_processing.c | 37 ++++++++++++++++++++++++++++++-------
>  2 files changed, 39 insertions(+), 7 deletions(-)
>
> diff --git a/doc/filters.texi b/doc/filters.texi
> index 33b7857..e3df8f9 100644
> --- a/doc/filters.texi
> +++ b/doc/filters.texi
> @@ -9155,6 +9155,7 @@ ffmpeg -i INPUT -f lavfi -i nullsrc=hd720,geq='r=128+80*(sin(sqrt((X-W/2)*(X-W/2
>  @end example
>  @end itemize
>
> + at anchor{dnn_processing}
>  @section dnn_processing
>
>  Do image processing with deep neural networks. It works together with another filter
> @@ -9216,6 +9217,12 @@ Handle the Y channel with srcnn.pb (see @ref{sr} filter) for frame with yuv420p
>  ./ffmpeg -i 480p.jpg -vf format=yuv420p,scale=w=iw*2:h=ih*2,dnn_processing=dnn_backend=tensorflow:model=srcnn.pb:input=x:output=y -y srcnn.jpg
>  @end example
>
> + at item
> +Handle the Y channel with espcn.pb (see @ref{sr} filter), which changes frame size, for format yuv420p (planar YUV formats supported):
> + at example
> +./ffmpeg -i 480p.jpg -vf format=yuv420p,dnn_processing=dnn_backend=tensorflow:model=espcn.pb:input=x:output=y -y tmp.espcn.jpg
> + at end example
> +
>  @end itemize
>
>  @section drawbox
> @@ -17369,6 +17376,8 @@ Default value is @code{2}. Scale factor is necessary for SRCNN model, because it
>  input upscaled using bicubic upscaling with proper scale factor.
>  @end table
>
> +This feature can also be finished with @ref{dnn_processing} filter.
> +
>  @section ssim
>
>  Obtain the SSIM (Structural SImilarity Metric) between two input videos.
> diff --git a/libavfilter/vf_dnn_processing.c b/libavfilter/vf_dnn_processing.c
> index f9458f0..7f40f85 100644
> --- a/libavfilter/vf_dnn_processing.c
> +++ b/libavfilter/vf_dnn_processing.c
> @@ -51,6 +51,8 @@ typedef struct DnnProcessingContext {
>
>      struct SwsContext *sws_gray8_to_grayf32;
>      struct SwsContext *sws_grayf32_to_gray8;
> +    struct SwsContext *sws_uv_scale;
> +    int sws_uv_height;
>  } DnnProcessingContext;
>
>  #define OFFSET(x) offsetof(DnnProcessingContext, x)
> @@ -274,6 +276,18 @@ static int prepare_sws_context(AVFilterLink *outlink)
>                                                     outlink->h,
>                                                     AV_PIX_FMT_GRAY8,
>                                                     0, NULL, NULL, NULL);
> +
> +        if (inlink->w != outlink->w || inlink->h != outlink->h) {
> +            const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(fmt);
> +            int sws_src_h = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
> +            int sws_src_w = AV_CEIL_RSHIFT(inlink->w, desc->log2_chroma_w);
> +            int sws_dst_h = AV_CEIL_RSHIFT(outlink->h, desc->log2_chroma_h);
> +            int sws_dst_w = AV_CEIL_RSHIFT(outlink->w, desc->log2_chroma_w);
> +            ctx->sws_uv_scale = sws_getContext(sws_src_w, sws_src_h, AV_PIX_FMT_GRAY8,
> +                                               sws_dst_w, sws_dst_h, AV_PIX_FMT_GRAY8,
> +                                               SWS_BICUBIC, NULL, NULL, NULL);
> +            ctx->sws_uv_height = sws_src_h;
> +        }
>          return 0;
>      default:
>          //do nothing
> @@ -404,13 +418,21 @@ static av_always_inline int isPlanarYUV(enum AVPixelFormat pix_fmt)
>
>  static int copy_uv_planes(DnnProcessingContext *ctx, AVFrame *out, const AVFrame *in)
>  {
> -    const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(in->format);
> -    int uv_height = AV_CEIL_RSHIFT(in->height, desc->log2_chroma_h);
> -    for (int i = 1; i < 3; ++i) {
> -        int bytewidth = av_image_get_linesize(in->format, in->width, i);
> -        av_image_copy_plane(out->data[i], out->linesize[i],
> -                            in->data[i], in->linesize[i],
> -                            bytewidth, uv_height);
> +    if (!ctx->sws_uv_scale) {
> +        av_assert0(in->height == out->height && in->width == out->width);
> +        const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(in->format);
> +        int uv_height = AV_CEIL_RSHIFT(in->height, desc->log2_chroma_h);
> +        for (int i = 1; i < 3; ++i) {
> +            int bytewidth = av_image_get_linesize(in->format, in->width, i);
> +            av_image_copy_plane(out->data[i], out->linesize[i],
> +                                in->data[i], in->linesize[i],
> +                                bytewidth, uv_height);
> +        }
> +    } else {
> +        sws_scale(ctx->sws_uv_scale, (const uint8_t **)(in->data + 1), in->linesize + 1,
> +                  0, ctx->sws_uv_height, out->data + 1, out->linesize + 1);
> +        sws_scale(ctx->sws_uv_scale, (const uint8_t **)(in->data + 2), in->linesize + 2,
> +                  0, ctx->sws_uv_height, out->data + 2, out->linesize + 2);
>      }
>
>      return 0;
> @@ -455,6 +477,7 @@ static av_cold void uninit(AVFilterContext *ctx)
>
>      sws_freeContext(context->sws_gray8_to_grayf32);
>      sws_freeContext(context->sws_grayf32_to_gray8);
> +    sws_freeContext(context->sws_uv_scale);
>
>      if (context->dnn_module)
>          (context->dnn_module->free_model)(&context->model);
> --
> 2.7.4
>


More information about the ffmpeg-devel mailing list