[FFmpeg-devel] [PATCH] lavc/vc1dsp: R-V V mspel_pixels

Rémi Denis-Courmont remi at remlab.net
Sat May 4 20:53:33 EEST 2024


Le lauantaina 4. toukokuuta 2024, 13.01.05 EEST uk7b at foxmail.com a écrit :
> From: sunyuechi <sunyuechi at iscas.ac.cn>
> 
> vc1dsp.avg_vc1_mspel_pixels_tab[0][0]_c: 869.7
> vc1dsp.avg_vc1_mspel_pixels_tab[0][0]_rvv_i32: 148.7
> vc1dsp.avg_vc1_mspel_pixels_tab[1][0]_c: 220.5
> vc1dsp.avg_vc1_mspel_pixels_tab[1][0]_rvv_i64: 56.2
> vc1dsp.put_vc1_mspel_pixels_tab[0][0]_c: 523.7
> vc1dsp.put_vc1_mspel_pixels_tab[0][0]_rvv_i32: 82.0
> vc1dsp.put_vc1_mspel_pixels_tab[1][0]_c: 138.5
> vc1dsp.put_vc1_mspel_pixels_tab[1][0]_rvv_i64: 23.7
> ---
>  libavcodec/riscv/vc1dsp_init.c |  8 +++++
>  libavcodec/riscv/vc1dsp_rvv.S  | 66 ++++++++++++++++++++++++++++++++++
>  2 files changed, 74 insertions(+)
> 
> diff --git a/libavcodec/riscv/vc1dsp_init.c b/libavcodec/riscv/vc1dsp_init.c
> index e47b644f80..610c43a1a3 100644
> --- a/libavcodec/riscv/vc1dsp_init.c
> +++ b/libavcodec/riscv/vc1dsp_init.c
> @@ -29,6 +29,10 @@ void ff_vc1_inv_trans_8x8_dc_rvv(uint8_t *dest, ptrdiff_t
> stride, int16_t *block void ff_vc1_inv_trans_4x8_dc_rvv(uint8_t *dest,
> ptrdiff_t stride, int16_t *block); void ff_vc1_inv_trans_8x4_dc_rvv(uint8_t
> *dest, ptrdiff_t stride, int16_t *block); void
> ff_vc1_inv_trans_4x4_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t
> *block); +void ff_put_pixels16x16_rvv(uint8_t *dst, const uint8_t *src,
> ptrdiff_t line_size, int rnd); +void ff_put_pixels8x8_rvv(uint8_t *dst,
> const uint8_t *src, ptrdiff_t line_size, int rnd); +void
> ff_avg_pixels16x16_rvv(uint8_t *dst, const uint8_t *src, ptrdiff_t
> line_size, int rnd); +void ff_avg_pixels8x8_rvv(uint8_t *dst, const uint8_t
> *src, ptrdiff_t line_size, int rnd);
> 
>  av_cold void ff_vc1dsp_init_riscv(VC1DSPContext *dsp)
>  {
> @@ -38,9 +42,13 @@ av_cold void ff_vc1dsp_init_riscv(VC1DSPContext *dsp)
>      if (flags & AV_CPU_FLAG_RVV_I32 && ff_get_rv_vlenb() >= 16) {
>          dsp->vc1_inv_trans_4x8_dc = ff_vc1_inv_trans_4x8_dc_rvv;
>          dsp->vc1_inv_trans_4x4_dc = ff_vc1_inv_trans_4x4_dc_rvv;
> +        dsp->put_vc1_mspel_pixels_tab[0][0] = ff_put_pixels16x16_rvv;
> +        dsp->avg_vc1_mspel_pixels_tab[0][0] = ff_avg_pixels16x16_rvv;
>          if (flags & AV_CPU_FLAG_RVV_I64) {
>              dsp->vc1_inv_trans_8x8_dc = ff_vc1_inv_trans_8x8_dc_rvv;
>              dsp->vc1_inv_trans_8x4_dc = ff_vc1_inv_trans_8x4_dc_rvv;
> +            dsp->put_vc1_mspel_pixels_tab[1][0] = ff_put_pixels8x8_rvv;
> +            dsp->avg_vc1_mspel_pixels_tab[1][0] = ff_avg_pixels8x8_rvv;
>          }
>      }
>  #endif
> diff --git a/libavcodec/riscv/vc1dsp_rvv.S b/libavcodec/riscv/vc1dsp_rvv.S
> index 4a00945ead..48244f91aa 100644
> --- a/libavcodec/riscv/vc1dsp_rvv.S
> +++ b/libavcodec/riscv/vc1dsp_rvv.S
> @@ -111,3 +111,69 @@ func ff_vc1_inv_trans_4x4_dc_rvv, zve32x
>          vsse32.v      v0, (a0), a1
>          ret
>  endfunc
> +
> +func ff_put_pixels16x16_rvv, zve32x
> +        vsetivli      zero, 16, e8, m1, ta, ma
> +        .irp n 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30
> +        vle8.v        v\n, (a1)
> +        add           a1, a1, a2
> +        .endr
> +        vle8.v        v31, (a1)

Is it not faster to compute the address ahead of time, e.g.:

add t1, a2, a1
vle8.v vN, (a1)
sh1add a1, a2, a1
vle8.v vN+1, (t1)

...and so on? Even on a reordering core, you can't eliminate stall on data 
dependency if there is nothing else to be done.

(Ditto below and in other patches.)

> +        .irp n 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30
> +        vse8.v        v\n, (a0)
> +        add           a0, a0, a2
> +        .endr
> +        vse8.v        v31, (a0)
> +
> +        ret
> +endfunc
> +
> +func ff_put_pixels8x8_rvv, zve64x
> +        vsetivli      zero, 8, e8, mf2, ta, ma
> +        vlse64.v      v8, (a1), a2
> +        vsse64.v      v8, (a0), a2

Copying 64-bit quantities should not need RVV at all. Maybe the C version 
needs to be improved instead, but if that is not possible, then an RVI version 
may be more portable and work just as well.

> +
> +        ret
> +endfunc
> +
> +func ff_avg_pixels16x16_rvv, zve32x
> +        csrwi         vxrm, 0
> +        vsetivli      zero, 16, e8, m1, ta, ma
> +        li            t0, 128
> +
> +        .irp n 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30
> +        vle8.v        v\n, (a1)
> +        add           a1, a1, a2
> +        .endr
> +        vle8.v        v31, (a1)
> +        .irp n 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14
> +        vle8.v        v\n, (a0)
> +        add           a0, a0, a2
> +        .endr
> +        vle8.v        v15, (a0)
> +        vsetvli       zero, t0, e8, m8, ta, ma
> +        vaaddu.vv     v0, v0, v16
> +        vaaddu.vv     v8, v8, v24
> +        vsetivli      zero, 16, e8, m1, ta, ma
> +        .irp n  15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1
> +        vse8.v        v\n, (a0)
> +        sub           a0, a0, a2
> +        .endr
> +        vse8.v        v0, (a0)
> +
> +        ret
> +endfunc
> +
> +func ff_avg_pixels8x8_rvv, zve64x
> +        csrwi         vxrm, 0
> +        li            t0, 64
> +        vsetivli      zero, 8, e8, mf2, ta, ma

Does MF2 actually improve perfs over M1 here?

> +        vlse64.v      v16, (a1), a2
> +        vlse64.v      v8, (a0), a2
> +        vsetvli       zero, t0, e8, m4, ta, ma
> +        vaaddu.vv     v16, v16, v8
> +        vsetivli      zero, 8, e8, mf2, ta, ma
> +        vsse64.v      v16, (a0), a2
> +
> +        ret
> +endfunc


-- 
レミ・デニ-クールモン
http://www.remlab.net/





More information about the ffmpeg-devel mailing list