diff options
author | Rémi Denis-Courmont <remi@remlab.net> | 2024-06-03 19:11:33 +0300 |
---|---|---|
committer | Rémi Denis-Courmont <remi@remlab.net> | 2024-06-07 17:53:05 +0300 |
commit | a169f3bca5937cf71d61de1fb85b522bd69c6cfc (patch) | |
tree | 9af7d306246820eaba02e83f653a92c9a1de8e2f | |
parent | 04397a29deccf72575a7b02ed56eb37bef2934cc (diff) | |
download | ffmpeg-a169f3bca5937cf71d61de1fb85b522bd69c6cfc.tar.gz |
lavc/vc1dsp: R-V V vc1_inv_trans_8x4
T-Head C908 (cycles):
vc1dsp.vc1_inv_trans_8x4_c: 626.2
vc1dsp.vc1_inv_trans_8x4_rvv_i32: 215.2
-rw-r--r-- | libavcodec/riscv/vc1dsp_init.c | 2 | ||||
-rw-r--r-- | libavcodec/riscv/vc1dsp_rvv.S | 73 |
2 files changed, 75 insertions, 0 deletions
diff --git a/libavcodec/riscv/vc1dsp_init.c b/libavcodec/riscv/vc1dsp_init.c index b8a1015ce5..e63870ad44 100644 --- a/libavcodec/riscv/vc1dsp_init.c +++ b/libavcodec/riscv/vc1dsp_init.c @@ -29,6 +29,7 @@ void ff_vc1_inv_trans_8x8_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block void ff_vc1_inv_trans_8x8_rvv(int16_t block[64]); void ff_vc1_inv_trans_4x8_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block); void ff_vc1_inv_trans_8x4_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block); +void ff_vc1_inv_trans_8x4_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block); void ff_vc1_inv_trans_4x4_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block); void ff_put_pixels16x16_rvi(uint8_t *dst, const uint8_t *src, ptrdiff_t line_size, int rnd); void ff_put_pixels8x8_rvi(uint8_t *dst, const uint8_t *src, ptrdiff_t line_size, int rnd); @@ -55,6 +56,7 @@ av_cold void ff_vc1dsp_init_riscv(VC1DSPContext *dsp) if (flags & AV_CPU_FLAG_RVV_I32) { if (ff_rv_vlen_least(128)) { dsp->vc1_inv_trans_8x8 = ff_vc1_inv_trans_8x8_rvv; + dsp->vc1_inv_trans_8x4 = ff_vc1_inv_trans_8x4_rvv; dsp->vc1_inv_trans_4x8_dc = ff_vc1_inv_trans_4x8_dc_rvv; dsp->vc1_inv_trans_4x4_dc = ff_vc1_inv_trans_4x4_dc_rvv; dsp->avg_vc1_mspel_pixels_tab[0][0] = ff_avg_pixels16x16_rvv; diff --git a/libavcodec/riscv/vc1dsp_rvv.S b/libavcodec/riscv/vc1dsp_rvv.S index e15783d113..d003185ade 100644 --- a/libavcodec/riscv/vc1dsp_rvv.S +++ b/libavcodec/riscv/vc1dsp_rvv.S @@ -173,6 +173,31 @@ func ff_vc1_inv_trans_8_rvv, zve32x jr t0 endfunc + .variant_cc ff_vc1_inv_trans_4_rvv +func ff_vc1_inv_trans_4_rvv, zve32x + li t3, 17 + vmul.vx v8, v0, t3 + li t4, 22 + vmul.vx v10, v2, t3 + li t2, 10 + vmul.vx v14, v1, t4 + vadd.vv v24, v8, v10 # t1 + vsub.vv v25, v8, v10 # t2 + vmul.vx v16, v3, t2 + vmul.vx v18, v3, t4 + vmul.vx v20, v1, t2 + vadd.vv v26, v14, v16 # t3 + vsub.vv v27, v18, v20 # t4 + vadd.vv v0, v24, v26 + vsub.vv v1, v25, v27 + vadd.vv v2, v25, v27 + vsub.vv v3, v24, v26 + .irp n,0,1,2,3 + vssra.vx v\n, v\n, t1 # + 4 >> 3 or + 64 >> 7 + .endr + jr t0 +endfunc + func ff_vc1_inv_trans_8x8_rvv, zve32x csrwi vxrm, 0 vsetivli zero, 8, e16, m1, ta, ma @@ -223,6 +248,54 @@ func ff_vc1_inv_trans_8x8_rvv, zve32x ret endfunc +func ff_vc1_inv_trans_8x4_rvv, zve32x + csrwi vxrm, 0 + vsetivli zero, 4, e16, mf2, ta, ma + vlseg8e16.v v0, (a2) + jal t0, ff_vc1_inv_trans_8_rvv + vsseg8e16.v v0, (a2) + addi a3, a2, 1 * 8 * 2 + vsetivli zero, 8, e16, m1, ta, ma + vle16.v v0, (a2) + addi a4, a2, 2 * 8 * 2 + vle16.v v1, (a3) + addi a5, a2, 3 * 8 * 2 + vle16.v v2, (a4) + vle16.v v3, (a5) + .irp n,0,1,2,3 + # shift 4 vectors of 8 elems after transpose instead of 8 of 4 + vssra.vi v\n, v\n, 3 + .endr + li t1, 7 + jal t0, ff_vc1_inv_trans_4_rvv + add a3, a1, a0 + vle8.v v8, (a0) + add a4, a1, a3 + vle8.v v9, (a3) + add a5, a1, a4 + vle8.v v10, (a4) + vle8.v v11, (a5) + vsetvli zero, zero, e8, mf2, ta, ma + vwaddu.wv v0, v0, v8 + vwaddu.wv v1, v1, v9 + vwaddu.wv v2, v2, v10 + vwaddu.wv v3, v3, v11 + vsetvli zero, zero, e16, m1, ta, ma + .irp n,0,1,2,3 + vmax.vx v\n, v\n, zero + .endr + vsetvli zero, zero, e8, mf2, ta, ma + vnclipu.wi v8, v0, 0 + vnclipu.wi v9, v1, 0 + vse8.v v8, (a0) + vnclipu.wi v10, v2, 0 + vse8.v v9, (a3) + vnclipu.wi v11, v3, 0 + vse8.v v10, (a4) + vse8.v v11, (a5) + ret +endfunc + .macro mspel_op op pos n1 n2 add t1, \pos, a2 v\op\()e8.v v\n1, (\pos) |