aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorRémi Denis-Courmont <remi@remlab.net>2024-06-03 19:11:33 +0300
committerRémi Denis-Courmont <remi@remlab.net>2024-06-07 17:53:05 +0300
commita169f3bca5937cf71d61de1fb85b522bd69c6cfc (patch)
tree9af7d306246820eaba02e83f653a92c9a1de8e2f
parent04397a29deccf72575a7b02ed56eb37bef2934cc (diff)
downloadffmpeg-a169f3bca5937cf71d61de1fb85b522bd69c6cfc.tar.gz
lavc/vc1dsp: R-V V vc1_inv_trans_8x4
T-Head C908 (cycles): vc1dsp.vc1_inv_trans_8x4_c: 626.2 vc1dsp.vc1_inv_trans_8x4_rvv_i32: 215.2
-rw-r--r--libavcodec/riscv/vc1dsp_init.c2
-rw-r--r--libavcodec/riscv/vc1dsp_rvv.S73
2 files changed, 75 insertions, 0 deletions
diff --git a/libavcodec/riscv/vc1dsp_init.c b/libavcodec/riscv/vc1dsp_init.c
index b8a1015ce5..e63870ad44 100644
--- a/libavcodec/riscv/vc1dsp_init.c
+++ b/libavcodec/riscv/vc1dsp_init.c
@@ -29,6 +29,7 @@ void ff_vc1_inv_trans_8x8_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block
void ff_vc1_inv_trans_8x8_rvv(int16_t block[64]);
void ff_vc1_inv_trans_4x8_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block);
void ff_vc1_inv_trans_8x4_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block);
+void ff_vc1_inv_trans_8x4_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block);
void ff_vc1_inv_trans_4x4_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block);
void ff_put_pixels16x16_rvi(uint8_t *dst, const uint8_t *src, ptrdiff_t line_size, int rnd);
void ff_put_pixels8x8_rvi(uint8_t *dst, const uint8_t *src, ptrdiff_t line_size, int rnd);
@@ -55,6 +56,7 @@ av_cold void ff_vc1dsp_init_riscv(VC1DSPContext *dsp)
if (flags & AV_CPU_FLAG_RVV_I32) {
if (ff_rv_vlen_least(128)) {
dsp->vc1_inv_trans_8x8 = ff_vc1_inv_trans_8x8_rvv;
+ dsp->vc1_inv_trans_8x4 = ff_vc1_inv_trans_8x4_rvv;
dsp->vc1_inv_trans_4x8_dc = ff_vc1_inv_trans_4x8_dc_rvv;
dsp->vc1_inv_trans_4x4_dc = ff_vc1_inv_trans_4x4_dc_rvv;
dsp->avg_vc1_mspel_pixels_tab[0][0] = ff_avg_pixels16x16_rvv;
diff --git a/libavcodec/riscv/vc1dsp_rvv.S b/libavcodec/riscv/vc1dsp_rvv.S
index e15783d113..d003185ade 100644
--- a/libavcodec/riscv/vc1dsp_rvv.S
+++ b/libavcodec/riscv/vc1dsp_rvv.S
@@ -173,6 +173,31 @@ func ff_vc1_inv_trans_8_rvv, zve32x
jr t0
endfunc
+ .variant_cc ff_vc1_inv_trans_4_rvv
+func ff_vc1_inv_trans_4_rvv, zve32x
+ li t3, 17
+ vmul.vx v8, v0, t3
+ li t4, 22
+ vmul.vx v10, v2, t3
+ li t2, 10
+ vmul.vx v14, v1, t4
+ vadd.vv v24, v8, v10 # t1
+ vsub.vv v25, v8, v10 # t2
+ vmul.vx v16, v3, t2
+ vmul.vx v18, v3, t4
+ vmul.vx v20, v1, t2
+ vadd.vv v26, v14, v16 # t3
+ vsub.vv v27, v18, v20 # t4
+ vadd.vv v0, v24, v26
+ vsub.vv v1, v25, v27
+ vadd.vv v2, v25, v27
+ vsub.vv v3, v24, v26
+ .irp n,0,1,2,3
+ vssra.vx v\n, v\n, t1 # + 4 >> 3 or + 64 >> 7
+ .endr
+ jr t0
+endfunc
+
func ff_vc1_inv_trans_8x8_rvv, zve32x
csrwi vxrm, 0
vsetivli zero, 8, e16, m1, ta, ma
@@ -223,6 +248,54 @@ func ff_vc1_inv_trans_8x8_rvv, zve32x
ret
endfunc
+func ff_vc1_inv_trans_8x4_rvv, zve32x
+ csrwi vxrm, 0
+ vsetivli zero, 4, e16, mf2, ta, ma
+ vlseg8e16.v v0, (a2)
+ jal t0, ff_vc1_inv_trans_8_rvv
+ vsseg8e16.v v0, (a2)
+ addi a3, a2, 1 * 8 * 2
+ vsetivli zero, 8, e16, m1, ta, ma
+ vle16.v v0, (a2)
+ addi a4, a2, 2 * 8 * 2
+ vle16.v v1, (a3)
+ addi a5, a2, 3 * 8 * 2
+ vle16.v v2, (a4)
+ vle16.v v3, (a5)
+ .irp n,0,1,2,3
+ # shift 4 vectors of 8 elems after transpose instead of 8 of 4
+ vssra.vi v\n, v\n, 3
+ .endr
+ li t1, 7
+ jal t0, ff_vc1_inv_trans_4_rvv
+ add a3, a1, a0
+ vle8.v v8, (a0)
+ add a4, a1, a3
+ vle8.v v9, (a3)
+ add a5, a1, a4
+ vle8.v v10, (a4)
+ vle8.v v11, (a5)
+ vsetvli zero, zero, e8, mf2, ta, ma
+ vwaddu.wv v0, v0, v8
+ vwaddu.wv v1, v1, v9
+ vwaddu.wv v2, v2, v10
+ vwaddu.wv v3, v3, v11
+ vsetvli zero, zero, e16, m1, ta, ma
+ .irp n,0,1,2,3
+ vmax.vx v\n, v\n, zero
+ .endr
+ vsetvli zero, zero, e8, mf2, ta, ma
+ vnclipu.wi v8, v0, 0
+ vnclipu.wi v9, v1, 0
+ vse8.v v8, (a0)
+ vnclipu.wi v10, v2, 0
+ vse8.v v9, (a3)
+ vnclipu.wi v11, v3, 0
+ vse8.v v10, (a4)
+ vse8.v v11, (a5)
+ ret
+endfunc
+
.macro mspel_op op pos n1 n2
add t1, \pos, a2
v\op\()e8.v v\n1, (\pos)