diff options
author | Rémi Denis-Courmont <remi@remlab.net> | 2024-07-27 18:34:58 +0300 |
---|---|---|
committer | Rémi Denis-Courmont <remi@remlab.net> | 2024-07-30 18:41:51 +0300 |
commit | 784a72a116c29a38ec55ce81e165a29bdd00c497 (patch) | |
tree | 5585dacea8cc3e0922f1e5ad32bb0e958b704488 /libavcodec | |
parent | bd0c3edb136fb79a8d99fd0dfc50c15e40a2355b (diff) | |
download | ffmpeg-784a72a116c29a38ec55ce81e165a29bdd00c497.tar.gz |
lavc/vc1dsp: unify R-V V DC bypass functions
Diffstat (limited to 'libavcodec')
-rw-r--r-- | libavcodec/riscv/vc1dsp_rvv.S | 126 |
1 files changed, 35 insertions, 91 deletions
diff --git a/libavcodec/riscv/vc1dsp_rvv.S b/libavcodec/riscv/vc1dsp_rvv.S index 5189d5e855..548ef9d3bf 100644 --- a/libavcodec/riscv/vc1dsp_rvv.S +++ b/libavcodec/riscv/vc1dsp_rvv.S @@ -21,101 +21,45 @@ #include "libavutil/riscv/asm.S" -func ff_vc1_inv_trans_8x8_dc_rvv, zve64x, zba +.macro inv_trans_dc rows, cols, w, mat_lmul, row_lmul +func ff_vc1_inv_trans_\cols\()x\rows\()_dc_rvv, zve64x, zba lpad 0 - lh t2, (a2) - vsetivli zero, 8, e8, mf2, ta, ma - vlse64.v v0, (a0), a1 - sh1add t2, t2, t2 - addi t2, t2, 1 - srai t2, t2, 1 - sh1add t2, t2, t2 - addi t2, t2, 16 - srai t2, t2, 5 - li t0, 8*8 - vsetvli zero, t0, e16, m8, ta, ma - vzext.vf2 v8, v0 - vadd.vx v8, v8, t2 - vmax.vx v8, v8, zero - vsetvli zero, zero, e8, m4, ta, ma - vnclipu.wi v0, v8, 0 - vsetivli zero, 8, e8, mf2, ta, ma - vsse64.v v0, (a0), a1 - ret -endfunc - -func ff_vc1_inv_trans_4x8_dc_rvv, zve32x, zba - lpad 0 - lh t2, (a2) - vsetivli zero, 8, e8, mf2, ta, ma - vlse32.v v0, (a0), a1 - slli t1, t2, 4 - add t2, t2, t1 - addi t2, t2, 4 - srai t2, t2, 3 - sh1add t2, t2, t2 - slli t2, t2, 2 - addi t2, t2, 64 - srai t2, t2, 7 - li t0, 4*8 - vsetvli zero, t0, e16, m4, ta, ma - vzext.vf2 v4, v0 - vadd.vx v4, v4, t2 - vmax.vx v4, v4, zero - vsetvli zero, zero, e8, m2, ta, ma - vnclipu.wi v0, v4, 0 - vsetivli zero, 8, e8, mf2, ta, ma - vsse32.v v0, (a0), a1 - ret -endfunc - -func ff_vc1_inv_trans_8x4_dc_rvv, zve64x, zba - lpad 0 - lh t2, (a2) - vsetivli zero, 4, e8, mf4, ta, ma - vlse64.v v0, (a0), a1 - sh1add t2, t2, t2 - addi t2, t2, 1 - srai t2, t2, 1 - slli t1, t2, 4 - add t2, t2, t1 - addi t2, t2, 64 - srai t2, t2, 7 - li t0, 8*4 - vsetvli zero, t0, e16, m4, ta, ma - vzext.vf2 v4, v0 - vadd.vx v4, v4, t2 - vmax.vx v4, v4, zero - vsetvli zero, zero, e8, m2, ta, ma - vnclipu.wi v0, v4, 0 - vsetivli zero, 4, e8, mf4, ta, ma - vsse64.v v0, (a0), a1 + lh t2, (a2) + li a4, 22 - (5 * \cols) / 4 + mul t2, t2, a4 + vsetivli zero, \rows, e8, m\row_lmul, ta, ma + vlse\w\().v v0, (a0), a1 + addi t2, t2, 4 + li a5, 22 - (5 * \rows) / 4 + srai t2, t2, 3 + mul t2, t2, a5 +.if \cols * \rows >= 32 + li t0, \cols * \rows +.endif + addi t2, t2, 64 + srai t2, t2, 7 +.if \rows * \cols == 64 + vsetvli zero, t0, e16, m8, ta, ma +.elseif \rows * \cols == 32 + vsetvli zero, t0, e16, m4, ta, ma +.else + vsetivli zero, \rows * \cols, e16, m2, ta, ma +.endif + vzext.vf2 v8, v0 + vadd.vx v8, v8, t2 + vmax.vx v8, v8, zero + vsetvli zero, zero, e8, m\mat_lmul, ta, ma + vnclipu.wi v0, v8, 0 + vsetivli zero, \rows, e8, m\row_lmul, ta, ma + vsse\w\().v v0, (a0), a1 ret endfunc +.endm -func ff_vc1_inv_trans_4x4_dc_rvv, zve32x - lpad 0 - lh t2, (a2) - vsetivli zero, 4, e8, mf4, ta, ma - vlse32.v v0, (a0), a1 - slli t1, t2, 4 - add t2, t2, t1 - addi t2, t2, 4 - srai t2, t2, 3 - slli t1, t2, 4 - add t2, t2, t1 - addi t2, t2, 64 - srai t2, t2, 7 - vsetivli zero, 4*4, e16, m2, ta, ma - vzext.vf2 v2, v0 - vadd.vx v2, v2, t2 - vmax.vx v2, v2, zero - vsetvli zero, zero, e8, m1, ta, ma - vnclipu.wi v0, v2, 0 - vsetivli zero, 4, e8, mf4, ta, ma - vsse32.v v0, (a0), a1 - ret -endfunc +inv_trans_dc 8, 8, 64, 4, f2 +inv_trans_dc 4, 8, 64, 2, f4 +inv_trans_dc 8, 4, 32, 2, f2 +inv_trans_dc 4, 4, 32, 1, f4 .variant_cc ff_vc1_inv_trans_8_rvv func ff_vc1_inv_trans_8_rvv, zve32x |