diff options
author | Martin Storsjö <martin@martin.st> | 2024-03-25 14:52:55 +0200 |
---|---|---|
committer | Martin Storsjö <martin@martin.st> | 2024-03-26 09:05:50 +0200 |
commit | 20c38f4b8d3eca93badef1bbe1306121dc5b776b (patch) | |
tree | d437cfd12ad632652e0356aa88fab0e05275cfe6 /libavcodec | |
parent | 4f71e4ebf23694ef8aa8d749d5e9aa219d165ce6 (diff) | |
download | ffmpeg-20c38f4b8d3eca93badef1bbe1306121dc5b776b.tar.gz |
aarch64: hevc: Reorder qpel_hv functions to prepare for templating
This is a pure reordering of code without changing anything in
the individual functions.
Signed-off-by: Martin Storsjö <martin@martin.st>
Diffstat (limited to 'libavcodec')
-rw-r--r-- | libavcodec/aarch64/hevcdsp_qpel_neon.S | 695 |
1 files changed, 355 insertions, 340 deletions
diff --git a/libavcodec/aarch64/hevcdsp_qpel_neon.S b/libavcodec/aarch64/hevcdsp_qpel_neon.S index 06832603d9..ad568e415b 100644 --- a/libavcodec/aarch64/hevcdsp_qpel_neon.S +++ b/libavcodec/aarch64/hevcdsp_qpel_neon.S @@ -2146,29 +2146,6 @@ function ff_hevc_put_hevc_qpel_uni_w_v64_8_neon, export=1 ret endfunc -#if HAVE_I8MM -ENABLE_I8MM - -function ff_hevc_put_hevc_qpel_uni_hv4_8_neon_i8mm, export=1 - add w10, w4, #7 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - str x30, [sp, #-48]! - stp x4, x6, [sp, #16] - stp x0, x1, [sp, #32] - sub x1, x2, x3, lsl #1 - sub x1, x1, x3 - add x0, sp, #48 - mov x2, x3 - add x3, x4, #7 - mov x4, x5 - bl X(ff_hevc_put_hevc_qpel_h4_8_neon_i8mm) - ldp x4, x6, [sp, #16] - ldp x0, x1, [sp, #32] - ldr x30, [sp], #48 - b hevc_put_hevc_qpel_uni_hv4_8_end_neon -endfunc - function hevc_put_hevc_qpel_uni_hv4_8_end_neon mov x9, #(MAX_PB_SIZE * 2) load_qpel_filterh x6, x5 @@ -2195,26 +2172,6 @@ function hevc_put_hevc_qpel_uni_hv4_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_uni_hv6_8_neon_i8mm, export=1 - add w10, w4, #7 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - str x30, [sp, #-48]! - stp x4, x6, [sp, #16] - stp x0, x1, [sp, #32] - sub x1, x2, x3, lsl #1 - sub x1, x1, x3 - add x0, sp, #48 - mov x2, x3 - add w3, w4, #7 - mov x4, x5 - bl X(ff_hevc_put_hevc_qpel_h6_8_neon_i8mm) - ldp x4, x6, [sp, #16] - ldp x0, x1, [sp, #32] - ldr x30, [sp], #48 - b hevc_put_hevc_qpel_uni_hv6_8_end_neon -endfunc - function hevc_put_hevc_qpel_uni_hv6_8_end_neon mov x9, #(MAX_PB_SIZE * 2) load_qpel_filterh x6, x5 @@ -2244,26 +2201,6 @@ function hevc_put_hevc_qpel_uni_hv6_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_uni_hv8_8_neon_i8mm, export=1 - add w10, w4, #7 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - str x30, [sp, #-48]! - stp x4, x6, [sp, #16] - stp x0, x1, [sp, #32] - sub x1, x2, x3, lsl #1 - sub x1, x1, x3 - add x0, sp, #48 - mov x2, x3 - add w3, w4, #7 - mov x4, x5 - bl X(ff_hevc_put_hevc_qpel_h8_8_neon_i8mm) - ldp x4, x6, [sp, #16] - ldp x0, x1, [sp, #32] - ldr x30, [sp], #48 - b hevc_put_hevc_qpel_uni_hv8_8_end_neon -endfunc - function hevc_put_hevc_qpel_uni_hv8_8_end_neon mov x9, #(MAX_PB_SIZE * 2) load_qpel_filterh x6, x5 @@ -2291,26 +2228,6 @@ function hevc_put_hevc_qpel_uni_hv8_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_uni_hv12_8_neon_i8mm, export=1 - add w10, w4, #7 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - stp x7, x30, [sp, #-48]! - stp x4, x6, [sp, #16] - stp x0, x1, [sp, #32] - sub x1, x2, x3, lsl #1 - sub x1, x1, x3 - mov x2, x3 - add x0, sp, #48 - add w3, w4, #7 - mov x4, x5 - bl X(ff_hevc_put_hevc_qpel_h12_8_neon_i8mm) - ldp x4, x6, [sp, #16] - ldp x0, x1, [sp, #32] - ldp x7, x30, [sp], #48 - b hevc_put_hevc_qpel_uni_hv12_8_end_neon -endfunc - function hevc_put_hevc_qpel_uni_hv12_8_end_neon mov x9, #(MAX_PB_SIZE * 2) load_qpel_filterh x6, x5 @@ -2338,26 +2255,6 @@ function hevc_put_hevc_qpel_uni_hv12_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_uni_hv16_8_neon_i8mm, export=1 - add w10, w4, #7 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - stp x7, x30, [sp, #-48]! - stp x4, x6, [sp, #16] - stp x0, x1, [sp, #32] - add x0, sp, #48 - sub x1, x2, x3, lsl #1 - sub x1, x1, x3 - mov x2, x3 - add w3, w4, #7 - mov x4, x5 - bl X(ff_hevc_put_hevc_qpel_h16_8_neon_i8mm) - ldp x4, x6, [sp, #16] - ldp x0, x1, [sp, #32] - ldp x7, x30, [sp], #48 - b hevc_put_hevc_qpel_uni_hv16_8_end_neon -endfunc - function hevc_put_hevc_qpel_uni_hv16_8_end_neon mov x9, #(MAX_PB_SIZE * 2) load_qpel_filterh x6, x5 @@ -2396,6 +2293,109 @@ function hevc_put_hevc_qpel_uni_hv16_8_end_neon ret endfunc +#if HAVE_I8MM +ENABLE_I8MM + +function ff_hevc_put_hevc_qpel_uni_hv4_8_neon_i8mm, export=1 + add w10, w4, #7 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + str x30, [sp, #-48]! + stp x4, x6, [sp, #16] + stp x0, x1, [sp, #32] + sub x1, x2, x3, lsl #1 + sub x1, x1, x3 + add x0, sp, #48 + mov x2, x3 + add x3, x4, #7 + mov x4, x5 + bl X(ff_hevc_put_hevc_qpel_h4_8_neon_i8mm) + ldp x4, x6, [sp, #16] + ldp x0, x1, [sp, #32] + ldr x30, [sp], #48 + b hevc_put_hevc_qpel_uni_hv4_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_uni_hv6_8_neon_i8mm, export=1 + add w10, w4, #7 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + str x30, [sp, #-48]! + stp x4, x6, [sp, #16] + stp x0, x1, [sp, #32] + sub x1, x2, x3, lsl #1 + sub x1, x1, x3 + add x0, sp, #48 + mov x2, x3 + add w3, w4, #7 + mov x4, x5 + bl X(ff_hevc_put_hevc_qpel_h6_8_neon_i8mm) + ldp x4, x6, [sp, #16] + ldp x0, x1, [sp, #32] + ldr x30, [sp], #48 + b hevc_put_hevc_qpel_uni_hv6_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_uni_hv8_8_neon_i8mm, export=1 + add w10, w4, #7 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + str x30, [sp, #-48]! + stp x4, x6, [sp, #16] + stp x0, x1, [sp, #32] + sub x1, x2, x3, lsl #1 + sub x1, x1, x3 + add x0, sp, #48 + mov x2, x3 + add w3, w4, #7 + mov x4, x5 + bl X(ff_hevc_put_hevc_qpel_h8_8_neon_i8mm) + ldp x4, x6, [sp, #16] + ldp x0, x1, [sp, #32] + ldr x30, [sp], #48 + b hevc_put_hevc_qpel_uni_hv8_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_uni_hv12_8_neon_i8mm, export=1 + add w10, w4, #7 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + stp x7, x30, [sp, #-48]! + stp x4, x6, [sp, #16] + stp x0, x1, [sp, #32] + sub x1, x2, x3, lsl #1 + sub x1, x1, x3 + mov x2, x3 + add x0, sp, #48 + add w3, w4, #7 + mov x4, x5 + bl X(ff_hevc_put_hevc_qpel_h12_8_neon_i8mm) + ldp x4, x6, [sp, #16] + ldp x0, x1, [sp, #32] + ldp x7, x30, [sp], #48 + b hevc_put_hevc_qpel_uni_hv12_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_uni_hv16_8_neon_i8mm, export=1 + add w10, w4, #7 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + stp x7, x30, [sp, #-48]! + stp x4, x6, [sp, #16] + stp x0, x1, [sp, #32] + add x0, sp, #48 + sub x1, x2, x3, lsl #1 + sub x1, x1, x3 + mov x2, x3 + add w3, w4, #7 + mov x4, x5 + bl X(ff_hevc_put_hevc_qpel_h16_8_neon_i8mm) + ldp x4, x6, [sp, #16] + ldp x0, x1, [sp, #32] + ldp x7, x30, [sp], #48 + b hevc_put_hevc_qpel_uni_hv16_8_end_neon +endfunc + function ff_hevc_put_hevc_qpel_uni_hv24_8_neon_i8mm, export=1 stp x4, x5, [sp, #-64]! stp x2, x3, [sp, #16] @@ -3779,25 +3779,10 @@ function ff_hevc_put_hevc_qpel_h64_8_neon_i8mm, export=1 b.ne 1b ret endfunc +DISABLE_I8MM +#endif -function ff_hevc_put_hevc_qpel_hv4_8_neon_i8mm, export=1 - add w10, w3, #7 - mov x7, #128 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - stp x5, x30, [sp, #-32]! - stp x0, x3, [sp, #16] - add x0, sp, #32 - sub x1, x1, x2, lsl #1 - add x3, x3, #7 - sub x1, x1, x2 - bl X(ff_hevc_put_hevc_qpel_h4_8_neon_i8mm) - ldp x0, x3, [sp, #16] - ldp x5, x30, [sp], #32 - b hevc_put_hevc_qpel_hv4_8_end_neon -endfunc - function hevc_put_hevc_qpel_hv4_8_end_neon load_qpel_filterh x5, x4 ldr d16, [sp] @@ -3822,23 +3807,6 @@ function hevc_put_hevc_qpel_hv4_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_hv6_8_neon_i8mm, export=1 - add w10, w3, #7 - mov x7, #128 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - stp x5, x30, [sp, #-32]! - stp x0, x3, [sp, #16] - add x0, sp, #32 - sub x1, x1, x2, lsl #1 - add x3, x3, #7 - sub x1, x1, x2 - bl X(ff_hevc_put_hevc_qpel_h6_8_neon_i8mm) - ldp x0, x3, [sp, #16] - ldp x5, x30, [sp], #32 - b hevc_put_hevc_qpel_hv6_8_end_neon -endfunc - function hevc_put_hevc_qpel_hv6_8_end_neon mov x8, #120 load_qpel_filterh x5, x4 @@ -3866,22 +3834,6 @@ function hevc_put_hevc_qpel_hv6_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_hv8_8_neon_i8mm, export=1 - add w10, w3, #7 - lsl x10, x10, #7 - sub x1, x1, x2, lsl #1 - sub sp, sp, x10 // tmp_array - stp x5, x30, [sp, #-32]! - stp x0, x3, [sp, #16] - add x0, sp, #32 - add x3, x3, #7 - sub x1, x1, x2 - bl X(ff_hevc_put_hevc_qpel_h8_8_neon_i8mm) - ldp x0, x3, [sp, #16] - ldp x5, x30, [sp], #32 - b hevc_put_hevc_qpel_hv8_8_end_neon -endfunc - function hevc_put_hevc_qpel_hv8_8_end_neon mov x7, #128 load_qpel_filterh x5, x4 @@ -3908,22 +3860,6 @@ function hevc_put_hevc_qpel_hv8_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_hv12_8_neon_i8mm, export=1 - add w10, w3, #7 - lsl x10, x10, #7 - sub x1, x1, x2, lsl #1 - sub sp, sp, x10 // tmp_array - stp x5, x30, [sp, #-32]! - stp x0, x3, [sp, #16] - add x0, sp, #32 - add x3, x3, #7 - sub x1, x1, x2 - bl X(ff_hevc_put_hevc_qpel_h12_8_neon_i8mm) - ldp x0, x3, [sp, #16] - ldp x5, x30, [sp], #32 - b hevc_put_hevc_qpel_hv12_8_end_neon -endfunc - function hevc_put_hevc_qpel_hv12_8_end_neon mov x7, #128 load_qpel_filterh x5, x4 @@ -3949,22 +3885,6 @@ function hevc_put_hevc_qpel_hv12_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_hv16_8_neon_i8mm, export=1 - add w10, w3, #7 - lsl x10, x10, #7 - sub x1, x1, x2, lsl #1 - sub sp, sp, x10 // tmp_array - stp x5, x30, [sp, #-32]! - stp x0, x3, [sp, #16] - add x3, x3, #7 - add x0, sp, #32 - sub x1, x1, x2 - bl X(ff_hevc_put_hevc_qpel_h16_8_neon_i8mm) - ldp x0, x3, [sp, #16] - ldp x5, x30, [sp], #32 - b hevc_put_hevc_qpel_hv16_8_end_neon -endfunc - function hevc_put_hevc_qpel_hv16_8_end_neon mov x7, #128 load_qpel_filterh x5, x4 @@ -3989,38 +3909,6 @@ function hevc_put_hevc_qpel_hv16_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_hv24_8_neon_i8mm, export=1 - stp x4, x5, [sp, #-64]! - stp x2, x3, [sp, #16] - stp x0, x1, [sp, #32] - str x30, [sp, #48] - bl X(ff_hevc_put_hevc_qpel_hv12_8_neon_i8mm) - ldp x0, x1, [sp, #32] - ldp x2, x3, [sp, #16] - ldp x4, x5, [sp], #48 - add x1, x1, #12 - add x0, x0, #24 - bl X(ff_hevc_put_hevc_qpel_hv12_8_neon_i8mm) - ldr x30, [sp], #16 - ret -endfunc - -function ff_hevc_put_hevc_qpel_hv32_8_neon_i8mm, export=1 - add w10, w3, #7 - sub x1, x1, x2, lsl #1 - lsl x10, x10, #7 - sub x1, x1, x2 - sub sp, sp, x10 // tmp_array - stp x5, x30, [sp, #-32]! - stp x0, x3, [sp, #16] - add x3, x3, #7 - add x0, sp, #32 - bl X(ff_hevc_put_hevc_qpel_h32_8_neon_i8mm) - ldp x0, x3, [sp, #16] - ldp x5, x30, [sp], #32 - b hevc_put_hevc_qpel_hv32_8_end_neon -endfunc - function hevc_put_hevc_qpel_hv32_8_end_neon mov x7, #128 load_qpel_filterh x5, x4 @@ -4056,6 +3944,122 @@ function hevc_put_hevc_qpel_hv32_8_end_neon ret endfunc +#if HAVE_I8MM +ENABLE_I8MM +function ff_hevc_put_hevc_qpel_hv4_8_neon_i8mm, export=1 + add w10, w3, #7 + mov x7, #128 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + stp x5, x30, [sp, #-32]! + stp x0, x3, [sp, #16] + add x0, sp, #32 + sub x1, x1, x2, lsl #1 + add x3, x3, #7 + sub x1, x1, x2 + bl X(ff_hevc_put_hevc_qpel_h4_8_neon_i8mm) + ldp x0, x3, [sp, #16] + ldp x5, x30, [sp], #32 + b hevc_put_hevc_qpel_hv4_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_hv6_8_neon_i8mm, export=1 + add w10, w3, #7 + mov x7, #128 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + stp x5, x30, [sp, #-32]! + stp x0, x3, [sp, #16] + add x0, sp, #32 + sub x1, x1, x2, lsl #1 + add x3, x3, #7 + sub x1, x1, x2 + bl X(ff_hevc_put_hevc_qpel_h6_8_neon_i8mm) + ldp x0, x3, [sp, #16] + ldp x5, x30, [sp], #32 + b hevc_put_hevc_qpel_hv6_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_hv8_8_neon_i8mm, export=1 + add w10, w3, #7 + lsl x10, x10, #7 + sub x1, x1, x2, lsl #1 + sub sp, sp, x10 // tmp_array + stp x5, x30, [sp, #-32]! + stp x0, x3, [sp, #16] + add x0, sp, #32 + add x3, x3, #7 + sub x1, x1, x2 + bl X(ff_hevc_put_hevc_qpel_h8_8_neon_i8mm) + ldp x0, x3, [sp, #16] + ldp x5, x30, [sp], #32 + b hevc_put_hevc_qpel_hv8_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_hv12_8_neon_i8mm, export=1 + add w10, w3, #7 + lsl x10, x10, #7 + sub x1, x1, x2, lsl #1 + sub sp, sp, x10 // tmp_array + stp x5, x30, [sp, #-32]! + stp x0, x3, [sp, #16] + add x0, sp, #32 + add x3, x3, #7 + sub x1, x1, x2 + bl X(ff_hevc_put_hevc_qpel_h12_8_neon_i8mm) + ldp x0, x3, [sp, #16] + ldp x5, x30, [sp], #32 + b hevc_put_hevc_qpel_hv12_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_hv16_8_neon_i8mm, export=1 + add w10, w3, #7 + lsl x10, x10, #7 + sub x1, x1, x2, lsl #1 + sub sp, sp, x10 // tmp_array + stp x5, x30, [sp, #-32]! + stp x0, x3, [sp, #16] + add x3, x3, #7 + add x0, sp, #32 + sub x1, x1, x2 + bl X(ff_hevc_put_hevc_qpel_h16_8_neon_i8mm) + ldp x0, x3, [sp, #16] + ldp x5, x30, [sp], #32 + b hevc_put_hevc_qpel_hv16_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_hv24_8_neon_i8mm, export=1 + stp x4, x5, [sp, #-64]! + stp x2, x3, [sp, #16] + stp x0, x1, [sp, #32] + str x30, [sp, #48] + bl X(ff_hevc_put_hevc_qpel_hv12_8_neon_i8mm) + ldp x0, x1, [sp, #32] + ldp x2, x3, [sp, #16] + ldp x4, x5, [sp], #48 + add x1, x1, #12 + add x0, x0, #24 + bl X(ff_hevc_put_hevc_qpel_hv12_8_neon_i8mm) + ldr x30, [sp], #16 + ret +endfunc + +function ff_hevc_put_hevc_qpel_hv32_8_neon_i8mm, export=1 + add w10, w3, #7 + sub x1, x1, x2, lsl #1 + lsl x10, x10, #7 + sub x1, x1, x2 + sub sp, sp, x10 // tmp_array + stp x5, x30, [sp, #-32]! + stp x0, x3, [sp, #16] + add x3, x3, #7 + add x0, sp, #32 + bl X(ff_hevc_put_hevc_qpel_h32_8_neon_i8mm) + ldp x0, x3, [sp, #16] + ldp x5, x30, [sp], #32 + b hevc_put_hevc_qpel_hv32_8_end_neon +endfunc + function ff_hevc_put_hevc_qpel_hv48_8_neon_i8mm, export=1 stp x4, x5, [sp, #-64]! stp x2, x3, [sp, #16] @@ -4089,6 +4093,8 @@ function ff_hevc_put_hevc_qpel_hv64_8_neon_i8mm, export=1 ldr x30, [sp], #16 ret endfunc +DISABLE_I8MM +#endif .macro QPEL_UNI_W_HV_HEADER width ldp x14, x15, [sp] // mx, my @@ -4168,11 +4174,6 @@ endfunc smlal2 \dst\().4s, \src7\().8h, v0.h[7] .endm -function ff_hevc_put_hevc_qpel_uni_w_hv4_8_neon_i8mm, export=1 - QPEL_UNI_W_HV_HEADER 4 - b hevc_put_hevc_qpel_uni_w_hv4_8_end_neon -endfunc - function hevc_put_hevc_qpel_uni_w_hv4_8_end_neon ldr d16, [sp] ldr d17, [sp, x10] @@ -4262,11 +4263,6 @@ endfunc st1 {v24.d}[0], [x20], x21 .endm -function ff_hevc_put_hevc_qpel_uni_w_hv8_8_neon_i8mm, export=1 - QPEL_UNI_W_HV_HEADER 8 - b hevc_put_hevc_qpel_uni_w_hv8_8_end_neon -endfunc - function hevc_put_hevc_qpel_uni_w_hv8_8_end_neon ldr q16, [sp] ldr q17, [sp, x10] @@ -4376,21 +4372,6 @@ endfunc st1 {v24.16b}, [x20], x21 .endm -function ff_hevc_put_hevc_qpel_uni_w_hv16_8_neon_i8mm, export=1 - QPEL_UNI_W_HV_HEADER 16 - b hevc_put_hevc_qpel_uni_w_hv16_8_end_neon -endfunc - -function ff_hevc_put_hevc_qpel_uni_w_hv32_8_neon_i8mm, export=1 - QPEL_UNI_W_HV_HEADER 32 - b hevc_put_hevc_qpel_uni_w_hv16_8_end_neon -endfunc - -function ff_hevc_put_hevc_qpel_uni_w_hv64_8_neon_i8mm, export=1 - QPEL_UNI_W_HV_HEADER 64 - b hevc_put_hevc_qpel_uni_w_hv16_8_end_neon -endfunc - function hevc_put_hevc_qpel_uni_w_hv16_8_end_neon mov x11, sp mov w12, w22 @@ -4503,26 +4484,37 @@ function hevc_put_hevc_qpel_uni_w_hv16_8_end_neon ret endfunc -function ff_hevc_put_hevc_qpel_bi_hv4_8_neon_i8mm, export=1 - add w10, w5, #7 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - stp x7, x30, [sp, #-48]! - stp x4, x5, [sp, #16] - stp x0, x1, [sp, #32] - sub x1, x2, x3, lsl #1 - sub x1, x1, x3 - add x0, sp, #48 - mov x2, x3 - add w3, w5, #7 - mov x4, x6 - bl X(ff_hevc_put_hevc_qpel_h4_8_neon_i8mm) - ldp x4, x5, [sp, #16] - ldp x0, x1, [sp, #32] - ldp x7, x30, [sp], #48 - b hevc_put_hevc_qpel_bi_hv4_8_end_neon +#if HAVE_I8MM +ENABLE_I8MM + +function ff_hevc_put_hevc_qpel_uni_w_hv4_8_neon_i8mm, export=1 + QPEL_UNI_W_HV_HEADER 4 + b hevc_put_hevc_qpel_uni_w_hv4_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_uni_w_hv8_8_neon_i8mm, export=1 + QPEL_UNI_W_HV_HEADER 8 + b hevc_put_hevc_qpel_uni_w_hv8_8_end_neon endfunc +function ff_hevc_put_hevc_qpel_uni_w_hv16_8_neon_i8mm, export=1 + QPEL_UNI_W_HV_HEADER 16 + b hevc_put_hevc_qpel_uni_w_hv16_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_uni_w_hv32_8_neon_i8mm, export=1 + QPEL_UNI_W_HV_HEADER 32 + b hevc_put_hevc_qpel_uni_w_hv16_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_uni_w_hv64_8_neon_i8mm, export=1 + QPEL_UNI_W_HV_HEADER 64 + b hevc_put_hevc_qpel_uni_w_hv16_8_end_neon +endfunc + +DISABLE_I8MM +#endif + function hevc_put_hevc_qpel_bi_hv4_8_end_neon mov x9, #(MAX_PB_SIZE * 2) load_qpel_filterh x7, x6 @@ -4548,26 +4540,6 @@ function hevc_put_hevc_qpel_bi_hv4_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_bi_hv6_8_neon_i8mm, export=1 - add w10, w5, #7 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - stp x7, x30, [sp, #-48]! - stp x4, x5, [sp, #16] - stp x0, x1, [sp, #32] - sub x1, x2, x3, lsl #1 - sub x1, x1, x3 - add x0, sp, #48 - mov x2, x3 - add x3, x5, #7 - mov x4, x6 - bl X(ff_hevc_put_hevc_qpel_h6_8_neon_i8mm) - ldp x4, x5, [sp, #16] - ldp x0, x1, [sp, #32] - ldp x7, x30, [sp], #48 - b hevc_put_hevc_qpel_bi_hv6_8_end_neon -endfunc - function hevc_put_hevc_qpel_bi_hv6_8_end_neon mov x9, #(MAX_PB_SIZE * 2) load_qpel_filterh x7, x6 @@ -4598,26 +4570,6 @@ function hevc_put_hevc_qpel_bi_hv6_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_bi_hv8_8_neon_i8mm, export=1 - add w10, w5, #7 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - stp x7, x30, [sp, #-48]! - stp x4, x5, [sp, #16] - stp x0, x1, [sp, #32] - sub x1, x2, x3, lsl #1 - sub x1, x1, x3 - add x0, sp, #48 - mov x2, x3 - add x3, x5, #7 - mov x4, x6 - bl X(ff_hevc_put_hevc_qpel_h8_8_neon_i8mm) - ldp x4, x5, [sp, #16] - ldp x0, x1, [sp, #32] - ldp x7, x30, [sp], #48 - b hevc_put_hevc_qpel_bi_hv8_8_end_neon -endfunc - function hevc_put_hevc_qpel_bi_hv8_8_end_neon mov x9, #(MAX_PB_SIZE * 2) load_qpel_filterh x7, x6 @@ -4646,46 +4598,6 @@ function hevc_put_hevc_qpel_bi_hv8_8_end_neon 2: ret endfunc -function ff_hevc_put_hevc_qpel_bi_hv12_8_neon_i8mm, export=1 - stp x6, x7, [sp, #-80]! - stp x4, x5, [sp, #16] - stp x2, x3, [sp, #32] - stp x0, x1, [sp, #48] - str x30, [sp, #64] - bl X(ff_hevc_put_hevc_qpel_bi_hv8_8_neon_i8mm) - ldp x4, x5, [sp, #16] - ldp x2, x3, [sp, #32] - ldp x0, x1, [sp, #48] - ldp x6, x7, [sp], #64 - add x4, x4, #16 - add x2, x2, #8 - add x0, x0, #8 - bl X(ff_hevc_put_hevc_qpel_bi_hv4_8_neon_i8mm) - ldr x30, [sp], #16 - ret -endfunc - -function ff_hevc_put_hevc_qpel_bi_hv16_8_neon_i8mm, export=1 - add w10, w5, #7 - lsl x10, x10, #7 - sub sp, sp, x10 // tmp_array - stp x7, x30, [sp, #-48]! - stp x4, x5, [sp, #16] - stp x0, x1, [sp, #32] - add x0, sp, #48 - sub x1, x2, x3, lsl #1 - sub x1, x1, x3 - mov x2, x3 - add w3, w5, #7 - mov x4, x6 - bl X(ff_hevc_put_hevc_qpel_h16_8_neon_i8mm) - ldp x4, x5, [sp, #16] - ldp x0, x1, [sp, #32] - ldp x7, x30, [sp], #48 - mov x6, #16 // width - b hevc_put_hevc_qpel_bi_hv16_8_end_neon -endfunc - function hevc_put_hevc_qpel_bi_hv16_8_end_neon load_qpel_filterh x7, x8 mov x9, #(MAX_PB_SIZE * 2) @@ -4735,6 +4647,109 @@ function hevc_put_hevc_qpel_bi_hv16_8_end_neon ret endfunc +#if HAVE_I8MM +ENABLE_I8MM + +function ff_hevc_put_hevc_qpel_bi_hv4_8_neon_i8mm, export=1 + add w10, w5, #7 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + stp x7, x30, [sp, #-48]! + stp x4, x5, [sp, #16] + stp x0, x1, [sp, #32] + sub x1, x2, x3, lsl #1 + sub x1, x1, x3 + add x0, sp, #48 + mov x2, x3 + add w3, w5, #7 + mov x4, x6 + bl X(ff_hevc_put_hevc_qpel_h4_8_neon_i8mm) + ldp x4, x5, [sp, #16] + ldp x0, x1, [sp, #32] + ldp x7, x30, [sp], #48 + b hevc_put_hevc_qpel_bi_hv4_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_bi_hv6_8_neon_i8mm, export=1 + add w10, w5, #7 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + stp x7, x30, [sp, #-48]! + stp x4, x5, [sp, #16] + stp x0, x1, [sp, #32] + sub x1, x2, x3, lsl #1 + sub x1, x1, x3 + add x0, sp, #48 + mov x2, x3 + add x3, x5, #7 + mov x4, x6 + bl X(ff_hevc_put_hevc_qpel_h6_8_neon_i8mm) + ldp x4, x5, [sp, #16] + ldp x0, x1, [sp, #32] + ldp x7, x30, [sp], #48 + b hevc_put_hevc_qpel_bi_hv6_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_bi_hv8_8_neon_i8mm, export=1 + add w10, w5, #7 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + stp x7, x30, [sp, #-48]! + stp x4, x5, [sp, #16] + stp x0, x1, [sp, #32] + sub x1, x2, x3, lsl #1 + sub x1, x1, x3 + add x0, sp, #48 + mov x2, x3 + add x3, x5, #7 + mov x4, x6 + bl X(ff_hevc_put_hevc_qpel_h8_8_neon_i8mm) + ldp x4, x5, [sp, #16] + ldp x0, x1, [sp, #32] + ldp x7, x30, [sp], #48 + b hevc_put_hevc_qpel_bi_hv8_8_end_neon +endfunc + +function ff_hevc_put_hevc_qpel_bi_hv12_8_neon_i8mm, export=1 + stp x6, x7, [sp, #-80]! + stp x4, x5, [sp, #16] + stp x2, x3, [sp, #32] + stp x0, x1, [sp, #48] + str x30, [sp, #64] + bl X(ff_hevc_put_hevc_qpel_bi_hv8_8_neon_i8mm) + ldp x4, x5, [sp, #16] + ldp x2, x3, [sp, #32] + ldp x0, x1, [sp, #48] + ldp x6, x7, [sp], #64 + add x4, x4, #16 + add x2, x2, #8 + add x0, x0, #8 + bl X(ff_hevc_put_hevc_qpel_bi_hv4_8_neon_i8mm) + ldr x30, [sp], #16 + ret +endfunc + +function ff_hevc_put_hevc_qpel_bi_hv16_8_neon_i8mm, export=1 + add w10, w5, #7 + lsl x10, x10, #7 + sub sp, sp, x10 // tmp_array + stp x7, x30, [sp, #-48]! + stp x4, x5, [sp, #16] + stp x0, x1, [sp, #32] + add x0, sp, #48 + sub x1, x2, x3, lsl #1 + sub x1, x1, x3 + mov x2, x3 + add w3, w5, #7 + mov x4, x6 + bl X(ff_hevc_put_hevc_qpel_h16_8_neon_i8mm) + ldp x4, x5, [sp, #16] + ldp x0, x1, [sp, #32] + ldp x7, x30, [sp], #48 + mov x6, #16 // width + b hevc_put_hevc_qpel_bi_hv16_8_end_neon +endfunc + function ff_hevc_put_hevc_qpel_bi_hv24_8_neon_i8mm, export=1 stp x6, x7, [sp, #-80]! stp x4, x5, [sp, #16] |