aboutsummaryrefslogtreecommitdiffstats
path: root/libavcodec/x86/hevc_idct.asm
diff options
context:
space:
mode:
authorAlexandra Hájková <alexandra.khirnova@gmail.com>2016-10-10 22:23:01 +0200
committerAnton Khirnov <anton@khirnov.net>2016-10-11 18:21:04 +0200
commit112cee0241f5799edff0e4682b9e8639b046dc78 (patch)
tree8b2d009250e8717de8487b3ec34917cbf029b2b4 /libavcodec/x86/hevc_idct.asm
parente3f941cb03b139b866a0ad6dc95fbe1b247d54af (diff)
downloadffmpeg-112cee0241f5799edff0e4682b9e8639b046dc78.tar.gz
hevc: Add SSE2 and AVX IDCT
Signed-off-by: Anton Khirnov <anton@khirnov.net>
Diffstat (limited to 'libavcodec/x86/hevc_idct.asm')
-rw-r--r--libavcodec/x86/hevc_idct.asm787
1 files changed, 765 insertions, 22 deletions
diff --git a/libavcodec/x86/hevc_idct.asm b/libavcodec/x86/hevc_idct.asm
index d662aa90ba..f397cc1097 100644
--- a/libavcodec/x86/hevc_idct.asm
+++ b/libavcodec/x86/hevc_idct.asm
@@ -2,6 +2,7 @@
;* SIMD-optimized IDCT functions for HEVC decoding
;* Copyright (c) 2014 Pierre-Edouard LEPERE
;* Copyright (c) 2014 James Almer
+;* Copyright (c) 2016 Alexandra Hájková
;*
;* This file is part of Libav.
;*
@@ -22,6 +23,217 @@
%include "libavutil/x86/x86util.asm"
+SECTION_RODATA
+
+pd_64: times 4 dd 64
+pd_2048: times 4 dd 2048
+pd_512: times 4 dd 512
+
+; 4x4 transform coeffs
+cextern pw_64
+pw_64_m64: times 4 dw 64, -64
+pw_83_36: times 4 dw 83, 36
+pw_36_m83: times 4 dw 36, -83
+
+; 8x8 transform coeffs
+pw_89_75: times 4 dw 89, 75
+pw_50_18: times 4 dw 50, 18
+
+pw_75_m18: times 4 dw 75, -18
+pw_m89_m50: times 4 dw -89, -50
+
+pw_50_m89: times 4 dw 50, -89
+pw_18_75: times 4 dw 18, 75
+
+pw_18_m50: times 4 dw 18, -50
+pw_75_m89: times 4 dw 75, -89
+
+; 16x16 transformation coeffs
+trans_coeffs16: times 4 dw 90, 87
+times 4 dw 80, 70
+times 4 dw 57, 43
+times 4 dw 25, 9
+
+times 4 dw 87, 57
+times 4 dw 9, -43
+times 4 dw -80, -90
+times 4 dw -70, -25
+
+times 4 dw 80, 9
+times 4 dw -70, -87
+times 4 dw -25, 57
+times 4 dw 90, 43
+
+times 4 dw 70, -43
+times 4 dw -87, 9
+times 4 dw 90, 25
+times 4 dw -80, -57
+
+times 4 dw 57, -80
+times 4 dw -25, 90
+times 4 dw -9, -87
+times 4 dw 43, 70
+
+times 4 dw 43, -90
+times 4 dw 57, 25
+times 4 dw -87, 70
+times 4 dw 9, -80
+
+times 4 dw 25, -70
+times 4 dw 90, -80
+times 4 dw 43, 9
+times 4 dw -57, 87
+
+times 4 dw 9, -25
+times 4 dw 43, -57
+times 4 dw 70, -80
+times 4 dw 87, -90
+
+; 32x32 transform coeffs
+trans_coeff32: times 8 dw 90
+times 4 dw 88, 85
+times 4 dw 82, 78
+times 4 dw 73, 67
+times 4 dw 61, 54
+times 4 dw 46, 38
+times 4 dw 31, 22
+times 4 dw 13, 4
+
+times 4 dw 90, 82
+times 4 dw 67, 46
+times 4 dw 22, -4
+times 4 dw -31, -54
+times 4 dw -73, -85
+times 4 dw -90, -88
+times 4 dw -78, -61
+times 4 dw -38, -13
+
+times 4 dw 88, 67
+times 4 dw 31, -13
+times 4 dw -54, -82
+times 4 dw -90, -78
+times 4 dw -46, -4
+times 4 dw 38, 73
+times 4 dw 90, 85
+times 4 dw 61, 22
+
+times 4 dw 85, 46
+times 4 dw -13, -67
+times 4 dw -90, -73
+times 4 dw -22, 38
+times 4 dw 82, 88
+times 4 dw 54, -4
+times 4 dw -61, -90
+times 4 dw -78, -31
+
+times 4 dw 82, 22
+times 4 dw -54, -90
+times 4 dw -61, 13
+times 4 dw 78, 85
+times 4 dw 31, -46
+times 4 dw -90, -67
+times 4 dw 4, 73
+times 4 dw 88, 38
+
+times 4 dw 78, -4
+times 4 dw -82, -73
+times 4 dw 13, 85
+times 4 dw 67, -22
+times 4 dw -88, -61
+times 4 dw 31, 90
+times 4 dw 54, -38
+times 4 dw -90, -46
+
+times 4 dw 73, -31
+times 4 dw -90, -22
+times 4 dw 78, 67
+times 4 dw -38, -90
+times 4 dw -13, 82
+times 4 dw 61, -46
+times 4 dw -88, -4
+times 4 dw 85, 54
+
+times 4 dw 67, -54
+times 4 dw -78, 38
+times 4 dw 85, -22
+times 4 dw -90, 4
+times 4 dw 90, 13
+times 4 dw -88, -31
+times 4 dw 82, 46
+times 4 dw -73, -61
+
+times 4 dw 61, -73
+times 4 dw -46, 82
+times 4 dw 31, -88
+times 4 dw -13, 90
+times 4 dw -4, -90
+times 4 dw 22, 85
+times 4 dw -38, -78
+times 4 dw 54, 67
+
+times 4 dw 54, -85
+times 4 dw -4, 88
+times 4 dw -46, -61
+times 4 dw 82, 13
+times 4 dw -90, 38
+times 4 dw 67, -78
+times 4 dw -22, 90
+times 4 dw -31, -73
+
+times 4 dw 46, -90
+times 4 dw 38, 54
+times 4 dw -90, 31
+times 4 dw 61, -88
+times 4 dw 22, 67
+times 4 dw -85, 13
+times 4 dw 73, -82
+times 4 dw 4, 78
+
+times 4 dw 38, -88
+times 4 dw 73, -4
+times 4 dw -67, 90
+times 4 dw -46, -31
+times 4 dw 85, -78
+times 4 dw 13, 61
+times 4 dw -90, 54
+times 4 dw 22, -82
+
+times 4 dw 31, -78
+times 4 dw 90, -61
+times 4 dw 4, 54
+times 4 dw -88, 82
+times 4 dw -38, -22
+times 4 dw 73, -90
+times 4 dw 67, -13
+times 4 dw -46, 85
+
+times 4 dw 22, -61
+times 4 dw 85, -90
+times 4 dw 73, -38
+times 4 dw -4, 46
+times 4 dw -78, 90
+times 4 dw -82, 54
+times 4 dw -13, -31
+times 4 dw 67, -88
+
+times 4 dw 13, -38
+times 4 dw 61, -78
+times 4 dw 88, -90
+times 4 dw 85, -73
+times 4 dw 54, -31
+times 4 dw 4, 22
+times 4 dw -46, 67
+times 4 dw -82, 90
+
+times 4 dw 4, -13
+times 4 dw 22, -31
+times 4 dw 38, -46
+times 4 dw 54, -61
+times 4 dw 67, -73
+times 4 dw 78, -82
+times 4 dw 85, -88
+times 4 dw 90, -90
+
section .text
; void ff_hevc_idctHxW_dc_{8,10}_<opt>(int16_t *coeffs)
@@ -74,34 +286,565 @@ cglobal hevc_idct_%1x%1_dc_%2, 1, 2, 1, coeff, tmp
RET
%endmacro
-; 8-bit
-INIT_MMX mmxext
-IDCT_DC_NL 4, 8
-IDCT_DC 8, 2, 8
+; IDCT 4x4, expects input in m0, m1
+; %1 - shift
+; %2 - 1/0 - SCALE and Transpose or not
+; %3 - 1/0 add constant or not
+%macro TR_4x4 3
+ ; interleaves src0 with src2 to m0
+ ; and src1 with scr3 to m2
+ ; src0: 00 01 02 03 m0: 00 20 01 21 02 22 03 23
+ ; src1: 10 11 12 13 -->
+ ; src2: 20 21 22 23 m1: 10 30 11 31 12 32 13 33
+ ; src3: 30 31 32 33
-INIT_XMM sse2
-IDCT_DC_NL 8, 8
-IDCT_DC 16, 4, 8
-IDCT_DC 32, 16, 8
+ SBUTTERFLY wd, 0, 1, 2
-%if HAVE_AVX2_EXTERNAL
-INIT_YMM avx2
-IDCT_DC 16, 2, 8
-IDCT_DC 32, 8, 8
-%endif ;HAVE_AVX2_EXTERNAL
+ pmaddwd m2, m0, [pw_64] ; e0
+ pmaddwd m3, m1, [pw_83_36] ; o0
+ pmaddwd m0, [pw_64_m64] ; e1
+ pmaddwd m1, [pw_36_m83] ; o1
+
+%if %3 == 1
+ %assign %%add 1 << (%1 - 1)
+ mova m4, [pd_ %+ %%add]
+ paddd m2, m4
+ paddd m0, m4
+%endif
+
+ SUMSUB_BADC d, 3, 2, 1, 0, 4
+
+%if %2 == 1
+ psrad m3, %1 ; e0 + o0
+ psrad m1, %1 ; e1 + o1
+ psrad m2, %1 ; e0 - o0
+ psrad m0, %1 ; e1 - o1
+ ;clip16
+ packssdw m3, m1
+ packssdw m0, m2
+ ; Transpose
+ SBUTTERFLY wd, 3, 0, 1
+ SBUTTERFLY wd, 3, 0, 1
+ SWAP 3, 1, 0
+%else
+ SWAP 3, 2, 0
+%endif
+%endmacro
+
+%macro DEFINE_BIAS 1
+ %assign shift (20 - %1)
+ %assign c_add (1 << (shift - 1))
+ %define arr_add pd_ %+ c_add
+%endmacro
+
+; %1 - bit_depth
+; %2 - register add constant
+; is loaded to
+; shift = 20 - bit_depth
+%macro LOAD_BIAS 2
+ DEFINE_BIAS %1
+ mova %2, [arr_add]
+%endmacro
+
+; %1, %2 - registers to load packed 16 bit values to
+; %3, %4, %5, %6 - vertical offsets
+; %7 - horizontal offset
+%macro LOAD_BLOCK 7
+ movq %1, [r0 + %3 + %7]
+ movhps %1, [r0 + %5 + %7]
+ movq %2, [r0 + %4 + %7]
+ movhps %2, [r0 + %6 + %7]
+%endmacro
+
+; void ff_hevc_idct_4x4__{8,10}_<opt>(int16_t *coeffs, int col_limit)
+; %1 = bitdepth
+%macro IDCT_4x4 1
+cglobal hevc_idct_4x4_%1, 1, 1, 5, coeffs
+ mova m0, [coeffsq]
+ mova m1, [coeffsq + 16]
+
+ TR_4x4 7, 1, 1
+ TR_4x4 20 - %1, 1, 1
+
+ mova [coeffsq], m0
+ mova [coeffsq + 16], m1
+ RET
+%endmacro
+
+; scale, pack (clip16) and store the residuals 0 e8[0] + o8[0] --> + %1
+; 4 at one time (4 columns) 1 e8[1] + o8[1]
+; from %5: e8/16 + o8/16, with %1 offset ...
+; and %3: e8/16 - o8/16, with %2 offset 6 e8[1] - o8[1]
+; %4 - shift 7 e8[0] - o8[0] --> + %2
+%macro STORE_8 7
+ psrad %5, %4
+ psrad %3, %4
+ packssdw %5, %3
+ movq [coeffsq + %1], %5
+ movhps [coeffsq + %2], %5
+%endmacro
+
+; %1 - horizontal offset
+; %2 - shift
+; %3, %4 - transform coeffs
+; %5 - vertical offset for e8 + o8
+; %6 - vertical offset for e8 - o8
+; %7 - register with e8 inside
+; %8 - block_size
+; %9 - register to store e8 +o8
+; %10 - register to store e8 - o8
+%macro E8_O8 10
+ pmaddwd m6, m4, %3
+ pmaddwd m7, m5, %4
+
+ paddd m6, m7
+ paddd m7, m6, %7 ; o8 + e8
+ psubd %7, m6 ; e8 - o8
+%if %8 == 8
+ STORE_8 %5 + %1, %6 + %1, %7, %2, m7, 0, 0
+%else
+ SWAP m7, %9
+ SWAP %7, %10
+%endif
+%endmacro
+
+; 8x4 residuals are processed and stored
+; %1 - horizontal offset
+; %2 - shift
+; %3 - offset of the even row
+; %4 - step: 1 for 8x8, 2 for 16x16, 4 for 32x32
+; %5 - offset of the odd row
+; %6 - block size
+; %7 - 1/0 add a constant in TR_4x4 or not
+; I want to add a constant for 8x8 transform but not for 16x16 and 32x32
+%macro TR_8x4 7
+ ; load 4 columns of even rows
+ LOAD_BLOCK m0, m1, 0, 2 * %4 * %3, %4 * %3, 3 * %4 * %3, %1
+
+ TR_4x4 %2, 0, %7 ; e8: m0, m1, m2, m3, for 4 columns only
+
+ ; load 4 columns of odd rows
+ LOAD_BLOCK m4, m5, %4 * %5, 3 * %4 * %5, 5 * %4 * %5, 7 * %4 * %5, %1
+
+ ; 00 01 02 03
+ ; 10 11 12 13 m4: 10 30 11 31 12 32 13 33
+
+ ; ... -- >
+ ; m5: 50 70 51 71 52 72 53 73
+ ; 70 71 72 73
+ SBUTTERFLY wd, 4, 5, 6
+
+ E8_O8 %1, %2, [pw_89_75], [pw_50_18], 0, %5 * 7, m0, %6, m8, m15
+ E8_O8 %1, %2, [pw_75_m18], [pw_m89_m50], %5, %5 * 6, m1, %6, m9, m14
+ E8_O8 %1, %2, [pw_50_m89], [pw_18_75], %5 * 2, %5 * 5, m2, %6, m10, m13
+ E8_O8 %1, %2, [pw_18_m50], [pw_75_m89], %5 * 3, %5 * 4, m3, %6, m11, m12
+%endmacro
+
+%macro STORE_PACKED 7
+ movq [r0 + %3 + %7], %1
+ movhps [r0 + %4 + %7], %1
+ movq [r0 + %5 + %7], %2
+ movhps [r0 + %6 + %7], %2
+%endmacro
+
+; transpose 4x4 block packed
+; in %1 and %2 registers
+; %3 - temporary register
+%macro TRANSPOSE_4x4 3
+ SBUTTERFLY wd, %1, %2, %3
+ SBUTTERFLY dq, %1, %2, %3
+%endmacro
+
+; %1 - horizontal offset of the block i
+; %2 - vertical offset of the block i
+; %3 - width in bytes
+; %4 - vertical offset for the block j
+; %5 - horizontal offset for the block j
+%macro SWAP_BLOCKS 5
+ ; M_j
+ LOAD_BLOCK m4, m5, %4, %4 + %3, %4 + 2 * %3, %4 + 3 * %3, %5
+ TRANSPOSE_4x4 4, 5, 6
+
+ ; M_i
+ LOAD_BLOCK m6, m7, %2, %2 + %3, %2 + 2 * %3, %2 + 3 * %3, %1
+
+ STORE_PACKED m4, m5, %2, %2 + %3, %2 + 2 * %3, %2 + 3 * %3, %1
+
+ ; transpose and store M_i
+ SWAP m6, m4
+ SWAP m7, m5
+ TRANSPOSE_4x4 4, 5, 6
+ STORE_PACKED m4, m5, %4, %4 + %3, %4 + 2 * %3, %4 + 3 * %3, %5
+%endmacro
+
+; %1 - horizontal offset
+; %2 - vertical offset of the block
+; %3 - width in bytes
+%macro TRANSPOSE_BLOCK 3
+ LOAD_BLOCK m4, m5, %2, %2 + %3, %2 + 2 * %3, %2 + 3 * %3, %1
+ TRANSPOSE_4x4 4, 5, 6
+ STORE_PACKED m4, m5, %2, %2 + %3, %2 + 2 * %3, %2 + 3 * %3, %1
+%endmacro
-; 10-bit
+%macro TRANSPOSE_8x8 0
+cglobal hevc_idct_transpose_8x8, 0, 0, 0
+ ; M1 M2 ^T = M1^t M3^t
+ ; M3 M4 M2^t M4^t
+
+ ; M1 4x4 block
+ TRANSPOSE_BLOCK 0, 0, 16
+
+ ; M2 and M3
+ SWAP_BLOCKS 0, 64, 16, 0, 8
+
+ ; M4
+ TRANSPOSE_BLOCK 8, 64, 16
+
+ ret
+%endmacro
+
+; void ff_hevc_idct_8x8_{8,10}_<opt>(int16_t *coeffs, int col_limit)
+; %1 = bitdepth
+%macro IDCT_8x8 1
+cglobal hevc_idct_8x8_%1, 1, 1, 8, coeffs
+ TR_8x4 0, 7, 32, 1, 16, 8, 1
+ TR_8x4 8, 7, 32, 1, 16, 8, 1
+
+ call hevc_idct_transpose_8x8_ %+ cpuname
+
+ DEFINE_BIAS %1
+ TR_8x4 0, shift, 32, 1, 16, 8, 1
+ TR_8x4 8, shift, 32, 1, 16, 8, 1
+
+ TAIL_CALL hevc_idct_transpose_8x8_ %+ cpuname, 1
+%endmacro
+
+; store intermedite e32 coeffs on stack
+; as 16x4 matrix
+; from m10: e8 + o8, with %6 offset
+; and %3: e8 - o8, with %7 offset
+; %4 - shift, unused here
+%macro STORE_16 7
+ mova [rsp + %6], %5
+ mova [rsp + %7], %3
+%endmacro
+
+; %1, %2 - transform constants
+; %3, %4 - regs with interleaved coeffs
+; %5 - 1/0 SWAP or add
+; %6, %7 - registers for intermidiate sums
+; %8 - accumulator register
+%macro ADD_ROWS 8
+ pmaddwd %6, %3, %1
+ pmaddwd %7, %4, %2
+ paddd %6, %7
+%if %5 == 1
+ SWAP %6, %8
+%else
+ paddd %8, %6
+%endif
+%endmacro
+
+; %1 - transform coeffs
+; %2, %3 offsets for storing e+o/e-o back to coeffsq
+; %4 - shift
+; %5 - add
+; %6 - block_size
+; %7 - register with e16
+; %8, %9 - stack offsets for storing e+o/e-o
+%macro E16_O16 9
+ ADD_ROWS [%1], [%1 + 16], m0, m1, 1, m5, m6, m7
+ ADD_ROWS [%1 + 2 * 16], [%1 + 3 * 16], m2, m3, 0, m5, m6, m7
+
+%if %6 == 8
+ paddd %7, %5
+%endif
+
+ paddd m4, m7, %7 ; o16 + e16
+ psubd %7, m7 ; e16 - o16
+ STORE_%6 %2, %3, %7, %4, m4, %8, %9
+%endmacro
+
+%macro TR_16x4 10
+ ; produce 8x4 matrix of e16 coeffs
+ ; for 4 first rows and store it on stack (128 bytes)
+ TR_8x4 %1, 7, %4, %5, %6, %8, 0
+
+ ; load 8 even rows
+ LOAD_BLOCK m0, m1, %9 * %6, %9 * 3 * %6, %9 * 5 * %6, %9 * 7 * %6, %1
+ LOAD_BLOCK m2, m3, %9 * 9 * %6, %9 * 11 * %6, %9 * 13 * %6, %9 * 15 * %6, %1
+
+ SBUTTERFLY wd, 0, 1, 4
+ SBUTTERFLY wd, 2, 3, 4
+
+ E16_O16 trans_coeffs16, 0 + %1, 15 * %6 + %1, %2, %3, %7, m8, 0, 15 * 16
+ mova m8, %3
+ E16_O16 trans_coeffs16 + 64, %6 + %1, 14 * %6 + %1, %2, m8, %7, m9, 16, 14 * 16
+ E16_O16 trans_coeffs16 + 2 * 64, 2 * %6 + %1, 13 * %6 + %1, %2, m8, %7, m10, 2 * 16, 13 * 16
+ E16_O16 trans_coeffs16 + 3 * 64, 3 * %6 + %1, 12 * %6 + %1, %2, m8, %7, m11, 3 * 16, 12 * 16
+ E16_O16 trans_coeffs16 + 4 * 64, 4 * %6 + %1, 11 * %6 + %1, %2, m8, %7, m12, 4 * 16, 11 * 16
+ E16_O16 trans_coeffs16 + 5 * 64, 5 * %6 + %1, 10 * %6 + %1, %2, m8, %7, m13, 5 * 16, 10 * 16
+ E16_O16 trans_coeffs16 + 6 * 64, 6 * %6 + %1, 9 * %6 + %1, %2, m8, %7, m14, 6 * 16, 9 * 16
+ E16_O16 trans_coeffs16 + 7 * 64, 7 * %6 + %1, 8 * %6 + %1, %2, m8, %7, m15, 7 * 16, 8 * 16
+%endmacro
+
+%macro TRANSPOSE_16x16 0
+cglobal hevc_idct_transpose_16x16, 0, 0, 0
+; M1 M2 M3 M4 ^T m1 m5 m9 m13 M_i^T = m_i
+; M5 M6 M7 M8 --> m2 m6 m10 m14
+; M9 M10 M11 M12 m3 m7 m11 m15
+; M13 M14 M15 M16 m4 m8 m12 m16
+
+ ; M1 4x4 block
+ TRANSPOSE_BLOCK 0, 0, 32
+
+ ; M5, M2
+ SWAP_BLOCKS 0, 128, 32, 0, 8
+ ; M9, M3
+ SWAP_BLOCKS 0, 256, 32, 0, 16
+ ; M13, M4
+ SWAP_BLOCKS 0, 384, 32, 0, 24
+
+ ;M6
+ TRANSPOSE_BLOCK 8, 128, 32
+
+ ; M10, M7
+ SWAP_BLOCKS 8, 256, 32, 128, 16
+ ; M14, M8
+ SWAP_BLOCKS 8, 384, 32, 128, 24
+
+ ;M11
+ TRANSPOSE_BLOCK 16, 256, 32
+
+ ; M15, M12
+ SWAP_BLOCKS 16, 384, 32, 256, 24
+
+ ;M16
+ TRANSPOSE_BLOCK 24, 384, 32
+
+ ret
+%endmacro
+
+; void ff_hevc_idct_16x16_{8,10}_<opt>(int16_t *coeffs, int col_limit)
+; %1 = bitdepth
+%macro IDCT_16x16 1
+cglobal hevc_idct_16x16_%1, 1, 2, 16, coeffs
+ mov r1d, 3
+.loop16:
+ TR_16x4 8 * r1, 7, [pd_64], 64, 2, 32, 8, 16, 1, 0
+ dec r1d
+ jge .loop16
+
+ call hevc_idct_transpose_16x16_ %+ cpuname
+
+ DEFINE_BIAS %1
+ mov r1d, 3
+.loop16_2:
+ TR_16x4 8 * r1, shift, [arr_add], 64, 2, 32, 8, 16, 1, 1
+ dec r1d
+ jge .loop16_2
+
+ TAIL_CALL hevc_idct_transpose_16x16_ %+ cpuname, 1
+%endmacro
+
+; scale, pack (clip16) and store the residuals 0 e32[0] + o32[0] --> %1
+; 4 at one time (4 columns) 1 e32[1] + o32[1]
+; %1 - address to store e32 + o32
+; %2 - address to store e32 - e32
+; %5 - reg with e32 + o32 ...
+; %3 - reg with e32 - o32 30 e32[1] - o32[1]
+; %4 - shift 31 e32[0] - o32[0] --> %2
+%macro STORE_32 5
+ psrad %5, %4
+ psrad %3, %4
+ packssdw %5, %3
+ movq [%1], %5
+ movhps [%2], %5
+%endmacro
+
+; %1 - transform coeffs
+; %2 - stack offset for e32
+; %2, %3 offsets for storing e+o/e-o back to coeffsq
+; %4 - shift
+; %5 - stack offset of e32
+%macro E32_O32 5
+ ADD_ROWS [%1], [%1 + 16], m0, m1, 1, m8, m9, m10
+ ADD_ROWS [%1 + 2 * 16], [%1 + 3 * 16], m2, m3, 0, m8, m9, m10
+ ADD_ROWS [%1 + 4 * 16], [%1 + 5 * 16], m4, m5, 0, m8, m9, m10
+ ADD_ROWS [%1 + 6 * 16], [%1 + 7 * 16], m6, m7, 0, m8, m9, m10
+
+ paddd m11, m14, [rsp + %5]
+ paddd m12, m10, m11 ; o32 + e32
+ psubd m11, m10 ; e32 - o32
+ STORE_32 %2, %3, m11, %4, m12
+%endmacro
+
+; %1 - horizontal offset
+; %2 - bitdepth
+%macro TR_32x4 3
+ TR_16x4 %1, 7, [pd_64], 128, 4, 64, 16, 16, 2, 0
+
+ LOAD_BLOCK m0, m1, 64, 3 * 64, 5 * 64, 7 * 64, %1
+ LOAD_BLOCK m2, m3, 9 * 64, 11 * 64, 13 * 64, 15 * 64, %1
+ LOAD_BLOCK m4, m5, 17 * 64, 19 * 64, 21 * 64, 23 * 64, %1
+ LOAD_BLOCK m6, m7, 25 * 64, 27 * 64, 29 * 64, 31 * 64, %1
+
+ SBUTTERFLY wd, 0, 1, 8
+ SBUTTERFLY wd, 2, 3, 8
+ SBUTTERFLY wd, 4, 5, 8
+ SBUTTERFLY wd, 6, 7, 8
+
+%if %3 == 1
+ %assign shift 7
+ mova m14, [pd_64]
+%else
+ LOAD_BIAS %2, m14
+%endif
+
+ lea r2, [trans_coeff32 + 15 * 128]
+ lea r3, [coeffsq + %1]
+ lea r4, [r3 + 16 * 64]
+ mov r5d, 15 * 16
+%%loop:
+ E32_O32 r2, r3 + r5 * 4, r4, shift, r5
+ sub r2, 128
+ add r4, 64
+ sub r5d, 16
+ jge %%loop
+%endmacro
+
+%macro TRANSPOSE_32x32 0
+cglobal hevc_idct_transpose_32x32, 0, 0, 0
+ ; M0 M1 ... M7
+ ; M8 M15
+ ;
+ ; ...
+ ;
+ ; M56 M63
+
+ TRANSPOSE_BLOCK 0, 0, 64 ; M1
+ mov r1d, 7
+ mov r2d, 7 * 256
+.loop_transpose:
+ SWAP_BLOCKS 0, r2, 64, 0, r1 * 8
+ sub r2d, 256
+ dec r1d
+ jg .loop_transpose
+
+ TRANSPOSE_BLOCK 8, 256, 64 ; M9
+ mov r1d, 6
+ mov r2d, 512
+ mov r3d, 16
+.loop_transpose2:
+ SWAP_BLOCKS 8, r2, 64, 256, r3
+ add r3d, 8
+ add r2d, 256
+ dec r1d
+ jg .loop_transpose2
+
+ TRANSPOSE_BLOCK 2 * 8, 2 * 256, 64 ; M9
+ mov r1d, 5
+ mov r2d, 768
+ mov r3d, 24
+.loop_transpose3:
+ SWAP_BLOCKS 2 * 8, r2, 64, 2 * 256, r3
+ add r3d, 8
+ add r2d, 256
+ dec r1d
+ jg .loop_transpose3
+
+ TRANSPOSE_BLOCK 3 * 8, 3 * 256, 64 ; M27
+ mov r1d, 4
+ mov r2d, 1024
+ mov r3d, 32
+.loop_transpose4:
+ SWAP_BLOCKS 3 * 8, r2, 64, 3 * 256, r3
+ add r3d, 8
+ add r2d, 256
+ dec r1d
+ jg .loop_transpose4
+
+ TRANSPOSE_BLOCK 4 * 8, 4 * 256, 64 ; M36
+ mov r1d, 3
+ mov r2d, 1280
+ mov r3d, 40
+.loop_transpose5:
+ SWAP_BLOCKS 4 * 8, r2, 64, 4 * 256, r3
+ add r3d, 8
+ add r2d, 256
+ dec r1d
+ jg .loop_transpose5
+
+ TRANSPOSE_BLOCK 5 * 8, 5 * 256, 64 ; M45
+ SWAP_BLOCKS 5 * 8, 6 * 256, 64, 5 * 256, 6 * 8
+ SWAP_BLOCKS 5 * 8, 7 * 256, 64, 5 * 256, 7 * 8
+
+ TRANSPOSE_BLOCK 6 * 8, 6 * 256, 64 ; M54
+ SWAP_BLOCKS 6 * 8, 7 * 256, 64, 6 * 256, 7 * 8
+
+ TRANSPOSE_BLOCK 7 * 8, 7 * 256, 64 ; M63
+
+ ret
+%endmacro
+
+; void ff_hevc_idct_32x32_{8,10}_<opt>(int16_t *coeffs, int col_limit)
+; %1 = bitdepth
+%macro IDCT_32x32 1
+cglobal hevc_idct_32x32_%1, 1, 6, 16, 256, coeffs
+ mov r1d, 7
+.loop32:
+ TR_32x4 8 * r1, %1, 1
+ dec r1d
+ jge .loop32
+
+ call hevc_idct_transpose_32x32_ %+ cpuname
+
+ mov r1d, 7
+.loop32_2:
+ TR_32x4 8 * r1, %1, 0
+ dec r1d
+ jge .loop32_2
+
+ TAIL_CALL hevc_idct_transpose_32x32_ %+ cpuname, 1
+%endmacro
+
+%macro INIT_IDCT_DC 1
INIT_MMX mmxext
-IDCT_DC_NL 4, 10
-IDCT_DC 8, 2, 10
+IDCT_DC_NL 4, %1
+IDCT_DC 8, 2, %1
INIT_XMM sse2
-IDCT_DC_NL 8, 10
-IDCT_DC 16, 4, 10
-IDCT_DC 32, 16, 10
+IDCT_DC_NL 8, %1
+IDCT_DC 16, 4, %1
+IDCT_DC 32, 16, %1
%if HAVE_AVX2_EXTERNAL
-INIT_YMM avx2
-IDCT_DC 16, 2, 10
-IDCT_DC 32, 8, 10
+ INIT_YMM avx2
+ IDCT_DC 16, 2, %1
+ IDCT_DC 32, 8, %1
%endif ;HAVE_AVX2_EXTERNAL
+%endmacro
+
+%macro INIT_IDCT 2
+INIT_XMM %2
+%if %1 == 8
+ TRANSPOSE_8x8
+ %if ARCH_X86_64
+ TRANSPOSE_16x16
+ TRANSPOSE_32x32
+ %endif
+%endif
+%if ARCH_X86_64
+ IDCT_32x32 %1
+ IDCT_16x16 %1
+%endif
+IDCT_8x8 %1
+IDCT_4x4 %1
+%endmacro
+
+INIT_IDCT_DC 8
+INIT_IDCT_DC 10
+INIT_IDCT 8, sse2
+INIT_IDCT 8, avx
+INIT_IDCT 10, sse2
+INIT_IDCT 10, avx