aboutsummaryrefslogtreecommitdiffstats
path: root/libavcodec/x86/hevc_add_res.asm
diff options
context:
space:
mode:
authorClément Bœsch <u@pkh.me>2017-03-24 12:29:21 +0100
committerClément Bœsch <u@pkh.me>2017-03-24 12:33:25 +0100
commit3d6535983282bea542dac2e568ae50da5796be34 (patch)
tree0dae743f60adad35e1bd088d1447eb99cc988d91 /libavcodec/x86/hevc_add_res.asm
parent40ac226014082682e504e8af613ed4169221ac36 (diff)
parent6d5636ad9ab6bd9bedf902051d88b7044385f88b (diff)
downloadffmpeg-3d6535983282bea542dac2e568ae50da5796be34.tar.gz
Merge commit '6d5636ad9ab6bd9bedf902051d88b7044385f88b'
* commit '6d5636ad9ab6bd9bedf902051d88b7044385f88b': hevc: x86: Add add_residual() SIMD optimizations See a6af4bf64dae46356a5f91537a1c8c5f86456b37 This merge is only cosmetics (renames, space shuffling, etc). The functionnal changes in the ASM are *not* merged: - unrolling with %rep is kept - ADD_RES_MMX_4_8 is left untouched: this needs investigation Merged-by: Clément Bœsch <u@pkh.me>
Diffstat (limited to 'libavcodec/x86/hevc_add_res.asm')
-rw-r--r--libavcodec/x86/hevc_add_res.asm269
1 files changed, 131 insertions, 138 deletions
diff --git a/libavcodec/x86/hevc_add_res.asm b/libavcodec/x86/hevc_add_res.asm
index 869288faf3..1ea15df0bd 100644
--- a/libavcodec/x86/hevc_add_res.asm
+++ b/libavcodec/x86/hevc_add_res.asm
@@ -1,4 +1,4 @@
-; /*
+; *****************************************************************************
; * Provide SIMD optimizations for add_residual functions for HEVC decoding
; * Copyright (c) 2014 Pierre-Edouard LEPERE
; *
@@ -17,7 +17,8 @@
; * You should have received a copy of the GNU Lesser General Public
; * License along with FFmpeg; if not, write to the Free Software
; * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
-; */
+; ******************************************************************************
+
%include "libavutil/x86/x86util.asm"
SECTION .text
@@ -25,9 +26,8 @@ SECTION .text
cextern pw_1023
%define max_pixels_10 pw_1023
-
-;the tr_add macros and functions were largely inspired by x264 project's code in the h264_idct.asm file
-%macro TR_ADD_MMX_4_8 0
+; the add_res macros and functions were largely inspired by h264_idct.asm from the x264 project
+%macro ADD_RES_MMX_4_8 0
mova m2, [r1]
mova m4, [r1+8]
pxor m3, m3
@@ -39,27 +39,27 @@ cextern pw_1023
packuswb m4, m4
packuswb m5, m5
- movh m0, [r0 ]
- movh m1, [r0+r2 ]
+ movh m0, [r0]
+ movh m1, [r0+r2]
paddusb m0, m2
paddusb m1, m4
psubusb m0, m3
psubusb m1, m5
- movh [r0 ], m0
- movh [r0+r2 ], m1
+ movh [r0], m0
+ movh [r0+r2], m1
%endmacro
INIT_MMX mmxext
-; void ff_hevc_tranform_add_8_mmxext(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
-cglobal hevc_add_residual4_8, 3, 4, 6
- TR_ADD_MMX_4_8
+; void ff_hevc_add_residual_4_8_mmxext(uint8_t *dst, int16_t *res, ptrdiff_t stride)
+cglobal hevc_add_residual_4_8, 3, 4, 6
+ ADD_RES_MMX_4_8
add r1, 16
lea r0, [r0+r2*2]
- TR_ADD_MMX_4_8
+ ADD_RES_MMX_4_8
RET
-%macro TR_ADD_SSE_8_8 0
+%macro ADD_RES_SSE_8_8 0
pxor m3, m3
mova m4, [r1]
mova m6, [r1+16]
@@ -74,22 +74,22 @@ cglobal hevc_add_residual4_8, 3, 4, 6
packuswb m6, m2
packuswb m7, m3
- movq m0, [r0 ]
- movq m1, [r0+r2 ]
- movhps m0, [r0+r2*2]
- movhps m1, [r0+r3 ]
- paddusb m0, m4
- paddusb m1, m6
- psubusb m0, m5
- psubusb m1, m7
- movq [r0 ], m0
- movq [r0+r2 ], m1
- movhps [r0+2*r2], m0
- movhps [r0+r3 ], m1
+ movq m0, [r0]
+ movq m1, [r0+r2]
+ movhps m0, [r0+r2*2]
+ movhps m1, [r0+r3]
+ paddusb m0, m4
+ paddusb m1, m6
+ psubusb m0, m5
+ psubusb m1, m7
+ movq [r0], m0
+ movq [r0+r2], m1
+ movhps [r0+2*r2], m0
+ movhps [r0+r3], m1
%endmacro
-%macro TR_ADD_SSE_16_32_8 3
- mova xm2, [r1+%1 ]
+%macro ADD_RES_SSE_16_32_8 3
+ mova xm2, [r1+%1]
mova xm6, [r1+%1+16]
%if cpuflag(avx2)
vinserti128 m2, m2, [r1+%1+32], 1
@@ -107,7 +107,7 @@ cglobal hevc_add_residual4_8, 3, 4, 6
packuswb m2, m6
packuswb m1, m5
- mova xm4, [r1+%1+mmsize*2 ]
+ mova xm4, [r1+%1+mmsize*2]
mova xm6, [r1+%1+mmsize*2+16]
%if cpuflag(avx2)
vinserti128 m4, m4, [r1+%1+96 ], 1
@@ -135,39 +135,39 @@ cglobal hevc_add_residual4_8, 3, 4, 6
%macro TRANSFORM_ADD_8 0
-; void ff_hevc_add_residual8_8_<opt>(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
-cglobal hevc_add_residual8_8, 3, 4, 8
+; void ff_hevc_add_residual_8_8_<opt>(uint8_t *dst, int16_t *res, ptrdiff_t stride)
+cglobal hevc_add_residual_8_8, 3, 4, 8
lea r3, [r2*3]
- TR_ADD_SSE_8_8
+ ADD_RES_SSE_8_8
add r1, 64
lea r0, [r0+r2*4]
- TR_ADD_SSE_8_8
+ ADD_RES_SSE_8_8
RET
-; void ff_hevc_add_residual16_8_<opt>(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
-cglobal hevc_add_residual16_8, 3, 4, 7
- pxor m0, m0
- lea r3, [r2*3]
- TR_ADD_SSE_16_32_8 0, r0, r0+r2
- TR_ADD_SSE_16_32_8 64, r0+r2*2, r0+r3
+; void ff_hevc_add_residual_16_8_<opt>(uint8_t *dst, int16_t *res, ptrdiff_t stride)
+cglobal hevc_add_residual_16_8, 3, 4, 7
+ pxor m0, m0
+ lea r3, [r2*3]
+ ADD_RES_SSE_16_32_8 0, r0, r0+r2
+ ADD_RES_SSE_16_32_8 64, r0+r2*2, r0+r3
%rep 3
- add r1, 128
- lea r0, [r0+r2*4]
- TR_ADD_SSE_16_32_8 0, r0, r0+r2
- TR_ADD_SSE_16_32_8 64, r0+r2*2, r0+r3
+ add r1, 128
+ lea r0, [r0+r2*4]
+ ADD_RES_SSE_16_32_8 0, r0, r0+r2
+ ADD_RES_SSE_16_32_8 64, r0+r2*2, r0+r3
%endrep
RET
-; void ff_hevc_add_residual32_8_<opt>(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
-cglobal hevc_add_residual32_8, 3, 4, 7
- pxor m0, m0
- TR_ADD_SSE_16_32_8 0, r0, r0+16
- TR_ADD_SSE_16_32_8 64, r0+r2, r0+r2+16
+; void ff_hevc_add_residual_32_8_<opt>(uint8_t *dst, int16_t *res, ptrdiff_t stride)
+cglobal hevc_add_residual_32_8, 3, 4, 7
+ pxor m0, m0
+ ADD_RES_SSE_16_32_8 0, r0, r0+16
+ ADD_RES_SSE_16_32_8 64, r0+r2, r0+r2+16
%rep 15
- add r1, 128
- lea r0, [r0+r2*2]
- TR_ADD_SSE_16_32_8 0, r0, r0+16
- TR_ADD_SSE_16_32_8 64, r0+r2, r0+r2+16
+ add r1, 128
+ lea r0, [r0+r2*2]
+ ADD_RES_SSE_16_32_8 0, r0, r0+16
+ ADD_RES_SSE_16_32_8 64, r0+r2, r0+r2+16
%endrep
RET
%endmacro
@@ -179,80 +179,77 @@ TRANSFORM_ADD_8
%if HAVE_AVX2_EXTERNAL
INIT_YMM avx2
-; void ff_hevc_add_residual32_8_avx2(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
-cglobal hevc_add_residual32_8, 3, 4, 7
- pxor m0, m0
- lea r3, [r2*3]
- TR_ADD_SSE_16_32_8 0, r0, r0+r2
- TR_ADD_SSE_16_32_8 128, r0+r2*2, r0+r3
+; void ff_hevc_add_residual_32_8_avx2(uint8_t *dst, int16_t *res, ptrdiff_t stride)
+cglobal hevc_add_residual_32_8, 3, 4, 7
+ pxor m0, m0
+ lea r3, [r2*3]
+ ADD_RES_SSE_16_32_8 0, r0, r0+r2
+ ADD_RES_SSE_16_32_8 128, r0+r2*2, r0+r3
%rep 7
- add r1, 256
- lea r0, [r0+r2*4]
- TR_ADD_SSE_16_32_8 0, r0, r0+r2
- TR_ADD_SSE_16_32_8 128, r0+r2*2, r0+r3
+ add r1, 256
+ lea r0, [r0+r2*4]
+ ADD_RES_SSE_16_32_8 0, r0, r0+r2
+ ADD_RES_SSE_16_32_8 128, r0+r2*2, r0+r3
%endrep
RET
%endif
-;-----------------------------------------------------------------------------
-; void ff_hevc_add_residual_10(pixel *dst, int16_t *block, int stride)
-;-----------------------------------------------------------------------------
-%macro TR_ADD_SSE_8_10 4
+%macro ADD_RES_SSE_8_10 4
mova m0, [%4]
mova m1, [%4+16]
mova m2, [%4+32]
mova m3, [%4+48]
- paddw m0, [%1+0 ]
- paddw m1, [%1+%2 ]
+ paddw m0, [%1+0]
+ paddw m1, [%1+%2]
paddw m2, [%1+%2*2]
- paddw m3, [%1+%3 ]
+ paddw m3, [%1+%3]
CLIPW m0, m4, m5
CLIPW m1, m4, m5
CLIPW m2, m4, m5
CLIPW m3, m4, m5
- mova [%1+0 ], m0
- mova [%1+%2 ], m1
+ mova [%1+0], m0
+ mova [%1+%2], m1
mova [%1+%2*2], m2
- mova [%1+%3 ], m3
+ mova [%1+%3], m3
%endmacro
-%macro TR_ADD_MMX4_10 3
- mova m0, [%1+0 ]
- mova m1, [%1+%2 ]
+%macro ADD_RES_MMX_4_10 3
+ mova m0, [%1+0]
+ mova m1, [%1+%2]
paddw m0, [%3]
paddw m1, [%3+8]
CLIPW m0, m2, m3
CLIPW m1, m2, m3
- mova [%1+0 ], m0
- mova [%1+%2 ], m1
+ mova [%1+0], m0
+ mova [%1+%2], m1
%endmacro
-%macro TRANS_ADD_SSE_16_10 3
+%macro ADD_RES_SSE_16_10 3
mova m0, [%3]
mova m1, [%3+16]
mova m2, [%3+32]
mova m3, [%3+48]
- paddw m0, [%1 ]
- paddw m1, [%1+16 ]
- paddw m2, [%1+%2 ]
+ paddw m0, [%1]
+ paddw m1, [%1+16]
+ paddw m2, [%1+%2]
paddw m3, [%1+%2+16]
CLIPW m0, m4, m5
CLIPW m1, m4, m5
CLIPW m2, m4, m5
CLIPW m3, m4, m5
- mova [%1 ], m0
- mova [%1+16 ], m1
- mova [%1+%2 ], m2
+ mova [%1], m0
+ mova [%1+16], m1
+ mova [%1+%2], m2
mova [%1+%2+16], m3
%endmacro
-%macro TRANS_ADD_SSE_32_10 2
+%macro ADD_RES_SSE_32_10 2
mova m0, [%2]
mova m1, [%2+16]
mova m2, [%2+32]
mova m3, [%2+48]
- paddw m0, [%1 ]
+ paddw m0, [%1]
paddw m1, [%1+16]
paddw m2, [%1+32]
paddw m3, [%1+48]
@@ -260,129 +257,125 @@ cglobal hevc_add_residual32_8, 3, 4, 7
CLIPW m1, m4, m5
CLIPW m2, m4, m5
CLIPW m3, m4, m5
- mova [%1 ], m0
+ mova [%1], m0
mova [%1+16], m1
mova [%1+32], m2
mova [%1+48], m3
%endmacro
-%macro TRANS_ADD16_AVX2 4
+%macro ADD_RES_AVX2_16_10 4
mova m0, [%4]
mova m1, [%4+32]
mova m2, [%4+64]
mova m3, [%4+96]
- paddw m0, [%1+0 ]
- paddw m1, [%1+%2 ]
+ paddw m0, [%1+0]
+ paddw m1, [%1+%2]
paddw m2, [%1+%2*2]
- paddw m3, [%1+%3 ]
+ paddw m3, [%1+%3]
CLIPW m0, m4, m5
CLIPW m1, m4, m5
CLIPW m2, m4, m5
CLIPW m3, m4, m5
- mova [%1+0 ], m0
- mova [%1+%2 ], m1
+ mova [%1+0], m0
+ mova [%1+%2], m1
mova [%1+%2*2], m2
- mova [%1+%3 ], m3
+ mova [%1+%3], m3
%endmacro
-%macro TRANS_ADD32_AVX2 3
+%macro ADD_RES_AVX2_32_10 3
mova m0, [%3]
mova m1, [%3+32]
mova m2, [%3+64]
mova m3, [%3+96]
- paddw m0, [%1 ]
- paddw m1, [%1+32 ]
- paddw m2, [%1+%2 ]
+ paddw m0, [%1]
+ paddw m1, [%1+32]
+ paddw m2, [%1+%2]
paddw m3, [%1+%2+32]
CLIPW m0, m4, m5
CLIPW m1, m4, m5
CLIPW m2, m4, m5
CLIPW m3, m4, m5
- mova [%1 ], m0
- mova [%1+32 ], m1
- mova [%1+%2 ], m2
+ mova [%1], m0
+ mova [%1+32], m1
+ mova [%1+%2], m2
mova [%1+%2+32], m3
%endmacro
-
+; void ff_hevc_add_residual_<4|8|16|32>_10(pixel *dst, int16_t *block, ptrdiff_t stride)
INIT_MMX mmxext
-cglobal hevc_add_residual4_10,3,4, 6
+cglobal hevc_add_residual_4_10, 3, 4, 6
pxor m2, m2
mova m3, [max_pixels_10]
- TR_ADD_MMX4_10 r0, r2, r1
+ ADD_RES_MMX_4_10 r0, r2, r1
add r1, 16
lea r0, [r0+2*r2]
- TR_ADD_MMX4_10 r0, r2, r1
+ ADD_RES_MMX_4_10 r0, r2, r1
RET
-;-----------------------------------------------------------------------------
-; void ff_hevc_add_residual_10(pixel *dst, int16_t *block, int stride)
-;-----------------------------------------------------------------------------
INIT_XMM sse2
-cglobal hevc_add_residual8_10,3,4,6
+cglobal hevc_add_residual_8_10, 3, 4, 6
pxor m4, m4
mova m5, [max_pixels_10]
lea r3, [r2*3]
- TR_ADD_SSE_8_10 r0, r2, r3, r1
+ ADD_RES_SSE_8_10 r0, r2, r3, r1
lea r0, [r0+r2*4]
add r1, 64
- TR_ADD_SSE_8_10 r0, r2, r3, r1
+ ADD_RES_SSE_8_10 r0, r2, r3, r1
RET
-cglobal hevc_add_residual16_10,3,4,6
+cglobal hevc_add_residual_16_10, 3, 4, 6
pxor m4, m4
mova m5, [max_pixels_10]
- TRANS_ADD_SSE_16_10 r0, r2, r1
+ ADD_RES_SSE_16_10 r0, r2, r1
%rep 7
- lea r0, [r0+r2*2]
- add r1, 64
- TRANS_ADD_SSE_16_10 r0, r2, r1
+ lea r0, [r0+r2*2]
+ add r1, 64
+ ADD_RES_SSE_16_10 r0, r2, r1
%endrep
RET
-cglobal hevc_add_residual32_10,3,4,6
+cglobal hevc_add_residual_32_10, 3, 4, 6
pxor m4, m4
mova m5, [max_pixels_10]
- TRANS_ADD_SSE_32_10 r0, r1
+ ADD_RES_SSE_32_10 r0, r1
%rep 31
- lea r0, [r0+r2]
- add r1, 64
- TRANS_ADD_SSE_32_10 r0, r1
+ lea r0, [r0+r2]
+ add r1, 64
+ ADD_RES_SSE_32_10 r0, r1
%endrep
RET
%if HAVE_AVX2_EXTERNAL
INIT_YMM avx2
+cglobal hevc_add_residual_16_10, 3, 4, 6
+ pxor m4, m4
+ mova m5, [max_pixels_10]
+ lea r3, [r2*3]
-cglobal hevc_add_residual16_10,3,4,6
- pxor m4, m4
- mova m5, [max_pixels_10]
- lea r3, [r2*3]
-
- TRANS_ADD16_AVX2 r0, r2, r3, r1
+ ADD_RES_AVX2_16_10 r0, r2, r3, r1
%rep 3
- lea r0, [r0+r2*4]
- add r1, 128
- TRANS_ADD16_AVX2 r0, r2, r3, r1
+ lea r0, [r0+r2*4]
+ add r1, 128
+ ADD_RES_AVX2_16_10 r0, r2, r3, r1
%endrep
RET
-cglobal hevc_add_residual32_10,3,4,6
- pxor m4, m4
- mova m5, [max_pixels_10]
+cglobal hevc_add_residual_32_10, 3, 4, 6
+ pxor m4, m4
+ mova m5, [max_pixels_10]
- TRANS_ADD32_AVX2 r0, r2, r1
+ ADD_RES_AVX2_32_10 r0, r2, r1
%rep 15
- lea r0, [r0+r2*2]
- add r1, 128
- TRANS_ADD32_AVX2 r0, r2, r1
+ lea r0, [r0+r2*2]
+ add r1, 128
+ ADD_RES_AVX2_32_10 r0, r2, r1
%endrep
RET
-%endif ;HAVE_AVX_EXTERNAL
+%endif ;HAVE_AVX2_EXTERNAL