diff options
author | Christophe Gisquet <christophe.gisquet@gmail.com> | 2015-10-12 19:37:44 +0200 |
---|---|---|
committer | Michael Niedermayer <michael@niedermayer.cc> | 2015-10-13 01:10:34 +0200 |
commit | e9a68b0316ab127098ac4c24a6762ce68980bd23 (patch) | |
tree | c981ce654068e58b938e134e3557ff57e6b0eca4 | |
parent | 93e6b23c9f489c1e0745769ae2ba00469f7cfb91 (diff) | |
download | ffmpeg-e9a68b0316ab127098ac4c24a6762ce68980bd23.tar.gz |
x86: prores: templatize 10 bits simple_idct
This should be reused for a generic simple_idct10 function.
Requires a bit of trickery to declare common constants in C.
Signed-off-by: Michael Niedermayer <michael@niedermayer.cc>
-rw-r--r-- | libavcodec/x86/constants.c | 28 | ||||
-rw-r--r-- | libavcodec/x86/constants.h | 16 | ||||
-rw-r--r-- | libavcodec/x86/proresdsp.asm | 263 | ||||
-rw-r--r-- | libavcodec/x86/simple_idct10_template.asm | 299 |
4 files changed, 346 insertions, 260 deletions
diff --git a/libavcodec/x86/constants.c b/libavcodec/x86/constants.c index 3f3ee0ffc1..9592fa73e1 100644 --- a/libavcodec/x86/constants.c +++ b/libavcodec/x86/constants.c @@ -87,3 +87,31 @@ DECLARE_ALIGNED(32, const ymm_reg, ff_pd_32) = { 0x0000002000000020ULL, 0x000 0x0000002000000020ULL, 0x0000002000000020ULL }; DECLARE_ALIGNED(32, const ymm_reg, ff_pd_65535)= { 0x0000ffff0000ffffULL, 0x0000ffff0000ffffULL, 0x0000ffff0000ffffULL, 0x0000ffff0000ffffULL }; + +/* simple idct 10 */ +#define W1sh2 22725 // W1 = 90901 = 22725<<2 + 1 +#define W2sh2 21407 // W2 = 85627 = 21407<<2 - 1 +#define W3sh2 19265 // W3 = 77062 = 19265<<2 + 2 +#define W4sh2 16384 // W4 = 65535 = 16384<<2 - 1 +#define W5sh2 12873 // W5 = 51491 = 12873<<2 - 1 +#define W6sh2 8867 // W6 = 35468 = 8867<<2 +#define W7sh2 4520 // W7 = 18081 = 4520<<2 + 1 + +#define TIMES4(a, b) \ + (((b)&0xFFFFLL)<<48)+(((a)&0xFFFFLL)<<32)+(((b)&0xFFFFLL)<<16)+((a)&0xFFFFLL), \ + (((b)&0xFFFFLL)<<48)+(((a)&0xFFFFLL)<<32)+(((b)&0xFFFFLL)<<16)+((a)&0xFFFFLL) + +#if ARCH_X86_64 +DECLARE_ALIGNED(16, const xmm_reg, ff_w4_plus_w2) = { TIMES4(W4sh2, W2sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w4_min_w2) = { TIMES4(W4sh2, -W2sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w4_plus_w6) = { TIMES4(W4sh2, +W6sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w4_min_w6) = { TIMES4(W4sh2, -W6sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w1_plus_w3) = { TIMES4(W1sh2, +W3sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w3_min_w1) = { TIMES4(W3sh2, -W1sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w7_plus_w3) = { TIMES4(W7sh2, +W3sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w3_min_w7) = { TIMES4(W3sh2, -W7sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w1_plus_w5) = { TIMES4(W1sh2, +W5sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w5_min_w1) = { TIMES4(W5sh2, -W1sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w5_plus_w7) = { TIMES4(W5sh2, +W7sh2) }; +DECLARE_ALIGNED(16, const xmm_reg, ff_w7_min_w5) = { TIMES4(W7sh2, -W5sh2) }; +#endif diff --git a/libavcodec/x86/constants.h b/libavcodec/x86/constants.h index ee8422eeb3..3605b63504 100644 --- a/libavcodec/x86/constants.h +++ b/libavcodec/x86/constants.h @@ -67,4 +67,20 @@ extern const ymm_reg ff_pd_16; extern const ymm_reg ff_pd_32; extern const ymm_reg ff_pd_65535; +# if ARCH_X86_64 +/* simple_idct10, used by prores and dnxhd */ +extern const xmm_reg ff_w4_plus_w2; +extern const xmm_reg ff_w4_min_w2; +extern const xmm_reg ff_w4_plus_w6; +extern const xmm_reg ff_w4_min_w6; +extern const xmm_reg ff_w1_plus_w3; +extern const xmm_reg ff_w3_min_w1; +extern const xmm_reg ff_w7_plus_w3; +extern const xmm_reg ff_w3_min_w7; +extern const xmm_reg ff_w1_plus_w5; +extern const xmm_reg ff_w5_min_w1; +extern const xmm_reg ff_w5_plus_w7; +extern const xmm_reg ff_w7_min_w5; +# endif + #endif /* AVCODEC_X86_CONSTANTS_H */ diff --git a/libavcodec/x86/proresdsp.asm b/libavcodec/x86/proresdsp.asm index 632ece6ebf..18cf15b3ca 100644 --- a/libavcodec/x86/proresdsp.asm +++ b/libavcodec/x86/proresdsp.asm @@ -24,279 +24,22 @@ %include "libavutil/x86/x86util.asm" -%define W1sh2 22725 ; W1 = 90901 = 22725<<2 + 1 -%define W2sh2 21407 ; W2 = 85627 = 21407<<2 - 1 -%define W3sh2 19265 ; W3 = 77062 = 19265<<2 + 2 -%define W4sh2 16384 ; W4 = 65535 = 16384<<2 - 1 -%define W5sh2 12873 ; W5 = 51491 = 12873<<2 - 1 -%define W6sh2 8867 ; W6 = 35468 = 8867<<2 -%define W7sh2 4520 ; W7 = 18081 = 4520<<2 + 1 - %if ARCH_X86_64 SECTION_RODATA -w4_plus_w2: times 4 dw W4sh2, +W2sh2 -w4_min_w2: times 4 dw W4sh2, -W2sh2 -w4_plus_w6: times 4 dw W4sh2, +W6sh2 -w4_min_w6: times 4 dw W4sh2, -W6sh2 -w1_plus_w3: times 4 dw W1sh2, +W3sh2 -w3_min_w1: times 4 dw W3sh2, -W1sh2 -w7_plus_w3: times 4 dw W7sh2, +W3sh2 -w3_min_w7: times 4 dw W3sh2, -W7sh2 -w1_plus_w5: times 4 dw W1sh2, +W5sh2 -w5_min_w1: times 4 dw W5sh2, -W1sh2 -w5_plus_w7: times 4 dw W5sh2, +W7sh2 -w7_min_w5: times 4 dw W7sh2, -W5sh2 pw_88: times 8 dw 0x2008 - cextern pw_1 cextern pw_4 -cextern pw_512 cextern pw_1019 -section .text align=16 - -; interleave data while maintaining source -; %1=type, %2=dstlo, %3=dsthi, %4=src, %5=interleave -%macro SBUTTERFLY3 5 - punpckl%1 m%2, m%4, m%5 - punpckh%1 m%3, m%4, m%5 -%endmacro - -; %1/%2=src1/dst1, %3/%4=dst2, %5/%6=src2, %7=shift -; action: %3/%4 = %1/%2 - %5/%6; %1/%2 += %5/%6 -; %1/%2/%3/%4 >>= %7; dword -> word (in %1/%3) -%macro SUMSUB_SHPK 7 - psubd %3, %1, %5 ; { a0 - b0 }[0-3] - psubd %4, %2, %6 ; { a0 - b0 }[4-7] - paddd %1, %5 ; { a0 + b0 }[0-3] - paddd %2, %6 ; { a0 + b0 }[4-7] - psrad %1, %7 - psrad %2, %7 - psrad %3, %7 - psrad %4, %7 - packssdw %1, %2 ; row[0] - packssdw %3, %4 ; row[7] -%endmacro - -; %1 = row or col (for rounding variable) -; %2 = number of bits to shift at the end -%macro IDCT_1D 2 - ; a0 = (W4 * row[0]) + (1 << (15 - 1)); - ; a1 = a0; - ; a2 = a0; - ; a3 = a0; - ; a0 += W2 * row[2]; - ; a1 += W6 * row[2]; - ; a2 -= W6 * row[2]; - ; a3 -= W2 * row[2]; -%ifidn %1, col - paddw m10,[pw_88] -%endif -%ifidn %1, row - paddw m10,[pw_1] -%endif - SBUTTERFLY3 wd, 0, 1, 10, 8 ; { row[0], row[2] }[0-3]/[4-7] - pmaddwd m2, m0, [w4_plus_w6] - pmaddwd m3, m1, [w4_plus_w6] - pmaddwd m4, m0, [w4_min_w6] - pmaddwd m5, m1, [w4_min_w6] - pmaddwd m6, m0, [w4_min_w2] - pmaddwd m7, m1, [w4_min_w2] - pmaddwd m0, [w4_plus_w2] - pmaddwd m1, [w4_plus_w2] - - ; a0: -1*row[0]-1*row[2] - ; a1: -1*row[0] - ; a2: -1*row[0] - ; a3: -1*row[0]+1*row[2] - - ; a0 += W4*row[4] + W6*row[6]; i.e. -1*row[4] - ; a1 -= W4*row[4] + W2*row[6]; i.e. -1*row[4]-1*row[6] - ; a2 -= W4*row[4] - W2*row[6]; i.e. -1*row[4]+1*row[6] - ; a3 += W4*row[4] - W6*row[6]; i.e. -1*row[4] - SBUTTERFLY3 wd, 8, 9, 13, 12 ; { row[4], row[6] }[0-3]/[4-7] - pmaddwd m10, m8, [w4_plus_w6] - pmaddwd m11, m9, [w4_plus_w6] - paddd m0, m10 ; a0[0-3] - paddd m1, m11 ; a0[4-7] - pmaddwd m10, m8, [w4_min_w6] - pmaddwd m11, m9, [w4_min_w6] - paddd m6, m10 ; a3[0-3] - paddd m7, m11 ; a3[4-7] - pmaddwd m10, m8, [w4_min_w2] - pmaddwd m11, m9, [w4_min_w2] - pmaddwd m8, [w4_plus_w2] - pmaddwd m9, [w4_plus_w2] - psubd m4, m10 ; a2[0-3] intermediate - psubd m5, m11 ; a2[4-7] intermediate - psubd m2, m8 ; a1[0-3] intermediate - psubd m3, m9 ; a1[4-7] intermediate - - ; load/store - mova [r2+ 0], m0 - mova [r2+ 32], m2 - mova [r2+ 64], m4 - mova [r2+ 96], m6 - mova m10,[r2+ 16] ; { row[1] }[0-7] - mova m8, [r2+ 48] ; { row[3] }[0-7] - mova m13,[r2+ 80] ; { row[5] }[0-7] - mova m14,[r2+112] ; { row[7] }[0-7] - mova [r2+ 16], m1 - mova [r2+ 48], m3 - mova [r2+ 80], m5 - mova [r2+112], m7 -%ifidn %1, row - pmullw m10,[r3+ 16] - pmullw m8, [r3+ 48] - pmullw m13,[r3+ 80] - pmullw m14,[r3+112] -%endif - - ; b0 = MUL(W1, row[1]); - ; MAC(b0, W3, row[3]); - ; b1 = MUL(W3, row[1]); - ; MAC(b1, -W7, row[3]); - ; b2 = MUL(W5, row[1]); - ; MAC(b2, -W1, row[3]); - ; b3 = MUL(W7, row[1]); - ; MAC(b3, -W5, row[3]); - SBUTTERFLY3 wd, 0, 1, 10, 8 ; { row[1], row[3] }[0-3]/[4-7] - pmaddwd m2, m0, [w3_min_w7] - pmaddwd m3, m1, [w3_min_w7] - pmaddwd m4, m0, [w5_min_w1] - pmaddwd m5, m1, [w5_min_w1] - pmaddwd m6, m0, [w7_min_w5] - pmaddwd m7, m1, [w7_min_w5] - pmaddwd m0, [w1_plus_w3] - pmaddwd m1, [w1_plus_w3] - - ; b0: +1*row[1]+2*row[3] - ; b1: +2*row[1]-1*row[3] - ; b2: -1*row[1]-1*row[3] - ; b3: +1*row[1]+1*row[3] - - ; MAC(b0, W5, row[5]); - ; MAC(b0, W7, row[7]); - ; MAC(b1, -W1, row[5]); - ; MAC(b1, -W5, row[7]); - ; MAC(b2, W7, row[5]); - ; MAC(b2, W3, row[7]); - ; MAC(b3, W3, row[5]); - ; MAC(b3, -W1, row[7]); - SBUTTERFLY3 wd, 8, 9, 13, 14 ; { row[5], row[7] }[0-3]/[4-7] - - ; b0: -1*row[5]+1*row[7] - ; b1: -1*row[5]+1*row[7] - ; b2: +1*row[5]+2*row[7] - ; b3: +2*row[5]-1*row[7] +%include "libavcodec/x86/simple_idct10_template.asm" - pmaddwd m10, m8, [w1_plus_w5] - pmaddwd m11, m9, [w1_plus_w5] - pmaddwd m12, m8, [w5_plus_w7] - pmaddwd m13, m9, [w5_plus_w7] - psubd m2, m10 ; b1[0-3] - psubd m3, m11 ; b1[4-7] - paddd m0, m12 ; b0[0-3] - paddd m1, m13 ; b0[4-7] - pmaddwd m12, m8, [w7_plus_w3] - pmaddwd m13, m9, [w7_plus_w3] - pmaddwd m8, [w3_min_w1] - pmaddwd m9, [w3_min_w1] - paddd m4, m12 ; b2[0-3] - paddd m5, m13 ; b2[4-7] - paddd m6, m8 ; b3[0-3] - paddd m7, m9 ; b3[4-7] - - ; row[0] = (a0 + b0) >> 15; - ; row[7] = (a0 - b0) >> 15; - ; row[1] = (a1 + b1) >> 15; - ; row[6] = (a1 - b1) >> 15; - ; row[2] = (a2 + b2) >> 15; - ; row[5] = (a2 - b2) >> 15; - ; row[3] = (a3 + b3) >> 15; - ; row[4] = (a3 - b3) >> 15; - mova m8, [r2+ 0] ; a0[0-3] - mova m9, [r2+16] ; a0[4-7] - SUMSUB_SHPK m8, m9, m10, m11, m0, m1, %2 - mova m0, [r2+32] ; a1[0-3] - mova m1, [r2+48] ; a1[4-7] - SUMSUB_SHPK m0, m1, m9, m11, m2, m3, %2 - mova m1, [r2+64] ; a2[0-3] - mova m2, [r2+80] ; a2[4-7] - SUMSUB_SHPK m1, m2, m11, m3, m4, m5, %2 - mova m2, [r2+96] ; a3[0-3] - mova m3, [r2+112] ; a3[4-7] - SUMSUB_SHPK m2, m3, m4, m5, m6, m7, %2 -%endmacro +section .text align=16 -; void ff_prores_idct_put_10_<opt>(uint8_t *pixels, int stride, -; int16_t *block, const int16_t *qmat); %macro idct_put_fn 1 cglobal prores_idct_put_10, 4, 4, %1 - movsxd r1, r1d - pxor m15, m15 ; zero - - ; for (i = 0; i < 8; i++) - ; idctRowCondDC(block + i*8); - mova m10,[r2+ 0] ; { row[0] }[0-7] - mova m8, [r2+32] ; { row[2] }[0-7] - mova m13,[r2+64] ; { row[4] }[0-7] - mova m12,[r2+96] ; { row[6] }[0-7] - - pmullw m10,[r3+ 0] - pmullw m8, [r3+32] - pmullw m13,[r3+64] - pmullw m12,[r3+96] - - IDCT_1D row, 15 - - ; transpose for second part of IDCT - TRANSPOSE8x8W 8, 0, 1, 2, 4, 11, 9, 10, 3 - mova [r2+ 16], m0 - mova [r2+ 48], m2 - mova [r2+ 80], m11 - mova [r2+112], m10 - SWAP 8, 10 - SWAP 1, 8 - SWAP 4, 13 - SWAP 9, 12 - - ; for (i = 0; i < 8; i++) - ; idctSparseColAdd(dest + i, line_size, block + i); - IDCT_1D col, 18 - - ; clip/store - mova m3, [pw_4] - mova m5, [pw_1019] - pmaxsw m8, m3 - pmaxsw m0, m3 - pmaxsw m1, m3 - pmaxsw m2, m3 - pmaxsw m4, m3 - pmaxsw m11, m3 - pmaxsw m9, m3 - pmaxsw m10, m3 - pminsw m8, m5 - pminsw m0, m5 - pminsw m1, m5 - pminsw m2, m5 - pminsw m4, m5 - pminsw m11, m5 - pminsw m9, m5 - pminsw m10, m5 - - lea r2, [r1*3] - mova [r0 ], m8 - mova [r0+r1 ], m0 - mova [r0+r1*2], m1 - mova [r0+r2 ], m2 - lea r0, [r0+r1*4] - mova [r0 ], m4 - mova [r0+r1 ], m11 - mova [r0+r1*2], m9 - mova [r0+r2 ], m10 + IDCT_PUT_FN pw_1, 15, pw_88, 18, pw_4, pw_1019, r3 RET %endmacro diff --git a/libavcodec/x86/simple_idct10_template.asm b/libavcodec/x86/simple_idct10_template.asm new file mode 100644 index 0000000000..968d280ba3 --- /dev/null +++ b/libavcodec/x86/simple_idct10_template.asm @@ -0,0 +1,299 @@ +;****************************************************************************** +;* x86-SIMD-optimized IDCT for prores +;* this is identical to "simple" IDCT written by Michael Niedermayer +;* except for the clip range +;* +;* Copyright (c) 2011 Ronald S. Bultje <rsbultje@gmail.com> +;* +;* This file is part of FFmpeg. +;* +;* FFmpeg is free software; you can redistribute it and/or +;* modify it under the terms of the GNU Lesser General Public +;* License as published by the Free Software Foundation; either +;* version 2.1 of the License, or (at your option) any later version. +;* +;* FFmpeg is distributed in the hope that it will be useful, +;* but WITHOUT ANY WARRANTY; without even the implied warranty of +;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU +;* Lesser General Public License for more details. +;* +;* You should have received a copy of the GNU Lesser General Public +;* License along with FFmpeg; if not, write to the Free Software +;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA +;****************************************************************************** + +; add SECTION_RODATA and proper include before including this file! + +%if ARCH_X86_64 + +cextern w4_plus_w2 +cextern w4_min_w2 +cextern w4_plus_w6 +cextern w4_min_w6 +cextern w1_plus_w3 +cextern w3_min_w1 +cextern w7_plus_w3 +cextern w3_min_w7 +cextern w1_plus_w5 +cextern w5_min_w1 +cextern w5_plus_w7 +cextern w7_min_w5 + +; interleave data while maintaining source +; %1=type, %2=dstlo, %3=dsthi, %4=src, %5=interleave +%macro SBUTTERFLY3 5 + punpckl%1 m%2, m%4, m%5 + punpckh%1 m%3, m%4, m%5 +%endmacro + +; %1/%2=src1/dst1, %3/%4=dst2, %5/%6=src2, %7=shift +; action: %3/%4 = %1/%2 - %5/%6; %1/%2 += %5/%6 +; %1/%2/%3/%4 >>= %7; dword -> word (in %1/%3) +%macro SUMSUB_SHPK 7 + psubd %3, %1, %5 ; { a0 - b0 }[0-3] + psubd %4, %2, %6 ; { a0 - b0 }[4-7] + paddd %1, %5 ; { a0 + b0 }[0-3] + paddd %2, %6 ; { a0 + b0 }[4-7] + psrad %1, %7 + psrad %2, %7 + psrad %3, %7 + psrad %4, %7 + packssdw %1, %2 ; row[0] + packssdw %3, %4 ; row[7] +%endmacro + +; %1 = initial bias ("" if nop) +; %2 = number of bits to shift at the end +; %3 = qmat (for prores) +%macro IDCT_1D 2-3 + ; a0 = (W4 * row[0]) + (1 << (15 - 1)); + ; a1 = a0; + ; a2 = a0; + ; a3 = a0; + ; a0 += W2 * row[2]; + ; a1 += W6 * row[2]; + ; a2 -= W6 * row[2]; + ; a3 -= W2 * row[2]; +%ifstr %1 +%else + paddw m10, [%1] +%endif + SBUTTERFLY3 wd, 0, 1, 10, 8 ; { row[0], row[2] }[0-3]/[4-7] + pmaddwd m2, m0, [w4_plus_w6] + pmaddwd m3, m1, [w4_plus_w6] + pmaddwd m4, m0, [w4_min_w6] + pmaddwd m5, m1, [w4_min_w6] + pmaddwd m6, m0, [w4_min_w2] + pmaddwd m7, m1, [w4_min_w2] + pmaddwd m0, [w4_plus_w2] + pmaddwd m1, [w4_plus_w2] + + ; a0: -1*row[0]-1*row[2] + ; a1: -1*row[0] + ; a2: -1*row[0] + ; a3: -1*row[0]+1*row[2] + + ; a0 += W4*row[4] + W6*row[6]; i.e. -1*row[4] + ; a1 -= W4*row[4] + W2*row[6]; i.e. -1*row[4]-1*row[6] + ; a2 -= W4*row[4] - W2*row[6]; i.e. -1*row[4]+1*row[6] + ; a3 += W4*row[4] - W6*row[6]; i.e. -1*row[4] + SBUTTERFLY3 wd, 8, 9, 13, 12 ; { row[4], row[6] }[0-3]/[4-7] + pmaddwd m10, m8, [w4_plus_w6] + pmaddwd m11, m9, [w4_plus_w6] + paddd m0, m10 ; a0[0-3] + paddd m1, m11 ; a0[4-7] + pmaddwd m10, m8, [w4_min_w6] + pmaddwd m11, m9, [w4_min_w6] + paddd m6, m10 ; a3[0-3] + paddd m7, m11 ; a3[4-7] + pmaddwd m10, m8, [w4_min_w2] + pmaddwd m11, m9, [w4_min_w2] + pmaddwd m8, [w4_plus_w2] + pmaddwd m9, [w4_plus_w2] + psubd m4, m10 ; a2[0-3] intermediate + psubd m5, m11 ; a2[4-7] intermediate + psubd m2, m8 ; a1[0-3] intermediate + psubd m3, m9 ; a1[4-7] intermediate + + ; load/store + mova [r2+ 0], m0 + mova [r2+ 32], m2 + mova [r2+ 64], m4 + mova [r2+ 96], m6 + mova m10,[r2+ 16] ; { row[1] }[0-7] + mova m8, [r2+ 48] ; { row[3] }[0-7] + mova m13,[r2+ 80] ; { row[5] }[0-7] + mova m14,[r2+112] ; { row[7] }[0-7] + mova [r2+ 16], m1 + mova [r2+ 48], m3 + mova [r2+ 80], m5 + mova [r2+112], m7 +%if %0 == 3 + pmullw m10,[%3+ 16] + pmullw m8, [%3+ 48] + pmullw m13,[%3+ 80] + pmullw m14,[%3+112] +%endif + + ; b0 = MUL(W1, row[1]); + ; MAC(b0, W3, row[3]); + ; b1 = MUL(W3, row[1]); + ; MAC(b1, -W7, row[3]); + ; b2 = MUL(W5, row[1]); + ; MAC(b2, -W1, row[3]); + ; b3 = MUL(W7, row[1]); + ; MAC(b3, -W5, row[3]); + SBUTTERFLY3 wd, 0, 1, 10, 8 ; { row[1], row[3] }[0-3]/[4-7] + pmaddwd m2, m0, [w3_min_w7] + pmaddwd m3, m1, [w3_min_w7] + pmaddwd m4, m0, [w5_min_w1] + pmaddwd m5, m1, [w5_min_w1] + pmaddwd m6, m0, [w7_min_w5] + pmaddwd m7, m1, [w7_min_w5] + pmaddwd m0, [w1_plus_w3] + pmaddwd m1, [w1_plus_w3] + + ; b0: +1*row[1]+2*row[3] + ; b1: +2*row[1]-1*row[3] + ; b2: -1*row[1]-1*row[3] + ; b3: +1*row[1]+1*row[3] + + ; MAC(b0, W5, row[5]); + ; MAC(b0, W7, row[7]); + ; MAC(b1, -W1, row[5]); + ; MAC(b1, -W5, row[7]); + ; MAC(b2, W7, row[5]); + ; MAC(b2, W3, row[7]); + ; MAC(b3, W3, row[5]); + ; MAC(b3, -W1, row[7]); + SBUTTERFLY3 wd, 8, 9, 13, 14 ; { row[5], row[7] }[0-3]/[4-7] + + ; b0: -1*row[5]+1*row[7] + ; b1: -1*row[5]+1*row[7] + ; b2: +1*row[5]+2*row[7] + ; b3: +2*row[5]-1*row[7] + + pmaddwd m10, m8, [w1_plus_w5] + pmaddwd m11, m9, [w1_plus_w5] + pmaddwd m12, m8, [w5_plus_w7] + pmaddwd m13, m9, [w5_plus_w7] + psubd m2, m10 ; b1[0-3] + psubd m3, m11 ; b1[4-7] + paddd m0, m12 ; b0[0-3] + paddd m1, m13 ; b0[4-7] + pmaddwd m12, m8, [w7_plus_w3] + pmaddwd m13, m9, [w7_plus_w3] + pmaddwd m8, [w3_min_w1] + pmaddwd m9, [w3_min_w1] + paddd m4, m12 ; b2[0-3] + paddd m5, m13 ; b2[4-7] + paddd m6, m8 ; b3[0-3] + paddd m7, m9 ; b3[4-7] + + ; row[0] = (a0 + b0) >> 15; + ; row[7] = (a0 - b0) >> 15; + ; row[1] = (a1 + b1) >> 15; + ; row[6] = (a1 - b1) >> 15; + ; row[2] = (a2 + b2) >> 15; + ; row[5] = (a2 - b2) >> 15; + ; row[3] = (a3 + b3) >> 15; + ; row[4] = (a3 - b3) >> 15; + mova m8, [r2+ 0] ; a0[0-3] + mova m9, [r2+16] ; a0[4-7] + SUMSUB_SHPK m8, m9, m10, m11, m0, m1, %2 + mova m0, [r2+32] ; a1[0-3] + mova m1, [r2+48] ; a1[4-7] + SUMSUB_SHPK m0, m1, m9, m11, m2, m3, %2 + mova m1, [r2+64] ; a2[0-3] + mova m2, [r2+80] ; a2[4-7] + SUMSUB_SHPK m1, m2, m11, m3, m4, m5, %2 + mova m2, [r2+96] ; a3[0-3] + mova m3, [r2+112] ; a3[4-7] + SUMSUB_SHPK m2, m3, m4, m5, m6, m7, %2 +%endmacro + +; void ff_prores_idct_put_10_<opt>(uint8_t *pixels, int stride, +; int16_t *block, const int16_t *qmat); + +; %1 = row shift +; %2 = row bias macro +; %3 = column shift +; %4 = column bias macro +; %5 = min pixel value +; %6 = max pixel value +; %7 = qmat (for prores) + +%macro IDCT_PUT_FN 6-7 + movsxd r1, r1d + pxor m15, m15 ; zero + + ; for (i = 0; i < 8; i++) + ; idctRowCondDC(block + i*8); + mova m10,[r2+ 0] ; { row[0] }[0-7] + mova m8, [r2+32] ; { row[2] }[0-7] + mova m13,[r2+64] ; { row[4] }[0-7] + mova m12,[r2+96] ; { row[6] }[0-7] + +%if %0 == 7 + pmullw m10,[%7+ 0] + pmullw m8, [%7+32] + pmullw m13,[%7+64] + pmullw m12,[%7+96] + + IDCT_1D %1, %2, %7 +%else + IDCT_1D %1, %2 +%endif + + ; transpose for second part of IDCT + TRANSPOSE8x8W 8, 0, 1, 2, 4, 11, 9, 10, 3 + mova [r2+ 16], m0 + mova [r2+ 48], m2 + mova [r2+ 80], m11 + mova [r2+112], m10 + SWAP 8, 10 + SWAP 1, 8 + SWAP 4, 13 + SWAP 9, 12 + + ; for (i = 0; i < 8; i++) + ; idctSparseColAdd(dest + i, line_size, block + i); + IDCT_1D %3, %4 + + ; clip/store +%ifidn %5, 0 + pxor m3, m3 +%else + mova m3, [%5] +%endif + mova m5, [%6] + pmaxsw m8, m3 + pmaxsw m0, m3 + pmaxsw m1, m3 + pmaxsw m2, m3 + pmaxsw m4, m3 + pmaxsw m11, m3 + pmaxsw m9, m3 + pmaxsw m10, m3 + pminsw m8, m5 + pminsw m0, m5 + pminsw m1, m5 + pminsw m2, m5 + pminsw m4, m5 + pminsw m11, m5 + pminsw m9, m5 + pminsw m10, m5 + + lea r2, [r1*3] + mova [r0 ], m8 + mova [r0+r1 ], m0 + mova [r0+r1*2], m1 + mova [r0+r2 ], m2 + lea r0, [r0+r1*4] + mova [r0 ], m4 + mova [r0+r1 ], m11 + mova [r0+r1*2], m9 + mova [r0+r2 ], m10 +%endmacro + +%endif |