aboutsummaryrefslogtreecommitdiffstats
path: root/libavcodec/x86
diff options
context:
space:
mode:
authorChristophe Gisquet <christophe.gisquet@gmail.com>2014-02-14 15:03:09 +0000
committerJanne Grunau <janne-libav@jannau.net>2014-02-28 13:00:48 +0100
commit08e3ea60ff4059341b74be04a428a38f7c3630b0 (patch)
treee9e639f3b99004cdf785a88b17e06b1c28fb1b66 /libavcodec/x86
parent57b1eb9f75b04571063ddec316e290c216c114ac (diff)
downloadffmpeg-08e3ea60ff4059341b74be04a428a38f7c3630b0.tar.gz
x86: synth filter float: implement SSE2 version
Timings for Arrandale: C SSE win32: 2108 334 win64: 1152 322 Factorizing the inner loop with a call/jmp is a >15 cycles cost, even with the jmp destination being aligned. Unrolling for ARCH_X86_64 is a 20 cycles gain. Signed-off-by: Janne Grunau <janne-libav@jannau.net>
Diffstat (limited to 'libavcodec/x86')
-rw-r--r--libavcodec/x86/dcadsp.asm152
-rw-r--r--libavcodec/x86/dcadsp_init.c28
2 files changed, 180 insertions, 0 deletions
diff --git a/libavcodec/x86/dcadsp.asm b/libavcodec/x86/dcadsp.asm
index de827d7676..c1db118676 100644
--- a/libavcodec/x86/dcadsp.asm
+++ b/libavcodec/x86/dcadsp.asm
@@ -175,3 +175,155 @@ cglobal dca_lfe_fir%1, 3,3,6-%1, out, in, cf0
INIT_XMM sse
DCA_LFE_FIR 0
DCA_LFE_FIR 1
+
+INIT_XMM sse2
+%macro INNER_LOOP 1
+ ; reading backwards: ptr1 = synth_buf + j + i; ptr2 = synth_buf + j - i
+ ;~ a += window[i + j] * (-synth_buf[15 - i + j])
+ ;~ b += window[i + j + 16] * (synth_buf[i + j])
+ pshufd m5, [ptr2 + j + (15 - 3) * 4], q0123
+ mova m6, [ptr1 + j]
+%if ARCH_X86_64
+ pshufd m11, [ptr2 + j + (15 - 3) * 4 - mmsize], q0123
+ mova m12, [ptr1 + j + mmsize]
+%endif
+ mulps m6, [win + %1 + j + 16 * 4]
+ mulps m5, [win + %1 + j]
+%if ARCH_X86_64
+ mulps m12, [win + %1 + j + mmsize + 16 * 4]
+ mulps m11, [win + %1 + j + mmsize]
+%endif
+ addps m2, m6
+ subps m1, m5
+%if ARCH_X86_64
+ addps m8, m12
+ subps m7, m11
+%endif
+ ;~ c += window[i + j + 32] * (synth_buf[16 + i + j])
+ ;~ d += window[i + j + 48] * (synth_buf[31 - i + j])
+ pshufd m6, [ptr2 + j + (31 - 3) * 4], q0123
+ mova m5, [ptr1 + j + 16 * 4]
+%if ARCH_X86_64
+ pshufd m12, [ptr2 + j + (31 - 3) * 4 - mmsize], q0123
+ mova m11, [ptr1 + j + mmsize + 16 * 4]
+%endif
+ mulps m5, [win + %1 + j + 32 * 4]
+ mulps m6, [win + %1 + j + 48 * 4]
+%if ARCH_X86_64
+ mulps m11, [win + %1 + j + mmsize + 32 * 4]
+ mulps m12, [win + %1 + j + mmsize + 48 * 4]
+%endif
+ addps m3, m5
+ addps m4, m6
+%if ARCH_X86_64
+ addps m9, m11
+ addps m10, m12
+%endif
+ sub j, 64 * 4
+%endmacro
+
+; void ff_synth_filter_inner_sse2(float *synth_buf, float synth_buf2[32],
+; const float window[512], float out[32],
+; intptr_t offset, float scale)
+cglobal synth_filter_inner, 0, 6 + 4 * ARCH_X86_64, 7 + 6 * ARCH_X86_64, \
+ synth_buf, synth_buf2, window, out, off, scale
+%define scale m0
+%if ARCH_X86_32 || WIN64
+ movd scale, scalem
+; Make sure offset is in a register and not on the stack
+%define OFFQ r4q
+%else
+%define OFFQ offq
+%endif
+ pshufd m0, m0, 0
+ ; prepare inner counter limit 1
+ mov r5q, 480
+ sub r5q, offmp
+ and r5q, -64
+ shl r5q, 2
+ mov OFFQ, r5q
+%define i r5q
+ mov i, 16 * 4 - (ARCH_X86_64 + 1) * mmsize ; main loop counter
+
+%define buf2 synth_buf2q
+%if ARCH_X86_32
+ mov buf2, synth_buf2mp
+%endif
+.mainloop
+ ; m1 = a m2 = b m3 = c m4 = d
+ pxor m3, m3
+ pxor m4, m4
+ mova m1, [buf2 + i]
+ mova m2, [buf2 + i + 16 * 4]
+%if ARCH_X86_32
+%define ptr1 r0q
+%define ptr2 r1q
+%define win r2q
+%define j r3q
+ mov win, windowm
+ mov ptr1, synth_bufm
+ add win, i
+ add ptr1, i
+%else ; ARCH_X86_64
+%define ptr1 r6q
+%define ptr2 r7q ; must be loaded
+%define win r8q
+%define j r9q
+ pxor m9, m9
+ pxor m10, m10
+ mova m7, [buf2 + i + mmsize]
+ mova m8, [buf2 + i + mmsize + 16 * 4]
+ lea win, [windowq + i]
+ lea ptr1, [synth_bufq + i]
+%endif
+ mov ptr2, synth_bufmp
+ ; prepare the inner loop counter
+ mov j, OFFQ
+ sub ptr2, i
+.loop1:
+ INNER_LOOP 0
+ jge .loop1
+
+ mov j, 448 * 4
+ sub j, OFFQ
+ jz .end
+ sub ptr1, j
+ sub ptr2, j
+ add win, OFFQ ; now at j-64, so define OFFSET
+ sub j, 64 * 4
+.loop2:
+ INNER_LOOP 64 * 4
+ jge .loop2
+
+.end:
+%if ARCH_X86_32
+ mov buf2, synth_buf2m ; needed for next iteration anyway
+ mov outq, outmp ; j, which will be set again during it
+%endif
+ ;~ out[i] = a * scale;
+ ;~ out[i + 16] = b * scale;
+ mulps m1, scale
+ mulps m2, scale
+%if ARCH_X86_64
+ mulps m7, scale
+ mulps m8, scale
+%endif
+ ;~ synth_buf2[i] = c;
+ ;~ synth_buf2[i + 16] = d;
+ mova [buf2 + i + 0 * 4], m3
+ mova [buf2 + i + 16 * 4], m4
+%if ARCH_X86_64
+ mova [buf2 + i + 0 * 4 + mmsize], m9
+ mova [buf2 + i + 16 * 4 + mmsize], m10
+%endif
+ ;~ out[i] = a;
+ ;~ out[i + 16] = a;
+ mova [outq + i + 0 * 4], m1
+ mova [outq + i + 16 * 4], m2
+%if ARCH_X86_64
+ mova [outq + i + 0 * 4 + mmsize], m7
+ mova [outq + i + 16 * 4 + mmsize], m8
+%endif
+ sub i, (ARCH_X86_64 + 1) * mmsize
+ jge .mainloop
+ RET
diff --git a/libavcodec/x86/dcadsp_init.c b/libavcodec/x86/dcadsp_init.c
index c234bd29f5..5f6e8c5a19 100644
--- a/libavcodec/x86/dcadsp_init.c
+++ b/libavcodec/x86/dcadsp_init.c
@@ -49,3 +49,31 @@ av_cold void ff_dcadsp_init_x86(DCADSPContext *s)
s->int8x8_fmul_int32 = ff_int8x8_fmul_int32_sse4;
}
}
+
+void ff_synth_filter_inner_sse2(float *synth_buf_ptr, float synth_buf2[32],
+ const float window[512],
+ float out[32], intptr_t offset, float scale);
+
+static void synth_filter_sse2(FFTContext *imdct,
+ float *synth_buf_ptr, int *synth_buf_offset,
+ float synth_buf2[32], const float window[512],
+ float out[32], const float in[32], float scale)
+{
+ float *synth_buf= synth_buf_ptr + *synth_buf_offset;
+
+ imdct->imdct_half(imdct, synth_buf, in);
+
+ ff_synth_filter_inner_sse2(synth_buf, synth_buf2, window,
+ out, *synth_buf_offset, scale);
+
+ *synth_buf_offset = (*synth_buf_offset - 32) & 511;
+}
+
+av_cold void ff_synth_filter_init_x86(SynthFilterContext *s)
+{
+ int cpu_flags = av_get_cpu_flags();
+
+ if (EXTERNAL_SSE2(cpu_flags)) {
+ s->synth_filter_float = synth_filter_sse2;
+ }
+}