diff options
author | plepere <pierre-edouard.lepere@insa-rennes.fr> | 2014-04-28 16:12:28 +0200 |
---|---|---|
committer | Michael Niedermayer <michaelni@gmx.at> | 2014-05-06 18:23:36 +0200 |
commit | 7a2491c436af41de103ca82294d79eb2031c6607 (patch) | |
tree | a23b3e9cea28f96463378ea45f61f7abb7d5e0d5 /libavcodec/x86 | |
parent | 83976e40e89655162e5394cf8915d9b6d89702d9 (diff) | |
download | ffmpeg-7a2491c436af41de103ca82294d79eb2031c6607.tar.gz |
HEVC : added assembly MC functions
pretty print x86
Reviewed-by: "Ronald S. Bultje" <rsbultje@gmail.com>
Signed-off-by: Michael Niedermayer <michaelni@gmx.at>
Diffstat (limited to 'libavcodec/x86')
-rw-r--r-- | libavcodec/x86/Makefile | 3 | ||||
-rw-r--r-- | libavcodec/x86/hevc_mc.asm | 1256 | ||||
-rw-r--r-- | libavcodec/x86/hevcdsp.h | 120 | ||||
-rw-r--r-- | libavcodec/x86/hevcdsp_init.c | 385 |
4 files changed, 1764 insertions, 0 deletions
diff --git a/libavcodec/x86/Makefile b/libavcodec/x86/Makefile index 2fd562c1e0..40f884060d 100644 --- a/libavcodec/x86/Makefile +++ b/libavcodec/x86/Makefile @@ -15,7 +15,9 @@ OBJS-$(CONFIG_H263DSP) += x86/h263dsp_init.o OBJS-$(CONFIG_H264CHROMA) += x86/h264chroma_init.o OBJS-$(CONFIG_H264DSP) += x86/h264dsp_init.o OBJS-$(CONFIG_H264PRED) += x86/h264_intrapred_init.o +OBJS-$(CONFIG_HEVC_DECODER) += x86/hevcdsp_init.o OBJS-$(CONFIG_H264QPEL) += x86/h264_qpel.o +OBJS-$(CONFIG_HEVC_DECODER) += x86/hevcdsp_init.o OBJS-$(CONFIG_HPELDSP) += x86/hpeldsp_init.o OBJS-$(CONFIG_LLVIDDSP) += x86/lossless_videodsp_init.o OBJS-$(CONFIG_LPC) += x86/lpc.o @@ -91,6 +93,7 @@ YASM-OBJS-$(CONFIG_H264QPEL) += x86/h264_qpel_8bit.o \ x86/h264_qpel_10bit.o \ x86/fpel.o \ x86/qpel.o +YASM-OBJS-$(CONFIG_HEVC_DECODER) += x86/hevc_mc.o YASM-OBJS-$(CONFIG_HPELDSP) += x86/fpel.o \ x86/hpeldsp.o YASM-OBJS-$(CONFIG_LLVIDDSP) += x86/lossless_videodsp.o diff --git a/libavcodec/x86/hevc_mc.asm b/libavcodec/x86/hevc_mc.asm new file mode 100644 index 0000000000..e83613b3ab --- /dev/null +++ b/libavcodec/x86/hevc_mc.asm @@ -0,0 +1,1256 @@ +; /* +; * Provide SSE luma and chroma mc functions for HEVC decoding +; * Copyright (c) 2013 Pierre-Edouard LEPERE +; * +; * This file is part of FFmpeg. +; * +; * FFmpeg is free software; you can redistribute it and/or +; * modify it under the terms of the GNU Lesser General Public +; * License as published by the Free Software Foundation; either +; * version 2.1 of the License, or (at your option) any later version. +; * +; * FFmpeg is distributed in the hope that it will be useful, +; * but WITHOUT ANY WARRANTY; without even the implied warranty of +; * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU +; * Lesser General Public License for more details. +; * +; * You should have received a copy of the GNU Lesser General Public +; * License along with FFmpeg; if not, write to the Free Software +; * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA +; */ +%include "libavutil/x86/x86util.asm" + +SECTION_RODATA +pw_8: times 8 dw 512 +pw_10: times 8 dw 2048 +pw_bi_8: times 8 dw 256 +pw_bi_10: times 8 dw 1024 +max_pixels_10: times 8 dw 1023 +zero: times 4 dd 0 +one_per_32: times 4 dd 1 + +SECTION .text +%macro EPEL_TABLE 4 +hevc_epel_filters_%4_%1 times %2 d%3 -2, 58 + times %2 d%3 10, -2 + times %2 d%3 -4, 54 + times %2 d%3 16, -2 + times %2 d%3 -6, 46 + times %2 d%3 28, -4 + times %2 d%3 -4, 36 + times %2 d%3 36, -4 + times %2 d%3 -4, 28 + times %2 d%3 46, -6 + times %2 d%3 -2, 16 + times %2 d%3 54, -4 + times %2 d%3 -2, 10 + times %2 d%3 58, -2 +%endmacro + + + +EPEL_TABLE 8, 8, b, sse4 +EPEL_TABLE 10, 4, w, sse4 + +%macro QPEL_TABLE 4 +hevc_qpel_filters_%4_%1 times %2 d%3 -1, 4 + times %2 d%3 -10, 58 + times %2 d%3 17, -5 + times %2 d%3 1, 0 + times %2 d%3 -1, 4 + times %2 d%3 -11, 40 + times %2 d%3 40,-11 + times %2 d%3 4, -1 + times %2 d%3 0, 1 + times %2 d%3 -5, 17 + times %2 d%3 58,-10 + times %2 d%3 4, -1 +%endmacro + +QPEL_TABLE 8, 8, b, sse4 +QPEL_TABLE 10, 4, w, sse4 + +%define hevc_qpel_filters_sse4_14 hevc_qpel_filters_sse4_10 + +%if ARCH_X86_64 + +%macro SIMPLE_BILOAD 4 ;width, tab, r1, r2 +%if %1 <= 4 + movq %3, [%2] ; load data from source2 +%elif %1 <= 8 + movdqa %3, [%2] ; load data from source2 +%elif %1 <= 12 + movdqa %3, [%2] ; load data from source2 + movq %4, [%2+16] ; load data from source2 +%else + movdqa %3, [%2] ; load data from source2 + movdqa %4, [%2+16] ; load data from source2 +%endif +%endmacro + +%macro SIMPLE_LOAD 4 ;width, bitd, tab, r1 +%if %1 == 2 || (%2 == 8 && %1 <= 4) + movd %4, [%3] ; load data from source +%elif %1 == 4 || (%2 == 8 && %1 <= 8) + movq %4, [%3] ; load data from source +%else + movdqu %4, [%3] ; load data from source +%endif +%endmacro + +%macro SIMPLE_8LOAD 5 ;width, bitd, tab, r1, r2 +%if %1 == 2 || (%2 == 8 && %1 <= 4) + movq %4, [%3] ; load data from source2 +%elif %1 == 4 || (%2 == 8 && %1 <= 8) + movdqa %4, [%3] ; load data from source2 +%elif %1 <= 12 + movdqa %4, [%3] ; load data from source2 + movq %5, [%3+16] ; load data from source2 +%else + movdqa %4, [%3] ; load data from source2 + movdqa %5, [%3+16] ; load data from source2 +%endif +%endmacro + +%macro EPEL_FILTER 2 ; bit depth, filter index +%ifdef PIC + lea rfilterq, [hevc_epel_filters_sse4_%1] +%else + %define rfilterq hevc_epel_filters_sse4_%1 +%endif + sub %2q, 1 + shl %2q, 5 ; multiply by 32 + movdqa m14, [rfilterq + %2q] ; get 2 first values of filters + movdqa m15, [rfilterq + %2q+16] ; get 2 last values of filters +%endmacro + +%macro EPEL_HV_FILTER 1 +%ifdef PIC + lea rfilterq, [hevc_epel_filters_sse4_%1] +%else + %define rfilterq hevc_epel_filters_sse4_%1 +%endif + sub mxq, 1 + sub myq, 1 + shl mxq, 5 ; multiply by 32 + shl myq, 5 ; multiply by 32 + movdqa m14, [rfilterq + mxq] ; get 2 first values of filters + movdqa m15, [rfilterq + mxq+16] ; get 2 last values of filters + lea r3srcq, [srcstrideq*3] + +%ifdef PIC + lea rfilterq, [hevc_epel_filters_sse4_10] +%else + %define rfilterq hevc_epel_filters_sse4_10 +%endif + movdqa m12, [rfilterq + myq] ; get 2 first values of filters + movdqa m13, [rfilterq + myq+16] ; get 2 last values of filters +%endmacro + +%macro QPEL_FILTER 2 +%ifdef PIC + lea rfilterq, [hevc_qpel_filters_sse4_%1] +%else + %define rfilterq hevc_qpel_filters_sse4_%1 +%endif + lea %2q, [%2q*8-8] + movdqa m12, [rfilterq + %2q*8] ; get 4 first values of filters + movdqa m13, [rfilterq + %2q*8 + 16] ; get 4 first values of filters + movdqa m14, [rfilterq + %2q*8 + 32] ; get 4 first values of filters + movdqa m15, [rfilterq + %2q*8 + 48] ; get 4 first values of filters +%endmacro + +%macro EPEL_LOAD 4 +%ifdef PIC + lea rfilterq, [%2] +%else + %define rfilterq %2 +%endif + movdqu m0, [rfilterq ] ;load 128bit of x +%ifnum %3 + movdqu m1, [rfilterq+ %3] ;load 128bit of x+stride + movdqu m2, [rfilterq+2*%3] ;load 128bit of x+2*stride + movdqu m3, [rfilterq+3*%3] ;load 128bit of x+3*stride +%else + movdqu m1, [rfilterq+ %3q] ;load 128bit of x+stride + movdqu m2, [rfilterq+2*%3q] ;load 128bit of x+2*stride + movdqu m3, [rfilterq+r3srcq] ;load 128bit of x+2*stride +%endif + +%if %1 == 8 +%if %4 > 8 + SBUTTERFLY bw, 0, 1, 10 + SBUTTERFLY bw, 2, 3, 10 +%else + punpcklbw m0, m1 + punpcklbw m2, m3 +%endif +%else +%if %4 > 4 + SBUTTERFLY wd, 0, 1, 10 + SBUTTERFLY wd, 2, 3, 10 +%else + punpcklwd m0, m1 + punpcklwd m2, m3 +%endif +%endif +%endmacro + + +%macro QPEL_H_LOAD 4 +%assign %%stride (%1+7)/8 +%if %1 == 8 +%if %3 <= 4 +%define %%load movd +%elif %3 == 8 +%define %%load movq +%else +%define %%load movdqu +%endif +%else +%if %3 == 2 +%define %%load movd +%elif %3 == 4 +%define %%load movq +%else +%define %%load movdqu +%endif +%endif + %%load m0, [%2-3*%%stride] ;load data from source + %%load m1, [%2-2*%%stride] + %%load m2, [%2-%%stride ] + %%load m3, [%2 ] + %%load m4, [%2+%%stride ] + %%load m5, [%2+2*%%stride] + %%load m6, [%2+3*%%stride] + %%load m7, [%2+4*%%stride] + +%if %1 == 8 +%if %3 > 8 + SBUTTERFLY wd, 0, 1, %4 + SBUTTERFLY wd, 2, 3, %4 + SBUTTERFLY wd, 4, 5, %4 + SBUTTERFLY wd, 6, 7, %4 +%else + punpcklwd m0, m1 + punpcklwd m2, m3 + punpcklwd m4, m5 + punpcklwd m6, m7 +%endif +%else +%if %3 > 4 + SBUTTERFLY dq, 0, 1, %4 + SBUTTERFLY dq, 2, 3, %4 + SBUTTERFLY dq, 4, 5, %4 + SBUTTERFLY dq, 6, 7, %4 +%else + punpckldq m0, m1 + punpckldq m2, m3 + punpckldq m4, m5 + punpckldq m6, m7 +%endif +%endif +%endmacro + +%macro QPEL_V_LOAD 4 + lea r12q, [%2] + sub r12q, r3srcq + movdqu m0, [r12 ] ;load x- 3*srcstride + movdqu m1, [r12+ %3q ] ;load x- 2*srcstride + movdqu m2, [r12+ 2*%3q ] ;load x-srcstride + movdqu m3, [%2 ] ;load x + movdqu m4, [%2+ %3q] ;load x+stride + movdqu m5, [%2+ 2*%3q] ;load x+2*stride + movdqu m6, [%2+r3srcq] ;load x+3*stride + movdqu m7, [%2+ 4*%3q] ;load x+4*stride +%if %1 == 8 +%if %4 > 8 + SBUTTERFLY bw, 0, 1, 8 + SBUTTERFLY bw, 2, 3, 8 + SBUTTERFLY bw, 4, 5, 8 + SBUTTERFLY bw, 6, 7, 8 +%else + punpcklbw m0, m1 + punpcklbw m2, m3 + punpcklbw m4, m5 + punpcklbw m6, m7 +%endif +%else +%if %4 > 4 + SBUTTERFLY wd, 0, 1, 8 + SBUTTERFLY wd, 2, 3, 8 + SBUTTERFLY wd, 4, 5, 8 + SBUTTERFLY wd, 6, 7, 8 +%else + punpcklwd m0, m1 + punpcklwd m2, m3 + punpcklwd m4, m5 + punpcklwd m6, m7 +%endif +%endif +%endmacro + +%macro PEL_10STORE2 3 + movd [%1], %2 +%endmacro +%macro PEL_10STORE4 3 + movq [%1], %2 +%endmacro +%macro PEL_10STORE6 3 + movq [%1], %2 + psrldq %2, 8 + movd [%1+8], %2 +%endmacro +%macro PEL_10STORE8 3 + movdqa [%1], %2 +%endmacro +%macro PEL_10STORE12 3 + movdqa [%1], %2 + movq [%1+16], %3 +%endmacro +%macro PEL_10STORE16 3 + PEL_10STORE8 %1, %2, %3 + movdqa [%1+16], %3 +%endmacro + +%macro PEL_8STORE2 3 + pextrw [%1], %2, 0 +%endmacro +%macro PEL_8STORE4 3 + movd [%1], %2 +%endmacro +%macro PEL_8STORE6 3 + movd [%1], %2 + pextrw [%1+4], %2, 2 +%endmacro +%macro PEL_8STORE8 3 + movq [%1], %2 +%endmacro +%macro PEL_8STORE12 3 + movq [%1], %2 + psrldq %2, 8 + movd [%1+8], %2 +%endmacro +%macro PEL_8STORE16 3 + movdqa [%1], %2 +%endmacro + +%macro LOOP_END 4 + lea %1q, [%1q+2*%2q] ; dst += dststride + lea %3q, [%3q+ %4q] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop +%endmacro + + +%macro MC_PIXEL_COMPUTE 2 ;width, bitdepth +%if %2 == 8 +%if %1 > 8 + punpckhbw m1, m0, m2 + psllw m1, 14-%2 +%endif + punpcklbw m0, m2 +%endif + psllw m0, 14-%2 +%endmacro + + +%macro EPEL_COMPUTE 4 ; bitdepth, width, filter1, filter2 +%if %1 == 8 + pmaddubsw m0, %3 ;x1*c1+x2*c2 + pmaddubsw m2, %4 ;x3*c3+x4*c4 + paddw m0, m2 +%if %2 > 8 + pmaddubsw m1, %3 + pmaddubsw m3, %4 + paddw m1, m3 +%endif +%else + pmaddwd m0, %3 + pmaddwd m2, %4 + paddd m0, m2 +%if %2 > 4 + pmaddwd m1, %3 + pmaddwd m3, %4 + paddd m1, m3 +%endif + psrad m0, %1-8 + psrad m1, %1-8 + packssdw m0, m1 +%endif +%endmacro + +%macro QPEL_HV_COMPUTE 4 ; width, bitdepth, filter idx +%ifdef PIC + lea rfilterq, [hevc_qpel_filters_sse4_%2] +%else + %define rfilterq hevc_qpel_filters_sse4_%2 +%endif + +%if %2 == 8 + pmaddubsw m0, [rfilterq + %3q*8 ] ;x1*c1+x2*c2 + pmaddubsw m2, [rfilterq + %3q*8+16] ;x3*c3+x4*c4 + pmaddubsw m4, [rfilterq + %3q*8+32] ;x5*c5+x6*c6 + pmaddubsw m6, [rfilterq + %3q*8+48] ;x7*c7+x8*c8 + paddw m0, m2 + paddw m4, m6 + paddw m0, m4 +%else + pmaddwd m0, [rfilterq + %3q*8 ] + pmaddwd m2, [rfilterq + %3q*8+16] + pmaddwd m4, [rfilterq + %3q*8+32] + pmaddwd m6, [rfilterq + %3q*8+48] + paddd m0, m2 + paddd m4, m6 + paddd m0, m4 + psrad m0, %2-8 +%if %1 > 4 + pmaddwd m1, [rfilterq + %3q*8 ] + pmaddwd m3, [rfilterq + %3q*8+16] + pmaddwd m5, [rfilterq + %3q*8+32] + pmaddwd m7, [rfilterq + %3q*8+48] + paddd m1, m3 + paddd m5, m7 + paddd m1, m5 + psrad m1, %2-8 +%endif + p%4 m0, m1 +%endif +%endmacro + +%macro QPEL_COMPUTE 2 ; width, bitdepth +%if %2 == 8 + pmaddubsw m0, m12 ;x1*c1+x2*c2 + pmaddubsw m2, m13 ;x3*c3+x4*c4 + pmaddubsw m4, m14 ;x5*c5+x6*c6 + pmaddubsw m6, m15 ;x7*c7+x8*c8 + paddw m0, m2 + paddw m4, m6 + paddw m0, m4 +%if %1 > 8 + pmaddubsw m1, m12 + pmaddubsw m3, m13 + pmaddubsw m5, m14 + pmaddubsw m7, m15 + paddw m1, m3 + paddw m5, m7 + paddw m1, m5 +%endif +%else + pmaddwd m0, m12 + pmaddwd m2, m13 + pmaddwd m4, m14 + pmaddwd m6, m15 + paddd m0, m2 + paddd m4, m6 + paddd m0, m4 + psrad m0, %2-8 +%if %1 > 4 + pmaddwd m1, m12 + pmaddwd m3, m13 + pmaddwd m5, m14 + pmaddwd m7, m15 + paddd m1, m3 + paddd m5, m7 + paddd m1, m5 + psrad m1, %2-8 +%endif +%endif +%endmacro + +%macro BI_COMPUTE 7 ; width, bitd, src1l, src1h, scr2l, scr2h, pw + paddsw %3, %5 +%if %1 > 8 + paddsw %4, %6 +%endif + UNI_COMPUTE %1, %2, %3, %4, %7 +%endmacro + +%macro UNI_COMPUTE 5 + pmulhrsw %3, %5 +%if %1 > 8 || (%2 > 8 && %1 > 4) + pmulhrsw %4, %5 +%endif +%if %2 == 8 + packuswb %3, %4 +%else + pminsw %3, [max_pixels_%2] + pmaxsw %3, [zero] +%if %1 > 8 + pminsw %4, [max_pixels_%2] + pmaxsw %4, [zero] +%endif +%endif +%endmacro + +INIT_XMM sse4 ; adds ff_ and _sse4 to function name +; ****************************** +; void put_hevc_mc_pixels(int16_t *dst, ptrdiff_t dststride, +; uint8_t *_src, ptrdiff_t _srcstride, +; int height, int mx, int my) +; ****************************** + +%macro HEVC_PUT_HEVC_PEL_PIXELS 2 +cglobal hevc_put_hevc_pel_pixels%1_%2, 5, 5, 3, dst, dststride, src, srcstride,height + pxor m2, m2 +.loop + SIMPLE_LOAD %1, %2, srcq, m0 + MC_PIXEL_COMPUTE %1, %2 + PEL_10STORE%1 dstq, m0, m1 + LOOP_END dst, dststride, src, srcstride + RET + +cglobal hevc_put_hevc_uni_pel_pixels%1_%2, 5, 5, 3, dst, dststride, src, srcstride,height + pxor m2, m2 +.loop + SIMPLE_LOAD %1, %2, srcq, m0 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + +cglobal hevc_put_hevc_bi_pel_pixels%1_%2, 7, 7, 6, dst, dststride, src, srcstride, src2, src2stride,height + pxor m2, m2 + movdqa m5, [pw_bi_%2] +.loop + SIMPLE_LOAD %1, %2, srcq, m0 + SIMPLE_BILOAD %1, src2q, m3, m4 + MC_PIXEL_COMPUTE %1, %2 + BI_COMPUTE %1, %2, m0, m1, m3, m4, m5 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + lea src2q, [src2q+2*src2strideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + +%endmacro + + +; ****************************** +; void put_hevc_epel_hX(int16_t *dst, ptrdiff_t dststride, +; uint8_t *_src, ptrdiff_t _srcstride, +; int width, int height, int mx, int my, +; int16_t* mcbuffer) +; ****************************** + + +%macro HEVC_PUT_HEVC_EPEL 2 +cglobal hevc_put_hevc_epel_h%1_%2, 6, 7, 15 , dst, dststride, src, srcstride, height, mx, rfilter +%assign %%stride ((%2 + 7)/8) + EPEL_FILTER %2, mx +.loop + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + PEL_10STORE%1 dstq, m0, m1 + LOOP_END dst, dststride, src, srcstride + RET + +cglobal hevc_put_hevc_uni_epel_h%1_%2, 6, 7, 15 , dst, dststride, src, srcstride, height, mx, rfilter +%assign %%stride ((%2 + 7)/8) + movdqa m9, [pw_%2] + EPEL_FILTER %2, mx +.loop + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + UNI_COMPUTE %1, %2, m0, m1, m9 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + +cglobal hevc_put_hevc_bi_epel_h%1_%2, 8, 8, 15, dst, dststride, src, srcstride, src2, src2stride,height, mx, rfilter + movdqa m9, [pw_bi_%2] + EPEL_FILTER %2, mx +.loop + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SIMPLE_BILOAD %1, src2q, m2, m3 + BI_COMPUTE %1, %2, m0, m1, m2, m3, m9 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + lea src2q, [src2q+2*src2strideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + +; ****************************** +; void put_hevc_epel_v(int16_t *dst, ptrdiff_t dststride, +; uint8_t *_src, ptrdiff_t _srcstride, +; int width, int height, int mx, int my, +; int16_t* mcbuffer) +; ****************************** + +cglobal hevc_put_hevc_epel_v%1_%2, 7, 8, 15 , dst, dststride, src, srcstride, height, r3src, my, rfilter + lea r3srcq, [srcstrideq*3] + sub srcq, srcstrideq + EPEL_FILTER %2, my +.loop + EPEL_LOAD %2, srcq, srcstride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + PEL_10STORE%1 dstq, m0, m1 + LOOP_END dst, dststride, src, srcstride + RET + +cglobal hevc_put_hevc_uni_epel_v%1_%2, 7, 8, 15 , dst, dststride, src, srcstride, height, r3src, my, rfilter + lea r3srcq, [srcstrideq*3] + movdqa m9, [pw_%2] + sub srcq, srcstrideq + EPEL_FILTER %2, my +.loop + EPEL_LOAD %2, srcq, srcstride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + UNI_COMPUTE %1, %2, m0, m1, m9 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + + +cglobal hevc_put_hevc_bi_epel_v%1_%2, 9, 10, 15, dst, dststride, src, srcstride, src2, src2stride,height, r3src, my, rfilter + lea r3srcq, [srcstrideq*3] + movdqa m9, [pw_bi_%2] + sub srcq, srcstrideq + EPEL_FILTER %2, my +.loop + EPEL_LOAD %2, srcq, srcstride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SIMPLE_BILOAD %1, src2q, m2, m3 + BI_COMPUTE %1, %2, m0, m1, m2, m3, m9 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + lea src2q, [src2q+2*src2strideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET +%endmacro + + +; ****************************** +; void put_hevc_epel_hv(int16_t *dst, ptrdiff_t dststride, +; uint8_t *_src, ptrdiff_t _srcstride, +; int width, int height, int mx, int my) +; ****************************** + +%macro HEVC_PUT_HEVC_EPEL_HV 2 +cglobal hevc_put_hevc_epel_hv%1_%2, 7, 9, 12 , dst, dststride, src, srcstride, height, mx, my, r3src, rfilter +%assign %%stride ((%2 + 7)/8) + sub srcq, srcstrideq + EPEL_HV_FILTER %2 + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m4, m0 + lea srcq, [srcq + srcstrideq] + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m5, m0 + lea srcq, [srcq + srcstrideq] + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m6, m0 + lea srcq, [srcq + srcstrideq] +.loop + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m7, m0 + punpcklwd m0, m4, m5 + punpcklwd m2, m6, m7 +%if %1 > 4 + punpckhwd m1, m4, m5 + punpckhwd m3, m6, m7 +%endif + EPEL_COMPUTE 14, %1, m12, m13 + PEL_10STORE%1 dstq, m0, m1 + movdqa m4, m5 + movdqa m5, m6 + movdqa m6, m7 + LOOP_END dst, dststride, src, srcstride + RET + +cglobal hevc_put_hevc_uni_epel_hv%1_%2, 7, 9, 12 , dst, dststride, src, srcstride, height, mx, my, r3src, rfilter +%assign %%stride ((%2 + 7)/8) + sub srcq, srcstrideq + EPEL_HV_FILTER %2 + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m4, m0 + lea srcq, [srcq + srcstrideq] + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m5, m0 + lea srcq, [srcq + srcstrideq] + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m6, m0 + lea srcq, [srcq + srcstrideq] +.loop + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m7, m0 + punpcklwd m0, m4, m5 + punpcklwd m2, m6, m7 +%if %1 > 4 + punpckhwd m1, m4, m5 + punpckhwd m3, m6, m7 +%endif + EPEL_COMPUTE 14, %1, m12, m13 + UNI_COMPUTE %1, %2, m0, m1, [pw_%2] + PEL_%2STORE%1 dstq, m0, m1 + movdqa m4, m5 + movdqa m5, m6 + movdqa m6, m7 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + + +cglobal hevc_put_hevc_bi_epel_hv%1_%2, 9, 11, 16, dst, dststride, src, srcstride, src2, src2stride, height, mx, my, r3src, rfilter +%assign %%stride ((%2 + 7)/8) + sub srcq, srcstrideq + EPEL_HV_FILTER %2 + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m4, m0 + lea srcq, [srcq + srcstrideq] + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m5, m0 + lea srcq, [srcq + srcstrideq] + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m6, m0 + lea srcq, [srcq + srcstrideq] +.loop + EPEL_LOAD %2, srcq-%%stride, %%stride, %1 + EPEL_COMPUTE %2, %1, m14, m15 + SWAP m7, m0 + punpcklwd m0, m4, m5 + punpcklwd m2, m6, m7 +%if %1 > 4 + punpckhwd m1, m4, m5 + punpckhwd m3, m6, m7 +%endif + EPEL_COMPUTE 14, %1, m12, m13 + SIMPLE_BILOAD %1, src2q, m8, m9 + BI_COMPUTE %1, %2, m0, m1, m8, m9, [pw_bi_%2] + PEL_%2STORE%1 dstq, m0, m1 + movdqa m4, m5 + movdqa m5, m6 + movdqa m6, m7 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + lea src2q, [src2q+2*src2strideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET +%endmacro + +; ****************************** +; void put_hevc_qpel_hX_X_X(int16_t *dst, ptrdiff_t dststride, +; uint8_t *_src, ptrdiff_t _srcstride, +; int width, int height, int mx, int my) +; ****************************** + +%macro HEVC_PUT_HEVC_QPEL 2 +cglobal hevc_put_hevc_qpel_h%1_%2, 6, 7, 15 , dst, dststride, src, srcstride, height, mx, rfilter + QPEL_FILTER %2, mx +.loop + QPEL_H_LOAD %2, srcq, %1, 10 + QPEL_COMPUTE %1, %2 +%if %2 > 8 + packssdw m0, m1 +%endif + PEL_10STORE%1 dstq, m0, m1 + LOOP_END dst, dststride, src, srcstride + RET + +cglobal hevc_put_hevc_uni_qpel_h%1_%2, 6, 7, 15 , dst, dststride, src, srcstride, height, mx, rfilter + movdqa m9, [pw_%2] + QPEL_FILTER %2, mx +.loop + QPEL_H_LOAD %2, srcq, %1, 10 + QPEL_COMPUTE %1, %2 +%if %2 > 8 + packssdw m0, m1 +%endif + UNI_COMPUTE %1, %2, m0, m1, m9 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + +cglobal hevc_put_hevc_bi_qpel_h%1_%2, 8, 9, 16 , dst, dststride, src, srcstride, src2, src2stride, height, mx, rfilter + movdqa m9, [pw_bi_%2] + QPEL_FILTER %2, mx +.loop + QPEL_H_LOAD %2, srcq, %1, 10 + QPEL_COMPUTE %1, %2 +%if %2 > 8 + packssdw m0, m1 +%endif + SIMPLE_BILOAD %1, src2q, m10, m11 + BI_COMPUTE %1, %2, m0, m1, m10, m11, m9 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + lea src2q, [src2q+2*src2strideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + + +; ****************************** +; void put_hevc_qpel_vX_X_X(int16_t *dst, ptrdiff_t dststride, +; uint8_t *_src, ptrdiff_t _srcstride, +; int width, int height, int mx, int my) +; ****************************** + +cglobal hevc_put_hevc_qpel_v%1_%2, 7, 14, 15 , dst, dststride, src, srcstride, height, r3src, my, rfilter + lea r3srcq, [srcstrideq*3] + QPEL_FILTER %2, my +.loop + QPEL_V_LOAD %2, srcq, srcstride, %1 + QPEL_COMPUTE %1, %2 +%if %2 > 8 + packssdw m0, m1 +%endif + PEL_10STORE%1 dstq, m0, m1 + LOOP_END dst, dststride, src, srcstride + RET + +cglobal hevc_put_hevc_uni_qpel_v%1_%2, 7, 14, 15 , dst, dststride, src, srcstride, height, r3src, my, rfilter + movdqa m9, [pw_%2] + lea r3srcq, [srcstrideq*3] + QPEL_FILTER %2, my +.loop + QPEL_V_LOAD %2, srcq, srcstride, %1 + QPEL_COMPUTE %1, %2 +%if %2 > 8 + packusdw m0, m1 +%endif + UNI_COMPUTE %1, %2, m0, m1, m9 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + +cglobal hevc_put_hevc_bi_qpel_v%1_%2, 9, 14, 16 , dst, dststride, src, srcstride, src2, src2stride, height, r3src, my, rfilter + movdqa m9, [pw_bi_%2] + lea r3srcq, [srcstrideq*3] + QPEL_FILTER %2, my +.loop + SIMPLE_BILOAD %1, src2q, m10, m11 + QPEL_V_LOAD %2, srcq, srcstride, %1 + QPEL_COMPUTE %1, %2 +%if %2 > 8 + packssdw m0, m1 +%endif + BI_COMPUTE %1, %2, m0, m1, m10, m11, m9 + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + lea src2q, [src2q+2*src2strideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET +%endmacro + + +; ****************************** +; void put_hevc_qpel_hvX_X(int16_t *dst, ptrdiff_t dststride, +; uint8_t *_src, ptrdiff_t _srcstride, +; int height, int mx, int my) +; ****************************** +%macro HEVC_PUT_HEVC_QPEL_HV 2 +cglobal hevc_put_hevc_qpel_hv%1_%2, 7, 9, 12 , dst, dststride, src, srcstride, height, mx, my, r3src, rfilter + lea mxq, [mxq*8-8] + lea myq, [myq*8-8] + lea r3srcq, [srcstrideq*3] + sub srcq, r3srcq + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m8, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m9, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m10, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m11, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m12, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m13, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m14, m0 + lea srcq, [srcq + srcstrideq] +.loop + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m15, m0 + punpcklwd m0, m8, m9 + punpcklwd m2, m10, m11 + punpcklwd m4, m12, m13 + punpcklwd m6, m14, m15 +%if %1 > 4 + punpckhwd m1, m8, m9 + punpckhwd m3, m10, m11 + punpckhwd m5, m12, m13 + punpckhwd m7, m14, m15 +%endif + QPEL_HV_COMPUTE %1, 14, my, ackssdw + PEL_10STORE%1 dstq, m0, m1 +%if %1 <= 4 + movq m8, m9 + movq m9, m10 + movq m10, m11 + movq m11, m12 + movq m12, m13 + movq m13, m14 + movq m14, m15 +%else + movdqa m8, m9 + movdqa m9, m10 + movdqa m10, m11 + movdqa m11, m12 + movdqa m12, m13 + movdqa m13, m14 + movdqa m14, m15 +%endif + LOOP_END dst, dststride, src, srcstride + RET + +cglobal hevc_put_hevc_uni_qpel_hv%1_%2, 7, 9, 12 , dst, dststride, src, srcstride, height, mx, my, r3src, rfilter + lea mxq, [mxq*8-8] + lea myq, [myq*8-8] + lea r3srcq, [srcstrideq*3] + sub srcq, r3srcq + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m8, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m9, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m10, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m11, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m12, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m13, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m14, m0 + lea srcq, [srcq + srcstrideq] +.loop + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m15, m0 + punpcklwd m0, m8, m9 + punpcklwd m2, m10, m11 + punpcklwd m4, m12, m13 + punpcklwd m6, m14, m15 +%if %1 > 4 + punpckhwd m1, m8, m9 + punpckhwd m3, m10, m11 + punpckhwd m5, m12, m13 + punpckhwd m7, m14, m15 +%endif + QPEL_HV_COMPUTE %1, 14, my, ackusdw + UNI_COMPUTE %1, %2, m0, m1, [pw_%2] + PEL_%2STORE%1 dstq, m0, m1 + +%if %1 <= 4 + movq m8, m9 + movq m9, m10 + movq m10, m11 + movq m11, m12 + movq m12, m13 + movq m13, m14 + movq m14, m15 +%else + movdqa m8, m9 + movdqa m9, m10 + movdqa m10, m11 + movdqa m11, m12 + movdqa m12, m13 + movdqa m13, m14 + movdqa m14, m15 +%endif + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + +cglobal hevc_put_hevc_bi_qpel_hv%1_%2, 9, 11, 16, dst, dststride, src, srcstride, src2, src2stride, height, mx, my, r3src, rfilter + lea mxq, [mxq*8-8] + lea myq, [myq*8-8] + lea r3srcq, [srcstrideq*3] + sub srcq, r3srcq + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m8, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m9, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m10, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m11, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m12, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m13, m0 + lea srcq, [srcq + srcstrideq] + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m14, m0 + lea srcq, [srcq + srcstrideq] +.loop + QPEL_H_LOAD %2, srcq, %1, 15 + QPEL_HV_COMPUTE %1, %2, mx, ackssdw + SWAP m15, m0 + punpcklwd m0, m8, m9 + punpcklwd m2, m10, m11 + punpcklwd m4, m12, m13 + punpcklwd m6, m14, m15 +%if %1 > 4 + punpckhwd m1, m8, m9 + punpckhwd m3, m10, m11 + punpckhwd m5, m12, m13 + punpckhwd m7, m14, m15 +%endif + QPEL_HV_COMPUTE %1, 14, my, ackssdw + SIMPLE_BILOAD %1, src2q, m8, m9 ;m9 not used in this case + BI_COMPUTE %1, %2, m0, m1, m8, m9, [pw_bi_%2] + PEL_%2STORE%1 dstq, m0, m1 + +%if %1 <= 4 + movq m8, m9 + movq m9, m10 + movq m10, m11 + movq m11, m12 + movq m12, m13 + movq m13, m14 + movq m14, m15 +%else + movdqa m8, m9 + movdqa m9, m10 + movdqa m10, m11 + movdqa m11, m12 + movdqa m12, m13 + movdqa m13, m14 + movdqa m14, m15 +%endif + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+srcstrideq] ; src += srcstride + lea src2q, [src2q+2*src2strideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET +%endmacro + +%macro WEIGHTING_FUNCS 2 +cglobal hevc_put_hevc_uni_w%1_%2, 8, 10, 11, dst, dststride, src, srcstride, height, denom, wx, ox, shift + lea shiftd, [denomd+14-%2] ; shift = 14 - bitd + denom + shl oxd, %2-8 ; ox << (bitd - 8) + movd m2, wxd ; WX + movd m3, oxd ; OX + movd m4, shiftd ; shift + punpcklwd m2, m2 + pshufd m3, m3, 0 + pshufd m2, m2, 0 + sub shiftd, 1 + movd m6, shiftd + movdqu m5, [one_per_32] + pslld m5, m6 +.loop + SIMPLE_LOAD %1, 10, srcq, m0 + pmulhw m6, m0, m2 + pmullw m0, m2 + punpckhwd m1, m0, m6 + punpcklwd m0, m6 + paddd m0, m5 + paddd m1, m5 + psrad m0, m4 + psrad m1, m4 + paddd m0, m3 + paddd m1, m3 + packusdw m0, m1 +%if %2 == 8 + packuswb m0, m0 +%else + pminsw m0, [max_pixels_%2] +%endif + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+2*srcstrideq] ; src += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET + +cglobal hevc_put_hevc_bi_w%1_%2, 12, 14, 14, dst, dststride, src, srcstride, src2, src2stride, height, denom, wx0, wx1, ox0, ox1, shift, temp + shl ox0d, %2-8 ; ox << (bitd - 8) + shl ox1d, %2-8 ; ox << (bitd - 8) + lea shiftd, [denomd+14-%2] ; shift = 14 - bitd + denom + movd m2, wx0d ; WX0 + movd m3, wx1d ; WX1 + punpcklwd m2, m2 + punpcklwd m3, m3 + pshufd m2, m2, 0 + pshufd m3, m3, 0 + add ox0d, ox1d + add ox0d, 1 + movd m4, ox0d ; offset + pshufd m4, m4, 0 + movd m5, shiftd ; shift + pslld m4, m5 + add shiftd, 1 + movd m5, shiftd ; shift + +.loop + SIMPLE_LOAD %1, 10, srcq, m0 + SIMPLE_LOAD %1, 10, src2q, m10 + pmulhw m6, m0, m3 + pmullw m0, m3 + pmulhw m7, m10, m2 + pmullw m10, m2 + punpckhwd m1, m0, m6 + punpcklwd m0, m6 + punpckhwd m11, m10, m7 + punpcklwd m10, m7 + paddd m0, m10 + paddd m1, m11 + paddd m0, m4 + paddd m1, m4 + psrad m0, m5 + psrad m1, m5 + packusdw m0, m1 +%if %2 == 8 + packuswb m0, m0 +%else + pminsw m0, [max_pixels_%2] +%endif + PEL_%2STORE%1 dstq, m0, m1 + lea dstq, [dstq+dststrideq] ; dst += dststride + lea srcq, [srcq+2*srcstrideq] ; src += srcstride + lea src2q, [src2q+2*src2strideq] ; src2 += srcstride + dec heightd ; cmp height + jnz .loop ; height loop + RET +%endmacro + +WEIGHTING_FUNCS 2, 8 +WEIGHTING_FUNCS 4, 8 +WEIGHTING_FUNCS 6, 8 +WEIGHTING_FUNCS 8, 8 + +WEIGHTING_FUNCS 2, 10 +WEIGHTING_FUNCS 4, 10 +WEIGHTING_FUNCS 6, 10 +WEIGHTING_FUNCS 8, 10 + +HEVC_PUT_HEVC_PEL_PIXELS 2, 8 +HEVC_PUT_HEVC_PEL_PIXELS 4, 8 +HEVC_PUT_HEVC_PEL_PIXELS 6, 8 +HEVC_PUT_HEVC_PEL_PIXELS 8, 8 +HEVC_PUT_HEVC_PEL_PIXELS 12, 8 +HEVC_PUT_HEVC_PEL_PIXELS 16, 8 + +HEVC_PUT_HEVC_PEL_PIXELS 2, 10 +HEVC_PUT_HEVC_PEL_PIXELS 4, 10 +HEVC_PUT_HEVC_PEL_PIXELS 6, 10 +HEVC_PUT_HEVC_PEL_PIXELS 8, 10 + + +HEVC_PUT_HEVC_EPEL 2, 8 +HEVC_PUT_HEVC_EPEL 4, 8 +HEVC_PUT_HEVC_EPEL 6, 8 +HEVC_PUT_HEVC_EPEL 8, 8 +HEVC_PUT_HEVC_EPEL 12, 8 +HEVC_PUT_HEVC_EPEL 16, 8 + + +HEVC_PUT_HEVC_EPEL 2, 10 +HEVC_PUT_HEVC_EPEL 4, 10 +HEVC_PUT_HEVC_EPEL 6, 10 +HEVC_PUT_HEVC_EPEL 8, 10 + + +HEVC_PUT_HEVC_EPEL_HV 2, 8 +HEVC_PUT_HEVC_EPEL_HV 4, 8 +HEVC_PUT_HEVC_EPEL_HV 6, 8 +HEVC_PUT_HEVC_EPEL_HV 8, 8 + +HEVC_PUT_HEVC_EPEL_HV 2, 10 +HEVC_PUT_HEVC_EPEL_HV 4, 10 +HEVC_PUT_HEVC_EPEL_HV 6, 10 +HEVC_PUT_HEVC_EPEL_HV 8, 10 + + +HEVC_PUT_HEVC_QPEL 4, 8 +HEVC_PUT_HEVC_QPEL 8, 8 +HEVC_PUT_HEVC_QPEL 12, 8 +HEVC_PUT_HEVC_QPEL 16, 8 + +HEVC_PUT_HEVC_QPEL 4, 10 +HEVC_PUT_HEVC_QPEL 8, 10 + +HEVC_PUT_HEVC_QPEL_HV 2, 8 +HEVC_PUT_HEVC_QPEL_HV 4, 8 +HEVC_PUT_HEVC_QPEL_HV 6, 8 +HEVC_PUT_HEVC_QPEL_HV 8, 8 + +HEVC_PUT_HEVC_QPEL_HV 2, 10 +HEVC_PUT_HEVC_QPEL_HV 4, 10 +HEVC_PUT_HEVC_QPEL_HV 6, 10 +HEVC_PUT_HEVC_QPEL_HV 8, 10 + +%endif ; ARCH_X86_64 diff --git a/libavcodec/x86/hevcdsp.h b/libavcodec/x86/hevcdsp.h new file mode 100644 index 0000000000..f1c2ffb06e --- /dev/null +++ b/libavcodec/x86/hevcdsp.h @@ -0,0 +1,120 @@ +/* + * HEVC video decoder + * + * Copyright (C) 2012 - 2013 Guillaume Martres + * Copyright (C) 2013 - 2014 Pierre-Edouard Lepere + * + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#ifndef AVCODEC_X86_HEVCDSP_H +#define AVCODEC_X86_HEVCDSP_H + +#define PEL_LINK(dst, idx1, idx2, idx3, name, D) \ +dst[idx1][idx2][idx3] = ff_hevc_put_hevc_ ## name ## _ ## D ## _sse4; \ +dst ## _bi[idx1][idx2][idx3] = ff_hevc_put_hevc_bi_ ## name ## _ ## D ## _sse4; \ +dst ## _uni[idx1][idx2][idx3] = ff_hevc_put_hevc_uni_ ## name ## _ ## D ## _sse4; \ +dst ## _uni_w[idx1][idx2][idx3] = ff_hevc_put_hevc_uni_w_ ## name ## _ ## D ## _sse4; \ +dst ## _bi_w[idx1][idx2][idx3] = ff_hevc_put_hevc_bi_w_ ## name ## _ ## D ## _sse4 + + +#define PEL_PROTOTYPE(name, D) \ +void ff_hevc_put_hevc_ ## name ## _ ## D ## _sse4(int16_t *dst, ptrdiff_t dststride,uint8_t *_src, ptrdiff_t _srcstride, int height, intptr_t mx, intptr_t my,int width); \ +void ff_hevc_put_hevc_bi_ ## name ## _ ## D ## _sse4(uint8_t *_dst, ptrdiff_t _dststride, uint8_t *_src, ptrdiff_t _srcstride, int16_t *src2, ptrdiff_t src2stride, int height, intptr_t mx, intptr_t my, int width); \ +void ff_hevc_put_hevc_uni_ ## name ## _ ## D ## _sse4(uint8_t *_dst, ptrdiff_t _dststride, uint8_t *_src, ptrdiff_t _srcstride, int height, intptr_t mx, intptr_t my, int width); \ +void ff_hevc_put_hevc_uni_w_ ## name ## _ ## D ## _sse4(uint8_t *_dst, ptrdiff_t _dststride, uint8_t *_src, ptrdiff_t _srcstride, int height, int denom, int wx, int ox, intptr_t mx, intptr_t my, int width); \ +void ff_hevc_put_hevc_bi_w_ ## name ## _ ## D ## _sse4(uint8_t *_dst, ptrdiff_t _dststride, uint8_t *_src, ptrdiff_t _srcstride, int16_t *src2, ptrdiff_t src2stride, int height, int denom, int wx0, int wx1, int ox0, int ox1, intptr_t mx, intptr_t my, int width) + + +/////////////////////////////////////////////////////////////////////////////// +// MC functions +/////////////////////////////////////////////////////////////////////////////// + +#define EPEL_PROTOTYPES(fname, bitd) \ + PEL_PROTOTYPE(fname##4, bitd); \ + PEL_PROTOTYPE(fname##6, bitd); \ + PEL_PROTOTYPE(fname##8, bitd); \ + PEL_PROTOTYPE(fname##12, bitd); \ + PEL_PROTOTYPE(fname##16, bitd); \ + PEL_PROTOTYPE(fname##24, bitd); \ + PEL_PROTOTYPE(fname##32, bitd); \ + PEL_PROTOTYPE(fname##48, bitd); \ + PEL_PROTOTYPE(fname##64, bitd) + +#define QPEL_PROTOTYPES(fname, bitd) \ + PEL_PROTOTYPE(fname##4, bitd); \ + PEL_PROTOTYPE(fname##8, bitd); \ + PEL_PROTOTYPE(fname##12, bitd); \ + PEL_PROTOTYPE(fname##16, bitd); \ + PEL_PROTOTYPE(fname##24, bitd); \ + PEL_PROTOTYPE(fname##32, bitd); \ + PEL_PROTOTYPE(fname##48, bitd); \ + PEL_PROTOTYPE(fname##64, bitd) + +#define WEIGHTING_PROTOTYPE(width, bitd) \ +void ff_hevc_put_hevc_uni_w##width##_##bitd##_sse4(uint8_t *dst, ptrdiff_t dststride, int16_t *_src, ptrdiff_t _srcstride, int height, int denom, int _wx, int _ox); \ +void ff_hevc_put_hevc_bi_w##width##_##bitd##_sse4(uint8_t *dst, ptrdiff_t dststride, int16_t *_src, ptrdiff_t _srcstride, int16_t *_src2, ptrdiff_t _src2stride, int height, int denom, int _wx0, int _wx1, int _ox0, int _ox1) + +#define WEIGHTING_PROTOTYPES(bitd) \ + WEIGHTING_PROTOTYPE(2, bitd); \ + WEIGHTING_PROTOTYPE(4, bitd); \ + WEIGHTING_PROTOTYPE(6, bitd); \ + WEIGHTING_PROTOTYPE(8, bitd); \ + WEIGHTING_PROTOTYPE(12, bitd); \ + WEIGHTING_PROTOTYPE(16, bitd); \ + WEIGHTING_PROTOTYPE(24, bitd); \ + WEIGHTING_PROTOTYPE(32, bitd); \ + WEIGHTING_PROTOTYPE(48, bitd); \ + WEIGHTING_PROTOTYPE(64, bitd) + + +/////////////////////////////////////////////////////////////////////////////// +// QPEL_PIXELS EPEL_PIXELS +/////////////////////////////////////////////////////////////////////////////// +EPEL_PROTOTYPES(pel_pixels , 8); +EPEL_PROTOTYPES(pel_pixels , 10); +/////////////////////////////////////////////////////////////////////////////// +// EPEL +/////////////////////////////////////////////////////////////////////////////// +EPEL_PROTOTYPES(epel_h , 8); +EPEL_PROTOTYPES(epel_h , 10); + +EPEL_PROTOTYPES(epel_v , 8); +EPEL_PROTOTYPES(epel_v , 10); + +EPEL_PROTOTYPES(epel_hv , 8); +EPEL_PROTOTYPES(epel_hv , 10); + +/////////////////////////////////////////////////////////////////////////////// +// QPEL +/////////////////////////////////////////////////////////////////////////////// +QPEL_PROTOTYPES(qpel_h , 8); +QPEL_PROTOTYPES(qpel_h , 10); + +QPEL_PROTOTYPES(qpel_v, 8); +QPEL_PROTOTYPES(qpel_v, 10); + +QPEL_PROTOTYPES(qpel_hv, 8); +QPEL_PROTOTYPES(qpel_hv, 10); + + +WEIGHTING_PROTOTYPES(8); +WEIGHTING_PROTOTYPES(10); + + +#endif // AVCODEC_X86_HEVCDSP_H diff --git a/libavcodec/x86/hevcdsp_init.c b/libavcodec/x86/hevcdsp_init.c new file mode 100644 index 0000000000..565c4a0a6e --- /dev/null +++ b/libavcodec/x86/hevcdsp_init.c @@ -0,0 +1,385 @@ +/* + * Copyright (c) 2013 Seppo Tomperi + * Copyright (c) 2013 - 2014 Pierre-Edouard Lepere + * + * + * This file is part of ffmpeg. + * + * ffmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * ffmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with ffmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include "config.h" +#include "libavutil/cpu.h" +#include "libavutil/x86/asm.h" +#include "libavutil/x86/cpu.h" +#include "libavcodec/get_bits.h" /* required for hevcdsp.h GetBitContext */ +#include "libavcodec/hevcdsp.h" +#include "libavcodec/x86/hevcdsp.h" + + +#define mc_rep_func(name, bitd, step, W) \ +void ff_hevc_put_hevc_##name##W##_##bitd##_sse4(int16_t *_dst, ptrdiff_t dststride, \ + uint8_t *_src, ptrdiff_t _srcstride, int height, \ + intptr_t mx, intptr_t my, int width) \ +{ \ + int i; \ + uint8_t *src; \ + int16_t *dst; \ + for (i = 0; i < W; i += step) { \ + src = _src + (i * ((bitd + 7) / 8)); \ + dst = _dst + i; \ + ff_hevc_put_hevc_##name##step##_##bitd##_sse4(dst, dststride, src, _srcstride, height, mx, my, width); \ + } \ +} +#define mc_rep_uni_func(name, bitd, step, W) \ +void ff_hevc_put_hevc_uni_##name##W##_##bitd##_sse4(uint8_t *_dst, ptrdiff_t dststride, \ + uint8_t *_src, ptrdiff_t _srcstride, int height, \ + intptr_t mx, intptr_t my, int width) \ +{ \ + int i; \ + uint8_t *src; \ + uint8_t *dst; \ + for (i = 0; i < W; i += step) { \ + src = _src + (i * ((bitd + 7) / 8)); \ + dst = _dst + (i * ((bitd + 7) / 8)); \ + ff_hevc_put_hevc_uni_##name##step##_##bitd##_sse4(dst, dststride, src, _srcstride, \ + height, mx, my, width); \ + } \ +} +#define mc_rep_bi_func(name, bitd, step, W) \ +void ff_hevc_put_hevc_bi_##name##W##_##bitd##_sse4(uint8_t *_dst, ptrdiff_t dststride, uint8_t *_src, \ + ptrdiff_t _srcstride, int16_t* _src2, ptrdiff_t _src2stride, \ + int height, intptr_t mx, intptr_t my, int width) \ +{ \ + int i; \ + uint8_t *src; \ + uint8_t *dst; \ + int16_t *src2; \ + for (i = 0; i < W ; i += step) { \ + src = _src + (i * ((bitd + 7) / 8)); \ + dst = _dst + (i * ((bitd + 7) / 8)); \ + src2 = _src2 + i; \ + ff_hevc_put_hevc_bi_##name##step##_##bitd##_sse4(dst, dststride, src, _srcstride, src2, \ + _src2stride, height, mx, my, width); \ + } \ +} + +#define mc_rep_funcs(name, bitd, step, W) \ + mc_rep_func(name, bitd, step, W); \ + mc_rep_uni_func(name, bitd, step, W); \ + mc_rep_bi_func(name, bitd, step, W) + + +mc_rep_funcs(pel_pixels, 8, 16, 64); +mc_rep_funcs(pel_pixels, 8, 16, 48); +mc_rep_funcs(pel_pixels, 8, 16, 32); +mc_rep_funcs(pel_pixels, 8, 8, 24); + +mc_rep_funcs(pel_pixels,10, 8, 64); +mc_rep_funcs(pel_pixels,10, 8, 48); +mc_rep_funcs(pel_pixels,10, 8, 32); +mc_rep_funcs(pel_pixels,10, 8, 24); +mc_rep_funcs(pel_pixels,10, 8, 16); +mc_rep_funcs(pel_pixels,10, 4, 12); + +mc_rep_funcs(epel_h, 8, 16, 64); +mc_rep_funcs(epel_h, 8, 16, 48); +mc_rep_funcs(epel_h, 8, 16, 32); +mc_rep_funcs(epel_h, 8, 8, 24); +mc_rep_funcs(epel_h,10, 8, 64); +mc_rep_funcs(epel_h,10, 8, 48); +mc_rep_funcs(epel_h,10, 8, 32); +mc_rep_funcs(epel_h,10, 8, 24); +mc_rep_funcs(epel_h,10, 8, 16); +mc_rep_funcs(epel_h,10, 4, 12); +mc_rep_funcs(epel_v, 8, 16, 64); +mc_rep_funcs(epel_v, 8, 16, 48); +mc_rep_funcs(epel_v, 8, 16, 32); +mc_rep_funcs(epel_v, 8, 8, 24); +mc_rep_funcs(epel_v,10, 8, 64); +mc_rep_funcs(epel_v,10, 8, 48); +mc_rep_funcs(epel_v,10, 8, 32); +mc_rep_funcs(epel_v,10, 8, 24); +mc_rep_funcs(epel_v,10, 8, 16); +mc_rep_funcs(epel_v,10, 4, 12); +mc_rep_funcs(epel_hv, 8, 8, 64); +mc_rep_funcs(epel_hv, 8, 8, 48); +mc_rep_funcs(epel_hv, 8, 8, 32); +mc_rep_funcs(epel_hv, 8, 8, 24); +mc_rep_funcs(epel_hv, 8, 8, 16); +mc_rep_funcs(epel_hv, 8, 4, 12); +mc_rep_funcs(epel_hv,10, 8, 64); +mc_rep_funcs(epel_hv,10, 8, 48); +mc_rep_funcs(epel_hv,10, 8, 32); +mc_rep_funcs(epel_hv,10, 8, 24); +mc_rep_funcs(epel_hv,10, 8, 16); +mc_rep_funcs(epel_hv,10, 4, 12); + +mc_rep_funcs(qpel_h, 8, 16, 64); +mc_rep_funcs(qpel_h, 8, 16, 48); +mc_rep_funcs(qpel_h, 8, 16, 32); +mc_rep_funcs(qpel_h, 8, 8, 24); +mc_rep_funcs(qpel_h,10, 8, 64); +mc_rep_funcs(qpel_h,10, 8, 48); +mc_rep_funcs(qpel_h,10, 8, 32); +mc_rep_funcs(qpel_h,10, 8, 24); +mc_rep_funcs(qpel_h,10, 8, 16); +mc_rep_funcs(qpel_h,10, 4, 12); +mc_rep_funcs(qpel_v, 8, 16, 64); +mc_rep_funcs(qpel_v, 8, 16, 48); +mc_rep_funcs(qpel_v, 8, 16, 32); +mc_rep_funcs(qpel_v, 8, 8, 24); +mc_rep_funcs(qpel_v,10, 8, 64); +mc_rep_funcs(qpel_v,10, 8, 48); +mc_rep_funcs(qpel_v,10, 8, 32); +mc_rep_funcs(qpel_v,10, 8, 24); +mc_rep_funcs(qpel_v,10, 8, 16); +mc_rep_funcs(qpel_v,10, 4, 12); +mc_rep_funcs(qpel_hv, 8, 8, 64); +mc_rep_funcs(qpel_hv, 8, 8, 48); +mc_rep_funcs(qpel_hv, 8, 8, 32); +mc_rep_funcs(qpel_hv, 8, 8, 24); +mc_rep_funcs(qpel_hv, 8, 8, 16); +mc_rep_funcs(qpel_hv, 8, 4, 12); +mc_rep_funcs(qpel_hv,10, 8, 64); +mc_rep_funcs(qpel_hv,10, 8, 48); +mc_rep_funcs(qpel_hv,10, 8, 32); +mc_rep_funcs(qpel_hv,10, 8, 24); +mc_rep_funcs(qpel_hv,10, 8, 16); +mc_rep_funcs(qpel_hv,10, 4, 12); + +#define mc_rep_uni_w(bitd, step, W) \ +void ff_hevc_put_hevc_uni_w##W##_##bitd##_sse4(uint8_t *_dst, ptrdiff_t dststride, int16_t *_src, ptrdiff_t _srcstride, \ + int height, int denom, int _wx, int _ox) \ +{ \ + int i; \ + int16_t *src; \ + uint8_t *dst; \ + for (i = 0; i < W; i += step) { \ + src= _src + i; \ + dst= _dst + (i * ((bitd + 7) / 8)); \ + ff_hevc_put_hevc_uni_w##step##_##bitd##_sse4(dst, dststride, src, _srcstride, \ + height, denom, _wx, _ox); \ + } \ +} + +mc_rep_uni_w(8, 6, 12); +mc_rep_uni_w(8, 8, 16); +mc_rep_uni_w(8, 8, 24); +mc_rep_uni_w(8, 8, 32); +mc_rep_uni_w(8, 8, 48); +mc_rep_uni_w(8, 8, 64); + +mc_rep_uni_w(10, 6, 12); +mc_rep_uni_w(10, 8, 16); +mc_rep_uni_w(10, 8, 24); +mc_rep_uni_w(10, 8, 32); +mc_rep_uni_w(10, 8, 48); +mc_rep_uni_w(10, 8, 64); + +#define mc_rep_bi_w(bitd, step, W) \ +void ff_hevc_put_hevc_bi_w##W##_##bitd##_sse4(uint8_t *_dst, ptrdiff_t dststride, int16_t *_src, ptrdiff_t _srcstride, \ + int16_t *_src2, ptrdiff_t _src2stride, int height, \ + int denom, int _wx0, int _wx1, int _ox0, int _ox1) \ +{ \ + int i; \ + int16_t *src; \ + int16_t *src2; \ + uint8_t *dst; \ + for (i = 0; i < W; i += step) { \ + src = _src + i; \ + src2 = _src2 + i; \ + dst = _dst + (i * ((bitd + 7) / 8)); \ + ff_hevc_put_hevc_bi_w##step##_##bitd##_sse4(dst, dststride, src, _srcstride, src2, _src2stride, \ + height, denom, _wx0, _wx1, _ox0, _ox1); \ + } \ +} + +mc_rep_bi_w(8, 6, 12); +mc_rep_bi_w(8, 8, 16); +mc_rep_bi_w(8, 8, 24); +mc_rep_bi_w(8, 8, 32); +mc_rep_bi_w(8, 8, 48); +mc_rep_bi_w(8, 8, 64); + +mc_rep_bi_w(10, 6, 12); +mc_rep_bi_w(10, 8, 16); +mc_rep_bi_w(10, 8, 24); +mc_rep_bi_w(10, 8, 32); +mc_rep_bi_w(10, 8, 48); +mc_rep_bi_w(10, 8, 64); + +#define mc_uni_w_func(name, bitd, W) \ +void ff_hevc_put_hevc_uni_w_##name##W##_##bitd##_sse4(uint8_t *_dst, ptrdiff_t _dststride, \ + uint8_t *_src, ptrdiff_t _srcstride, \ + int height, int denom, \ + int _wx, int _ox, \ + intptr_t mx, intptr_t my, int width) \ +{ \ + LOCAL_ALIGNED_16(int16_t, temp, [71 * 64]); \ + ff_hevc_put_hevc_##name##W##_##bitd##_sse4(temp, 64, _src, _srcstride, height, mx, my, width); \ + ff_hevc_put_hevc_uni_w##W##_##bitd##_sse4(_dst, _dststride, temp, 64, height, denom, _wx, _ox); \ +} + +#define mc_uni_w_funcs(name, bitd) \ + mc_uni_w_func(name, bitd, 4); \ + mc_uni_w_func(name, bitd, 8); \ + mc_uni_w_func(name, bitd, 12); \ + mc_uni_w_func(name, bitd, 16); \ + mc_uni_w_func(name, bitd, 24); \ + mc_uni_w_func(name, bitd, 32); \ + mc_uni_w_func(name, bitd, 48); \ + mc_uni_w_func(name, bitd, 64) + +mc_uni_w_funcs(pel_pixels, 8); +mc_uni_w_func(pel_pixels, 8, 6); +mc_uni_w_funcs(epel_h, 8); +mc_uni_w_func(epel_h, 8, 6); +mc_uni_w_funcs(epel_v, 8); +mc_uni_w_func(epel_v, 8, 6); +mc_uni_w_funcs(epel_hv, 8); +mc_uni_w_func(epel_hv, 8, 6); +mc_uni_w_funcs(qpel_h, 8); +mc_uni_w_funcs(qpel_v, 8); +mc_uni_w_funcs(qpel_hv, 8); + +mc_uni_w_funcs(pel_pixels, 10); +mc_uni_w_func(pel_pixels, 10, 6); +mc_uni_w_funcs(epel_h, 10); +mc_uni_w_func(epel_h, 10, 6); +mc_uni_w_funcs(epel_v, 10); +mc_uni_w_func(epel_v, 10, 6); +mc_uni_w_funcs(epel_hv, 10); +mc_uni_w_func(epel_hv, 10, 6); +mc_uni_w_funcs(qpel_h, 10); +mc_uni_w_funcs(qpel_v, 10); +mc_uni_w_funcs(qpel_hv, 10); + + +#define mc_bi_w_func(name, bitd, W) \ +void ff_hevc_put_hevc_bi_w_##name##W##_##bitd##_sse4(uint8_t *_dst, ptrdiff_t _dststride, \ + uint8_t *_src, ptrdiff_t _srcstride, \ + int16_t *_src2, ptrdiff_t _src2stride, \ + int height, int denom, \ + int _wx0, int _wx1, int _ox0, int _ox1, \ + intptr_t mx, intptr_t my, int width) \ +{ \ + LOCAL_ALIGNED_16(int16_t, temp, [71 * 64]); \ + ff_hevc_put_hevc_##name##W##_##bitd##_sse4(temp, 64, _src, _srcstride, height, mx, my, width); \ + ff_hevc_put_hevc_bi_w##W##_##bitd##_sse4(_dst, _dststride, temp, 64, _src2, _src2stride, \ + height, denom, _wx0, _wx1, _ox0, _ox1); \ +} + +#define mc_bi_w_funcs(name, bitd) \ + mc_bi_w_func(name, bitd, 4); \ + mc_bi_w_func(name, bitd, 8); \ + mc_bi_w_func(name, bitd, 12); \ + mc_bi_w_func(name, bitd, 16); \ + mc_bi_w_func(name, bitd, 24); \ + mc_bi_w_func(name, bitd, 32); \ + mc_bi_w_func(name, bitd, 48); \ + mc_bi_w_func(name, bitd, 64) + +mc_bi_w_funcs(pel_pixels, 8); +mc_bi_w_func(pel_pixels, 8, 6); +mc_bi_w_funcs(epel_h, 8); +mc_bi_w_func(epel_h, 8, 6); +mc_bi_w_funcs(epel_v, 8); +mc_bi_w_func(epel_v, 8, 6); +mc_bi_w_funcs(epel_hv, 8); +mc_bi_w_func(epel_hv, 8, 6); +mc_bi_w_funcs(qpel_h, 8); +mc_bi_w_funcs(qpel_v, 8); +mc_bi_w_funcs(qpel_hv, 8); + +mc_bi_w_funcs(pel_pixels, 10); +mc_bi_w_func(pel_pixels, 10, 6); +mc_bi_w_funcs(epel_h, 10); +mc_bi_w_func(epel_h, 10, 6); +mc_bi_w_funcs(epel_v, 10); +mc_bi_w_func(epel_v, 10, 6); +mc_bi_w_funcs(epel_hv, 10); +mc_bi_w_func(epel_hv, 10, 6); +mc_bi_w_funcs(qpel_h, 10); +mc_bi_w_funcs(qpel_v, 10); +mc_bi_w_funcs(qpel_hv, 10); + + +#define EPEL_LINKS(pointer, my, mx, fname, bitd) \ + PEL_LINK(pointer, 1, my , mx , fname##4 , bitd ); \ + PEL_LINK(pointer, 2, my , mx , fname##6 , bitd ); \ + PEL_LINK(pointer, 3, my , mx , fname##8 , bitd ); \ + PEL_LINK(pointer, 4, my , mx , fname##12, bitd ); \ + PEL_LINK(pointer, 5, my , mx , fname##16, bitd ); \ + PEL_LINK(pointer, 6, my , mx , fname##24, bitd ); \ + PEL_LINK(pointer, 7, my , mx , fname##32, bitd ); \ + PEL_LINK(pointer, 8, my , mx , fname##48, bitd ); \ + PEL_LINK(pointer, 9, my , mx , fname##64, bitd ) +#define QPEL_LINKS(pointer, my, mx, fname, bitd) \ + PEL_LINK(pointer, 1, my , mx , fname##4 , bitd ); \ + PEL_LINK(pointer, 3, my , mx , fname##8 , bitd ); \ + PEL_LINK(pointer, 4, my , mx , fname##12, bitd ); \ + PEL_LINK(pointer, 5, my , mx , fname##16, bitd ); \ + PEL_LINK(pointer, 6, my , mx , fname##24, bitd ); \ + PEL_LINK(pointer, 7, my , mx , fname##32, bitd ); \ + PEL_LINK(pointer, 8, my , mx , fname##48, bitd ); \ + PEL_LINK(pointer, 9, my , mx , fname##64, bitd ) + + +void ff_hevcdsp_init_x86(HEVCDSPContext *c, const int bit_depth) +{ + int mm_flags = av_get_cpu_flags(); + + if (bit_depth == 8) { + if (EXTERNAL_MMX(mm_flags)) { + + if (EXTERNAL_MMXEXT(mm_flags)) { + + if (EXTERNAL_SSSE3(mm_flags) && ARCH_X86_64) { + + EPEL_LINKS(c->put_hevc_epel, 0, 0, pel_pixels, 8); + EPEL_LINKS(c->put_hevc_epel, 0, 1, epel_h, 8); + EPEL_LINKS(c->put_hevc_epel, 1, 0, epel_v, 8); + EPEL_LINKS(c->put_hevc_epel, 1, 1, epel_hv, 8); + + QPEL_LINKS(c->put_hevc_qpel, 0, 0, pel_pixels, 8); + QPEL_LINKS(c->put_hevc_qpel, 0, 1, qpel_h, 8); + QPEL_LINKS(c->put_hevc_qpel, 1, 0, qpel_v, 8); + QPEL_LINKS(c->put_hevc_qpel, 1, 1, qpel_hv, 8); + + } + } + } + } else if (bit_depth == 10) { + if (EXTERNAL_MMX(mm_flags)) { + if (EXTERNAL_MMXEXT(mm_flags) && ARCH_X86_64) { + + if (EXTERNAL_SSSE3(mm_flags)) { + + EPEL_LINKS(c->put_hevc_epel, 0, 0, pel_pixels, 10); + EPEL_LINKS(c->put_hevc_epel, 0, 1, epel_h, 10); + EPEL_LINKS(c->put_hevc_epel, 1, 0, epel_v, 10); + EPEL_LINKS(c->put_hevc_epel, 1, 1, epel_hv, 10); + + QPEL_LINKS(c->put_hevc_qpel, 0, 0, pel_pixels, 10); + QPEL_LINKS(c->put_hevc_qpel, 0, 1, qpel_h, 10); + QPEL_LINKS(c->put_hevc_qpel, 1, 0, qpel_v, 10); + QPEL_LINKS(c->put_hevc_qpel, 1, 1, qpel_hv, 10); + } + + } + } + } +} |