diff options
author | Michael Niedermayer <michaelni@gmx.at> | 2013-03-10 16:48:31 -0700 |
---|---|---|
committer | Michael Niedermayer <michaelni@gmx.at> | 2013-03-13 04:07:41 +0100 |
commit | a0565a2b656f51a942c0113b7591e7f2d255a173 (patch) | |
tree | 8d66709ccbd9da0170e2a68258345147ac68edc6 /libavcodec | |
parent | e9e608ad5804d483de23e43b98d7af4d1b51e2e4 (diff) | |
download | ffmpeg-a0565a2b656f51a942c0113b7591e7f2d255a173.tar.gz |
Move sh4 half-pel assembly from dsputil to hpeldsp.
Signed-off-by: Michael Niedermayer <michaelni@gmx.at>
Diffstat (limited to 'libavcodec')
-rw-r--r-- | libavcodec/hpeldsp.c | 2 | ||||
-rw-r--r-- | libavcodec/sh4/Makefile | 2 | ||||
-rw-r--r-- | libavcodec/sh4/dsputil_align.c | 94 | ||||
-rw-r--r-- | libavcodec/sh4/dsputil_sh4.h | 6 | ||||
-rw-r--r-- | libavcodec/sh4/hpeldsp.c | 347 |
5 files changed, 365 insertions, 86 deletions
diff --git a/libavcodec/hpeldsp.c b/libavcodec/hpeldsp.c index b323d98b7f..b55d036070 100644 --- a/libavcodec/hpeldsp.c +++ b/libavcodec/hpeldsp.c @@ -60,8 +60,6 @@ av_cold void ff_hpeldsp_init(HpelDSPContext* c, int flags) if (ARCH_ALPHA) ff_hpeldsp_init_alpha (c, flags); #endif if (ARCH_PPC) ff_hpeldsp_init_ppc (c, flags); -#if 0 if (ARCH_SH4) ff_hpeldsp_init_sh4 (c, flags); -#endif if (ARCH_BFIN) ff_hpeldsp_init_bfin (c, flags); } diff --git a/libavcodec/sh4/Makefile b/libavcodec/sh4/Makefile index f907408d9f..a24b298ead 100644 --- a/libavcodec/sh4/Makefile +++ b/libavcodec/sh4/Makefile @@ -3,3 +3,5 @@ OBJS += sh4/dsputil_align.o \ sh4/idct_sh4.o \ OBJS-$(CONFIG_H264CHROMA) += sh4/h264chroma_init.o \ + +OBJS-$(CONFIG_HPELDSP) += sh4/hpeldsp.o diff --git a/libavcodec/sh4/dsputil_align.c b/libavcodec/sh4/dsputil_align.c index 46981f6c87..d6bd479fca 100644 --- a/libavcodec/sh4/dsputil_align.c +++ b/libavcodec/sh4/dsputil_align.c @@ -245,55 +245,15 @@ static void op##_##rnd##_pixels##sz##_##xy (uint8_t * dest, const uint8_t * ref, } \ } -#define OP put - -DEFFUNC(put, rnd,o,8,OP_C,avg32) -DEFFUNC(put, rnd,x,8,OP_X,avg32) -DEFFUNC(put,no_rnd,x,8,OP_X,avg32) -DEFFUNC(put, rnd,y,8,OP_Y,avg32) -DEFFUNC(put,no_rnd,y,8,OP_Y,avg32) -DEFFUNC(put, rnd,xy,8,OP_XY,PACK) -DEFFUNC(put,no_rnd,xy,8,OP_XY,PACK) -DEFFUNC(put, rnd,o,16,OP_C,avg32) -DEFFUNC(put, rnd,x,16,OP_X,avg32) -DEFFUNC(put,no_rnd,x,16,OP_X,avg32) -DEFFUNC(put, rnd,y,16,OP_Y,avg32) -DEFFUNC(put,no_rnd,y,16,OP_Y,avg32) -DEFFUNC(put, rnd,xy,16,OP_XY,PACK) -DEFFUNC(put,no_rnd,xy,16,OP_XY,PACK) - -#undef OP -#define OP avg - -DEFFUNC(avg, rnd,o,8,OP_C,avg32) -DEFFUNC(avg, rnd,x,8,OP_X,avg32) -DEFFUNC(avg, rnd,y,8,OP_Y,avg32) -DEFFUNC(avg, rnd,xy,8,OP_XY,PACK) -DEFFUNC(avg, rnd,o,16,OP_C,avg32) -DEFFUNC(avg, rnd,x,16,OP_X,avg32) -DEFFUNC(avg,no_rnd,x,16,OP_X,avg32) -DEFFUNC(avg, rnd,y,16,OP_Y,avg32) -DEFFUNC(avg,no_rnd,y,16,OP_Y,avg32) -DEFFUNC(avg, rnd,xy,16,OP_XY,PACK) -DEFFUNC(avg,no_rnd,xy,16,OP_XY,PACK) - -#undef OP - -#define put_no_rnd_pixels8_o put_rnd_pixels8_o -#define put_no_rnd_pixels16_o put_rnd_pixels16_o -#define avg_no_rnd_pixels16_o avg_rnd_pixels16_o - -#define put_pixels8_c put_rnd_pixels8_o -#define put_pixels16_c put_rnd_pixels16_o -#define avg_pixels8_c avg_rnd_pixels8_o -#define avg_pixels16_c avg_rnd_pixels16_o -#define put_no_rnd_pixels8_c put_rnd_pixels8_o -#define put_no_rnd_pixels16_c put_rnd_pixels16_o -#define avg_no_rnd_pixels16_c avg_rnd_pixels16_o - -#define QPEL - -#ifdef QPEL +#define put_pixels8_c ff_put_rnd_pixels8_o +#define put_pixels16_c ff_put_rnd_pixels16_o +#define avg_pixels8_c ff_avg_rnd_pixels8_o +#define avg_pixels16_c ff_avg_rnd_pixels16_o +#define put_no_rnd_pixels8_c ff_put_rnd_pixels8_o +#define put_no_rnd_pixels16_c ff_put_rnd_pixels16_o +#define avg_no_rnd_pixels16_c ff_avg_rnd_pixels16_o + +#if CONFIG_H264QPEL #include "qpel.c" @@ -303,41 +263,7 @@ av_cold void ff_dsputil_init_align(DSPContext *c, AVCodecContext *avctx) { const int high_bit_depth = avctx->bits_per_raw_sample > 8; - if (!high_bit_depth) { - c->put_pixels_tab[0][0] = put_rnd_pixels16_o; - c->put_pixels_tab[0][1] = put_rnd_pixels16_x; - c->put_pixels_tab[0][2] = put_rnd_pixels16_y; - c->put_pixels_tab[0][3] = put_rnd_pixels16_xy; - c->put_pixels_tab[1][0] = put_rnd_pixels8_o; - c->put_pixels_tab[1][1] = put_rnd_pixels8_x; - c->put_pixels_tab[1][2] = put_rnd_pixels8_y; - c->put_pixels_tab[1][3] = put_rnd_pixels8_xy; - - c->put_no_rnd_pixels_tab[0][0] = put_no_rnd_pixels16_o; - c->put_no_rnd_pixels_tab[0][1] = put_no_rnd_pixels16_x; - c->put_no_rnd_pixels_tab[0][2] = put_no_rnd_pixels16_y; - c->put_no_rnd_pixels_tab[0][3] = put_no_rnd_pixels16_xy; - c->put_no_rnd_pixels_tab[1][0] = put_no_rnd_pixels8_o; - c->put_no_rnd_pixels_tab[1][1] = put_no_rnd_pixels8_x; - c->put_no_rnd_pixels_tab[1][2] = put_no_rnd_pixels8_y; - c->put_no_rnd_pixels_tab[1][3] = put_no_rnd_pixels8_xy; - - c->avg_pixels_tab[0][0] = avg_rnd_pixels16_o; - c->avg_pixels_tab[0][1] = avg_rnd_pixels16_x; - c->avg_pixels_tab[0][2] = avg_rnd_pixels16_y; - c->avg_pixels_tab[0][3] = avg_rnd_pixels16_xy; - c->avg_pixels_tab[1][0] = avg_rnd_pixels8_o; - c->avg_pixels_tab[1][1] = avg_rnd_pixels8_x; - c->avg_pixels_tab[1][2] = avg_rnd_pixels8_y; - c->avg_pixels_tab[1][3] = avg_rnd_pixels8_xy; - - c->avg_no_rnd_pixels_tab[0] = avg_no_rnd_pixels16_o; - c->avg_no_rnd_pixels_tab[1] = avg_no_rnd_pixels16_x; - c->avg_no_rnd_pixels_tab[2] = avg_no_rnd_pixels16_y; - c->avg_no_rnd_pixels_tab[3] = avg_no_rnd_pixels16_xy; - } - -#ifdef QPEL +#if CONFIG_H264QPEL #define dspfunc(PFX, IDX, NUM) \ c->PFX ## _pixels_tab[IDX][ 0] = PFX ## NUM ## _mc00_sh4; \ diff --git a/libavcodec/sh4/dsputil_sh4.h b/libavcodec/sh4/dsputil_sh4.h index 1c041ae27d..bd97a5b91f 100644 --- a/libavcodec/sh4/dsputil_sh4.h +++ b/libavcodec/sh4/dsputil_sh4.h @@ -21,8 +21,14 @@ #include "libavcodec/avcodec.h" #include "libavcodec/dsputil.h" +#include "libavcodec/hpeldsp.h" void ff_idct_sh4(int16_t *block); void ff_dsputil_init_align(DSPContext* c, AVCodecContext *avctx); +void ff_put_rnd_pixels8_o (uint8_t * dest, const uint8_t * ref, const int stride, int height); +void ff_put_rnd_pixels16_o(uint8_t * dest, const uint8_t * ref, const int stride, int height); +void ff_avg_rnd_pixels8_o (uint8_t * dest, const uint8_t * ref, const int stride, int height); +void ff_avg_rnd_pixels16_o(uint8_t * dest, const uint8_t * ref, const int stride, int height); + #endif /* AVCODEC_SH4_DSPUTIL_SH4_H */ diff --git a/libavcodec/sh4/hpeldsp.c b/libavcodec/sh4/hpeldsp.c new file mode 100644 index 0000000000..b524003f3e --- /dev/null +++ b/libavcodec/sh4/hpeldsp.c @@ -0,0 +1,347 @@ +/* + * aligned/packed access motion + * + * Copyright (c) 2001-2003 BERO <bero@geocities.co.jp> + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + + +#include "libavcodec/avcodec.h" +#include "libavcodec/dsputil.h" +#include "libavcodec/bit_depth_template.c" // for BYTE_VEC32 +#include "dsputil_sh4.h" + + +#define LP(p) *(uint32_t*)(p) +#define LPC(p) *(const uint32_t*)(p) + + +#define UNPACK(ph,pl,tt0,tt1) do { \ + uint32_t t0,t1; t0=tt0;t1=tt1; \ + ph = ( (t0 & ~BYTE_VEC32(0x03))>>2) + ( (t1 & ~BYTE_VEC32(0x03))>>2); \ + pl = (t0 & BYTE_VEC32(0x03)) + (t1 & BYTE_VEC32(0x03)); } while(0) + +#define rnd_PACK(ph,pl,nph,npl) ph + nph + (((pl + npl + BYTE_VEC32(0x02))>>2) & BYTE_VEC32(0x03)) +#define no_rnd_PACK(ph,pl,nph,npl) ph + nph + (((pl + npl + BYTE_VEC32(0x01))>>2) & BYTE_VEC32(0x03)) + +/* little-endian */ +#define MERGE1(a,b,ofs) (ofs==0)?a:( ((a)>>(8*ofs))|((b)<<(32-8*ofs)) ) +#define MERGE2(a,b,ofs) (ofs==3)?b:( ((a)>>(8*(ofs+1)))|((b)<<(32-8*(ofs+1))) ) +/* big +#define MERGE1(a,b,ofs) (ofs==0)?a:( ((a)<<(8*ofs))|((b)>>(32-8*ofs)) ) +#define MERGE2(a,b,ofs) (ofs==3)?b:( ((a)<<(8+8*ofs))|((b)>>(32-8-8*ofs)) ) +*/ + + +#define put(d,s) d = s +#define avg(d,s) d = rnd_avg32(s,d) + +#define OP_C4(ofs) \ + ref-=ofs; \ + do { \ + OP(LP(dest),MERGE1(LPC(ref),LPC(ref+4),ofs)); \ + ref+=stride; \ + dest+=stride; \ + } while(--height) + +#define OP_C40() \ + do { \ + OP(LP(dest),LPC(ref)); \ + ref+=stride; \ + dest+=stride; \ + } while(--height) + + +#define OP put + +static void put_pixels4_c(uint8_t *dest,const uint8_t *ref, const int stride,int height) +{ + switch((int)ref&3){ + case 0: OP_C40(); return; + case 1: OP_C4(1); return; + case 2: OP_C4(2); return; + case 3: OP_C4(3); return; + } +} + +#undef OP +#define OP avg + +static void avg_pixels4_c(uint8_t *dest,const uint8_t *ref, const int stride,int height) +{ + switch((int)ref&3){ + case 0: OP_C40(); return; + case 1: OP_C4(1); return; + case 2: OP_C4(2); return; + case 3: OP_C4(3); return; + } +} + +#undef OP + +#define OP_C(ofs,sz,avg2) \ +{ \ + ref-=ofs; \ + do { \ + uint32_t t0,t1; \ + t0 = LPC(ref+0); \ + t1 = LPC(ref+4); \ + OP(LP(dest+0), MERGE1(t0,t1,ofs)); \ + t0 = LPC(ref+8); \ + OP(LP(dest+4), MERGE1(t1,t0,ofs)); \ +if (sz==16) { \ + t1 = LPC(ref+12); \ + OP(LP(dest+8), MERGE1(t0,t1,ofs)); \ + t0 = LPC(ref+16); \ + OP(LP(dest+12), MERGE1(t1,t0,ofs)); \ +} \ + ref+=stride; \ + dest+= stride; \ + } while(--height); \ +} + +/* aligned */ +#define OP_C0(sz,avg2) \ +{ \ + do { \ + OP(LP(dest+0), LPC(ref+0)); \ + OP(LP(dest+4), LPC(ref+4)); \ +if (sz==16) { \ + OP(LP(dest+8), LPC(ref+8)); \ + OP(LP(dest+12), LPC(ref+12)); \ +} \ + ref+=stride; \ + dest+= stride; \ + } while(--height); \ +} + +#define OP_X(ofs,sz,avg2) \ +{ \ + ref-=ofs; \ + do { \ + uint32_t t0,t1; \ + t0 = LPC(ref+0); \ + t1 = LPC(ref+4); \ + OP(LP(dest+0), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \ + t0 = LPC(ref+8); \ + OP(LP(dest+4), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \ +if (sz==16) { \ + t1 = LPC(ref+12); \ + OP(LP(dest+8), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \ + t0 = LPC(ref+16); \ + OP(LP(dest+12), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \ +} \ + ref+=stride; \ + dest+= stride; \ + } while(--height); \ +} + +/* aligned */ +#define OP_Y0(sz,avg2) \ +{ \ + uint32_t t0,t1,t2,t3,t; \ +\ + t0 = LPC(ref+0); \ + t1 = LPC(ref+4); \ +if (sz==16) { \ + t2 = LPC(ref+8); \ + t3 = LPC(ref+12); \ +} \ + do { \ + ref += stride; \ +\ + t = LPC(ref+0); \ + OP(LP(dest+0), avg2(t0,t)); t0 = t; \ + t = LPC(ref+4); \ + OP(LP(dest+4), avg2(t1,t)); t1 = t; \ +if (sz==16) { \ + t = LPC(ref+8); \ + OP(LP(dest+8), avg2(t2,t)); t2 = t; \ + t = LPC(ref+12); \ + OP(LP(dest+12), avg2(t3,t)); t3 = t; \ +} \ + dest+= stride; \ + } while(--height); \ +} + +#define OP_Y(ofs,sz,avg2) \ +{ \ + uint32_t t0,t1,t2,t3,t,w0,w1; \ +\ + ref-=ofs; \ + w0 = LPC(ref+0); \ + w1 = LPC(ref+4); \ + t0 = MERGE1(w0,w1,ofs); \ + w0 = LPC(ref+8); \ + t1 = MERGE1(w1,w0,ofs); \ +if (sz==16) { \ + w1 = LPC(ref+12); \ + t2 = MERGE1(w0,w1,ofs); \ + w0 = LPC(ref+16); \ + t3 = MERGE1(w1,w0,ofs); \ +} \ + do { \ + ref += stride; \ +\ + w0 = LPC(ref+0); \ + w1 = LPC(ref+4); \ + t = MERGE1(w0,w1,ofs); \ + OP(LP(dest+0), avg2(t0,t)); t0 = t; \ + w0 = LPC(ref+8); \ + t = MERGE1(w1,w0,ofs); \ + OP(LP(dest+4), avg2(t1,t)); t1 = t; \ +if (sz==16) { \ + w1 = LPC(ref+12); \ + t = MERGE1(w0,w1,ofs); \ + OP(LP(dest+8), avg2(t2,t)); t2 = t; \ + w0 = LPC(ref+16); \ + t = MERGE1(w1,w0,ofs); \ + OP(LP(dest+12), avg2(t3,t)); t3 = t; \ +} \ + dest+=stride; \ + } while(--height); \ +} + +#define OP_X0(sz,avg2) OP_X(0,sz,avg2) +#define OP_XY0(sz,PACK) OP_XY(0,sz,PACK) +#define OP_XY(ofs,sz,PACK) \ +{ \ + uint32_t t2,t3,w0,w1; \ + uint32_t a0,a1,a2,a3,a4,a5,a6,a7; \ +\ + ref -= ofs; \ + w0 = LPC(ref+0); \ + w1 = LPC(ref+4); \ + UNPACK(a0,a1,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \ + w0 = LPC(ref+8); \ + UNPACK(a2,a3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \ +if (sz==16) { \ + w1 = LPC(ref+12); \ + UNPACK(a4,a5,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \ + w0 = LPC(ref+16); \ + UNPACK(a6,a7,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \ +} \ + do { \ + ref+=stride; \ + w0 = LPC(ref+0); \ + w1 = LPC(ref+4); \ + UNPACK(t2,t3,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \ + OP(LP(dest+0),PACK(a0,a1,t2,t3)); \ + a0 = t2; a1 = t3; \ + w0 = LPC(ref+8); \ + UNPACK(t2,t3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \ + OP(LP(dest+4),PACK(a2,a3,t2,t3)); \ + a2 = t2; a3 = t3; \ +if (sz==16) { \ + w1 = LPC(ref+12); \ + UNPACK(t2,t3,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \ + OP(LP(dest+8),PACK(a4,a5,t2,t3)); \ + a4 = t2; a5 = t3; \ + w0 = LPC(ref+16); \ + UNPACK(t2,t3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \ + OP(LP(dest+12),PACK(a6,a7,t2,t3)); \ + a6 = t2; a7 = t3; \ +} \ + dest+=stride; \ + } while(--height); \ +} + +#define DEFFUNC(prefix, op,rnd,xy,sz,OP_N,avgfunc) \ +prefix void op##_##rnd##_pixels##sz##_##xy (uint8_t * dest, const uint8_t * ref, \ + const int stride, int height) \ +{ \ + switch((int)ref&3) { \ + case 0:OP_N##0(sz,rnd##_##avgfunc); return; \ + case 1:OP_N(1,sz,rnd##_##avgfunc); return; \ + case 2:OP_N(2,sz,rnd##_##avgfunc); return; \ + case 3:OP_N(3,sz,rnd##_##avgfunc); return; \ + } \ +} + +#define OP put + +DEFFUNC( ,ff_put,rnd,o,8,OP_C,avg32) +DEFFUNC(static,put, rnd,x,8,OP_X,avg32) +DEFFUNC(static,put,no_rnd,x,8,OP_X,avg32) +DEFFUNC(static,put, rnd,y,8,OP_Y,avg32) +DEFFUNC(static,put,no_rnd,y,8,OP_Y,avg32) +DEFFUNC(static,put, rnd,xy,8,OP_XY,PACK) +DEFFUNC(static,put,no_rnd,xy,8,OP_XY,PACK) +DEFFUNC( ,ff_put,rnd,o,16,OP_C,avg32) +DEFFUNC(static,put, rnd,x,16,OP_X,avg32) +DEFFUNC(static,put,no_rnd,x,16,OP_X,avg32) +DEFFUNC(static,put, rnd,y,16,OP_Y,avg32) +DEFFUNC(static,put,no_rnd,y,16,OP_Y,avg32) +DEFFUNC(static,put, rnd,xy,16,OP_XY,PACK) +DEFFUNC(static,put,no_rnd,xy,16,OP_XY,PACK) + +#undef OP +#define OP avg + +DEFFUNC( ,ff_avg,rnd,o,8,OP_C,avg32) +DEFFUNC(static,avg, rnd,x,8,OP_X,avg32) +DEFFUNC(static,avg, rnd,y,8,OP_Y,avg32) +DEFFUNC(static,avg, rnd,xy,8,OP_XY,PACK) +DEFFUNC( ,ff_avg,rnd,o,16,OP_C,avg32) +DEFFUNC(static,avg, rnd,x,16,OP_X,avg32) +DEFFUNC(static,avg,no_rnd,x,16,OP_X,avg32) +DEFFUNC(static,avg, rnd,y,16,OP_Y,avg32) +DEFFUNC(static,avg,no_rnd,y,16,OP_Y,avg32) +DEFFUNC(static,avg, rnd,xy,16,OP_XY,PACK) +DEFFUNC(static,avg,no_rnd,xy,16,OP_XY,PACK) + +#undef OP + +#define ff_put_no_rnd_pixels8_o ff_put_rnd_pixels8_o +#define ff_put_no_rnd_pixels16_o ff_put_rnd_pixels16_o +#define ff_avg_no_rnd_pixels16_o ff_avg_rnd_pixels16_o + +void ff_hpeldsp_init_sh4(HpelDSPContext* c, int flags) +{ + c->put_pixels_tab[0][0] = ff_put_rnd_pixels16_o; + c->put_pixels_tab[0][1] = put_rnd_pixels16_x; + c->put_pixels_tab[0][2] = put_rnd_pixels16_y; + c->put_pixels_tab[0][3] = put_rnd_pixels16_xy; + c->put_pixels_tab[1][0] = ff_put_rnd_pixels8_o; + c->put_pixels_tab[1][1] = put_rnd_pixels8_x; + c->put_pixels_tab[1][2] = put_rnd_pixels8_y; + c->put_pixels_tab[1][3] = put_rnd_pixels8_xy; + + c->put_no_rnd_pixels_tab[0][0] = ff_put_no_rnd_pixels16_o; + c->put_no_rnd_pixels_tab[0][1] = put_no_rnd_pixels16_x; + c->put_no_rnd_pixels_tab[0][2] = put_no_rnd_pixels16_y; + c->put_no_rnd_pixels_tab[0][3] = put_no_rnd_pixels16_xy; + c->put_no_rnd_pixels_tab[1][0] = ff_put_no_rnd_pixels8_o; + c->put_no_rnd_pixels_tab[1][1] = put_no_rnd_pixels8_x; + c->put_no_rnd_pixels_tab[1][2] = put_no_rnd_pixels8_y; + c->put_no_rnd_pixels_tab[1][3] = put_no_rnd_pixels8_xy; + + c->avg_pixels_tab[0][0] = ff_avg_rnd_pixels16_o; + c->avg_pixels_tab[0][1] = avg_rnd_pixels16_x; + c->avg_pixels_tab[0][2] = avg_rnd_pixels16_y; + c->avg_pixels_tab[0][3] = avg_rnd_pixels16_xy; + c->avg_pixels_tab[1][0] = ff_avg_rnd_pixels8_o; + c->avg_pixels_tab[1][1] = avg_rnd_pixels8_x; + c->avg_pixels_tab[1][2] = avg_rnd_pixels8_y; + c->avg_pixels_tab[1][3] = avg_rnd_pixels8_xy; + + c->avg_no_rnd_pixels_tab[0] = ff_avg_no_rnd_pixels16_o; + c->avg_no_rnd_pixels_tab[1] = avg_no_rnd_pixels16_x; + c->avg_no_rnd_pixels_tab[2] = avg_no_rnd_pixels16_y; + c->avg_no_rnd_pixels_tab[3] = avg_no_rnd_pixels16_xy; +} |