diff options
author | ZhouXiaoyong <zhouxiaoyong@loongson.cn> | 2015-07-09 14:34:21 +0800 |
---|---|---|
committer | Michael Niedermayer <michael@niedermayer.cc> | 2015-07-15 15:51:07 +0200 |
commit | 95418634b1e8ff5167952727c1f26ae4d4d6d797 (patch) | |
tree | ecca7483a3ad715807e6ce431c6da0a451487007 /libavcodec/mips/mpegvideo_mmi.c | |
parent | a0ad6c85aca95418739005d4fe3d68a550959435 (diff) | |
download | ffmpeg-95418634b1e8ff5167952727c1f26ae4d4d6d797.tar.gz |
avcodec: loongson optimized mpegvideo dct unquantize with mmi
Signed-off-by: ZhouXiaoyong <zhouxiaoyong@loongson.cn>
Signed-off-by: Michael Niedermayer <michael@niedermayer.cc>
Diffstat (limited to 'libavcodec/mips/mpegvideo_mmi.c')
-rw-r--r-- | libavcodec/mips/mpegvideo_mmi.c | 303 |
1 files changed, 303 insertions, 0 deletions
diff --git a/libavcodec/mips/mpegvideo_mmi.c b/libavcodec/mips/mpegvideo_mmi.c new file mode 100644 index 0000000000..b3d58bd32c --- /dev/null +++ b/libavcodec/mips/mpegvideo_mmi.c @@ -0,0 +1,303 @@ +/* + * Loongson SIMD optimized mpegvideo + * + * Copyright (c) 2015 Loongson Technology Corporation Limited + * Copyright (c) 2015 Zhou Xiaoyong <zhouxiaoyong@loongson.cn> + * Zhang Shuangshuang <zhangshuangshuang@ict.ac.cn> + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include "mpegvideo_mips.h" + +void ff_dct_unquantize_h263_intra_mmi(MpegEncContext *s, int16_t *block, + int n, int qscale) +{ + int64_t level, qmul, qadd, nCoeffs; + + qmul = qscale << 1; + assert(s->block_last_index[n]>=0 || s->h263_aic); + + if (!s->h263_aic) { + if (n<4) + level = block[0] * s->y_dc_scale; + else + level = block[0] * s->c_dc_scale; + qadd = (qscale-1) | 1; + } else { + qadd = 0; + level = block[0]; + } + + if(s->ac_pred) + nCoeffs = 63; + else + nCoeffs = s->inter_scantable.raster_end[s->block_last_index[n]]; + + __asm__ volatile ( + "xor $f12, $f12, $f12 \r\n" + "lwc1 $f12, %1 \n\r" + "xor $f10, $f10, $f10 \r\n" + "lwc1 $f10, %2 \r\n" + "xor $f14, $f14, $f14 \r\n" + "packsswh $f12, $f12, $f12 \r\n" + "packsswh $f12, $f12, $f12 \r\n" + "packsswh $f10, $f10, $f10 \r\n" + "packsswh $f10, $f10, $f10 \r\n" + "psubh $f14, $f14, $f10 \r\n" + "xor $f8, $f8, $f8 \r\n" + ".p2align 4 \r\n" + "1: \r\n" + "daddu $8, %0, %3 \r\n" + "gsldlc1 $f0, 7($8) \r\n" + "gsldrc1 $f0, 0($8) \r\n" + "gsldlc1 $f2, 15($8) \r\n" + "gsldrc1 $f2, 8($8) \r\n" + "mov.d $f4, $f0 \r\n" + "mov.d $f6, $f2 \r\n" + "pmullh $f0, $f0, $f12 \r\n" + "pmullh $f2, $f2, $f12 \r\n" + "pcmpgth $f4, $f4, $f8 \r\n" + "pcmpgth $f6, $f6, $f8 \r\n" + "xor $f0, $f0, $f4 \r\n" + "xor $f2, $f2, $f6 \r\n" + "paddh $f0, $f0, $f14 \r\n" + "paddh $f2, $f2, $f14 \r\n" + "xor $f4, $f4, $f0 \r\n" + "xor $f6, $f6, $f2 \r\n" + "pcmpeqh $f0, $f0, $f14 \r\n" + "pcmpeqh $f2, $f2, $f14 \r\n" + "pandn $f0, $f0, $f4 \r\n" + "pandn $f2, $f2, $f6 \r\n" + "gssdlc1 $f0, 7($8) \r\n" + "gssdrc1 $f0, 0($8) \r\n" + "gssdlc1 $f2, 15($8) \r\n" + "gssdrc1 $f2, 8($8) \r\n" + "addi %3, %3, 16 \r\n" + "blez %3, 1b \r\n" + ::"r"(block+nCoeffs),"m"(qmul),"m"(qadd),"r"(2*(-nCoeffs)) + :"$8","memory" + ); + + block[0] = level; +} + +void ff_dct_unquantize_h263_inter_mmi(MpegEncContext *s, int16_t *block, + int n, int qscale) +{ + int64_t qmul, qadd, nCoeffs; + + qmul = qscale << 1; + qadd = (qscale - 1) | 1; + assert(s->block_last_index[n]>=0 || s->h263_aic); + nCoeffs = s->inter_scantable.raster_end[s->block_last_index[n]]; + + __asm__ volatile ( + "xor $f12, $f12, $f12 \r\n" + "lwc1 $f12, %1 \r\n" + "xor $f10, $f10, $f10 \r\n" + "lwc1 $f10, %2 \r\n" + "packsswh $f12, $f12, $f12 \r\n" + "packsswh $f12, $f12, $f12 \r\n" + "xor $f14, $f14, $f14 \r\n" + "packsswh $f10, $f10, $f10 \r\n" + "packsswh $f10, $f10, $f10 \r\n" + "psubh $f14, $f14, $f10 \r\n" + "xor $f8, $f8, $f8 \r\n" + ".p2align 4 \r\n" + "1: \r\n" + "daddu $8, %0, %3 \r\n" + "gsldlc1 $f0, 7($8) \r\n" + "gsldrc1 $f0, 0($8) \r\n" + "gsldlc1 $f2, 15($8) \r\n" + "gsldrc1 $f2, 8($8) \r\n" + "mov.d $f4, $f0 \r\n" + "mov.d $f6, $f2 \r\n" + "pmullh $f0, $f0, $f12 \r\n" + "pmullh $f2, $f2, $f12 \r\n" + "pcmpgth $f4, $f4, $f8 \r\n" + "pcmpgth $f6, $f6, $f8 \r\n" + "xor $f0, $f0, $f4 \r\n" + "xor $f2, $f2, $f6 \r\n" + "paddh $f0, $f0, $f14 \r\n" + "paddh $f2, $f2, $f14 \r\n" + "xor $f4, $f4, $f0 \r\n" + "xor $f6, $f6, $f2 \r\n" + "pcmpeqh $f0, $f0, $f14 \r\n" + "pcmpeqh $f2, $f2, $f14 \r\n" + "pandn $f0, $f0, $f4 \r\n" + "pandn $f2, $f2, $f6 \r\n" + "gssdlc1 $f0, 7($8) \r\n" + "gssdrc1 $f0, 0($8) \r\n" + "gssdlc1 $f2, 15($8) \r\n" + "gssdrc1 $f2, 8($8) \r\n" + "addi %3, %3, 16 \r\n" + "blez %3, 1b \r\n" + ::"r"(block+nCoeffs),"m"(qmul),"m"(qadd),"r"(2*(-nCoeffs)) + : "$8","memory" + ); +} + +void ff_dct_unquantize_mpeg1_intra_mmi(MpegEncContext *s, int16_t *block, + int n, int qscale) +{ + int64_t nCoeffs; + const uint16_t *quant_matrix; + int block0; + + assert(s->block_last_index[n]>=0); + nCoeffs = s->intra_scantable.raster_end[s->block_last_index[n]] + 1; + + if (n<4) + block0 = block[0] * s->y_dc_scale; + else + block0 = block[0] * s->c_dc_scale; + + /* XXX: only mpeg1 */ + quant_matrix = s->intra_matrix; + + __asm__ volatile ( + "pcmpeqh $f14, $f14, $f14 \r\n" + "dli $10, 15 \r\n" + "dmtc1 $10, $f16 \r\n" + "xor $f12, $f12, $f12 \r\n" + "lwc1 $f12, %2 \r\n" + "psrlh $f14, $f14, $f16 \r\n" + "packsswh $f12, $f12, $f12 \r\n" + "packsswh $f12, $f12, $f12 \r\n" + "or $8, %3, $0 \r\n" + ".p2align 4 \r\n" + "1: \r\n" + "gsldxc1 $f0, 0($8, %0) \r\n" + "gsldxc1 $f2, 8($8, %0) \r\n" + "mov.d $f16, $f0 \r\n" + "mov.d $f18, $f2 \r\n" + "gsldxc1 $f8, 0($8, %1) \r\n" + "gsldxc1 $f10, 8($8, %1) \r\n" + "pmullh $f8, $f8, $f12 \r\n" + "pmullh $f10, $f10, $f12 \r\n" + "xor $f4, $f4, $f4 \r\n" + "xor $f6, $f6, $f6 \r\n" + "pcmpgth $f4, $f4, $f0 \r\n" + "pcmpgth $f6, $f6, $f2 \r\n" + "xor $f0, $f0, $f4 \r\n" + "xor $f2, $f2, $f6 \r\n" + "psubh $f0, $f0, $f4 \r\n" + "psubh $f2, $f2, $f6 \r\n" + "pmullh $f0, $f0, $f8 \r\n" + "pmullh $f2, $f2, $f10 \r\n" + "xor $f8, $f8, $f8 \r\n" + "xor $f10, $f10, $f10 \r\n" + "pcmpeqh $f8, $f8, $f16 \r\n" + "pcmpeqh $f10, $f10, $f18 \r\n" + "dli $10, 3 \r\n" + "dmtc1 $10, $f16 \r\n" + "psrah $f0, $f0, $f16 \r\n" + "psrah $f2, $f2, $f16 \r\n" + "psubh $f0, $f0, $f14 \r\n" + "psubh $f2, $f2, $f14 \r\n" + "or $f0, $f0, $f14 \r\n" + "or $f2, $f2, $f14 \r\n" + "xor $f0, $f0, $f4 \r\n" + "xor $f2, $f2, $f6 \r\n" + "psubh $f0, $f0, $f4 \r\n" + "psubh $f2, $f2, $f6 \r\n" + "pandn $f8, $f8, $f0 \r\n" + "pandn $f10, $f10, $f2 \r\n" + "gssdxc1 $f8, 0($8, %0) \r\n" + "gssdxc1 $f10, 8($8, %0) \r\n" + "addi $8, $8, 16 \r\n" + "bltz $8, 1b \r\n" + ::"r"(block+nCoeffs),"r"(quant_matrix+nCoeffs),"m"(qscale), + "g"(-2*nCoeffs) + : "$8","$10","memory" + ); + + block[0] = block0; +} + +void ff_dct_unquantize_mpeg1_inter_mmi(MpegEncContext *s, int16_t *block, + int n, int qscale) +{ + int64_t nCoeffs; + const uint16_t *quant_matrix; + + assert(s->block_last_index[n] >= 0); + nCoeffs = s->intra_scantable.raster_end[s->block_last_index[n]] + 1; + quant_matrix = s->inter_matrix; + + __asm__ volatile ( + "pcmpeqh $f14, $f14, $f14 \r\n" + "dli $10, 15 \r\n" + "dmtc1 $10, $f16 \r\n" + "xor $f12, $f12, $f12 \r\n" + "lwc1 $f12, %2 \r\n" + "psrlh $f14, $f14, $f16 \r\n" + "packsswh $f12, $f12, $f12 \r\n" + "packsswh $f12, $f12, $f12 \r\n" + "or $8, %3, $0 \r\n" + ".p2align 4 \r\n" + "1: \r\n" + "gsldxc1 $f0, 0($8, %0) \r\n" + "gsldxc1 $f2, 8($8, %0) \r\n" + "mov.d $f16, $f0 \r\n" + "mov.d $f18, $f2 \r\n" + "gsldxc1 $f8, 0($8, %1) \r\n" + "gsldxc1 $f10, 8($8, %1) \r\n" + "pmullh $f8, $f8, $f12 \r\n" + "pmullh $f10, $f10, $f12 \r\n" + "xor $f4, $f4, $f4 \r\n" + "xor $f6, $f6, $f6 \r\n" + "pcmpgth $f4, $f4, $f0 \r\n" + "pcmpgth $f6, $f6, $f2 \r\n" + "xor $f0, $f0, $f4 \r\n" + "xor $f2, $f2, $f6 \r\n" + "psubh $f0, $f0, $f4 \r\n" + "psubh $f2, $f2, $f6 \r\n" + "paddh $f0, $f0, $f0 \r\n" + "paddh $f2, $f2, $f2 \r\n" + "paddh $f0, $f0, $f14 \r\n" + "paddh $f2, $f2, $f14 \r\n" + "pmullh $f0, $f0, $f8 \r\n" + "pmullh $f2, $f2, $f10 \r\n" + "xor $f8, $f8, $f8 \r\n" + "xor $f10, $f10, $f10 \r\n" + "pcmpeqh $f8, $f8, $f16 \r\n" + "pcmpeqh $f10, $f10, $f18 \r\n" + "dli $10, 4 \r\n" + "dmtc1 $10, $f16 \r\n" + "psrah $f0, $f0, $f16 \r\n" + "psrah $f2, $f2, $f16 \r\n" + "psubh $f0, $f0, $f14 \r\n" + "psubh $f2, $f2, $f14 \r\n" + "or $f0, $f0, $f14 \r\n" + "or $f2, $f2, $f14 \r\n" + "xor $f0, $f0, $f4 \r\n" + "xor $f2, $f2, $f6 \r\n" + "psubh $f0, $f0, $f4 \r\n" + "psubh $f2, $f2, $f6 \r\n" + "pandn $f8, $f8, $f0 \r\n" + "pandn $f10, $f10, $f2 \r\n" + "gssdxc1 $f8, 0($8, %0) \r\n" + "gssdxc1 $f10, 8($8, %0) \r\n" + "addi $8, $8, 16 \r\n" + "bltz $8, 1b \r\n" + ::"r"(block+nCoeffs),"r"(quant_matrix+nCoeffs),"m"(qscale), + "g"(-2*nCoeffs) + :"$8","$10","memory" + ); +} |