diff options
author | Michael Niedermayer <michaelni@gmx.at> | 2014-03-14 01:01:37 +0100 |
---|---|---|
committer | Michael Niedermayer <michaelni@gmx.at> | 2014-03-14 01:01:37 +0100 |
commit | d926c4b240edd86d9ee5c1b449968ac91eeaa918 (patch) | |
tree | ba72934d12b504eb23a9cd6e79c7588dd3dcc74f | |
parent | 55f53f6c29acd459df64abf20cbd016c420b6cf6 (diff) | |
parent | 38675229a879aa5258a8c71891fc8cbf74cf139f (diff) | |
download | ffmpeg-d926c4b240edd86d9ee5c1b449968ac91eeaa918.tar.gz |
Merge commit '38675229a879aa5258a8c71891fc8cbf74cf139f'
* commit '38675229a879aa5258a8c71891fc8cbf74cf139f':
dsputil_mmx: K&R formatting cosmetics
Conflicts:
libavcodec/x86/dsputil_mmx.c
Merged-by: Michael Niedermayer <michaelni@gmx.at>
-rw-r--r-- | libavcodec/x86/dsputil_mmx.c | 107 |
1 files changed, 50 insertions, 57 deletions
diff --git a/libavcodec/x86/dsputil_mmx.c b/libavcodec/x86/dsputil_mmx.c index df8cfdb464..47835acf2b 100644 --- a/libavcodec/x86/dsputil_mmx.c +++ b/libavcodec/x86/dsputil_mmx.c @@ -60,8 +60,8 @@ void ff_put_pixels_clamped_mmx(const int16_t *block, uint8_t *pixels, "movq %%mm2, (%0, %1) \n\t" "movq %%mm4, (%0, %1, 2) \n\t" "movq %%mm6, (%0, %2) \n\t" - :: "r"(pix), "r"((x86_reg)line_size), "r"((x86_reg)line_size * 3), - "r"(p) + :: "r" (pix), "r" ((x86_reg) line_size), "r" ((x86_reg) line_size * 3), + "r" (p) : "memory"); pix += line_size * 4; p += 32; @@ -86,7 +86,8 @@ void ff_put_pixels_clamped_mmx(const int16_t *block, uint8_t *pixels, "movq %%mm2, (%0, %1) \n\t" "movq %%mm4, (%0, %1, 2) \n\t" "movq %%mm6, (%0, %2) \n\t" - :: "r"(pix), "r"((x86_reg)line_size), "r"((x86_reg)line_size * 3), "r"(p) + :: "r" (pix), "r" ((x86_reg) line_size), "r" ((x86_reg) line_size * 3), + "r" (p) : "memory"); } @@ -120,8 +121,8 @@ void ff_put_signed_pixels_clamped_mmx(const int16_t *block, uint8_t *pixels, put_signed_pixels_clamped_mmx_half(0) "lea (%0, %3, 4), %0 \n\t" put_signed_pixels_clamped_mmx_half(64) - : "+&r"(pixels), "=&r"(line_skip3) - : "r"(block), "r"(line_skip) + : "+&r" (pixels), "=&r" (line_skip3) + : "r" (block), "r" (line_skip) : "memory"); } @@ -159,8 +160,8 @@ void ff_add_pixels_clamped_mmx(const int16_t *block, uint8_t *pixels, "packuswb %%mm3, %%mm2 \n\t" "movq %%mm0, %0 \n\t" "movq %%mm2, %1 \n\t" - : "+m"(*pix), "+m"(*(pix + line_size)) - : "r"(p) + : "+m" (*pix), "+m" (*(pix + line_size)) + : "r" (p) : "memory"); pix += line_size * 2; p += 16; @@ -180,10 +181,9 @@ void name(int16_t *blocks) \ "movq %%mm7, 24(%0, %%"REG_a") \n\t" \ "add $32, %%"REG_a" \n\t" \ "js 1b \n\t" \ - :: "r"(((uint8_t *)blocks) + 128 * n), \ + :: "r"(((uint8_t *) blocks) + 128 * n), \ "i"(-128 * n) \ - : "%"REG_a \ - ); \ + : "%"REG_a); \ } CLEAR_BLOCKS(ff_clear_blocks_mmx, 6) CLEAR_BLOCKS(ff_clear_block_mmx, 1) @@ -200,9 +200,8 @@ void ff_clear_block_sse(int16_t *block) "movaps %%xmm0, 80(%0) \n" "movaps %%xmm0, 96(%0) \n" "movaps %%xmm0, 112(%0) \n" - :: "r"(block) - : "memory" - ); + :: "r" (block) + : "memory"); } void ff_clear_blocks_sse(int16_t *blocks) @@ -221,15 +220,14 @@ void ff_clear_blocks_sse(int16_t *blocks) "movaps %%xmm0, 112(%0, %%"REG_a") \n" "add $128, %%"REG_a" \n" "js 1b \n" - :: "r"(((uint8_t *)blocks) + 128 * 6), - "i"(-128 * 6) - : "%"REG_a - ); + :: "r"(((uint8_t *) blocks) + 128 * 6), "i"(-128 * 6) + : "%"REG_a); } void ff_add_bytes_mmx(uint8_t *dst, uint8_t *src, int w) { x86_reg i = 0; + __asm__ volatile ( "jmp 2f \n\t" "1: \n\t" @@ -245,10 +243,10 @@ void ff_add_bytes_mmx(uint8_t *dst, uint8_t *src, int w) "2: \n\t" "cmp %3, %0 \n\t" "js 1b \n\t" - : "+r"(i) - : "r"(src), "r"(dst), "r"((x86_reg)w - 15) - ); - for ( ; i < w; i++) + : "+r" (i) + : "r" (src), "r" (dst), "r" ((x86_reg) w - 15)); + + for (; i < w; i++) dst[i + 0] += src[i + 0]; } @@ -279,10 +277,10 @@ void ff_draw_edges_mmx(uint8_t *buf, int wrap, int width, int height, "add %1, %0 \n\t" "cmp %3, %0 \n\t" "jb 1b \n\t" - : "+r"(ptr) - : "r"((x86_reg)wrap), "r"((x86_reg)width), "r"(ptr + wrap * height) - ); - } else if(w==16){ + : "+r" (ptr) + : "r" ((x86_reg) wrap), "r" ((x86_reg) width), + "r" (ptr + wrap * height)); + } else if (w == 16) { __asm__ volatile ( "1: \n\t" "movd (%0), %%mm0 \n\t" @@ -319,9 +317,9 @@ void ff_draw_edges_mmx(uint8_t *buf, int wrap, int width, int height, "add %1, %0 \n\t" "cmp %3, %0 \n\t" "jb 1b \n\t" - : "+r"(ptr) - : "r"((x86_reg)wrap), "r"((x86_reg)width), "r"(ptr + wrap * height) - ); + : "+r" (ptr) + : "r" ((x86_reg) wrap), "r" ((x86_reg) width), + "r" (ptr + wrap * height)); } /* top and bottom (and hopefully also the corners) */ @@ -338,10 +336,10 @@ void ff_draw_edges_mmx(uint8_t *buf, int wrap, int width, int height, "add $8, %0 \n\t" "cmp %4, %0 \n\t" "jb 1b \n\t" - : "+r"(ptr) - : "r"((x86_reg)buf - (x86_reg)ptr - w), "r"((x86_reg) -wrap), - "r"((x86_reg) -wrap * 3), "r"(ptr + width + 2 * w) - ); + : "+r" (ptr) + : "r" ((x86_reg) buf - (x86_reg) ptr - w), + "r" ((x86_reg) - wrap), "r" ((x86_reg) - wrap * 3), + "r" (ptr + width + 2 * w)); } } @@ -358,11 +356,10 @@ void ff_draw_edges_mmx(uint8_t *buf, int wrap, int width, int height, "add $8, %0 \n\t" "cmp %4, %0 \n\t" "jb 1b \n\t" - : "+r"(ptr) - : "r"((x86_reg)last_line - (x86_reg)ptr - w), - "r"((x86_reg)wrap), "r"((x86_reg)wrap * 3), - "r"(ptr + width + 2 * w) - ); + : "+r" (ptr) + : "r" ((x86_reg) last_line - (x86_reg) ptr - w), + "r" ((x86_reg) wrap), "r" ((x86_reg) wrap * 3), + "r" (ptr + width + 2 * w)); } } } @@ -391,7 +388,7 @@ static av_always_inline void gmc(uint8_t *dst, uint8_t *src, const uint16_t r4[4] = { r, r, r, r }; const uint16_t dxy4[4] = { dxys, dxys, dxys, dxys }; const uint16_t dyy4[4] = { dyys, dyys, dyys, dyys }; - const uint64_t shift2 = 2 * shift; + const uint64_t shift2 = 2 * shift; #define MAX_STRIDE 4096U #define MAX_H 8U uint8_t edge_buf[(MAX_H + 1) * MAX_STRIDE]; @@ -401,15 +398,15 @@ static av_always_inline void gmc(uint8_t *dst, uint8_t *src, const int dyh = (dyy - (1 << (16 + shift))) * (h - 1); const int dxh = dxy * (h - 1); const int dyw = dyx * (w - 1); - int need_emu = (unsigned)ix >= width - w || - (unsigned)iy >= height - h; + int need_emu = (unsigned) ix >= width - w || + (unsigned) iy >= height - h; if ( // non-constant fullpel offset (3% of blocks) ((ox ^ (ox + dxw)) | (ox ^ (ox + dxh)) | (ox ^ (ox + dxw + dxh)) | - (oy ^ (oy + dyw)) | (oy ^ (oy + dyh)) | (oy ^ (oy + dyw + dyh))) >> (16 + shift) + (oy ^ (oy + dyw)) | (oy ^ (oy + dyh)) | (oy ^ (oy + dyw + dyh))) >> (16 + shift) || // uses more than 16 bits of subpel mv (only at huge resolution) - || (dxx | dxy | dyx | dyy) & 15 - || (need_emu && (h > MAX_H || stride > MAX_STRIDE))) { + (dxx | dxy | dyx | dyy) & 15 || + (need_emu && (h > MAX_H || stride > MAX_STRIDE))) { // FIXME could still use mmx for some of the rows ff_gmc_c(dst, src, stride, h, ox, oy, dxx, dxy, dyx, dyy, shift, r, width, height); @@ -427,8 +424,7 @@ static av_always_inline void gmc(uint8_t *dst, uint8_t *src, "pxor %%mm7, %%mm7 \n\t" "punpcklwd %%mm6, %%mm6 \n\t" "punpcklwd %%mm6, %%mm6 \n\t" - :: "r"(1<<shift) - ); + :: "r" (1 << shift)); for (x = 0; x < w; x += 4) { uint16_t dx4[4] = { oxs - dxys + dxxs * (x + 0), @@ -450,9 +446,8 @@ static av_always_inline void gmc(uint8_t *dst, uint8_t *src, "movq %%mm5, %1 \n\t" "psrlw $12, %%mm4 \n\t" "psrlw $12, %%mm5 \n\t" - : "+m"(*dx4), "+m"(*dy4) - : "m"(*dxy4), "m"(*dyy4) - ); + : "+m" (*dx4), "+m" (*dy4) + : "m" (*dxy4), "m" (*dyy4)); __asm__ volatile ( "movq %%mm6, %%mm2 \n\t" @@ -488,11 +483,10 @@ static av_always_inline void gmc(uint8_t *dst, uint8_t *src, "packuswb %%mm0, %%mm0 \n\t" "movd %%mm0, %0 \n\t" - : "=m"(dst[x + y * stride]) - : "m"(src[0]), "m"(src[1]), - "m"(src[stride]), "m"(src[stride + 1]), - "m"(*r4), "m"(shift2) - ); + : "=m" (dst[x + y * stride]) + : "m" (src[0]), "m" (src[1]), + "m" (src[stride]), "m" (src[stride + 1]), + "m" (*r4), "m" (shift2)); src += stride; } src += 4 - h * stride; @@ -629,10 +623,9 @@ void ff_vector_clipf_sse(float *dst, const float *src, "movaps %%xmm3, 48(%1, %0) \n\t" "sub $64, %0 \n\t" "jge 1b \n\t" - : "+&r"(i) - : "r"(dst), "r"(src), "m"(min), "m"(max) - : "memory" - ); + : "+&r" (i) + : "r" (dst), "r" (src), "m" (min), "m" (max) + : "memory"); } #endif /* HAVE_INLINE_ASM */ |