/* * Alpha optimized DSP utils * Copyright (c) 2002 Falk Hueffner <falk@debian.org> * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #ifndef FFMPEG_ASM_H #define FFMPEG_ASM_H #include <inttypes.h> #if defined __GNUC__ # define GNUC_PREREQ(maj, min) \ ((__GNUC__ << 16) + __GNUC_MINOR__ >= ((maj) << 16) + (min)) #else # define GNUC_PREREQ(maj, min) 0 #endif #if GNUC_PREREQ(2,96) # define likely(x) __builtin_expect((x) != 0, 1) # define unlikely(x) __builtin_expect((x) != 0, 0) #else # define likely(x) (x) # define unlikely(x) (x) #endif #define AMASK_BWX (1 << 0) #define AMASK_FIX (1 << 1) #define AMASK_CIX (1 << 2) #define AMASK_MVI (1 << 8) static inline uint64_t BYTE_VEC(uint64_t x) { x |= x << 8; x |= x << 16; x |= x << 32; return x; } static inline uint64_t WORD_VEC(uint64_t x) { x |= x << 16; x |= x << 32; return x; } #define sextw(x) ((int16_t) (x)) #ifdef __GNUC__ #define ldq(p) \ (((union { \ uint64_t __l; \ __typeof__(*(p)) __s[sizeof (uint64_t) / sizeof *(p)]; \ } *) (p))->__l) #define ldl(p) \ (((union { \ int32_t __l; \ __typeof__(*(p)) __s[sizeof (int32_t) / sizeof *(p)]; \ } *) (p))->__l) #define stq(l, p) \ do { \ (((union { \ uint64_t __l; \ __typeof__(*(p)) __s[sizeof (uint64_t) / sizeof *(p)]; \ } *) (p))->__l) = l; \ } while (0) #define stl(l, p) \ do { \ (((union { \ int32_t __l; \ __typeof__(*(p)) __s[sizeof (int32_t) / sizeof *(p)]; \ } *) (p))->__l) = l; \ } while (0) struct unaligned_long { uint64_t l; } __attribute__((packed)); #define ldq_u(p) (*(const uint64_t *) (((uint64_t) (p)) & ~7ul)) #define uldq(a) (((const struct unaligned_long *) (a))->l) #if GNUC_PREREQ(3,3) #define prefetch(p) __builtin_prefetch((p), 0, 1) #define prefetch_en(p) __builtin_prefetch((p), 0, 0) #define prefetch_m(p) __builtin_prefetch((p), 1, 1) #define prefetch_men(p) __builtin_prefetch((p), 1, 0) #define cmpbge __builtin_alpha_cmpbge /* Avoid warnings. */ #define extql(a, b) __builtin_alpha_extql(a, (uint64_t) (b)) #define extwl(a, b) __builtin_alpha_extwl(a, (uint64_t) (b)) #define extqh(a, b) __builtin_alpha_extqh(a, (uint64_t) (b)) #define zap __builtin_alpha_zap #define zapnot __builtin_alpha_zapnot #define amask __builtin_alpha_amask #define implver __builtin_alpha_implver #define rpcc __builtin_alpha_rpcc #else #define prefetch(p) asm volatile("ldl $31,%0" : : "m"(*(const char *) (p)) : "memory") #define prefetch_en(p) asm volatile("ldq $31,%0" : : "m"(*(const char *) (p)) : "memory") #define prefetch_m(p) asm volatile("lds $f31,%0" : : "m"(*(const char *) (p)) : "memory") #define prefetch_men(p) asm volatile("ldt $f31,%0" : : "m"(*(const char *) (p)) : "memory") #define cmpbge(a, b) ({ uint64_t __r; asm ("cmpbge %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define extql(a, b) ({ uint64_t __r; asm ("extql %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define extwl(a, b) ({ uint64_t __r; asm ("extwl %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define extqh(a, b) ({ uint64_t __r; asm ("extqh %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define zap(a, b) ({ uint64_t __r; asm ("zap %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define zapnot(a, b) ({ uint64_t __r; asm ("zapnot %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define amask(a) ({ uint64_t __r; asm ("amask %1,%0" : "=r" (__r) : "rI" (a)); __r; }) #define implver() ({ uint64_t __r; asm ("implver %0" : "=r" (__r)); __r; }) #define rpcc() ({ uint64_t __r; asm volatile ("rpcc %0" : "=r" (__r)); __r; }) #endif #define wh64(p) asm volatile("wh64 (%0)" : : "r"(p) : "memory") #if GNUC_PREREQ(3,3) && defined(__alpha_max__) #define minub8 __builtin_alpha_minub8 #define minsb8 __builtin_alpha_minsb8 #define minuw4 __builtin_alpha_minuw4 #define minsw4 __builtin_alpha_minsw4 #define maxub8 __builtin_alpha_maxub8 #define maxsb8 __builtin_alpha_maxsb8 #define maxuw4 __builtin_alpha_maxuw4 #define maxsw4 __builtin_alpha_maxsw4 #define perr __builtin_alpha_perr #define pklb __builtin_alpha_pklb #define pkwb __builtin_alpha_pkwb #define unpkbl __builtin_alpha_unpkbl #define unpkbw __builtin_alpha_unpkbw #else #define minub8(a, b) ({ uint64_t __r; asm (".arch ev6; minub8 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define minsb8(a, b) ({ uint64_t __r; asm (".arch ev6; minsb8 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define minuw4(a, b) ({ uint64_t __r; asm (".arch ev6; minuw4 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define minsw4(a, b) ({ uint64_t __r; asm (".arch ev6; minsw4 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define maxub8(a, b) ({ uint64_t __r; asm (".arch ev6; maxub8 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define maxsb8(a, b) ({ uint64_t __r; asm (".arch ev6; maxsb8 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define maxuw4(a, b) ({ uint64_t __r; asm (".arch ev6; maxuw4 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define maxsw4(a, b) ({ uint64_t __r; asm (".arch ev6; maxsw4 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define perr(a, b) ({ uint64_t __r; asm (".arch ev6; perr %r1,%r2,%0" : "=r" (__r) : "%rJ" (a), "rJ" (b)); __r; }) #define pklb(a) ({ uint64_t __r; asm (".arch ev6; pklb %r1,%0" : "=r" (__r) : "rJ" (a)); __r; }) #define pkwb(a) ({ uint64_t __r; asm (".arch ev6; pkwb %r1,%0" : "=r" (__r) : "rJ" (a)); __r; }) #define unpkbl(a) ({ uint64_t __r; asm (".arch ev6; unpkbl %r1,%0" : "=r" (__r) : "rJ" (a)); __r; }) #define unpkbw(a) ({ uint64_t __r; asm (".arch ev6; unpkbw %r1,%0" : "=r" (__r) : "rJ" (a)); __r; }) #endif #elif defined(__DECC) /* Digital/Compaq/hp "ccc" compiler */ #include <c_asm.h> #define ldq(p) (*(const uint64_t *) (p)) #define ldl(p) (*(const int32_t *) (p)) #define stq(l, p) do { *(uint64_t *) (p) = (l); } while (0) #define stl(l, p) do { *(int32_t *) (p) = (l); } while (0) #define ldq_u(a) asm ("ldq_u %v0,0(%a0)", a) #define uldq(a) (*(const __unaligned uint64_t *) (a)) #define cmpbge(a, b) asm ("cmpbge %a0,%a1,%v0", a, b) #define extql(a, b) asm ("extql %a0,%a1,%v0", a, b) #define extwl(a, b) asm ("extwl %a0,%a1,%v0", a, b) #define extqh(a, b) asm ("extqh %a0,%a1,%v0", a, b) #define zap(a, b) asm ("zap %a0,%a1,%v0", a, b) #define zapnot(a, b) asm ("zapnot %a0,%a1,%v0", a, b) #define amask(a) asm ("amask %a0,%v0", a) #define implver() asm ("implver %v0") #define rpcc() asm ("rpcc %v0") #define minub8(a, b) asm ("minub8 %a0,%a1,%v0", a, b) #define minsb8(a, b) asm ("minsb8 %a0,%a1,%v0", a, b) #define minuw4(a, b) asm ("minuw4 %a0,%a1,%v0", a, b) #define minsw4(a, b) asm ("minsw4 %a0,%a1,%v0", a, b) #define maxub8(a, b) asm ("maxub8 %a0,%a1,%v0", a, b) #define maxsb8(a, b) asm ("maxsb8 %a0,%a1,%v0", a, b) #define maxuw4(a, b) asm ("maxuw4 %a0,%a1,%v0", a, b) #define maxsw4(a, b) asm ("maxsw4 %a0,%a1,%v0", a, b) #define perr(a, b) asm ("perr %a0,%a1,%v0", a, b) #define pklb(a) asm ("pklb %a0,%v0", a) #define pkwb(a) asm ("pkwb %a0,%v0", a) #define unpkbl(a) asm ("unpkbl %a0,%v0", a) #define unpkbw(a) asm ("unpkbw %a0,%v0", a) #define wh64(a) asm ("wh64 %a0", a) #else #error "Unknown compiler!" #endif #endif /* FFMPEG_ASM_H */