CVS: main/libmpeg2 alloc.c,NONE,1.1 alpha_asm.h,NONE,1.1 convert.h,NONE,1.1 cpu_accel.c,NONE,1.1 cpu_state.c,NONE,1.1 decode.c,1.23,1.24 idct_alpha.c,NONE,1.1 idct_altivec.c,NONE,1.1 motion_comp_alpha.c,NONE,1.1 motion_comp_altivec.c,NONE,1.1
Update of /cvsroot/mplayer/main/libmpeg2 In directory mail:/var/tmp.root/cvs-serv12821 Added Files: alloc.c alpha_asm.h convert.h cpu_accel.c cpu_state.c decode.c idct_alpha.c idct_altivec.c motion_comp_alpha.c motion_comp_altivec.c Log Message: Importing libmpeg2 from mpeg2dec-0.3.1 --- NEW FILE --- /* * alloc.c * Copyright (C) 2000-2002 Michel Lespinasse <walken@zoy.org> * Copyright (C) 1999-2000 Aaron Holtzman <aholtzma@ess.engr.uvic.ca> * * This file is part of mpeg2dec, a free MPEG-2 video stream decoder. * See http://libmpeg2.sourceforge.net/ for updates. * * mpeg2dec is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * mpeg2dec is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "config.h" #include <stdlib.h> #include <inttypes.h> #include "mpeg2.h" #include "mpeg2_internal.h" #if defined(HAVE_MEMALIGN) && !defined(__cplusplus) /* some systems have memalign() but no declaration for it */ void * memalign (size_t align, size_t size); #endif void * (* mpeg2_malloc_hook) (int size, int reason) = NULL; int (* mpeg2_free_hook) (void * buf) = NULL; void * mpeg2_malloc (int size, int reason) { char * buf; if (mpeg2_malloc_hook) { buf = (char *) mpeg2_malloc_hook (size, reason); if (buf) return buf; } #if defined(HAVE_MEMALIGN) && !defined(__cplusplus) && !defined(DEBUG) return memalign (16, size); #else buf = (char *) malloc (size + 15 + sizeof (void **)); if (buf) { char * align_buf; align_buf = buf + 15 + sizeof (void **); align_buf -= (long)align_buf & 15; *(((void **)align_buf) - 1) = buf; return align_buf; } return NULL; #endif } void mpeg2_free (void * buf) { if (mpeg2_free_hook && mpeg2_free_hook (buf)) return; #if defined(HAVE_MEMALIGN) && !defined(__cplusplus) && !defined(DEBUG) free (buf); #else free (*(((void **)buf) - 1)); #endif } --- NEW FILE --- /* * Alpha assembly macros * Copyright (c) 2002 Falk Hueffner <falk@debian.org> * * This file is part of mpeg2dec, a free MPEG-2 video stream decoder. * See http://libmpeg2.sourceforge.net/ for updates. * * mpeg2dec is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * mpeg2dec is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #ifndef ALPHA_ASM_H #define ALPHA_ASM_H #include <inttypes.h> #if defined __GNUC__ # define GNUC_PREREQ(maj, min) \ ((__GNUC__ << 16) + __GNUC_MINOR__ >= ((maj) << 16) + (min)) #else # define GNUC_PREREQ(maj, min) 0 #endif #define AMASK_BWX (1 << 0) #define AMASK_FIX (1 << 1) #define AMASK_CIX (1 << 2) #define AMASK_MVI (1 << 8) #ifdef __alpha_bwx__ # define HAVE_BWX() 1 #else # define HAVE_BWX() (amask(AMASK_BWX) == 0) #endif #ifdef __alpha_fix__ # define HAVE_FIX() 1 #else # define HAVE_FIX() (amask(AMASK_FIX) == 0) #endif #ifdef __alpha_max__ # define HAVE_MVI() 1 #else # define HAVE_MVI() (amask(AMASK_MVI) == 0) #endif #ifdef __alpha_cix__ # define HAVE_CIX() 1 #else # define HAVE_CIX() (amask(AMASK_CIX) == 0) #endif inline static uint64_t BYTE_VEC(uint64_t x) { x |= x << 8; x |= x << 16; x |= x << 32; return x; } inline static uint64_t WORD_VEC(uint64_t x) { x |= x << 16; x |= x << 32; return x; } #define ldq(p) (*(const uint64_t *) (p)) #define ldl(p) (*(const int32_t *) (p)) #define stl(l, p) do { *(uint32_t *) (p) = (l); } while (0) #define stq(l, p) do { *(uint64_t *) (p) = (l); } while (0) #define sextw(x) ((int16_t) (x)) #ifdef __GNUC__ struct unaligned_long { uint64_t l; } __attribute__((packed)); #define ldq_u(p) (*(const uint64_t *) (((uint64_t) (p)) & ~7ul)) #define uldq(a) (((const struct unaligned_long *) (a))->l) #if GNUC_PREREQ(3,0) /* Unfortunately, __builtin_prefetch is slightly buggy on Alpha. The defines here are kludged so we still get the right instruction. This needs to be adapted as soon as gcc is fixed. */ # define prefetch(p) __builtin_prefetch((p), 0, 1) # define prefetch_en(p) __builtin_prefetch((p), 1, 1) # define prefetch_m(p) __builtin_prefetch((p), 0, 0) # define prefetch_men(p) __builtin_prefetch((p), 1, 0) #else # define prefetch(p) asm volatile("ldl $31,%0" : : "m"(*(const char *) (p)) : "memory") # define prefetch_en(p) asm volatile("ldq $31,%0" : : "m"(*(const char *) (p)) : "memory") # define prefetch_m(p) asm volatile("lds $f31,%0" : : "m"(*(const char *) (p)) : "memory") # define prefetch_men(p) asm volatile("ldt $f31,%0" : : "m"(*(const char *) (p)) : "memory") #endif #if GNUC_PREREQ(3,3) #define cmpbge __builtin_alpha_cmpbge /* Avoid warnings. */ #define extql(a, b) __builtin_alpha_extql(a, (uint64_t) (b)) #define extwl(a, b) __builtin_alpha_extwl(a, (uint64_t) (b)) #define extqh(a, b) __builtin_alpha_extqh(a, (uint64_t) (b)) #define zap __builtin_alpha_zap #define zapnot __builtin_alpha_zapnot #define amask __builtin_alpha_amask #define implver __builtin_alpha_implver #define rpcc __builtin_alpha_rpcc #define minub8 __builtin_alpha_minub8 #define minsb8 __builtin_alpha_minsb8 #define minuw4 __builtin_alpha_minuw4 #define minsw4 __builtin_alpha_minsw4 #define maxub8 __builtin_alpha_maxub8 #define maxsb8 __builtin_alpha_maxsb8 #define maxuw4 __builtin_alpha_maxuw4 #define maxsw4 __builtin_alpha_maxsw4 #define perr __builtin_alpha_perr #define pklb __builtin_alpha_pklb #define pkwb __builtin_alpha_pkwb #define unpkbl __builtin_alpha_unpkbl #define unpkbw __builtin_alpha_unpkbw #else #define cmpbge(a, b) ({ uint64_t __r; asm ("cmpbge %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define extql(a, b) ({ uint64_t __r; asm ("extql %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define extwl(a, b) ({ uint64_t __r; asm ("extwl %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define extqh(a, b) ({ uint64_t __r; asm ("extqh %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define zap(a, b) ({ uint64_t __r; asm ("zap %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define zapnot(a, b) ({ uint64_t __r; asm ("zapnot %r1,%2,%0" : "=r" (__r) : "rJ" (a), "rI" (b)); __r; }) #define amask(a) ({ uint64_t __r; asm ("amask %1,%0" : "=r" (__r) : "rI" (a)); __r; }) #define implver() ({ uint64_t __r; asm ("implver %0" : "=r" (__r)); __r; }) #define rpcc() ({ uint64_t __r; asm volatile ("rpcc %0" : "=r" (__r)); __r; }) #define minub8(a, b) ({ uint64_t __r; asm ("minub8 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define minsb8(a, b) ({ uint64_t __r; asm ("minsb8 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define minuw4(a, b) ({ uint64_t __r; asm ("minuw4 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define minsw4(a, b) ({ uint64_t __r; asm ("minsw4 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define maxub8(a, b) ({ uint64_t __r; asm ("maxub8 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define maxsb8(a, b) ({ uint64_t __r; asm ("maxsb8 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define maxuw4(a, b) ({ uint64_t __r; asm ("maxuw4 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define maxsw4(a, b) ({ uint64_t __r; asm ("maxsw4 %r1,%2,%0" : "=r" (__r) : "%rJ" (a), "rI" (b)); __r; }) #define perr(a, b) ({ uint64_t __r; asm ("perr %r1,%r2,%0" : "=r" (__r) : "%rJ" (a), "rJ" (b)); __r; }) #define pklb(a) ({ uint64_t __r; asm ("pklb %r1,%0" : "=r" (__r) : "rJ" (a)); __r; }) #define pkwb(a) ({ uint64_t __r; asm ("pkwb %r1,%0" : "=r" (__r) : "rJ" (a)); __r; }) #define unpkbl(a) ({ uint64_t __r; asm ("unpkbl %r1,%0" : "=r" (__r) : "rJ" (a)); __r; }) #define unpkbw(a) ({ uint64_t __r; asm ("unpkbw %r1,%0" : "=r" (__r) : "rJ" (a)); __r; }) #endif #define wh64(p) asm volatile("wh64 (%0)" : : "r"(p) : "memory") #elif defined(__DECC) /* Digital/Compaq/hp "ccc" compiler */ #include <c_asm.h> #define ldq_u(a) asm ("ldq_u %v0,0(%a0)", a) #define uldq(a) (*(const __unaligned uint64_t *) (a)) #define cmpbge(a, b) asm ("cmpbge %a0,%a1,%v0", a, b) #define extql(a, b) asm ("extql %a0,%a1,%v0", a, b) #define extwl(a, b) asm ("extwl %a0,%a1,%v0", a, b) #define extqh(a, b) asm ("extqh %a0,%a1,%v0", a, b) #define zap(a, b) asm ("zap %a0,%a1,%v0", a, b) #define zapnot(a, b) asm ("zapnot %a0,%a1,%v0", a, b) #define amask(a) asm ("amask %a0,%v0", a) #define implver() asm ("implver %v0") #define rpcc() asm ("rpcc %v0") #define minub8(a, b) asm ("minub8 %a0,%a1,%v0", a, b) #define minsb8(a, b) asm ("minsb8 %a0,%a1,%v0", a, b) #define minuw4(a, b) asm ("minuw4 %a0,%a1,%v0", a, b) #define minsw4(a, b) asm ("minsw4 %a0,%a1,%v0", a, b) #define maxub8(a, b) asm ("maxub8 %a0,%a1,%v0", a, b) #define maxsb8(a, b) asm ("maxsb8 %a0,%a1,%v0", a, b) #define maxuw4(a, b) asm ("maxuw4 %a0,%a1,%v0", a, b) #define maxsw4(a, b) asm ("maxsw4 %a0,%a1,%v0", a, b) #define perr(a, b) asm ("perr %a0,%a1,%v0", a, b) #define pklb(a) asm ("pklb %a0,%v0", a) #define pkwb(a) asm ("pkwb %a0,%v0", a) #define unpkbl(a) asm ("unpkbl %a0,%v0", a) #define unpkbw(a) asm ("unpkbw %a0,%v0", a) #define wh64(a) asm ("wh64 %a0", a) #else #error "Unknown compiler!" #endif #endif /* ALPHA_ASM_H */ --- NEW FILE --- /* * convert.h * Copyright (C) 2000-2002 Michel Lespinasse <walken@zoy.org> * Copyright (C) 1999-2000 Aaron Holtzman <aholtzma@ess.engr.uvic.ca> * * This file is part of mpeg2dec, a free MPEG-2 video stream decoder. * See http://libmpeg2.sourceforge.net/ for updates. * * mpeg2dec is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * mpeg2dec is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #ifndef CONVERT_H #define CONVERT_H #define CONVERT_FRAME 0 #define CONVERT_TOP_FIELD 1 #define CONVERT_BOTTOM_FIELD 2 #define CONVERT_BOTH_FIELDS 3 typedef struct convert_init_s { void * id; int id_size; int buf_size[3]; void (* start) (void * id, uint8_t * const * dest, int flags); void (* copy) (void * id, uint8_t * const * src, unsigned int v_offset); } convert_init_t; typedef void convert_t (int width, int height, uint32_t accel, void * arg, convert_init_t * result); convert_t convert_rgb32; convert_t convert_rgb24; convert_t convert_rgb16; convert_t convert_rgb15; convert_t convert_bgr32; convert_t convert_bgr24; convert_t convert_bgr16; convert_t convert_bgr15; #define CONVERT_RGB 0 #define CONVERT_BGR 1 convert_t * convert_rgb (int order, int bpp); #endif /* CONVERT_H */ --- NEW FILE --- /* * cpu_accel.c * Copyright (C) 2000-2002 Michel Lespinasse <walken@zoy.org> * Copyright (C) 1999-2000 Aaron Holtzman <aholtzma@ess.engr.uvic.ca> * * This file is part of mpeg2dec, a free MPEG-2 video stream decoder. * See http://libmpeg2.sourceforge.net/ for updates. * * mpeg2dec is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * mpeg2dec is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "config.h" #include <inttypes.h> #include "mpeg2.h" #ifdef ACCEL_DETECT #ifdef ARCH_X86 static inline uint32_t arch_accel (void) { uint32_t eax, ebx, ecx, edx; int AMD; uint32_t caps; #ifndef PIC #define cpuid(op,eax,ebx,ecx,edx) \ __asm__ ("cpuid" \ : "=a" (eax), \ "=b" (ebx), \ "=c" (ecx), \ "=d" (edx) \ : "a" (op) \ : "cc") #else /* PIC version : save ebx */ #define cpuid(op,eax,ebx,ecx,edx) \ __asm__ ("push %%ebx\n\t" \ "cpuid\n\t" \ "movl %%ebx,%1\n\t" \ "pop %%ebx" \ : "=a" (eax), \ "=r" (ebx), \ "=c" (ecx), \ "=d" (edx) \ : "a" (op) \ : "cc") #endif __asm__ ("pushf\n\t" "pushf\n\t" "pop %0\n\t" "movl %0,%1\n\t" "xorl $0x200000,%0\n\t" "push %0\n\t" "popf\n\t" "pushf\n\t" "pop %0\n\t" "popf" : "=r" (eax), "=r" (ebx) : : "cc"); if (eax == ebx) /* no cpuid */ return 0; cpuid (0x00000000, eax, ebx, ecx, edx); if (!eax) /* vendor string only */ return 0; AMD = (ebx == 0x68747541) && (ecx == 0x444d4163) && (edx == 0x69746e65); cpuid (0x00000001, eax, ebx, ecx, edx); if (! (edx & 0x00800000)) /* no MMX */ return 0; caps = MPEG2_ACCEL_X86_MMX; if (edx & 0x02000000) /* SSE - identical to AMD MMX extensions */ caps = MPEG2_ACCEL_X86_MMX | MPEG2_ACCEL_X86_MMXEXT; cpuid (0x80000000, eax, ebx, ecx, edx); if (eax < 0x80000001) /* no extended capabilities */ return caps; cpuid (0x80000001, eax, ebx, ecx, edx); if (edx & 0x80000000) caps |= MPEG2_ACCEL_X86_3DNOW; if (AMD && (edx & 0x00400000)) /* AMD MMX extensions */ caps |= MPEG2_ACCEL_X86_MMXEXT; return caps; } #endif /* ARCH_X86 */ #ifdef ARCH_PPC #include <signal.h> #include <setjmp.h> static sigjmp_buf jmpbuf; static volatile sig_atomic_t canjump = 0; static RETSIGTYPE sigill_handler (int sig) { if (!canjump) { signal (sig, SIG_DFL); raise (sig); } canjump = 0; siglongjmp (jmpbuf, 1); } static inline uint32_t arch_accel (void) { signal (SIGILL, sigill_handler); if (sigsetjmp (jmpbuf, 1)) { signal (SIGILL, SIG_DFL); return 0; } canjump = 1; asm volatile ("mtspr 256, %0\n\t" "vand %%v0, %%v0, %%v0" : : "r" (-1)); signal (SIGILL, SIG_DFL); return MPEG2_ACCEL_PPC_ALTIVEC; } #endif /* ARCH_PPC */ #ifdef ARCH_ALPHA static inline uint32_t arch_accel (void) { uint64_t no_mvi; asm volatile ("amask %1, %0" : "=r" (no_mvi) : "rI" (256)); /* AMASK_MVI */ return no_mvi ? MPEG2_ACCEL_ALPHA : (MPEG2_ACCEL_ALPHA | MPEG2_ACCEL_ALPHA_MVI); } #endif /* ARCH_ALPHA */ #endif uint32_t mpeg2_detect_accel (void) { uint32_t accel; accel = 0; #ifdef ACCEL_DETECT #ifdef LIBMPEG2_MLIB accel = MPEG2_ACCEL_MLIB; #endif #if defined (ARCH_X86) || defined (ARCH_PPC) || defined (ARCH_ALPHA) accel |= arch_accel (); #endif #endif return accel; } --- NEW FILE --- /* * cpu_state.c * Copyright (C) 2000-2002 Michel Lespinasse <walken@zoy.org> * Copyright (C) 1999-2000 Aaron Holtzman <aholtzma@ess.engr.uvic.ca> * * This file is part of mpeg2dec, a free MPEG-2 video stream decoder. * See http://libmpeg2.sourceforge.net/ for updates. * * mpeg2dec is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * mpeg2dec is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "config.h" #include <stdlib.h> #include <inttypes.h> #include "mpeg2.h" #include "mpeg2_internal.h" #include "attributes.h" #ifdef ARCH_X86 #include "mmx.h" #endif void (* mpeg2_cpu_state_save) (cpu_state_t * state) = NULL; void (* mpeg2_cpu_state_restore) (cpu_state_t * state) = NULL; #ifdef ARCH_X86 static void state_restore_mmx (cpu_state_t * state) { emms (); } #endif #ifdef ARCH_PPC static void state_save_altivec (cpu_state_t * state) { asm (" \n" " li %r9, 16 \n" " stvx %v20, 0, %r3 \n" " li %r11, 32 \n" " stvx %v21, %r9, %r3 \n" " li %r9, 48 \n" " stvx %v22, %r11, %r3 \n" " li %r11, 64 \n" " stvx %v23, %r9, %r3 \n" " li %r9, 80 \n" " stvx %v24, %r11, %r3 \n" " li %r11, 96 \n" " stvx %v25, %r9, %r3 \n" " li %r9, 112 \n" " stvx %v26, %r11, %r3 \n" " li %r11, 128 \n" " stvx %v27, %r9, %r3 \n" " li %r9, 144 \n" " stvx %v28, %r11, %r3 \n" " li %r11, 160 \n" " stvx %v29, %r9, %r3 \n" " li %r9, 176 \n" " stvx %v30, %r11, %r3 \n" " stvx %v31, %r9, %r3 \n" ); } static void state_restore_altivec (cpu_state_t * state) { asm (" \n" " li %r9, 16 \n" " lvx %v20, 0, %r3 \n" " li %r11, 32 \n" " lvx %v21, %r9, %r3 \n" " li %r9, 48 \n" " lvx %v22, %r11, %r3 \n" " li %r11, 64 \n" " lvx %v23, %r9, %r3 \n" " li %r9, 80 \n" " lvx %v24, %r11, %r3 \n" " li %r11, 96 \n" " lvx %v25, %r9, %r3 \n" " li %r9, 112 \n" " lvx %v26, %r11, %r3 \n" " li %r11, 128 \n" " lvx %v27, %r9, %r3 \n" " li %r9, 144 \n" " lvx %v28, %r11, %r3 \n" " li %r11, 160 \n" " lvx %v29, %r9, %r3 \n" " li %r9, 176 \n" " lvx %v30, %r11, %r3 \n" " lvx %v31, %r9, %r3 \n" ); } #endif void mpeg2_cpu_state_init (uint32_t accel) { #ifdef ARCH_X86 if (accel & MPEG2_ACCEL_X86_MMX) { mpeg2_cpu_state_restore = state_restore_mmx; } #endif #ifdef ARCH_PPC if (accel & MPEG2_ACCEL_PPC_ALTIVEC) { mpeg2_cpu_state_save = state_save_altivec; mpeg2_cpu_state_restore = state_restore_altivec; } #endif } --- NEW FILE --- /* * idct_alpha.c * Copyright (C) 2002 Falk Hueffner <falk@debian.org> * Copyright (C) 2000-2002 Michel Lespinasse <walken@zoy.org> * Copyright (C) 1999-2000 Aaron Holtzman <aholtzma@ess.engr.uvic.ca> * * This file is part of mpeg2dec, a free MPEG-2 video stream decoder. * See http://libmpeg2.sourceforge.net/ for updates. * * mpeg2dec is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * mpeg2dec is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "config.h" #ifdef ARCH_ALPHA #include <stdlib.h> #include <inttypes.h> #include "alpha_asm.h" #include "attributes.h" #define W1 2841 /* 2048*sqrt (2)*cos (1*pi/16) */ #define W2 2676 /* 2048*sqrt (2)*cos (2*pi/16) */ #define W3 2408 /* 2048*sqrt (2)*cos (3*pi/16) */ #define W5 1609 /* 2048*sqrt (2)*cos (5*pi/16) */ #define W6 1108 /* 2048*sqrt (2)*cos (6*pi/16) */ #define W7 565 /* 2048*sqrt (2)*cos (7*pi/16) */ static uint8_t clip_lut[1024]; #define CLIP(i) ((clip_lut+384)[(i)]) #if 0 #define BUTTERFLY(t0,t1,W0,W1,d0,d1) \ do { \ t0 = W0*d0 + W1*d1; \ t1 = W0*d1 - W1*d0; \ } while (0) #else #define BUTTERFLY(t0,t1,W0,W1,d0,d1) \ do { \ int_fast32_t tmp = W0 * (d0 + d1); \ t0 = tmp + (W1 - W0) * d1; \ t1 = tmp - (W1 + W0) * d0; \ } while (0) #endif static void inline idct_row (int16_t * const block) { uint64_t l, r; int_fast32_t d0, d1, d2, d3; int_fast32_t a0, a1, a2, a3, b0, b1, b2, b3; int_fast32_t t0, t1, t2, t3; l = ldq (block); r = ldq (block + 4); /* shortcut */ if (likely (!((l & ~0xffffUL) | r))) { uint64_t tmp = (uint16_t) (l << 3); tmp |= tmp << 16; tmp |= tmp << 32; ((int32_t *)block)[0] = tmp; ((int32_t *)block)[1] = tmp; ((int32_t *)block)[2] = tmp; ((int32_t *)block)[3] = tmp; return; } d0 = (sextw (l) << 11) + 128; d1 = sextw (extwl (l, 2)); d2 = sextw (extwl (l, 4)) << 11; d3 = sextw (extwl (l, 6)); t0 = d0 + d2; t1 = d0 - d2; BUTTERFLY (t2, t3, W6, W2, d3, d1); a0 = t0 + t2; a1 = t1 + t3; a2 = t1 - t3; a3 = t0 - t2; d0 = sextw (r); d1 = sextw (extwl (r, 2)); d2 = sextw (extwl (r, 4)); d3 = sextw (extwl (r, 6)); BUTTERFLY (t0, t1, W7, W1, d3, d0); BUTTERFLY (t2, t3, W3, W5, d1, d2); b0 = t0 + t2; b3 = t1 + t3; t0 -= t2; t1 -= t3; b1 = ((t0 + t1) * 181) >> 8; b2 = ((t0 - t1) * 181) >> 8; block[0] = (a0 + b0) >> 8; block[1] = (a1 + b1) >> 8; block[2] = (a2 + b2) >> 8; block[3] = (a3 + b3) >> 8; block[4] = (a3 - b3) >> 8; block[5] = (a2 - b2) >> 8; block[6] = (a1 - b1) >> 8; block[7] = (a0 - b0) >> 8; } static void inline idct_col (int16_t * const block) { int_fast32_t d0, d1, d2, d3; int_fast32_t a0, a1, a2, a3, b0, b1, b2, b3; int_fast32_t t0, t1, t2, t3; d0 = (block[8*0] << 11) + 65536; d1 = block[8*1]; d2 = block[8*2] << 11; d3 = block[8*3]; t0 = d0 + d2; t1 = d0 - d2; BUTTERFLY (t2, t3, W6, W2, d3, d1); a0 = t0 + t2; a1 = t1 + t3; a2 = t1 - t3; a3 = t0 - t2; d0 = block[8*4]; d1 = block[8*5]; d2 = block[8*6]; d3 = block[8*7]; BUTTERFLY (t0, t1, W7, W1, d3, d0); BUTTERFLY (t2, t3, W3, W5, d1, d2); b0 = t0 + t2; b3 = t1 + t3; t0 = (t0 - t2) >> 8; t1 = (t1 - t3) >> 8; b1 = (t0 + t1) * 181; b2 = (t0 - t1) * 181; block[8*0] = (a0 + b0) >> 17; block[8*1] = (a1 + b1) >> 17; block[8*2] = (a2 + b2) >> 17; block[8*3] = (a3 + b3) >> 17; block[8*4] = (a3 - b3) >> 17; block[8*5] = (a2 - b2) >> 17; block[8*6] = (a1 - b1) >> 17; block[8*7] = (a0 - b0) >> 17; } void mpeg2_idct_copy_mvi (int16_t * block, uint8_t * dest, const int stride) { uint64_t clampmask; int i; for (i = 0; i < 8; i++) idct_row (block + 8 * i); for (i = 0; i < 8; i++) idct_col (block + i); clampmask = zap (-1, 0xaa); /* 0x00ff00ff00ff00ff */ do { uint64_t shorts0, shorts1; shorts0 = ldq (block); shorts0 = maxsw4 (shorts0, 0); shorts0 = minsw4 (shorts0, clampmask); stl (pkwb (shorts0), dest); shorts1 = ldq (block + 4); shorts1 = maxsw4 (shorts1, 0); shorts1 = minsw4 (shorts1, clampmask); stl (pkwb (shorts1), dest + 4); stq (0, block); stq (0, block + 4); dest += stride; block += 8; } while (--i); } void mpeg2_idct_add_mvi (const int last, int16_t * block, uint8_t * dest, const int stride) { uint64_t clampmask; uint64_t signmask; int i; if (last != 129 || (block[0] & 7) == 4) { for (i = 0; i < 8; i++) idct_row (block + 8 * i); for (i = 0; i < 8; i++) idct_col (block + i); clampmask = zap (-1, 0xaa); /* 0x00ff00ff00ff00ff */ signmask = zap (-1, 0x33); signmask ^= signmask >> 1; /* 0x8000800080008000 */ do { uint64_t shorts0, pix0, signs0; uint64_t shorts1, pix1, signs1; shorts0 = ldq (block); shorts1 = ldq (block + 4); pix0 = unpkbw (ldl (dest)); /* signed subword add (MMX paddw). */ signs0 = shorts0 & signmask; shorts0 &= ~signmask; shorts0 += pix0; shorts0 ^= signs0; /* clamp. */ shorts0 = maxsw4 (shorts0, 0); shorts0 = minsw4 (shorts0, clampmask); /* next 4. */ pix1 = unpkbw (ldl (dest + 4)); signs1 = shorts1 & signmask; shorts1 &= ~signmask; shorts1 += pix1; shorts1 ^= signs1; shorts1 = maxsw4 (shorts1, 0); shorts1 = minsw4 (shorts1, clampmask); stl (pkwb (shorts0), dest); stl (pkwb (shorts1), dest + 4); stq (0, block); stq (0, block + 4); dest += stride; block += 8; } while (--i); } else { int DC; uint64_t p0, p1, p2, p3, p4, p5, p6, p7; uint64_t DCs; DC = (block[0] + 4) >> 3; block[0] = block[63] = 0; p0 = ldq (dest + 0 * stride); p1 = ldq (dest + 1 * stride); p2 = ldq (dest + 2 * stride); p3 = ldq (dest + 3 * stride); p4 = ldq (dest + 4 * stride); p5 = ldq (dest + 5 * stride); p6 = ldq (dest + 6 * stride); p7 = ldq (dest + 7 * stride); if (DC > 0) { DCs = BYTE_VEC (likely (DC <= 255) ? DC : 255); p0 += minub8 (DCs, ~p0); p1 += minub8 (DCs, ~p1); p2 += minub8 (DCs, ~p2); p3 += minub8 (DCs, ~p3); p4 += minub8 (DCs, ~p4); p5 += minub8 (DCs, ~p5); p6 += minub8 (DCs, ~p6); p7 += minub8 (DCs, ~p7); } else { DCs = BYTE_VEC (likely (-DC <= 255) ? -DC : 255); p0 -= minub8 (DCs, p0); p1 -= minub8 (DCs, p1); p2 -= minub8 (DCs, p2); p3 -= minub8 (DCs, p3); p4 -= minub8 (DCs, p4); p5 -= minub8 (DCs, p5); p6 -= minub8 (DCs, p6); p7 -= minub8 (DCs, p7); } stq (p0, dest + 0 * stride); stq (p1, dest + 1 * stride); stq (p2, dest + 2 * stride); stq (p3, dest + 3 * stride); stq (p4, dest + 4 * stride); stq (p5, dest + 5 * stride); stq (p6, dest + 6 * stride); stq (p7, dest + 7 * stride); } } void mpeg2_idct_copy_alpha (int16_t * block, uint8_t * dest, const int stride) { int i; for (i = 0; i < 8; i++) idct_row (block + 8 * i); for (i = 0; i < 8; i++) idct_col (block + i); do { dest[0] = CLIP (block[0]); dest[1] = CLIP (block[1]); dest[2] = CLIP (block[2]); dest[3] = CLIP (block[3]); dest[4] = CLIP (block[4]); dest[5] = CLIP (block[5]); dest[6] = CLIP (block[6]); dest[7] = CLIP (block[7]); stq(0, block); stq(0, block + 4); dest += stride; block += 8; } while (--i); } void mpeg2_idct_add_alpha (const int last, int16_t * block, uint8_t * dest, const int stride) { int i; if (last != 129 || (block[0] & 7) == 4) { for (i = 0; i < 8; i++) idct_row (block + 8 * i); for (i = 0; i < 8; i++) idct_col (block + i); do { dest[0] = CLIP (block[0] + dest[0]); dest[1] = CLIP (block[1] + dest[1]); dest[2] = CLIP (block[2] + dest[2]); dest[3] = CLIP (block[3] + dest[3]); dest[4] = CLIP (block[4] + dest[4]); dest[5] = CLIP (block[5] + dest[5]); dest[6] = CLIP (block[6] + dest[6]); dest[7] = CLIP (block[7] + dest[7]); stq(0, block); stq(0, block + 4); dest += stride; block += 8; } while (--i); } else { int DC; DC = (block[0] + 4) >> 3; block[0] = block[63] = 0; i = 8; do { dest[0] = CLIP (DC + dest[0]); dest[1] = CLIP (DC + dest[1]); dest[2] = CLIP (DC + dest[2]); dest[3] = CLIP (DC + dest[3]); dest[4] = CLIP (DC + dest[4]); dest[5] = CLIP (DC + dest[5]); dest[6] = CLIP (DC + dest[6]); dest[7] = CLIP (DC + dest[7]); dest += stride; } while (--i); } } void mpeg2_idct_alpha_init(int no_mvi) { extern uint8_t mpeg2_scan_norm[64]; extern uint8_t mpeg2_scan_alt[64]; int i, j; if (no_mvi) for (i = -384; i < 640; i++) clip_lut[i + 384] = (i < 0) ? 0 : ((i > 255) ? 255 : i); for (i = 0; i < 64; i++) { j = mpeg2_scan_norm[i]; mpeg2_scan_norm[i] = ((j & 0x36) >> 1) | ((j & 0x09) << 2); j = mpeg2_scan_alt[i]; mpeg2_scan_alt[i] = ((j & 0x36) >> 1) | ((j & 0x09) << 2); } } #endif /* ARCH_ALPHA */ --- NEW FILE --- /* * idct_altivec.c * Copyright (C) 2000-2002 Michel Lespinasse <walken@zoy.org> * Copyright (C) 1999-2000 Aaron Holtzman <aholtzma@ess.engr.uvic.ca> * * This file is part of mpeg2dec, a free MPEG-2 video stream decoder. * See http://libmpeg2.sourceforge.net/ for updates. * * mpeg2dec is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * mpeg2dec is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #ifndef __ALTIVEC__ #include "config.h" #ifdef ARCH_PPC #include <inttypes.h> #include "mpeg2.h" #include "mpeg2_internal.h" #include "attributes.h" static const int16_t constants[5][8] ATTR_ALIGN(16) = { {23170, 13573, 6518, 21895, -23170, -21895, 32, 31}, {16384, 22725, 21407, 19266, 16384, 19266, 21407, 22725}, {22725, 31521, 29692, 26722, 22725, 26722, 29692, 31521}, {21407, 29692, 27969, 25172, 21407, 25172, 27969, 29692}, {19266, 26722, 25172, 22654, 19266, 22654, 25172, 26722} }; /* * The asm code is generated with: * * gcc-2.95 -fvec -D__ALTIVEC__ -O9 -fomit-frame-pointer -mregnames -S * idct_altivec.c * * awk '{args=""; len=split ($2, arg, ","); * for (i=1; i<=len; i++) { a=arg[i]; if (i<len) a=a","; * args = args sprintf ("%-6s", a) } * printf ("\t\"\t%-16s%-24s\\n\"\n", $1, args) }' idct_altivec.s | * unexpand -a * * I then do some simple trimming on the function prolog/trailers */ void mpeg2_idct_copy_altivec (int16_t * block, uint8_t * dest, int stride) { asm (" \n" "# stwu %r1, -128(%r1) \n" "# mflr %r0 \n" "# stw %r0, 132(%r1) \n" "# addi %r0, %r1, 128 \n" "# bl _savev25 \n" " addi %r9, %r3, 112 \n" " vspltish %v25, 4 \n" " vxor %v13, %v13, %v13 \n" " lis %r10, constants@ha \n" " lvx %v1, 0, %r9 \n" " la %r10, constants@l(%r10) \n" " lvx %v5, 0, %r3 \n" " addi %r9, %r3, 16 \n" " lvx %v8, 0, %r10 \n" " addi %r11, %r10, 32 \n" " lvx %v12, 0, %r9 \n" " lvx %v6, 0, %r11 \n" " addi %r8, %r3, 48 \n" " vslh %v1, %v1, %v25 \n" " addi %r9, %r3, 80 \n" " lvx %v11, 0, %r8 \n" " vslh %v5, %v5, %v25 \n" " lvx %v0, 0, %r9 \n" " addi %r11, %r10, 64 \n" " vsplth %v3, %v8, 2 \n" " lvx %v7, 0, %r11 \n" " addi %r9, %r3, 96 \n" " vslh %v12, %v12, %v25 \n" " vmhraddshs %v27, %v1, %v6, %v13 \n" " addi %r8, %r3, 32 \n" " vsplth %v2, %v8, 5 \n" " lvx %v1, 0, %r9 \n" " vslh %v11, %v11, %v25 \n" " addi %r3, %r3, 64 \n" " lvx %v9, 0, %r8 \n" " addi %r9, %r10, 48 \n" " vslh %v0, %v0, %v25 \n" " lvx %v4, 0, %r9 \n" " vmhraddshs %v31, %v12, %v6, %v13 \n" " addi %r10, %r10, 16 \n" " vmhraddshs %v30, %v0, %v7, %v13 \n" " lvx %v10, 0, %r3 \n" " vsplth %v19, %v8, 3 \n" " vmhraddshs %v15, %v11, %v7, %v13 \n" " lvx %v12, 0, %r10 \n" " vsplth %v6, %v8, 4 \n" " vslh %v1, %v1, %v25 \n" " vsplth %v11, %v8, 1 \n" " li %r9, 4 \n" " vslh %v9, %v9, %v25 \n" " vsplth %v7, %v8, 0 \n" " vmhraddshs %v18, %v1, %v4, %v13 \n" " vspltw %v8, %v8, 3 \n" " vsubshs %v0, %v13, %v27 \n" " vmhraddshs %v1, %v9, %v4, %v13 \n" " vmhraddshs %v17, %v3, %v31, %v0 \n" " vmhraddshs %v4, %v2, %v15, %v30 \n" " vslh %v10, %v10, %v25 \n" " vmhraddshs %v9, %v5, %v12, %v13 \n" " vspltish %v25, 6 \n" " vmhraddshs %v5, %v10, %v12, %v13 \n" " vmhraddshs %v28, %v19, %v30, %v15 \n" " vmhraddshs %v27, %v3, %v27, %v31 \n" " vsubshs %v0, %v13, %v18 \n" " vmhraddshs %v18, %v11, %v18, %v1 \n" " vaddshs %v30, %v17, %v4 \n" " vmhraddshs %v12, %v11, %v1, %v0 \n" " vsubshs %v4, %v17, %v4 \n" " vaddshs %v10, %v9, %v5 \n" " vsubshs %v17, %v27, %v28 \n" " vaddshs %v27, %v27, %v28 \n" " vsubshs %v1, %v9, %v5 \n" " vaddshs %v28, %v10, %v18 \n" " vsubshs %v18, %v10, %v18 \n" " vaddshs %v10, %v1, %v12 \n" " vsubshs %v1, %v1, %v12 \n" " vsubshs %v12, %v17, %v4 \n" " vaddshs %v4, %v17, %v4 \n" " vmhraddshs %v5, %v7, %v12, %v1 \n" " vmhraddshs %v26, %v6, %v4, %v10 \n" " vmhraddshs %v29, %v6, %v12, %v1 \n" " vmhraddshs %v14, %v7, %v4, %v10 \n" " vsubshs %v12, %v18, %v30 \n" " vaddshs %v9, %v28, %v27 \n" " vaddshs %v16, %v18, %v30 \n" " vsubshs %v10, %v28, %v27 \n" " vmrglh %v31, %v9, %v12 \n" " vmrglh %v30, %v5, %v26 \n" " vmrglh %v15, %v14, %v29 \n" " vmrghh %v5, %v5, %v26 \n" " vmrglh %v27, %v16, %v10 \n" " vmrghh %v9, %v9, %v12 \n" " vmrghh %v18, %v16, %v10 \n" " vmrghh %v1, %v14, %v29 \n" " vmrglh %v14, %v9, %v5 \n" " vmrglh %v16, %v31, %v30 \n" " vmrglh %v10, %v15, %v27 \n" " vmrghh %v9, %v9, %v5 \n" " vmrghh %v26, %v15, %v27 \n" " vmrglh %v27, %v16, %v10 \n" " vmrghh %v12, %v1, %v18 \n" " vmrglh %v29, %v1, %v18 \n" " vsubshs %v0, %v13, %v27 \n" " vmrghh %v5, %v31, %v30 \n" " vmrglh %v31, %v9, %v12 \n" " vmrglh %v30, %v5, %v26 \n" " vmrglh %v15, %v14, %v29 \n" " vmhraddshs %v17, %v3, %v31, %v0 \n" " vmrghh %v18, %v16, %v10 \n" " vmhraddshs %v27, %v3, %v27, %v31 \n" " vmhraddshs %v4, %v2, %v15, %v30 \n" " vmrghh %v1, %v14, %v29 \n" " vmhraddshs %v28, %v19, %v30, %v15 \n" " vmrghh %v0, %v9, %v12 \n" " vsubshs %v13, %v13, %v18 \n" " vmrghh %v5, %v5, %v26 \n" " vmhraddshs %v18, %v11, %v18, %v1 \n" " vaddshs %v9, %v0, %v8 \n" " vaddshs %v30, %v17, %v4 \n" " vmhraddshs %v12, %v11, %v1, %v13 \n" " vsubshs %v4, %v17, %v4 \n" " vaddshs %v10, %v9, %v5 \n" " vsubshs %v17, %v27, %v28 \n" " vaddshs %v27, %v27, %v28 \n" " vsubshs %v1, %v9, %v5 \n" " vaddshs %v28, %v10, %v18 \n" " vsubshs %v18, %v10, %v18 \n" " vaddshs %v10, %v1, %v12 \n" " vsubshs %v1, %v1, %v12 \n" " vsubshs %v12, %v17, %v4 \n" " vaddshs %v4, %v17, %v4 \n" " vaddshs %v9, %v28, %v27 \n" " vmhraddshs %v14, %v7, %v4, %v10 \n" " vsrah %v9, %v9, %v25 \n" " vmhraddshs %v5, %v7, %v12, %v1 \n" " vpkshus %v0, %v9, %v9 \n" " vmhraddshs %v29, %v6, %v12, %v1 \n" " stvewx %v0, 0, %r4 \n" " vaddshs %v16, %v18, %v30 \n" " vsrah %v31, %v14, %v25 \n" " stvewx %v0, %r9, %r4 \n" " add %r4, %r4, %r5 \n" " vsrah %v15, %v16, %v25 \n" " vpkshus %v0, %v31, %v31 \n" " vsrah %v1, %v5, %v25 \n" " stvewx %v0, 0, %r4 \n" " vsubshs %v12, %v18, %v30 \n" " stvewx %v0, %r9, %r4 \n" " vmhraddshs %v26, %v6, %v4, %v10 \n" " vpkshus %v0, %v1, %v1 \n" " add %r4, %r4, %r5 \n" " vsrah %v5, %v12, %v25 \n" " stvewx %v0, 0, %r4 \n" " vsrah %v30, %v29, %v25 \n" " stvewx %v0, %r9, %r4 \n" " vsubshs %v10, %v28, %v27 \n" " vpkshus %v0, %v15, %v15 \n" " add %r4, %r4, %r5 \n" " stvewx %v0, 0, %r4 \n" " vsrah %v18, %v26, %v25 \n" " stvewx %v0, %r9, %r4 \n" " vsrah %v27, %v10, %v25 \n" " vpkshus %v0, %v5, %v5 \n" " add %r4, %r4, %r5 \n" " stvewx %v0, 0, %r4 \n" " stvewx %v0, %r9, %r4 \n" " vpkshus %v0, %v30, %v30 \n" " add %r4, %r4, %r5 \n" " stvewx %v0, 0, %r4 \n" " stvewx %v0, %r9, %r4 \n" " vpkshus %v0, %v18, %v18 \n" " add %r4, %r4, %r5 \n" " stvewx %v0, 0, %r4 \n" " stvewx %v0, %r9, %r4 \n" " add %r4, %r4, %r5 \n" " vpkshus %v0, %v27, %v27 \n" " stvewx %v0, 0, %r4 \n" " stvewx %v0, %r9, %r4 \n" "# addi %r0, %r1, 128 \n" "# bl _restv25 \n" "# lwz %r0, 132(%r1) \n" "# mtlr %r0 \n" "# la %r1, 128(%r1) \n" " vxor %v1, %v1, %v1 \n" " addi %r9, %r3, 16 \n" " stvx %v1, 0, %r3 \n" " stvx %v1, 0, %r9 \n" " addi %r11, %r3, 32 \n" " stvx %v1, 0, %r11 \n" " addi %r9, %r3, 48 \n" " stvx %v1, 0, %r9 \n" " addi %r11, %r3, -64 \n" " stvx %v1, 0, %r11 \n" " addi %r9, %r3, -48 \n" " stvx %v1, 0, %r9 \n" " addi %r11, %r3, -32 \n" " stvx %v1, 0, %r11 \n" " addi %r3, %r3, -16 \n" " stvx %v1, 0, %r3 \n" ); } void mpeg2_idct_add_altivec (int last, int16_t * block, uint8_t * dest, int stride) { asm (" \n" "# stwu %r1, -192(%r1) \n" "# mflr %r0 \n" "# stw %r0, 196(%r1) \n" "# addi %r0, %r1, 192 \n" "# bl _savev21 \n" " addi %r9, %r4, 112 \n" " vspltish %v21, 4 \n" " vxor %v1, %v1, %v1 \n" " lvx %v13, 0, %r9 \n" " lis %r10, constants@ha \n" " vspltisw %v3, -1 \n" " la %r10, constants@l(%r10) \n" " lvx %v5, 0, %r4 \n" " addi %r9, %r4, 16 \n" " lvx %v8, 0, %r10 \n" " lvx %v12, 0, %r9 \n" " addi %r11, %r10, 32 \n" " lvx %v6, 0, %r11 \n" " addi %r8, %r4, 48 \n" " vslh %v13, %v13, %v21 \n" " addi %r9, %r4, 80 \n" " lvx %v11, 0, %r8 \n" " vslh %v5, %v5, %v21 \n" " lvx %v0, 0, %r9 \n" " addi %r11, %r10, 64 \n" " vsplth %v2, %v8, 2 \n" " lvx %v7, 0, %r11 \n" " vslh %v12, %v12, %v21 \n" " addi %r9, %r4, 96 \n" " vmhraddshs %v24, %v13, %v6, %v1 \n" " addi %r8, %r4, 32 \n" " vsplth %v17, %v8, 5 \n" " lvx %v13, 0, %r9 \n" " vslh %v11, %v11, %v21 \n" " addi %r4, %r4, 64 \n" " lvx %v10, 0, %r8 \n" " vslh %v0, %v0, %v21 \n" " addi %r9, %r10, 48 \n" " vmhraddshs %v31, %v12, %v6, %v1 \n" " lvx %v4, 0, %r9 \n" " addi %r10, %r10, 16 \n" " vmhraddshs %v26, %v0, %v7, %v1 \n" " lvx %v9, 0, %r4 \n" " vsplth %v16, %v8, 3 \n" " vmhraddshs %v22, %v11, %v7, %v1 \n" " lvx %v6, 0, %r10 \n" " lvsl %v19, 0, %r5 \n" " vsubshs %v12, %v1, %v24 \n" " lvsl %v0, %r6, %r5 \n" " vsplth %v11, %v8, 1 \n" " vslh %v10, %v10, %v21 \n" " vmrghb %v19, %v3, %v19 \n" " lvx %v15, 0, %r5 \n" " vslh %v13, %v13, %v21 \n" " vmrghb %v3, %v3, %v0 \n" " li %r9, 4 \n" " vmhraddshs %v14, %v2, %v31, %v12 \n" " vsplth %v7, %v8, 0 \n" " vmhraddshs %v23, %v13, %v4, %v1 \n" " vsplth %v18, %v8, 4 \n" " vmhraddshs %v27, %v10, %v4, %v1 \n" " vspltw %v8, %v8, 3 \n" " vmhraddshs %v12, %v17, %v22, %v26 \n" " vperm %v15, %v15, %v1, %v19 \n" " vslh %v9, %v9, %v21 \n" " vmhraddshs %v10, %v5, %v6, %v1 \n" " vspltish %v21, 6 \n" " vmhraddshs %v30, %v9, %v6, %v1 \n" " vmhraddshs %v26, %v16, %v26, %v22 \n" " vmhraddshs %v24, %v2, %v24, %v31 \n" " vmhraddshs %v31, %v11, %v23, %v27 \n" " vsubshs %v0, %v1, %v23 \n" " vaddshs %v23, %v14, %v12 \n" " vmhraddshs %v9, %v11, %v27, %v0 \n" " vsubshs %v12, %v14, %v12 \n" " vaddshs %v6, %v10, %v30 \n" " vsubshs %v14, %v24, %v26 \n" " vaddshs %v24, %v24, %v26 \n" " vsubshs %v13, %v10, %v30 \n" " vaddshs %v26, %v6, %v31 \n" " vsubshs %v31, %v6, %v31 \n" " vaddshs %v6, %v13, %v9 \n" " vsubshs %v13, %v13, %v9 \n" " vsubshs %v9, %v14, %v12 \n" " vaddshs %v12, %v14, %v12 \n" " vmhraddshs %v30, %v7, %v9, %v13 \n" " vmhraddshs %v25, %v18, %v12, %v6 \n" " vmhraddshs %v28, %v18, %v9, %v13 \n" " vmhraddshs %v29, %v7, %v12, %v6 \n" " vaddshs %v10, %v26, %v24 \n" " vsubshs %v5, %v31, %v23 \n" " vsubshs %v13, %v26, %v24 \n" " vaddshs %v4, %v31, %v23 \n" " vmrglh %v26, %v30, %v25 \n" " vmrglh %v31, %v10, %v5 \n" " vmrglh %v22, %v29, %v28 \n" " vmrghh %v30, %v30, %v25 \n" " vmrglh %v24, %v4, %v13 \n" " vmrghh %v10, %v10, %v5 \n" " vmrghh %v23, %v4, %v13 \n" " vmrghh %v27, %v29, %v28 \n" " vmrglh %v29, %v10, %v30 \n" " vmrglh %v4, %v31, %v26 \n" " vmrglh %v13, %v22, %v24 \n" " vmrghh %v10, %v10, %v30 \n" " vmrghh %v25, %v22, %v24 \n" " vmrglh %v24, %v4, %v13 \n" " vmrghh %v5, %v27, %v23 \n" " vmrglh %v28, %v27, %v23 \n" " vsubshs %v0, %v1, %v24 \n" " vmrghh %v30, %v31, %v26 \n" " vmrglh %v31, %v10, %v5 \n" " vmrglh %v26, %v30, %v25 \n" " vmrglh %v22, %v29, %v28 \n" " vmhraddshs %v14, %v2, %v31, %v0 \n" " vmrghh %v23, %v4, %v13 \n" " vmhraddshs %v24, %v2, %v24, %v31 \n" " vmhraddshs %v12, %v17, %v22, %v26 \n" " vmrghh %v27, %v29, %v28 \n" " vmhraddshs %v26, %v16, %v26, %v22 \n" " vmrghh %v0, %v10, %v5 \n" " vmhraddshs %v31, %v11, %v23, %v27 \n" " vmrghh %v30, %v30, %v25 \n" " vsubshs %v13, %v1, %v23 \n" " vaddshs %v10, %v0, %v8 \n" " vaddshs %v23, %v14, %v12 \n" " vsubshs %v12, %v14, %v12 \n" " vaddshs %v6, %v10, %v30 \n" " vsubshs %v14, %v24, %v26 \n" " vmhraddshs %v9, %v11, %v27, %v13 \n" " vaddshs %v24, %v24, %v26 \n" " vaddshs %v26, %v6, %v31 \n" " vsubshs %v13, %v10, %v30 \n" " vaddshs %v10, %v26, %v24 \n" " vsubshs %v31, %v6, %v31 \n" " vaddshs %v6, %v13, %v9 \n" " vsrah %v10, %v10, %v21 \n" " vsubshs %v13, %v13, %v9 \n" " vaddshs %v0, %v15, %v10 \n" " vsubshs %v9, %v14, %v12 \n" " vaddshs %v12, %v14, %v12 \n" " vpkshus %v15, %v0, %v0 \n" " stvewx %v15, 0, %r5 \n" " vaddshs %v4, %v31, %v23 \n" " vmhraddshs %v29, %v7, %v12, %v6 \n" " stvewx %v15, %r9, %r5 \n" " add %r5, %r5, %r6 \n" " vsubshs %v5, %v31, %v23 \n" " lvx %v15, 0, %r5 \n" " vmhraddshs %v30, %v7, %v9, %v13 \n" " vsrah %v22, %v4, %v21 \n" " vperm %v15, %v15, %v1, %v3 \n" " vmhraddshs %v28, %v18, %v9, %v13 \n" " vsrah %v31, %v29, %v21 \n" " vsubshs %v13, %v26, %v24 \n" " vaddshs %v0, %v15, %v31 \n" " vsrah %v27, %v30, %v21 \n" " vpkshus %v15, %v0, %v0 \n" " vsrah %v30, %v5, %v21 \n" " stvewx %v15, 0, %r5 \n" " vsrah %v26, %v28, %v21 \n" " stvewx %v15, %r9, %r5 \n" " vmhraddshs %v25, %v18, %v12, %v6 \n" " add %r5, %r5, %r6 \n" " vsrah %v24, %v13, %v21 \n" " lvx %v15, 0, %r5 \n" " vperm %v15, %v15, %v1, %v19 \n" " vsrah %v23, %v25, %v21 \n" " vaddshs %v0, %v15, %v27 \n" " vpkshus %v15, %v0, %v0 \n" " stvewx %v15, 0, %r5 \n" " stvewx %v15, %r9, %r5 \n" " add %r5, %r5, %r6 \n" " lvx %v15, 0, %r5 \n" " vperm %v15, %v15, %v1, %v3 \n" " vaddshs %v0, %v15, %v22 \n" " vpkshus %v15, %v0, %v0 \n" " stvewx %v15, 0, %r5 \n" " stvewx %v15, %r9, %r5 \n" " add %r5, %r5, %r6 \n" " lvx %v15, 0, %r5 \n" " vperm %v15, %v15, %v1, %v19 \n" " vaddshs %v0, %v15, %v30 \n" " vpkshus %v15, %v0, %v0 \n" " stvewx %v15, 0, %r5 \n" " stvewx %v15, %r9, %r5 \n" " add %r5, %r5, %r6 \n" " lvx %v15, 0, %r5 \n" " vperm %v15, %v15, %v1, %v3 \n" " vaddshs %v0, %v15, %v26 \n" " vpkshus %v15, %v0, %v0 \n" " stvewx %v15, 0, %r5 \n" " stvewx %v15, %r9, %r5 \n" " add %r5, %r5, %r6 \n" " lvx %v15, 0, %r5 \n" " vperm %v15, %v15, %v1, %v19 \n" " vaddshs %v0, %v15, %v23 \n" " vpkshus %v15, %v0, %v0 \n" " stvewx %v15, 0, %r5 \n" " stvewx %v15, %r9, %r5 \n" " add %r5, %r5, %r6 \n" " lvx %v15, 0, %r5 \n" " vperm %v15, %v15, %v1, %v3 \n" " vaddshs %v0, %v15, %v24 \n" " vpkshus %v15, %v0, %v0 \n" " stvewx %v15, 0, %r5 \n" " stvewx %v15, %r9, %r5 \n" "# addi %r0, %r1, 192 \n" "# bl _restv21 \n" "# lwz %r0, 196(%r1) \n" "# mtlr %r0 \n" "# la %r1, 192(%r1) \n" " addi %r9, %r4, 16 \n" " stvx %v1, 0, %r4 \n" " stvx %v1, 0, %r9 \n" " addi %r11, %r4, 32 \n" " stvx %v1, 0, %r11 \n" " addi %r9, %r4, 48 \n" " stvx %v1, 0, %r9 \n" " addi %r11, %r4, -64 \n" " stvx %v1, 0, %r11 \n" " addi %r9, %r4, -48 \n" " stvx %v1, 0, %r9 \n" " addi %r11, %r4, -32 \n" " stvx %v1, 0, %r11 \n" " addi %r4, %r4, -16 \n" " stvx %v1, 0, %r4 \n" ); } void mpeg2_idct_altivec_init (void) { extern uint8_t mpeg2_scan_norm[64]; extern uint8_t mpeg2_scan_alt[64]; int i, j; i = constants[0][0]; /* just pretending - keeps gcc happy */ /* the altivec idct uses a transposed input, so we patch scan tables */ for (i = 0; i < 64; i++) { j = mpeg2_scan_norm[i]; mpeg2_scan_norm[i] = (j >> 3) | ((j & 7) << 3); j = mpeg2_scan_alt[i]; mpeg2_scan_alt[i] = (j >> 3) | ((j & 7) << 3); } } #endif /* ARCH_PPC */ #else /* __ALTIVEC__ */ #define vector_s16_t vector signed short #define vector_u16_t vector unsigned short #define vector_s8_t vector signed char #define vector_u8_t vector unsigned char #define vector_s32_t vector signed int #define vector_u32_t vector unsigned int #define IDCT_HALF \ /* 1st stage */ \ t1 = vec_mradds (a1, vx7, vx1 ); \ t8 = vec_mradds (a1, vx1, vec_subs (zero, vx7)); \ t7 = vec_mradds (a2, vx5, vx3); \ t3 = vec_mradds (ma2, vx3, vx5); \ \ /* 2nd stage */ \ t5 = vec_adds (vx0, vx4); \ t0 = vec_subs (vx0, vx4); \ t2 = vec_mradds (a0, vx6, vx2); \ t4 = vec_mradds (a0, vx2, vec_subs (zero, vx6)); \ t6 = vec_adds (t8, t3); \ t3 = vec_subs (t8, t3); \ t8 = vec_subs (t1, t7); \ t1 = vec_adds (t1, t7); \ \ /* 3rd stage */ \ t7 = vec_adds (t5, t2); \ t2 = vec_subs (t5, t2); \ t5 = vec_adds (t0, t4); \ t0 = vec_subs (t0, t4); \ t4 = vec_subs (t8, t3); \ t3 = vec_adds (t8, t3); \ \ /* 4th stage */ \ vy0 = vec_adds (t7, t1); \ vy7 = vec_subs (t7, t1); \ vy1 = vec_mradds (c4, t3, t5); \ vy6 = vec_mradds (mc4, t3, t5); \ vy2 = vec_mradds (c4, t4, t0); \ vy5 = vec_mradds (mc4, t4, t0); \ vy3 = vec_adds (t2, t6); \ vy4 = vec_subs (t2, t6); #define IDCT \ vector_s16_t vx0, vx1, vx2, vx3, vx4, vx5, vx6, vx7; \ vector_s16_t vy0, vy1, vy2, vy3, vy4, vy5, vy6, vy7; \ vector_s16_t a0, a1, a2, ma2, c4, mc4, zero, bias; \ vector_s16_t t0, t1, t2, t3, t4, t5, t6, t7, t8; \ vector_u16_t shift; \ \ c4 = vec_splat (constants[0], 0); \ a0 = vec_splat (constants[0], 1); \ a1 = vec_splat (constants[0], 2); \ a2 = vec_splat (constants[0], 3); \ mc4 = vec_splat (constants[0], 4); \ ma2 = vec_splat (constants[0], 5); \ bias = (vector_s16_t)vec_splat ((vector_s32_t)constants[0], 3); \ \ zero = vec_splat_s16 (0); \ shift = vec_splat_u16 (4); \ \ vx0 = vec_mradds (vec_sl (block[0], shift), constants[1], zero); \ vx1 = vec_mradds (vec_sl (block[1], shift), constants[2], zero); \ vx2 = vec_mradds (vec_sl (block[2], shift), constants[3], zero); \ vx3 = vec_mradds (vec_sl (block[3], shift), constants[4], zero); \ vx4 = vec_mradds (vec_sl (block[4], shift), constants[1], zero); \ vx5 = vec_mradds (vec_sl (block[5], shift), constants[4], zero); \ vx6 = vec_mradds (vec_sl (block[6], shift), constants[3], zero); \ vx7 = vec_mradds (vec_sl (block[7], shift), constants[2], zero); \ \ IDCT_HALF \ \ vx0 = vec_mergeh (vy0, vy4); \ vx1 = vec_mergel (vy0, vy4); \ vx2 = vec_mergeh (vy1, vy5); \ vx3 = vec_mergel (vy1, vy5); \ vx4 = vec_mergeh (vy2, vy6); \ vx5 = vec_mergel (vy2, vy6); \ vx6 = vec_mergeh (vy3, vy7); \ vx7 = vec_mergel (vy3, vy7); \ \ vy0 = vec_mergeh (vx0, vx4); \ vy1 = vec_mergel (vx0, vx4); \ vy2 = vec_mergeh (vx1, vx5); \ vy3 = vec_mergel (vx1, vx5); \ vy4 = vec_mergeh (vx2, vx6); \ vy5 = vec_mergel (vx2, vx6); \ vy6 = vec_mergeh (vx3, vx7); \ vy7 = vec_mergel (vx3, vx7); \ \ vx0 = vec_adds (vec_mergeh (vy0, vy4), bias); \ vx1 = vec_mergel (vy0, vy4); \ vx2 = vec_mergeh (vy1, vy5); \ vx3 = vec_mergel (vy1, vy5); \ vx4 = vec_mergeh (vy2, vy6); \ vx5 = vec_mergel (vy2, vy6); \ vx6 = vec_mergeh (vy3, vy7); \ vx7 = vec_mergel (vy3, vy7); \ \ IDCT_HALF \ \ shift = vec_splat_u16 (6); \ vx0 = vec_sra (vy0, shift); \ vx1 = vec_sra (vy1, shift); \ vx2 = vec_sra (vy2, shift); \ vx3 = vec_sra (vy3, shift); \ vx4 = vec_sra (vy4, shift); \ vx5 = vec_sra (vy5, shift); \ vx6 = vec_sra (vy6, shift); \ vx7 = vec_sra (vy7, shift); static const vector_s16_t constants[5] = { (vector_s16_t)(23170, 13573, 6518, 21895, -23170, -21895, 32, 31), (vector_s16_t)(16384, 22725, 21407, 19266, 16384, 19266, 21407, 22725), (vector_s16_t)(22725, 31521, 29692, 26722, 22725, 26722, 29692, 31521), (vector_s16_t)(21407, 29692, 27969, 25172, 21407, 25172, 27969, 29692), (vector_s16_t)(19266, 26722, 25172, 22654, 19266, 22654, 25172, 26722) }; void mpeg2_idct_copy_altivec (vector_s16_t * const block, unsigned char * dest, const int stride) { vector_u8_t tmp; IDCT #define COPY(dest,src) \ tmp = vec_packsu (src, src); \ vec_ste ((vector_u32_t)tmp, 0, (unsigned int *)dest); \ vec_ste ((vector_u32_t)tmp, 4, (unsigned int *)dest); COPY (dest, vx0) dest += stride; COPY (dest, vx1) dest += stride; COPY (dest, vx2) dest += stride; COPY (dest, vx3) dest += stride; COPY (dest, vx4) dest += stride; COPY (dest, vx5) dest += stride; COPY (dest, vx6) dest += stride; COPY (dest, vx7) memset (block, 0, 64 * sizeof (signed short)); } void mpeg2_idct_add_altivec (const int last, vector_s16_t * const block, unsigned char * dest, const int stride) { vector_u8_t tmp; vector_s16_t tmp2, tmp3; vector_u8_t perm0; vector_u8_t perm1; vector_u8_t p0, p1, p; IDCT p0 = vec_lvsl (0, dest); p1 = vec_lvsl (stride, dest); p = vec_splat_u8 (-1); perm0 = vec_mergeh (p, p0); perm1 = vec_mergeh (p, p1); #define ADD(dest,src,perm) \ /* *(uint64_t *)&tmp = *(uint64_t *)dest; */ \ tmp = vec_ld (0, dest); \ tmp2 = (vector_s16_t)vec_perm (tmp, (vector_u8_t)zero, perm); \ tmp3 = vec_adds (tmp2, src); \ tmp = vec_packsu (tmp3, tmp3); \ vec_ste ((vector_u32_t)tmp, 0, (unsigned int *)dest); \ vec_ste ((vector_u32_t)tmp, 4, (unsigned int *)dest); ADD (dest, vx0, perm0) dest += stride; ADD (dest, vx1, perm1) dest += stride; ADD (dest, vx2, perm0) dest += stride; ADD (dest, vx3, perm1) dest += stride; ADD (dest, vx4, perm0) dest += stride; ADD (dest, vx5, perm1) dest += stride; ADD (dest, vx6, perm0) dest += stride; ADD (dest, vx7, perm1) memset (block, 0, 64 * sizeof (signed short)); } #endif /* __ALTIVEC__ */ --- NEW FILE --- /* * motion_comp_alpha.c * Copyright (C) 2002 Falk Hueffner <falk@debian.org> * * This file is part of mpeg2dec, a free MPEG-2 video stream decoder. * See http://libmpeg2.sourceforge.net/ for updates. * * mpeg2dec is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * mpeg2dec is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "config.h" #ifdef ARCH_ALPHA #include <inttypes.h> #include "mpeg2.h" #include "mpeg2_internal.h" #include "alpha_asm.h" static inline uint64_t avg2(uint64_t a, uint64_t b) { return (a | b) - (((a ^ b) & BYTE_VEC(0xfe)) >> 1); } // Load two unaligned quadwords from addr. This macro only works if // addr is actually unaligned. #define ULOAD16(ret_l, ret_r, addr) \ do { \ uint64_t _l = ldq_u(addr + 0); \ uint64_t _m = ldq_u(addr + 8); \ uint64_t _r = ldq_u(addr + 16); \ ret_l = extql(_l, addr) | extqh(_m, addr); \ ret_r = extql(_m, addr) | extqh(_r, addr); \ } while (0) // Load two aligned quadwords from addr. #define ALOAD16(ret_l, ret_r, addr) \ do { \ ret_l = ldq(addr); \ ret_r = ldq(addr + 8); \ } while (0) #define OP8(LOAD, LOAD16, STORE) \ do { \ STORE(LOAD(pixels), block); \ pixels += line_size; \ block += line_size; \ } while (--h) #define OP16(LOAD, LOAD16, STORE) \ do { \ uint64_t l, r; \ LOAD16(l, r, pixels); \ STORE(l, block); \ STORE(r, block + 8); \ pixels += line_size; \ block += line_size; \ } while (--h) #define OP8_X2(LOAD, LOAD16, STORE) \ do { \ uint64_t p0, p1; \ \ p0 = LOAD(pixels); \ p1 = p0 >> 8 | ((uint64_t) pixels[8] << 56); \ STORE(avg2(p0, p1), block); \ pixels += line_size; \ block += line_size; \ } while (--h) #define OP16_X2(LOAD, LOAD16, STORE) \ do { \ uint64_t p0, p1; \ \ LOAD16(p0, p1, pixels); \ STORE(avg2(p0, p0 >> 8 | p1 << 56), block); \ STORE(avg2(p1, p1 >> 8 | (uint64_t) pixels[16] << 56), \ block + 8); \ pixels += line_size; \ block += line_size; \ } while (--h) #define OP8_Y2(LOAD, LOAD16, STORE) \ do { \ uint64_t p0, p1; \ p0 = LOAD(pixels); \ pixels += line_size; \ p1 = LOAD(pixels); \ do { \ uint64_t av = avg2(p0, p1); \ if (--h == 0) line_size = 0; \ pixels += line_size; \ p0 = p1; \ p1 = LOAD(pixels); \ STORE(av, block); \ block += line_size; \ } while (h); \ } while (0) #define OP16_Y2(LOAD, LOAD16, STORE) \ do { \ uint64_t p0l, p0r, p1l, p1r; \ LOAD16(p0l, p0r, pixels); \ pixels += line_size; \ LOAD16(p1l, p1r, pixels); \ do { \ uint64_t avl, avr; \ if (--h == 0) line_size = 0; \ avl = avg2(p0l, p1l); \ avr = avg2(p0r, p1r); \ p0l = p1l; \ p0r = p1r; \ pixels += line_size; \ LOAD16(p1l, p1r, pixels); \ STORE(avl, block); \ STORE(avr, block + 8); \ block += line_size; \ } while (h); \ } while (0) #define OP8_XY2(LOAD, LOAD16, STORE) \ do { \ uint64_t pl, ph; \ uint64_t p1 = LOAD(pixels); \ uint64_t p2 = p1 >> 8 | ((uint64_t) pixels[8] << 56); \ \ ph = ((p1 & ~BYTE_VEC(0x03)) >> 2) \ + ((p2 & ~BYTE_VEC(0x03)) >> 2); \ pl = (p1 & BYTE_VEC(0x03)) \ + (p2 & BYTE_VEC(0x03)); \ \ do { \ uint64_t npl, nph; \ \ pixels += line_size; \ p1 = LOAD(pixels); \ p2 = (p1 >> 8) | ((uint64_t) pixels[8] << 56); \ nph = ((p1 & ~BYTE_VEC(0x03)) >> 2) \ + ((p2 & ~BYTE_VEC(0x03)) >> 2); \ npl = (p1 & BYTE_VEC(0x03)) \ + (p2 & BYTE_VEC(0x03)); \ \ STORE(ph + nph \ + (((pl + npl + BYTE_VEC(0x02)) >> 2) \ & BYTE_VEC(0x03)), block); \ \ block += line_size; \ pl = npl; \ ph = nph; \ } while (--h); \ } while (0) #define OP16_XY2(LOAD, LOAD16, STORE) \ do { \ uint64_t p0, p1, p2, p3, pl_l, ph_l, pl_r, ph_r; \ LOAD16(p0, p2, pixels); \ p1 = p0 >> 8 | (p2 << 56); \ p3 = p2 >> 8 | ((uint64_t) pixels[16] << 56); \ \ ph_l = ((p0 & ~BYTE_VEC(0x03)) >> 2) \ + ((p1 & ~BYTE_VEC(0x03)) >> 2); \ pl_l = (p0 & BYTE_VEC(0x03)) \ + (p1 & BYTE_VEC(0x03)); \ ph_r = ((p2 & ~BYTE_VEC(0x03)) >> 2) \ + ((p3 & ~BYTE_VEC(0x03)) >> 2); \ pl_r = (p2 & BYTE_VEC(0x03)) \ + (p3 & BYTE_VEC(0x03)); \ \ do { \ uint64_t npl_l, nph_l, npl_r, nph_r; \ \ pixels += line_size; \ LOAD16(p0, p2, pixels); \ p1 = p0 >> 8 | (p2 << 56); \ p3 = p2 >> 8 | ((uint64_t) pixels[16] << 56); \ nph_l = ((p0 & ~BYTE_VEC(0x03)) >> 2) \ + ((p1 & ~BYTE_VEC(0x03)) >> 2); \ npl_l = (p0 & BYTE_VEC(0x03)) \ + (p1 & BYTE_VEC(0x03)); \ nph_r = ((p2 & ~BYTE_VEC(0x03)) >> 2) \ + ((p3 & ~BYTE_VEC(0x03)) >> 2); \ npl_r = (p2 & BYTE_VEC(0x03)) \ + (p3 & BYTE_VEC(0x03)); \ \ STORE(ph_l + nph_l \ + (((pl_l + npl_l + BYTE_VEC(0x02)) >> 2) \ & BYTE_VEC(0x03)), block); \ STORE(ph_r + nph_r \ + (((pl_r + npl_r + BYTE_VEC(0x02)) >> 2) \ & BYTE_VEC(0x03)), block + 8); \ \ block += line_size; \ pl_l = npl_l; \ ph_l = nph_l; \ pl_r = npl_r; \ ph_r = nph_r; \ } while (--h); \ } while (0) #define MAKE_OP(OPNAME, SIZE, SUFF, OPKIND, STORE) \ static void MC_ ## OPNAME ## _ ## SUFF ## _ ## SIZE ## _alpha \ (uint8_t *restrict block, const uint8_t *restrict pixels, \ int line_size, int h) \ { \ if ((uint64_t) pixels & 0x7) { \ OPKIND(uldq, ULOAD16, STORE); \ } else { \ OPKIND(ldq, ALOAD16, STORE); \ } \ } #define PIXOP(OPNAME, STORE) \ MAKE_OP(OPNAME, 8, o, OP8, STORE); \ MAKE_OP(OPNAME, 8, x, OP8_X2, STORE); \ MAKE_OP(OPNAME, 8, y, OP8_Y2, STORE); \ MAKE_OP(OPNAME, 8, xy, OP8_XY2, STORE); \ MAKE_OP(OPNAME, 16, o, OP16, STORE); \ MAKE_OP(OPNAME, 16, x, OP16_X2, STORE); \ MAKE_OP(OPNAME, 16, y, OP16_Y2, STORE); \ MAKE_OP(OPNAME, 16, xy, OP16_XY2, STORE); #define STORE(l, b) stq(l, b) PIXOP(put, STORE); #undef STORE #define STORE(l, b) stq(avg2(l, ldq(b)), b); PIXOP(avg, STORE); mpeg2_mc_t mpeg2_mc_alpha = { { MC_put_o_16_alpha, MC_put_x_16_alpha, MC_put_y_16_alpha, MC_put_xy_16_alpha, MC_put_o_8_alpha, MC_put_x_8_alpha, MC_put_y_8_alpha, MC_put_xy_8_alpha }, { MC_avg_o_16_alpha, MC_avg_x_16_alpha, MC_avg_y_16_alpha, MC_avg_xy_16_alpha, MC_avg_o_8_alpha, MC_avg_x_8_alpha, MC_avg_y_8_alpha, MC_avg_xy_8_alpha } }; #endif --- NEW FILE --- /* * motion_comp_altivec.c * Copyright (C) 2000-2002 Michel Lespinasse <walken@zoy.org> * Copyright (C) 1999-2000 Aaron Holtzman <aholtzma@ess.engr.uvic.ca> * * This file is part of mpeg2dec, a free MPEG-2 video stream decoder. * See http://libmpeg2.sourceforge.net/ for updates. * * mpeg2dec is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * mpeg2dec is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License [...1980 lines suppressed...] } while (--height); ref0 = vec_ld (0, ref); ref1 = vec_ld (8, ref); prev = vec_ld (stride, dest); vec_ste ((vector_u32_t)tmp, 0, (unsigned int *)dest); vec_ste ((vector_u32_t)tmp, 4, (unsigned int *)dest); dest += stride; A = vec_perm (ref0, ref1, perm0A); B = vec_perm (ref0, ref1, perm0B); avg0 = vec_avg (A, B); xor0 = vec_xor (A, B); tmp = vec_avg (prev, vec_sub (vec_avg (avg0, avg1), vec_and (vec_and (ones, vec_or (xor0, xor1)), vec_xor (avg0, avg1)))); vec_ste ((vector_u32_t)tmp, 0, (unsigned int *)dest); vec_ste ((vector_u32_t)tmp, 4, (unsigned int *)dest); } #endif /* __ALTIVEC__ */
participants (1)
-
Arpi of Ize