[FFmpeg-devel] [PATCH] This patch addresses Trac ticket #5570. The optimized functions are in file libswscale/ppc/input_vsx.c. Each optimized function name is a concatenation of the corresponding name in libswscale/input.c with suffix _vsx. From bountysource https://www.bountysource.com/issues/34315232-power8-vsx-vectorization-libswscale-input-c (reverse_forever)
Вячеслав
pestov.vyach at yandex.ru
Fri Mar 29 16:00:38 EET 2019
---
libswscale/ppc/Makefile | 3 +-
libswscale/ppc/input_vsx.c | 3801 +++++++++++++++++++++++++++++++++++++++++
libswscale/swscale.c | 3 +
libswscale/swscale_internal.h | 1 +
4 files changed, 3807 insertions(+), 1 deletion(-)
create mode 100644 libswscale/ppc/input_vsx.c
diff --git a/libswscale/ppc/Makefile b/libswscale/ppc/Makefile
index 0a31a3025b..4d821240be 100644
--- a/libswscale/ppc/Makefile
+++ b/libswscale/ppc/Makefile
@@ -1,4 +1,5 @@
OBJS += ppc/swscale_altivec.o \
ppc/yuv2rgb_altivec.o \
ppc/yuv2yuv_altivec.o \
- ppc/swscale_vsx.o
+ ppc/swscale_vsx.o \
+ ppc/input_vsx.o
diff --git a/libswscale/ppc/input_vsx.c b/libswscale/ppc/input_vsx.c
new file mode 100644
index 0000000000..b652ed17ca
--- /dev/null
+++ b/libswscale/ppc/input_vsx.c
@@ -0,0 +1,3801 @@
+/*
+ * POWER8 VSX vectorization libswscale/input.c
+ * Written by Vyacheslav Pestov.
+ *
+ * This file is part of FFmpeg.
+ *
+ * FFmpeg is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * FFmpeg is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with FFmpeg; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+/**
+ * @file
+ * POWER8 VSX vectorization libswscale/input.c
+ * @author Vyacheslav Pestov
+ */
+#include <math.h>
+#include <stdint.h>
+#include <stdio.h>
+#include <string.h>
+
+
+#include "libavutil/avutil.h"
+#include "libavutil/bswap.h"
+#include "libavutil/cpu.h"
+#include "libavutil/intreadwrite.h"
+#include "libavutil/mathematics.h"
+#include "libavutil/pixdesc.h"
+#include "libavutil/avassert.h"
+#include "libavutil/ppc/util_altivec.h"
+#include "config.h"
+#include "../rgb2rgb.h"
+#include "../swscale.h"
+#include "../swscale_internal.h"
+
+
+
+#if HAVE_VSX
+#if !HAVE_BIGENDIAN
+
+#define input_pixel(pos) (isBE(origin) ? AV_RB16(pos) : AV_RL16(pos))
+
+#define r ((origin == AV_PIX_FMT_BGR48BE || origin == AV_PIX_FMT_BGR48LE || \
+ origin == AV_PIX_FMT_BGRA64BE || origin == AV_PIX_FMT_BGRA64LE) \
+ ? b_r : r_b)
+#define b ((origin == AV_PIX_FMT_BGR48BE || origin == AV_PIX_FMT_BGR48LE || \
+ origin == AV_PIX_FMT_BGRA64BE || origin == AV_PIX_FMT_BGRA64LE) \
+ ? r_b : b_r)
+#define v_r1 ((origin == AV_PIX_FMT_BGR48BE || origin == AV_PIX_FMT_BGR48LE || \
+ origin == AV_PIX_FMT_BGRA64BE || origin == AV_PIX_FMT_BGRA64LE) \
+ ? v_b_r1 : v_r_b1)
+#define v_b1 ((origin == AV_PIX_FMT_BGR48BE || origin == AV_PIX_FMT_BGR48LE || \
+ origin == AV_PIX_FMT_BGRA64BE || origin == AV_PIX_FMT_BGRA64LE) \
+ ? v_r_b1 : v_b_r1)
+#define v_r2 ((origin == AV_PIX_FMT_BGR48BE || origin == AV_PIX_FMT_BGR48LE || \
+ origin == AV_PIX_FMT_BGRA64BE || origin == AV_PIX_FMT_BGRA64LE) \
+ ? v_b_r2 : v_r_b2)
+#define v_b2 ((origin == AV_PIX_FMT_BGR48BE || origin == AV_PIX_FMT_BGR48LE || \
+ origin == AV_PIX_FMT_BGRA64BE || origin == AV_PIX_FMT_BGRA64LE) \
+ ? v_r_b2 : v_b_r2)
+
+static av_always_inline void
+rgb64ToY_c_template_vsx(uint16_t *dst, const uint16_t *src, int width,
+ enum AVPixelFormat origin, int32_t *rgb2yuv)
+{
+ int i, width_adj, is_BE;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_rd3, v_nul;
+ vector unsigned short v_b_r1, v_b_r2, v_r_b1, v_r_b2, v_g1, v_g2;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ vector signed int v_ry, v_gy, v_by;
+ int32_t ry = rgb2yuv[RY_IDX], gy = rgb2yuv[GY_IDX], by = rgb2yuv[BY_IDX];
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x2001<<(RGB2YUV_SHIFT-1)));
+ shift2 = vec_splats((unsigned int)RGB2YUV_SHIFT);
+ v_ry = vec_splats((signed int)ry);
+ v_gy = vec_splats((signed int)gy);
+ v_by = vec_splats((signed int)by);
+ is_BE = isBE(origin);
+ v_nul = vec_splats((unsigned short)0x0000);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr+16));
+ v_rd2 = vec_vsx_ld(0, (unsigned short *)(src_addr+32));
+ v_rd3 = vec_vsx_ld(0, (unsigned short *)(src_addr+48));
+
+ if(is_BE){
+ v_r_b1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){1, 0, 9, 8, 17, 16, 25, 24}));
+ v_r_b2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){1, 0, 9, 8, 17, 16, 25, 24}));
+ v_g1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){3, 2, 11, 10, 19, 18, 27, 26}));
+ v_g2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){3, 2, 11, 10, 19, 18, 27, 26}));
+ v_b_r1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){5, 4, 13, 12, 21, 20, 29, 28}));
+ v_b_r2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){5, 4, 13, 12, 21, 20, 29, 28}));
+ }else{
+ v_r_b1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 1, 8, 9, 16, 17, 24, 25}));
+ v_r_b2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){0, 1, 8, 9, 16, 17, 24, 25}));
+ v_g1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){2, 3, 10, 11, 18, 19, 26, 27}));
+ v_g2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){2, 3, 10, 11, 18, 19, 26, 27}));
+ v_b_r1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){4, 5, 12, 13, 20, 21, 28, 29}));
+ v_b_r2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){4, 5, 12, 13, 20, 21, 28, 29}));
+ }
+
+
+ v_r_b1 = vec_mergeh(v_r_b1, v_nul);
+ v_g1 = vec_mergeh(v_g1, v_nul);
+ v_b_r1 = vec_mergeh(v_b_r1, v_nul);
+
+ v_r_b2 = vec_mergeh(v_r_b2, v_nul);
+ v_g2 = vec_mergeh(v_g2, v_nul);
+ v_b_r2 = vec_mergeh(v_b_r2, v_nul);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_ry);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gy ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_by ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r2, v_ry);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g2, v_gy ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b2, v_by ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2, ((vector unsigned char)
+ {0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dst_addr);
+
+ src_addr += 64;
+ dst_addr += 16;
+ }
+
+ for (i = width_adj; i < width; i++) {
+ unsigned int r_b = input_pixel(&src[i*4+0]);
+ unsigned int g = input_pixel(&src[i*4+1]);
+ unsigned int b_r = input_pixel(&src[i*4+2]);
+
+ dst[i] = (ry*r + gy*g + by*b + (0x2001<<(RGB2YUV_SHIFT-1))) >> RGB2YUV_SHIFT;
+ }
+}
+
+static av_always_inline void
+rgb64ToUV_c_template_vsx(uint16_t *dstU, uint16_t *dstV,
+ const uint16_t *src1, const uint16_t *src2,
+ int width, enum AVPixelFormat origin, int32_t *rgb2yuv)
+{
+
+ int i, width_adj, is_BE ;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_rd3, v_nul;
+ vector unsigned short v_b_r1, v_b_r2, v_r_b1, v_r_b2, v_g1, v_g2;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ vector signed int v_ru, v_gu, v_bu, v_rv, v_gv, v_bv;
+ int32_t ru = rgb2yuv[RU_IDX], gu = rgb2yuv[GU_IDX], bu = rgb2yuv[BU_IDX];
+ int32_t rv = rgb2yuv[RV_IDX], gv = rgb2yuv[GV_IDX], bv = rgb2yuv[BV_IDX];
+ av_assert1(src1==src2);
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x10001<<(RGB2YUV_SHIFT-1)));
+ shift2 = vec_splats((unsigned int)RGB2YUV_SHIFT);
+ v_ru = vec_splats((signed int)ru);
+ v_gu = vec_splats((signed int)gu);
+ v_bu = vec_splats((signed int)bu);
+ v_rv = vec_splats((signed int)rv);
+ v_gv = vec_splats((signed int)gv);
+ v_bv = vec_splats((signed int)bv);
+ is_BE = isBE(origin);
+ v_nul = vec_splats((unsigned short)0x0000);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr+16));
+ v_rd2 = vec_vsx_ld(0, (unsigned short *)(src_addr+32));
+ v_rd3 = vec_vsx_ld(0, (unsigned short *)(src_addr+48));
+
+ if(is_BE){
+ v_r_b1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){1, 0, 9, 8, 17, 16, 25, 24}));
+ v_r_b2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){1, 0, 9, 8, 17, 16, 25, 24}));
+ v_g1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){3, 2, 11, 10, 19, 18, 27, 26}));
+ v_g2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){3, 2, 11, 10, 19, 18, 27, 26}));
+ v_b_r1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){5, 4, 13, 12, 21, 20, 29, 28}));
+ v_b_r2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){5, 4, 13, 12, 21, 20, 29, 28}));
+ }else{
+ v_r_b1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 1, 8, 9, 16, 17, 24, 25}));
+ v_r_b2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){0, 1, 8, 9, 16, 17, 24, 25}));
+ v_g1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){2, 3, 10, 11, 18, 19, 26, 27}));
+ v_g2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){2, 3, 10, 11, 18, 19, 26, 27}));
+ v_b_r1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){4, 5, 12, 13, 20, 21, 28, 29}));
+ v_b_r2 = vec_perm(v_rd2, v_rd3,
+ ((vector unsigned char){4, 5, 12, 13, 20, 21, 28, 29}));
+ }
+
+
+ v_r_b1 = vec_mergeh(v_r_b1, v_nul);
+ v_g1 = vec_mergeh(v_g1, v_nul);
+ v_b_r1 = vec_mergeh(v_b_r1, v_nul);
+
+ v_r_b2 = vec_mergeh(v_r_b2, v_nul);
+ v_g2 = vec_mergeh(v_g2, v_nul);
+ v_b_r2 = vec_mergeh(v_b_r2, v_nul);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_ru);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gu ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_bu ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r2, v_ru);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g2, v_gu ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b2, v_bu ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstU_addr);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_rv);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gv ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_bv ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r2, v_rv);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g2, v_gv ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b2, v_bv ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 64;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+ for (i = width_adj; i < width; i++) {
+ int r_b = input_pixel(&src1[i*4+0]);
+ int g = input_pixel(&src1[i*4+1]);
+ int b_r = input_pixel(&src1[i*4+2]);
+
+ dstU[i] = (ru*r + gu*g + bu*b + (0x10001<<(RGB2YUV_SHIFT-1))) >> RGB2YUV_SHIFT;
+ dstV[i] = (rv*r + gv*g + bv*b + (0x10001<<(RGB2YUV_SHIFT-1))) >> RGB2YUV_SHIFT;
+ }
+}
+
+
+#define rgb64funcs(pattern, BE_LE, origin) \
+static void pattern ## 64 ## BE_LE ## ToY_c_vsx(uint8_t *_dst, const uint8_t *_src, \
+ const uint8_t *unused0, const uint8_t *unused1, \
+ int width, uint32_t *rgb2yuv) \
+{ \
+ const uint16_t *src = (const uint16_t *) _src; \
+ uint16_t *dst = (uint16_t *) _dst; \
+ rgb64ToY_c_template_vsx(dst, src, width, origin, rgb2yuv); \
+} \
+ \
+static void pattern ## 64 ## BE_LE ## ToUV_c_vsx(uint8_t *_dstU, uint8_t *_dstV, \
+ const uint8_t *unused0, const uint8_t *_src1, const uint8_t *_src2, \
+ int width, uint32_t *rgb2yuv) \
+{ \
+ const uint16_t *src1 = (const uint16_t *) _src1, \
+ *src2 = (const uint16_t *) _src2; \
+ uint16_t *dstU = (uint16_t *) _dstU, *dstV = (uint16_t *) _dstV; \
+ rgb64ToUV_c_template_vsx(dstU, dstV, src1, src2, width, origin, rgb2yuv); \
+}
+
+rgb64funcs(rgb, LE, AV_PIX_FMT_RGBA64LE)
+rgb64funcs(rgb, BE, AV_PIX_FMT_RGBA64BE)
+rgb64funcs(bgr, LE, AV_PIX_FMT_BGRA64LE)
+rgb64funcs(bgr, BE, AV_PIX_FMT_BGRA64BE)
+
+static av_always_inline void rgb48ToY_c_template_vsx(uint16_t *dst,
+ const uint16_t *src, int width,
+ enum AVPixelFormat origin,
+ int32_t *rgb2yuv)
+{
+ int i, width_adj, is_BE;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_nul;
+ vector unsigned short v_b_r1, v_b_r2, v_r_b1, v_r_b2, v_g1, v_g2;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ vector signed int v_ry, v_gy, v_by;
+ int32_t ry = rgb2yuv[RY_IDX], gy = rgb2yuv[GY_IDX], by = rgb2yuv[BY_IDX];
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x2001<<(RGB2YUV_SHIFT-1)));
+ shift2 = vec_splats((unsigned int)RGB2YUV_SHIFT);
+ v_ry = vec_splats((signed int)ry);
+ v_gy = vec_splats((signed int)gy);
+ v_by = vec_splats((signed int)by);
+ is_BE = isBE(origin);
+ v_nul = vec_splats((unsigned short)0x0000);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr+16));
+ v_rd2 = vec_vsx_ld(0, (unsigned short *)(src_addr+32));
+
+ if(is_BE){
+ v_r_b1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){1, 0, 7, 6, 13, 12, 19, 18}));
+ v_r_b2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){9, 8, 15, 14, 21, 20, 27, 26}));
+ v_g1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){3, 2, 9, 8, 15, 14, 21, 20}));
+ v_g2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){11, 10, 17, 16, 23, 22, 29, 28}));
+ v_b_r1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){5, 4, 11, 10, 17, 16, 23, 22}));
+ v_b_r2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){13, 12, 19, 18, 25, 24, 31, 30}));
+ }else{
+ v_r_b1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 1, 6, 7, 12, 13, 18, 19}));
+ v_r_b2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){8, 9, 14, 15, 20, 21, 26, 27}));
+ v_g1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){2, 3, 8, 9, 14, 15, 20, 21}));
+ v_g2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){10, 11, 16, 17, 22, 23, 28, 29}));
+ v_b_r1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){4, 5, 10, 11, 16, 17, 22, 23}));
+ v_b_r2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){12, 13, 18, 19, 24, 25, 30, 31}));
+ }
+
+
+ v_r_b1 = vec_mergeh(v_r_b1, v_nul);
+ v_g1 = vec_mergeh(v_g1, v_nul);
+ v_b_r1 = vec_mergeh(v_b_r1, v_nul);
+
+ v_r_b2 = vec_mergeh(v_r_b2, v_nul);
+ v_g2 = vec_mergeh(v_g2, v_nul);
+ v_b_r2 = vec_mergeh(v_b_r2, v_nul);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_ry);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gy ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_by ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r2, v_ry);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g2, v_gy ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b2, v_by ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dst_addr);
+
+ src_addr += 48;
+ dst_addr += 16;
+ }
+
+ for (i = 0; i < width; i++) {
+ unsigned int r_b = input_pixel(&src[i * 3 + 0]);
+ unsigned int g = input_pixel(&src[i * 3 + 1]);
+ unsigned int b_r = input_pixel(&src[i * 3 + 2]);
+
+ dst[i] = (ry*r + gy*g + by*b + (0x2001 << (RGB2YUV_SHIFT - 1))) >> RGB2YUV_SHIFT;
+ }
+}
+
+static av_always_inline void rgb48ToUV_c_template_vsx(uint16_t *dstU,
+ uint16_t *dstV,
+ const uint16_t *src1,
+ const uint16_t *src2,
+ int width,
+ enum AVPixelFormat origin,
+ int32_t *rgb2yuv)
+{
+
+ int i, width_adj, is_BE ;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_nul;
+ vector unsigned short v_b_r1, v_b_r2, v_r_b1, v_r_b2, v_g1, v_g2;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ vector signed int v_ru, v_gu, v_bu, v_rv, v_gv, v_bv;
+ int32_t ru = rgb2yuv[RU_IDX], gu = rgb2yuv[GU_IDX], bu = rgb2yuv[BU_IDX],
+ rv = rgb2yuv[RV_IDX], gv = rgb2yuv[GV_IDX], bv = rgb2yuv[BV_IDX];
+ av_assert1(src1==src2);
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x10001<<(RGB2YUV_SHIFT-1)));
+ shift2 = vec_splats((unsigned int)RGB2YUV_SHIFT);
+ v_ru = vec_splats((signed int)ru);
+ v_gu = vec_splats((signed int)gu);
+ v_bu = vec_splats((signed int)bu);
+ v_rv = vec_splats((signed int)rv);
+ v_gv = vec_splats((signed int)gv);
+ v_bv = vec_splats((signed int)bv);
+ is_BE = isBE(origin);
+ v_nul = vec_splats((unsigned short)0x0000);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr+16));
+ v_rd2 = vec_vsx_ld(0, (unsigned short *)(src_addr+32));
+
+ if(is_BE){
+ v_r_b1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){1, 0, 7, 6, 13, 12, 19, 18}));
+ v_r_b2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){9, 8, 15, 14, 21, 20, 27, 26}));
+ v_g1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){3, 2, 9, 8, 15, 14, 21, 20}));
+ v_g2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){11, 10, 17, 16, 23, 22, 29, 28}));
+ v_b_r1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){5, 4, 11, 10, 17, 16, 23, 22}));
+ v_b_r2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){13, 12, 19, 18, 25, 24, 31, 30}));
+ }else{
+ v_r_b1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 1, 6, 7, 12, 13, 18, 19}));
+ v_r_b2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){8, 9, 14, 15, 20, 21, 26, 27}));
+ v_g1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){2, 3, 8, 9, 14, 15, 20, 21}));
+ v_g2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){10, 11, 16, 17, 22, 23, 28, 29}));
+ v_b_r1 = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){4, 5, 10, 11, 16, 17, 22, 23}));
+ v_b_r2 = vec_perm(v_rd1, v_rd2,
+ ((vector unsigned char){12, 13, 18, 19, 24, 25, 30, 31}));
+ }
+
+
+ v_r_b1 = vec_mergeh(v_r_b1, v_nul);
+ v_g1 = vec_mergeh(v_g1, v_nul);
+ v_b_r1 = vec_mergeh(v_b_r1, v_nul);
+
+ v_r_b2 = vec_mergeh(v_r_b2, v_nul);
+ v_g2 = vec_mergeh(v_g2, v_nul);
+ v_b_r2 = vec_mergeh(v_b_r2, v_nul);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_ru);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gu ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_bu ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r2, v_ru);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g2, v_gu ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b2, v_bu ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstU_addr);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_rv);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gv ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_bv ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r2, v_rv);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g2, v_gv ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b2, v_bv ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 48;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+ for (i = width_adj; i < width; i++) {
+ int r_b = input_pixel(&src1[i * 3 + 0]);
+ int g = input_pixel(&src1[i * 3 + 1]);
+ int b_r = input_pixel(&src1[i * 3 + 2]);
+
+ dstU[i] = (ru*r + gu*g + bu*b + (0x10001 << (RGB2YUV_SHIFT - 1))) >> RGB2YUV_SHIFT;
+ dstV[i] = (rv*r + gv*g + bv*b + (0x10001 << (RGB2YUV_SHIFT - 1))) >> RGB2YUV_SHIFT;
+ }
+}
+#undef r
+#undef b
+#undef v_r1
+#undef v_b1
+#undef v_r2
+#undef v_b2
+#undef input_pixel
+
+#define rgb48funcs(pattern, BE_LE, origin) \
+static void pattern ## 48 ## BE_LE ## ToY_c_vsx(uint8_t *_dst, \
+ const uint8_t *_src, \
+ const uint8_t *unused0, const uint8_t *unused1,\
+ int width, \
+ uint32_t *rgb2yuv) \
+{ \
+ const uint16_t *src = (const uint16_t *)_src; \
+ uint16_t *dst = (uint16_t *)_dst; \
+ rgb48ToY_c_template_vsx(dst, src, width, origin, rgb2yuv); \
+} \
+ \
+static void pattern ## 48 ## BE_LE ## ToUV_c_vsx(uint8_t *_dstU, \
+ uint8_t *_dstV, \
+ const uint8_t *unused0, \
+ const uint8_t *_src1, \
+ const uint8_t *_src2, \
+ int width, \
+ uint32_t *rgb2yuv) \
+{ \
+ const uint16_t *src1 = (const uint16_t *)_src1, \
+ *src2 = (const uint16_t *)_src2; \
+ uint16_t *dstU = (uint16_t *)_dstU, \
+ *dstV = (uint16_t *)_dstV; \
+ rgb48ToUV_c_template_vsx(dstU, dstV, src1, src2, width, origin, rgb2yuv); \
+}
+
+rgb48funcs(rgb, LE, AV_PIX_FMT_RGB48LE)
+rgb48funcs(rgb, BE, AV_PIX_FMT_RGB48BE)
+rgb48funcs(bgr, LE, AV_PIX_FMT_BGR48LE)
+rgb48funcs(bgr, BE, AV_PIX_FMT_BGR48BE)
+
+#define input_pixel(i) ((origin == AV_PIX_FMT_RGBA || \
+ origin == AV_PIX_FMT_BGRA || \
+ origin == AV_PIX_FMT_ARGB || \
+ origin == AV_PIX_FMT_ABGR) \
+ ? AV_RN32A(&src[(i) * 4]) \
+ : (isBE(origin) ? AV_RB16(&src[(i) * 2]) \
+ : AV_RL16(&src[(i) * 2])))
+
+static av_always_inline void rgb16_32ToY_c_template_vsx(int16_t *dst,
+ const uint8_t *src,
+ int width,
+ enum AVPixelFormat origin,
+ int shr, int shg,
+ int shb, int shp,
+ int maskr, int maskg,
+ int maskb, int rsh,
+ int gsh, int bsh, int S,
+ int32_t *rgb2yuv)
+{
+ int i, width_adj, is_DW, is_BE;
+ vector unsigned short v_rd0, v_rd1, v_px, v_nul;
+ vector unsigned short v_r1, v_r2, v_b1, v_b2, v_g1, v_g2;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ const int ry = rgb2yuv[RY_IDX]<<rsh, gy = rgb2yuv[GY_IDX]<<gsh,
+ by = rgb2yuv[BY_IDX]<<bsh;
+ const unsigned rnd = (32<<((S)-1)) + (1<<(S-7));
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)rnd);
+ shift2 = vec_splats((unsigned int)((S)-6));
+ is_DW = (origin == AV_PIX_FMT_RGBA || origin == AV_PIX_FMT_BGRA ||
+ origin == AV_PIX_FMT_ARGB || origin == AV_PIX_FMT_ABGR);
+ if(is_DW)
+ is_BE = isBE(origin);
+ v_nul = vec_splats((unsigned short)0x0000);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+
+ if(is_DW){
+ src_addr += 16;
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_px = (vector unsigned short)vec_sr((vector unsigned int)v_rd0,
+ vec_splats((unsigned int)shp));
+ v_b1 = (vector unsigned short)vec_and((vector unsigned int)v_px,
+ vec_splats((unsigned int)maskb));
+ v_b1 = (vector unsigned short)vec_sr((vector unsigned int)v_b1,
+ vec_splats((unsigned int)shb));
+ v_g1 = (vector unsigned short)vec_and((vector unsigned int)v_px,
+ vec_splats((unsigned int)maskg));
+ v_g1 = (vector unsigned short)vec_sr((vector unsigned int)v_g1,
+ vec_splats((unsigned int)shg));
+ v_r1 = (vector unsigned short)vec_and((vector unsigned int)v_px,
+ vec_splats((unsigned int)maskr));
+ v_r1 = (vector unsigned short)vec_sr((vector unsigned int)v_r1,
+ vec_splats((unsigned int)shr));
+
+ v_px = (vector unsigned short)vec_sr((vector unsigned int)v_rd1,
+ vec_splats((unsigned int)shp));
+ v_b2 = (vector unsigned short)vec_and((vector unsigned int)v_px,
+ vec_splats((unsigned int)maskb));
+ v_b2 = (vector unsigned short)vec_sr((vector unsigned int)v_b2,
+ vec_splats((unsigned int)shb));
+ v_g2 = (vector unsigned short)vec_and((vector unsigned int)v_px,
+ vec_splats((unsigned int)maskg));
+ v_g2 = (vector unsigned short)vec_sr((vector unsigned int)v_g2,
+ vec_splats((unsigned int)shg));
+ v_r2 = (vector unsigned short)vec_and((vector unsigned int)v_px,
+ vec_splats((unsigned int)maskr));
+ v_r2 = (vector unsigned short)vec_sr((vector unsigned int)v_r2,
+ vec_splats((unsigned int)shr));
+ }else{
+ if(is_BE){
+ v_px = vec_perm(v_rd0, v_rd0,
+ ((vector unsigned char){1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14}));
+ v_px = vec_sr(v_px, vec_splats((unsigned short)shp));
+ v_b1 = vec_and(v_px, vec_splats((unsigned short)maskb));
+ v_b1 = vec_sr(v_b1, vec_splats((unsigned short)shb));
+ v_g1 = vec_and(v_px, vec_splats((unsigned short)maskg));
+ v_g1 = vec_sr(v_g1, vec_splats((unsigned short)shg));
+ v_r1 = vec_and(v_px, vec_splats((unsigned short)maskr));
+ v_r1 = vec_sr(v_r1, vec_splats((unsigned short)shr));
+ }else{
+ v_px = vec_sr(v_rd0, vec_splats((unsigned short)shp));
+ v_b1 = vec_and(v_px, vec_splats((unsigned short)maskb));
+ v_b1 = vec_sr(v_b1, vec_splats((unsigned short)shb));
+ v_g1 = vec_and(v_px, vec_splats((unsigned short)maskg));
+ v_g1 = vec_sr(v_g1, vec_splats((unsigned short)shg));
+ v_r1 = vec_and(v_px, vec_splats((unsigned short)maskr));
+ v_r1 = vec_sr(v_r1, vec_splats((unsigned short)shr));
+
+ }
+ v_b2 = v_b1;
+ v_g2 = v_g1;
+ v_r2 = v_r1;
+ v_b1 = vec_mergeh(v_b1, v_nul);
+ v_g1 = vec_mergeh(v_g1, v_nul);
+ v_r1 = vec_mergeh(v_r1, v_nul);
+ v_b2 = vec_mergel(v_b2, v_nul);
+ v_g2 = vec_mergel(v_g2, v_nul);
+ v_r2 = vec_mergel(v_r2, v_nul);
+ }
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1,
+ vec_splats((signed int)ry));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1,
+ vec_splats((signed int)gy)) );
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1,
+ vec_splats((signed int)by)) );
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r2,
+ vec_splats((signed int)ry));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g2,
+ vec_splats((signed int)gy)) );
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b2,
+ vec_splats((signed int)by)) );
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dst_addr);
+
+ dst_addr += 16;
+ src_addr += 16;
+
+ }
+
+ for (i = width_adj; i < width; i++) {
+ int px = input_pixel(i) >> shp;
+ int b = (px & maskb) >> shb;
+ int g = (px & maskg) >> shg;
+ int r = (px & maskr) >> shr;
+
+ dst[i] = (ry * r + gy * g + by * b + rnd) >> ((S)-6);
+ }
+}
+
+
+static av_always_inline void rgb16_32ToUV_c_template_vsx(int16_t *dstU,
+ int16_t *dstV,
+ const uint8_t *src,
+ int width,
+ enum AVPixelFormat origin,
+ int shr, int shg,
+ int shb, int shp,
+ int maskr, int maskg,
+ int maskb, int rsh,
+ int gsh, int bsh, int S,
+ int32_t *rgb2yuv)
+{
+ int i, width_adj, is_DW, is_BE;
+ vector unsigned short v_rd0, v_rd1, v_px, v_nul;
+ vector unsigned short v_r1, v_r2, v_b1, v_b2, v_g1, v_g2;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ const int ru = rgb2yuv[RU_IDX] << rsh, gu = rgb2yuv[GU_IDX] << gsh,
+ bu = rgb2yuv[BU_IDX] << bsh, rv = rgb2yuv[RV_IDX] << rsh,
+ gv = rgb2yuv[GV_IDX] << gsh, bv = rgb2yuv[BV_IDX] << bsh;
+ const unsigned rnd = (256u<<((S)-1)) + (1<<(S-7));
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)rnd);
+ shift2 = vec_splats((unsigned int)((S)-6));
+ is_DW = (origin == AV_PIX_FMT_RGBA || origin == AV_PIX_FMT_BGRA ||
+ origin == AV_PIX_FMT_ARGB || origin == AV_PIX_FMT_ABGR);
+ if(!is_DW)
+ is_BE = isBE(origin);
+ else
+ v_nul = vec_splats((unsigned short)0x0000);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+
+ if(is_DW){
+ src_addr += 16;
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd0 = (vector unsigned short)vec_sr((vector unsigned int)v_rd0,
+ vec_splats((unsigned int)shp));
+ v_b1 = (vector unsigned short)vec_and((vector unsigned int)v_rd0,
+ vec_splats((unsigned int)maskb));
+ v_b1 = (vector unsigned short)vec_sr((vector unsigned int)v_b1,
+ vec_splats((unsigned int)shb));
+ v_g1 = (vector unsigned short)vec_and((vector unsigned int)v_rd0,
+ vec_splats((unsigned int)maskg));
+ v_g1 = (vector unsigned short)vec_sr((vector unsigned int)v_g1,
+ vec_splats((unsigned int)shg));
+ v_r1 = (vector unsigned short)vec_and((vector unsigned int)v_rd0,
+ vec_splats((unsigned int)maskr));
+ v_r1 = (vector unsigned short)vec_sr((vector unsigned int)v_r1,
+ vec_splats((unsigned int)shr));
+
+ v_rd1 = (vector unsigned short)vec_sr((vector unsigned int)v_rd1,
+ vec_splats((unsigned int)shp));
+ v_b2 = (vector unsigned short)vec_and((vector unsigned int)v_rd1,
+ vec_splats((unsigned int)maskb));
+ v_b2 = (vector unsigned short)vec_sr((vector unsigned int)v_b2,
+ vec_splats((unsigned int)shb));
+ v_g2 = (vector unsigned short)vec_and((vector unsigned int)v_rd1,
+ vec_splats((unsigned int)maskg));
+ v_g2 = (vector unsigned short)vec_sr((vector unsigned int)v_g2,
+ vec_splats((unsigned int)shg));
+ v_r2 = (vector unsigned short)vec_and((vector unsigned int)v_rd1,
+ vec_splats((unsigned int)maskr));
+ v_r2 = (vector unsigned short)vec_sr((vector unsigned int)v_r2,
+ vec_splats((unsigned int)shr));
+ }else{
+ if(is_BE){
+ v_px = vec_perm(v_rd0, v_rd0,
+ ((vector unsigned char){1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14}));
+ v_px = vec_sr(v_px, vec_splats((unsigned short)shp));
+ v_b1 = vec_and(v_px, vec_splats((unsigned short)maskb));
+ v_b1 = vec_sr(v_b1, vec_splats((unsigned short)shb));
+ v_g1 = vec_and(v_px, vec_splats((unsigned short)maskg));
+ v_g1 = vec_sr(v_g1, vec_splats((unsigned short)shg));
+ v_r1 = vec_and(v_px, vec_splats((unsigned short)maskr));
+ v_r1 = vec_sr(v_r1, vec_splats((unsigned short)shr));
+ }else{
+ v_px = vec_sr(v_rd0, vec_splats((unsigned short)shp));
+ v_b1 = vec_and(v_px, vec_splats((unsigned short)maskb));
+ v_b1 = vec_sr(v_b1, vec_splats((unsigned short)shb));
+ v_g1 = vec_and(v_px, vec_splats((unsigned short)maskg));
+ v_g1 = vec_sr(v_g1, vec_splats((unsigned short)shg));
+ v_r1 = vec_and(v_px, vec_splats((unsigned short)maskr));
+ v_r1 = vec_sr(v_r1, vec_splats((unsigned short)shr));
+
+ }
+ v_b2 = v_b1;
+ v_g2 = v_g1;
+ v_r2 = v_r1;
+ v_b1 = vec_mergeh(v_b1, v_nul);
+ v_g1 = vec_mergeh(v_g1, v_nul);
+ v_r1 = vec_mergeh(v_r1, v_nul);
+ v_b2 = vec_mergel(v_b2, v_nul);
+ v_g2 = vec_mergel(v_g2, v_nul);
+ v_r2 = vec_mergel(v_r2, v_nul);
+ }
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1,
+ vec_splats((signed int)ru));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1,
+ vec_splats((signed int)gu) ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1,
+ vec_splats((signed int)bu) ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r2,
+ vec_splats((signed int)ru));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g2,
+ vec_splats((signed int)gu) ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b2,
+ vec_splats((signed int)bu) ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstU_addr);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1,
+ vec_splats((signed int)rv));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1,
+ vec_splats((signed int)gv) ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1,
+ vec_splats((signed int)bv) ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r2, vec_splats((signed int)rv));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g2,
+ vec_splats((signed int)gv) ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b2,
+ vec_splats((signed int)bv) ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstV_addr);
+
+ dstU_addr += 16;
+ dstV_addr += 16;
+ src_addr += 16;
+
+ }
+
+ for (i = width_adj; i < width; i++) {
+ int px = input_pixel(i) >> shp;
+ int b = (px & maskb) >> shb;
+ int g = (px & maskg) >> shg;
+ int r = (px & maskr) >> shr;
+
+ dstU[i] = (ru * r + gu * g + bu * b + rnd) >> ((S)-6);
+ dstV[i] = (rv * r + gv * g + bv * b + rnd) >> ((S)-6);
+ }
+}
+
+#undef input_pixel
+
+#define rgb16_32_wrapper(fmt, name, shr, shg, shb, shp, maskr, \
+ maskg, maskb, rsh, gsh, bsh, S) \
+static void name ## ToY_c_vsx(uint8_t *dst, const uint8_t *src, const uint8_t *unused1, const uint8_t *unused2, \
+ int width, uint32_t *tab) \
+{ \
+ rgb16_32ToY_c_template_vsx((int16_t*)dst, src, width, fmt, shr, shg, shb, shp, \
+ maskr, maskg, maskb, rsh, gsh, bsh, S, tab); \
+} \
+ \
+static void name ## ToUV_c_vsx(uint8_t *dstU, uint8_t *dstV, \
+ const uint8_t *unused0, const uint8_t *src, const uint8_t *dummy, \
+ int width, uint32_t *tab) \
+{ \
+ rgb16_32ToUV_c_template_vsx((int16_t*)dstU, (int16_t*)dstV, src, width, fmt, \
+ shr, shg, shb, shp, \
+ maskr, maskg, maskb, rsh, gsh, bsh, S, tab);\
+}
+
+rgb16_32_wrapper(AV_PIX_FMT_BGR32, bgr32, 16, 0, 0, 0, 0xFF0000, 0xFF00, 0x00FF, 8, 0, 8, RGB2YUV_SHIFT + 8)
+rgb16_32_wrapper(AV_PIX_FMT_BGR32_1, bgr321, 16, 0, 0, 8, 0xFF0000, 0xFF00, 0x00FF, 8, 0, 8, RGB2YUV_SHIFT + 8)
+rgb16_32_wrapper(AV_PIX_FMT_RGB32, rgb32, 0, 0, 16, 0, 0x00FF, 0xFF00, 0xFF0000, 8, 0, 8, RGB2YUV_SHIFT + 8)
+rgb16_32_wrapper(AV_PIX_FMT_RGB32_1, rgb321, 0, 0, 16, 8, 0x00FF, 0xFF00, 0xFF0000, 8, 0, 8, RGB2YUV_SHIFT + 8)
+rgb16_32_wrapper(AV_PIX_FMT_BGR565LE, bgr16le, 0, 0, 0, 0, 0x001F, 0x07E0, 0xF800, 11, 5, 0, RGB2YUV_SHIFT + 8)
+rgb16_32_wrapper(AV_PIX_FMT_BGR555LE, bgr15le, 0, 0, 0, 0, 0x001F, 0x03E0, 0x7C00, 10, 5, 0, RGB2YUV_SHIFT + 7)
+rgb16_32_wrapper(AV_PIX_FMT_BGR444LE, bgr12le, 0, 0, 0, 0, 0x000F, 0x00F0, 0x0F00, 8, 4, 0, RGB2YUV_SHIFT + 4)
+rgb16_32_wrapper(AV_PIX_FMT_RGB565LE, rgb16le, 0, 0, 0, 0, 0xF800, 0x07E0, 0x001F, 0, 5, 11, RGB2YUV_SHIFT + 8)
+rgb16_32_wrapper(AV_PIX_FMT_RGB555LE, rgb15le, 0, 0, 0, 0, 0x7C00, 0x03E0, 0x001F, 0, 5, 10, RGB2YUV_SHIFT + 7)
+rgb16_32_wrapper(AV_PIX_FMT_RGB444LE, rgb12le, 0, 0, 0, 0, 0x0F00, 0x00F0, 0x000F, 0, 4, 8, RGB2YUV_SHIFT + 4)
+rgb16_32_wrapper(AV_PIX_FMT_BGR565BE, bgr16be, 0, 0, 0, 0, 0x001F, 0x07E0, 0xF800, 11, 5, 0, RGB2YUV_SHIFT + 8)
+rgb16_32_wrapper(AV_PIX_FMT_BGR555BE, bgr15be, 0, 0, 0, 0, 0x001F, 0x03E0, 0x7C00, 10, 5, 0, RGB2YUV_SHIFT + 7)
+rgb16_32_wrapper(AV_PIX_FMT_BGR444BE, bgr12be, 0, 0, 0, 0, 0x000F, 0x00F0, 0x0F00, 8, 4, 0, RGB2YUV_SHIFT + 4)
+rgb16_32_wrapper(AV_PIX_FMT_RGB565BE, rgb16be, 0, 0, 0, 0, 0xF800, 0x07E0, 0x001F, 0, 5, 11, RGB2YUV_SHIFT + 8)
+rgb16_32_wrapper(AV_PIX_FMT_RGB555BE, rgb15be, 0, 0, 0, 0, 0x7C00, 0x03E0, 0x001F, 0, 5, 10, RGB2YUV_SHIFT + 7)
+rgb16_32_wrapper(AV_PIX_FMT_RGB444BE, rgb12be, 0, 0, 0, 0, 0x0F00, 0x00F0, 0x000F, 0, 4, 8, RGB2YUV_SHIFT + 4)
+
+static void gbr24pToUV_half_c_vsx(uint8_t *_dstU, uint8_t *_dstV,
+ const uint8_t *gsrc, const uint8_t *bsrc, const uint8_t *rsrc,
+ int width, uint32_t *rgb2yuv)
+{
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_rd00, v_rd01, v_rd02;
+ int i, width_adj;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ uint16_t *dstU = (uint16_t *)_dstU;
+ uint16_t *dstV = (uint16_t *)_dstV;
+ const int ru = rgb2yuv[RU_IDX], gu = rgb2yuv[GU_IDX], bu = rgb2yuv[BU_IDX];
+ const int rv = rgb2yuv[RV_IDX], gv = rgb2yuv[GV_IDX], bv = rgb2yuv[BV_IDX];
+
+ uintptr_t gsrc_addr = (uintptr_t)gsrc;
+ uintptr_t bsrc_addr = (uintptr_t)bsrc;
+ uintptr_t rsrc_addr = (uintptr_t)rsrc;
+ uintptr_t dstU_addr = (uintptr_t)_dstU;
+ uintptr_t dstV_addr = (uintptr_t)_dstV;
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = ((vector unsigned int){(0x4001<<(RGB2YUV_SHIFT-6)),(0x4001<<(RGB2YUV_SHIFT-6)),
+ (0x4001<<(RGB2YUV_SHIFT-6)),(0x4001<<(RGB2YUV_SHIFT-6))} );
+ shift2 = vec_splat_u32((RGB2YUV_SHIFT-6+1));
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)gsrc_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)bsrc_addr);
+ v_rd2 = vec_vsx_ld(0, (unsigned short *)rsrc_addr);
+
+ v_rd00 = vec_sr(v_rd0, vec_splats((unsigned short)8));
+ v_rd01 = vec_sr(v_rd1, vec_splats((unsigned short)8));
+ v_rd02 = vec_sr(v_rd2, vec_splats((unsigned short)8));
+ v_rd0 = vec_and(v_rd0, vec_splats((unsigned short)0xFF));
+ v_rd1 = vec_and(v_rd1, vec_splats((unsigned short)0xFF));
+ v_rd2 = vec_and(v_rd2, vec_splats((unsigned short)0xFF));
+
+ v_rd0 = vec_add(v_rd0, v_rd00);
+ v_rd1 = vec_add(v_rd1, v_rd01);
+ v_rd2 = vec_add(v_rd2, v_rd02);
+
+ v_rd00 = vec_mergeh(v_rd0, vec_splat_u16(0x0000));
+ v_rd01 = vec_mergeh(v_rd1, vec_splat_u16(0x0000));
+ v_rd02 = vec_mergeh(v_rd2, vec_splat_u16(0x0000));
+ v_rd0 = vec_mergel(v_rd0, vec_splat_u16(0x0000));
+ v_rd1 = vec_mergel(v_rd1, vec_splat_u16(0x0000));
+ v_rd2 = vec_mergel(v_rd2, vec_splat_u16(0x0000));
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_rd02,
+ ((vector signed int){ru,ru,ru,ru}));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd00,
+ ((vector signed int){gu,gu,gu,gu}) ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd01,
+ ((vector signed int){bu,bu,bu,bu})));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_rd2,
+ ((vector signed int){ru,ru,ru,ru}));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_rd0,
+ ((vector signed int){gu,gu,gu,gu})));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_rd1,
+ ((vector signed int){bu,bu,bu,bu})));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstU_addr); dstU_addr+=16;
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_rd02,
+ ((vector signed int){rv,rv,rv,rv}));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd00,
+ ((vector signed int){gv,gv,gv,gv})));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd01,
+ ((vector signed int){bv,bv,bv,bv})));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_rd2,
+ ((vector signed int){rv,rv,rv,rv}));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_rd0,
+ ((vector signed int){gv,gv,gv,gv})));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_rd1,
+ ((vector signed int){bv,bv,bv,bv})));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstV_addr); dstV_addr+=16;
+
+ gsrc_addr += 16;
+ bsrc_addr += 16;
+ rsrc_addr += 16;
+ }
+ for (i=width_adj; i< width; i++) {
+ unsigned int g = gsrc[2*i] + gsrc[2*i+1];
+ unsigned int b = bsrc[2*i] + bsrc[2*i+1];
+ unsigned int r = rsrc[2*i] + rsrc[2*i+1];
+
+ dstU[i] = (ru*r + gu*g + bu*b + (0x4001<<(RGB2YUV_SHIFT-6))) >> (RGB2YUV_SHIFT-6+1);
+ dstV[i] = (rv*r + gv*g + bv*b + (0x4001<<(RGB2YUV_SHIFT-6))) >> (RGB2YUV_SHIFT-6+1);
+ }
+}
+
+static void rgba64leToA_c_vsx(uint8_t *_dst, const uint8_t *_src, const uint8_t *unused1,
+ const uint8_t *unused2, int width, uint32_t *unused)
+{
+ int16_t *dst = (int16_t *)_dst;
+ const uint16_t *src = (const uint16_t *)_src;
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_rd3, v_dst;
+
+ uintptr_t src_addr = (uintptr_t)_src;
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+ // compute integral number of vector-length items and length of final fragment
+ width_adj = width & (~(int)0x07);
+
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr + 16));
+ v_rd2 = vec_vsx_ld(0, (unsigned short *)(src_addr + 32));
+ v_rd3 = vec_vsx_ld(0, (unsigned short *)(src_addr + 48));
+
+ v_dst = vec_perm(v_rd0, v_rd1, ((vector unsigned char){6, 7, 14, 15, 22, 23, 30, 31}));
+ v_rd0 = vec_perm(v_rd2, v_rd3, ((vector unsigned char){6, 7, 14, 15, 22, 23, 30, 31}));
+ v_dst = vec_sld(v_dst, v_dst, 8);
+ v_dst = vec_sld(v_rd0, v_dst, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 64;
+ dst_addr += 16;
+ }
+
+ for (i=width_adj; i< width; i++) {
+ dst[i]= AV_RL16(src + 4*i + 3);
+ }
+}
+
+static void rgba64beToA_c_vsx(uint8_t *_dst, const uint8_t *_src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int16_t *dst = (int16_t *)_dst;
+ const uint16_t *src = (const uint16_t *)_src;
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_rd3, v_dst;
+
+ uintptr_t src_addr = (uintptr_t)_src;
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+ // compute integral number of vector-length items and length of final fragment
+ width_adj = width & (~(int)0x07);
+
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr + 16));
+ v_rd2 = vec_vsx_ld(0, (unsigned short *)(src_addr + 32));
+ v_rd3 = vec_vsx_ld(0, (unsigned short *)(src_addr + 48));
+
+ v_dst = vec_perm(v_rd0, v_rd1, ((vector unsigned char){7, 6, 15, 14, 23, 22, 31, 30}));
+ v_rd0 = vec_perm(v_rd2, v_rd3, ((vector unsigned char){7, 6, 15, 14, 23, 22, 31, 30}));
+ v_dst = vec_sld(v_dst, v_dst, 8);
+ v_dst = vec_sld(v_rd0, v_dst, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 64;
+ dst_addr += 16;
+ }
+
+ for (i=width_adj; i< width; i++) {
+ dst[i]= AV_RB16(src + 4*i + 3);
+ }
+}
+
+static void abgrToA_c_vsx(uint8_t *_dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int16_t *dst = (int16_t *)_dst;
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_dst;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ // compute integral number of vector-length items and length of final fragment
+ width_adj = width & (~(int)0x07);
+
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)(src_addr));
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr + 16));
+
+ v_rd0 = vec_and(v_rd0, vec_splats((unsigned short)0x00ff));
+ v_rd1 = vec_and(v_rd1, vec_splats((unsigned short)0x00ff));
+
+ v_rd0 = vec_sl(v_rd0, vec_splats((unsigned short)6));
+ v_rd1 = vec_sl(v_rd1, vec_splats((unsigned short)6));
+
+ v_dst = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+
+ for (i = width_adj; i < width; i++) {
+ dst[i]= src[4*i]<<6;
+ }
+}
+
+static void rgbaToA_c_vsx(uint8_t *_dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int16_t *dst = (int16_t *)_dst;
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_dst;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ // compute integral number of vector-length items and length of final fragment
+ width_adj = width & (~(int)0x07);
+
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)(src_addr));
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr + 16));
+
+ v_rd0 = vec_sld(v_rd0, v_rd0, 13);
+ v_rd1 = vec_sld(v_rd1, v_rd1, 13);
+
+ v_rd0 = vec_and(v_rd0, vec_splats((unsigned short)0x00FF));
+ v_rd1 = vec_and(v_rd1, vec_splats((unsigned short)0x00FF));
+
+ v_rd0 = vec_sl(v_rd0, vec_splats((unsigned short)6));
+ v_rd1 = vec_sl(v_rd1, vec_splats((unsigned short)6));
+
+ v_dst = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+
+ for (i = width_adj; i < width; i++) {
+ dst[i]= src[4*i+3]<<6;
+ }
+}
+
+static void palToA_c_vsx(uint8_t *_dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *pal)
+{
+ int16_t *dst = (int16_t *)_dst;
+ int i, j, d, width_adj;
+ uint32_t _pal[8];
+
+ vector unsigned short v_dst;
+ vector unsigned int v_rd0, v_rd1, v_rd3, v_rd4;
+ vector unsigned char sample;
+ vector unsigned int shift1;
+ vector unsigned short shift2;
+
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample = ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29});
+ shift1 = ((vector unsigned int){24, 24, 24, 24});
+ shift2 = vec_splat_u16(6);
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ for( j=0; j<8; ++j)
+ _pal[j] = pal[src[j]];
+
+ v_rd0 = vec_vsx_ld(0, (unsigned int *)_pal);
+ v_rd1 = vec_vsx_ld(0, (unsigned int *)(&_pal[4]));
+ v_rd3 = vec_sr(v_rd0, shift1);
+ v_rd4 = vec_sr(v_rd1, shift1);
+ v_rd0 = vec_perm(v_rd3, v_rd4, sample);
+ v_dst = vec_sl((vector unsigned short)v_rd0, shift2);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src += 8;
+ dst_addr += 16;
+
+ }
+
+ for (i=width_adj; i< width; i++) {
+ d = *src;
+ dst[i]= (pal[d] >> 24)<<6;
+ ++src;
+ }
+}
+
+static void palToY_c_vsx(uint8_t *_dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *pal)
+{
+ int16_t *dst = (int16_t *)_dst;
+ int i, j, d, width_adj;
+ uint32_t _pal[8];
+
+ vector unsigned short v_dst;
+ vector unsigned int v_rd0, v_rd1, v_rd3, v_rd4;
+ vector unsigned char sample;
+ vector unsigned int v_and;
+ vector unsigned short shift;
+
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample = ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29});
+ v_and = ((vector unsigned int){0xFF, 0xFF, 0xFF, 0xFF});
+ shift = vec_splat_u16(6);
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ for( j=0; j<8; ++j)
+ _pal[j] = pal[src[j]];
+
+ v_rd0 = vec_vsx_ld(0, (unsigned int *)_pal);
+ v_rd1 = vec_vsx_ld(0, (unsigned int *)(&_pal[4]));
+ v_rd3 = vec_and(v_rd0, v_and);
+ v_rd4 = vec_and(v_rd1, v_and);
+ v_rd0 = vec_perm(v_rd3, v_rd4, sample);
+ v_dst = vec_sl((vector unsigned short)v_rd0, shift);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src += 8;
+ dst_addr += 16;
+
+ }
+
+ for (i=width_adj; i< width; i++) {
+ d= *src;
+ dst[i] = (pal[d] & 0xFF)<<6;
+ src++;
+ }
+}
+
+static void palToUV_c_vsx(uint8_t *_dstU, uint8_t *_dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *pal)
+{
+ uint16_t *dstU = (uint16_t *)_dstU;
+ int16_t *dstV = (int16_t *)_dstV;
+ int i, j, d, width_adj;
+ uint32_t _pal[8];
+
+ vector unsigned short v_dst, v_tmp;
+ vector unsigned int v_rd0, v_rd1, v_rd3, v_rd4;
+ vector unsigned char sample;
+ vector unsigned short shift1, shift3;
+
+ uintptr_t dstU_addr = (uintptr_t)_dstU;
+ uintptr_t dstV_addr = (uintptr_t)_dstV;
+
+ av_assert1(src1 == src2);
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample = ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29});
+ shift1 = vec_splat_u16(8);
+ //shift2 = ((vector unsigned int){16, 16, 16, 16});
+ shift3 = vec_splat_u16(6);
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ for( j = 0; j < 8; ++j)
+ _pal[j] = pal[src1[j]];
+
+ v_rd0 = vec_vsx_ld(0, (unsigned int *)_pal);
+ v_rd1 = vec_vsx_ld(0, (unsigned int *)(&_pal[4]));
+ v_tmp = (vector unsigned short)vec_perm(v_rd0, v_rd1, sample);
+ v_tmp = vec_sr(v_tmp, shift1);
+ v_dst = vec_sl((vector unsigned short)v_tmp, shift3);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+ v_tmp = (vector unsigned short)vec_perm(v_rd0, v_rd1, sample);
+ v_tmp = vec_sr(v_tmp, shift1);
+ v_dst = vec_sl((vector unsigned short)v_tmp, shift3);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+
+ src1 += 8;
+ dstU_addr += 16;
+ dstV_addr += 16;
+
+ }
+
+ for (i=width_adj; i< width; i++) {
+ d = pal[*src1];;
+ dstU[i] = (uint8_t)(d>> 8)<<6;
+ dstV[i] = (uint8_t)(d>>16)<<6;
+ src1++;
+ }
+}
+
+static void monowhite2Y_c_vsx(uint8_t *_dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+
+ int16_t *dst = (int16_t *)_dst;
+ int i, j;
+ vector unsigned short v_rd0, v_dst;
+
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width = (width + 7) >> 3;
+ for (i = 0; i < width; i++) {
+ v_rd0 = vec_splats((unsigned short)~src[i]);
+
+ v_dst = vec_sr(v_rd0, ((vector unsigned short){7, 6, 5, 4, 3, 2, 1, 0}));
+ v_dst = vec_and(v_dst, vec_splats((unsigned short)0x01));
+ v_dst = vec_mul(v_dst, vec_splats((unsigned short)16383));
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+ dst_addr += 16;
+ }
+ if(width&7){
+ int d= ~src[i];
+ for (j = 0; j < (width&7); j++)
+ dst[8*i+j]= ((d>>(7-j))&1) * 16383;
+ }
+}
+
+static void monoblack2Y_c_vsx(uint8_t *_dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int16_t *dst = (int16_t *)_dst;
+ int i, j;
+ vector unsigned short v_rd0, v_dst;
+
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width = (width + 7) >> 3;
+ for (i = 0; i < width; i++) {
+ v_rd0 = vec_splats((unsigned short)src[i]);
+
+ v_dst = vec_sr(v_rd0, ((vector unsigned short){7, 6, 5, 4, 3, 2, 1, 0}));
+ v_dst = vec_and(v_dst, vec_splats((unsigned short)0x01));
+ v_dst = vec_mul(v_dst, vec_splats((unsigned short)16383));
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+ dst_addr += 16;
+ }
+ if(width&7){
+ int d= src[i];
+ for (j = 0; j < (width&7); j++)
+ dst[8*i+j]= ((d>>(7-j))&1) * 16383;
+ }
+}
+
+static void yuy2ToY_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int i, width_adj;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x0F);
+
+ for ( i = 0; i < width_adj; i += 16) {
+ vector int v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ vector int v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+
+ vector int v_dst = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30}));
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+
+ for (i=width_adj; i< width; i++) {
+ dst[i] = src[2 * i];
+ }
+}
+
+static void yuy2ToUV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *unused)
+{
+ int i, width_adj;
+ vector int v_rd0, v_rd1, v_rd2, v_rd3, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width & (~(int)0x0F);
+
+ vec_and(v_dst, vec_splats(0x000));
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){1, 5, 9, 13, 17, 21, 25, 29, 3, 7, 11, 15, 19, 23, 27, 31});
+ sample2 = ((vector unsigned char){3, 7, 11, 15, 19, 23, 27, 31, 1, 5, 9, 13, 17, 21, 25, 29});
+ }
+ for ( i = 0; i < width_adj; i += 16) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_rd2 = vec_perm(v_rd0, v_rd1, sample1);
+ src_addr += 32;
+
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_rd3 = vec_perm(v_rd0, v_rd1, sample2);
+ v_dst = vec_sld(v_rd2, v_rd3, 8);
+ v_dst = vec_sld(v_dst, v_dst, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+ v_dst = vec_sld(v_rd3, v_rd2, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 32;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+ for (i=width_adj; i< width; i++) {
+ dstU[i] = src1[4 * i + 1];
+ dstV[i] = src1[4 * i + 3];
+ }
+
+ av_assert1(src1 == src2);
+}
+
+static void yvy2ToUV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *unused)
+{
+ int i, width_adj;
+ vector int v_rd0, v_rd1, v_rd2, v_rd3, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width & (~(int)0x0F);
+
+ vec_and(v_dst, vec_splats(0x000));
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){1, 5, 9, 13, 17, 21, 25, 29, 3, 7, 11, 15, 19, 23, 27, 31});
+ sample2 = ((vector unsigned char){3, 7, 11, 15, 19, 23, 27, 31, 1, 5, 9, 13, 17, 21, 25, 29});
+ }
+ for ( i = 0; i < width_adj; i += 16) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_rd2 = vec_perm(v_rd0, v_rd1, sample1);
+ src_addr += 32;
+
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_rd3 = vec_perm(v_rd0, v_rd1, sample2);
+ v_dst = vec_sld(v_rd2, v_rd3, 8);
+ v_dst = vec_sld(v_dst, v_dst, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+ v_dst = vec_sld(v_rd3, v_rd2, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+
+ src_addr += 32;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ dstV[i] = src1[4 * i + 1];
+ dstU[i] = src1[4 * i + 3];
+ }
+
+ av_assert1(src1 == src2);
+}
+static void bswap16Y_c_vsx(uint8_t *_dst, const uint8_t *_src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_dst, v_shift;
+
+ const uint16_t *src = (const uint16_t *)_src;
+ uint16_t *dst = (uint16_t *)_dst;
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj)
+ v_shift = (vector unsigned short)vec_splat_u16((unsigned)8);
+ for ( i = 0; i < width_adj; i += 8) {
+ v_dst = vec_vsx_ld(0, (unsigned short *)src_addr);
+
+ v_rd0 = vec_sl((vector unsigned short)v_dst, v_shift);
+ v_rd1 = vec_sr((vector unsigned short)v_dst, v_shift);
+ v_dst = vec_or(v_rd0, v_rd1);
+
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 16;
+ dst_addr += 16;
+ }
+
+ for (i=width_adj; i< width; i++) {
+ dst[i] = (src[i]>>8) | (src[i]<<8);
+ }
+}
+
+static void bswap16UV_c_vsx(uint8_t *_dstU, uint8_t *_dstV,
+ const uint8_t *unused0, const uint8_t *_src1,
+ const uint8_t *_src2, int width, uint32_t *unused)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_dst, v_shift;
+
+ const uint16_t *src1 = (const uint16_t *)_src1,
+ *src2 = (const uint16_t *)_src2;
+ uint16_t *dstU = (uint16_t *)_dstU,
+ *dstV = (uint16_t *)_dstV;
+ uintptr_t src1_addr = (uintptr_t)_src1,
+ src2_addr = (uintptr_t)_src2;
+ uintptr_t dstU_addr = (uintptr_t)dstU,
+ dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj)
+ v_shift = (vector unsigned short)vec_splat_u16((unsigned)8);
+ for ( i = 0; i < width_adj; i += 8) {
+ // load to dstU
+ v_dst = vec_vsx_ld(0, (unsigned short *)src1_addr);
+ v_rd0 = vec_sl((vector unsigned short)v_dst, v_shift);
+ v_rd1 = vec_sr((vector unsigned short)v_dst, v_shift);
+ v_dst = vec_or(v_rd0, v_rd1);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+ // load to dstV
+ v_dst = vec_vsx_ld(0, (unsigned short *)src2_addr);
+ v_rd0 = vec_sl((vector unsigned short)v_dst, v_shift);
+ v_rd1 = vec_sr((vector unsigned short)v_dst, v_shift);
+ v_dst = vec_or(v_rd0, v_rd1);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+ //
+ src1_addr += 16;
+ src2_addr += 16;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+ for (i=width_adj; i< width; i++) {
+ dstU[i] = (src1[i]>>8) | (src1[i]<<8);
+ dstV[i] = (src2[i]>>8) | (src2[i]<<8);
+ }
+}
+
+static void read_ya16le_gray_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int i, width_adj;
+ vector int v_rd0, v_rd1, v_dst;
+ vector unsigned char sample;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample = ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29});
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_dst = vec_perm(v_rd0, v_rd1, sample);
+
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dst + i * 2, AV_RL16(src + i * 4));
+ }
+}
+
+static void read_ya16le_alpha_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int i, width_adj;
+ vector int v_rd0, v_rd1, v_dst;
+ vector unsigned char sample;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample = ((vector unsigned char){2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31});
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_dst = vec_perm(v_rd0, v_rd1, sample);
+
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dst + i * 2, AV_RL16(src + i * 4 + 2));
+ }
+}
+
+static void read_ya16be_gray_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+
+ int i, width_adj;
+
+ vector int v_rd0, v_rd1, v_dst;
+ vector unsigned char sample;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample = ((vector unsigned char){1, 0, 5, 4, 9, 8, 13, 12, 17, 16, 21, 20, 25, 24, 29, 28});
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_dst = vec_perm(v_rd0, v_rd1, sample);
+
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dst + i * 2, AV_RB16(src + i * 4));
+ }
+}
+
+static void read_ya16be_alpha_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int i, width_adj;
+
+ vector int v_rd0, v_rd1, v_dst;
+ vector unsigned char sample;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample = ((vector unsigned char){3, 2, 7, 6, 11, 10, 15, 14, 19, 18, 23, 22, 27, 26, 31, 30});
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_dst = vec_perm(v_rd0, v_rd1, sample);
+
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dst + i * 2, AV_RB16(src + i * 4 + 2));
+ }
+
+}
+
+static void read_ayuv64le_Y_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused0, const uint8_t *unused1,
+ int width, uint32_t *unused2)
+{
+ int i, width_adj;
+
+ vector int v_rd0, v_rd1, v_rd2, v_rd3, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){0, 0, 0, 0, 0, 0, 0, 0, 2, 3, 10, 11, 18, 19, 26, 27});
+ sample2 = ((vector unsigned char){2, 3, 10, 11, 18, 19, 26, 27, 0, 0, 0, 0, 0, 0, 0, 0});
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_rd2 = vec_vsx_ld(0, (int *)(src_addr+32));
+ v_rd3 = vec_vsx_ld(0, (int *)(src_addr + 48));
+
+ v_rd0 = vec_perm(v_rd0, v_rd1, sample1);
+ v_rd2 = vec_perm(v_rd2, v_rd3, sample2);
+ v_dst = vec_sld(v_rd2, v_rd0, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 64;
+ dst_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dst + i * 2, AV_RL16(src + i * 8 + 2));
+ }
+}
+
+
+static void read_ayuv64le_UV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src,
+ const uint8_t *unused1, int width, uint32_t *unused2)
+{
+ int i, width_adj;
+
+ vector int v_rd0, v_rd1, v_rd2, v_rd3, v_rd4, v_rd5, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){6, 7, 14, 15, 22, 23, 30, 31, 4, 5, 12, 13, 20, 21, 28, 29});
+ sample2 = ((vector unsigned char){4, 5, 12, 13, 20, 21, 28, 29, 6, 7, 14, 15, 22, 23, 30, 31});
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_rd2 = vec_vsx_ld(0, (int *)(src_addr+32));
+ v_rd3 = vec_vsx_ld(0, (int *)(src_addr + 48));
+
+ v_rd4 = vec_perm(v_rd0, v_rd1, sample1);
+ v_rd5 = vec_perm(v_rd2, v_rd3, sample2);
+ v_dst = vec_sld(v_rd5, v_rd4, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+ v_dst = vec_sld(v_rd4, v_rd5, 8);
+ v_dst = vec_sld(v_dst, v_dst, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 64;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dstU + i * 2, AV_RL16(src + i * 8 + 4));
+ AV_WN16(dstV + i * 2, AV_RL16(src + i * 8 + 6));
+ }
+}
+
+static void read_ayuv64le_A_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused0, const uint8_t *unused1,
+ int width, uint32_t *unused2)
+{
+ int i, width_adj;
+
+ vector int v_rd0, v_rd1, v_rd2, v_rd3, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 8, 9, 16, 17, 24, 25});
+ sample2 = ((vector unsigned char){0, 1, 8, 9, 16, 17, 24, 25, 0, 0, 0, 0, 0, 0, 0, 0});
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_rd2 = vec_vsx_ld(0, (int *)(src_addr+32));
+ v_rd3 = vec_vsx_ld(0, (int *)(src_addr + 48));
+
+ v_rd0 = vec_perm(v_rd0, v_rd1, sample1);
+ v_rd2 = vec_perm(v_rd2, v_rd3, sample2);
+ v_dst = vec_sld(v_rd2, v_rd0, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 64;
+ dst_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dst + i * 2, AV_RL16(src + i * 8));
+ }
+}
+
+/* This is almost identical to the previous, end exists only because
+ * yuy2ToY/UV)(dst, src + 1, ...) would have 100% unaligned accesses. */
+static void uyvyToY_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int i, width_adj;
+
+ vector int v_rd0, v_rd1, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x0F);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31});
+ }
+ for ( i = 0; i < width_adj; i += 16) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+
+ v_dst = vec_perm(v_rd0, v_rd1, sample1);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ dst[i] = src[2 * i + 1];
+ }
+}
+
+static void uyvyToUV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *unused)
+{
+ int i, width_adj;
+
+ vector int v_rd0, v_rd1, v_rd2, v_rd3, v_rd4, v_rd5, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width & (~(int)0x0F);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){2, 6, 10, 14, 18, 22, 26, 30, 0, 4, 8, 12, 16, 20, 24, 28});
+ sample2 = ((vector unsigned char){0, 4, 8, 12, 16, 20, 24, 28, 2, 6, 10, 14, 18, 22, 26, 30});
+ }
+ for ( i = 0; i < width_adj; i += 16) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+ v_rd2 = vec_vsx_ld(0, (int *)(src_addr+32));
+ v_rd3 = vec_vsx_ld(0, (int *)(src_addr + 48));
+
+ v_rd4 = vec_perm(v_rd0, v_rd1, sample1);
+ v_rd5 = vec_perm(v_rd2, v_rd3, sample2);
+ v_dst = vec_sld(v_rd5, v_rd4, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+ v_dst = vec_sld(v_rd4, v_rd5, 8);
+ v_dst = vec_sld(v_dst, v_dst, 8);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 64;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ dstU[i] = src1[4 * i + 0];
+ dstV[i] = src1[4 * i + 2];
+ }
+
+ av_assert1(src1 == src2);
+}
+
+static av_always_inline void nvXXtoUV_c_vsx(uint8_t *dst1, uint8_t *dst2,
+ const uint8_t *src, int width)
+{
+ int i, width_adj;
+
+ vector int v_rd0, v_rd1, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst1_addr = (uintptr_t)dst1;
+ uintptr_t dst2_addr = (uintptr_t)dst2;
+
+ width_adj = width & (~(int)0x0F);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30});
+ sample2 = ((vector unsigned char){1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31});
+ }
+ for ( i = 0; i < width_adj; i += 16) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr + 16));
+
+ v_dst = vec_perm(v_rd0, v_rd1, sample1);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst1_addr);
+ v_dst = vec_perm(v_rd0, v_rd1, sample2);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst2_addr);
+
+ src_addr += 32;
+ dst1_addr += 16;
+ dst2_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ dst1[i] = src[2 * i + 0];
+ dst2[i] = src[2 * i + 1];
+ }
+}
+
+static void nv12ToUV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *unused)
+{
+ nvXXtoUV_c_vsx(dstU, dstV, src1, width);
+}
+
+static void nv21ToUV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *unused)
+{
+ nvXXtoUV_c_vsx(dstV, dstU, src1, width);
+}
+
+static void p010LEToY_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_dst, shift;
+ vector unsigned char sample;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj)
+ shift = vec_splat_u16(6);
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+
+ v_dst = vec_sr(v_rd0, vec_splat_u16(6));
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 16;
+ dst_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dst + i * 2, AV_RL16(src + i * 2) >> 6);
+ }
+}
+
+static void p010BEToY_c_vsx(uint8_t *dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_dst, shift;
+ vector unsigned char sample;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)dst;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample = ((vector unsigned char){1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14});
+ shift = vec_splat_u16(6);
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+
+ v_rd1 = vec_perm(v_rd0, v_rd0, sample);
+ v_dst = vec_sr(v_rd1, shift);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 16;
+ dst_addr += 16;
+ }
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dst + i * 2, AV_RB16(src + i * 2) >> 6);
+ }
+}
+
+static void p010LEToUV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *unused)
+{
+ int i, width_adj;
+
+ vector unsigned short v_rd0, v_rd1, v_dst;
+ vector unsigned char sample1, sample2;
+ vector unsigned short shift;
+
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29});
+ sample2 = ((vector unsigned char){2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31});
+ shift = vec_splat_u16(6);
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr + 16));
+
+ v_dst = vec_perm(v_rd0, v_rd1, sample1);
+ v_dst = vec_sr(v_dst, shift);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+ v_dst = vec_perm(v_rd0, v_rd1, sample2);
+ v_dst = vec_sr(v_dst, shift);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 32;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dstU + i * 2, AV_RL16(src1 + i * 4 + 0) >> 6);
+ AV_WN16(dstV + i * 2, AV_RL16(src1 + i * 4 + 2) >> 6);
+ }
+}
+
+static void p010BEToUV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *unused)
+{
+ int i, width_adj;
+
+ vector unsigned short v_rd0, v_rd1, v_dst;
+ vector unsigned char sample1, sample2;
+ vector unsigned short shift;
+
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){1, 0, 5, 4, 9, 8, 13, 12, 17, 16, 21, 22, 25, 24, 29, 28});
+ sample2 = ((vector unsigned char){3, 2, 7, 6, 11, 10, 15, 14, 19, 18, 23, 22, 27, 26, 31, 30});
+ shift = vec_splat_u16(6);
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr + 16));
+
+ v_dst = vec_perm(v_rd0, v_rd1, sample1);
+ v_dst = vec_sr(v_dst, shift);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+ v_dst = vec_perm(v_rd0, v_rd1, sample2);
+ v_dst = vec_sr(v_dst, shift);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 32;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dstU + i * 2, AV_RB16(src1 + i * 4 + 0) >> 6);
+ AV_WN16(dstV + i * 2, AV_RB16(src1 + i * 4 + 2) >> 6);
+ }
+}
+
+static void p016LEToUV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *unused)
+{
+ int i, width_adj;
+
+ vector unsigned short v_rd0, v_rd1, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29});
+ sample2 = ((vector unsigned char){2, 3, 6, 7, 10, 11, 14, 15, 18, 19, 22, 23, 26, 27, 30, 31});
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr + 16));
+
+ v_dst = vec_perm(v_rd0, v_rd1, sample1);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+ v_dst = vec_perm(v_rd0, v_rd1, sample2);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 32;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dstU + i * 2, AV_RL16(src1 + i * 4 + 0));
+ AV_WN16(dstV + i * 2, AV_RL16(src1 + i * 4 + 2));
+ }
+}
+
+static void p016BEToUV_c_vsx(uint8_t *dstU, uint8_t *dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *unused)
+{
+ int i, width_adj;
+
+ vector unsigned short v_rd0, v_rd1, v_dst;
+ vector unsigned char sample1, sample2;
+
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)dstU;
+ uintptr_t dstV_addr = (uintptr_t)dstV;
+
+ width_adj = width & (~(int)0x07);
+
+ if(width_adj){
+ sample1 = ((vector unsigned char){1, 0, 5, 4, 9, 8, 13, 12, 17, 16, 21, 22, 25, 24, 29, 28});
+ sample2 = ((vector unsigned char){3, 2, 7, 6, 11, 10, 15, 14, 19, 18, 23, 22, 27, 26, 31, 30});
+ }
+ for ( i = 0; i < width_adj; i += 8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr + 16));
+
+ v_dst = vec_perm(v_rd0, v_rd1, sample1);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstU_addr);
+ v_dst = vec_perm(v_rd0, v_rd1, sample2);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 32;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+
+
+ for (i=width_adj; i< width; i++) {
+ AV_WN16(dstU + i * 2, AV_RB16(src1 + i * 4 + 0));
+ AV_WN16(dstV + i * 2, AV_RB16(src1 + i * 4 + 2));;
+ }
+}
+
+#define input_pixel(pos) (isBE(origin) ? AV_RB16(pos) : AV_RL16(pos))
+
+static void bgr24ToY_c_vsx(uint8_t *_dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *rgb2yuv)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_g, v_b, v_r;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ int16_t *dst = (int16_t *)_dst;
+ vector signed int v_ry, v_gy, v_by;
+ int32_t ry = rgb2yuv[RY_IDX], gy = rgb2yuv[GY_IDX], by = rgb2yuv[BY_IDX];
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x801<<(RGB2YUV_SHIFT-7)));
+ shift2 = vec_splats((unsigned int)(RGB2YUV_SHIFT-6));
+ v_ry = vec_splats((signed int)ry);
+ v_gy = vec_splats((signed int)gy);
+ v_by = vec_splats((signed int)by);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr+16));
+
+ v_b = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 0, 3, 0, 6, 0, 9, 0, 12, 0, 15, 0, 18, 0, 21, 0}));
+ v_g = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){1, 0, 4, 0, 7, 0, 10, 0, 13, 0, 16, 0, 19, 0, 22, 0}));
+ v_r = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){2, 0, 5, 0, 8, 0, 11, 0, 14, 0, 17, 0, 20, 0, 23, 0}));
+
+ v_b = vec_and(v_b, vec_splats((unsigned short)0x00FF));
+ v_g = vec_and(v_g, vec_splats((unsigned short)0x00FF));
+ v_r = vec_and(v_r, vec_splats((unsigned short)0x00FF));
+
+ v_rd0 = vec_mergeh(v_b, vec_splats((unsigned short)0x0000));
+ v_rd1 = vec_mergeh(v_g, vec_splats((unsigned short)0x0000));
+ v_rd2 = vec_mergeh(v_r, vec_splats((unsigned short)0x0000));
+
+ v_g = vec_mergel(v_g, vec_splats((unsigned short)0x0000));
+ v_b = vec_mergel(v_b, vec_splats((unsigned short)0x0000));
+ v_r = vec_mergel(v_r, vec_splats((unsigned short)0x0000));
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_rd2, v_ry);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd1, v_gy ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd0, v_by ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_ry);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gy ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_by ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dst_addr);
+
+ src_addr += 24;
+ dst_addr += 16;
+ }
+ for (i=width_adj; i< width; i++) {
+ unsigned int b = src[3*i];
+ unsigned int g = src[3*i + 1];
+ unsigned int r = src[3*i + 2];
+
+ dst[i] = ((ry*r + gy*g + by*b + (0x801<<(RGB2YUV_SHIFT-7)))>>(RGB2YUV_SHIFT-6));
+ }
+}
+
+static void bgr24ToUV_c(uint8_t *_dstU, uint8_t *_dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *rgb2yuv)
+{
+ int16_t *dstU = (int16_t *)_dstU;
+ int16_t *dstV = (int16_t *)_dstV;
+ int32_t ru = rgb2yuv[RU_IDX], gu = rgb2yuv[GU_IDX], bu = rgb2yuv[BU_IDX];
+ int32_t rv = rgb2yuv[RV_IDX], gv = rgb2yuv[GV_IDX], bv = rgb2yuv[BV_IDX];
+ int i;
+ for (i = 0; i < width; i++) {
+ int b = src1[3 * i + 0];
+ int g = src1[3 * i + 1];
+ int r = src1[3 * i + 2];
+
+ dstU[i] = (ru*r + gu*g + bu*b + (256<<(RGB2YUV_SHIFT-1)) +
+ (1<<(RGB2YUV_SHIFT-7)))>>(RGB2YUV_SHIFT-6);
+ dstV[i] = (rv*r + gv*g + bv*b + (256<<(RGB2YUV_SHIFT-1)) +
+ (1<<(RGB2YUV_SHIFT-7)))>>(RGB2YUV_SHIFT-6);
+ }
+ av_assert1(src1 == src2);
+}
+
+static void bgr24ToUV_c_vsx(uint8_t *_dstU, uint8_t *_dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *rgb2yuv)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_g, v_b, v_r;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ int16_t *dstU = (int16_t *)_dstU;
+ int16_t *dstV = (int16_t *)_dstV;
+ vector signed int v_ru, v_gu, v_bu, v_rv, v_gv, v_bv;
+ int32_t ru = rgb2yuv[RU_IDX], gu = rgb2yuv[GU_IDX], bu = rgb2yuv[BU_IDX];
+ int32_t rv = rgb2yuv[RV_IDX], gv = rgb2yuv[GV_IDX], bv = rgb2yuv[BV_IDX];
+
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)_dstU;
+ uintptr_t dstV_addr = (uintptr_t)_dstV;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x4001<<(RGB2YUV_SHIFT-7)));
+ shift2 = vec_splats((unsigned int)(RGB2YUV_SHIFT-6));
+ v_ru = vec_splats((signed int)ru);
+ v_gu = vec_splats((signed int)gu);
+ v_bu = vec_splats((signed int)bu);
+ v_rv = vec_splats((signed int)rv);
+ v_gv = vec_splats((signed int)gv);
+ v_bv = vec_splats((signed int)bv);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr+16));
+
+ v_b = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 0, 3, 0, 6, 0, 9, 0, 12, 0, 15, 0, 18, 0, 21, 0}));
+ v_g = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){1, 0, 4, 0, 7, 0, 10, 0, 13, 0, 16, 0, 19, 0, 22, 0}));
+ v_r = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){2, 0, 5, 0, 8, 0, 11, 0, 14, 0, 17, 0, 20, 0, 23, 0}));
+
+ v_b = vec_and(v_b, vec_splats((unsigned short)0x00FF));
+ v_g = vec_and(v_g, vec_splats((unsigned short)0x00FF));
+ v_r = vec_and(v_r, vec_splats((unsigned short)0x00FF));
+
+ v_rd0 = vec_mergeh(v_b, vec_splats((unsigned short)0x0000));
+ v_rd1 = vec_mergeh(v_g, vec_splats((unsigned short)0x0000));
+ v_rd2 = vec_mergeh(v_r, vec_splats((unsigned short)0x0000));
+
+ v_g = vec_mergel(v_g, vec_splats((unsigned short)0x0000));
+ v_b = vec_mergel(v_b, vec_splats((unsigned short)0x0000));
+ v_r = vec_mergel(v_r, vec_splats((unsigned short)0x0000));
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_rd2, v_ru);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd1, v_gu ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd0, v_bu ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_ru);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gu ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_bu ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstU_addr);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_rd2, v_rv);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd1, v_gv ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd0, v_bv ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_rv);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gv ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_bv ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 24;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+ for (i=width_adj; i< width; i++) {
+ int b = src1[3 * i + 0];
+ int g = src1[3 * i + 1];
+ int r = src1[3 * i + 2];
+
+ dstU[i] = (ru*r + gu*g + bu*b + (0x4001<<(RGB2YUV_SHIFT-7)))>>(RGB2YUV_SHIFT-6);
+ dstV[i] = (rv*r + gv*g + bv*b + (0x4001<<(RGB2YUV_SHIFT-7)))>>(RGB2YUV_SHIFT-6);
+ }
+ av_assert1(src1 == src2);
+}
+
+static void rgb24ToY_c_vsx(uint8_t *_dst, const uint8_t *src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *rgb2yuv)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_g, v_b, v_r;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ int16_t *dst = (int16_t *)_dst;
+ vector signed int v_ry, v_gy, v_by;
+ int32_t ry = rgb2yuv[RY_IDX], gy = rgb2yuv[GY_IDX], by = rgb2yuv[BY_IDX];
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x801<<(RGB2YUV_SHIFT-7)));
+ shift2 = vec_splats((unsigned int)(RGB2YUV_SHIFT-6));
+ v_ry = vec_splats((signed int)ry);
+ v_gy = vec_splats((signed int)gy);
+ v_by = vec_splats((signed int)by);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr+16));
+
+ v_r = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 0, 3, 0, 6, 0, 9, 0, 12, 0, 15, 0, 18, 0, 21, 0}));
+ v_g = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){1, 0, 4, 0, 7, 0, 10, 0, 13, 0, 16, 0, 19, 0, 22, 0}));
+ v_b = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){2, 0, 5, 0, 8, 0, 11, 0, 14, 0, 17, 0, 20, 0, 23, 0}));
+
+ v_b = vec_and(v_b, vec_splats((unsigned short)0x00FF));
+ v_g = vec_and(v_g, vec_splats((unsigned short)0x00FF));
+ v_r = vec_and(v_r, vec_splats((unsigned short)0x00FF));
+
+ v_rd0 = vec_mergeh(v_b, vec_splats((unsigned short)0x0000));
+ v_rd1 = vec_mergeh(v_g, vec_splats((unsigned short)0x0000));
+ v_rd2 = vec_mergeh(v_r, vec_splats((unsigned short)0x0000));
+
+ v_g = vec_mergel(v_g, vec_splats((unsigned short)0x0000));
+ v_b = vec_mergel(v_b, vec_splats((unsigned short)0x0000));
+ v_r = vec_mergel(v_r, vec_splats((unsigned short)0x0000));
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_rd2, v_ry);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd1, v_gy ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd0, v_by ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_ry);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gy ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_by ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dst_addr);
+
+ src_addr += 24;
+ dst_addr += 16;
+ }
+ for (i=width_adj; i< width; i++) {
+ unsigned int r = src[3*i];
+ unsigned int g = src[3*i + 1];
+ unsigned int b = src[3*i + 2];
+
+ //dst[i] = ((ry*r + gy*g + by*b + (0x4001<<(RGB2YUV_SHIFT-7)))>>(RGB2YUV_SHIFT-6));
+ dst[i] = ((ry*r + gy*g + by*b + (0x801<<(RGB2YUV_SHIFT-7)))>>(RGB2YUV_SHIFT-6));
+ }
+}
+
+static void rgb24ToUV_c_vsx(uint8_t *_dstU, uint8_t *_dstV,
+ const uint8_t *unused0, const uint8_t *src1,
+ const uint8_t *src2, int width, uint32_t *rgb2yuv)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_g, v_b, v_r;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ int16_t *dstU = (int16_t *)_dstU;
+ int16_t *dstV = (int16_t *)_dstV;
+ vector signed int v_ru, v_gu, v_bu, v_rv, v_gv, v_bv;
+ int32_t ru = rgb2yuv[RU_IDX], gu = rgb2yuv[GU_IDX], bu = rgb2yuv[BU_IDX];
+ int32_t rv = rgb2yuv[RV_IDX], gv = rgb2yuv[GV_IDX], bv = rgb2yuv[BV_IDX];
+
+ uintptr_t src_addr = (uintptr_t)src1;
+ uintptr_t dstU_addr = (uintptr_t)_dstU;
+ uintptr_t dstV_addr = (uintptr_t)_dstV;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x4001<<(RGB2YUV_SHIFT-7)));
+ shift2 = vec_splats((unsigned int)(RGB2YUV_SHIFT-6));
+ v_ru = vec_splats((signed int)ru);
+ v_gu = vec_splats((signed int)gu);
+ v_bu = vec_splats((signed int)bu);
+ v_rv = vec_splats((signed int)rv);
+ v_gv = vec_splats((signed int)gv);
+ v_bv = vec_splats((signed int)bv);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src_addr+16));
+
+ v_r = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 0, 3, 0, 6, 0, 9, 0, 12, 0, 15, 0, 18, 0, 21, 0}));
+ v_g = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){1, 0, 4, 0, 7, 0, 10, 0, 13, 0, 16, 0, 19, 0, 22, 0}));
+ v_b = vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){2, 0, 5, 0, 8, 0, 11, 0, 14, 0, 17, 0, 20, 0, 23, 0}));
+
+ v_r = vec_and(v_r, vec_splats((unsigned short)0x00FF));
+ v_g = vec_and(v_g, vec_splats((unsigned short)0x00FF));
+ v_b = vec_and(v_b, vec_splats((unsigned short)0x00FF));
+
+ v_rd0 = vec_mergeh(v_r, vec_splats((unsigned short)0x0000));
+ v_rd1 = vec_mergeh(v_g, vec_splats((unsigned short)0x0000));
+ v_rd2 = vec_mergeh(v_b, vec_splats((unsigned short)0x0000));
+
+ v_r = vec_mergel(v_r, vec_splats((unsigned short)0x0000));
+ v_g = vec_mergel(v_g, vec_splats((unsigned short)0x0000));
+ v_b = vec_mergel(v_b, vec_splats((unsigned short)0x0000));
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_rd0, v_ru);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd1, v_gu ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd2, v_bu ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_ru);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gu ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_bu ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstU_addr);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_rd0, v_rv);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd1, v_gv ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_rd2, v_bv ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_rv);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gv ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_bv ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstV_addr);
+
+ src_addr += 24;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+ for (i=width_adj; i< width; i++) {
+ int r = src1[3 * i + 0];
+ int g = src1[3 * i + 1];
+ int b = src1[3 * i + 2];
+
+ dstU[i] = (ru*r + gu*g + bu*b + (0x4001<<(RGB2YUV_SHIFT-7)))>>(RGB2YUV_SHIFT-6);
+ dstV[i] = (rv*r + gv*g + bv*b + (0x4001<<(RGB2YUV_SHIFT-7)))>>(RGB2YUV_SHIFT-6);
+ }
+ av_assert1(src1 == src2);
+}
+
+static void planar_rgb_to_y_vsx(uint8_t *_dst, const uint8_t *src[4],
+ int width, int32_t *rgb2yuv)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_g, v_b, v_r, v_g1, v_b1, v_r1;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ int16_t *dst = (int16_t *)_dst;
+ vector signed int v_ry, v_gy, v_by;
+ int32_t ry = rgb2yuv[RY_IDX], gy = rgb2yuv[GY_IDX], by = rgb2yuv[BY_IDX];
+
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x801<<(RGB2YUV_SHIFT-7)));
+ shift2 = vec_splats((unsigned int)(RGB2YUV_SHIFT-6));
+ v_ry = vec_splats((signed int)ry);
+ v_gy = vec_splats((signed int)gy);
+ v_by = vec_splats((signed int)by);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ if(i&1){
+ v_rd0 = vec_sld(v_rd0, v_rd0, 8);
+ v_rd1 = vec_sld(v_rd1, v_rd1, 8);
+ v_rd2 = vec_sld(v_rd2, v_rd2, 8);
+ }else{
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src[0]);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src[1]));
+ v_rd2 = vec_vsx_ld(0, (unsigned short *)(src[2]));
+ }
+
+ v_g = vec_perm(v_rd0, v_rd0,
+ ((vector unsigned char){0, 0, 1, 0, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0}));
+ v_b = vec_perm(v_rd1, v_rd1,
+ ((vector unsigned char){0, 0, 1, 0, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0}));
+ v_r = vec_perm(v_rd2, v_rd2,
+ ((vector unsigned char){0, 0, 1, 0, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0}));
+
+ v_b = vec_and(v_b, vec_splats((unsigned short)0x00FF));
+ v_g = vec_and(v_g, vec_splats((unsigned short)0x00FF));
+ v_r = vec_and(v_r, vec_splats((unsigned short)0x00FF));
+
+ v_b1 = vec_mergeh(v_b, vec_splats((unsigned short)0x0000));
+ v_g1 = vec_mergeh(v_g, vec_splats((unsigned short)0x0000));
+ v_r1 = vec_mergeh(v_r, vec_splats((unsigned short)0x0000));
+
+ v_g = vec_mergel(v_g, vec_splats((unsigned short)0x0000));
+ v_b = vec_mergel(v_b, vec_splats((unsigned short)0x0000));
+ v_r = vec_mergel(v_r, vec_splats((unsigned short)0x0000));
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_ry);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gy ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_by ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_ry);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gy ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_by ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dst_addr);
+
+ src[0] += 8;
+ src[1] += 8;
+ src[2] += 8;
+ dst_addr += 16;
+ }
+
+ for (i=width_adj; i< width; i++) {
+ int g = src[0][0];
+ int b = src[1][0];
+ int r = src[2][0];
+ dst[i] = (ry*r + gy*g + by*b + (0x801<<(RGB2YUV_SHIFT-7))) >> (RGB2YUV_SHIFT-6);
+ ++src[0];
+ ++src[1];
+ ++src[2];
+ }
+}
+
+static void planar_rgb_to_a_vsx(uint8_t *_dst, const uint8_t *src[4],
+ int width, int32_t *unused)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_a, v_dst;
+ int16_t *dst = (int16_t *)_dst;
+
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+
+ width_adj = width&(~(int)0x07);
+
+ for (i = 0; i < width_adj; i+=8) {
+ if(i&1)
+ v_rd0 = vec_sld(v_rd0, v_rd0, 8);
+ else
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src[3]);
+
+ v_a = vec_perm(v_rd0, v_rd0,
+ ((vector unsigned char){0, 0, 1, 0, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0}));
+ v_dst = vec_sl(v_a, vec_splats((unsigned short)6));
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src[3] += 8;
+ dst_addr += 16;
+ }
+ for (i=width_adj; i< width; i++){
+ dst[i] = src[3][0] << 6;
+ ++src[3];
+ }
+}
+
+
+static void planar_rgb_to_uv_vsx(uint8_t *_dstU, uint8_t *_dstV,
+ const uint8_t *src[4], int width, int32_t *rgb2yuv)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_g, v_b, v_r, v_g1, v_b1, v_r1;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ uint16_t *dstU = (uint16_t *)_dstU;
+ uint16_t *dstV = (uint16_t *)_dstV;
+ vector signed int v_ru, v_gu, v_bu, v_rv, v_gv, v_bv;
+ int32_t ru = rgb2yuv[RU_IDX], gu = rgb2yuv[GU_IDX], bu = rgb2yuv[BU_IDX];
+ int32_t rv = rgb2yuv[RV_IDX], gv = rgb2yuv[GV_IDX], bv = rgb2yuv[BV_IDX];
+
+ uintptr_t dstU_addr = (uintptr_t)_dstU;
+ uintptr_t dstV_addr = (uintptr_t)_dstV;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(0x4001<<(RGB2YUV_SHIFT-7)));
+ shift2 = vec_splats((unsigned int)(RGB2YUV_SHIFT-6));
+ v_ru = vec_splats((signed int)ru);
+ v_gu = vec_splats((signed int)gu);
+ v_bu = vec_splats((signed int)bu);
+ v_rv = vec_splats((signed int)rv);
+ v_gv = vec_splats((signed int)gv);
+ v_bv = vec_splats((signed int)bv);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ if(i&1){
+ v_rd0 = vec_sld(v_rd0, v_rd0, 8);
+ v_rd1 = vec_sld(v_rd1, v_rd1, 8);
+ v_rd2 = vec_sld(v_rd2, v_rd2, 8);
+ }else{
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src[0]);
+ v_rd1 = vec_vsx_ld(0, (unsigned short *)(src[1]));
+ v_rd2 = vec_vsx_ld(0, (unsigned short *)(src[2]));
+ }
+
+ v_g = vec_perm(v_rd0, v_rd0,
+ ((vector unsigned char){0, 0, 1, 0, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0}));
+ v_b = vec_perm(v_rd1, v_rd1,
+ ((vector unsigned char){0, 0, 1, 0, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0}));
+ v_r = vec_perm(v_rd2, v_rd2,
+ ((vector unsigned char){0, 0, 1, 0, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0}));
+
+ v_b = vec_and(v_b, vec_splats((unsigned short)0x00FF));
+ v_g = vec_and(v_g, vec_splats((unsigned short)0x00FF));
+ v_r = vec_and(v_r, vec_splats((unsigned short)0x00FF));
+
+ v_b1 = vec_mergeh(v_b, vec_splats((unsigned short)0x0000));
+ v_g1 = vec_mergeh(v_g, vec_splats((unsigned short)0x0000));
+ v_r1 = vec_mergeh(v_r, vec_splats((unsigned short)0x0000));
+
+ v_g = vec_mergel(v_g, vec_splats((unsigned short)0x0000));
+ v_b = vec_mergel(v_b, vec_splats((unsigned short)0x0000));
+ v_r = vec_mergel(v_r, vec_splats((unsigned short)0x0000));
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_ru);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gu ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_bu ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_ru);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gu ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_bu ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstU_addr);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_rv);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gv ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_bv ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_rv);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gv ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_bv ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstV_addr);
+
+ src[0] += 8;
+ src[1] += 8;
+ src[2] += 8;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+ for (i=width_adj; i< width; i++) {
+ int g = src[0][0];
+ int b = src[1][0];
+ int r = src[2][0];
+
+ dstU[i] = (ru*r + gu*g + bu*b + (0x4001<<(RGB2YUV_SHIFT-7))) >> (RGB2YUV_SHIFT-6);
+ dstV[i] = (rv*r + gv*g + bv*b + (0x4001<<(RGB2YUV_SHIFT-7))) >> (RGB2YUV_SHIFT-6);
+ ++src[0];
+ ++src[1];
+ ++src[2];
+
+ }
+}
+
+#define rdpx(src) \
+ is_be ? AV_RB16(src) : AV_RL16(src)
+static av_always_inline
+void planar_rgb16_to_y_vsx(uint8_t *_dst, const uint8_t *_src[4],
+ int width, int bpc, int is_be, int32_t *rgb2yuv)
+{
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_g, v_b, v_r, v_g1, v_b1, v_r1;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ int16_t *dst = (int16_t *)_dst;
+ const uint16_t **src = (const uint16_t **)_src;
+ vector signed int v_ry, v_gy, v_by;
+ int32_t ry = rgb2yuv[RY_IDX], gy = rgb2yuv[GY_IDX], by = rgb2yuv[BY_IDX];
+ int sh = bpc < 16 ? bpc : 14;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(33 << (RGB2YUV_SHIFT + bpc - 9)));
+ shift2 = vec_splats((unsigned int)(RGB2YUV_SHIFT + sh - 14));
+ v_ry = vec_splats((signed int)ry);
+ v_gy = vec_splats((signed int)gy);
+ v_by = vec_splats((signed int)by);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+
+ v_g = vec_vsx_ld(0, (unsigned short *)src[0]);
+ v_b = vec_vsx_ld(0, (unsigned short *)(src[1]));
+ v_r = vec_vsx_ld(0, (unsigned short *)(src[2]));
+ if(is_be){
+ v_g = vec_perm(v_g, v_g,
+ ((vector unsigned char){1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14}));
+ v_b = vec_perm(v_b, v_b,
+ ((vector unsigned char){1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14}));
+ v_r = vec_perm(v_r, v_r,
+ ((vector unsigned char){1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14}));
+ }
+
+
+ v_b1 = vec_mergeh(v_b, vec_splats((unsigned short)0x0000));
+ v_g1 = vec_mergeh(v_g, vec_splats((unsigned short)0x0000));
+ v_r1 = vec_mergeh(v_r, vec_splats((unsigned short)0x0000));
+
+ v_g = vec_mergel(v_g, vec_splats((unsigned short)0x0000));
+ v_b = vec_mergel(v_b, vec_splats((unsigned short)0x0000));
+ v_r = vec_mergel(v_r, vec_splats((unsigned short)0x0000));
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_ry);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gy ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_by ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_ry);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gy ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_by ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dst_addr);
+
+ src[0] += 8;
+ src[1] += 8;
+ src[2] += 8;
+ dst_addr += 16;
+ }
+ for (i=width_adj; i< width; i++) {
+ int g = rdpx(src[0]);
+ int b = rdpx(src[1]);
+ int r = rdpx(src[2]);
+
+ dst[i] = ((ry*r + gy*g + by*b +
+ (33 << (RGB2YUV_SHIFT + bpc - 9))) >> (RGB2YUV_SHIFT + sh - 14));
+ ++src[0];
+ ++src[1];
+ ++src[2];
+ }
+}
+
+static av_always_inline
+void planar_rgb16_to_a_vsx(uint8_t *_dst, const uint8_t *_src[4],
+ int width, int bpc, int is_be, int32_t *rgb2yuv)
+{
+
+ int i, width_adj;
+ vector unsigned short v_rd0, v_a, v_dst, shift;
+ const uint16_t **src = (const uint16_t **)_src;
+ uint16_t *dst = (uint16_t *)_dst;
+ int sh = bpc < 16 ? bpc : 14;
+
+ uintptr_t dst_addr = (uintptr_t)_dst;
+
+
+ width_adj = width&(~(int)0x07);
+ if(width_adj)
+ shift = vec_splats((unsigned short)(14 - sh));
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (unsigned short *)src[3]);
+ if(is_be)
+ v_rd0 = vec_perm(v_rd0, v_rd0,
+ ((vector unsigned char){1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14}));
+
+
+ v_dst = vec_sl(v_rd0, shift);
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src[3] += 8;
+ dst_addr += 16;
+ }
+ for (i=width_adj; i< width; i++){
+ dst[i] = rdpx(src[3]) << (14 - sh);
+ ++src[3];
+ }
+}
+
+static av_always_inline
+void planar_rgb16_to_uv_vsx(uint8_t *_dstU, uint8_t *_dstV,
+ const uint8_t *_src[4], int width,
+ int bpc, int is_be, int32_t *rgb2yuv)
+{
+
+ int i, width_adj;
+ vector unsigned short v_rd0, v_rd1, v_rd2, v_g, v_b, v_r, v_g1, v_b1, v_r1;
+ vector unsigned int v_dst1, v_dst2;
+ vector unsigned int shift1, shift2;
+ const uint16_t **src = (const uint16_t **)_src;
+ uint16_t *dstU = (uint16_t *)_dstU;
+ uint16_t *dstV = (uint16_t *)_dstV;
+ vector signed int v_ru, v_gu, v_bu, v_rv, v_gv, v_bv;
+ int32_t ru = rgb2yuv[RU_IDX], gu = rgb2yuv[GU_IDX], bu = rgb2yuv[BU_IDX];
+ int32_t rv = rgb2yuv[RV_IDX], gv = rgb2yuv[GV_IDX], bv = rgb2yuv[BV_IDX];
+ int sh = bpc < 16 ? bpc : 14;
+
+ uintptr_t src_addr = (uintptr_t)src;
+ uintptr_t dstU_addr = (uintptr_t)_dstU;
+ uintptr_t dstV_addr = (uintptr_t)_dstV;
+
+
+ width_adj = width&(~(int)0x07);
+
+ if(width_adj){
+ shift1 = vec_splats((unsigned int)(257 << (RGB2YUV_SHIFT + bpc - 9)));
+ shift2 = vec_splats((unsigned int)(RGB2YUV_SHIFT + sh - 14));
+ v_ru = vec_splats((signed int)ru);
+ v_gu = vec_splats((signed int)gu);
+ v_bu = vec_splats((signed int)bu);
+ v_rv = vec_splats((signed int)rv);
+ v_gv = vec_splats((signed int)gv);
+ v_bv = vec_splats((signed int)bv);
+ }
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_g = vec_vsx_ld(0, (unsigned short *)src[0]);
+ v_b = vec_vsx_ld(0, (unsigned short *)(src[1]));
+ v_r = vec_vsx_ld(0, (unsigned short *)(src[2]));
+ if(is_be){
+ v_g = vec_perm(v_g, v_g,
+ ((vector unsigned char){1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14}));
+ v_b = vec_perm(v_b, v_b,
+ ((vector unsigned char){1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14}));
+ v_r = vec_perm(v_r, v_r,
+ ((vector unsigned char){1,0,3,2,5,4,7,6,9,8,11,10,13,12,15,14}));
+ }
+
+
+
+ v_b1 = vec_mergeh(v_b, vec_splats((unsigned short)0x0000));
+ v_g1 = vec_mergeh(v_g, vec_splats((unsigned short)0x0000));
+ v_r1 = vec_mergeh(v_r, vec_splats((unsigned short)0x0000));
+
+ v_g = vec_mergel(v_g, vec_splats((unsigned short)0x0000));
+ v_b = vec_mergel(v_b, vec_splats((unsigned short)0x0000));
+ v_r = vec_mergel(v_r, vec_splats((unsigned short)0x0000));
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_ru);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gu ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_bu ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_ru);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gu ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_bu ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstU_addr);
+
+ v_dst1 = (vector unsigned int)vec_mul((vector signed int)v_r1, v_rv);
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_g1, v_gv ));
+ v_dst1 = (vector unsigned int)vec_add((vector signed int)v_dst1,
+ vec_mul((vector signed int)v_b1, v_bv ));
+ v_dst1 = vec_add(v_dst1, shift1);
+ v_dst1 = vec_sr(v_dst1, shift2);
+ v_dst2 = (vector unsigned int)vec_mul((vector signed int)v_r, v_rv);
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_g, v_gv ));
+ v_dst2 = (vector unsigned int)vec_add((vector signed int)v_dst2,
+ vec_mul((vector signed int)v_b, v_bv ));
+ v_dst2 = vec_add(v_dst2, shift1);
+ v_dst2 = vec_sr(v_dst2, shift2);
+ v_dst1 = vec_perm(v_dst1, v_dst2,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst1, 0, (unsigned char *)dstV_addr);
+
+ src[0] += 8;
+ src[1] += 8;
+ src[2] += 8;
+ dstU_addr += 16;
+ dstV_addr += 16;
+ }
+ for (i=width_adj; i< width; i++) {
+ int g = rdpx(src[0]);
+ int b = rdpx(src[1]);
+ int r = rdpx(src[2]);
+
+ dstU[i] = (ru*r + gu*g + bu*b +
+ (257 << (RGB2YUV_SHIFT + bpc - 9))) >> (RGB2YUV_SHIFT + sh - 14);
+ dstV[i] = (rv*r + gv*g + bv*b +
+ (257 << (RGB2YUV_SHIFT + bpc - 9))) >> (RGB2YUV_SHIFT + sh - 14);
+ ++src[0];
+ ++src[1];
+ ++src[2];
+ }
+}
+#undef rdpx
+
+static av_always_inline
+void grayf32ToY16_c_vsx(uint8_t *_dst, const uint8_t *_src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int i, width_adj;
+ vector float v_rd0, v_rd1;
+ vector signed int v_rd00, v_rd01, v_rd02, v_rd03;
+ vector unsigned short v_dst;
+ const float *src = (const float *)_src;
+ uint16_t *dst = (uint16_t *)_dst;
+
+ uintptr_t dst_addr = (uintptr_t)_dst;
+ uintptr_t src_addr = (uintptr_t)_src;
+
+
+ width_adj = width&(~(int)0x07);
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (float *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (float *)(src_addr+16));
+
+ v_rd0 = vec_rint(vec_mul(v_rd0, vec_splats((float)65535.0f)));
+ v_rd1 = vec_rint(vec_mul(v_rd1, vec_splats((float)65535.0f)));
+ v_rd00 = (vector signed int)vec_cts(v_rd0, 0);
+ v_rd01 = (vector signed int)vec_cts(v_rd1, 0);
+ v_rd02 = (vector signed int)vec_cmpgt((vector unsigned int)v_rd00,
+ vec_splats((unsigned int)0xFFFF));
+ v_rd03 = (vector signed int)vec_cmpgt((vector unsigned int)v_rd01,
+ vec_splats((unsigned int)0xFFFF));
+ v_rd00 = vec_or(v_rd00, v_rd02);
+ v_rd01 = vec_or(v_rd01, v_rd03);
+
+ v_dst = (vector unsigned short)vec_perm(v_rd00, v_rd01,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+ for (i=width_adj; i< width; i++){
+ dst[i] = av_clip_uint16(lrintf(65535.0f * src[i]));
+ }
+}
+static av_always_inline
+void grayf32ToY16_bswap_c_vsx(uint8_t *_dst, const uint8_t *_src,
+ const uint8_t *unused1, const uint8_t *unused2,
+ int width, uint32_t *unused)
+{
+ int i, width_adj;
+ vector signed int v_rd0, v_rd1, v_rd2, v_rd3;
+ vector float v_rd00, v_rd01;
+ vector unsigned short v_dst;
+ const uint32_t *src = (const float *)_src;
+ uint16_t *dst = (uint16_t *)_dst;
+
+ uintptr_t dst_addr = (uintptr_t)_dst;
+ uintptr_t src_addr = (uintptr_t)_src;
+
+
+ width_adj = width&(~(int)0x07);
+
+ for (i = 0; i < width_adj; i+=8) {
+ v_rd0 = vec_vsx_ld(0, (int *)src_addr);
+ v_rd1 = vec_vsx_ld(0, (int *)(src_addr+16));
+
+ v_rd0 = vec_perm(v_rd0, v_rd0,
+ ((vector unsigned char){3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12}));
+ v_rd1 = vec_perm(v_rd1, v_rd1,
+ ((vector unsigned char){3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12}));
+ v_rd00 = vec_rint(vec_mul((vector float)v_rd0, vec_splats((float)65535.0f)));
+ v_rd01 = vec_rint(vec_mul((vector float)v_rd1, vec_splats((float)65535.0f)));
+ v_rd0 = vec_cts(v_rd00, 0);
+ v_rd1 = vec_cts(v_rd01, 0);
+ v_rd2 = (vector signed int)vec_cmpgt((vector unsigned int)v_rd0,
+ vec_splats((unsigned int)0xFFFF));
+ v_rd3 = (vector signed int)vec_cmpgt((vector unsigned int)v_rd1,
+ vec_splats((unsigned int)0xFFFF));
+ v_rd0 = vec_or(v_rd0, v_rd2);
+ v_rd1 = vec_or(v_rd1, v_rd3);
+
+ v_dst = (vector unsigned short)vec_perm(v_rd0, v_rd1,
+ ((vector unsigned char){0, 1, 4, 5, 8, 9, 12, 13, 16, 17, 20, 21, 24, 25, 28, 29}));
+ vec_vsx_st((vector unsigned char)v_dst, 0, (unsigned char *)dst_addr);
+
+ src_addr += 32;
+ dst_addr += 16;
+ }
+ for (i=width_adj; i< width; i++){
+ dst[i] = av_clip_uint16(lrintf(65535.0f * av_int2float(av_bswap32(src[i]))));
+ }
+}
+
+#define rgb9plus_planar_funcs_endian(nbits, endian_name, endian) \
+static void planar_rgb##nbits##endian_name##_to_y_vsx(uint8_t *dst, const uint8_t *src[4], \
+ int w, int32_t *rgb2yuv) \
+{ \
+ planar_rgb16_to_y_vsx(dst, src, w, nbits, endian, rgb2yuv); \
+} \
+static void planar_rgb##nbits##endian_name##_to_uv_vsx(uint8_t *dstU, uint8_t *dstV, \
+ const uint8_t *src[4], int w, int32_t *rgb2yuv) \
+{ \
+ planar_rgb16_to_uv_vsx(dstU, dstV, src, w, nbits, endian, rgb2yuv); \
+} \
+
+
+#define rgb9plus_planar_transparency_funcs(nbits) \
+static void planar_rgb##nbits##le_to_a_vsx(uint8_t *dst, const uint8_t *src[4], \
+ int w, int32_t *rgb2yuv) \
+{ \
+ planar_rgb16_to_a_vsx(dst, src, w, nbits, 0, rgb2yuv); \
+} \
+static void planar_rgb##nbits##be_to_a_vsx(uint8_t *dst, const uint8_t *src[4], \
+ int w, int32_t *rgb2yuv) \
+{ \
+ planar_rgb16_to_a_vsx(dst, src, w, nbits, 1, rgb2yuv); \
+}
+
+#define rgb9plus_planar_funcs(nbits) \
+ rgb9plus_planar_funcs_endian(nbits, le, 0) \
+ rgb9plus_planar_funcs_endian(nbits, be, 1)
+
+rgb9plus_planar_funcs(9)
+rgb9plus_planar_funcs(10)
+rgb9plus_planar_funcs(12)
+rgb9plus_planar_funcs(14)
+rgb9plus_planar_funcs(16)
+
+rgb9plus_planar_transparency_funcs(10)
+rgb9plus_planar_transparency_funcs(12)
+rgb9plus_planar_transparency_funcs(16)
+
+av_cold void ff_sws_init_input_funcs_vsx(SwsContext *c)
+{
+ enum AVPixelFormat srcFormat = c->srcFormat;
+
+ c->chrToYV12 = NULL;
+ switch (srcFormat) {
+ case AV_PIX_FMT_YUYV422:
+ c->chrToYV12 = yuy2ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_YVYU422:
+ c->chrToYV12 = yvy2ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_UYVY422:
+ c->chrToYV12 = uyvyToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_NV12:
+ c->chrToYV12 = nv12ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_NV21:
+ c->chrToYV12 = nv21ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB8:
+ case AV_PIX_FMT_BGR8:
+ case AV_PIX_FMT_PAL8:
+ case AV_PIX_FMT_BGR4_BYTE:
+ case AV_PIX_FMT_RGB4_BYTE:
+ c->chrToYV12 = palToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_GBRP9LE:
+ c->readChrPlanar = planar_rgb9le_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP10LE:
+ case AV_PIX_FMT_GBRP10LE:
+ c->readChrPlanar = planar_rgb10le_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP12LE:
+ case AV_PIX_FMT_GBRP12LE:
+ c->readChrPlanar = planar_rgb12le_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRP14LE:
+ c->readChrPlanar = planar_rgb14le_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP16LE:
+ case AV_PIX_FMT_GBRP16LE:
+ c->readChrPlanar = planar_rgb16le_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRP9BE:
+ c->readChrPlanar = planar_rgb9be_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP10BE:
+ case AV_PIX_FMT_GBRP10BE:
+ c->readChrPlanar = planar_rgb10be_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP12BE:
+ case AV_PIX_FMT_GBRP12BE:
+ c->readChrPlanar = planar_rgb12be_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRP14BE:
+ c->readChrPlanar = planar_rgb14be_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP16BE:
+ case AV_PIX_FMT_GBRP16BE:
+ c->readChrPlanar = planar_rgb16be_to_uv_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP:
+ case AV_PIX_FMT_GBRP:
+ c->readChrPlanar = planar_rgb_to_uv_vsx;
+ break;
+#if HAVE_BIGENDIAN
+ case AV_PIX_FMT_YUV420P9LE:
+ case AV_PIX_FMT_YUV422P9LE:
+ case AV_PIX_FMT_YUV444P9LE:
+ case AV_PIX_FMT_YUV420P10LE:
+ case AV_PIX_FMT_YUV422P10LE:
+ case AV_PIX_FMT_YUV440P10LE:
+ case AV_PIX_FMT_YUV444P10LE:
+ case AV_PIX_FMT_YUV420P12LE:
+ case AV_PIX_FMT_YUV422P12LE:
+ case AV_PIX_FMT_YUV440P12LE:
+ case AV_PIX_FMT_YUV444P12LE:
+ case AV_PIX_FMT_YUV420P14LE:
+ case AV_PIX_FMT_YUV422P14LE:
+ case AV_PIX_FMT_YUV444P14LE:
+ case AV_PIX_FMT_YUV420P16LE:
+ case AV_PIX_FMT_YUV422P16LE:
+ case AV_PIX_FMT_YUV444P16LE:
+
+ case AV_PIX_FMT_YUVA420P9LE:
+ case AV_PIX_FMT_YUVA422P9LE:
+ case AV_PIX_FMT_YUVA444P9LE:
+ case AV_PIX_FMT_YUVA420P10LE:
+ case AV_PIX_FMT_YUVA422P10LE:
+ case AV_PIX_FMT_YUVA444P10LE:
+ case AV_PIX_FMT_YUVA422P12LE:
+ case AV_PIX_FMT_YUVA444P12LE:
+ case AV_PIX_FMT_YUVA420P16LE:
+ case AV_PIX_FMT_YUVA422P16LE:
+ case AV_PIX_FMT_YUVA444P16LE:
+ c->chrToYV12 = bswap16UV_c_vsx;
+ break;
+#else
+ case AV_PIX_FMT_YUV420P9BE:
+ case AV_PIX_FMT_YUV422P9BE:
+ case AV_PIX_FMT_YUV444P9BE:
+ case AV_PIX_FMT_YUV420P10BE:
+ case AV_PIX_FMT_YUV422P10BE:
+ case AV_PIX_FMT_YUV440P10BE:
+ case AV_PIX_FMT_YUV444P10BE:
+ case AV_PIX_FMT_YUV420P12BE:
+ case AV_PIX_FMT_YUV422P12BE:
+ case AV_PIX_FMT_YUV440P12BE:
+ case AV_PIX_FMT_YUV444P12BE:
+ case AV_PIX_FMT_YUV420P14BE:
+ case AV_PIX_FMT_YUV422P14BE:
+ case AV_PIX_FMT_YUV444P14BE:
+ case AV_PIX_FMT_YUV420P16BE:
+ case AV_PIX_FMT_YUV422P16BE:
+ case AV_PIX_FMT_YUV444P16BE:
+
+ case AV_PIX_FMT_YUVA420P9BE:
+ case AV_PIX_FMT_YUVA422P9BE:
+ case AV_PIX_FMT_YUVA444P9BE:
+ case AV_PIX_FMT_YUVA420P10BE:
+ case AV_PIX_FMT_YUVA422P10BE:
+ case AV_PIX_FMT_YUVA444P10BE:
+ case AV_PIX_FMT_YUVA422P12BE:
+ case AV_PIX_FMT_YUVA444P12BE:
+ case AV_PIX_FMT_YUVA420P16BE:
+ case AV_PIX_FMT_YUVA422P16BE:
+ case AV_PIX_FMT_YUVA444P16BE:
+ c->chrToYV12 = bswap16UV_c_vsx;
+ break;
+#endif
+ case AV_PIX_FMT_AYUV64LE:
+ c->chrToYV12 = read_ayuv64le_UV_c_vsx;
+ break;
+ case AV_PIX_FMT_P010LE:
+ c->chrToYV12 = p010LEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_P010BE:
+ c->chrToYV12 = p010BEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_P016LE:
+ c->chrToYV12 = p016LEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_P016BE:
+ c->chrToYV12 = p016BEToUV_c_vsx;
+ break;
+ }
+ if (c->chrSrcHSubSample) {
+ switch (srcFormat) {
+ case AV_PIX_FMT_GBRAP:
+ case AV_PIX_FMT_GBRP:
+ c->chrToYV12 = gbr24pToUV_half_c_vsx;
+ break;
+ }
+ } else {
+ switch (srcFormat) {
+ case AV_PIX_FMT_RGBA64BE:
+ c->chrToYV12 = rgb64BEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGBA64LE:
+ c->chrToYV12 = rgb64LEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGRA64BE:
+ c->chrToYV12 = bgr64BEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGRA64LE:
+ c->chrToYV12 = bgr64LEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB48BE:
+ c->chrToYV12 = rgb48BEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB48LE:
+ c->chrToYV12 = rgb48LEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR48BE:
+ c->chrToYV12 = bgr48BEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR48LE:
+ c->chrToYV12 = bgr48LEToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB32:
+ c->chrToYV12 = bgr32ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB32_1:
+ c->chrToYV12 = bgr321ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR24:
+ c->chrToYV12 = bgr24ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR565LE:
+ c->chrToYV12 = bgr16leToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR565BE:
+ c->chrToYV12 = bgr16beToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR555LE:
+ c->chrToYV12 = bgr15leToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR555BE:
+ c->chrToYV12 = bgr15beToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR444LE:
+ c->chrToYV12 = bgr12leToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR444BE:
+ c->chrToYV12 = bgr12beToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR32:
+ c->chrToYV12 = rgb32ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR32_1:
+ c->chrToYV12 = rgb321ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB24:
+ c->chrToYV12 = rgb24ToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB565LE:
+ c->chrToYV12 = rgb16leToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB565BE:
+ c->chrToYV12 = rgb16beToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB555LE:
+ c->chrToYV12 = rgb15leToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB555BE:
+ c->chrToYV12 = rgb15beToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB444LE:
+ c->chrToYV12 = rgb12leToUV_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB444BE:
+ c->chrToYV12 = rgb12beToUV_c_vsx;
+ break;
+ }
+ }
+
+ c->lumToYV12 = NULL;
+ c->alpToYV12 = NULL;
+ switch (srcFormat) {
+ case AV_PIX_FMT_GBRP9LE:
+ c->readLumPlanar = planar_rgb9le_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP10LE:
+ c->readAlpPlanar = planar_rgb10le_to_a_vsx;
+ case AV_PIX_FMT_GBRP10LE:
+ c->readLumPlanar = planar_rgb10le_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP12LE:
+ c->readAlpPlanar = planar_rgb12le_to_a_vsx;
+ case AV_PIX_FMT_GBRP12LE:
+ c->readLumPlanar = planar_rgb12le_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRP14LE:
+ c->readLumPlanar = planar_rgb14le_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP16LE:
+ c->readAlpPlanar = planar_rgb16le_to_a_vsx;
+ case AV_PIX_FMT_GBRP16LE:
+ c->readLumPlanar = planar_rgb16le_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRP9BE:
+ c->readLumPlanar = planar_rgb9be_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP10BE:
+ c->readAlpPlanar = planar_rgb10be_to_a_vsx;
+ case AV_PIX_FMT_GBRP10BE:
+ c->readLumPlanar = planar_rgb10be_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP12BE:
+ c->readAlpPlanar = planar_rgb12be_to_a_vsx;
+ case AV_PIX_FMT_GBRP12BE:
+ c->readLumPlanar = planar_rgb12be_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRP14BE:
+ c->readLumPlanar = planar_rgb14be_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP16BE:
+ c->readAlpPlanar = planar_rgb16be_to_a_vsx;
+ case AV_PIX_FMT_GBRP16BE:
+ c->readLumPlanar = planar_rgb16be_to_y_vsx;
+ break;
+ case AV_PIX_FMT_GBRAP:
+ c->readAlpPlanar = planar_rgb_to_a_vsx;
+ case AV_PIX_FMT_GBRP:
+ c->readLumPlanar = planar_rgb_to_y_vsx;
+ break;
+#if HAVE_BIGENDIAN
+ case AV_PIX_FMT_YUV420P9LE:
+ case AV_PIX_FMT_YUV422P9LE:
+ case AV_PIX_FMT_YUV444P9LE:
+ case AV_PIX_FMT_YUV420P10LE:
+ case AV_PIX_FMT_YUV422P10LE:
+ case AV_PIX_FMT_YUV440P10LE:
+ case AV_PIX_FMT_YUV444P10LE:
+ case AV_PIX_FMT_YUV420P12LE:
+ case AV_PIX_FMT_YUV422P12LE:
+ case AV_PIX_FMT_YUV440P12LE:
+ case AV_PIX_FMT_YUV444P12LE:
+ case AV_PIX_FMT_YUV420P14LE:
+ case AV_PIX_FMT_YUV422P14LE:
+ case AV_PIX_FMT_YUV444P14LE:
+ case AV_PIX_FMT_YUV420P16LE:
+ case AV_PIX_FMT_YUV422P16LE:
+ case AV_PIX_FMT_YUV444P16LE:
+
+ case AV_PIX_FMT_GRAY9LE:
+ case AV_PIX_FMT_GRAY10LE:
+ case AV_PIX_FMT_GRAY12LE:
+ case AV_PIX_FMT_GRAY14LE:
+ case AV_PIX_FMT_GRAY16LE:
+
+ case AV_PIX_FMT_P016LE:
+ c->lumToYV12 = bswap16Y_c_vsx;
+ break;
+ case AV_PIX_FMT_YUVA420P9LE:
+ case AV_PIX_FMT_YUVA422P9LE:
+ case AV_PIX_FMT_YUVA444P9LE:
+ case AV_PIX_FMT_YUVA420P10LE:
+ case AV_PIX_FMT_YUVA422P10LE:
+ case AV_PIX_FMT_YUVA444P10LE:
+ case AV_PIX_FMT_YUVA422P12LE:
+ case AV_PIX_FMT_YUVA444P12LE:
+ case AV_PIX_FMT_YUVA420P16LE:
+ case AV_PIX_FMT_YUVA422P16LE:
+ case AV_PIX_FMT_YUVA444P16LE:
+ c->lumToYV12 = bswap16Y_c_vsx;
+ c->alpToYV12 = bswap16Y_c_vsx;
+ break;
+#else
+ case AV_PIX_FMT_YUV420P9BE:
+ case AV_PIX_FMT_YUV422P9BE:
+ case AV_PIX_FMT_YUV444P9BE:
+ case AV_PIX_FMT_YUV420P10BE:
+ case AV_PIX_FMT_YUV422P10BE:
+ case AV_PIX_FMT_YUV440P10BE:
+ case AV_PIX_FMT_YUV444P10BE:
+ case AV_PIX_FMT_YUV420P12BE:
+ case AV_PIX_FMT_YUV422P12BE:
+ case AV_PIX_FMT_YUV440P12BE:
+ case AV_PIX_FMT_YUV444P12BE:
+ case AV_PIX_FMT_YUV420P14BE:
+ case AV_PIX_FMT_YUV422P14BE:
+ case AV_PIX_FMT_YUV444P14BE:
+ case AV_PIX_FMT_YUV420P16BE:
+ case AV_PIX_FMT_YUV422P16BE:
+ case AV_PIX_FMT_YUV444P16BE:
+
+ case AV_PIX_FMT_GRAY9BE:
+ case AV_PIX_FMT_GRAY10BE:
+ case AV_PIX_FMT_GRAY12BE:
+ case AV_PIX_FMT_GRAY14BE:
+ case AV_PIX_FMT_GRAY16BE:
+
+ case AV_PIX_FMT_P016BE:
+ c->lumToYV12 = bswap16Y_c_vsx;
+ break;
+ case AV_PIX_FMT_YUVA420P9BE:
+ case AV_PIX_FMT_YUVA422P9BE:
+ case AV_PIX_FMT_YUVA444P9BE:
+ case AV_PIX_FMT_YUVA420P10BE:
+ case AV_PIX_FMT_YUVA422P10BE:
+ case AV_PIX_FMT_YUVA444P10BE:
+ case AV_PIX_FMT_YUVA422P12BE:
+ case AV_PIX_FMT_YUVA444P12BE:
+ case AV_PIX_FMT_YUVA420P16BE:
+ case AV_PIX_FMT_YUVA422P16BE:
+ case AV_PIX_FMT_YUVA444P16BE:
+ c->lumToYV12 = bswap16Y_c_vsx;
+ c->alpToYV12 = bswap16Y_c_vsx;
+ break;
+#endif
+ case AV_PIX_FMT_YA16LE:
+ c->lumToYV12 = read_ya16le_gray_c_vsx;
+ break;
+ case AV_PIX_FMT_YA16BE:
+ c->lumToYV12 = read_ya16be_gray_c_vsx;
+ break;
+ case AV_PIX_FMT_AYUV64LE:
+ c->lumToYV12 = read_ayuv64le_Y_c_vsx;
+ break;
+ case AV_PIX_FMT_YUYV422:
+ case AV_PIX_FMT_YVYU422:
+ case AV_PIX_FMT_YA8:
+ c->lumToYV12 = yuy2ToY_c_vsx;
+ break;
+ case AV_PIX_FMT_UYVY422:
+ c->lumToYV12 = uyvyToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR24:
+ c->lumToYV12 = bgr24ToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR565LE:
+ c->lumToYV12 = bgr16leToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR565BE:
+ c->lumToYV12 = bgr16beToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR555LE:
+ c->lumToYV12 = bgr15leToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR555BE:
+ c->lumToYV12 = bgr15beToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR444LE:
+ c->lumToYV12 = bgr12leToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR444BE:
+ c->lumToYV12 = bgr12beToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB24:
+ c->lumToYV12 = rgb24ToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB565LE:
+ c->lumToYV12 = rgb16leToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB565BE:
+ c->lumToYV12 = rgb16beToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB555LE:
+ c->lumToYV12 = rgb15leToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB555BE:
+ c->lumToYV12 = rgb15beToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB444LE:
+ c->lumToYV12 = rgb12leToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB444BE:
+ c->lumToYV12 = rgb12beToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB8:
+ case AV_PIX_FMT_BGR8:
+ case AV_PIX_FMT_PAL8:
+ case AV_PIX_FMT_BGR4_BYTE:
+ case AV_PIX_FMT_RGB4_BYTE:
+ c->lumToYV12 = palToY_c_vsx;
+ break;
+ case AV_PIX_FMT_MONOBLACK:
+ c->lumToYV12 = monoblack2Y_c_vsx;
+ break;
+ case AV_PIX_FMT_MONOWHITE:
+ c->lumToYV12 = monowhite2Y_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB32:
+ c->lumToYV12 = bgr32ToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB32_1:
+ c->lumToYV12 = bgr321ToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR32:
+ c->lumToYV12 = rgb32ToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR32_1:
+ c->lumToYV12 = rgb321ToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB48BE:
+ c->lumToYV12 = rgb48BEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGB48LE:
+ c->lumToYV12 = rgb48LEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR48BE:
+ c->lumToYV12 = bgr48BEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGR48LE:
+ c->lumToYV12 = bgr48LEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGBA64BE:
+ c->lumToYV12 = rgb64BEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_RGBA64LE:
+ c->lumToYV12 = rgb64LEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGRA64BE:
+ c->lumToYV12 = bgr64BEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_BGRA64LE:
+ c->lumToYV12 = bgr64LEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_P010LE:
+ c->lumToYV12 = p010LEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_P010BE:
+ c->lumToYV12 = p010BEToY_c_vsx;
+ break;
+ case AV_PIX_FMT_GRAYF32LE:
+#if HAVE_BIGENDIAN
+ c->lumToYV12 = grayf32ToY16_bswap_c_vsx;
+#else
+ c->lumToYV12 = grayf32ToY16_c_vsx;
+#endif
+ break;
+ case AV_PIX_FMT_GRAYF32BE:
+#if HAVE_BIGENDIAN
+ c->lumToYV12 = grayf32ToY16_c_vsx;
+#else
+ c->lumToYV12 = grayf32ToY16_bswap_c_vsx;
+#endif
+ break;
+ }
+ if (c->needAlpha) {
+ if (is16BPS(srcFormat) || isNBPS(srcFormat)) {
+ if (HAVE_BIGENDIAN == !isBE(srcFormat) && !c->readAlpPlanar)
+ c->alpToYV12 = bswap16Y_c_vsx;
+ }
+ switch (srcFormat) {
+ case AV_PIX_FMT_BGRA64LE:
+ case AV_PIX_FMT_RGBA64LE: c->alpToYV12 = rgba64leToA_c_vsx; break;
+ case AV_PIX_FMT_BGRA64BE:
+ case AV_PIX_FMT_RGBA64BE: c->alpToYV12 = rgba64beToA_c_vsx; break;
+ case AV_PIX_FMT_BGRA:
+ case AV_PIX_FMT_RGBA:
+ c->alpToYV12 = rgbaToA_c_vsx;
+ break;
+ case AV_PIX_FMT_ABGR:
+ case AV_PIX_FMT_ARGB:
+ c->alpToYV12 = abgrToA_c_vsx;
+ break;
+ case AV_PIX_FMT_YA8:
+ c->alpToYV12 = uyvyToY_c_vsx;
+ break;
+ case AV_PIX_FMT_YA16LE:
+ c->alpToYV12 = read_ya16le_alpha_c_vsx;
+ break;
+ case AV_PIX_FMT_YA16BE:
+ c->alpToYV12 = read_ya16be_alpha_c_vsx;
+ break;
+ case AV_PIX_FMT_AYUV64LE:
+ c->alpToYV12 = read_ayuv64le_A_c_vsx;
+ break;
+ case AV_PIX_FMT_PAL8 :
+ c->alpToYV12 = palToA_c_vsx;
+ break;
+ }
+ }
+}
+#endif //!HAVE_BIGENDIAN
+#endif //HAVE_VSX
\ No newline at end of file
diff --git a/libswscale/swscale.c b/libswscale/swscale.c
index 40695503ad..a4ab8eba4c 100644
--- a/libswscale/swscale.c
+++ b/libswscale/swscale.c
@@ -571,6 +571,9 @@ static av_cold void sws_init_swscale(SwsContext *c)
&c->yuv2packed2, &c->yuv2packedX, &c->yuv2anyX);
ff_sws_init_input_funcs(c);
+#if HAVE_VSX
+ ff_sws_init_input_funcs_vsx(c);
+#endif
if (c->srcBpc == 8) {
diff --git a/libswscale/swscale_internal.h b/libswscale/swscale_internal.h
index a59d12745a..e5f0e9d93d 100644
--- a/libswscale/swscale_internal.h
+++ b/libswscale/swscale_internal.h
@@ -859,6 +859,7 @@ void ff_get_unscaled_swscale_aarch64(SwsContext *c);
SwsFunc ff_getSwsFunc(SwsContext *c);
void ff_sws_init_input_funcs(SwsContext *c);
+void ff_sws_init_input_funcs_vsx(SwsContext *c);
void ff_sws_init_output_funcs(SwsContext *c,
yuv2planar1_fn *yuv2plane1,
yuv2planarX_fn *yuv2planeX,
--
2.14.1
More information about the ffmpeg-devel
mailing list