[FFmpeg-cvslog] avcodec/x86/hevc: updated to use x86util macros
plepere
git at videolan.org
Fri May 16 21:13:06 CEST 2014
ffmpeg | branch: master | plepere <pierre-edouard.lepere at insa-rennes.fr> | Fri May 16 11:44:20 2014 +0200| [ef7c4cd001a8c7c3349161789c6c0bb1fe0f8730] | committer: Michael Niedermayer
avcodec/x86/hevc: updated to use x86util macros
Reviewed-by: James Almer <jamrial at gmail.com>
Reviewed-by: Ronald S. Bultje
Signed-off-by: Michael Niedermayer <michaelni at gmx.at>
> http://git.videolan.org/gitweb.cgi/ffmpeg.git/?a=commit;h=ef7c4cd001a8c7c3349161789c6c0bb1fe0f8730
---
libavcodec/x86/hevc_deblock.asm | 155 +++++++++++++++++----------------------
1 file changed, 67 insertions(+), 88 deletions(-)
diff --git a/libavcodec/x86/hevc_deblock.asm b/libavcodec/x86/hevc_deblock.asm
index d27b131..b79ed40 100644
--- a/libavcodec/x86/hevc_deblock.asm
+++ b/libavcodec/x86/hevc_deblock.asm
@@ -57,11 +57,10 @@ INIT_XMM sse2
punpcklbw m5, m7
punpcklwd m4, m5
- movdqa m2, m0
+ punpckhdq m2, m0, m4
punpckldq m0, m4
- punpckhdq m2, m4
- movdqa m1, m0
- movdqa m3, m2
+ mova m1, m0
+ mova m3, m2
pxor m5, m5
punpcklbw m0, m5
@@ -81,10 +80,8 @@ INIT_XMM sse2
punpcklbw m0, m1
punpcklbw m2, m3
- movdqa m6, m0
-
+ punpckhwd m6, m0, m2
punpcklwd m0, m2
- punpckhwd m6, m2
movd %1, m0
pshufd m0, m0, 0x39
@@ -113,9 +110,8 @@ INIT_XMM sse2
punpcklwd m0, m2
punpcklwd m1, m3
- movdqa m2, m0
+ punpckhdq m2, m0, m1
punpckldq m0, m1
- punpckhdq m2, m1
movq m4, %5
movq m6, %6
@@ -124,16 +120,13 @@ INIT_XMM sse2
punpcklwd m4, m6
punpcklwd m5, m7
- movdqa m6, m4
+ punpckhdq m6, m4, m5
punpckldq m4, m5
- punpckhdq m6, m5
- movdqa m1, m0
+ punpckhqdq m1, m0, m4
punpcklqdq m0, m4
- punpckhqdq m1, m4
- movdqa m3, m2
+ punpckhqdq m3, m2, m6
punpcklqdq m2, m6
- punpckhqdq m3, m6
%endmacro
@@ -146,14 +139,13 @@ INIT_XMM sse2
CLIPW m2, m5, [pw_pixel_max]
CLIPW m3, m5, [pw_pixel_max]
- movdqa m4, m0
- movdqa m5, m2
+ mova m4, m0
+ mova m5, m2
punpcklwd m0, m1
punpcklwd m2, m3
- movdqa m6, m0
+ punpckhdq m6, m0, m2
punpckldq m0, m2
- punpckhdq m6, m2
movq %1, m0
punpckhqdq m0, m0
@@ -164,9 +156,8 @@ INIT_XMM sse2
punpckhwd m4, m1
punpckhwd m5, m3
- movdqa m6, m4
+ punpckhdq m6, m4, m5
punpckldq m4, m5
- punpckhdq m6, m5
movq %5, m4
punpckhqdq m4, m4
@@ -186,8 +177,7 @@ INIT_XMM sse2
punpcklbw m7, m2
punpcklbw m1, m3
- movdqa m3, m7
- punpcklwd m3, m1
+ punpcklwd m3, m7, m1
punpckhwd m7, m1
movq m4, %5
@@ -197,34 +187,27 @@ INIT_XMM sse2
punpcklbw m4, m6
punpcklbw m5, m15
- movdqa m9, m4
- punpcklwd m9, m5
+ punpcklwd m9, m4, m5
punpckhwd m4, m5
- movdqa m1, m3
- punpckldq m1, m9; 0, 1
+ punpckldq m1, m3, m9; 0, 1
punpckhdq m3, m9; 2, 3
- movdqa m5, m7
- punpckldq m5, m4; 4, 5
+ punpckldq m5, m7, m4; 4, 5
punpckhdq m7, m4; 6, 7
pxor m13, m13
- movdqa m0, m1
- punpcklbw m0, m13; 0 in 16 bit
+ punpcklbw m0, m1, m13; 0 in 16 bit
punpckhbw m1, m13; 1 in 16 bit
- movdqa m2, m3;
- punpcklbw m2, m13; 2
+ punpcklbw m2, m3, m13; 2
punpckhbw m3, m13; 3
- movdqa m4, m5;
- punpcklbw m4, m13; 4
+ punpcklbw m4, m5, m13; 4
punpckhbw m5, m13; 5
- movdqa m6, m7
- punpcklbw m6, m13; 6
+ punpcklbw m6, m7, m13; 6
punpckhbw m7, m13; 7
%endmacro
@@ -244,23 +227,19 @@ INIT_XMM sse2
punpcklbw m0, m1
punpcklbw m2, m3
- movdqa m8, m0
+ punpckhwd m8, m0, m2
punpcklwd m0, m2
- punpckhwd m8, m2
punpcklbw m4, m5
punpcklbw m6, m7
- movdqa m9, m4
+ punpckhwd m9, m4, m6
punpcklwd m4, m6
- punpckhwd m9, m6
- movdqa m10, m0
+ punpckhdq m10, m0, m4; 2, 3
punpckldq m0, m4; 0, 1
- punpckhdq m10, m4; 2, 3
- movdqa m11, m8
- punpckldq m11, m9; 4, 5
+ punpckldq m11, m8, m9; 4, 5
punpckhdq m8, m9; 6, 7
movq %1, m0
pshufd m0, m0, 0x4E
@@ -322,10 +301,10 @@ INIT_XMM sse2
; clobbers m10
%macro MASKED_COPY 2
pand %2, m11 ; and mask
- movdqa m10, m11
+ mova m10, m11
pandn m10, %1; and -mask
por %2, m10
- movdqa %1, %2
+ mova %1, %2
%endmacro
; in: %2 clobbered
@@ -335,14 +314,14 @@ INIT_XMM sse2
pand %2, %3 ; and mask
pandn %3, %1; and -mask
por %2, %3
- movdqa %1, %2
+ mova %1, %2
%endmacro
ALIGN 16
; input in m0 ... m3 and tcs in r2. Output in m1 and m2
%macro CHROMA_DEBLOCK_BODY 1
- movdqa m4, m2; temp copy of q0
- movdqa m5, m0; temp copy of p1
+ mova m4, m2; temp copy of q0
+ mova m5, m0; temp copy of p1
psubw m4, m1; q0 - p0
psubw m5, m3; p1 - q1
psllw m4, 2; << 2
@@ -355,7 +334,7 @@ ALIGN 16
movd m7, [r2]; tc1
punpcklwd m7, m7
shufps m6, m7, 0; tc0, tc1
- movdqa m4, m6
+ mova m4, m6
pcmpeqw m7, m7; set all bits to 1
pxor m4, m7; flip all bits of first reg
psrlw m7, 15; 1 in every cell
@@ -376,19 +355,19 @@ ALIGN 16
; input in m0 ... m7, betas in r2 tcs in r3. Output in m1...m6
%macro LUMA_DEBLOCK_BODY 2
- movdqa m9, m2
+ mova m9, m2
psllw m9, 1; *2
- movdqa m10, m1
+ mova m10, m1
psubw m10, m9
paddw m10, m3
- pabsw m10, m10 ; 0dp0, 0dp3 , 1dp0, 1dp3
+ ABS1 m10, m10 ; 0dp0, 0dp3 , 1dp0, 1dp3
- movdqa m9, m5
+ mova m9, m5
psllw m9, 1; *2
- movdqa m11, m6
+ mova m11, m6
psubw m11, m9
paddw m11, m4
- pabsw m11, m11 ; 0dq0, 0dq3 , 1dq0, 1dq3
+ ABS1 m11, m11 ; 0dq0, 0dq3 , 1dq0, 1dq3
;beta calculations
mov r11, [betaq];
@@ -403,7 +382,7 @@ ALIGN 16
pshufd m13, m14, 0; beta0, beta1
;end beta calculations
- movdqa m9, m10
+ mova m9, m10
paddw m9, m11; 0d0, 0d3 , 1d0, 1d3
pshufhw m14, m9, 0x0f ;0b00001111; 0d3 0d3 0d0 0d0 in high
@@ -413,7 +392,7 @@ ALIGN 16
pshuflw m9, m9, 0xf0 ;0b11110000; 1d0 1d0 1d3 1d3
paddw m14, m9; 0d0+0d3, 1d0+1d3
- movdqa m15, m13; beta0, beta1
+ mova m15, m13; beta0, beta1
;compare
pcmpgtw m15, m14
@@ -422,9 +401,9 @@ ALIGN 16
je bypasswrite_macro_%2%1
;weak / strong decision compare to beta_2
- movdqa m15, m13; beta0, beta1
+ mova m15, m13; beta0, beta1
psraw m15, 2; beta >> 2
- movdqa m8, m9;
+ mova m8, m9;
psllw m8, 1;
pcmpgtw m15, m8; (d0 << 1) < beta_2, (d3 << 1) < beta_2
movmskps r14, m15;
@@ -481,19 +460,19 @@ ALIGN 16
je bypasswrite_macro_%2%1
punpcklwd m9, m9
shufps m8, m9, 0; tc0, tc1
- movdqa m9, m8
+ mova m9, m8
psllw m8, 2; tc << 2
pavgw m8, m9; tc25 = ((tc * 5 + 1) >> 1)
;end tc25 calculations
;----beta_3 comparison-----
- movdqa m12, m0; p3
+ mova m12, m0; p3
psubw m12, m3; p3 - p0
- pabsw m12, m12; abs(p3 - p0)
+ ABS1 m12, m12; abs(p3 - p0)
- movdqa m15, m7; q3
+ mova m15, m7; q3
psubw m15, m4; q3 - q0
- pabsw m15, m15; abs(q3 - q0)
+ ABS1 m15, m15; abs(q3 - q0)
paddw m12, m15; abs(p3 - p0) + abs(q3 - q0)
@@ -506,9 +485,9 @@ ALIGN 16
and r14, r2; strong mask , beta_2 and beta_3 comparisons
;----beta_3 comparison end-----
;----tc25 comparison---
- movdqa m12, m3; p0
+ mova m12, m3; p0
psubw m12, m4; p0 - q0
- pabsw m12, m12; abs(p0 - q0)
+ ABS1 m12, m12; abs(p0 - q0)
pshufhw m12, m12, 0xf0 ;0b11110000;
pshuflw m12, m12, 0xf0 ;0b11110000;
@@ -522,7 +501,7 @@ ALIGN 16
and r14, r2; strong mask, bits 2 and 0
pcmpeqw m13, m13; set all bits to 1
- movdqa m14, m9; tc
+ mova m14, m9; tc
pxor m14, m13; invert bits
psrlw m13, 15; 1 in every cell
paddw m14, m13; -tc
@@ -549,10 +528,10 @@ ALIGN 16
psllw m13, 2; 4 in every cell
pand m11, m10; combine filtering mask and strong mask
- movdqa m12, m2; p1
+ mova m12, m2; p1
paddw m12, m3; p1 + p0
paddw m12, m4; p1 + p0 + q0
- movdqa m10, m12; copy
+ mova m10, m12; copy
psllw m12, 1; 2*p1 + 2*p0 + 2*q0
paddw m12, m1; p2 + 2*p1 + 2*p0 + 2*q0
paddw m12, m5; p2 + 2*p1 + 2*p0 + 2*q0 + q1
@@ -563,7 +542,7 @@ ALIGN 16
pminsw m12, m9; av_clip( , -2 * tc, 2 * tc)
paddw m12, m3; p0'
- movdqa m15, m1; p2
+ mova m15, m1; p2
paddw m15, m10; p2 + p1 + p0 + q0
psrlw m13, 1; 2 in every cell
paddw m15, m13; p2 + p1 + p0 + q0 + 2
@@ -573,7 +552,7 @@ ALIGN 16
pminsw m15, m9; av_clip( , -2 * tc, 2 * tc)
paddw m15, m2; p1'
- movdqa m8, m1; p2
+ mova m8, m1; p2
paddw m8, m0; p3 + p2
psllw m8, 1; 2*p3 + 2*p2
paddw m8, m1; 2*p3 + 3*p2
@@ -587,7 +566,7 @@ ALIGN 16
paddw m8, m1; p2'
MASKED_COPY m1, m8
- movdqa m8, m3; p0
+ mova m8, m3; p0
paddw m8, m4; p0 + q0
paddw m8, m5; p0 + q0 + q1
psllw m8, 1; 2*p0 + 2*q0 + 2*q1
@@ -601,10 +580,10 @@ ALIGN 16
paddw m8, m4; q0'
MASKED_COPY m2, m15
- movdqa m15, m3; p0
+ mova m15, m3; p0
paddw m15, m4; p0 + q0
paddw m15, m5; p0 + q0 + q1
- movdqa m10, m15;
+ mova m10, m15;
paddw m15, m6; p0 + q0 + q1 + q2
psrlw m13, 1; 2 in every cell
paddw m15, m13; p0 + q0 + q1 + q2 + 2
@@ -662,24 +641,24 @@ weakfilter_macro_%2%1:
psrlw m13, 15; 1 in every cell
psllw m13, 3; 8 in every cell
- movdqa m12, m4 ; q0
+ mova m12, m4 ; q0
psubw m12, m3 ; q0 - p0
- movdqa m10, m12
+ mova m10, m12
psllw m10, 3; 8 * (q0 - p0)
paddw m12, m10 ; 9 * (q0 - p0)
- movdqa m10, m5 ; q1
+ mova m10, m5 ; q1
psubw m10, m2 ; q1 - p1
- movdqa m8, m10
+ mova m8, m10
psllw m8, 1; 2 * ( q1 - p1 )
paddw m10, m8; 3 * ( q1 - p1 )
psubw m12, m10; 9 * (q0 - p0) - 3 * ( q1 - p1 )
paddw m12, m13; + 8
psraw m12, 4; >> 4 , delta0
- pabsw m13, m12; abs(delta0)
+ PABSW m13, m12; abs(delta0)
- movdqa m10, m9; 2*tc
+ mova m10, m9; 2*tc
psllw m10, 2; 8 * tc
paddw m10, m9; 10 * tc
pcmpgtw m10, m13
@@ -693,12 +672,12 @@ weakfilter_macro_%2%1:
pcmpeqw m13, m13; set all bits to 1
psraw m9, 1; tc -> tc / 2
- movdqa m14, m9;
+ mova m14, m9;
pxor m14, m13; complement -tc
psrlw m13, 15; set all cells to 1
paddw m14, m13; add 1, -tc / 2
- movdqa m15, m1; p2
+ mova m15, m1; p2
pavgw m15, m3; (p2 + p0 + 1) >> 1
psubw m15, m2; ((p2 + p0 + 1) >> 1) - p1
paddw m15, m12; ((p2 + p0 + 1) >> 1) - p1 + delta0
@@ -719,13 +698,13 @@ weakfilter_macro_%2%1:
punpcklwd m8, m8
punpcklwd m13, m13
shufps m13, m8, 0;
- movdqa m8, m10; copy of beta
+ mova m8, m10; copy of beta
pcmpgtw m8, m13
pand m8, m11
;end beta calculations
MASKED_COPY2 m2, m15, m8; write p1'
- movdqa m8, m6; q2
+ mova m8, m6; q2
pavgw m8, m4; (q2 + q0 + 1) >> 1
psubw m8, m5; ((q2 + q0 + 1) >> 1) - q1
psubw m8, m12; ((q2 + q0 + 1) >> 1) - q1 - delta0)
@@ -744,11 +723,11 @@ weakfilter_macro_%2%1:
pand m10, m11
MASKED_COPY2 m5, m8, m10; write q1'
- movdqa m15, m3 ; p0
+ mova m15, m3 ; p0
paddw m15, m12 ; p0 + delta0
MASKED_COPY m3, m15
- movdqa m8, m4 ; q0
+ mova m8, m4 ; q0
psubw m8, m12 ; q0 - delta0
MASKED_COPY m4, m8
ready_macro_%2%1:
More information about the ffmpeg-cvslog
mailing list