From ef7c4cd001a8c7c3349161789c6c0bb1fe0f8730 Mon Sep 17 00:00:00 2001 From: plepere Date: Fri, 16 May 2014 11:44:20 +0200 Subject: [PATCH] avcodec/x86/hevc: updated to use x86util macros Reviewed-by: James Almer Reviewed-by: Ronald S. Bultje Signed-off-by: Michael Niedermayer --- libavcodec/x86/hevc_deblock.asm | 155 ++++++++++++++------------------ 1 file changed, 67 insertions(+), 88 deletions(-) diff --git a/libavcodec/x86/hevc_deblock.asm b/libavcodec/x86/hevc_deblock.asm index d27b131586..b79ed40d3b 100644 --- a/libavcodec/x86/hevc_deblock.asm +++ b/libavcodec/x86/hevc_deblock.asm @@ -57,11 +57,10 @@ INIT_XMM sse2 punpcklbw m5, m7 punpcklwd m4, m5 - movdqa m2, m0 + punpckhdq m2, m0, m4 punpckldq m0, m4 - punpckhdq m2, m4 - movdqa m1, m0 - movdqa m3, m2 + mova m1, m0 + mova m3, m2 pxor m5, m5 punpcklbw m0, m5 @@ -81,10 +80,8 @@ INIT_XMM sse2 punpcklbw m0, m1 punpcklbw m2, m3 - movdqa m6, m0 - + punpckhwd m6, m0, m2 punpcklwd m0, m2 - punpckhwd m6, m2 movd %1, m0 pshufd m0, m0, 0x39 @@ -113,9 +110,8 @@ INIT_XMM sse2 punpcklwd m0, m2 punpcklwd m1, m3 - movdqa m2, m0 + punpckhdq m2, m0, m1 punpckldq m0, m1 - punpckhdq m2, m1 movq m4, %5 movq m6, %6 @@ -124,16 +120,13 @@ INIT_XMM sse2 punpcklwd m4, m6 punpcklwd m5, m7 - movdqa m6, m4 + punpckhdq m6, m4, m5 punpckldq m4, m5 - punpckhdq m6, m5 - movdqa m1, m0 + punpckhqdq m1, m0, m4 punpcklqdq m0, m4 - punpckhqdq m1, m4 - movdqa m3, m2 + punpckhqdq m3, m2, m6 punpcklqdq m2, m6 - punpckhqdq m3, m6 %endmacro @@ -146,14 +139,13 @@ INIT_XMM sse2 CLIPW m2, m5, [pw_pixel_max] CLIPW m3, m5, [pw_pixel_max] - movdqa m4, m0 - movdqa m5, m2 + mova m4, m0 + mova m5, m2 punpcklwd m0, m1 punpcklwd m2, m3 - movdqa m6, m0 + punpckhdq m6, m0, m2 punpckldq m0, m2 - punpckhdq m6, m2 movq %1, m0 punpckhqdq m0, m0 @@ -164,9 +156,8 @@ INIT_XMM sse2 punpckhwd m4, m1 punpckhwd m5, m3 - movdqa m6, m4 + punpckhdq m6, m4, m5 punpckldq m4, m5 - punpckhdq m6, m5 movq %5, m4 punpckhqdq m4, m4 @@ -186,8 +177,7 @@ INIT_XMM sse2 punpcklbw m7, m2 punpcklbw m1, m3 - movdqa m3, m7 - punpcklwd m3, m1 + punpcklwd m3, m7, m1 punpckhwd m7, m1 movq m4, %5 @@ -197,34 +187,27 @@ INIT_XMM sse2 punpcklbw m4, m6 punpcklbw m5, m15 - movdqa m9, m4 - punpcklwd m9, m5 + punpcklwd m9, m4, m5 punpckhwd m4, m5 - movdqa m1, m3 - punpckldq m1, m9; 0, 1 + punpckldq m1, m3, m9; 0, 1 punpckhdq m3, m9; 2, 3 - movdqa m5, m7 - punpckldq m5, m4; 4, 5 + punpckldq m5, m7, m4; 4, 5 punpckhdq m7, m4; 6, 7 pxor m13, m13 - movdqa m0, m1 - punpcklbw m0, m13; 0 in 16 bit + punpcklbw m0, m1, m13; 0 in 16 bit punpckhbw m1, m13; 1 in 16 bit - movdqa m2, m3; - punpcklbw m2, m13; 2 + punpcklbw m2, m3, m13; 2 punpckhbw m3, m13; 3 - movdqa m4, m5; - punpcklbw m4, m13; 4 + punpcklbw m4, m5, m13; 4 punpckhbw m5, m13; 5 - movdqa m6, m7 - punpcklbw m6, m13; 6 + punpcklbw m6, m7, m13; 6 punpckhbw m7, m13; 7 %endmacro @@ -244,23 +227,19 @@ INIT_XMM sse2 punpcklbw m0, m1 punpcklbw m2, m3 - movdqa m8, m0 + punpckhwd m8, m0, m2 punpcklwd m0, m2 - punpckhwd m8, m2 punpcklbw m4, m5 punpcklbw m6, m7 - movdqa m9, m4 + punpckhwd m9, m4, m6 punpcklwd m4, m6 - punpckhwd m9, m6 - movdqa m10, m0 + punpckhdq m10, m0, m4; 2, 3 punpckldq m0, m4; 0, 1 - punpckhdq m10, m4; 2, 3 - movdqa m11, m8 - punpckldq m11, m9; 4, 5 + punpckldq m11, m8, m9; 4, 5 punpckhdq m8, m9; 6, 7 movq %1, m0 pshufd m0, m0, 0x4E @@ -322,10 +301,10 @@ INIT_XMM sse2 ; clobbers m10 %macro MASKED_COPY 2 pand %2, m11 ; and mask - movdqa m10, m11 + mova m10, m11 pandn m10, %1; and -mask por %2, m10 - movdqa %1, %2 + mova %1, %2 %endmacro ; in: %2 clobbered @@ -335,14 +314,14 @@ INIT_XMM sse2 pand %2, %3 ; and mask pandn %3, %1; and -mask por %2, %3 - movdqa %1, %2 + mova %1, %2 %endmacro ALIGN 16 ; input in m0 ... m3 and tcs in r2. Output in m1 and m2 %macro CHROMA_DEBLOCK_BODY 1 - movdqa m4, m2; temp copy of q0 - movdqa m5, m0; temp copy of p1 + mova m4, m2; temp copy of q0 + mova m5, m0; temp copy of p1 psubw m4, m1; q0 - p0 psubw m5, m3; p1 - q1 psllw m4, 2; << 2 @@ -355,7 +334,7 @@ ALIGN 16 movd m7, [r2]; tc1 punpcklwd m7, m7 shufps m6, m7, 0; tc0, tc1 - movdqa m4, m6 + mova m4, m6 pcmpeqw m7, m7; set all bits to 1 pxor m4, m7; flip all bits of first reg psrlw m7, 15; 1 in every cell @@ -376,19 +355,19 @@ ALIGN 16 ; input in m0 ... m7, betas in r2 tcs in r3. Output in m1...m6 %macro LUMA_DEBLOCK_BODY 2 - movdqa m9, m2 + mova m9, m2 psllw m9, 1; *2 - movdqa m10, m1 + mova m10, m1 psubw m10, m9 paddw m10, m3 - pabsw m10, m10 ; 0dp0, 0dp3 , 1dp0, 1dp3 + ABS1 m10, m10 ; 0dp0, 0dp3 , 1dp0, 1dp3 - movdqa m9, m5 + mova m9, m5 psllw m9, 1; *2 - movdqa m11, m6 + mova m11, m6 psubw m11, m9 paddw m11, m4 - pabsw m11, m11 ; 0dq0, 0dq3 , 1dq0, 1dq3 + ABS1 m11, m11 ; 0dq0, 0dq3 , 1dq0, 1dq3 ;beta calculations mov r11, [betaq]; @@ -403,7 +382,7 @@ ALIGN 16 pshufd m13, m14, 0; beta0, beta1 ;end beta calculations - movdqa m9, m10 + mova m9, m10 paddw m9, m11; 0d0, 0d3 , 1d0, 1d3 pshufhw m14, m9, 0x0f ;0b00001111; 0d3 0d3 0d0 0d0 in high @@ -413,7 +392,7 @@ ALIGN 16 pshuflw m9, m9, 0xf0 ;0b11110000; 1d0 1d0 1d3 1d3 paddw m14, m9; 0d0+0d3, 1d0+1d3 - movdqa m15, m13; beta0, beta1 + mova m15, m13; beta0, beta1 ;compare pcmpgtw m15, m14 @@ -422,9 +401,9 @@ ALIGN 16 je bypasswrite_macro_%2%1 ;weak / strong decision compare to beta_2 - movdqa m15, m13; beta0, beta1 + mova m15, m13; beta0, beta1 psraw m15, 2; beta >> 2 - movdqa m8, m9; + mova m8, m9; psllw m8, 1; pcmpgtw m15, m8; (d0 << 1) < beta_2, (d3 << 1) < beta_2 movmskps r14, m15; @@ -481,19 +460,19 @@ ALIGN 16 je bypasswrite_macro_%2%1 punpcklwd m9, m9 shufps m8, m9, 0; tc0, tc1 - movdqa m9, m8 + mova m9, m8 psllw m8, 2; tc << 2 pavgw m8, m9; tc25 = ((tc * 5 + 1) >> 1) ;end tc25 calculations ;----beta_3 comparison----- - movdqa m12, m0; p3 + mova m12, m0; p3 psubw m12, m3; p3 - p0 - pabsw m12, m12; abs(p3 - p0) + ABS1 m12, m12; abs(p3 - p0) - movdqa m15, m7; q3 + mova m15, m7; q3 psubw m15, m4; q3 - q0 - pabsw m15, m15; abs(q3 - q0) + ABS1 m15, m15; abs(q3 - q0) paddw m12, m15; abs(p3 - p0) + abs(q3 - q0) @@ -506,9 +485,9 @@ ALIGN 16 and r14, r2; strong mask , beta_2 and beta_3 comparisons ;----beta_3 comparison end----- ;----tc25 comparison--- - movdqa m12, m3; p0 + mova m12, m3; p0 psubw m12, m4; p0 - q0 - pabsw m12, m12; abs(p0 - q0) + ABS1 m12, m12; abs(p0 - q0) pshufhw m12, m12, 0xf0 ;0b11110000; pshuflw m12, m12, 0xf0 ;0b11110000; @@ -522,7 +501,7 @@ ALIGN 16 and r14, r2; strong mask, bits 2 and 0 pcmpeqw m13, m13; set all bits to 1 - movdqa m14, m9; tc + mova m14, m9; tc pxor m14, m13; invert bits psrlw m13, 15; 1 in every cell paddw m14, m13; -tc @@ -549,10 +528,10 @@ ALIGN 16 psllw m13, 2; 4 in every cell pand m11, m10; combine filtering mask and strong mask - movdqa m12, m2; p1 + mova m12, m2; p1 paddw m12, m3; p1 + p0 paddw m12, m4; p1 + p0 + q0 - movdqa m10, m12; copy + mova m10, m12; copy psllw m12, 1; 2*p1 + 2*p0 + 2*q0 paddw m12, m1; p2 + 2*p1 + 2*p0 + 2*q0 paddw m12, m5; p2 + 2*p1 + 2*p0 + 2*q0 + q1 @@ -563,7 +542,7 @@ ALIGN 16 pminsw m12, m9; av_clip( , -2 * tc, 2 * tc) paddw m12, m3; p0' - movdqa m15, m1; p2 + mova m15, m1; p2 paddw m15, m10; p2 + p1 + p0 + q0 psrlw m13, 1; 2 in every cell paddw m15, m13; p2 + p1 + p0 + q0 + 2 @@ -573,7 +552,7 @@ ALIGN 16 pminsw m15, m9; av_clip( , -2 * tc, 2 * tc) paddw m15, m2; p1' - movdqa m8, m1; p2 + mova m8, m1; p2 paddw m8, m0; p3 + p2 psllw m8, 1; 2*p3 + 2*p2 paddw m8, m1; 2*p3 + 3*p2 @@ -587,7 +566,7 @@ ALIGN 16 paddw m8, m1; p2' MASKED_COPY m1, m8 - movdqa m8, m3; p0 + mova m8, m3; p0 paddw m8, m4; p0 + q0 paddw m8, m5; p0 + q0 + q1 psllw m8, 1; 2*p0 + 2*q0 + 2*q1 @@ -601,10 +580,10 @@ ALIGN 16 paddw m8, m4; q0' MASKED_COPY m2, m15 - movdqa m15, m3; p0 + mova m15, m3; p0 paddw m15, m4; p0 + q0 paddw m15, m5; p0 + q0 + q1 - movdqa m10, m15; + mova m10, m15; paddw m15, m6; p0 + q0 + q1 + q2 psrlw m13, 1; 2 in every cell paddw m15, m13; p0 + q0 + q1 + q2 + 2 @@ -662,24 +641,24 @@ weakfilter_macro_%2%1: psrlw m13, 15; 1 in every cell psllw m13, 3; 8 in every cell - movdqa m12, m4 ; q0 + mova m12, m4 ; q0 psubw m12, m3 ; q0 - p0 - movdqa m10, m12 + mova m10, m12 psllw m10, 3; 8 * (q0 - p0) paddw m12, m10 ; 9 * (q0 - p0) - movdqa m10, m5 ; q1 + mova m10, m5 ; q1 psubw m10, m2 ; q1 - p1 - movdqa m8, m10 + mova m8, m10 psllw m8, 1; 2 * ( q1 - p1 ) paddw m10, m8; 3 * ( q1 - p1 ) psubw m12, m10; 9 * (q0 - p0) - 3 * ( q1 - p1 ) paddw m12, m13; + 8 psraw m12, 4; >> 4 , delta0 - pabsw m13, m12; abs(delta0) + PABSW m13, m12; abs(delta0) - movdqa m10, m9; 2*tc + mova m10, m9; 2*tc psllw m10, 2; 8 * tc paddw m10, m9; 10 * tc pcmpgtw m10, m13 @@ -693,12 +672,12 @@ weakfilter_macro_%2%1: pcmpeqw m13, m13; set all bits to 1 psraw m9, 1; tc -> tc / 2 - movdqa m14, m9; + mova m14, m9; pxor m14, m13; complement -tc psrlw m13, 15; set all cells to 1 paddw m14, m13; add 1, -tc / 2 - movdqa m15, m1; p2 + mova m15, m1; p2 pavgw m15, m3; (p2 + p0 + 1) >> 1 psubw m15, m2; ((p2 + p0 + 1) >> 1) - p1 paddw m15, m12; ((p2 + p0 + 1) >> 1) - p1 + delta0 @@ -719,13 +698,13 @@ weakfilter_macro_%2%1: punpcklwd m8, m8 punpcklwd m13, m13 shufps m13, m8, 0; - movdqa m8, m10; copy of beta + mova m8, m10; copy of beta pcmpgtw m8, m13 pand m8, m11 ;end beta calculations MASKED_COPY2 m2, m15, m8; write p1' - movdqa m8, m6; q2 + mova m8, m6; q2 pavgw m8, m4; (q2 + q0 + 1) >> 1 psubw m8, m5; ((q2 + q0 + 1) >> 1) - q1 psubw m8, m12; ((q2 + q0 + 1) >> 1) - q1 - delta0) @@ -744,11 +723,11 @@ weakfilter_macro_%2%1: pand m10, m11 MASKED_COPY2 m5, m8, m10; write q1' - movdqa m15, m3 ; p0 + mova m15, m3 ; p0 paddw m15, m12 ; p0 + delta0 MASKED_COPY m3, m15 - movdqa m8, m4 ; q0 + mova m8, m4 ; q0 psubw m8, m12 ; q0 - delta0 MASKED_COPY m4, m8 ready_macro_%2%1: