;****************************************************************************** ;* H.264 intra prediction asm optimizations ;* Copyright (c) 2010 Fiona Glaser ;* Copyright (c) 2010 Holger Lubitz ;* Copyright (c) 2010 Loren Merritt ;* Copyright (c) 2010 Ronald S. Bultje ;* ;* This file is part of FFmpeg. ;* ;* FFmpeg is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* FFmpeg is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with FFmpeg; if not, write to the Free Software ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "libavutil/x86/x86util.asm" SECTION_RODATA tm_shuf: times 8 db 0x03, 0x80 pw_ff00: times 8 dw 0xff00 plane_shuf: db -8, -7, -6, -5, -4, -3, -2, -1 db 1, 2, 3, 4, 5, 6, 7, 8 plane8_shuf: db -4, -3, -2, -1, 0, 0, 0, 0 db 1, 2, 3, 4, 0, 0, 0, 0 pw_0to7: dw 0, 1, 2, 3, 4, 5, 6, 7 pw_1to8: dw 1, 2, 3, 4, 5, 6, 7, 8 pw_m8tom1: dw -8, -7, -6, -5, -4, -3, -2, -1 pw_m4to4: dw -4, -3, -2, -1, 1, 2, 3, 4 SECTION .text cextern pb_1 cextern pb_3 cextern pw_4 cextern pw_8 ;----------------------------------------------------------------------------- ; void ff_pred16x16_vertical_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_XMM sse cglobal pred16x16_vertical_8, 2,3 sub r0, r1 mov r2, 4 movaps xmm0, [r0] .loop: movaps [r0+r1*1], xmm0 movaps [r0+r1*2], xmm0 lea r0, [r0+r1*2] movaps [r0+r1*1], xmm0 movaps [r0+r1*2], xmm0 lea r0, [r0+r1*2] dec r2 jg .loop RET ;----------------------------------------------------------------------------- ; void ff_pred16x16_horizontal_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED16x16_H 0 cglobal pred16x16_horizontal_8, 2,3 mov r2, 8 %if cpuflag(ssse3) mova m2, [pb_3] %endif .loop: movd m0, [r0+r1*0-4] movd m1, [r0+r1*1-4] %if cpuflag(ssse3) pshufb m0, m2 pshufb m1, m2 %else punpcklbw m0, m0 punpcklbw m1, m1 SPLATW m0, m0, 3 SPLATW m1, m1, 3 mova [r0+r1*0+8], m0 mova [r0+r1*1+8], m1 %endif mova [r0+r1*0], m0 mova [r0+r1*1], m1 lea r0, [r0+r1*2] dec r2 jg .loop RET %endmacro INIT_MMX mmxext PRED16x16_H INIT_XMM ssse3 PRED16x16_H ;----------------------------------------------------------------------------- ; void ff_pred16x16_dc_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED16x16_DC 0 cglobal pred16x16_dc_8, 2,7 mov r4, r0 sub r0, r1 pxor mm0, mm0 pxor mm1, mm1 psadbw mm0, [r0+0] psadbw mm1, [r0+8] dec r0 movzx r5d, byte [r0+r1*1] paddw mm0, mm1 movd r6d, mm0 lea r0, [r0+r1*2] %rep 7 movzx r2d, byte [r0+r1*0] movzx r3d, byte [r0+r1*1] add r5d, r2d add r6d, r3d lea r0, [r0+r1*2] %endrep movzx r2d, byte [r0+r1*0] add r5d, r6d lea r2d, [r2+r5+16] shr r2d, 5 %if cpuflag(ssse3) pxor m1, m1 %endif SPLATB_REG m0, r2, m1 mov r3d, 4 .loop: mova [r4+r1*0], m0 mova [r4+r1*1], m0 lea r4, [r4+r1*2] mova [r4+r1*0], m0 mova [r4+r1*1], m0 lea r4, [r4+r1*2] dec r3d jg .loop RET %endmacro INIT_XMM sse2 PRED16x16_DC INIT_XMM ssse3 PRED16x16_DC ;----------------------------------------------------------------------------- ; void ff_pred16x16_tm_vp8_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_XMM sse2 cglobal pred16x16_tm_vp8_8, 2,6,6 sub r0, r1 pxor xmm2, xmm2 movdqa xmm0, [r0] movdqa xmm1, xmm0 punpcklbw xmm0, xmm2 punpckhbw xmm1, xmm2 movzx r4d, byte [r0-1] mov r5d, 8 .loop: movzx r2d, byte [r0+r1*1-1] movzx r3d, byte [r0+r1*2-1] sub r2d, r4d sub r3d, r4d movd xmm2, r2d movd xmm4, r3d pshuflw xmm2, xmm2, 0 pshuflw xmm4, xmm4, 0 punpcklqdq xmm2, xmm2 punpcklqdq xmm4, xmm4 movdqa xmm3, xmm2 movdqa xmm5, xmm4 paddw xmm2, xmm0 paddw xmm3, xmm1 paddw xmm4, xmm0 paddw xmm5, xmm1 packuswb xmm2, xmm3 packuswb xmm4, xmm5 movdqa [r0+r1*1], xmm2 movdqa [r0+r1*2], xmm4 lea r0, [r0+r1*2] dec r5d jg .loop RET %if HAVE_AVX2_EXTERNAL INIT_YMM avx2 cglobal pred16x16_tm_vp8_8, 2, 4, 5, dst, stride, stride3, iteration sub dstq, strideq pmovzxbw m0, [dstq] vpbroadcastb xm1, [r0-1] pmovzxbw m1, xm1 psubw m0, m1 mov iterationd, 4 lea stride3q, [strideq*3] .loop: vpbroadcastb xm1, [dstq+strideq*1-1] vpbroadcastb xm2, [dstq+strideq*2-1] vpbroadcastb xm3, [dstq+stride3q-1] vpbroadcastb xm4, [dstq+strideq*4-1] pmovzxbw m1, xm1 pmovzxbw m2, xm2 pmovzxbw m3, xm3 pmovzxbw m4, xm4 paddw m1, m0 paddw m2, m0 paddw m3, m0 paddw m4, m0 vpackuswb m1, m1, m2 vpackuswb m3, m3, m4 vpermq m1, m1, q3120 vpermq m3, m3, q3120 movdqa [dstq+strideq*1], xm1 vextracti128 [dstq+strideq*2], m1, 1 movdqa [dstq+stride3q*1], xm3 vextracti128 [dstq+strideq*4], m3, 1 lea dstq, [dstq+strideq*4] dec iterationd jg .loop RET %endif ;----------------------------------------------------------------------------- ; void ff_pred16x16_plane_*_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro H264_PRED16x16_PLANE 1 cglobal pred16x16_plane_%1_8, 2,9,7 mov r2, r1 ; +stride neg r1 ; -stride movh m0, [r0+r1 -1] %if cpuflag(ssse3) movhps m0, [r0+r1 +8] pmaddubsw m0, [plane_shuf] ; H coefficients %else ; sse2 pxor m2, m2 movh m1, [r0+r1 +8] punpcklbw m0, m2 punpcklbw m1, m2 pmullw m0, [pw_m8tom1] pmullw m1, [pw_1to8] paddw m0, m1 %endif movhlps m1, m0 paddw m0, m1 PSHUFLW m1, m0, 0xE paddw m0, m1 PSHUFLW m1, m0, 0x1 paddw m0, m1 ; sum of H coefficients lea r4, [r0+r2*8-1] lea r3, [r0+r2*4-1] add r4, r2 %if ARCH_X86_64 %define e_reg r8 %else %define e_reg r0 %endif movzx e_reg, byte [r3+r2*2 ] movzx r5, byte [r4+r1 ] sub r5, e_reg movzx e_reg, byte [r3+r2 ] movzx r6, byte [r4 ] sub r6, e_reg lea r5, [r5+r6*2] movzx e_reg, byte [r3+r1 ] movzx r6, byte [r4+r2*2 ] sub r6, e_reg lea r5, [r5+r6*4] movzx e_reg, byte [r3 ] %if ARCH_X86_64 movzx r7, byte [r4+r2 ] sub r7, e_reg %else movzx r6, byte [r4+r2 ] sub r6, e_reg lea r5, [r5+r6*4] sub r5, r6 %endif lea e_reg, [r3+r1*4] lea r3, [r4+r2*4] movzx r4, byte [e_reg+r2 ] movzx r6, byte [r3 ] sub r6, r4 %if ARCH_X86_64 lea r6, [r7+r6*2] lea r5, [r5+r6*2] add r5, r6 %else lea r5, [r5+r6*4] lea r5, [r5+r6*2] %endif movzx r4, byte [e_reg ] %if ARCH_X86_64 movzx r7, byte [r3 +r2 ] sub r7, r4 sub r5, r7 %else movzx r6, byte [r3 +r2 ] sub r6, r4 lea r5, [r5+r6*8] sub r5, r6 %endif movzx r4, byte [e_reg+r1 ] movzx r6, byte [r3 +r2*2] sub r6, r4 %if ARCH_X86_64 add r6, r7 %endif lea r5, [r5+r6*8] movzx r4, byte [e_reg+r2*2] movzx r6, byte [r3 +r1 ] sub r6, r4 lea r5, [r5+r6*4] add r5, r6 ; sum of V coefficients %if ARCH_X86_64 == 0 mov r0, r0m %endif %ifidn %1, h264 lea r5, [r5*5+32] sar r5, 6 %elifidn %1, rv40 lea r5, [r5*5] sar r5, 6 %elifidn %1, svq3 test r5, r5 lea r6, [r5+3] cmovs r5, r6 sar r5, 2 ; V/4 lea r5, [r5*5] ; 5*(V/4) test r5, r5 lea r6, [r5+15] cmovs r5, r6 sar r5, 4 ; (5*(V/4))/16 %endif movzx r4, byte [r0+r1 +15] movzx r3, byte [r3+r2*2 ] lea r3, [r3+r4+1] shl r3, 4 movd r1d, m0 movsx r1d, r1w %ifnidn %1, svq3 %ifidn %1, h264 lea r1d, [r1d*5+32] %else ; rv40 lea r1d, [r1d*5] %endif sar r1d, 6 %else ; svq3 test r1d, r1d lea r4d, [r1d+3] cmovs r1d, r4d sar r1d, 2 ; H/4 lea r1d, [r1d*5] ; 5*(H/4) test r1d, r1d lea r4d, [r1d+15] cmovs r1d, r4d sar r1d, 4 ; (5*(H/4))/16 %endif movd m0, r1d add r1d, r5d add r3d, r1d shl r1d, 3 sub r3d, r1d ; a movd m1, r5d movd m3, r3d SPLATW m0, m0, 0 ; H SPLATW m1, m1, 0 ; V SPLATW m3, m3, 0 ; a %ifidn %1, svq3 SWAP 0, 1 %endif mova m2, m0 pmullw m0, [pw_0to7] ; 0*H, 1*H, ..., 7*H (words) psllw m2, 3 paddw m0, m3 ; a + {0,1,2,3,4,5,6,7}*H paddw m2, m0 ; a + {8,9,10,11,12,13,14,15}*H mov r4, 8 .loop: mova m3, m0 ; b[0..7] mova m4, m2 ; b[8..15] psraw m3, 5 psraw m4, 5 packuswb m3, m4 mova [r0], m3 paddw m0, m1 paddw m2, m1 mova m3, m0 ; b[0..7] mova m4, m2 ; b[8..15] psraw m3, 5 psraw m4, 5 packuswb m3, m4 mova [r0+r2], m3 paddw m0, m1 paddw m2, m1 lea r0, [r0+r2*2] dec r4 jg .loop RET %endmacro INIT_XMM sse2 H264_PRED16x16_PLANE h264 H264_PRED16x16_PLANE rv40 H264_PRED16x16_PLANE svq3 INIT_XMM ssse3 H264_PRED16x16_PLANE h264 H264_PRED16x16_PLANE rv40 H264_PRED16x16_PLANE svq3 ;----------------------------------------------------------------------------- ; void ff_pred8x8_plane_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro H264_PRED8x8_PLANE 0 cglobal pred8x8_plane_8, 2,9,7 mov r2, r1 ; +stride neg r1 ; -stride movd m0, [r0+r1 -1] %if cpuflag(ssse3) movhps m0, [r0+r1 +4] ; this reads 4 bytes more than necessary pmaddubsw m0, [plane8_shuf] ; H coefficients %else ; sse2 pxor m2, m2 movd m1, [r0+r1 +4] punpckldq m0, m1 punpcklbw m0, m2 pmullw m0, [pw_m4to4] %endif movhlps m1, m0 paddw m0, m1 %if notcpuflag(ssse3) PSHUFLW m1, m0, 0xE paddw m0, m1 %endif ; !ssse3 PSHUFLW m1, m0, 0x1 paddw m0, m1 ; sum of H coefficients lea r4, [r0+r2*4-1] lea r3, [r0 -1] add r4, r2 %if ARCH_X86_64 %define e_reg r8 %else %define e_reg r0 %endif movzx e_reg, byte [r3+r2*2 ] movzx r5, byte [r4+r1 ] sub r5, e_reg movzx e_reg, byte [r3 ] %if ARCH_X86_64 movzx r7, byte [r4+r2 ] sub r7, e_reg sub r5, r7 %else movzx r6, byte [r4+r2 ] sub r6, e_reg lea r5, [r5+r6*4] sub r5, r6 %endif movzx e_reg, byte [r3+r1 ] movzx r6, byte [r4+r2*2 ] sub r6, e_reg %if ARCH_X86_64 add r6, r7 %endif lea r5, [r5+r6*4] movzx e_reg, byte [r3+r2 ] movzx r6, byte [r4 ] sub r6, e_reg lea r6, [r5+r6*2] lea r5, [r6*9+16] lea r5, [r5+r6*8] sar r5, 5 %if ARCH_X86_64 == 0 mov r0, r0m %endif movzx r3, byte [r4+r2*2 ] movzx r4, byte [r0+r1 +7] lea r3, [r3+r4+1] shl r3, 4 movd r1d, m0 movsx r1d, r1w imul r1d, 17 add r1d, 16 sar r1d, 5 movd m0, r1d add r1d, r5d sub r3d, r1d add r1d, r1d sub r3d, r1d ; a movd m1, r5d movd m3, r3d SPLATW m0, m0, 0 ; H SPLATW m1, m1, 0 ; V SPLATW m3, m3, 0 ; a pmullw m0, [pw_0to7] ; 0*H, 1*H, ..., 7*H (words) paddw m0, m3 ; a + {0,1,2,3,4,5,6,7}*H mov r4, 4 ALIGN 16 .loop: mova m3, m0 ; b[0..7] paddw m0, m1 psraw m3, 5 mova m4, m0 ; V+b[0..7] paddw m0, m1 psraw m4, 5 packuswb m3, m4 movh [r0], m3 movhps [r0+r2], m3 lea r0, [r0+r2*2] dec r4 jg .loop RET %endmacro INIT_XMM sse2 H264_PRED8x8_PLANE INIT_XMM ssse3 H264_PRED8x8_PLANE ;----------------------------------------------------------------------------- ; void ff_pred8x8_vertical_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_MMX mmx cglobal pred8x8_vertical_8, 2,2 sub r0, r1 movq mm0, [r0] %rep 3 movq [r0+r1*1], mm0 movq [r0+r1*2], mm0 lea r0, [r0+r1*2] %endrep movq [r0+r1*1], mm0 movq [r0+r1*2], mm0 RET ;----------------------------------------------------------------------------- ; void ff_pred8x8_horizontal_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8_H 0 cglobal pred8x8_horizontal_8, 2,3 mov r2, 4 %if cpuflag(ssse3) mova m2, [pb_3] %endif .loop: SPLATB_LOAD m0, r0+r1*0-1, m2 SPLATB_LOAD m1, r0+r1*1-1, m2 mova [r0+r1*0], m0 mova [r0+r1*1], m1 lea r0, [r0+r1*2] dec r2 jg .loop RET %endmacro INIT_MMX mmxext PRED8x8_H INIT_MMX ssse3 PRED8x8_H ;----------------------------------------------------------------------------- ; void ff_pred8x8_top_dc_8_mmxext(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_MMX mmxext cglobal pred8x8_top_dc_8, 2,5 sub r0, r1 movq mm0, [r0] pxor mm1, mm1 pxor mm2, mm2 lea r2, [r0+r1*2] punpckhbw mm1, mm0 punpcklbw mm0, mm2 psadbw mm1, mm2 ; s1 lea r3, [r2+r1*2] psadbw mm0, mm2 ; s0 psrlw mm1, 1 psrlw mm0, 1 pavgw mm1, mm2 lea r4, [r3+r1*2] pavgw mm0, mm2 pshufw mm1, mm1, 0 pshufw mm0, mm0, 0 ; dc0 (w) packuswb mm0, mm1 ; dc0,dc1 (b) movq [r0+r1*1], mm0 movq [r0+r1*2], mm0 lea r0, [r3+r1*2] movq [r2+r1*1], mm0 movq [r2+r1*2], mm0 movq [r3+r1*1], mm0 movq [r3+r1*2], mm0 movq [r0+r1*1], mm0 movq [r0+r1*2], mm0 RET ;----------------------------------------------------------------------------- ; void ff_pred8x8_dc_8_mmxext(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_MMX mmxext cglobal pred8x8_dc_8, 2,5 sub r0, r1 pxor m7, m7 movd m0, [r0+0] movd m1, [r0+4] psadbw m0, m7 ; s0 mov r4, r0 psadbw m1, m7 ; s1 movzx r2d, byte [r0+r1*1-1] movzx r3d, byte [r0+r1*2-1] lea r0, [r0+r1*2] add r2d, r3d movzx r3d, byte [r0+r1*1-1] add r2d, r3d movzx r3d, byte [r0+r1*2-1] add r2d, r3d lea r0, [r0+r1*2] movd m2, r2d ; s2 movzx r2d, byte [r0+r1*1-1] movzx r3d, byte [r0+r1*2-1] lea r0, [r0+r1*2] add r2d, r3d movzx r3d, byte [r0+r1*1-1] add r2d, r3d movzx r3d, byte [r0+r1*2-1] add r2d, r3d movd m3, r2d ; s3 punpcklwd m0, m1 mov r0, r4 punpcklwd m2, m3 punpckldq m0, m2 ; s0, s1, s2, s3 pshufw m3, m0, 11110110b ; s2, s1, s3, s3 lea r2, [r0+r1*2] pshufw m0, m0, 01110100b ; s0, s1, s3, s1 paddw m0, m3 lea r3, [r2+r1*2] psrlw m0, 2 pavgw m0, m7 ; s0+s2, s1, s3, s1+s3 lea r4, [r3+r1*2] packuswb m0, m0 punpcklbw m0, m0 movq m1, m0 punpcklbw m0, m0 punpckhbw m1, m1 movq [r0+r1*1], m0 movq [r0+r1*2], m0 movq [r2+r1*1], m0 movq [r2+r1*2], m0 movq [r3+r1*1], m1 movq [r3+r1*2], m1 movq [r4+r1*1], m1 movq [r4+r1*2], m1 RET ;----------------------------------------------------------------------------- ; void ff_pred8x8_dc_rv40_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_MMX mmxext cglobal pred8x8_dc_rv40_8, 2,7 mov r4, r0 sub r0, r1 pxor mm0, mm0 psadbw mm0, [r0] dec r0 movzx r5d, byte [r0+r1*1] movd r6d, mm0 lea r0, [r0+r1*2] %rep 3 movzx r2d, byte [r0+r1*0] movzx r3d, byte [r0+r1*1] add r5d, r2d add r6d, r3d lea r0, [r0+r1*2] %endrep movzx r2d, byte [r0+r1*0] add r5d, r6d lea r2d, [r2+r5+8] shr r2d, 4 movd mm0, r2d punpcklbw mm0, mm0 pshufw mm0, mm0, 0 mov r3d, 4 .loop: movq [r4+r1*0], mm0 movq [r4+r1*1], mm0 lea r4, [r4+r1*2] dec r3d jg .loop RET ;----------------------------------------------------------------------------- ; void ff_pred8x8_tm_vp8_8(uint8_t *src, ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_XMM sse2 cglobal pred8x8_tm_vp8_8, 2,6,4 sub r0, r1 pxor xmm1, xmm1 movq xmm0, [r0] punpcklbw xmm0, xmm1 movzx r4d, byte [r0-1] mov r5d, 4 .loop: movzx r2d, byte [r0+r1*1-1] movzx r3d, byte [r0+r1*2-1] sub r2d, r4d sub r3d, r4d movd xmm2, r2d movd xmm3, r3d pshuflw xmm2, xmm2, 0 pshuflw xmm3, xmm3, 0 punpcklqdq xmm2, xmm2 punpcklqdq xmm3, xmm3 paddw xmm2, xmm0 paddw xmm3, xmm0 packuswb xmm2, xmm3 movq [r0+r1*1], xmm2 movhps [r0+r1*2], xmm2 lea r0, [r0+r1*2] dec r5d jg .loop RET INIT_XMM ssse3 cglobal pred8x8_tm_vp8_8, 2,3,6 sub r0, r1 movdqa xmm4, [tm_shuf] pxor xmm1, xmm1 movq xmm0, [r0] punpcklbw xmm0, xmm1 movd xmm5, [r0-4] pshufb xmm5, xmm4 mov r2d, 4 .loop: movd xmm2, [r0+r1*1-4] movd xmm3, [r0+r1*2-4] pshufb xmm2, xmm4 pshufb xmm3, xmm4 psubw xmm2, xmm5 psubw xmm3, xmm5 paddw xmm2, xmm0 paddw xmm3, xmm0 packuswb xmm2, xmm3 movq [r0+r1*1], xmm2 movhps [r0+r1*2], xmm2 lea r0, [r0+r1*2] dec r2d jg .loop RET ; dest, left, right, src, tmp ; output: %1 = (t[n-1] + t[n]*2 + t[n+1] + 2) >> 2 %macro PRED4x4_LOWPASS 5 mova %5, %2 pavgb %2, %3 pxor %3, %5 mova %1, %4 pand %3, [pb_1] psubusb %2, %3 pavgb %1, %2 %endmacro ;----------------------------------------------------------------------------- ; void ff_pred8x8l_top_dc_8(uint8_t *src, int has_topleft, int has_topright, ; ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_TOP_DC 0 cglobal pred8x8l_top_dc_8, 4,4 sub r0, r3 pxor mm7, mm7 movq mm0, [r0-8] movq mm3, [r0] movq mm1, [r0+8] movq mm2, mm3 movq mm4, mm3 PALIGNR mm2, mm0, 7, mm0 PALIGNR mm1, mm4, 1, mm4 test r1d, r1d ; top_left jz .fix_lt_2 test r2d, r2d ; top_right jz .fix_tr_1 jmp .body .fix_lt_2: movq mm5, mm3 pxor mm5, mm2 psllq mm5, 56 psrlq mm5, 56 pxor mm2, mm5 test r2d, r2d ; top_right jnz .body .fix_tr_1: movq mm5, mm3 pxor mm5, mm1 psrlq mm5, 56 psllq mm5, 56 pxor mm1, mm5 .body: PRED4x4_LOWPASS mm0, mm2, mm1, mm3, mm5 psadbw mm7, mm0 paddw mm7, [pw_4] psrlw mm7, 3 pshufw mm7, mm7, 0 packuswb mm7, mm7 %rep 3 movq [r0+r3*1], mm7 movq [r0+r3*2], mm7 lea r0, [r0+r3*2] %endrep movq [r0+r3*1], mm7 movq [r0+r3*2], mm7 RET %endmacro INIT_MMX mmxext PRED8x8L_TOP_DC INIT_MMX ssse3 PRED8x8L_TOP_DC ;----------------------------------------------------------------------------- ; void ff_pred8x8l_dc_8(uint8_t *src, int has_topleft, int has_topright, ; ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_DC 0 cglobal pred8x8l_dc_8, 4,5 sub r0, r3 lea r4, [r0+r3*2] movq mm0, [r0+r3*1-8] punpckhbw mm0, [r0+r3*0-8] movq mm1, [r4+r3*1-8] punpckhbw mm1, [r0+r3*2-8] mov r4, r0 punpckhwd mm1, mm0 lea r0, [r0+r3*4] movq mm2, [r0+r3*1-8] punpckhbw mm2, [r0+r3*0-8] lea r0, [r0+r3*2] movq mm3, [r0+r3*1-8] punpckhbw mm3, [r0+r3*0-8] punpckhwd mm3, mm2 punpckhdq mm3, mm1 lea r0, [r0+r3*2] movq mm0, [r0+r3*0-8] movq mm1, [r4] mov r0, r4 movq mm4, mm3 movq mm2, mm3 PALIGNR mm4, mm0, 7, mm0 PALIGNR mm1, mm2, 1, mm2 test r1d, r1d jnz .do_left .fix_lt_1: movq mm5, mm3 pxor mm5, mm4 psrlq mm5, 56 psllq mm5, 48 pxor mm1, mm5 jmp .do_left .fix_lt_2: movq mm5, mm3 pxor mm5, mm2 psllq mm5, 56 psrlq mm5, 56 pxor mm2, mm5 test r2d, r2d jnz .body .fix_tr_1: movq mm5, mm3 pxor mm5, mm1 psrlq mm5, 56 psllq mm5, 56 pxor mm1, mm5 jmp .body .do_left: movq mm0, mm4 PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 movq mm4, mm0 movq mm7, mm2 PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 psllq mm1, 56 PALIGNR mm7, mm1, 7, mm3 movq mm0, [r0-8] movq mm3, [r0] movq mm1, [r0+8] movq mm2, mm3 movq mm4, mm3 PALIGNR mm2, mm0, 7, mm0 PALIGNR mm1, mm4, 1, mm4 test r1d, r1d jz .fix_lt_2 test r2d, r2d jz .fix_tr_1 .body: lea r1, [r0+r3*2] PRED4x4_LOWPASS mm6, mm2, mm1, mm3, mm5 pxor mm0, mm0 pxor mm1, mm1 lea r2, [r1+r3*2] psadbw mm0, mm7 psadbw mm1, mm6 paddw mm0, [pw_8] paddw mm0, mm1 lea r4, [r2+r3*2] psrlw mm0, 4 pshufw mm0, mm0, 0 packuswb mm0, mm0 movq [r0+r3*1], mm0 movq [r0+r3*2], mm0 movq [r1+r3*1], mm0 movq [r1+r3*2], mm0 movq [r2+r3*1], mm0 movq [r2+r3*2], mm0 movq [r4+r3*1], mm0 movq [r4+r3*2], mm0 RET %endmacro INIT_MMX mmxext PRED8x8L_DC INIT_MMX ssse3 PRED8x8L_DC ;----------------------------------------------------------------------------- ; void ff_pred8x8l_horizontal_8(uint8_t *src, int has_topleft, ; int has_topright, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_HORIZONTAL 0 cglobal pred8x8l_horizontal_8, 4,4 sub r0, r3 lea r2, [r0+r3*2] movq mm0, [r0+r3*1-8] test r1d, r1d lea r1, [r0+r3] cmovnz r1, r0 punpckhbw mm0, [r1+r3*0-8] movq mm1, [r2+r3*1-8] punpckhbw mm1, [r0+r3*2-8] mov r2, r0 punpckhwd mm1, mm0 lea r0, [r0+r3*4] movq mm2, [r0+r3*1-8] punpckhbw mm2, [r0+r3*0-8] lea r0, [r0+r3*2] movq mm3, [r0+r3*1-8] punpckhbw mm3, [r0+r3*0-8] punpckhwd mm3, mm2 punpckhdq mm3, mm1 lea r0, [r0+r3*2] movq mm0, [r0+r3*0-8] movq mm1, [r1+r3*0-8] mov r0, r2 movq mm4, mm3 movq mm2, mm3 PALIGNR mm4, mm0, 7, mm0 PALIGNR mm1, mm2, 1, mm2 movq mm0, mm4 PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 movq mm4, mm0 movq mm7, mm2 PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 psllq mm1, 56 PALIGNR mm7, mm1, 7, mm3 movq mm3, mm7 lea r1, [r0+r3*2] movq mm7, mm3 punpckhbw mm3, mm3 punpcklbw mm7, mm7 pshufw mm0, mm3, 0xff pshufw mm1, mm3, 0xaa lea r2, [r1+r3*2] pshufw mm2, mm3, 0x55 pshufw mm3, mm3, 0x00 pshufw mm4, mm7, 0xff pshufw mm5, mm7, 0xaa pshufw mm6, mm7, 0x55 pshufw mm7, mm7, 0x00 movq [r0+r3*1], mm0 movq [r0+r3*2], mm1 movq [r1+r3*1], mm2 movq [r1+r3*2], mm3 movq [r2+r3*1], mm4 movq [r2+r3*2], mm5 lea r0, [r2+r3*2] movq [r0+r3*1], mm6 movq [r0+r3*2], mm7 RET %endmacro INIT_MMX mmxext PRED8x8L_HORIZONTAL INIT_MMX ssse3 PRED8x8L_HORIZONTAL ;----------------------------------------------------------------------------- ; void ff_pred8x8l_vertical_8(uint8_t *src, int has_topleft, int has_topright, ; ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_VERTICAL 0 cglobal pred8x8l_vertical_8, 4,4 sub r0, r3 movq mm0, [r0-8] movq mm3, [r0] movq mm1, [r0+8] movq mm2, mm3 movq mm4, mm3 PALIGNR mm2, mm0, 7, mm0 PALIGNR mm1, mm4, 1, mm4 test r1d, r1d ; top_left jz .fix_lt_2 test r2d, r2d ; top_right jz .fix_tr_1 jmp .body .fix_lt_2: movq mm5, mm3 pxor mm5, mm2 psllq mm5, 56 psrlq mm5, 56 pxor mm2, mm5 test r2d, r2d ; top_right jnz .body .fix_tr_1: movq mm5, mm3 pxor mm5, mm1 psrlq mm5, 56 psllq mm5, 56 pxor mm1, mm5 .body: PRED4x4_LOWPASS mm0, mm2, mm1, mm3, mm5 %rep 3 movq [r0+r3*1], mm0 movq [r0+r3*2], mm0 lea r0, [r0+r3*2] %endrep movq [r0+r3*1], mm0 movq [r0+r3*2], mm0 RET %endmacro INIT_MMX mmxext PRED8x8L_VERTICAL INIT_MMX ssse3 PRED8x8L_VERTICAL ;----------------------------------------------------------------------------- ; void ff_pred8x8l_down_left_8(uint8_t *src, int has_topleft, ; int has_topright, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_DOWN_LEFT 0 cglobal pred8x8l_down_left_8, 4,4 sub r0, r3 movq mm0, [r0-8] movq mm3, [r0] movq mm1, [r0+8] movq mm2, mm3 movq mm4, mm3 PALIGNR mm2, mm0, 7, mm0 PALIGNR mm1, mm4, 1, mm4 test r1d, r1d ; top_left jz .fix_lt_2 test r2d, r2d ; top_right jz .fix_tr_1 jmp .do_top .fix_lt_2: movq mm5, mm3 pxor mm5, mm2 psllq mm5, 56 psrlq mm5, 56 pxor mm2, mm5 test r2d, r2d ; top_right jnz .do_top .fix_tr_1: movq mm5, mm3 pxor mm5, mm1 psrlq mm5, 56 psllq mm5, 56 pxor mm1, mm5 jmp .do_top .fix_tr_2: punpckhbw mm3, mm3 pshufw mm1, mm3, 0xFF jmp .do_topright .do_top: PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5 movq2dq xmm3, mm4 test r2d, r2d ; top_right jz .fix_tr_2 movq mm0, [r0+8] movq mm5, mm0 movq mm2, mm0 movq mm4, mm0 psrlq mm5, 56 PALIGNR mm2, mm3, 7, mm3 PALIGNR mm5, mm4, 1, mm4 PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4 .do_topright: movq2dq xmm4, mm1 psrlq mm1, 56 movq2dq xmm5, mm1 lea r1, [r0+r3*2] pslldq xmm4, 8 por xmm3, xmm4 movdqa xmm2, xmm3 psrldq xmm2, 1 pslldq xmm5, 15 por xmm2, xmm5 lea r2, [r1+r3*2] movdqa xmm1, xmm3 pslldq xmm1, 1 INIT_XMM cpuname PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm3, xmm4 psrldq xmm0, 1 movq [r0+r3*1], xmm0 psrldq xmm0, 1 movq [r0+r3*2], xmm0 psrldq xmm0, 1 lea r0, [r2+r3*2] movq [r1+r3*1], xmm0 psrldq xmm0, 1 movq [r1+r3*2], xmm0 psrldq xmm0, 1 movq [r2+r3*1], xmm0 psrldq xmm0, 1 movq [r2+r3*2], xmm0 psrldq xmm0, 1 movq [r0+r3*1], xmm0 psrldq xmm0, 1 movq [r0+r3*2], xmm0 RET %endmacro INIT_MMX sse2 PRED8x8L_DOWN_LEFT INIT_MMX ssse3 PRED8x8L_DOWN_LEFT ;----------------------------------------------------------------------------- ; void ff_pred8x8l_down_right_8(uint8_t *src, int has_topleft, ; int has_topright, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_DOWN_RIGHT 0 cglobal pred8x8l_down_right_8, 4,5 sub r0, r3 lea r4, [r0+r3*2] movq mm0, [r0+r3*1-8] punpckhbw mm0, [r0+r3*0-8] movq mm1, [r4+r3*1-8] punpckhbw mm1, [r0+r3*2-8] mov r4, r0 punpckhwd mm1, mm0 lea r0, [r0+r3*4] movq mm2, [r0+r3*1-8] punpckhbw mm2, [r0+r3*0-8] lea r0, [r0+r3*2] movq mm3, [r0+r3*1-8] punpckhbw mm3, [r0+r3*0-8] punpckhwd mm3, mm2 punpckhdq mm3, mm1 lea r0, [r0+r3*2] movq mm0, [r0+r3*0-8] movq mm1, [r4] mov r0, r4 movq mm4, mm3 movq mm2, mm3 PALIGNR mm4, mm0, 7, mm0 PALIGNR mm1, mm2, 1, mm2 test r1d, r1d jz .fix_lt_1 jmp .do_left .fix_lt_1: movq mm5, mm3 pxor mm5, mm4 psrlq mm5, 56 psllq mm5, 48 pxor mm1, mm5 jmp .do_left .fix_lt_2: movq mm5, mm3 pxor mm5, mm2 psllq mm5, 56 psrlq mm5, 56 pxor mm2, mm5 test r2d, r2d jnz .do_top .fix_tr_1: movq mm5, mm3 pxor mm5, mm1 psrlq mm5, 56 psllq mm5, 56 pxor mm1, mm5 jmp .do_top .do_left: movq mm0, mm4 PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 movq mm4, mm0 movq mm7, mm2 movq2dq xmm3, mm2 PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 psllq mm1, 56 PALIGNR mm7, mm1, 7, mm3 movq2dq xmm1, mm7 movq mm0, [r0-8] movq mm3, [r0] movq mm1, [r0+8] movq mm2, mm3 movq mm4, mm3 PALIGNR mm2, mm0, 7, mm0 PALIGNR mm1, mm4, 1, mm4 test r1d, r1d jz .fix_lt_2 test r2d, r2d jz .fix_tr_1 .do_top: PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5 movq2dq xmm4, mm4 lea r1, [r0+r3*2] movdqa xmm0, xmm3 pslldq xmm4, 8 por xmm3, xmm4 lea r2, [r1+r3*2] pslldq xmm4, 1 por xmm1, xmm4 psrldq xmm0, 7 pslldq xmm0, 15 psrldq xmm0, 7 por xmm1, xmm0 lea r0, [r2+r3*2] movdqa xmm2, xmm3 psrldq xmm2, 1 INIT_XMM cpuname PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm3, xmm4 movdqa xmm1, xmm0 psrldq xmm1, 1 movq [r0+r3*2], xmm0 movq [r0+r3*1], xmm1 psrldq xmm0, 2 psrldq xmm1, 2 movq [r2+r3*2], xmm0 movq [r2+r3*1], xmm1 psrldq xmm0, 2 psrldq xmm1, 2 movq [r1+r3*2], xmm0 movq [r1+r3*1], xmm1 psrldq xmm0, 2 psrldq xmm1, 2 movq [r4+r3*2], xmm0 movq [r4+r3*1], xmm1 RET %endmacro INIT_MMX sse2 PRED8x8L_DOWN_RIGHT INIT_MMX ssse3 PRED8x8L_DOWN_RIGHT ;----------------------------------------------------------------------------- ; void ff_pred8x8l_vertical_right_8(uint8_t *src, int has_topleft, ; int has_topright, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_VERTICAL_RIGHT 0 cglobal pred8x8l_vertical_right_8, 4,5,7 ; manually spill XMM registers for Win64 because ; the code here is initialized with INIT_MMX WIN64_SPILL_XMM 7 sub r0, r3 lea r4, [r0+r3*2] movq mm0, [r0+r3*1-8] punpckhbw mm0, [r0+r3*0-8] movq mm1, [r4+r3*1-8] punpckhbw mm1, [r0+r3*2-8] mov r4, r0 punpckhwd mm1, mm0 lea r0, [r0+r3*4] movq mm2, [r0+r3*1-8] punpckhbw mm2, [r0+r3*0-8] lea r0, [r0+r3*2] movq mm3, [r0+r3*1-8] punpckhbw mm3, [r0+r3*0-8] punpckhwd mm3, mm2 punpckhdq mm3, mm1 lea r0, [r0+r3*2] movq mm0, [r0+r3*0-8] movq mm1, [r4] mov r0, r4 movq mm4, mm3 movq mm2, mm3 PALIGNR mm4, mm0, 7, mm0 PALIGNR mm1, mm2, 1, mm2 test r1d, r1d jnz .do_left .fix_lt_1: movq mm5, mm3 pxor mm5, mm4 psrlq mm5, 56 psllq mm5, 48 pxor mm1, mm5 jmp .do_left .fix_lt_2: movq mm5, mm3 pxor mm5, mm2 psllq mm5, 56 psrlq mm5, 56 pxor mm2, mm5 test r2d, r2d jnz .do_top .fix_tr_1: movq mm5, mm3 pxor mm5, mm1 psrlq mm5, 56 psllq mm5, 56 pxor mm1, mm5 jmp .do_top .do_left: movq mm0, mm4 PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 movq2dq xmm0, mm2 movq mm0, [r0-8] movq mm3, [r0] movq mm1, [r0+8] movq mm2, mm3 movq mm4, mm3 PALIGNR mm2, mm0, 7, mm0 PALIGNR mm1, mm4, 1, mm4 test r1d, r1d jz .fix_lt_2 test r2d, r2d jz .fix_tr_1 .do_top: PRED4x4_LOWPASS mm6, mm2, mm1, mm3, mm5 lea r1, [r0+r3*2] movq2dq xmm4, mm6 pslldq xmm4, 8 por xmm0, xmm4 movdqa xmm6, [pw_ff00] movdqa xmm1, xmm0 lea r2, [r1+r3*2] movdqa xmm2, xmm0 movdqa xmm3, xmm0 pslldq xmm0, 1 pslldq xmm1, 2 pavgb xmm2, xmm0 INIT_XMM cpuname PRED4x4_LOWPASS xmm4, xmm3, xmm1, xmm0, xmm5 pandn xmm6, xmm4 movdqa xmm5, xmm4 psrlw xmm4, 8 packuswb xmm6, xmm4 movhlps xmm4, xmm6 movhps [r0+r3*2], xmm5 movhps [r0+r3*1], xmm2 psrldq xmm5, 4 movss xmm5, xmm6 psrldq xmm2, 4 movss xmm2, xmm4 lea r0, [r2+r3*2] psrldq xmm5, 1 psrldq xmm2, 1 movq [r0+r3*2], xmm5 movq [r0+r3*1], xmm2 psrldq xmm5, 1 psrldq xmm2, 1 movq [r2+r3*2], xmm5 movq [r2+r3*1], xmm2 psrldq xmm5, 1 psrldq xmm2, 1 movq [r1+r3*2], xmm5 movq [r1+r3*1], xmm2 RET %endmacro INIT_MMX sse2 PRED8x8L_VERTICAL_RIGHT INIT_MMX ssse3 PRED8x8L_VERTICAL_RIGHT ;----------------------------------------------------------------------------- ; void ff_pred8x8l_vertical_left_8(uint8_t *src, int has_topleft, ; int has_topright, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_VERTICAL_LEFT 0 cglobal pred8x8l_vertical_left_8, 4,4 sub r0, r3 movq mm0, [r0-8] movq mm3, [r0] movq mm1, [r0+8] movq mm2, mm3 movq mm4, mm3 PALIGNR mm2, mm0, 7, mm0 PALIGNR mm1, mm4, 1, mm4 test r1d, r1d jz .fix_lt_2 test r2d, r2d jz .fix_tr_1 jmp .do_top .fix_lt_2: movq mm5, mm3 pxor mm5, mm2 psllq mm5, 56 psrlq mm5, 56 pxor mm2, mm5 test r2d, r2d jnz .do_top .fix_tr_1: movq mm5, mm3 pxor mm5, mm1 psrlq mm5, 56 psllq mm5, 56 pxor mm1, mm5 jmp .do_top .fix_tr_2: punpckhbw mm3, mm3 pshufw mm1, mm3, 0xFF jmp .do_topright .do_top: PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5 movq2dq xmm4, mm4 test r2d, r2d jz .fix_tr_2 movq mm0, [r0+8] movq mm5, mm0 movq mm2, mm0 movq mm4, mm0 psrlq mm5, 56 PALIGNR mm2, mm3, 7, mm3 PALIGNR mm5, mm4, 1, mm4 PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4 .do_topright: movq2dq xmm3, mm1 lea r1, [r0+r3*2] pslldq xmm3, 8 por xmm4, xmm3 movdqa xmm2, xmm4 movdqa xmm1, xmm4 movdqa xmm3, xmm4 psrldq xmm2, 1 pslldq xmm1, 1 pavgb xmm3, xmm2 lea r2, [r1+r3*2] INIT_XMM cpuname PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm4, xmm5 psrldq xmm0, 1 movq [r0+r3*1], xmm3 movq [r0+r3*2], xmm0 lea r0, [r2+r3*2] psrldq xmm3, 1 psrldq xmm0, 1 movq [r1+r3*1], xmm3 movq [r1+r3*2], xmm0 psrldq xmm3, 1 psrldq xmm0, 1 movq [r2+r3*1], xmm3 movq [r2+r3*2], xmm0 psrldq xmm3, 1 psrldq xmm0, 1 movq [r0+r3*1], xmm3 movq [r0+r3*2], xmm0 RET %endmacro INIT_MMX sse2 PRED8x8L_VERTICAL_LEFT INIT_MMX ssse3 PRED8x8L_VERTICAL_LEFT ;----------------------------------------------------------------------------- ; void ff_pred8x8l_horizontal_up_8(uint8_t *src, int has_topleft, ; int has_topright, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_HORIZONTAL_UP 0 cglobal pred8x8l_horizontal_up_8, 4,4 sub r0, r3 lea r2, [r0+r3*2] movq mm0, [r0+r3*1-8] test r1d, r1d lea r1, [r0+r3] cmovnz r1, r0 punpckhbw mm0, [r1+r3*0-8] movq mm1, [r2+r3*1-8] punpckhbw mm1, [r0+r3*2-8] mov r2, r0 punpckhwd mm1, mm0 lea r0, [r0+r3*4] movq mm2, [r0+r3*1-8] punpckhbw mm2, [r0+r3*0-8] lea r0, [r0+r3*2] movq mm3, [r0+r3*1-8] punpckhbw mm3, [r0+r3*0-8] punpckhwd mm3, mm2 punpckhdq mm3, mm1 lea r0, [r0+r3*2] movq mm0, [r0+r3*0-8] movq mm1, [r1+r3*0-8] mov r0, r2 movq mm4, mm3 movq mm2, mm3 PALIGNR mm4, mm0, 7, mm0 PALIGNR mm1, mm2, 1, mm2 movq mm0, mm4 PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 movq mm4, mm0 movq mm7, mm2 PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 psllq mm1, 56 PALIGNR mm7, mm1, 7, mm3 lea r1, [r0+r3*2] pshufw mm0, mm7, 00011011b ; l6 l7 l4 l5 l2 l3 l0 l1 psllq mm7, 56 ; l7 .. .. .. .. .. .. .. movq mm2, mm0 psllw mm0, 8 psrlw mm2, 8 por mm2, mm0 ; l7 l6 l5 l4 l3 l2 l1 l0 movq mm3, mm2 movq mm4, mm2 movq mm5, mm2 psrlq mm2, 8 psrlq mm3, 16 lea r2, [r1+r3*2] por mm2, mm7 ; l7 l7 l6 l5 l4 l3 l2 l1 punpckhbw mm7, mm7 por mm3, mm7 ; l7 l7 l7 l6 l5 l4 l3 l2 pavgb mm4, mm2 PRED4x4_LOWPASS mm1, mm3, mm5, mm2, mm6 movq mm5, mm4 punpcklbw mm4, mm1 ; p4 p3 p2 p1 punpckhbw mm5, mm1 ; p8 p7 p6 p5 movq mm6, mm5 movq mm7, mm5 movq mm0, mm5 PALIGNR mm5, mm4, 2, mm1 pshufw mm1, mm6, 11111001b PALIGNR mm6, mm4, 4, mm2 pshufw mm2, mm7, 11111110b PALIGNR mm7, mm4, 6, mm3 pshufw mm3, mm0, 11111111b movq [r0+r3*1], mm4 movq [r0+r3*2], mm5 lea r0, [r2+r3*2] movq [r1+r3*1], mm6 movq [r1+r3*2], mm7 movq [r2+r3*1], mm0 movq [r2+r3*2], mm1 movq [r0+r3*1], mm2 movq [r0+r3*2], mm3 RET %endmacro INIT_MMX mmxext PRED8x8L_HORIZONTAL_UP INIT_MMX ssse3 PRED8x8L_HORIZONTAL_UP ;----------------------------------------------------------------------------- ; void ff_pred8x8l_horizontal_down_8(uint8_t *src, int has_topleft, ; int has_topright, ptrdiff_t stride) ;----------------------------------------------------------------------------- %macro PRED8x8L_HORIZONTAL_DOWN 0 cglobal pred8x8l_horizontal_down_8, 4,5 sub r0, r3 lea r4, [r0+r3*2] movq mm0, [r0+r3*1-8] punpckhbw mm0, [r0+r3*0-8] movq mm1, [r4+r3*1-8] punpckhbw mm1, [r0+r3*2-8] mov r4, r0 punpckhwd mm1, mm0 lea r0, [r0+r3*4] movq mm2, [r0+r3*1-8] punpckhbw mm2, [r0+r3*0-8] lea r0, [r0+r3*2] movq mm3, [r0+r3*1-8] punpckhbw mm3, [r0+r3*0-8] punpckhwd mm3, mm2 punpckhdq mm3, mm1 lea r0, [r0+r3*2] movq mm0, [r0+r3*0-8] movq mm1, [r4] mov r0, r4 movq mm4, mm3 movq mm2, mm3 PALIGNR mm4, mm0, 7, mm0 PALIGNR mm1, mm2, 1, mm2 test r1d, r1d jnz .do_left .fix_lt_1: movq mm5, mm3 pxor mm5, mm4 psrlq mm5, 56 psllq mm5, 48 pxor mm1, mm5 jmp .do_left .fix_lt_2: movq mm5, mm3 pxor mm5, mm2 psllq mm5, 56 psrlq mm5, 56 pxor mm2, mm5 test r2d, r2d jnz .do_top .fix_tr_1: movq mm5, mm3 pxor mm5, mm1 psrlq mm5, 56 psllq mm5, 56 pxor mm1, mm5 jmp .do_top .fix_tr_2: punpckhbw mm3, mm3 pshufw mm1, mm3, 0xFF jmp .do_topright .do_left: movq mm0, mm4 PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 movq2dq xmm0, mm2 pslldq xmm0, 8 movq mm4, mm0 PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 movq2dq xmm2, mm1 pslldq xmm2, 15 psrldq xmm2, 8 por xmm0, xmm2 movq mm0, [r0-8] movq mm3, [r0] movq mm1, [r0+8] movq mm2, mm3 movq mm4, mm3 PALIGNR mm2, mm0, 7, mm0 PALIGNR mm1, mm4, 1, mm4 test r1d, r1d jz .fix_lt_2 test r2d, r2d jz .fix_tr_1 .do_top: PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5 movq2dq xmm1, mm4 test r2d, r2d jz .fix_tr_2 movq mm0, [r0+8] movq mm5, mm0 movq mm2, mm0 movq mm4, mm0 psrlq mm5, 56 PALIGNR mm2, mm3, 7, mm3 PALIGNR mm5, mm4, 1, mm4 PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4 .do_topright: movq2dq xmm5, mm1 pslldq xmm5, 8 por xmm1, xmm5 INIT_XMM cpuname lea r2, [r4+r3*2] movdqa xmm2, xmm1 movdqa xmm3, xmm1 PALIGNR xmm1, xmm0, 7, xmm4 PALIGNR xmm2, xmm0, 9, xmm5 lea r1, [r2+r3*2] PALIGNR xmm3, xmm0, 8, xmm0 movdqa xmm4, xmm1 pavgb xmm4, xmm3 lea r0, [r1+r3*2] PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm3, xmm5 punpcklbw xmm4, xmm0 movhlps xmm0, xmm4 movq [r0+r3*2], xmm4 movq [r2+r3*2], xmm0 psrldq xmm4, 2 psrldq xmm0, 2 movq [r0+r3*1], xmm4 movq [r2+r3*1], xmm0 psrldq xmm4, 2 psrldq xmm0, 2 movq [r1+r3*2], xmm4 movq [r4+r3*2], xmm0 psrldq xmm4, 2 psrldq xmm0, 2 movq [r1+r3*1], xmm4 movq [r4+r3*1], xmm0 RET %endmacro INIT_MMX sse2 PRED8x8L_HORIZONTAL_DOWN INIT_MMX ssse3 PRED8x8L_HORIZONTAL_DOWN ;------------------------------------------------------------------------------- ; void ff_pred4x4_dc_8_mmxext(uint8_t *src, const uint8_t *topright, ; ptrdiff_t stride) ;------------------------------------------------------------------------------- INIT_MMX mmxext cglobal pred4x4_dc_8, 3,5 pxor mm7, mm7 mov r4, r0 sub r0, r2 movd mm0, [r0] psadbw mm0, mm7 movzx r1d, byte [r0+r2*1-1] movd r3d, mm0 add r3d, r1d movzx r1d, byte [r0+r2*2-1] lea r0, [r0+r2*2] add r3d, r1d movzx r1d, byte [r0+r2*1-1] add r3d, r1d movzx r1d, byte [r0+r2*2-1] add r3d, r1d add r3d, 4 shr r3d, 3 imul r3d, 0x01010101 mov [r4+r2*0], r3d mov [r0+r2*0], r3d mov [r0+r2*1], r3d mov [r0+r2*2], r3d RET ;----------------------------------------------------------------------------- ; void ff_pred4x4_tm_vp8_8_mmxext(uint8_t *src, const uint8_t *topright, ; ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_MMX mmxext cglobal pred4x4_tm_vp8_8, 3,6 sub r0, r2 pxor mm7, mm7 movd mm0, [r0] punpcklbw mm0, mm7 movzx r4d, byte [r0-1] mov r5d, 2 .loop: movzx r1d, byte [r0+r2*1-1] movzx r3d, byte [r0+r2*2-1] sub r1d, r4d sub r3d, r4d movd mm2, r1d movd mm4, r3d pshufw mm2, mm2, 0 pshufw mm4, mm4, 0 paddw mm2, mm0 paddw mm4, mm0 packuswb mm2, mm2 packuswb mm4, mm4 movd [r0+r2*1], mm2 movd [r0+r2*2], mm4 lea r0, [r0+r2*2] dec r5d jg .loop RET INIT_XMM ssse3 cglobal pred4x4_tm_vp8_8, 3,3 sub r0, r2 movq mm6, [tm_shuf] pxor mm1, mm1 movd mm0, [r0] punpcklbw mm0, mm1 movd mm7, [r0-4] pshufb mm7, mm6 lea r1, [r0+r2*2] movd mm2, [r0+r2*1-4] movd mm3, [r0+r2*2-4] movd mm4, [r1+r2*1-4] movd mm5, [r1+r2*2-4] pshufb mm2, mm6 pshufb mm3, mm6 pshufb mm4, mm6 pshufb mm5, mm6 psubw mm0, mm7 paddw mm2, mm0 paddw mm3, mm0 paddw mm4, mm0 paddw mm5, mm0 packuswb mm2, mm2 packuswb mm3, mm3 packuswb mm4, mm4 packuswb mm5, mm5 movd [r0+r2*1], mm2 movd [r0+r2*2], mm3 movd [r1+r2*1], mm4 movd [r1+r2*2], mm5 RET ;----------------------------------------------------------------------------- ; void ff_pred4x4_vertical_vp8_8_mmxext(uint8_t *src, const uint8_t *topright, ; ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_MMX mmxext cglobal pred4x4_vertical_vp8_8, 3,3 sub r0, r2 movd m1, [r0-1] movd m0, [r0] mova m2, m0 ;t0 t1 t2 t3 punpckldq m0, [r1] ;t0 t1 t2 t3 t4 t5 t6 t7 lea r1, [r0+r2*2] psrlq m0, 8 ;t1 t2 t3 t4 PRED4x4_LOWPASS m3, m1, m0, m2, m4 movd [r0+r2*1], m3 movd [r0+r2*2], m3 movd [r1+r2*1], m3 movd [r1+r2*2], m3 RET ;----------------------------------------------------------------------------- ; void ff_pred4x4_down_left_8_mmxext(uint8_t *src, const uint8_t *topright, ; ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_MMX mmxext cglobal pred4x4_down_left_8, 3,3 sub r0, r2 movq m1, [r0] punpckldq m1, [r1] movq m2, m1 movq m3, m1 psllq m1, 8 pxor m2, m1 psrlq m2, 8 pxor m2, m3 PRED4x4_LOWPASS m0, m1, m2, m3, m4 lea r1, [r0+r2*2] psrlq m0, 8 movd [r0+r2*1], m0 psrlq m0, 8 movd [r0+r2*2], m0 psrlq m0, 8 movd [r1+r2*1], m0 psrlq m0, 8 movd [r1+r2*2], m0 RET ;------------------------------------------------------------------------------ ; void ff_pred4x4_vertical_left_8_mmxext(uint8_t *src, const uint8_t *topright, ; ptrdiff_t stride) ;------------------------------------------------------------------------------ INIT_MMX mmxext cglobal pred4x4_vertical_left_8, 3,3 sub r0, r2 movq m1, [r0] punpckldq m1, [r1] movq m3, m1 movq m2, m1 psrlq m3, 8 psrlq m2, 16 movq m4, m3 pavgb m4, m1 PRED4x4_LOWPASS m0, m1, m2, m3, m5 lea r1, [r0+r2*2] movh [r0+r2*1], m4 movh [r0+r2*2], m0 psrlq m4, 8 psrlq m0, 8 movh [r1+r2*1], m4 movh [r1+r2*2], m0 RET ;------------------------------------------------------------------------------ ; void ff_pred4x4_horizontal_up_8_mmxext(uint8_t *src, const uint8_t *topright, ; ptrdiff_t stride) ;------------------------------------------------------------------------------ INIT_MMX mmxext cglobal pred4x4_horizontal_up_8, 3,3 sub r0, r2 lea r1, [r0+r2*2] movd m0, [r0+r2*1-4] punpcklbw m0, [r0+r2*2-4] movd m1, [r1+r2*1-4] punpcklbw m1, [r1+r2*2-4] punpckhwd m0, m1 movq m1, m0 punpckhbw m1, m1 pshufw m1, m1, 0xFF punpckhdq m0, m1 movq m2, m0 movq m3, m0 movq m7, m0 psrlq m2, 16 psrlq m3, 8 pavgb m7, m3 PRED4x4_LOWPASS m4, m0, m2, m3, m5 punpcklbw m7, m4 movd [r0+r2*1], m7 psrlq m7, 16 movd [r0+r2*2], m7 psrlq m7, 16 movd [r1+r2*1], m7 movd [r1+r2*2], m1 RET ;------------------------------------------------------------------------------ ; void ff_pred4x4_horizontal_down_8_mmxext(uint8_t *src, ; const uint8_t *topright, ; ptrdiff_t stride) ;------------------------------------------------------------------------------ INIT_MMX mmxext cglobal pred4x4_horizontal_down_8, 3,3 sub r0, r2 lea r1, [r0+r2*2] movh m0, [r0-4] ; lt .. punpckldq m0, [r0] ; t3 t2 t1 t0 lt .. .. .. psllq m0, 8 ; t2 t1 t0 lt .. .. .. .. movd m1, [r1+r2*2-4] ; l3 punpcklbw m1, [r1+r2*1-4] ; l2 l3 movd m2, [r0+r2*2-4] ; l1 punpcklbw m2, [r0+r2*1-4] ; l0 l1 punpckhwd m1, m2 ; l0 l1 l2 l3 punpckhdq m1, m0 ; t2 t1 t0 lt l0 l1 l2 l3 movq m0, m1 movq m2, m1 movq m5, m1 psrlq m0, 16 ; .. .. t2 t1 t0 lt l0 l1 psrlq m2, 8 ; .. t2 t1 t0 lt l0 l1 l2 pavgb m5, m2 PRED4x4_LOWPASS m3, m1, m0, m2, m4 punpcklbw m5, m3 psrlq m3, 32 PALIGNR m3, m5, 6, m4 movh [r1+r2*2], m5 psrlq m5, 16 movh [r1+r2*1], m5 psrlq m5, 16 movh [r0+r2*2], m5 movh [r0+r2*1], m3 RET ;----------------------------------------------------------------------------- ; void ff_pred4x4_vertical_right_8_mmxext(uint8_t *src, ; const uint8_t *topright, ; ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_MMX mmxext cglobal pred4x4_vertical_right_8, 3,3 sub r0, r2 lea r1, [r0+r2*2] movh m0, [r0] ; ........t3t2t1t0 movq m5, m0 PALIGNR m0, [r0-8], 7, m1 ; ......t3t2t1t0lt pavgb m5, m0 PALIGNR m0, [r0+r2*1-8], 7, m1 ; ....t3t2t1t0ltl0 movq m1, m0 PALIGNR m0, [r0+r2*2-8], 7, m2 ; ..t3t2t1t0ltl0l1 movq m2, m0 PALIGNR m0, [r1+r2*1-8], 7, m3 ; t3t2t1t0ltl0l1l2 PRED4x4_LOWPASS m3, m1, m0, m2, m4 movq m1, m3 psrlq m3, 16 psllq m1, 48 movh [r0+r2*1], m5 movh [r0+r2*2], m3 PALIGNR m5, m1, 7, m2 psllq m1, 8 movh [r1+r2*1], m5 PALIGNR m3, m1, 7, m1 movh [r1+r2*2], m3 RET ;----------------------------------------------------------------------------- ; void ff_pred4x4_down_right_8_mmxext(uint8_t *src, const uint8_t *topright, ; ptrdiff_t stride) ;----------------------------------------------------------------------------- INIT_MMX mmxext cglobal pred4x4_down_right_8, 3,3 sub r0, r2 lea r1, [r0+r2*2] movq m1, [r1-8] movq m2, [r0+r2*1-8] punpckhbw m2, [r0-8] movh m3, [r0] punpckhwd m1, m2 PALIGNR m3, m1, 5, m1 movq m1, m3 PALIGNR m3, [r1+r2*1-8], 7, m4 movq m2, m3 PALIGNR m3, [r1+r2*2-8], 7, m4 PRED4x4_LOWPASS m0, m3, m1, m2, m4 movh [r1+r2*2], m0 psrlq m0, 8 movh [r1+r2*1], m0 psrlq m0, 8 movh [r0+r2*2], m0 psrlq m0, 8 movh [r0+r2*1], m0 RET