;***************************************************************************** ;* SSE2-optimized weighted prediction code ;***************************************************************************** ;* Copyright (c) 2004-2005 Michael Niedermayer, Loren Merritt ;* Copyright (C) 2010 Eli Friedman ;* ;* This file is part of FFmpeg. ;* ;* FFmpeg is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* FFmpeg is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with FFmpeg; if not, write to the Free Software ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "libavutil/x86/x86util.asm" SECTION .text ;----------------------------------------------------------------------------- ; biweight pred: ; ; void ff_h264_biweight_16_sse2(uint8_t *dst, uint8_t *src, int stride, ; int height, int log2_denom, int weightd, ; int weights, int offset); ; and ; void ff_h264_weight_16_sse2(uint8_t *dst, int stride, int height, ; int log2_denom, int weight, int offset); ;----------------------------------------------------------------------------- %macro WEIGHT_SETUP 0 add r5, r5 inc r5 movd m3, r4d movd m5, r5d movd m6, r3d pslld m5, m6 psrld m5, 1 %if mmsize == 16 pshuflw m3, m3, 0 pshuflw m5, m5, 0 punpcklqdq m3, m3 punpcklqdq m5, m5 %else pshufw m3, m3, 0 pshufw m5, m5, 0 %endif pxor m7, m7 %endmacro %macro WEIGHT_OP 2 movh m0, [r0+%1] movh m1, [r0+%2] punpcklbw m0, m7 punpcklbw m1, m7 pmullw m0, m3 pmullw m1, m3 paddsw m0, m5 paddsw m1, m5 psraw m0, m6 psraw m1, m6 packuswb m0, m1 %endmacro %macro WEIGHT_FUNC_MM 2 cglobal h264_weight_%1, 6, 6, %2 WEIGHT_SETUP .nextrow: WEIGHT_OP 0, mmsize/2 mova [r0], m0 add r0, r1 dec r2d jnz .nextrow RET %endmacro INIT_XMM sse2 WEIGHT_FUNC_MM 16, 8 %macro WEIGHT_FUNC_HALF_MM 2 cglobal h264_weight_%1, 6, 6, %2 WEIGHT_SETUP sar r2d, 1 lea r3, [r1*2] .nextrow: WEIGHT_OP 0, r1 movh [r0], m0 %if mmsize == 16 movhps [r0+r1], m0 %else psrlq m0, 32 movh [r0+r1], m0 %endif add r0, r3 dec r2d jnz .nextrow RET %endmacro INIT_MMX mmxext WEIGHT_FUNC_HALF_MM 4, 0 INIT_XMM sse2 WEIGHT_FUNC_HALF_MM 8, 8 %macro BIWEIGHT_SETUP 0 %if ARCH_X86_64 %define off_regd r7d %else %define off_regd r3d %endif mov off_regd, r7m add off_regd, 1 or off_regd, 1 add r4d, 1 cmp r6d, 128 je .nonnormal cmp r5d, 128 jne .normal .nonnormal: sar r5d, 1 sar r6d, 1 sar off_regd, 1 sub r4d, 1 .normal: %if cpuflag(ssse3) movd m4, r5d movd m0, r6d %else movd m3, r5d movd m4, r6d %endif movd m5, off_regd movd m6, r4d pslld m5, m6 psrld m5, 1 %if cpuflag(ssse3) punpcklbw m4, m0 pshuflw m4, m4, 0 pshuflw m5, m5, 0 punpcklqdq m4, m4 punpcklqdq m5, m5 %else %if mmsize == 16 pshuflw m3, m3, 0 pshuflw m4, m4, 0 pshuflw m5, m5, 0 punpcklqdq m3, m3 punpcklqdq m4, m4 punpcklqdq m5, m5 %else pshufw m3, m3, 0 pshufw m4, m4, 0 pshufw m5, m5, 0 %endif pxor m7, m7 %endif %endmacro %macro BIWEIGHT_STEPA 3 movh m%1, [r0+%3] movh m%2, [r1+%3] punpcklbw m%1, m7 punpcklbw m%2, m7 pmullw m%1, m3 pmullw m%2, m4 paddsw m%1, m%2 %endmacro %macro BIWEIGHT_STEPB 0 paddsw m0, m5 paddsw m1, m5 psraw m0, m6 psraw m1, m6 packuswb m0, m1 %endmacro %macro BIWEIGHT_FUNC_MM 2 cglobal h264_biweight_%1, 7, 8, %2 BIWEIGHT_SETUP movifnidn r3d, r3m .nextrow: BIWEIGHT_STEPA 0, 1, 0 BIWEIGHT_STEPA 1, 2, mmsize/2 BIWEIGHT_STEPB mova [r0], m0 add r0, r2 add r1, r2 dec r3d jnz .nextrow RET %endmacro INIT_XMM sse2 BIWEIGHT_FUNC_MM 16, 8 %macro BIWEIGHT_FUNC_HALF_MM 2 cglobal h264_biweight_%1, 7, 8, %2 BIWEIGHT_SETUP movifnidn r3d, r3m sar r3d, 1 lea r4, [r2*2] .nextrow: BIWEIGHT_STEPA 0, 1, 0 BIWEIGHT_STEPA 1, 2, r2 BIWEIGHT_STEPB movh [r0], m0 %if mmsize == 16 movhps [r0+r2], m0 %else psrlq m0, 32 movh [r0+r2], m0 %endif add r0, r4 add r1, r4 dec r3d jnz .nextrow RET %endmacro INIT_MMX mmxext BIWEIGHT_FUNC_HALF_MM 4, 0 INIT_XMM sse2 BIWEIGHT_FUNC_HALF_MM 8, 8 %macro BIWEIGHT_SSSE3_OP 0 pmaddubsw m0, m4 pmaddubsw m2, m4 paddsw m0, m5 paddsw m2, m5 psraw m0, m6 psraw m2, m6 packuswb m0, m2 %endmacro INIT_XMM ssse3 cglobal h264_biweight_16, 7, 8, 8 BIWEIGHT_SETUP movifnidn r3d, r3m .nextrow: movh m0, [r0] movh m2, [r0+8] movh m3, [r1+8] punpcklbw m0, [r1] punpcklbw m2, m3 BIWEIGHT_SSSE3_OP mova [r0], m0 add r0, r2 add r1, r2 dec r3d jnz .nextrow RET INIT_XMM ssse3 cglobal h264_biweight_8, 7, 8, 8 BIWEIGHT_SETUP movifnidn r3d, r3m sar r3d, 1 lea r4, [r2*2] .nextrow: movh m0, [r0] movh m1, [r1] movh m2, [r0+r2] movh m3, [r1+r2] punpcklbw m0, m1 punpcklbw m2, m3 BIWEIGHT_SSSE3_OP movh [r0], m0 movhps [r0+r2], m0 add r0, r4 add r1, r4 dec r3d jnz .nextrow RET