;****************************************************************************** ;* Copyright (c) 2010 David Conrad ;* ;* This file is part of FFmpeg. ;* ;* FFmpeg is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* FFmpeg is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with FFmpeg; if not, write to the Free Software ;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "libavutil/x86/x86util.asm" SECTION_RODATA pw_7: times 8 dw 7 convert_to_unsigned_10bit: times 4 dd 0x200 clip_10bit: times 8 dw 0x3ff cextern pw_3 cextern pw_16 cextern pw_32 cextern pb_80 SECTION .text %macro UNPACK_ADD 6 mov%5 %1, %3 mov%6 m5, %4 mova m4, %1 mova %2, m5 punpcklbw %1, m7 punpcklbw m5, m7 punpckhbw m4, m7 punpckhbw %2, m7 paddw %1, m5 paddw %2, m4 %endmacro %macro HPEL_FILTER 1 ; dirac_hpel_filter_v_sse2(uint8_t *dst, uint8_t *src, int stride, int width); cglobal dirac_hpel_filter_v_%1, 4,6,8, dst, src, stride, width, src0, stridex3 mov src0q, srcq lea stridex3q, [3*strideq] sub src0q, stridex3q pxor m7, m7 .loop: ; 7*(src[0] + src[1]) UNPACK_ADD m0, m1, [srcq], [srcq + strideq], a,a pmullw m0, [pw_7] pmullw m1, [pw_7] ; 3*( ... + src[-2] + src[3]) UNPACK_ADD m2, m3, [src0q + strideq], [srcq + stridex3q], a,a paddw m0, m2 paddw m1, m3 pmullw m0, [pw_3] pmullw m1, [pw_3] ; ... - 7*(src[-1] + src[2]) UNPACK_ADD m2, m3, [src0q + strideq*2], [srcq + strideq*2], a,a pmullw m2, [pw_7] pmullw m3, [pw_7] psubw m0, m2 psubw m1, m3 ; ... - (src[-3] + src[4]) UNPACK_ADD m2, m3, [src0q], [srcq + strideq*4], a,a psubw m0, m2 psubw m1, m3 paddw m0, [pw_16] paddw m1, [pw_16] psraw m0, 5 psraw m1, 5 packuswb m0, m1 mova [dstq], m0 add dstq, mmsize add srcq, mmsize add src0q, mmsize sub widthd, mmsize jg .loop RET ; dirac_hpel_filter_h_sse2(uint8_t *dst, uint8_t *src, int width); cglobal dirac_hpel_filter_h_%1, 3,3,8, dst, src, width dec widthd pxor m7, m7 and widthd, ~(mmsize-1) .loop: ; 7*(src[0] + src[1]) UNPACK_ADD m0, m1, [srcq + widthq], [srcq + widthq + 1], u,u pmullw m0, [pw_7] pmullw m1, [pw_7] ; 3*( ... + src[-2] + src[3]) UNPACK_ADD m2, m3, [srcq + widthq - 2], [srcq + widthq + 3], u,u paddw m0, m2 paddw m1, m3 pmullw m0, [pw_3] pmullw m1, [pw_3] ; ... - 7*(src[-1] + src[2]) UNPACK_ADD m2, m3, [srcq + widthq - 1], [srcq + widthq + 2], u,u pmullw m2, [pw_7] pmullw m3, [pw_7] psubw m0, m2 psubw m1, m3 ; ... - (src[-3] + src[4]) UNPACK_ADD m2, m3, [srcq + widthq - 3], [srcq + widthq + 4], u,u psubw m0, m2 psubw m1, m3 paddw m0, [pw_16] paddw m1, [pw_16] psraw m0, 5 psraw m1, 5 packuswb m0, m1 mova [dstq + widthq], m0 sub widthd, mmsize jge .loop RET %endmacro %macro PUT_RECT 1 ; void put_rect_clamped(uint8_t *dst, int dst_stride, int16_t *src, int src_stride, int width, int height) cglobal put_signed_rect_clamped_%1, 5,9,3, dst, dst_stride, src, src_stride, w, dst2, src2 mova m0, [pb_80] add wd, (mmsize-1) and wd, ~(mmsize-1) %if ARCH_X86_64 movsxd dst_strideq, dst_strided movsxd src_strideq, src_strided mov r7d, r5m mov r8d, wd %define wspill r8d %define hd r7d %else mov r4m, wd %define wspill r4m %define hd r5mp %endif .loopy: lea src2q, [srcq+src_strideq] lea dst2q, [dstq+dst_strideq] .loopx: sub wd, mmsize mova m1, [srcq +2*wq] mova m2, [src2q+2*wq] packsswb m1, [srcq +2*wq+mmsize] packsswb m2, [src2q+2*wq+mmsize] paddb m1, m0 paddb m2, m0 mova [dstq +wq], m1 mova [dst2q+wq], m2 jg .loopx lea srcq, [srcq+src_strideq*2] lea dstq, [dstq+dst_strideq*2] sub hd, 2 mov wd, wspill jg .loopy RET %endm %macro ADD_RECT 1 ; void add_rect_clamped(uint8_t *dst, uint16_t *src, int stride, int16_t *idwt, int idwt_stride, int width, int height) cglobal add_rect_clamped_%1, 7,9,3, dst, src, stride, idwt, idwt_stride, w, h mova m0, [pw_32] add wd, (mmsize-1) and wd, ~(mmsize-1) %if ARCH_X86_64 movsxd strideq, strided movsxd idwt_strideq, idwt_strided mov r8d, wd %define wspill r8d %else mov r5m, wd %define wspill r5m %endif .loop: sub wd, mmsize movu m1, [srcq +2*wq] ; FIXME: ensure alignment paddw m1, m0 psraw m1, 6 movu m2, [srcq +2*wq+mmsize] ; FIXME: ensure alignment paddw m2, m0 psraw m2, 6 paddw m1, [idwtq+2*wq] paddw m2, [idwtq+2*wq+mmsize] packuswb m1, m2 mova [dstq +wq], m1 jg .loop lea srcq, [srcq + 2*strideq] add dstq, strideq lea idwtq, [idwtq+ 2*idwt_strideq] sub hd, 1 mov wd, wspill jg .loop RET %endm %macro ADD_OBMC 2 ; void add_obmc(uint16_t *dst, uint8_t *src, int stride, uint8_t *obmc_weight, int yblen) cglobal add_dirac_obmc%1_%2, 6,6,5, dst, src, stride, obmc, yblen pxor m4, m4 .loop: %assign i 0 %rep %1 / mmsize mova m0, [srcq+i] mova m1, m0 punpcklbw m0, m4 punpckhbw m1, m4 mova m2, [obmcq+i] mova m3, m2 punpcklbw m2, m4 punpckhbw m3, m4 pmullw m0, m2 pmullw m1, m3 movu m2, [dstq+2*i] movu m3, [dstq+2*i+mmsize] paddw m0, m2 paddw m1, m3 movu [dstq+2*i], m0 movu [dstq+2*i+mmsize], m1 %assign i i+mmsize %endrep lea srcq, [srcq+strideq] lea dstq, [dstq+2*strideq] add obmcq, 32 sub yblend, 1 jg .loop RET %endm INIT_MMX %if ARCH_X86_64 == 0 PUT_RECT mmx ADD_RECT mmx HPEL_FILTER mmx ADD_OBMC 32, mmx ADD_OBMC 16, mmx %endif ADD_OBMC 8, mmx INIT_XMM PUT_RECT sse2 ADD_RECT sse2 HPEL_FILTER sse2 ADD_OBMC 32, sse2 ADD_OBMC 16, sse2 INIT_XMM sse4 ; void dequant_subband_32(uint8_t *src, uint8_t *dst, ptrdiff_t stride, const int qf, const int qs, int tot_v, int tot_h) cglobal dequant_subband_32, 7, 7, 4, src, dst, stride, qf, qs, tot_v, tot_h movd m2, qfd movd m3, qsd SPLATD m2 SPLATD m3 mov r4, tot_hq mov r3, dstq .loop_v: mov tot_hq, r4 mov dstq, r3 .loop_h: movu m0, [srcq] pabsd m1, m0 pmulld m1, m2 paddd m1, m3 psrld m1, 2 psignd m1, m0 movu [dstq], m1 add srcq, mmsize add dstq, mmsize sub tot_hq, 4 jg .loop_h lea srcq, [srcq + 4*tot_hq] add r3, strideq dec tot_vd jg .loop_v RET INIT_XMM sse4 ; void put_signed_rect_clamped_10(uint8_t *dst, int dst_stride, const uint8_t *src, int src_stride, int width, int height) %if ARCH_X86_64 cglobal put_signed_rect_clamped_10, 6, 8, 5, dst, dst_stride, src, src_stride, w, h, t1, t2 %else cglobal put_signed_rect_clamped_10, 5, 7, 5, dst, dst_stride, src, src_stride, w, t1, t2 %define hd r5mp %endif shl wd, 2 add srcq, wq neg wq mov t2q, dstq mov t1q, wq pxor m2, m2 mova m3, [clip_10bit] mova m4, [convert_to_unsigned_10bit] .loop_h: mov dstq, t2q mov wq, t1q .loop_w: movu m0, [srcq+wq+0*mmsize] movu m1, [srcq+wq+1*mmsize] paddd m0, m4 paddd m1, m4 packusdw m0, m0, m1 CLIPW m0, m2, m3 ; packusdw saturates so it's fine movu [dstq], m0 add dstq, 1*mmsize add wq, 2*mmsize jl .loop_w add srcq, src_strideq add t2q, dst_strideq sub hd, 1 jg .loop_h RET