mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-28 18:32:22 +00:00
264441715b
* commit 'f5fa03660db16f9d78abc5a626438b4d0b54f563': vble: Do not abort decoding when version is not 1 lavr: do not pass consumed samples as a parameter to ff_audio_resample() lavr: correct the documentation for the ff_audio_resample() return value lavr: do not pass sample count as a parameter to ff_audio_convert() x86: h264_weight: port to cpuflags configure: Enable avconv filter dependencies automatically Conflicts: configure libavcodec/x86/h264_weight.asm Merged-by: Michael Niedermayer <michaelni@gmx.at>
318 lines
6.9 KiB
NASM
318 lines
6.9 KiB
NASM
;*****************************************************************************
|
|
;* SSE2-optimized weighted prediction code
|
|
;*****************************************************************************
|
|
;* Copyright (c) 2004-2005 Michael Niedermayer, Loren Merritt
|
|
;* Copyright (C) 2010 Eli Friedman <eli.friedman@gmail.com>
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION .text
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; biweight pred:
|
|
;
|
|
; void h264_biweight_16_sse2(uint8_t *dst, uint8_t *src, int stride,
|
|
; int height, int log2_denom, int weightd,
|
|
; int weights, int offset);
|
|
; and
|
|
; void h264_weight_16_sse2(uint8_t *dst, int stride, int height,
|
|
; int log2_denom, int weight, int offset);
|
|
;-----------------------------------------------------------------------------
|
|
|
|
%macro WEIGHT_SETUP 0
|
|
add r5, r5
|
|
inc r5
|
|
movd m3, r4d
|
|
movd m5, r5d
|
|
movd m6, r3d
|
|
pslld m5, m6
|
|
psrld m5, 1
|
|
%if mmsize == 16
|
|
pshuflw m3, m3, 0
|
|
pshuflw m5, m5, 0
|
|
punpcklqdq m3, m3
|
|
punpcklqdq m5, m5
|
|
%else
|
|
pshufw m3, m3, 0
|
|
pshufw m5, m5, 0
|
|
%endif
|
|
pxor m7, m7
|
|
%endmacro
|
|
|
|
%macro WEIGHT_OP 2
|
|
movh m0, [r0+%1]
|
|
movh m1, [r0+%2]
|
|
punpcklbw m0, m7
|
|
punpcklbw m1, m7
|
|
pmullw m0, m3
|
|
pmullw m1, m3
|
|
paddsw m0, m5
|
|
paddsw m1, m5
|
|
psraw m0, m6
|
|
psraw m1, m6
|
|
packuswb m0, m1
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
cglobal h264_weight_16, 6, 6, 0
|
|
WEIGHT_SETUP
|
|
.nextrow:
|
|
WEIGHT_OP 0, 4
|
|
mova [r0 ], m0
|
|
WEIGHT_OP 8, 12
|
|
mova [r0+8], m0
|
|
add r0, r1
|
|
dec r2d
|
|
jnz .nextrow
|
|
REP_RET
|
|
|
|
%macro WEIGHT_FUNC_MM 2
|
|
cglobal h264_weight_%1, 6, 6, %2
|
|
WEIGHT_SETUP
|
|
.nextrow:
|
|
WEIGHT_OP 0, mmsize/2
|
|
mova [r0], m0
|
|
add r0, r1
|
|
dec r2d
|
|
jnz .nextrow
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
WEIGHT_FUNC_MM 8, 0
|
|
INIT_XMM sse2
|
|
WEIGHT_FUNC_MM 16, 8
|
|
|
|
%macro WEIGHT_FUNC_HALF_MM 2
|
|
cglobal h264_weight_%1, 6, 6, %2
|
|
WEIGHT_SETUP
|
|
sar r2d, 1
|
|
lea r3, [r1*2]
|
|
.nextrow:
|
|
WEIGHT_OP 0, r1
|
|
movh [r0], m0
|
|
%if mmsize == 16
|
|
movhps [r0+r1], m0
|
|
%else
|
|
psrlq m0, 32
|
|
movh [r0+r1], m0
|
|
%endif
|
|
add r0, r3
|
|
dec r2d
|
|
jnz .nextrow
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
WEIGHT_FUNC_HALF_MM 4, 0
|
|
INIT_XMM sse2
|
|
WEIGHT_FUNC_HALF_MM 8, 8
|
|
|
|
%macro BIWEIGHT_SETUP 0
|
|
%if ARCH_X86_64
|
|
%define off_regd r7d
|
|
%else
|
|
%define off_regd r3d
|
|
%endif
|
|
mov off_regd, r7m
|
|
add off_regd, 1
|
|
or off_regd, 1
|
|
add r4, 1
|
|
cmp r5, 128
|
|
jne .normal
|
|
sar r5, 1
|
|
sar r6, 1
|
|
sar off_regd, 1
|
|
sub r4, 1
|
|
.normal
|
|
%if cpuflag(ssse3)
|
|
movd m4, r5d
|
|
movd m0, r6d
|
|
%else
|
|
movd m3, r5d
|
|
movd m4, r6d
|
|
%endif
|
|
movd m5, off_regd
|
|
movd m6, r4d
|
|
pslld m5, m6
|
|
psrld m5, 1
|
|
%if cpuflag(ssse3)
|
|
punpcklbw m4, m0
|
|
pshuflw m4, m4, 0
|
|
pshuflw m5, m5, 0
|
|
punpcklqdq m4, m4
|
|
punpcklqdq m5, m5
|
|
|
|
%else
|
|
%if mmsize == 16
|
|
pshuflw m3, m3, 0
|
|
pshuflw m4, m4, 0
|
|
pshuflw m5, m5, 0
|
|
punpcklqdq m3, m3
|
|
punpcklqdq m4, m4
|
|
punpcklqdq m5, m5
|
|
%else
|
|
pshufw m3, m3, 0
|
|
pshufw m4, m4, 0
|
|
pshufw m5, m5, 0
|
|
%endif
|
|
pxor m7, m7
|
|
%endif
|
|
%endmacro
|
|
|
|
%macro BIWEIGHT_STEPA 3
|
|
movh m%1, [r0+%3]
|
|
movh m%2, [r1+%3]
|
|
punpcklbw m%1, m7
|
|
punpcklbw m%2, m7
|
|
pmullw m%1, m3
|
|
pmullw m%2, m4
|
|
paddsw m%1, m%2
|
|
%endmacro
|
|
|
|
%macro BIWEIGHT_STEPB 0
|
|
paddsw m0, m5
|
|
paddsw m1, m5
|
|
psraw m0, m6
|
|
psraw m1, m6
|
|
packuswb m0, m1
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
cglobal h264_biweight_16, 7, 8, 0
|
|
BIWEIGHT_SETUP
|
|
movifnidn r3d, r3m
|
|
.nextrow:
|
|
BIWEIGHT_STEPA 0, 1, 0
|
|
BIWEIGHT_STEPA 1, 2, 4
|
|
BIWEIGHT_STEPB
|
|
mova [r0], m0
|
|
BIWEIGHT_STEPA 0, 1, 8
|
|
BIWEIGHT_STEPA 1, 2, 12
|
|
BIWEIGHT_STEPB
|
|
mova [r0+8], m0
|
|
add r0, r2
|
|
add r1, r2
|
|
dec r3d
|
|
jnz .nextrow
|
|
REP_RET
|
|
|
|
%macro BIWEIGHT_FUNC_MM 2
|
|
cglobal h264_biweight_%1, 7, 8, %2
|
|
BIWEIGHT_SETUP
|
|
movifnidn r3d, r3m
|
|
.nextrow:
|
|
BIWEIGHT_STEPA 0, 1, 0
|
|
BIWEIGHT_STEPA 1, 2, mmsize/2
|
|
BIWEIGHT_STEPB
|
|
mova [r0], m0
|
|
add r0, r2
|
|
add r1, r2
|
|
dec r3d
|
|
jnz .nextrow
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
BIWEIGHT_FUNC_MM 8, 0
|
|
INIT_XMM sse2
|
|
BIWEIGHT_FUNC_MM 16, 8
|
|
|
|
%macro BIWEIGHT_FUNC_HALF_MM 2
|
|
cglobal h264_biweight_%1, 7, 8, %2
|
|
BIWEIGHT_SETUP
|
|
movifnidn r3d, r3m
|
|
sar r3, 1
|
|
lea r4, [r2*2]
|
|
.nextrow:
|
|
BIWEIGHT_STEPA 0, 1, 0
|
|
BIWEIGHT_STEPA 1, 2, r2
|
|
BIWEIGHT_STEPB
|
|
movh [r0], m0
|
|
%if mmsize == 16
|
|
movhps [r0+r2], m0
|
|
%else
|
|
psrlq m0, 32
|
|
movh [r0+r2], m0
|
|
%endif
|
|
add r0, r4
|
|
add r1, r4
|
|
dec r3d
|
|
jnz .nextrow
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
BIWEIGHT_FUNC_HALF_MM 4, 0
|
|
INIT_XMM sse2
|
|
BIWEIGHT_FUNC_HALF_MM 8, 8
|
|
|
|
%macro BIWEIGHT_SSSE3_OP 0
|
|
pmaddubsw m0, m4
|
|
pmaddubsw m2, m4
|
|
paddsw m0, m5
|
|
paddsw m2, m5
|
|
psraw m0, m6
|
|
psraw m2, m6
|
|
packuswb m0, m2
|
|
%endmacro
|
|
|
|
INIT_XMM ssse3
|
|
cglobal h264_biweight_16, 7, 8, 8
|
|
BIWEIGHT_SETUP
|
|
movifnidn r3d, r3m
|
|
|
|
.nextrow:
|
|
movh m0, [r0]
|
|
movh m2, [r0+8]
|
|
movh m3, [r1+8]
|
|
punpcklbw m0, [r1]
|
|
punpcklbw m2, m3
|
|
BIWEIGHT_SSSE3_OP
|
|
mova [r0], m0
|
|
add r0, r2
|
|
add r1, r2
|
|
dec r3d
|
|
jnz .nextrow
|
|
REP_RET
|
|
|
|
INIT_XMM ssse3
|
|
cglobal h264_biweight_8, 7, 8, 8
|
|
BIWEIGHT_SETUP
|
|
movifnidn r3d, r3m
|
|
sar r3, 1
|
|
lea r4, [r2*2]
|
|
|
|
.nextrow:
|
|
movh m0, [r0]
|
|
movh m1, [r1]
|
|
movh m2, [r0+r2]
|
|
movh m3, [r1+r2]
|
|
punpcklbw m0, m1
|
|
punpcklbw m2, m3
|
|
BIWEIGHT_SSSE3_OP
|
|
movh [r0], m0
|
|
movhps [r0+r2], m0
|
|
add r0, r4
|
|
add r1, r4
|
|
dec r3d
|
|
jnz .nextrow
|
|
REP_RET
|