mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-27 18:02:11 +00:00
98a6806fdd
* commit '368f50359eb328b0b9d67451f56fda20b3255f9a': dsputil: Split off quarterpel bits into their own context Conflicts: configure libavcodec/dsputil.c libavcodec/h263dec.c libavcodec/mpegvideo.c libavcodec/mpegvideo_enc.c libavcodec/vc1dec.c libavcodec/vc1dsp.c libavcodec/x86/dsputil_init.c libavcodec/x86/qpeldsp.asm Merged-by: Michael Niedermayer <michaelni@gmx.at>
561 lines
14 KiB
NASM
561 lines
14 KiB
NASM
;******************************************************************************
|
|
;* mpeg4 qpel
|
|
;* Copyright (c) 2003 Michael Niedermayer <michaelni@gmx.at>
|
|
;* Copyright (c) 2008 Loren Merritt
|
|
;* Copyright (c) 2013 Daniel Kang
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION_RODATA
|
|
cextern pb_1
|
|
cextern pw_3
|
|
cextern pw_15
|
|
cextern pw_16
|
|
cextern pw_20
|
|
|
|
|
|
SECTION_TEXT
|
|
|
|
; void ff_put_no_rnd_pixels8_l2(uint8_t *dst, uint8_t *src1, uint8_t *src2, int dstStride, int src1Stride, int h)
|
|
%macro PUT_NO_RND_PIXELS8_L2 0
|
|
cglobal put_no_rnd_pixels8_l2, 6,6
|
|
movsxdifnidn r4, r4d
|
|
movsxdifnidn r3, r3d
|
|
pcmpeqb m6, m6
|
|
test r5d, 1
|
|
je .loop
|
|
mova m0, [r1]
|
|
mova m1, [r2]
|
|
add r1, r4
|
|
add r2, 8
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
PAVGB m0, m1
|
|
pxor m0, m6
|
|
mova [r0], m0
|
|
add r0, r3
|
|
dec r5d
|
|
.loop:
|
|
mova m0, [r1]
|
|
add r1, r4
|
|
mova m1, [r1]
|
|
add r1, r4
|
|
mova m2, [r2]
|
|
mova m3, [r2+8]
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
pxor m2, m6
|
|
pxor m3, m6
|
|
PAVGB m0, m2
|
|
PAVGB m1, m3
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
mova [r0], m0
|
|
add r0, r3
|
|
mova [r0], m1
|
|
add r0, r3
|
|
mova m0, [r1]
|
|
add r1, r4
|
|
mova m1, [r1]
|
|
add r1, r4
|
|
mova m2, [r2+16]
|
|
mova m3, [r2+24]
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
pxor m2, m6
|
|
pxor m3, m6
|
|
PAVGB m0, m2
|
|
PAVGB m1, m3
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
mova [r0], m0
|
|
add r0, r3
|
|
mova [r0], m1
|
|
add r0, r3
|
|
add r2, 32
|
|
sub r5d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PUT_NO_RND_PIXELS8_L2
|
|
|
|
|
|
; void ff_put_no_rnd_pixels16_l2(uint8_t *dst, uint8_t *src1, uint8_t *src2, int dstStride, int src1Stride, int h)
|
|
%macro PUT_NO_RND_PIXELS16_l2 0
|
|
cglobal put_no_rnd_pixels16_l2, 6,6
|
|
movsxdifnidn r3, r3d
|
|
movsxdifnidn r4, r4d
|
|
pcmpeqb m6, m6
|
|
test r5d, 1
|
|
je .loop
|
|
mova m0, [r1]
|
|
mova m1, [r1+8]
|
|
mova m2, [r2]
|
|
mova m3, [r2+8]
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
pxor m2, m6
|
|
pxor m3, m6
|
|
PAVGB m0, m2
|
|
PAVGB m1, m3
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
add r1, r4
|
|
add r2, 16
|
|
mova [r0], m0
|
|
mova [r0+8], m1
|
|
add r0, r3
|
|
dec r5d
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m1, [r1+8]
|
|
add r1, r4
|
|
mova m2, [r2]
|
|
mova m3, [r2+8]
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
pxor m2, m6
|
|
pxor m3, m6
|
|
PAVGB m0, m2
|
|
PAVGB m1, m3
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
mova [r0], m0
|
|
mova [r0+8], m1
|
|
add r0, r3
|
|
mova m0, [r1]
|
|
mova m1, [r1+8]
|
|
add r1, r4
|
|
mova m2, [r2+16]
|
|
mova m3, [r2+24]
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
pxor m2, m6
|
|
pxor m3, m6
|
|
PAVGB m0, m2
|
|
PAVGB m1, m3
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
mova [r0], m0
|
|
mova [r0+8], m1
|
|
add r0, r3
|
|
add r2, 32
|
|
sub r5d, 2
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PUT_NO_RND_PIXELS16_l2
|
|
INIT_MMX 3dnow
|
|
PUT_NO_RND_PIXELS16_l2
|
|
|
|
%macro MPEG4_QPEL16_H_LOWPASS 1
|
|
cglobal %1_mpeg4_qpel16_h_lowpass, 5, 5, 0, 16
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
pxor m7, m7
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m1, m0
|
|
mova m2, m0
|
|
punpcklbw m0, m7
|
|
punpckhbw m1, m7
|
|
pshufw m5, m0, 0x90
|
|
pshufw m6, m0, 0x41
|
|
mova m3, m2
|
|
mova m4, m2
|
|
psllq m2, 8
|
|
psllq m3, 16
|
|
psllq m4, 24
|
|
punpckhbw m2, m7
|
|
punpckhbw m3, m7
|
|
punpckhbw m4, m7
|
|
paddw m5, m3
|
|
paddw m6, m2
|
|
paddw m5, m5
|
|
psubw m6, m5
|
|
pshufw m5, m0, 6
|
|
pmullw m6, [pw_3]
|
|
paddw m0, m4
|
|
paddw m5, m1
|
|
pmullw m0, [pw_20]
|
|
psubw m0, m5
|
|
paddw m6, [PW_ROUND]
|
|
paddw m0, m6
|
|
psraw m0, 5
|
|
mova [rsp+8], m0
|
|
mova m0, [r1+5]
|
|
mova m5, m0
|
|
mova m6, m0
|
|
psrlq m0, 8
|
|
psrlq m5, 16
|
|
punpcklbw m0, m7
|
|
punpcklbw m5, m7
|
|
paddw m2, m0
|
|
paddw m3, m5
|
|
paddw m2, m2
|
|
psubw m3, m2
|
|
mova m2, m6
|
|
psrlq m6, 24
|
|
punpcklbw m2, m7
|
|
punpcklbw m6, m7
|
|
pmullw m3, [pw_3]
|
|
paddw m1, m2
|
|
paddw m4, m6
|
|
pmullw m1, [pw_20]
|
|
psubw m3, m4
|
|
paddw m1, [PW_ROUND]
|
|
paddw m3, m1
|
|
psraw m3, 5
|
|
mova m1, [rsp+8]
|
|
packuswb m1, m3
|
|
OP_MOV [r0], m1, m4
|
|
mova m1, [r1+9]
|
|
mova m4, m1
|
|
mova m3, m1
|
|
psrlq m1, 8
|
|
psrlq m4, 16
|
|
punpcklbw m1, m7
|
|
punpcklbw m4, m7
|
|
paddw m5, m1
|
|
paddw m0, m4
|
|
paddw m5, m5
|
|
psubw m0, m5
|
|
mova m5, m3
|
|
psrlq m3, 24
|
|
pmullw m0, [pw_3]
|
|
punpcklbw m3, m7
|
|
paddw m2, m3
|
|
psubw m0, m2
|
|
mova m2, m5
|
|
punpcklbw m2, m7
|
|
punpckhbw m5, m7
|
|
paddw m6, m2
|
|
pmullw m6, [pw_20]
|
|
paddw m0, [PW_ROUND]
|
|
paddw m0, m6
|
|
psraw m0, 5
|
|
paddw m3, m5
|
|
pshufw m6, m5, 0xf9
|
|
paddw m6, m4
|
|
pshufw m4, m5, 0xbe
|
|
pshufw m5, m5, 0x6f
|
|
paddw m4, m1
|
|
paddw m5, m2
|
|
paddw m6, m6
|
|
psubw m4, m6
|
|
pmullw m3, [pw_20]
|
|
pmullw m4, [pw_3]
|
|
psubw m3, m5
|
|
paddw m4, [PW_ROUND]
|
|
paddw m4, m3
|
|
psraw m4, 5
|
|
packuswb m0, m4
|
|
OP_MOV [r0+8], m0, m4
|
|
add r1, r3
|
|
add r0, r2
|
|
dec r4d
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
%macro PUT_OP 2-3
|
|
mova %1, %2
|
|
%endmacro
|
|
|
|
%macro AVG_OP 2-3
|
|
mova %3, %1
|
|
pavgb %2, %3
|
|
mova %1, %2
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
%define PW_ROUND pw_16
|
|
%define OP_MOV PUT_OP
|
|
MPEG4_QPEL16_H_LOWPASS put
|
|
%define PW_ROUND pw_16
|
|
%define OP_MOV AVG_OP
|
|
MPEG4_QPEL16_H_LOWPASS avg
|
|
%define PW_ROUND pw_15
|
|
%define OP_MOV PUT_OP
|
|
MPEG4_QPEL16_H_LOWPASS put_no_rnd
|
|
|
|
|
|
|
|
%macro MPEG4_QPEL8_H_LOWPASS 1
|
|
cglobal %1_mpeg4_qpel8_h_lowpass, 5, 5, 0, 8
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
pxor m7, m7
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m1, m0
|
|
mova m2, m0
|
|
punpcklbw m0, m7
|
|
punpckhbw m1, m7
|
|
pshufw m5, m0, 0x90
|
|
pshufw m6, m0, 0x41
|
|
mova m3, m2
|
|
mova m4, m2
|
|
psllq m2, 8
|
|
psllq m3, 16
|
|
psllq m4, 24
|
|
punpckhbw m2, m7
|
|
punpckhbw m3, m7
|
|
punpckhbw m4, m7
|
|
paddw m5, m3
|
|
paddw m6, m2
|
|
paddw m5, m5
|
|
psubw m6, m5
|
|
pshufw m5, m0, 0x6
|
|
pmullw m6, [pw_3]
|
|
paddw m0, m4
|
|
paddw m5, m1
|
|
pmullw m0, [pw_20]
|
|
psubw m0, m5
|
|
paddw m6, [PW_ROUND]
|
|
paddw m0, m6
|
|
psraw m0, 5
|
|
movh m5, [r1+5]
|
|
punpcklbw m5, m7
|
|
pshufw m6, m5, 0xf9
|
|
paddw m1, m5
|
|
paddw m2, m6
|
|
pshufw m6, m5, 0xbe
|
|
pshufw m5, m5, 0x6f
|
|
paddw m3, m6
|
|
paddw m4, m5
|
|
paddw m2, m2
|
|
psubw m3, m2
|
|
pmullw m1, [pw_20]
|
|
pmullw m3, [pw_3]
|
|
psubw m3, m4
|
|
paddw m1, [PW_ROUND]
|
|
paddw m3, m1
|
|
psraw m3, 5
|
|
packuswb m0, m3
|
|
OP_MOV [r0], m0, m4
|
|
add r1, r3
|
|
add r0, r2
|
|
dec r4d
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
%define PW_ROUND pw_16
|
|
%define OP_MOV PUT_OP
|
|
MPEG4_QPEL8_H_LOWPASS put
|
|
%define PW_ROUND pw_16
|
|
%define OP_MOV AVG_OP
|
|
MPEG4_QPEL8_H_LOWPASS avg
|
|
%define PW_ROUND pw_15
|
|
%define OP_MOV PUT_OP
|
|
MPEG4_QPEL8_H_LOWPASS put_no_rnd
|
|
|
|
|
|
|
|
%macro QPEL_V_LOW 5
|
|
paddw m0, m1
|
|
mova m4, [pw_20]
|
|
pmullw m4, m0
|
|
mova m0, %4
|
|
mova m5, %1
|
|
paddw m5, m0
|
|
psubw m4, m5
|
|
mova m5, %2
|
|
mova m6, %3
|
|
paddw m5, m3
|
|
paddw m6, m2
|
|
paddw m6, m6
|
|
psubw m5, m6
|
|
pmullw m5, [pw_3]
|
|
paddw m4, [PW_ROUND]
|
|
paddw m5, m4
|
|
psraw m5, 5
|
|
packuswb m5, m5
|
|
OP_MOV %5, m5, m7
|
|
SWAP 0,1,2,3
|
|
%endmacro
|
|
|
|
%macro MPEG4_QPEL16_V_LOWPASS 1
|
|
cglobal %1_mpeg4_qpel16_v_lowpass, 4, 6, 0, 544
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
|
|
mov r4d, 17
|
|
mov r5, rsp
|
|
pxor m7, m7
|
|
.looph:
|
|
mova m0, [r1]
|
|
mova m1, [r1]
|
|
mova m2, [r1+8]
|
|
mova m3, [r1+8]
|
|
punpcklbw m0, m7
|
|
punpckhbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpckhbw m3, m7
|
|
mova [r5], m0
|
|
mova [r5+0x88], m1
|
|
mova [r5+0x110], m2
|
|
mova [r5+0x198], m3
|
|
add r5, 8
|
|
add r1, r3
|
|
dec r4d
|
|
jne .looph
|
|
|
|
|
|
; NOTE: r1 CHANGES VALUES: r1 -> 4 - 14*dstStride
|
|
mov r4d, 4
|
|
mov r1, 4
|
|
neg r2
|
|
lea r1, [r1+r2*8]
|
|
lea r1, [r1+r2*4]
|
|
lea r1, [r1+r2*2]
|
|
neg r2
|
|
mov r5, rsp
|
|
.loopv:
|
|
pxor m7, m7
|
|
mova m0, [r5+ 0x0]
|
|
mova m1, [r5+ 0x8]
|
|
mova m2, [r5+0x10]
|
|
mova m3, [r5+0x18]
|
|
QPEL_V_LOW [r5+0x10], [r5+ 0x8], [r5+ 0x0], [r5+0x20], [r0]
|
|
QPEL_V_LOW [r5+ 0x8], [r5+ 0x0], [r5+ 0x0], [r5+0x28], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+ 0x0], [r5+ 0x0], [r5+ 0x8], [r5+0x30], [r0]
|
|
QPEL_V_LOW [r5+ 0x0], [r5+ 0x8], [r5+0x10], [r5+0x38], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+ 0x8], [r5+0x10], [r5+0x18], [r5+0x40], [r0]
|
|
QPEL_V_LOW [r5+0x10], [r5+0x18], [r5+0x20], [r5+0x48], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+0x18], [r5+0x20], [r5+0x28], [r5+0x50], [r0]
|
|
QPEL_V_LOW [r5+0x20], [r5+0x28], [r5+0x30], [r5+0x58], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+0x28], [r5+0x30], [r5+0x38], [r5+0x60], [r0]
|
|
QPEL_V_LOW [r5+0x30], [r5+0x38], [r5+0x40], [r5+0x68], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+0x38], [r5+0x40], [r5+0x48], [r5+0x70], [r0]
|
|
QPEL_V_LOW [r5+0x40], [r5+0x48], [r5+0x50], [r5+0x78], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+0x48], [r5+0x50], [r5+0x58], [r5+0x80], [r0]
|
|
QPEL_V_LOW [r5+0x50], [r5+0x58], [r5+0x60], [r5+0x80], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+0x58], [r5+0x60], [r5+0x68], [r5+0x78], [r0]
|
|
QPEL_V_LOW [r5+0x60], [r5+0x68], [r5+0x70], [r5+0x70], [r0+r2]
|
|
|
|
add r5, 0x88
|
|
add r0, r1
|
|
dec r4d
|
|
jne .loopv
|
|
REP_RET
|
|
%endmacro
|
|
|
|
%macro PUT_OPH 2-3
|
|
movh %1, %2
|
|
%endmacro
|
|
|
|
%macro AVG_OPH 2-3
|
|
movh %3, %1
|
|
pavgb %2, %3
|
|
movh %1, %2
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
%define PW_ROUND pw_16
|
|
%define OP_MOV PUT_OPH
|
|
MPEG4_QPEL16_V_LOWPASS put
|
|
%define PW_ROUND pw_16
|
|
%define OP_MOV AVG_OPH
|
|
MPEG4_QPEL16_V_LOWPASS avg
|
|
%define PW_ROUND pw_15
|
|
%define OP_MOV PUT_OPH
|
|
MPEG4_QPEL16_V_LOWPASS put_no_rnd
|
|
|
|
|
|
|
|
%macro MPEG4_QPEL8_V_LOWPASS 1
|
|
cglobal %1_mpeg4_qpel8_v_lowpass, 4, 6, 0, 288
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
|
|
mov r4d, 9
|
|
mov r5, rsp
|
|
pxor m7, m7
|
|
.looph:
|
|
mova m0, [r1]
|
|
mova m1, [r1]
|
|
punpcklbw m0, m7
|
|
punpckhbw m1, m7
|
|
mova [r5], m0
|
|
mova [r5+0x48], m1
|
|
add r5, 8
|
|
add r1, r3
|
|
dec r4d
|
|
jne .looph
|
|
|
|
|
|
; NOTE: r1 CHANGES VALUES: r1 -> 4 - 6*dstStride
|
|
mov r4d, 2
|
|
mov r1, 4
|
|
neg r2
|
|
lea r1, [r1+r2*4]
|
|
lea r1, [r1+r2*2]
|
|
neg r2
|
|
mov r5, rsp
|
|
.loopv:
|
|
pxor m7, m7
|
|
mova m0, [r5+ 0x0]
|
|
mova m1, [r5+ 0x8]
|
|
mova m2, [r5+0x10]
|
|
mova m3, [r5+0x18]
|
|
QPEL_V_LOW [r5+0x10], [r5+ 0x8], [r5+ 0x0], [r5+0x20], [r0]
|
|
QPEL_V_LOW [r5+ 0x8], [r5+ 0x0], [r5+ 0x0], [r5+0x28], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+ 0x0], [r5+ 0x0], [r5+ 0x8], [r5+0x30], [r0]
|
|
QPEL_V_LOW [r5+ 0x0], [r5+ 0x8], [r5+0x10], [r5+0x38], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+ 0x8], [r5+0x10], [r5+0x18], [r5+0x40], [r0]
|
|
QPEL_V_LOW [r5+0x10], [r5+0x18], [r5+0x20], [r5+0x40], [r0+r2]
|
|
lea r0, [r0+r2*2]
|
|
QPEL_V_LOW [r5+0x18], [r5+0x20], [r5+0x28], [r5+0x38], [r0]
|
|
QPEL_V_LOW [r5+0x20], [r5+0x28], [r5+0x30], [r5+0x30], [r0+r2]
|
|
|
|
add r5, 0x48
|
|
add r0, r1
|
|
dec r4d
|
|
jne .loopv
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
%define PW_ROUND pw_16
|
|
%define OP_MOV PUT_OPH
|
|
MPEG4_QPEL8_V_LOWPASS put
|
|
%define PW_ROUND pw_16
|
|
%define OP_MOV AVG_OPH
|
|
MPEG4_QPEL8_V_LOWPASS avg
|
|
%define PW_ROUND pw_15
|
|
%define OP_MOV PUT_OPH
|
|
MPEG4_QPEL8_V_LOWPASS put_no_rnd
|