mirror of https://git.ffmpeg.org/ffmpeg.git
130 lines
3.5 KiB
NASM
130 lines
3.5 KiB
NASM
;*****************************************************************************
|
|
;* SIMD-optimized pixel operations
|
|
;*****************************************************************************
|
|
;* Copyright (c) 2000, 2001 Fabrice Bellard
|
|
;* Copyright (c) 2002-2004 Michael Niedermayer <michaelni@gmx.at>
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;*****************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION .text
|
|
|
|
INIT_MMX mmx
|
|
; void ff_get_pixels_mmx(int16_t *block, const uint8_t *pixels, ptrdiff_t line_size)
|
|
cglobal get_pixels, 3,4
|
|
add r0, 128
|
|
mov r3, -128
|
|
pxor m7, m7
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m2, [r1+r2]
|
|
mova m1, m0
|
|
mova m3, m2
|
|
punpcklbw m0, m7
|
|
punpckhbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpckhbw m3, m7
|
|
mova [r0+r3+ 0], m0
|
|
mova [r0+r3+ 8], m1
|
|
mova [r0+r3+16], m2
|
|
mova [r0+r3+24], m3
|
|
lea r1, [r1+r2*2]
|
|
add r3, 32
|
|
js .loop
|
|
REP_RET
|
|
|
|
INIT_XMM sse2
|
|
cglobal get_pixels, 3, 4, 5
|
|
lea r3, [r2*3]
|
|
pxor m4, m4
|
|
movh m0, [r1]
|
|
movh m1, [r1+r2]
|
|
movh m2, [r1+r2*2]
|
|
movh m3, [r1+r3]
|
|
lea r1, [r1+r2*4]
|
|
punpcklbw m0, m4
|
|
punpcklbw m1, m4
|
|
punpcklbw m2, m4
|
|
punpcklbw m3, m4
|
|
mova [r0], m0
|
|
mova [r0+0x10], m1
|
|
mova [r0+0x20], m2
|
|
mova [r0+0x30], m3
|
|
movh m0, [r1]
|
|
movh m1, [r1+r2*1]
|
|
movh m2, [r1+r2*2]
|
|
movh m3, [r1+r3]
|
|
punpcklbw m0, m4
|
|
punpcklbw m1, m4
|
|
punpcklbw m2, m4
|
|
punpcklbw m3, m4
|
|
mova [r0+0x40], m0
|
|
mova [r0+0x50], m1
|
|
mova [r0+0x60], m2
|
|
mova [r0+0x70], m3
|
|
RET
|
|
|
|
; void ff_diff_pixels_mmx(int16_t *block, const uint8_t *s1, const uint8_t *s2,
|
|
; int stride);
|
|
%macro DIFF_PIXELS 0
|
|
cglobal diff_pixels, 4,5,5
|
|
movsxdifnidn r3, r3d
|
|
pxor m4, m4
|
|
add r0, 128
|
|
mov r4, -128
|
|
.loop:
|
|
movq m0, [r1]
|
|
movq m2, [r2]
|
|
%if mmsize == 8
|
|
movq m1, m0
|
|
movq m3, m2
|
|
punpcklbw m0, m4
|
|
punpckhbw m1, m4
|
|
punpcklbw m2, m4
|
|
punpckhbw m3, m4
|
|
%else
|
|
movq m1, [r1+r3]
|
|
movq m3, [r2+r3]
|
|
punpcklbw m0, m4
|
|
punpcklbw m1, m4
|
|
punpcklbw m2, m4
|
|
punpcklbw m3, m4
|
|
%endif
|
|
psubw m0, m2
|
|
psubw m1, m3
|
|
mova [r0+r4+0], m0
|
|
mova [r0+r4+mmsize], m1
|
|
%if mmsize == 8
|
|
add r1, r3
|
|
add r2, r3
|
|
%else
|
|
lea r1, [r1+r3*2]
|
|
lea r2, [r2+r3*2]
|
|
%endif
|
|
add r4, 2 * mmsize
|
|
jne .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmx
|
|
DIFF_PIXELS
|
|
|
|
INIT_XMM sse2
|
|
DIFF_PIXELS
|