mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-27 09:52:17 +00:00
92b5800277
x64 always has MMX, MMXEXT, SSE and SSE2 and this means that some functions for MMX, MMXEXT and 3dnow are always overridden by other functions (unless one e.g. explicitly disables SSE2) for x64. So given that the only systems that benefit from these functions are truely ancient 32bit x86s they are removed. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
84 lines
2.5 KiB
NASM
84 lines
2.5 KiB
NASM
;*****************************************************************************
|
|
;* SIMD-optimized pixel operations
|
|
;*****************************************************************************
|
|
;* Copyright (c) 2000, 2001 Fabrice Bellard
|
|
;* Copyright (c) 2002-2004 Michael Niedermayer <michaelni@gmx.at>
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;*****************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION .text
|
|
|
|
INIT_XMM sse2
|
|
cglobal get_pixels, 3, 4, 5
|
|
lea r3, [r2*3]
|
|
pxor m4, m4
|
|
movh m0, [r1]
|
|
movh m1, [r1+r2]
|
|
movh m2, [r1+r2*2]
|
|
movh m3, [r1+r3]
|
|
lea r1, [r1+r2*4]
|
|
punpcklbw m0, m4
|
|
punpcklbw m1, m4
|
|
punpcklbw m2, m4
|
|
punpcklbw m3, m4
|
|
mova [r0], m0
|
|
mova [r0+0x10], m1
|
|
mova [r0+0x20], m2
|
|
mova [r0+0x30], m3
|
|
movh m0, [r1]
|
|
movh m1, [r1+r2*1]
|
|
movh m2, [r1+r2*2]
|
|
movh m3, [r1+r3]
|
|
punpcklbw m0, m4
|
|
punpcklbw m1, m4
|
|
punpcklbw m2, m4
|
|
punpcklbw m3, m4
|
|
mova [r0+0x40], m0
|
|
mova [r0+0x50], m1
|
|
mova [r0+0x60], m2
|
|
mova [r0+0x70], m3
|
|
RET
|
|
|
|
; void ff_diff_pixels(int16_t *block, const uint8_t *s1, const uint8_t *s2,
|
|
; ptrdiff_t stride);
|
|
INIT_XMM sse2
|
|
cglobal diff_pixels, 4,5,5
|
|
pxor m4, m4
|
|
add r0, 128
|
|
mov r4, -128
|
|
.loop:
|
|
movq m0, [r1]
|
|
movq m2, [r2]
|
|
movq m1, [r1+r3]
|
|
movq m3, [r2+r3]
|
|
punpcklbw m0, m4
|
|
punpcklbw m1, m4
|
|
punpcklbw m2, m4
|
|
punpcklbw m3, m4
|
|
psubw m0, m2
|
|
psubw m1, m3
|
|
mova [r0+r4+0], m0
|
|
mova [r0+r4+mmsize], m1
|
|
lea r1, [r1+r3*2]
|
|
lea r2, [r2+r3*2]
|
|
add r4, 2 * mmsize
|
|
jne .loop
|
|
RET
|