2014-06-23 20:01:03 +00:00
|
|
|
;******************************************************************************
|
2014-07-01 21:10:15 +00:00
|
|
|
;* SIMD-optimized IDCT-related routines
|
2014-06-23 20:01:03 +00:00
|
|
|
;* Copyright (c) 2008 Loren Merritt
|
|
|
|
;* Copyright (c) 2003-2013 Michael Niedermayer
|
|
|
|
;* Copyright (c) 2013 Daniel Kang
|
|
|
|
;*
|
|
|
|
;* This file is part of FFmpeg.
|
|
|
|
;*
|
|
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
|
|
;* License as published by the Free Software Foundation; either
|
|
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
;*
|
|
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
;* Lesser General Public License for more details.
|
|
|
|
;*
|
|
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
;******************************************************************************
|
|
|
|
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
|
|
|
|
SECTION_RODATA
|
|
|
|
|
|
|
|
cextern pb_80
|
|
|
|
|
|
|
|
SECTION_TEXT
|
|
|
|
|
|
|
|
;--------------------------------------------------------------------------
|
|
|
|
;void ff_put_signed_pixels_clamped(const int16_t *block, uint8_t *pixels,
|
2014-09-24 22:51:02 +00:00
|
|
|
; ptrdiff_t line_size)
|
2014-06-23 20:01:03 +00:00
|
|
|
;--------------------------------------------------------------------------
|
|
|
|
|
|
|
|
%macro PUT_SIGNED_PIXELS_CLAMPED_HALF 1
|
|
|
|
mova m1, [blockq+mmsize*0+%1]
|
|
|
|
mova m2, [blockq+mmsize*2+%1]
|
|
|
|
%if mmsize == 8
|
|
|
|
mova m3, [blockq+mmsize*4+%1]
|
|
|
|
mova m4, [blockq+mmsize*6+%1]
|
|
|
|
%endif
|
|
|
|
packsswb m1, [blockq+mmsize*1+%1]
|
|
|
|
packsswb m2, [blockq+mmsize*3+%1]
|
|
|
|
%if mmsize == 8
|
|
|
|
packsswb m3, [blockq+mmsize*5+%1]
|
|
|
|
packsswb m4, [blockq+mmsize*7+%1]
|
|
|
|
%endif
|
|
|
|
paddb m1, m0
|
|
|
|
paddb m2, m0
|
|
|
|
%if mmsize == 8
|
|
|
|
paddb m3, m0
|
|
|
|
paddb m4, m0
|
|
|
|
movq [pixelsq+lsizeq*0], m1
|
|
|
|
movq [pixelsq+lsizeq*1], m2
|
|
|
|
movq [pixelsq+lsizeq*2], m3
|
|
|
|
movq [pixelsq+lsize3q ], m4
|
|
|
|
%else
|
|
|
|
movq [pixelsq+lsizeq*0], m1
|
|
|
|
movhps [pixelsq+lsizeq*1], m1
|
|
|
|
movq [pixelsq+lsizeq*2], m2
|
|
|
|
movhps [pixelsq+lsize3q ], m2
|
|
|
|
%endif
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro PUT_SIGNED_PIXELS_CLAMPED 1
|
|
|
|
cglobal put_signed_pixels_clamped, 3, 4, %1, block, pixels, lsize, lsize3
|
|
|
|
mova m0, [pb_80]
|
|
|
|
lea lsize3q, [lsizeq*3]
|
|
|
|
PUT_SIGNED_PIXELS_CLAMPED_HALF 0
|
|
|
|
lea pixelsq, [pixelsq+lsizeq*4]
|
|
|
|
PUT_SIGNED_PIXELS_CLAMPED_HALF 64
|
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmx
|
|
|
|
PUT_SIGNED_PIXELS_CLAMPED 0
|
|
|
|
INIT_XMM sse2
|
|
|
|
PUT_SIGNED_PIXELS_CLAMPED 3
|
2014-09-24 22:53:07 +00:00
|
|
|
|
|
|
|
;--------------------------------------------------------------------------
|
|
|
|
; void ff_put_pixels_clamped(const int16_t *block, uint8_t *pixels,
|
|
|
|
; ptrdiff_t line_size);
|
|
|
|
;--------------------------------------------------------------------------
|
|
|
|
; %1 = block offset
|
|
|
|
%macro PUT_PIXELS_CLAMPED_HALF 1
|
|
|
|
mova m0, [blockq+mmsize*0+%1]
|
|
|
|
mova m1, [blockq+mmsize*2+%1]
|
|
|
|
%if mmsize == 8
|
|
|
|
mova m2, [blockq+mmsize*4+%1]
|
|
|
|
mova m3, [blockq+mmsize*6+%1]
|
|
|
|
%endif
|
|
|
|
packuswb m0, [blockq+mmsize*1+%1]
|
|
|
|
packuswb m1, [blockq+mmsize*3+%1]
|
|
|
|
%if mmsize == 8
|
|
|
|
packuswb m2, [blockq+mmsize*5+%1]
|
|
|
|
packuswb m3, [blockq+mmsize*7+%1]
|
|
|
|
movq [pixelsq], m0
|
|
|
|
movq [lsizeq+pixelsq], m1
|
|
|
|
movq [2*lsizeq+pixelsq], m2
|
|
|
|
movq [lsize3q+pixelsq], m3
|
|
|
|
%else
|
|
|
|
movq [pixelsq], m0
|
|
|
|
movhps [lsizeq+pixelsq], m0
|
|
|
|
movq [2*lsizeq+pixelsq], m1
|
|
|
|
movhps [lsize3q+pixelsq], m1
|
|
|
|
%endif
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro PUT_PIXELS_CLAMPED 0
|
|
|
|
cglobal put_pixels_clamped, 3, 4, 2, block, pixels, lsize, lsize3
|
|
|
|
lea lsize3q, [lsizeq*3]
|
|
|
|
PUT_PIXELS_CLAMPED_HALF 0
|
|
|
|
lea pixelsq, [pixelsq+lsizeq*4]
|
|
|
|
PUT_PIXELS_CLAMPED_HALF 64
|
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmx
|
|
|
|
PUT_PIXELS_CLAMPED
|
|
|
|
INIT_XMM sse2
|
|
|
|
PUT_PIXELS_CLAMPED
|
|
|
|
|
|
|
|
;--------------------------------------------------------------------------
|
|
|
|
; void ff_add_pixels_clamped(const int16_t *block, uint8_t *pixels,
|
|
|
|
; ptrdiff_t line_size);
|
|
|
|
;--------------------------------------------------------------------------
|
|
|
|
; %1 = block offset
|
|
|
|
%macro ADD_PIXELS_CLAMPED 1
|
|
|
|
mova m0, [blockq+mmsize*0+%1]
|
|
|
|
mova m1, [blockq+mmsize*1+%1]
|
|
|
|
%if mmsize == 8
|
|
|
|
mova m5, [blockq+mmsize*2+%1]
|
|
|
|
mova m6, [blockq+mmsize*3+%1]
|
|
|
|
%endif
|
|
|
|
movq m2, [pixelsq]
|
|
|
|
movq m3, [pixelsq+lsizeq]
|
|
|
|
%if mmsize == 8
|
|
|
|
mova m7, m2
|
|
|
|
punpcklbw m2, m4
|
|
|
|
punpckhbw m7, m4
|
|
|
|
paddsw m0, m2
|
|
|
|
paddsw m1, m7
|
|
|
|
mova m7, m3
|
|
|
|
punpcklbw m3, m4
|
|
|
|
punpckhbw m7, m4
|
|
|
|
paddsw m5, m3
|
|
|
|
paddsw m6, m7
|
|
|
|
%else
|
|
|
|
punpcklbw m2, m4
|
|
|
|
punpcklbw m3, m4
|
|
|
|
paddsw m0, m2
|
|
|
|
paddsw m1, m3
|
|
|
|
%endif
|
|
|
|
packuswb m0, m1
|
|
|
|
%if mmsize == 8
|
|
|
|
packuswb m5, m6
|
|
|
|
movq [pixelsq], m0
|
|
|
|
movq [pixelsq+lsizeq], m5
|
|
|
|
%else
|
|
|
|
movq [pixelsq], m0
|
|
|
|
movhps [pixelsq+lsizeq], m0
|
|
|
|
%endif
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro ADD_PIXELS_CLAMPED 0
|
|
|
|
cglobal add_pixels_clamped, 3, 3, 5, block, pixels, lsize
|
|
|
|
pxor m4, m4
|
|
|
|
ADD_PIXELS_CLAMPED 0
|
|
|
|
lea pixelsq, [pixelsq+lsizeq*2]
|
|
|
|
ADD_PIXELS_CLAMPED 32
|
|
|
|
lea pixelsq, [pixelsq+lsizeq*2]
|
|
|
|
ADD_PIXELS_CLAMPED 64
|
|
|
|
lea pixelsq, [pixelsq+lsizeq*2]
|
|
|
|
ADD_PIXELS_CLAMPED 96
|
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmx
|
|
|
|
ADD_PIXELS_CLAMPED
|
|
|
|
INIT_XMM sse2
|
|
|
|
ADD_PIXELS_CLAMPED
|