mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-30 11:22:14 +00:00
52ec81c67d
Should fix compilation with old Yasm/Nasm versions. Signed-off-by: James Almer <jamrial@gmail.com>
389 lines
11 KiB
NASM
389 lines
11 KiB
NASM
; /*
|
|
; * Provide SIMD optimizations for transform_add functions for HEVC decoding
|
|
; * Copyright (c) 2014 Pierre-Edouard LEPERE
|
|
; *
|
|
; * This file is part of FFmpeg.
|
|
; *
|
|
; * FFmpeg is free software; you can redistribute it and/or
|
|
; * modify it under the terms of the GNU Lesser General Public
|
|
; * License as published by the Free Software Foundation; either
|
|
; * version 2.1 of the License, or (at your option) any later version.
|
|
; *
|
|
; * FFmpeg is distributed in the hope that it will be useful,
|
|
; * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
; * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
; * Lesser General Public License for more details.
|
|
; *
|
|
; * You should have received a copy of the GNU Lesser General Public
|
|
; * License along with FFmpeg; if not, write to the Free Software
|
|
; * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
; */
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION_RODATA 32
|
|
max_pixels_10: times 16 dw ((1 << 10)-1)
|
|
|
|
|
|
SECTION .text
|
|
|
|
;the tr_add macros and functions were largely inspired by x264 project's code in the h264_idct.asm file
|
|
%macro TR_ADD_MMX_4_8 0
|
|
mova m2, [r1]
|
|
mova m4, [r1+8]
|
|
pxor m3, m3
|
|
psubw m3, m2
|
|
packuswb m2, m2
|
|
packuswb m3, m3
|
|
pxor m5, m5
|
|
psubw m5, m4
|
|
packuswb m4, m4
|
|
packuswb m5, m5
|
|
|
|
movh m0, [r0 ]
|
|
movh m1, [r0+r2 ]
|
|
paddusb m0, m2
|
|
paddusb m1, m4
|
|
psubusb m0, m3
|
|
psubusb m1, m5
|
|
movh [r0 ], m0
|
|
movh [r0+r2 ], m1
|
|
%endmacro
|
|
|
|
|
|
INIT_MMX mmxext
|
|
; void ff_hevc_tranform_add_8_mmxext(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
|
|
cglobal hevc_transform_add4_8, 3, 4, 6
|
|
TR_ADD_MMX_4_8
|
|
add r1, 16
|
|
lea r0, [r0+r2*2]
|
|
TR_ADD_MMX_4_8
|
|
RET
|
|
|
|
%macro TR_ADD_SSE_8_8 0
|
|
pxor m3, m3
|
|
mova m4, [r1]
|
|
mova m6, [r1+16]
|
|
mova m0, [r1+32]
|
|
mova m2, [r1+48]
|
|
psubw m5, m3, m4
|
|
psubw m7, m3, m6
|
|
psubw m1, m3, m0
|
|
packuswb m4, m0
|
|
packuswb m5, m1
|
|
psubw m3, m2
|
|
packuswb m6, m2
|
|
packuswb m7, m3
|
|
|
|
movq m0, [r0 ]
|
|
movq m1, [r0+r2 ]
|
|
movhps m0, [r0+r2*2]
|
|
movhps m1, [r0+r3 ]
|
|
paddusb m0, m4
|
|
paddusb m1, m6
|
|
psubusb m0, m5
|
|
psubusb m1, m7
|
|
movq [r0 ], m0
|
|
movq [r0+r2 ], m1
|
|
movhps [r0+2*r2], m0
|
|
movhps [r0+r3 ], m1
|
|
%endmacro
|
|
|
|
%macro TR_ADD_SSE_16_32_8 3
|
|
mova xm2, [r1+%1 ]
|
|
mova xm6, [r1+%1+16]
|
|
%if cpuflag(avx2)
|
|
vinserti128 m2, m2, [r1+%1+32], 1
|
|
vinserti128 m6, m6, [r1+%1+48], 1
|
|
%endif
|
|
%if cpuflag(avx)
|
|
psubw m1, m0, m2
|
|
psubw m5, m0, m6
|
|
%else
|
|
mova m1, m0
|
|
mova m5, m0
|
|
psubw m1, m2
|
|
psubw m5, m6
|
|
%endif
|
|
packuswb m2, m6
|
|
packuswb m1, m5
|
|
|
|
mova xm4, [r1+%1+mmsize*2 ]
|
|
mova xm6, [r1+%1+mmsize*2+16]
|
|
%if cpuflag(avx2)
|
|
vinserti128 m4, m4, [r1+%1+96 ], 1
|
|
vinserti128 m6, m6, [r1+%1+112], 1
|
|
%endif
|
|
%if cpuflag(avx)
|
|
psubw m3, m0, m4
|
|
psubw m5, m0, m6
|
|
%else
|
|
mova m3, m0
|
|
mova m5, m0
|
|
psubw m3, m4
|
|
psubw m5, m6
|
|
%endif
|
|
packuswb m4, m6
|
|
packuswb m3, m5
|
|
|
|
paddusb m2, [%2]
|
|
paddusb m4, [%3]
|
|
psubusb m2, m1
|
|
psubusb m4, m3
|
|
mova [%2], m2
|
|
mova [%3], m4
|
|
%endmacro
|
|
|
|
|
|
%macro TRANSFORM_ADD_8 0
|
|
; void ff_hevc_transform_add8_8_<opt>(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
|
|
cglobal hevc_transform_add8_8, 3, 4, 8
|
|
lea r3, [r2*3]
|
|
TR_ADD_SSE_8_8
|
|
add r1, 64
|
|
lea r0, [r0+r2*4]
|
|
TR_ADD_SSE_8_8
|
|
RET
|
|
|
|
; void ff_hevc_transform_add16_8_<opt>(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
|
|
cglobal hevc_transform_add16_8, 3, 4, 7
|
|
pxor m0, m0
|
|
lea r3, [r2*3]
|
|
TR_ADD_SSE_16_32_8 0, r0, r0+r2
|
|
TR_ADD_SSE_16_32_8 64, r0+r2*2, r0+r3
|
|
%rep 3
|
|
add r1, 128
|
|
lea r0, [r0+r2*4]
|
|
TR_ADD_SSE_16_32_8 0, r0, r0+r2
|
|
TR_ADD_SSE_16_32_8 64, r0+r2*2, r0+r3
|
|
%endrep
|
|
RET
|
|
|
|
; void ff_hevc_transform_add32_8_<opt>(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
|
|
cglobal hevc_transform_add32_8, 3, 4, 7
|
|
pxor m0, m0
|
|
TR_ADD_SSE_16_32_8 0, r0, r0+16
|
|
TR_ADD_SSE_16_32_8 64, r0+r2, r0+r2+16
|
|
%rep 15
|
|
add r1, 128
|
|
lea r0, [r0+r2*2]
|
|
TR_ADD_SSE_16_32_8 0, r0, r0+16
|
|
TR_ADD_SSE_16_32_8 64, r0+r2, r0+r2+16
|
|
%endrep
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
TRANSFORM_ADD_8
|
|
INIT_XMM avx
|
|
TRANSFORM_ADD_8
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
INIT_YMM avx2
|
|
; void ff_hevc_transform_add32_8_avx2(uint8_t *dst, int16_t *coeffs, ptrdiff_t stride)
|
|
cglobal hevc_transform_add32_8, 3, 4, 7
|
|
pxor m0, m0
|
|
lea r3, [r2*3]
|
|
TR_ADD_SSE_16_32_8 0, r0, r0+r2
|
|
TR_ADD_SSE_16_32_8 128, r0+r2*2, r0+r3
|
|
%rep 7
|
|
add r1, 256
|
|
lea r0, [r0+r2*4]
|
|
TR_ADD_SSE_16_32_8 0, r0, r0+r2
|
|
TR_ADD_SSE_16_32_8 128, r0+r2*2, r0+r3
|
|
%endrep
|
|
RET
|
|
%endif
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_hevc_transform_add_10(pixel *dst, int16_t *block, int stride)
|
|
;-----------------------------------------------------------------------------
|
|
%macro TR_ADD_SSE_8_10 4
|
|
mova m0, [%4]
|
|
mova m1, [%4+16]
|
|
mova m2, [%4+32]
|
|
mova m3, [%4+48]
|
|
paddw m0, [%1+0 ]
|
|
paddw m1, [%1+%2 ]
|
|
paddw m2, [%1+%2*2]
|
|
paddw m3, [%1+%3 ]
|
|
CLIPW m0, m4, m5
|
|
CLIPW m1, m4, m5
|
|
CLIPW m2, m4, m5
|
|
CLIPW m3, m4, m5
|
|
mova [%1+0 ], m0
|
|
mova [%1+%2 ], m1
|
|
mova [%1+%2*2], m2
|
|
mova [%1+%3 ], m3
|
|
%endmacro
|
|
|
|
%macro TR_ADD_MMX4_10 3
|
|
mova m0, [%1+0 ]
|
|
mova m1, [%1+%2 ]
|
|
paddw m0, [%3]
|
|
paddw m1, [%3+8]
|
|
CLIPW m0, m2, m3
|
|
CLIPW m1, m2, m3
|
|
mova [%1+0 ], m0
|
|
mova [%1+%2 ], m1
|
|
%endmacro
|
|
|
|
%macro TRANS_ADD_SSE_16_10 3
|
|
mova m0, [%3]
|
|
mova m1, [%3+16]
|
|
mova m2, [%3+32]
|
|
mova m3, [%3+48]
|
|
paddw m0, [%1 ]
|
|
paddw m1, [%1+16 ]
|
|
paddw m2, [%1+%2 ]
|
|
paddw m3, [%1+%2+16]
|
|
CLIPW m0, m4, m5
|
|
CLIPW m1, m4, m5
|
|
CLIPW m2, m4, m5
|
|
CLIPW m3, m4, m5
|
|
mova [%1 ], m0
|
|
mova [%1+16 ], m1
|
|
mova [%1+%2 ], m2
|
|
mova [%1+%2+16], m3
|
|
%endmacro
|
|
|
|
%macro TRANS_ADD_SSE_32_10 2
|
|
mova m0, [%2]
|
|
mova m1, [%2+16]
|
|
mova m2, [%2+32]
|
|
mova m3, [%2+48]
|
|
|
|
paddw m0, [%1 ]
|
|
paddw m1, [%1+16]
|
|
paddw m2, [%1+32]
|
|
paddw m3, [%1+48]
|
|
CLIPW m0, m4, m5
|
|
CLIPW m1, m4, m5
|
|
CLIPW m2, m4, m5
|
|
CLIPW m3, m4, m5
|
|
mova [%1 ], m0
|
|
mova [%1+16], m1
|
|
mova [%1+32], m2
|
|
mova [%1+48], m3
|
|
%endmacro
|
|
|
|
%macro TRANS_ADD16_AVX2 4
|
|
mova m0, [%4]
|
|
mova m1, [%4+32]
|
|
mova m2, [%4+64]
|
|
mova m3, [%4+96]
|
|
|
|
paddw m0, [%1+0 ]
|
|
paddw m1, [%1+%2 ]
|
|
paddw m2, [%1+%2*2]
|
|
paddw m3, [%1+%3 ]
|
|
|
|
CLIPW m0, m4, m5
|
|
CLIPW m1, m4, m5
|
|
CLIPW m2, m4, m5
|
|
CLIPW m3, m4, m5
|
|
mova [%1+0 ], m0
|
|
mova [%1+%2 ], m1
|
|
mova [%1+%2*2], m2
|
|
mova [%1+%3 ], m3
|
|
%endmacro
|
|
|
|
%macro TRANS_ADD32_AVX2 3
|
|
mova m0, [%3]
|
|
mova m1, [%3+32]
|
|
mova m2, [%3+64]
|
|
mova m3, [%3+96]
|
|
|
|
paddw m0, [%1 ]
|
|
paddw m1, [%1+32 ]
|
|
paddw m2, [%1+%2 ]
|
|
paddw m3, [%1+%2+32]
|
|
|
|
CLIPW m0, m4, m5
|
|
CLIPW m1, m4, m5
|
|
CLIPW m2, m4, m5
|
|
CLIPW m3, m4, m5
|
|
mova [%1 ], m0
|
|
mova [%1+32 ], m1
|
|
mova [%1+%2 ], m2
|
|
mova [%1+%2+32], m3
|
|
%endmacro
|
|
|
|
|
|
INIT_MMX mmxext
|
|
cglobal hevc_transform_add4_10,3,4, 6
|
|
pxor m2, m2
|
|
mova m3, [max_pixels_10]
|
|
TR_ADD_MMX4_10 r0, r2, r1
|
|
add r1, 16
|
|
lea r0, [r0+2*r2]
|
|
TR_ADD_MMX4_10 r0, r2, r1
|
|
RET
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_hevc_transform_add_10(pixel *dst, int16_t *block, int stride)
|
|
;-----------------------------------------------------------------------------
|
|
INIT_XMM sse2
|
|
cglobal hevc_transform_add8_10,3,4,6
|
|
pxor m4, m4
|
|
mova m5, [max_pixels_10]
|
|
lea r3, [r2*3]
|
|
|
|
TR_ADD_SSE_8_10 r0, r2, r3, r1
|
|
lea r0, [r0+r2*4]
|
|
add r1, 64
|
|
TR_ADD_SSE_8_10 r0, r2, r3, r1
|
|
RET
|
|
|
|
cglobal hevc_transform_add16_10,3,4,6
|
|
pxor m4, m4
|
|
mova m5, [max_pixels_10]
|
|
|
|
TRANS_ADD_SSE_16_10 r0, r2, r1
|
|
%rep 7
|
|
lea r0, [r0+r2*2]
|
|
add r1, 64
|
|
TRANS_ADD_SSE_16_10 r0, r2, r1
|
|
%endrep
|
|
RET
|
|
|
|
cglobal hevc_transform_add32_10,3,4,6
|
|
pxor m4, m4
|
|
mova m5, [max_pixels_10]
|
|
|
|
TRANS_ADD_SSE_32_10 r0, r1
|
|
%rep 31
|
|
lea r0, [r0+r2]
|
|
add r1, 64
|
|
TRANS_ADD_SSE_32_10 r0, r1
|
|
%endrep
|
|
RET
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
INIT_YMM avx2
|
|
|
|
cglobal hevc_transform_add16_10,3,4,6
|
|
pxor m4, m4
|
|
mova m5, [max_pixels_10]
|
|
lea r3, [r2*3]
|
|
|
|
TRANS_ADD16_AVX2 r0, r2, r3, r1
|
|
%rep 3
|
|
lea r0, [r0+r2*4]
|
|
add r1, 128
|
|
TRANS_ADD16_AVX2 r0, r2, r3, r1
|
|
%endrep
|
|
RET
|
|
|
|
cglobal hevc_transform_add32_10,3,4,6
|
|
pxor m4, m4
|
|
mova m5, [max_pixels_10]
|
|
|
|
TRANS_ADD32_AVX2 r0, r2, r1
|
|
%rep 15
|
|
lea r0, [r0+r2*2]
|
|
add r1, 128
|
|
TRANS_ADD32_AVX2 r0, r2, r1
|
|
%endrep
|
|
RET
|
|
%endif ;HAVE_AVX_EXTERNAL
|