mirror of https://git.ffmpeg.org/ffmpeg.git
308 lines
7.3 KiB
NASM
308 lines
7.3 KiB
NASM
;******************************************************************************
|
|
;* MMX optimized discrete wavelet trasnform
|
|
;* Copyright (c) 2010 David Conrad
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION_RODATA
|
|
pw_1991: times 4 dw 9,-1
|
|
|
|
cextern pw_1
|
|
cextern pw_2
|
|
cextern pw_8
|
|
cextern pw_16
|
|
|
|
section .text
|
|
|
|
; %1 -= (%2 + %3 + 2)>>2 %4 is pw_2
|
|
%macro COMPOSE_53iL0 4
|
|
paddw %2, %3
|
|
paddw %2, %4
|
|
psraw %2, 2
|
|
psubw %1, %2
|
|
%endm
|
|
|
|
; m1 = %1 + (-m0 + 9*m1 + 9*%2 -%3 + 8)>>4
|
|
; if %4 is supplied, %1 is loaded unaligned from there
|
|
; m2: clobbered m3: pw_8 m4: pw_1991
|
|
%macro COMPOSE_DD97iH0 3-4
|
|
paddw m0, %3
|
|
paddw m1, %2
|
|
psubw m0, m3
|
|
mova m2, m1
|
|
punpcklwd m1, m0
|
|
punpckhwd m2, m0
|
|
pmaddwd m1, m4
|
|
pmaddwd m2, m4
|
|
%if %0 > 3
|
|
movu %1, %4
|
|
%endif
|
|
psrad m1, 4
|
|
psrad m2, 4
|
|
packssdw m1, m2
|
|
paddw m1, %1
|
|
%endm
|
|
|
|
%macro COMPOSE_VERTICAL 1
|
|
; void vertical_compose53iL0(IDWTELEM *b0, IDWTELEM *b1, IDWTELEM *b2,
|
|
; int width)
|
|
cglobal vertical_compose53iL0_%1, 4,4,1, b0, b1, b2, width
|
|
mova m2, [pw_2]
|
|
%if ARCH_X86_64
|
|
mov widthd, widthd
|
|
%endif
|
|
.loop:
|
|
sub widthq, mmsize/2
|
|
mova m1, [b0q+2*widthq]
|
|
mova m0, [b1q+2*widthq]
|
|
COMPOSE_53iL0 m0, m1, [b2q+2*widthq], m2
|
|
mova [b1q+2*widthq], m0
|
|
jg .loop
|
|
REP_RET
|
|
|
|
; void vertical_compose_dirac53iH0(IDWTELEM *b0, IDWTELEM *b1, IDWTELEM *b2,
|
|
; int width)
|
|
cglobal vertical_compose_dirac53iH0_%1, 4,4,1, b0, b1, b2, width
|
|
mova m1, [pw_1]
|
|
%if ARCH_X86_64
|
|
mov widthd, widthd
|
|
%endif
|
|
.loop:
|
|
sub widthq, mmsize/2
|
|
mova m0, [b0q+2*widthq]
|
|
paddw m0, [b2q+2*widthq]
|
|
paddw m0, m1
|
|
psraw m0, 1
|
|
paddw m0, [b1q+2*widthq]
|
|
mova [b1q+2*widthq], m0
|
|
jg .loop
|
|
REP_RET
|
|
|
|
; void vertical_compose_dd97iH0(IDWTELEM *b0, IDWTELEM *b1, IDWTELEM *b2,
|
|
; IDWTELEM *b3, IDWTELEM *b4, int width)
|
|
cglobal vertical_compose_dd97iH0_%1, 6,6,5, b0, b1, b2, b3, b4, width
|
|
mova m3, [pw_8]
|
|
mova m4, [pw_1991]
|
|
%if ARCH_X86_64
|
|
mov widthd, widthd
|
|
%endif
|
|
.loop:
|
|
sub widthq, mmsize/2
|
|
mova m0, [b0q+2*widthq]
|
|
mova m1, [b1q+2*widthq]
|
|
COMPOSE_DD97iH0 [b2q+2*widthq], [b3q+2*widthq], [b4q+2*widthq]
|
|
mova [b2q+2*widthq], m1
|
|
jg .loop
|
|
REP_RET
|
|
|
|
; void vertical_compose_dd137iL0(IDWTELEM *b0, IDWTELEM *b1, IDWTELEM *b2,
|
|
; IDWTELEM *b3, IDWTELEM *b4, int width)
|
|
cglobal vertical_compose_dd137iL0_%1, 6,6,6, b0, b1, b2, b3, b4, width
|
|
mova m3, [pw_16]
|
|
mova m4, [pw_1991]
|
|
%if ARCH_X86_64
|
|
mov widthd, widthd
|
|
%endif
|
|
.loop:
|
|
sub widthq, mmsize/2
|
|
mova m0, [b0q+2*widthq]
|
|
mova m1, [b1q+2*widthq]
|
|
mova m5, [b2q+2*widthq]
|
|
paddw m0, [b4q+2*widthq]
|
|
paddw m1, [b3q+2*widthq]
|
|
psubw m0, m3
|
|
mova m2, m1
|
|
punpcklwd m1, m0
|
|
punpckhwd m2, m0
|
|
pmaddwd m1, m4
|
|
pmaddwd m2, m4
|
|
psrad m1, 5
|
|
psrad m2, 5
|
|
packssdw m1, m2
|
|
psubw m5, m1
|
|
mova [b2q+2*widthq], m5
|
|
jg .loop
|
|
REP_RET
|
|
|
|
; void vertical_compose_haar(IDWTELEM *b0, IDWTELEM *b1, int width)
|
|
cglobal vertical_compose_haar_%1, 3,4,3, b0, b1, width
|
|
mova m3, [pw_1]
|
|
%if ARCH_X86_64
|
|
mov widthd, widthd
|
|
%endif
|
|
.loop:
|
|
sub widthq, mmsize/2
|
|
mova m1, [b1q+2*widthq]
|
|
mova m0, [b0q+2*widthq]
|
|
mova m2, m1
|
|
paddw m1, m3
|
|
psraw m1, 1
|
|
psubw m0, m1
|
|
mova [b0q+2*widthq], m0
|
|
paddw m2, m0
|
|
mova [b1q+2*widthq], m2
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
; extend the left and right edges of the tmp array by %1 and %2 respectively
|
|
%macro EDGE_EXTENSION 3
|
|
mov %3, [tmpq]
|
|
%assign %%i 1
|
|
%rep %1
|
|
mov [tmpq-2*%%i], %3
|
|
%assign %%i %%i+1
|
|
%endrep
|
|
mov %3, [tmpq+2*w2q-2]
|
|
%assign %%i 0
|
|
%rep %2
|
|
mov [tmpq+2*w2q+2*%%i], %3
|
|
%assign %%i %%i+1
|
|
%endrep
|
|
%endmacro
|
|
|
|
|
|
%macro HAAR_HORIZONTAL 2
|
|
; void horizontal_compose_haari(IDWTELEM *b, IDWTELEM *tmp, int width)
|
|
cglobal horizontal_compose_haar%2i_%1, 3,6,4, b, tmp, w, x, w2, b_w2
|
|
mov w2d, wd
|
|
xor xq, xq
|
|
shr w2d, 1
|
|
lea b_w2q, [bq+wq]
|
|
mova m3, [pw_1]
|
|
.lowpass_loop:
|
|
movu m1, [b_w2q + 2*xq]
|
|
mova m0, [bq + 2*xq]
|
|
paddw m1, m3
|
|
psraw m1, 1
|
|
psubw m0, m1
|
|
mova [tmpq + 2*xq], m0
|
|
add xq, mmsize/2
|
|
cmp xq, w2q
|
|
jl .lowpass_loop
|
|
|
|
xor xq, xq
|
|
and w2q, ~(mmsize/2 - 1)
|
|
cmp w2q, mmsize/2
|
|
jl .end
|
|
|
|
.highpass_loop:
|
|
movu m1, [b_w2q + 2*xq]
|
|
mova m0, [tmpq + 2*xq]
|
|
paddw m1, m0
|
|
|
|
; shift and interleave
|
|
%if %2 == 1
|
|
paddw m0, m3
|
|
paddw m1, m3
|
|
psraw m0, 1
|
|
psraw m1, 1
|
|
%endif
|
|
mova m2, m0
|
|
punpcklwd m0, m1
|
|
punpckhwd m2, m1
|
|
mova [bq+4*xq], m0
|
|
mova [bq+4*xq+mmsize], m2
|
|
|
|
add xq, mmsize/2
|
|
cmp xq, w2q
|
|
jl .highpass_loop
|
|
.end:
|
|
REP_RET
|
|
%endmacro
|
|
|
|
|
|
INIT_XMM
|
|
; void horizontal_compose_dd97i(IDWTELEM *b, IDWTELEM *tmp, int width)
|
|
cglobal horizontal_compose_dd97i_ssse3, 3,6,8, b, tmp, w, x, w2, b_w2
|
|
mov w2d, wd
|
|
xor xd, xd
|
|
shr w2d, 1
|
|
lea b_w2q, [bq+wq]
|
|
movu m4, [bq+wq]
|
|
mova m7, [pw_2]
|
|
pslldq m4, 14
|
|
.lowpass_loop:
|
|
movu m1, [b_w2q + 2*xq]
|
|
mova m0, [bq + 2*xq]
|
|
mova m2, m1
|
|
palignr m1, m4, 14
|
|
mova m4, m2
|
|
COMPOSE_53iL0 m0, m1, m2, m7
|
|
mova [tmpq + 2*xq], m0
|
|
add xd, mmsize/2
|
|
cmp xd, w2d
|
|
jl .lowpass_loop
|
|
|
|
EDGE_EXTENSION 1, 2, xw
|
|
; leave the last up to 7 (sse) or 3 (mmx) values for C
|
|
xor xd, xd
|
|
and w2d, ~(mmsize/2 - 1)
|
|
cmp w2d, mmsize/2
|
|
jl .end
|
|
|
|
mova m7, [tmpq-mmsize]
|
|
mova m0, [tmpq]
|
|
mova m5, [pw_1]
|
|
mova m3, [pw_8]
|
|
mova m4, [pw_1991]
|
|
.highpass_loop:
|
|
mova m6, m0
|
|
palignr m0, m7, 14
|
|
mova m7, [tmpq + 2*xq + 16]
|
|
mova m1, m7
|
|
mova m2, m7
|
|
palignr m1, m6, 2
|
|
palignr m2, m6, 4
|
|
COMPOSE_DD97iH0 m0, m6, m2, [b_w2q + 2*xq]
|
|
mova m0, m7
|
|
mova m7, m6
|
|
|
|
; shift and interleave
|
|
paddw m6, m5
|
|
paddw m1, m5
|
|
psraw m6, 1
|
|
psraw m1, 1
|
|
mova m2, m6
|
|
punpcklwd m6, m1
|
|
punpckhwd m2, m1
|
|
mova [bq+4*xq], m6
|
|
mova [bq+4*xq+mmsize], m2
|
|
|
|
add xd, mmsize/2
|
|
cmp xd, w2d
|
|
jl .highpass_loop
|
|
.end:
|
|
REP_RET
|
|
|
|
|
|
%if ARCH_X86_64 == 0
|
|
INIT_MMX
|
|
COMPOSE_VERTICAL mmx
|
|
HAAR_HORIZONTAL mmx, 0
|
|
HAAR_HORIZONTAL mmx, 1
|
|
%endif
|
|
|
|
;;INIT_XMM
|
|
INIT_XMM
|
|
COMPOSE_VERTICAL sse2
|
|
HAAR_HORIZONTAL sse2, 0
|
|
HAAR_HORIZONTAL sse2, 1
|