mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-25 08:42:39 +00:00
6660bc034d
Clang supports the macro expansion counter (used for making unique labels within macro expansions), but not when targeting darwin. Convert uses of the counter into normal local labels, as used elsewhere. Since Xcode 9.3, the bundled clang supports altmacro and doesn't require using gas-preprocessor any longer. Signed-off-by: Martin Storsjö <martin@martin.st>
386 lines
10 KiB
ArmAsm
386 lines
10 KiB
ArmAsm
/*
|
|
* Copyright (c) 2014 Seppo Tomperi <seppo.tomperi@vtt.fi>
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
|
|
#include "libavutil/arm/asm.S"
|
|
#include "neon.S"
|
|
|
|
.macro hevc_loop_filter_chroma_start
|
|
ldr r12, [r2]
|
|
ldr r3, [r2, #4]
|
|
add r2, r3, r12
|
|
cmp r2, #0
|
|
it eq
|
|
bxeq lr
|
|
.endm
|
|
|
|
.macro hevc_loop_filter_chroma_body
|
|
vsubl.u8 q3, d4, d2
|
|
vsubl.u8 q11, d18, d19
|
|
vshl.i16 q3, #2
|
|
vadd.i16 q11, q3
|
|
vdup.16 d0, r12
|
|
vdup.16 d1, r3
|
|
vrshr.s16 q11, q11, #3
|
|
vneg.s16 q12, q0
|
|
vmovl.u8 q2, d4
|
|
vmin.s16 q11, q11, q0
|
|
vmax.s16 q11, q11, q12
|
|
vaddw.u8 q1, q11, d2
|
|
vsub.i16 q2, q11
|
|
vqmovun.s16 d2, q1
|
|
vqmovun.s16 d4, q2
|
|
.endm
|
|
|
|
.macro hevc_loop_filter_luma_start
|
|
ldr r12, [r3]
|
|
ldr r3, [r3, #4]
|
|
lsl r3, #16
|
|
orr r3, r12
|
|
cmp r3, #0
|
|
it eq
|
|
bxeq lr
|
|
lsr r3, #16
|
|
.endm
|
|
|
|
.macro hevc_loop_filter_luma_body
|
|
vmovl.u8 q8, d16
|
|
vmovl.u8 q9, d18
|
|
vmovl.u8 q10, d20
|
|
vmovl.u8 q11, d22
|
|
vmovl.u8 q12, d24
|
|
vmovl.u8 q13, d26
|
|
vmovl.u8 q14, d28
|
|
vmovl.u8 q15, d30
|
|
|
|
vadd.i16 q7, q9, q11
|
|
vadd.i16 q6, q14, q12
|
|
vsub.i16 q7, q10
|
|
vsub.i16 q6, q13
|
|
vabd.s16 q7, q7, q10
|
|
vabd.s16 q6, q6, q13
|
|
|
|
|
|
vdup.16 q0, r2
|
|
vmov q4, q7
|
|
vmov q5, q6
|
|
vdup.16 d4, r12
|
|
vtrn.16 q7, q4
|
|
vtrn.16 q6, q5
|
|
|
|
vshl.u64 q7, #32
|
|
vshr.u64 q4, #32
|
|
vshl.u64 q6, #32
|
|
vshr.u64 q5, #32
|
|
vshr.u64 q7, #32
|
|
vshr.u64 q6, #32
|
|
vshl.u64 q5, #32
|
|
vshl.u64 q4, #32
|
|
vorr q6, q5
|
|
vorr q7, q4
|
|
vdup.16 d5, r3
|
|
vadd.i16 q5, q7, q6
|
|
|
|
vmov q4, q5
|
|
vmov q3, q5
|
|
vtrn.32 q3, q4
|
|
|
|
vadd.i16 q4, q3
|
|
|
|
vshl.s16 q5, q5, #1
|
|
vcgt.s16 q3, q0, q4
|
|
|
|
vmovn.i16 d6, q3
|
|
vshr.s16 q1, q0, #2
|
|
vmovn.i16 d6, q3
|
|
vcgt.s16 q5, q1, q5
|
|
vmov r7, s12
|
|
cmp r7, #0
|
|
beq bypasswrite
|
|
|
|
vpadd.i32 d0, d14, d12
|
|
vpadd.i32 d1, d15, d13
|
|
vmov q4, q2
|
|
vshl.s16 q2, #2
|
|
vshr.s16 q1, q1, #1
|
|
vrhadd.s16 q2, q4
|
|
|
|
vabd.s16 q7, q8, q11
|
|
vaba.s16 q7, q15, q12
|
|
|
|
vmovn.i32 d0, q0
|
|
vmov r5, r6, s0, s1
|
|
vcgt.s16 q6, q1, q7
|
|
vand q5, q5, q6
|
|
vabd.s16 q7, q11, q12
|
|
vcgt.s16 q6, q2, q7
|
|
vand q5, q5, q6
|
|
|
|
vmov q2, q5
|
|
vtrn.s16 q5, q2
|
|
vshr.u64 q2, #32
|
|
vshl.u64 q5, #32
|
|
vshl.u64 q2, #32
|
|
vshr.u64 q5, #32
|
|
vorr q5, q2
|
|
|
|
vmov q2, q5
|
|
vshl.i16 q7, q4, #1
|
|
vtrn.32 q2, q5
|
|
vand q5, q2
|
|
vneg.s16 q6, q7
|
|
vmovn.i16 d4, q5
|
|
vmovn.i16 d4, q2
|
|
vmov r8, s8
|
|
|
|
and r9, r8, r7
|
|
cmp r9, #0
|
|
beq 1f
|
|
|
|
vadd.i16 q2, q11, q12
|
|
vadd.i16 q4, q9, q8
|
|
vadd.i16 q1, q2, q10
|
|
vdup.16 d10, r9
|
|
vadd.i16 q0, q1, q9
|
|
vshl.i16 q4, #1
|
|
lsr r9, #16
|
|
vadd.i16 q1, q0
|
|
vrshr.s16 q3, q0, #2
|
|
vadd.i16 q1, q13
|
|
vadd.i16 q4, q0
|
|
vsub.i16 q3, q10
|
|
vrshr.s16 q1, #3
|
|
vrshr.s16 q4, #3
|
|
vmax.s16 q3, q6
|
|
vsub.i16 q1, q11
|
|
vsub.i16 q4, q9
|
|
vmin.s16 q3, q7
|
|
vmax.s16 q4, q6
|
|
vmax.s16 q1, q6
|
|
vadd.i16 q3, q10
|
|
vmin.s16 q4, q7
|
|
vmin.s16 q1, q7
|
|
vdup.16 d11, r9
|
|
vadd.i16 q4, q9
|
|
vadd.i16 q1, q11
|
|
vbit q9, q4, q5
|
|
vadd.i16 q4, q2, q13
|
|
vbit q11, q1, q5
|
|
vadd.i16 q0, q4, q14
|
|
vadd.i16 q2, q15, q14
|
|
vadd.i16 q4, q0
|
|
|
|
vshl.i16 q2, #1
|
|
vadd.i16 q4, q10
|
|
vbit q10, q3, q5
|
|
vrshr.s16 q4, #3
|
|
vadd.i16 q2, q0
|
|
vrshr.s16 q3, q0, #2
|
|
vsub.i16 q4, q12
|
|
vrshr.s16 q2, #3
|
|
vsub.i16 q3, q13
|
|
vmax.s16 q4, q6
|
|
vsub.i16 q2, q14
|
|
vmax.s16 q3, q6
|
|
vmin.s16 q4, q7
|
|
vmax.s16 q2, q6
|
|
vmin.s16 q3, q7
|
|
vadd.i16 q4, q12
|
|
vmin.s16 q2, q7
|
|
vadd.i16 q3, q13
|
|
vbit q12, q4, q5
|
|
vadd.i16 q2, q14
|
|
vbit q13, q3, q5
|
|
vbit q14, q2, q5
|
|
|
|
1:
|
|
mvn r8, r8
|
|
and r9, r8, r7
|
|
cmp r9, #0
|
|
beq 2f
|
|
|
|
vdup.16 q4, r2
|
|
|
|
vdup.16 d10, r9
|
|
lsr r9, #16
|
|
vmov q1, q4
|
|
vdup.16 d11, r9
|
|
vshr.s16 q1, #1
|
|
vsub.i16 q2, q12, q11
|
|
vadd.i16 q4, q1
|
|
vshl.s16 q0, q2, #3
|
|
vshr.s16 q4, #3
|
|
vadd.i16 q2, q0
|
|
vsub.i16 q0, q13, q10
|
|
vsub.i16 q2, q0
|
|
vshl.i16 q0, q0, #1
|
|
vsub.i16 q2, q0
|
|
vshl.s16 q1, q7, 2
|
|
vrshr.s16 q2, q2, #4
|
|
vadd.i16 q1, q7
|
|
vabs.s16 q3, q2
|
|
vshr.s16 q6, q6, #1
|
|
vcgt.s16 q1, q1, q3
|
|
vand q5, q1
|
|
vshr.s16 q7, q7, #1
|
|
vmax.s16 q2, q2, q6
|
|
vmin.s16 q2, q2, q7
|
|
|
|
vshr.s16 q7, q7, #1
|
|
vrhadd.s16 q3, q9, q11
|
|
vneg.s16 q6, q7
|
|
vsub.s16 q3, q10
|
|
vdup.16 d2, r5
|
|
vhadd.s16 q3, q2
|
|
vdup.16 d3, r6
|
|
vmax.s16 q3, q3, q6
|
|
vcgt.s16 q1, q4, q1
|
|
vmin.s16 q3, q3, q7
|
|
vand q1, q5
|
|
vadd.i16 q3, q10
|
|
lsr r5, #16
|
|
lsr r6, #16
|
|
vbit q10, q3, q1
|
|
|
|
vrhadd.s16 q3, q14, q12
|
|
vdup.16 d2, r5
|
|
vsub.s16 q3, q13
|
|
vdup.16 d3, r6
|
|
vhsub.s16 q3, q2
|
|
vcgt.s16 q1, q4, q1
|
|
vmax.s16 q3, q3, q6
|
|
vand q1, q5
|
|
vmin.s16 q3, q3, q7
|
|
vadd.i16 q3, q13
|
|
vbit q13, q3, q1
|
|
vadd.i16 q0, q11, q2
|
|
vsub.i16 q4, q12, q2
|
|
vbit q11, q0, q5
|
|
vbit q12, q4, q5
|
|
|
|
2:
|
|
vqmovun.s16 d16, q8
|
|
vqmovun.s16 d18, q9
|
|
vqmovun.s16 d20, q10
|
|
vqmovun.s16 d22, q11
|
|
vqmovun.s16 d24, q12
|
|
vqmovun.s16 d26, q13
|
|
vqmovun.s16 d28, q14
|
|
vqmovun.s16 d30, q15
|
|
.endm
|
|
|
|
function ff_hevc_v_loop_filter_luma_neon, export=1
|
|
hevc_loop_filter_luma_start
|
|
push {r5-r11}
|
|
vpush {d8-d15}
|
|
sub r0, #4
|
|
vld1.8 {d16}, [r0], r1
|
|
vld1.8 {d18}, [r0], r1
|
|
vld1.8 {d20}, [r0], r1
|
|
vld1.8 {d22}, [r0], r1
|
|
vld1.8 {d24}, [r0], r1
|
|
vld1.8 {d26}, [r0], r1
|
|
vld1.8 {d28}, [r0], r1
|
|
vld1.8 {d30}, [r0], r1
|
|
sub r0, r0, r1, lsl #3
|
|
transpose_8x8 d16, d18, d20, d22, d24, d26, d28, d30
|
|
hevc_loop_filter_luma_body
|
|
transpose_8x8 d16, d18, d20, d22, d24, d26, d28, d30
|
|
vst1.8 {d16}, [r0], r1
|
|
vst1.8 {d18}, [r0], r1
|
|
vst1.8 {d20}, [r0], r1
|
|
vst1.8 {d22}, [r0], r1
|
|
vst1.8 {d24}, [r0], r1
|
|
vst1.8 {d26}, [r0], r1
|
|
vst1.8 {d28}, [r0], r1
|
|
vst1.8 {d30}, [r0]
|
|
vpop {d8-d15}
|
|
pop {r5-r11}
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_h_loop_filter_luma_neon, export=1
|
|
hevc_loop_filter_luma_start
|
|
push {r5-r11}
|
|
vpush {d8-d15}
|
|
sub r0, r0, r1, lsl #2
|
|
vld1.8 {d16}, [r0], r1
|
|
vld1.8 {d18}, [r0], r1
|
|
vld1.8 {d20}, [r0], r1
|
|
vld1.8 {d22}, [r0], r1
|
|
vld1.8 {d24}, [r0], r1
|
|
vld1.8 {d26}, [r0], r1
|
|
vld1.8 {d28}, [r0], r1
|
|
vld1.8 {d30}, [r0], r1
|
|
sub r0, r0, r1, lsl #3
|
|
add r0, r1
|
|
hevc_loop_filter_luma_body
|
|
vst1.8 {d18}, [r0], r1
|
|
vst1.8 {d20}, [r0], r1
|
|
vst1.8 {d22}, [r0], r1
|
|
vst1.8 {d24}, [r0], r1
|
|
vst1.8 {d26}, [r0], r1
|
|
vst1.8 {d28}, [r0]
|
|
bypasswrite:
|
|
vpop {d8-d15}
|
|
pop {r5-r11}
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_v_loop_filter_chroma_neon, export=1
|
|
hevc_loop_filter_chroma_start
|
|
sub r0, #4
|
|
vld1.8 {d16}, [r0], r1
|
|
vld1.8 {d17}, [r0], r1
|
|
vld1.8 {d18}, [r0], r1
|
|
vld1.8 {d2}, [r0], r1
|
|
vld1.8 {d4}, [r0], r1
|
|
vld1.8 {d19}, [r0], r1
|
|
vld1.8 {d20}, [r0], r1
|
|
vld1.8 {d21}, [r0], r1
|
|
sub r0, r0, r1, lsl #3
|
|
transpose_8x8 d16, d17, d18, d2, d4, d19, d20, d21
|
|
hevc_loop_filter_chroma_body
|
|
transpose_8x8 d16, d17, d18, d2, d4, d19, d20, d21
|
|
vst1.8 {d16}, [r0], r1
|
|
vst1.8 {d17}, [r0], r1
|
|
vst1.8 {d18}, [r0], r1
|
|
vst1.8 {d2}, [r0], r1
|
|
vst1.8 {d4}, [r0], r1
|
|
vst1.8 {d19}, [r0], r1
|
|
vst1.8 {d20}, [r0], r1
|
|
vst1.8 {d21}, [r0]
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_h_loop_filter_chroma_neon, export=1
|
|
hevc_loop_filter_chroma_start
|
|
sub r0, r0, r1, lsl #1
|
|
vld1.8 {d18}, [r0], r1
|
|
vld1.8 {d2}, [r0], r1
|
|
vld1.8 {d4}, [r0], r1
|
|
vld1.8 {d19}, [r0]
|
|
sub r0, r0, r1, lsl #1
|
|
hevc_loop_filter_chroma_body
|
|
vst1.8 {d2}, [r0], r1
|
|
vst1.8 {d4}, [r0]
|
|
bx lr
|
|
endfunc
|