ffmpeg/libavcodec/aarch64/vc1dsp_neon.S
Ben Avison 6eee650289 avcodec/vc1: Arm 64-bit NEON unescape fast path
checkasm benchmarks on 1.5 GHz Cortex-A72 are as follows.

vc1dsp.vc1_unescape_buffer_c: 655617.7
vc1dsp.vc1_unescape_buffer_neon: 118237.0

Signed-off-by: Ben Avison <bavison@riscosopen.org>
Signed-off-by: Martin Storsjö <martin@martin.st>
2022-04-01 10:03:34 +03:00

1547 lines
92 KiB
ArmAsm

/*
* VC1 AArch64 NEON optimisations
*
* Copyright (c) 2022 Ben Avison <bavison@riscosopen.org>
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "libavutil/aarch64/asm.S"
// VC-1 8x8 inverse transform
// On entry:
// x0 -> array of 16-bit inverse transform coefficients, in column-major order
// On exit:
// array at x0 updated to hold transformed block; also now held in row-major order
function ff_vc1_inv_trans_8x8_neon, export=1
ld1 {v1.16b, v2.16b}, [x0], #32
ld1 {v3.16b, v4.16b}, [x0], #32
ld1 {v5.16b, v6.16b}, [x0], #32
shl v1.8h, v1.8h, #2 // 8/2 * src[0]
sub x1, x0, #3*32
ld1 {v16.16b, v17.16b}, [x0]
shl v7.8h, v2.8h, #4 // 16 * src[8]
shl v18.8h, v2.8h, #2 // 4 * src[8]
shl v19.8h, v4.8h, #4 // 16 * src[24]
ldr d0, .Lcoeffs_it8
shl v5.8h, v5.8h, #2 // 8/2 * src[32]
shl v20.8h, v6.8h, #4 // 16 * src[40]
shl v21.8h, v6.8h, #2 // 4 * src[40]
shl v22.8h, v17.8h, #4 // 16 * src[56]
ssra v20.8h, v19.8h, #2 // 4 * src[24] + 16 * src[40]
mul v23.8h, v3.8h, v0.h[0] // 6/2 * src[16]
sub v19.8h, v19.8h, v21.8h // 16 * src[24] - 4 * src[40]
ssra v7.8h, v22.8h, #2 // 16 * src[8] + 4 * src[56]
sub v18.8h, v22.8h, v18.8h // - 4 * src[8] + 16 * src[56]
shl v3.8h, v3.8h, #3 // 16/2 * src[16]
mls v20.8h, v2.8h, v0.h[2] // - 15 * src[8] + 4 * src[24] + 16 * src[40]
ssra v1.8h, v1.8h, #1 // 12/2 * src[0]
ssra v5.8h, v5.8h, #1 // 12/2 * src[32]
mla v7.8h, v4.8h, v0.h[2] // 16 * src[8] + 15 * src[24] + 4 * src[56]
shl v21.8h, v16.8h, #3 // 16/2 * src[48]
mls v19.8h, v2.8h, v0.h[1] // - 9 * src[8] + 16 * src[24] - 4 * src[40]
sub v2.8h, v23.8h, v21.8h // t4/2 = 6/2 * src[16] - 16/2 * src[48]
mla v18.8h, v4.8h, v0.h[1] // - 4 * src[8] + 9 * src[24] + 16 * src[56]
add v4.8h, v1.8h, v5.8h // t1/2 = 12/2 * src[0] + 12/2 * src[32]
sub v1.8h, v1.8h, v5.8h // t2/2 = 12/2 * src[0] - 12/2 * src[32]
mla v3.8h, v16.8h, v0.h[0] // t3/2 = 16/2 * src[16] + 6/2 * src[48]
mla v7.8h, v6.8h, v0.h[1] // t1 = 16 * src[8] + 15 * src[24] + 9 * src[40] + 4 * src[56]
add v5.8h, v1.8h, v2.8h // t6/2 = t2/2 + t4/2
sub v16.8h, v1.8h, v2.8h // t7/2 = t2/2 - t4/2
mla v20.8h, v17.8h, v0.h[1] // -t2 = - 15 * src[8] + 4 * src[24] + 16 * src[40] + 9 * src[56]
add v21.8h, v1.8h, v2.8h // t6/2 = t2/2 + t4/2
add v22.8h, v4.8h, v3.8h // t5/2 = t1/2 + t3/2
mls v19.8h, v17.8h, v0.h[2] // -t3 = - 9 * src[8] + 16 * src[24] - 4 * src[40] - 15 * src[56]
sub v17.8h, v4.8h, v3.8h // t8/2 = t1/2 - t3/2
add v23.8h, v4.8h, v3.8h // t5/2 = t1/2 + t3/2
mls v18.8h, v6.8h, v0.h[2] // -t4 = - 4 * src[8] + 9 * src[24] - 15 * src[40] + 16 * src[56]
sub v1.8h, v1.8h, v2.8h // t7/2 = t2/2 - t4/2
sub v2.8h, v4.8h, v3.8h // t8/2 = t1/2 - t3/2
neg v3.8h, v7.8h // -t1
neg v4.8h, v20.8h // +t2
neg v6.8h, v19.8h // +t3
ssra v22.8h, v7.8h, #1 // (t5 + t1) >> 1
ssra v1.8h, v19.8h, #1 // (t7 - t3) >> 1
neg v7.8h, v18.8h // +t4
ssra v5.8h, v4.8h, #1 // (t6 + t2) >> 1
ssra v16.8h, v6.8h, #1 // (t7 + t3) >> 1
ssra v2.8h, v18.8h, #1 // (t8 - t4) >> 1
ssra v17.8h, v7.8h, #1 // (t8 + t4) >> 1
ssra v21.8h, v20.8h, #1 // (t6 - t2) >> 1
ssra v23.8h, v3.8h, #1 // (t5 - t1) >> 1
srshr v3.8h, v22.8h, #2 // (t5 + t1 + 4) >> 3
srshr v4.8h, v5.8h, #2 // (t6 + t2 + 4) >> 3
srshr v5.8h, v16.8h, #2 // (t7 + t3 + 4) >> 3
srshr v6.8h, v17.8h, #2 // (t8 + t4 + 4) >> 3
srshr v2.8h, v2.8h, #2 // (t8 - t4 + 4) >> 3
srshr v1.8h, v1.8h, #2 // (t7 - t3 + 4) >> 3
srshr v7.8h, v21.8h, #2 // (t6 - t2 + 4) >> 3
srshr v16.8h, v23.8h, #2 // (t5 - t1 + 4) >> 3
trn2 v17.8h, v3.8h, v4.8h
trn2 v18.8h, v5.8h, v6.8h
trn2 v19.8h, v2.8h, v1.8h
trn2 v20.8h, v7.8h, v16.8h
trn1 v21.4s, v17.4s, v18.4s
trn2 v17.4s, v17.4s, v18.4s
trn1 v18.4s, v19.4s, v20.4s
trn2 v19.4s, v19.4s, v20.4s
trn1 v3.8h, v3.8h, v4.8h
trn2 v4.2d, v21.2d, v18.2d
trn1 v20.2d, v17.2d, v19.2d
trn1 v5.8h, v5.8h, v6.8h
trn1 v1.8h, v2.8h, v1.8h
trn1 v2.8h, v7.8h, v16.8h
trn1 v6.2d, v21.2d, v18.2d
trn2 v7.2d, v17.2d, v19.2d
shl v16.8h, v20.8h, #4 // 16 * src[24]
shl v17.8h, v4.8h, #4 // 16 * src[40]
trn1 v18.4s, v3.4s, v5.4s
trn1 v19.4s, v1.4s, v2.4s
shl v21.8h, v7.8h, #4 // 16 * src[56]
shl v22.8h, v6.8h, #2 // 4 * src[8]
shl v23.8h, v4.8h, #2 // 4 * src[40]
trn2 v3.4s, v3.4s, v5.4s
trn2 v1.4s, v1.4s, v2.4s
shl v2.8h, v6.8h, #4 // 16 * src[8]
sub v5.8h, v16.8h, v23.8h // 16 * src[24] - 4 * src[40]
ssra v17.8h, v16.8h, #2 // 4 * src[24] + 16 * src[40]
sub v16.8h, v21.8h, v22.8h // - 4 * src[8] + 16 * src[56]
trn1 v22.2d, v18.2d, v19.2d
trn2 v18.2d, v18.2d, v19.2d
trn1 v19.2d, v3.2d, v1.2d
ssra v2.8h, v21.8h, #2 // 16 * src[8] + 4 * src[56]
mls v17.8h, v6.8h, v0.h[2] // - 15 * src[8] + 4 * src[24] + 16 * src[40]
shl v21.8h, v22.8h, #2 // 8/2 * src[0]
shl v18.8h, v18.8h, #2 // 8/2 * src[32]
mls v5.8h, v6.8h, v0.h[1] // - 9 * src[8] + 16 * src[24] - 4 * src[40]
shl v6.8h, v19.8h, #3 // 16/2 * src[16]
trn2 v1.2d, v3.2d, v1.2d
mla v16.8h, v20.8h, v0.h[1] // - 4 * src[8] + 9 * src[24] + 16 * src[56]
ssra v21.8h, v21.8h, #1 // 12/2 * src[0]
ssra v18.8h, v18.8h, #1 // 12/2 * src[32]
mul v3.8h, v19.8h, v0.h[0] // 6/2 * src[16]
shl v19.8h, v1.8h, #3 // 16/2 * src[48]
mla v2.8h, v20.8h, v0.h[2] // 16 * src[8] + 15 * src[24] + 4 * src[56]
add v20.8h, v21.8h, v18.8h // t1/2 = 12/2 * src[0] + 12/2 * src[32]
mla v6.8h, v1.8h, v0.h[0] // t3/2 = 16/2 * src[16] + 6/2 * src[48]
sub v1.8h, v21.8h, v18.8h // t2/2 = 12/2 * src[0] - 12/2 * src[32]
sub v3.8h, v3.8h, v19.8h // t4/2 = 6/2 * src[16] - 16/2 * src[48]
mla v17.8h, v7.8h, v0.h[1] // -t2 = - 15 * src[8] + 4 * src[24] + 16 * src[40] + 9 * src[56]
mls v5.8h, v7.8h, v0.h[2] // -t3 = - 9 * src[8] + 16 * src[24] - 4 * src[40] - 15 * src[56]
add v7.8h, v1.8h, v3.8h // t6/2 = t2/2 + t4/2
add v18.8h, v20.8h, v6.8h // t5/2 = t1/2 + t3/2
mls v16.8h, v4.8h, v0.h[2] // -t4 = - 4 * src[8] + 9 * src[24] - 15 * src[40] + 16 * src[56]
sub v19.8h, v1.8h, v3.8h // t7/2 = t2/2 - t4/2
neg v21.8h, v17.8h // +t2
mla v2.8h, v4.8h, v0.h[1] // t1 = 16 * src[8] + 15 * src[24] + 9 * src[40] + 4 * src[56]
sub v0.8h, v20.8h, v6.8h // t8/2 = t1/2 - t3/2
neg v4.8h, v5.8h // +t3
sub v22.8h, v1.8h, v3.8h // t7/2 = t2/2 - t4/2
sub v23.8h, v20.8h, v6.8h // t8/2 = t1/2 - t3/2
neg v24.8h, v16.8h // +t4
add v6.8h, v20.8h, v6.8h // t5/2 = t1/2 + t3/2
add v1.8h, v1.8h, v3.8h // t6/2 = t2/2 + t4/2
ssra v7.8h, v21.8h, #1 // (t6 + t2) >> 1
neg v3.8h, v2.8h // -t1
ssra v18.8h, v2.8h, #1 // (t5 + t1) >> 1
ssra v19.8h, v4.8h, #1 // (t7 + t3) >> 1
ssra v0.8h, v24.8h, #1 // (t8 + t4) >> 1
srsra v23.8h, v16.8h, #1 // (t8 - t4 + 1) >> 1
srsra v22.8h, v5.8h, #1 // (t7 - t3 + 1) >> 1
srsra v1.8h, v17.8h, #1 // (t6 - t2 + 1) >> 1
srsra v6.8h, v3.8h, #1 // (t5 - t1 + 1) >> 1
srshr v2.8h, v18.8h, #6 // (t5 + t1 + 64) >> 7
srshr v3.8h, v7.8h, #6 // (t6 + t2 + 64) >> 7
srshr v4.8h, v19.8h, #6 // (t7 + t3 + 64) >> 7
srshr v5.8h, v0.8h, #6 // (t8 + t4 + 64) >> 7
srshr v16.8h, v23.8h, #6 // (t8 - t4 + 65) >> 7
srshr v17.8h, v22.8h, #6 // (t7 - t3 + 65) >> 7
st1 {v2.16b, v3.16b}, [x1], #32
srshr v0.8h, v1.8h, #6 // (t6 - t2 + 65) >> 7
srshr v1.8h, v6.8h, #6 // (t5 - t1 + 65) >> 7
st1 {v4.16b, v5.16b}, [x1], #32
st1 {v16.16b, v17.16b}, [x1], #32
st1 {v0.16b, v1.16b}, [x1]
ret
endfunc
// VC-1 8x4 inverse transform
// On entry:
// x0 -> array of 8-bit samples, in row-major order
// x1 = row stride for 8-bit sample array
// x2 -> array of 16-bit inverse transform coefficients, in row-major order
// On exit:
// array at x0 updated by saturated addition of (narrowed) transformed block
function ff_vc1_inv_trans_8x4_neon, export=1
ld1 {v1.8b, v2.8b, v3.8b, v4.8b}, [x2], #32
mov x3, x0
ld1 {v16.8b, v17.8b, v18.8b, v19.8b}, [x2]
ldr q0, .Lcoeffs_it8 // includes 4-point coefficients in upper half of vector
ld1 {v5.8b}, [x0], x1
trn2 v6.4h, v1.4h, v3.4h
trn2 v7.4h, v2.4h, v4.4h
trn1 v1.4h, v1.4h, v3.4h
trn1 v2.4h, v2.4h, v4.4h
trn2 v3.4h, v16.4h, v18.4h
trn2 v4.4h, v17.4h, v19.4h
trn1 v16.4h, v16.4h, v18.4h
trn1 v17.4h, v17.4h, v19.4h
ld1 {v18.8b}, [x0], x1
trn1 v19.2s, v6.2s, v3.2s
trn2 v3.2s, v6.2s, v3.2s
trn1 v6.2s, v7.2s, v4.2s
trn2 v4.2s, v7.2s, v4.2s
trn1 v7.2s, v1.2s, v16.2s
trn1 v20.2s, v2.2s, v17.2s
shl v21.4h, v19.4h, #4 // 16 * src[1]
trn2 v1.2s, v1.2s, v16.2s
shl v16.4h, v3.4h, #4 // 16 * src[3]
trn2 v2.2s, v2.2s, v17.2s
shl v17.4h, v6.4h, #4 // 16 * src[5]
ld1 {v22.8b}, [x0], x1
shl v23.4h, v4.4h, #4 // 16 * src[7]
mul v24.4h, v1.4h, v0.h[0] // 6/2 * src[2]
ld1 {v25.8b}, [x0]
shl v26.4h, v19.4h, #2 // 4 * src[1]
shl v27.4h, v6.4h, #2 // 4 * src[5]
ssra v21.4h, v23.4h, #2 // 16 * src[1] + 4 * src[7]
ssra v17.4h, v16.4h, #2 // 4 * src[3] + 16 * src[5]
sub v23.4h, v23.4h, v26.4h // - 4 * src[1] + 16 * src[7]
sub v16.4h, v16.4h, v27.4h // 16 * src[3] - 4 * src[5]
shl v7.4h, v7.4h, #2 // 8/2 * src[0]
shl v20.4h, v20.4h, #2 // 8/2 * src[4]
mla v21.4h, v3.4h, v0.h[2] // 16 * src[1] + 15 * src[3] + 4 * src[7]
shl v1.4h, v1.4h, #3 // 16/2 * src[2]
mls v17.4h, v19.4h, v0.h[2] // - 15 * src[1] + 4 * src[3] + 16 * src[5]
ssra v7.4h, v7.4h, #1 // 12/2 * src[0]
mls v16.4h, v19.4h, v0.h[1] // - 9 * src[1] + 16 * src[3] - 4 * src[5]
ssra v20.4h, v20.4h, #1 // 12/2 * src[4]
mla v23.4h, v3.4h, v0.h[1] // - 4 * src[1] + 9 * src[3] + 16 * src[7]
shl v3.4h, v2.4h, #3 // 16/2 * src[6]
mla v1.4h, v2.4h, v0.h[0] // t3/2 = 16/2 * src[2] + 6/2 * src[6]
mla v21.4h, v6.4h, v0.h[1] // t1 = 16 * src[1] + 15 * src[3] + 9 * src[5] + 4 * src[7]
mla v17.4h, v4.4h, v0.h[1] // -t2 = - 15 * src[1] + 4 * src[3] + 16 * src[5] + 9 * src[7]
sub v2.4h, v24.4h, v3.4h // t4/2 = 6/2 * src[2] - 16/2 * src[6]
mls v16.4h, v4.4h, v0.h[2] // -t3 = - 9 * src[1] + 16 * src[3] - 4 * src[5] - 15 * src[7]
add v3.4h, v7.4h, v20.4h // t1/2 = 12/2 * src[0] + 12/2 * src[4]
mls v23.4h, v6.4h, v0.h[2] // -t4 = - 4 * src[1] + 9 * src[3] - 15 * src[5] + 16 * src[7]
sub v4.4h, v7.4h, v20.4h // t2/2 = 12/2 * src[0] - 12/2 * src[4]
neg v6.4h, v21.4h // -t1
add v7.4h, v3.4h, v1.4h // t5/2 = t1/2 + t3/2
sub v19.4h, v3.4h, v1.4h // t8/2 = t1/2 - t3/2
add v20.4h, v4.4h, v2.4h // t6/2 = t2/2 + t4/2
sub v24.4h, v4.4h, v2.4h // t7/2 = t2/2 - t4/2
add v26.4h, v3.4h, v1.4h // t5/2 = t1/2 + t3/2
add v27.4h, v4.4h, v2.4h // t6/2 = t2/2 + t4/2
sub v2.4h, v4.4h, v2.4h // t7/2 = t2/2 - t4/2
sub v1.4h, v3.4h, v1.4h // t8/2 = t1/2 - t3/2
neg v3.4h, v17.4h // +t2
neg v4.4h, v16.4h // +t3
neg v28.4h, v23.4h // +t4
ssra v7.4h, v21.4h, #1 // (t5 + t1) >> 1
ssra v1.4h, v23.4h, #1 // (t8 - t4) >> 1
ssra v20.4h, v3.4h, #1 // (t6 + t2) >> 1
ssra v24.4h, v4.4h, #1 // (t7 + t3) >> 1
ssra v19.4h, v28.4h, #1 // (t8 + t4) >> 1
ssra v2.4h, v16.4h, #1 // (t7 - t3) >> 1
ssra v27.4h, v17.4h, #1 // (t6 - t2) >> 1
ssra v26.4h, v6.4h, #1 // (t5 - t1) >> 1
trn1 v1.2d, v7.2d, v1.2d
trn1 v2.2d, v20.2d, v2.2d
trn1 v3.2d, v24.2d, v27.2d
trn1 v4.2d, v19.2d, v26.2d
srshr v1.8h, v1.8h, #2 // (t5 + t1 + 4) >> 3, (t8 - t4 + 4) >> 3
srshr v2.8h, v2.8h, #2 // (t6 + t2 + 4) >> 3, (t7 - t3 + 4) >> 3
srshr v3.8h, v3.8h, #2 // (t7 + t3 + 4) >> 3, (t6 - t2 + 4) >> 3
srshr v4.8h, v4.8h, #2 // (t8 + t4 + 4) >> 3, (t5 - t1 + 4) >> 3
trn2 v6.8h, v1.8h, v2.8h
trn1 v1.8h, v1.8h, v2.8h
trn2 v2.8h, v3.8h, v4.8h
trn1 v3.8h, v3.8h, v4.8h
trn2 v4.4s, v6.4s, v2.4s
trn1 v7.4s, v1.4s, v3.4s
trn2 v1.4s, v1.4s, v3.4s
mul v3.8h, v4.8h, v0.h[5] // 22/2 * src[24]
trn1 v2.4s, v6.4s, v2.4s
mul v4.8h, v4.8h, v0.h[4] // 10/2 * src[24]
mul v6.8h, v7.8h, v0.h[6] // 17 * src[0]
mul v1.8h, v1.8h, v0.h[6] // 17 * src[16]
mls v3.8h, v2.8h, v0.h[4] // t4/2 = - 10/2 * src[8] + 22/2 * src[24]
mla v4.8h, v2.8h, v0.h[5] // t3/2 = 22/2 * src[8] + 10/2 * src[24]
add v0.8h, v6.8h, v1.8h // t1 = 17 * src[0] + 17 * src[16]
sub v1.8h, v6.8h, v1.8h // t2 = 17 * src[0] - 17 * src[16]
neg v2.8h, v3.8h // -t4/2
neg v6.8h, v4.8h // -t3/2
ssra v4.8h, v0.8h, #1 // (t1 + t3) >> 1
ssra v2.8h, v1.8h, #1 // (t2 - t4) >> 1
ssra v3.8h, v1.8h, #1 // (t2 + t4) >> 1
ssra v6.8h, v0.8h, #1 // (t1 - t3) >> 1
srshr v0.8h, v4.8h, #6 // (t1 + t3 + 64) >> 7
srshr v1.8h, v2.8h, #6 // (t2 - t4 + 64) >> 7
srshr v2.8h, v3.8h, #6 // (t2 + t4 + 64) >> 7
srshr v3.8h, v6.8h, #6 // (t1 - t3 + 64) >> 7
uaddw v0.8h, v0.8h, v5.8b
uaddw v1.8h, v1.8h, v18.8b
uaddw v2.8h, v2.8h, v22.8b
uaddw v3.8h, v3.8h, v25.8b
sqxtun v0.8b, v0.8h
sqxtun v1.8b, v1.8h
sqxtun v2.8b, v2.8h
sqxtun v3.8b, v3.8h
st1 {v0.8b}, [x3], x1
st1 {v1.8b}, [x3], x1
st1 {v2.8b}, [x3], x1
st1 {v3.8b}, [x3]
ret
endfunc
// VC-1 4x8 inverse transform
// On entry:
// x0 -> array of 8-bit samples, in row-major order
// x1 = row stride for 8-bit sample array
// x2 -> array of 16-bit inverse transform coefficients, in row-major order (row stride is 8 coefficients)
// On exit:
// array at x0 updated by saturated addition of (narrowed) transformed block
function ff_vc1_inv_trans_4x8_neon, export=1
mov x3, #16
ldr q0, .Lcoeffs_it8 // includes 4-point coefficients in upper half of vector
mov x4, x0
ld1 {v1.d}[0], [x2], x3 // 00 01 02 03
ld1 {v2.d}[0], [x2], x3 // 10 11 12 13
ld1 {v3.d}[0], [x2], x3 // 20 21 22 23
ld1 {v4.d}[0], [x2], x3 // 30 31 32 33
ld1 {v1.d}[1], [x2], x3 // 40 41 42 43
ld1 {v2.d}[1], [x2], x3 // 50 51 52 53
ld1 {v3.d}[1], [x2], x3 // 60 61 62 63
ld1 {v4.d}[1], [x2] // 70 71 72 73
ld1 {v5.s}[0], [x0], x1
ld1 {v6.s}[0], [x0], x1
ld1 {v7.s}[0], [x0], x1
trn2 v16.8h, v1.8h, v2.8h // 01 11 03 13 41 51 43 53
trn1 v1.8h, v1.8h, v2.8h // 00 10 02 12 40 50 42 52
trn2 v2.8h, v3.8h, v4.8h // 21 31 23 33 61 71 63 73
trn1 v3.8h, v3.8h, v4.8h // 20 30 22 32 60 70 62 72
ld1 {v4.s}[0], [x0], x1
trn2 v17.4s, v16.4s, v2.4s // 03 13 23 33 43 53 63 73
trn1 v18.4s, v1.4s, v3.4s // 00 10 20 30 40 50 60 70
trn1 v2.4s, v16.4s, v2.4s // 01 11 21 31 41 51 61 71
mul v16.8h, v17.8h, v0.h[4] // 10/2 * src[3]
ld1 {v5.s}[1], [x0], x1
mul v17.8h, v17.8h, v0.h[5] // 22/2 * src[3]
ld1 {v6.s}[1], [x0], x1
trn2 v1.4s, v1.4s, v3.4s // 02 12 22 32 42 52 62 72
mul v3.8h, v18.8h, v0.h[6] // 17 * src[0]
ld1 {v7.s}[1], [x0], x1
mul v1.8h, v1.8h, v0.h[6] // 17 * src[2]
ld1 {v4.s}[1], [x0]
mla v16.8h, v2.8h, v0.h[5] // t3/2 = 22/2 * src[1] + 10/2 * src[3]
mls v17.8h, v2.8h, v0.h[4] // t4/2 = - 10/2 * src[1] + 22/2 * src[3]
add v2.8h, v3.8h, v1.8h // t1 = 17 * src[0] + 17 * src[2]
sub v1.8h, v3.8h, v1.8h // t2 = 17 * src[0] - 17 * src[2]
neg v3.8h, v16.8h // -t3/2
ssra v16.8h, v2.8h, #1 // (t1 + t3) >> 1
neg v18.8h, v17.8h // -t4/2
ssra v17.8h, v1.8h, #1 // (t2 + t4) >> 1
ssra v3.8h, v2.8h, #1 // (t1 - t3) >> 1
ssra v18.8h, v1.8h, #1 // (t2 - t4) >> 1
srshr v1.8h, v16.8h, #2 // (t1 + t3 + 64) >> 3
srshr v2.8h, v17.8h, #2 // (t2 + t4 + 64) >> 3
srshr v3.8h, v3.8h, #2 // (t1 - t3 + 64) >> 3
srshr v16.8h, v18.8h, #2 // (t2 - t4 + 64) >> 3
trn2 v17.8h, v2.8h, v3.8h // 12 13 32 33 52 53 72 73
trn2 v18.8h, v1.8h, v16.8h // 10 11 30 31 50 51 70 71
trn1 v1.8h, v1.8h, v16.8h // 00 01 20 21 40 41 60 61
trn1 v2.8h, v2.8h, v3.8h // 02 03 22 23 42 43 62 63
trn1 v3.4s, v18.4s, v17.4s // 10 11 12 13 50 51 52 53
trn2 v16.4s, v18.4s, v17.4s // 30 31 32 33 70 71 72 73
trn1 v17.4s, v1.4s, v2.4s // 00 01 02 03 40 41 42 43
mov d18, v3.d[1] // 50 51 52 53
shl v19.4h, v3.4h, #4 // 16 * src[8]
mov d20, v16.d[1] // 70 71 72 73
shl v21.4h, v16.4h, #4 // 16 * src[24]
mov d22, v17.d[1] // 40 41 42 43
shl v23.4h, v3.4h, #2 // 4 * src[8]
shl v24.4h, v18.4h, #4 // 16 * src[40]
shl v25.4h, v20.4h, #4 // 16 * src[56]
shl v26.4h, v18.4h, #2 // 4 * src[40]
trn2 v1.4s, v1.4s, v2.4s // 20 21 22 23 60 61 62 63
ssra v24.4h, v21.4h, #2 // 4 * src[24] + 16 * src[40]
sub v2.4h, v25.4h, v23.4h // - 4 * src[8] + 16 * src[56]
shl v17.4h, v17.4h, #2 // 8/2 * src[0]
sub v21.4h, v21.4h, v26.4h // 16 * src[24] - 4 * src[40]
shl v22.4h, v22.4h, #2 // 8/2 * src[32]
mov d23, v1.d[1] // 60 61 62 63
ssra v19.4h, v25.4h, #2 // 16 * src[8] + 4 * src[56]
mul v25.4h, v1.4h, v0.h[0] // 6/2 * src[16]
shl v1.4h, v1.4h, #3 // 16/2 * src[16]
mls v24.4h, v3.4h, v0.h[2] // - 15 * src[8] + 4 * src[24] + 16 * src[40]
ssra v17.4h, v17.4h, #1 // 12/2 * src[0]
mls v21.4h, v3.4h, v0.h[1] // - 9 * src[8] + 16 * src[24] - 4 * src[40]
ssra v22.4h, v22.4h, #1 // 12/2 * src[32]
mla v2.4h, v16.4h, v0.h[1] // - 4 * src[8] + 9 * src[24] + 16 * src[56]
shl v3.4h, v23.4h, #3 // 16/2 * src[48]
mla v19.4h, v16.4h, v0.h[2] // 16 * src[8] + 15 * src[24] + 4 * src[56]
mla v1.4h, v23.4h, v0.h[0] // t3/2 = 16/2 * src[16] + 6/2 * src[48]
mla v24.4h, v20.4h, v0.h[1] // -t2 = - 15 * src[8] + 4 * src[24] + 16 * src[40] + 9 * src[56]
add v16.4h, v17.4h, v22.4h // t1/2 = 12/2 * src[0] + 12/2 * src[32]
sub v3.4h, v25.4h, v3.4h // t4/2 = 6/2 * src[16] - 16/2 * src[48]
sub v17.4h, v17.4h, v22.4h // t2/2 = 12/2 * src[0] - 12/2 * src[32]
mls v21.4h, v20.4h, v0.h[2] // -t3 = - 9 * src[8] + 16 * src[24] - 4 * src[40] - 15 * src[56]
mla v19.4h, v18.4h, v0.h[1] // t1 = 16 * src[8] + 15 * src[24] + 9 * src[40] + 4 * src[56]
add v20.4h, v16.4h, v1.4h // t5/2 = t1/2 + t3/2
mls v2.4h, v18.4h, v0.h[2] // -t4 = - 4 * src[8] + 9 * src[24] - 15 * src[40] + 16 * src[56]
sub v0.4h, v16.4h, v1.4h // t8/2 = t1/2 - t3/2
add v18.4h, v17.4h, v3.4h // t6/2 = t2/2 + t4/2
sub v22.4h, v17.4h, v3.4h // t7/2 = t2/2 - t4/2
neg v23.4h, v24.4h // +t2
sub v25.4h, v17.4h, v3.4h // t7/2 = t2/2 - t4/2
add v3.4h, v17.4h, v3.4h // t6/2 = t2/2 + t4/2
neg v17.4h, v21.4h // +t3
sub v26.4h, v16.4h, v1.4h // t8/2 = t1/2 - t3/2
add v1.4h, v16.4h, v1.4h // t5/2 = t1/2 + t3/2
neg v16.4h, v19.4h // -t1
neg v27.4h, v2.4h // +t4
ssra v20.4h, v19.4h, #1 // (t5 + t1) >> 1
srsra v0.4h, v2.4h, #1 // (t8 - t4 + 1) >> 1
ssra v18.4h, v23.4h, #1 // (t6 + t2) >> 1
srsra v22.4h, v21.4h, #1 // (t7 - t3 + 1) >> 1
ssra v25.4h, v17.4h, #1 // (t7 + t3) >> 1
srsra v3.4h, v24.4h, #1 // (t6 - t2 + 1) >> 1
ssra v26.4h, v27.4h, #1 // (t8 + t4) >> 1
srsra v1.4h, v16.4h, #1 // (t5 - t1 + 1) >> 1
trn1 v0.2d, v20.2d, v0.2d
trn1 v2.2d, v18.2d, v22.2d
trn1 v3.2d, v25.2d, v3.2d
trn1 v1.2d, v26.2d, v1.2d
srshr v0.8h, v0.8h, #6 // (t5 + t1 + 64) >> 7, (t8 - t4 + 65) >> 7
srshr v2.8h, v2.8h, #6 // (t6 + t2 + 64) >> 7, (t7 - t3 + 65) >> 7
srshr v3.8h, v3.8h, #6 // (t7 + t3 + 64) >> 7, (t6 - t2 + 65) >> 7
srshr v1.8h, v1.8h, #6 // (t8 + t4 + 64) >> 7, (t5 - t1 + 65) >> 7
uaddw v0.8h, v0.8h, v5.8b
uaddw v2.8h, v2.8h, v6.8b
uaddw v3.8h, v3.8h, v7.8b
uaddw v1.8h, v1.8h, v4.8b
sqxtun v0.8b, v0.8h
sqxtun v2.8b, v2.8h
sqxtun v3.8b, v3.8h
sqxtun v1.8b, v1.8h
st1 {v0.s}[0], [x4], x1
st1 {v2.s}[0], [x4], x1
st1 {v3.s}[0], [x4], x1
st1 {v1.s}[0], [x4], x1
st1 {v0.s}[1], [x4], x1
st1 {v2.s}[1], [x4], x1
st1 {v3.s}[1], [x4], x1
st1 {v1.s}[1], [x4]
ret
endfunc
// VC-1 4x4 inverse transform
// On entry:
// x0 -> array of 8-bit samples, in row-major order
// x1 = row stride for 8-bit sample array
// x2 -> array of 16-bit inverse transform coefficients, in row-major order (row stride is 8 coefficients)
// On exit:
// array at x0 updated by saturated addition of (narrowed) transformed block
function ff_vc1_inv_trans_4x4_neon, export=1
mov x3, #16
ldr d0, .Lcoeffs_it4
mov x4, x0
ld1 {v1.d}[0], [x2], x3 // 00 01 02 03
ld1 {v2.d}[0], [x2], x3 // 10 11 12 13
ld1 {v3.d}[0], [x2], x3 // 20 21 22 23
ld1 {v4.d}[0], [x2] // 30 31 32 33
ld1 {v5.s}[0], [x0], x1
ld1 {v5.s}[1], [x0], x1
ld1 {v6.s}[0], [x0], x1
trn2 v7.4h, v1.4h, v2.4h // 01 11 03 13
trn1 v1.4h, v1.4h, v2.4h // 00 10 02 12
ld1 {v6.s}[1], [x0]
trn2 v2.4h, v3.4h, v4.4h // 21 31 23 33
trn1 v3.4h, v3.4h, v4.4h // 20 30 22 32
trn2 v4.2s, v7.2s, v2.2s // 03 13 23 33
trn1 v16.2s, v1.2s, v3.2s // 00 10 20 30
trn1 v2.2s, v7.2s, v2.2s // 01 11 21 31
trn2 v1.2s, v1.2s, v3.2s // 02 12 22 32
mul v3.4h, v4.4h, v0.h[0] // 10/2 * src[3]
mul v4.4h, v4.4h, v0.h[1] // 22/2 * src[3]
mul v7.4h, v16.4h, v0.h[2] // 17 * src[0]
mul v1.4h, v1.4h, v0.h[2] // 17 * src[2]
mla v3.4h, v2.4h, v0.h[1] // t3/2 = 22/2 * src[1] + 10/2 * src[3]
mls v4.4h, v2.4h, v0.h[0] // t4/2 = - 10/2 * src[1] + 22/2 * src[3]
add v2.4h, v7.4h, v1.4h // t1 = 17 * src[0] + 17 * src[2]
sub v1.4h, v7.4h, v1.4h // t2 = 17 * src[0] - 17 * src[2]
neg v7.4h, v3.4h // -t3/2
neg v16.4h, v4.4h // -t4/2
ssra v3.4h, v2.4h, #1 // (t1 + t3) >> 1
ssra v4.4h, v1.4h, #1 // (t2 + t4) >> 1
ssra v16.4h, v1.4h, #1 // (t2 - t4) >> 1
ssra v7.4h, v2.4h, #1 // (t1 - t3) >> 1
srshr v1.4h, v3.4h, #2 // (t1 + t3 + 64) >> 3
srshr v2.4h, v4.4h, #2 // (t2 + t4 + 64) >> 3
srshr v3.4h, v16.4h, #2 // (t2 - t4 + 64) >> 3
srshr v4.4h, v7.4h, #2 // (t1 - t3 + 64) >> 3
trn2 v7.4h, v1.4h, v3.4h // 10 11 30 31
trn1 v1.4h, v1.4h, v3.4h // 00 01 20 21
trn2 v3.4h, v2.4h, v4.4h // 12 13 32 33
trn1 v2.4h, v2.4h, v4.4h // 02 03 22 23
trn2 v4.2s, v7.2s, v3.2s // 30 31 32 33
trn1 v16.2s, v1.2s, v2.2s // 00 01 02 03
trn1 v3.2s, v7.2s, v3.2s // 10 11 12 13
trn2 v1.2s, v1.2s, v2.2s // 20 21 22 23
mul v2.4h, v4.4h, v0.h[1] // 22/2 * src[24]
mul v4.4h, v4.4h, v0.h[0] // 10/2 * src[24]
mul v7.4h, v16.4h, v0.h[2] // 17 * src[0]
mul v1.4h, v1.4h, v0.h[2] // 17 * src[16]
mls v2.4h, v3.4h, v0.h[0] // t4/2 = - 10/2 * src[8] + 22/2 * src[24]
mla v4.4h, v3.4h, v0.h[1] // t3/2 = 22/2 * src[8] + 10/2 * src[24]
add v0.4h, v7.4h, v1.4h // t1 = 17 * src[0] + 17 * src[16]
sub v1.4h, v7.4h, v1.4h // t2 = 17 * src[0] - 17 * src[16]
neg v3.4h, v2.4h // -t4/2
neg v7.4h, v4.4h // -t3/2
ssra v4.4h, v0.4h, #1 // (t1 + t3) >> 1
ssra v3.4h, v1.4h, #1 // (t2 - t4) >> 1
ssra v2.4h, v1.4h, #1 // (t2 + t4) >> 1
ssra v7.4h, v0.4h, #1 // (t1 - t3) >> 1
trn1 v0.2d, v4.2d, v3.2d
trn1 v1.2d, v2.2d, v7.2d
srshr v0.8h, v0.8h, #6 // (t1 + t3 + 64) >> 7, (t2 - t4 + 64) >> 7
srshr v1.8h, v1.8h, #6 // (t2 + t4 + 64) >> 7, (t1 - t3 + 64) >> 7
uaddw v0.8h, v0.8h, v5.8b
uaddw v1.8h, v1.8h, v6.8b
sqxtun v0.8b, v0.8h
sqxtun v1.8b, v1.8h
st1 {v0.s}[0], [x4], x1
st1 {v0.s}[1], [x4], x1
st1 {v1.s}[0], [x4], x1
st1 {v1.s}[1], [x4]
ret
endfunc
// VC-1 8x8 inverse transform, DC case
// On entry:
// x0 -> array of 8-bit samples, in row-major order
// x1 = row stride for 8-bit sample array
// x2 -> 16-bit inverse transform DC coefficient
// On exit:
// array at x0 updated by saturated addition of (narrowed) transformed block
function ff_vc1_inv_trans_8x8_dc_neon, export=1
ldrsh w2, [x2]
mov x3, x0
ld1 {v0.8b}, [x0], x1
ld1 {v1.8b}, [x0], x1
ld1 {v2.8b}, [x0], x1
add w2, w2, w2, lsl #1
ld1 {v3.8b}, [x0], x1
ld1 {v4.8b}, [x0], x1
add w2, w2, #1
ld1 {v5.8b}, [x0], x1
asr w2, w2, #1
ld1 {v6.8b}, [x0], x1
add w2, w2, w2, lsl #1
ld1 {v7.8b}, [x0]
add w0, w2, #16
asr w0, w0, #5
dup v16.8h, w0
uaddw v0.8h, v16.8h, v0.8b
uaddw v1.8h, v16.8h, v1.8b
uaddw v2.8h, v16.8h, v2.8b
uaddw v3.8h, v16.8h, v3.8b
uaddw v4.8h, v16.8h, v4.8b
uaddw v5.8h, v16.8h, v5.8b
sqxtun v0.8b, v0.8h
uaddw v6.8h, v16.8h, v6.8b
sqxtun v1.8b, v1.8h
uaddw v7.8h, v16.8h, v7.8b
sqxtun v2.8b, v2.8h
sqxtun v3.8b, v3.8h
sqxtun v4.8b, v4.8h
st1 {v0.8b}, [x3], x1
sqxtun v0.8b, v5.8h
st1 {v1.8b}, [x3], x1
sqxtun v1.8b, v6.8h
st1 {v2.8b}, [x3], x1
sqxtun v2.8b, v7.8h
st1 {v3.8b}, [x3], x1
st1 {v4.8b}, [x3], x1
st1 {v0.8b}, [x3], x1
st1 {v1.8b}, [x3], x1
st1 {v2.8b}, [x3]
ret
endfunc
// VC-1 8x4 inverse transform, DC case
// On entry:
// x0 -> array of 8-bit samples, in row-major order
// x1 = row stride for 8-bit sample array
// x2 -> 16-bit inverse transform DC coefficient
// On exit:
// array at x0 updated by saturated addition of (narrowed) transformed block
function ff_vc1_inv_trans_8x4_dc_neon, export=1
ldrsh w2, [x2]
mov x3, x0
ld1 {v0.8b}, [x0], x1
ld1 {v1.8b}, [x0], x1
ld1 {v2.8b}, [x0], x1
add w2, w2, w2, lsl #1
ld1 {v3.8b}, [x0]
add w0, w2, #1
asr w0, w0, #1
add w0, w0, w0, lsl #4
add w0, w0, #64
asr w0, w0, #7
dup v4.8h, w0
uaddw v0.8h, v4.8h, v0.8b
uaddw v1.8h, v4.8h, v1.8b
uaddw v2.8h, v4.8h, v2.8b
uaddw v3.8h, v4.8h, v3.8b
sqxtun v0.8b, v0.8h
sqxtun v1.8b, v1.8h
sqxtun v2.8b, v2.8h
sqxtun v3.8b, v3.8h
st1 {v0.8b}, [x3], x1
st1 {v1.8b}, [x3], x1
st1 {v2.8b}, [x3], x1
st1 {v3.8b}, [x3]
ret
endfunc
// VC-1 4x8 inverse transform, DC case
// On entry:
// x0 -> array of 8-bit samples, in row-major order
// x1 = row stride for 8-bit sample array
// x2 -> 16-bit inverse transform DC coefficient
// On exit:
// array at x0 updated by saturated addition of (narrowed) transformed block
function ff_vc1_inv_trans_4x8_dc_neon, export=1
ldrsh w2, [x2]
mov x3, x0
ld1 {v0.s}[0], [x0], x1
ld1 {v1.s}[0], [x0], x1
ld1 {v2.s}[0], [x0], x1
add w2, w2, w2, lsl #4
ld1 {v3.s}[0], [x0], x1
add w2, w2, #4
asr w2, w2, #3
add w2, w2, w2, lsl #1
ld1 {v0.s}[1], [x0], x1
add w2, w2, #16
asr w2, w2, #5
dup v4.8h, w2
ld1 {v1.s}[1], [x0], x1
ld1 {v2.s}[1], [x0], x1
ld1 {v3.s}[1], [x0]
uaddw v0.8h, v4.8h, v0.8b
uaddw v1.8h, v4.8h, v1.8b
uaddw v2.8h, v4.8h, v2.8b
uaddw v3.8h, v4.8h, v3.8b
sqxtun v0.8b, v0.8h
sqxtun v1.8b, v1.8h
sqxtun v2.8b, v2.8h
sqxtun v3.8b, v3.8h
st1 {v0.s}[0], [x3], x1
st1 {v1.s}[0], [x3], x1
st1 {v2.s}[0], [x3], x1
st1 {v3.s}[0], [x3], x1
st1 {v0.s}[1], [x3], x1
st1 {v1.s}[1], [x3], x1
st1 {v2.s}[1], [x3], x1
st1 {v3.s}[1], [x3]
ret
endfunc
// VC-1 4x4 inverse transform, DC case
// On entry:
// x0 -> array of 8-bit samples, in row-major order
// x1 = row stride for 8-bit sample array
// x2 -> 16-bit inverse transform DC coefficient
// On exit:
// array at x0 updated by saturated addition of (narrowed) transformed block
function ff_vc1_inv_trans_4x4_dc_neon, export=1
ldrsh w2, [x2]
mov x3, x0
ld1 {v0.s}[0], [x0], x1
ld1 {v1.s}[0], [x0], x1
ld1 {v0.s}[1], [x0], x1
add w2, w2, w2, lsl #4
ld1 {v1.s}[1], [x0]
add w0, w2, #4
asr w0, w0, #3
add w0, w0, w0, lsl #4
add w0, w0, #64
asr w0, w0, #7
dup v2.8h, w0
uaddw v0.8h, v2.8h, v0.8b
uaddw v1.8h, v2.8h, v1.8b
sqxtun v0.8b, v0.8h
sqxtun v1.8b, v1.8h
st1 {v0.s}[0], [x3], x1
st1 {v1.s}[0], [x3], x1
st1 {v0.s}[1], [x3], x1
st1 {v1.s}[1], [x3]
ret
endfunc
.align 5
.Lcoeffs_it8:
.quad 0x000F00090003
.Lcoeffs_it4:
.quad 0x0011000B0005
.Lcoeffs:
.quad 0x00050002
// VC-1 in-loop deblocking filter for 4 pixel pairs at boundary of vertically-neighbouring blocks
// On entry:
// x0 -> top-left pel of lower block
// x1 = row stride, bytes
// w2 = PQUANT bitstream parameter
function ff_vc1_v_loop_filter4_neon, export=1
sub x3, x0, w1, sxtw #2
ldr d0, .Lcoeffs
ld1 {v1.s}[0], [x0], x1 // P5
ld1 {v2.s}[0], [x3], x1 // P1
ld1 {v3.s}[0], [x3], x1 // P2
ld1 {v4.s}[0], [x0], x1 // P6
ld1 {v5.s}[0], [x3], x1 // P3
ld1 {v6.s}[0], [x0], x1 // P7
ld1 {v7.s}[0], [x3] // P4
ld1 {v16.s}[0], [x0] // P8
ushll v17.8h, v1.8b, #1 // 2*P5
dup v18.8h, w2 // pq
ushll v2.8h, v2.8b, #1 // 2*P1
uxtl v3.8h, v3.8b // P2
uxtl v4.8h, v4.8b // P6
uxtl v19.8h, v5.8b // P3
mls v2.4h, v3.4h, v0.h[1] // 2*P1-5*P2
uxtl v3.8h, v6.8b // P7
mls v17.4h, v4.4h, v0.h[1] // 2*P5-5*P6
ushll v5.8h, v5.8b, #1 // 2*P3
uxtl v6.8h, v7.8b // P4
mla v17.4h, v3.4h, v0.h[1] // 2*P5-5*P6+5*P7
uxtl v3.8h, v16.8b // P8
mla v2.4h, v19.4h, v0.h[1] // 2*P1-5*P2+5*P3
uxtl v1.8h, v1.8b // P5
mls v5.4h, v6.4h, v0.h[1] // 2*P3-5*P4
mls v17.4h, v3.4h, v0.h[0] // 2*P5-5*P6+5*P7-2*P8
sub v3.4h, v6.4h, v1.4h // P4-P5
mls v2.4h, v6.4h, v0.h[0] // 2*P1-5*P2+5*P3-2*P4
mla v5.4h, v1.4h, v0.h[1] // 2*P3-5*P4+5*P5
mls v5.4h, v4.4h, v0.h[0] // 2*P3-5*P4+5*P5-2*P6
abs v4.4h, v3.4h
srshr v7.4h, v17.4h, #3
srshr v2.4h, v2.4h, #3
sshr v4.4h, v4.4h, #1 // clip
srshr v5.4h, v5.4h, #3
abs v7.4h, v7.4h // a2
sshr v3.4h, v3.4h, #8 // clip_sign
abs v2.4h, v2.4h // a1
cmeq v16.4h, v4.4h, #0 // test clip == 0
abs v17.4h, v5.4h // a0
sshr v5.4h, v5.4h, #8 // a0_sign
cmhs v19.4h, v2.4h, v7.4h // test a1 >= a2
cmhs v18.4h, v17.4h, v18.4h // test a0 >= pq
sub v3.4h, v3.4h, v5.4h // clip_sign - a0_sign
bsl v19.8b, v7.8b, v2.8b // a3
orr v2.8b, v16.8b, v18.8b // test clip == 0 || a0 >= pq
uqsub v5.4h, v17.4h, v19.4h // a0 >= a3 ? a0-a3 : 0 (a0 > a3 in all cases where filtering is enabled, so makes more sense to subtract this way round than the opposite and then taking the abs)
cmhs v7.4h, v19.4h, v17.4h // test a3 >= a0
mul v0.4h, v5.4h, v0.h[1] // a0 >= a3 ? 5*(a0-a3) : 0
orr v5.8b, v2.8b, v7.8b // test clip == 0 || a0 >= pq || a3 >= a0
mov w0, v5.s[1] // move to gp reg
ushr v0.4h, v0.4h, #3 // a0 >= a3 ? (5*(a0-a3))>>3 : 0
cmhs v5.4h, v0.4h, v4.4h
tbnz w0, #0, 1f // none of the 4 pixel pairs should be updated if this one is not filtered
bsl v5.8b, v4.8b, v0.8b // FFMIN(d, clip)
bic v0.8b, v5.8b, v2.8b // set each d to zero if it should not be filtered because clip == 0 || a0 >= pq (a3 > a0 case already zeroed by saturating sub)
mls v6.4h, v0.4h, v3.4h // invert d depending on clip_sign & a0_sign, or zero it if they match, and accumulate into P4
mla v1.4h, v0.4h, v3.4h // invert d depending on clip_sign & a0_sign, or zero it if they match, and accumulate into P5
sqxtun v0.8b, v6.8h
sqxtun v1.8b, v1.8h
st1 {v0.s}[0], [x3], x1
st1 {v1.s}[0], [x3]
1: ret
endfunc
// VC-1 in-loop deblocking filter for 4 pixel pairs at boundary of horizontally-neighbouring blocks
// On entry:
// x0 -> top-left pel of right block
// x1 = row stride, bytes
// w2 = PQUANT bitstream parameter
function ff_vc1_h_loop_filter4_neon, export=1
sub x3, x0, #4 // where to start reading
ldr d0, .Lcoeffs
ld1 {v1.8b}, [x3], x1
sub x0, x0, #1 // where to start writing
ld1 {v2.8b}, [x3], x1
ld1 {v3.8b}, [x3], x1
ld1 {v4.8b}, [x3]
dup v5.8h, w2 // pq
trn1 v6.8b, v1.8b, v2.8b
trn2 v1.8b, v1.8b, v2.8b
trn1 v2.8b, v3.8b, v4.8b
trn2 v3.8b, v3.8b, v4.8b
trn1 v4.4h, v6.4h, v2.4h // P1, P5
trn1 v7.4h, v1.4h, v3.4h // P2, P6
trn2 v2.4h, v6.4h, v2.4h // P3, P7
trn2 v1.4h, v1.4h, v3.4h // P4, P8
ushll v3.8h, v4.8b, #1 // 2*P1, 2*P5
uxtl v6.8h, v7.8b // P2, P6
uxtl v7.8h, v2.8b // P3, P7
uxtl v1.8h, v1.8b // P4, P8
mls v3.8h, v6.8h, v0.h[1] // 2*P1-5*P2, 2*P5-5*P6
ushll v2.8h, v2.8b, #1 // 2*P3, 2*P7
uxtl v4.8h, v4.8b // P1, P5
mla v3.8h, v7.8h, v0.h[1] // 2*P1-5*P2+5*P3, 2*P5-5*P6+5*P7
mov d6, v6.d[1] // P6
mls v3.8h, v1.8h, v0.h[0] // 2*P1-5*P2+5*P3-2*P4, 2*P5-5*P6+5*P7-2*P8
mov d4, v4.d[1] // P5
mls v2.4h, v1.4h, v0.h[1] // 2*P3-5*P4
mla v2.4h, v4.4h, v0.h[1] // 2*P3-5*P4+5*P5
sub v7.4h, v1.4h, v4.4h // P4-P5
mls v2.4h, v6.4h, v0.h[0] // 2*P3-5*P4+5*P5-2*P6
srshr v3.8h, v3.8h, #3
abs v6.4h, v7.4h
sshr v7.4h, v7.4h, #8 // clip_sign
srshr v2.4h, v2.4h, #3
abs v3.8h, v3.8h // a1, a2
sshr v6.4h, v6.4h, #1 // clip
mov d16, v3.d[1] // a2
abs v17.4h, v2.4h // a0
cmeq v18.4h, v6.4h, #0 // test clip == 0
sshr v2.4h, v2.4h, #8 // a0_sign
cmhs v19.4h, v3.4h, v16.4h // test a1 >= a2
cmhs v5.4h, v17.4h, v5.4h // test a0 >= pq
sub v2.4h, v7.4h, v2.4h // clip_sign - a0_sign
bsl v19.8b, v16.8b, v3.8b // a3
orr v3.8b, v18.8b, v5.8b // test clip == 0 || a0 >= pq
uqsub v5.4h, v17.4h, v19.4h // a0 >= a3 ? a0-a3 : 0 (a0 > a3 in all cases where filtering is enabled, so makes more sense to subtract this way round than the opposite and then taking the abs)
cmhs v7.4h, v19.4h, v17.4h // test a3 >= a0
mul v0.4h, v5.4h, v0.h[1] // a0 >= a3 ? 5*(a0-a3) : 0
orr v5.8b, v3.8b, v7.8b // test clip == 0 || a0 >= pq || a3 >= a0
mov w2, v5.s[1] // move to gp reg
ushr v0.4h, v0.4h, #3 // a0 >= a3 ? (5*(a0-a3))>>3 : 0
cmhs v5.4h, v0.4h, v6.4h
tbnz w2, #0, 1f // none of the 4 pixel pairs should be updated if this one is not filtered
bsl v5.8b, v6.8b, v0.8b // FFMIN(d, clip)
bic v0.8b, v5.8b, v3.8b // set each d to zero if it should not be filtered because clip == 0 || a0 >= pq (a3 > a0 case already zeroed by saturating sub)
mla v4.4h, v0.4h, v2.4h // invert d depending on clip_sign & a0_sign, or zero it if they match, and accumulate into P5
mls v1.4h, v0.4h, v2.4h // invert d depending on clip_sign & a0_sign, or zero it if they match, and accumulate into P4
sqxtun v3.8b, v4.8h
sqxtun v2.8b, v1.8h
st2 {v2.b, v3.b}[0], [x0], x1
st2 {v2.b, v3.b}[1], [x0], x1
st2 {v2.b, v3.b}[2], [x0], x1
st2 {v2.b, v3.b}[3], [x0]
1: ret
endfunc
// VC-1 in-loop deblocking filter for 8 pixel pairs at boundary of vertically-neighbouring blocks
// On entry:
// x0 -> top-left pel of lower block
// x1 = row stride, bytes
// w2 = PQUANT bitstream parameter
function ff_vc1_v_loop_filter8_neon, export=1
sub x3, x0, w1, sxtw #2
ldr d0, .Lcoeffs
ld1 {v1.8b}, [x0], x1 // P5
movi v2.2d, #0x0000ffff00000000
ld1 {v3.8b}, [x3], x1 // P1
ld1 {v4.8b}, [x3], x1 // P2
ld1 {v5.8b}, [x0], x1 // P6
ld1 {v6.8b}, [x3], x1 // P3
ld1 {v7.8b}, [x0], x1 // P7
ushll v16.8h, v1.8b, #1 // 2*P5
ushll v3.8h, v3.8b, #1 // 2*P1
ld1 {v17.8b}, [x3] // P4
uxtl v4.8h, v4.8b // P2
ld1 {v18.8b}, [x0] // P8
uxtl v5.8h, v5.8b // P6
dup v19.8h, w2 // pq
uxtl v20.8h, v6.8b // P3
mls v3.8h, v4.8h, v0.h[1] // 2*P1-5*P2
uxtl v4.8h, v7.8b // P7
ushll v6.8h, v6.8b, #1 // 2*P3
mls v16.8h, v5.8h, v0.h[1] // 2*P5-5*P6
uxtl v7.8h, v17.8b // P4
uxtl v17.8h, v18.8b // P8
mla v16.8h, v4.8h, v0.h[1] // 2*P5-5*P6+5*P7
uxtl v1.8h, v1.8b // P5
mla v3.8h, v20.8h, v0.h[1] // 2*P1-5*P2+5*P3
sub v4.8h, v7.8h, v1.8h // P4-P5
mls v6.8h, v7.8h, v0.h[1] // 2*P3-5*P4
mls v16.8h, v17.8h, v0.h[0] // 2*P5-5*P6+5*P7-2*P8
abs v17.8h, v4.8h
sshr v4.8h, v4.8h, #8 // clip_sign
mls v3.8h, v7.8h, v0.h[0] // 2*P1-5*P2+5*P3-2*P4
sshr v17.8h, v17.8h, #1 // clip
mla v6.8h, v1.8h, v0.h[1] // 2*P3-5*P4+5*P5
srshr v16.8h, v16.8h, #3
mls v6.8h, v5.8h, v0.h[0] // 2*P3-5*P4+5*P5-2*P6
cmeq v5.8h, v17.8h, #0 // test clip == 0
srshr v3.8h, v3.8h, #3
abs v16.8h, v16.8h // a2
abs v3.8h, v3.8h // a1
srshr v6.8h, v6.8h, #3
cmhs v18.8h, v3.8h, v16.8h // test a1 >= a2
abs v20.8h, v6.8h // a0
sshr v6.8h, v6.8h, #8 // a0_sign
bsl v18.16b, v16.16b, v3.16b // a3
cmhs v3.8h, v20.8h, v19.8h // test a0 >= pq
sub v4.8h, v4.8h, v6.8h // clip_sign - a0_sign
uqsub v6.8h, v20.8h, v18.8h // a0 >= a3 ? a0-a3 : 0 (a0 > a3 in all cases where filtering is enabled, so makes more sense to subtract this way round than the opposite and then taking the abs)
cmhs v16.8h, v18.8h, v20.8h // test a3 >= a0
orr v3.16b, v5.16b, v3.16b // test clip == 0 || a0 >= pq
mul v0.8h, v6.8h, v0.h[1] // a0 >= a3 ? 5*(a0-a3) : 0
orr v5.16b, v3.16b, v16.16b // test clip == 0 || a0 >= pq || a3 >= a0
cmtst v2.2d, v5.2d, v2.2d // if 2nd of each group of is not filtered, then none of the others in the group should be either
mov w0, v5.s[1] // move to gp reg
ushr v0.8h, v0.8h, #3 // a0 >= a3 ? (5*(a0-a3))>>3 : 0
mov w2, v5.s[3]
orr v2.16b, v3.16b, v2.16b
cmhs v3.8h, v0.8h, v17.8h
and w0, w0, w2
bsl v3.16b, v17.16b, v0.16b // FFMIN(d, clip)
tbnz w0, #0, 1f // none of the 8 pixel pairs should be updated in this case
bic v0.16b, v3.16b, v2.16b // set each d to zero if it should not be filtered
mls v7.8h, v0.8h, v4.8h // invert d depending on clip_sign & a0_sign, or zero it if they match, and accumulate into P4
mla v1.8h, v0.8h, v4.8h // invert d depending on clip_sign & a0_sign, or zero it if they match, and accumulate into P5
sqxtun v0.8b, v7.8h
sqxtun v1.8b, v1.8h
st1 {v0.8b}, [x3], x1
st1 {v1.8b}, [x3]
1: ret
endfunc
// VC-1 in-loop deblocking filter for 8 pixel pairs at boundary of horizontally-neighbouring blocks
// On entry:
// x0 -> top-left pel of right block
// x1 = row stride, bytes
// w2 = PQUANT bitstream parameter
function ff_vc1_h_loop_filter8_neon, export=1
sub x3, x0, #4 // where to start reading
ldr d0, .Lcoeffs
ld1 {v1.8b}, [x3], x1 // P1[0], P2[0]...
sub x0, x0, #1 // where to start writing
ld1 {v2.8b}, [x3], x1
add x4, x0, x1, lsl #2
ld1 {v3.8b}, [x3], x1
ld1 {v4.8b}, [x3], x1
ld1 {v5.8b}, [x3], x1
ld1 {v6.8b}, [x3], x1
ld1 {v7.8b}, [x3], x1
trn1 v16.8b, v1.8b, v2.8b // P1[0], P1[1], P3[0]...
ld1 {v17.8b}, [x3]
trn2 v1.8b, v1.8b, v2.8b // P2[0], P2[1], P4[0]...
trn1 v2.8b, v3.8b, v4.8b // P1[2], P1[3], P3[2]...
trn2 v3.8b, v3.8b, v4.8b // P2[2], P2[3], P4[2]...
dup v4.8h, w2 // pq
trn1 v18.8b, v5.8b, v6.8b // P1[4], P1[5], P3[4]...
trn2 v5.8b, v5.8b, v6.8b // P2[4], P2[5], P4[4]...
trn1 v6.4h, v16.4h, v2.4h // P1[0], P1[1], P1[2], P1[3], P5[0]...
trn1 v19.4h, v1.4h, v3.4h // P2[0], P2[1], P2[2], P2[3], P6[0]...
trn1 v20.8b, v7.8b, v17.8b // P1[6], P1[7], P3[6]...
trn2 v7.8b, v7.8b, v17.8b // P2[6], P2[7], P4[6]...
trn2 v2.4h, v16.4h, v2.4h // P3[0], P3[1], P3[2], P3[3], P7[0]...
trn2 v1.4h, v1.4h, v3.4h // P4[0], P4[1], P4[2], P4[3], P8[0]...
trn1 v3.4h, v18.4h, v20.4h // P1[4], P1[5], P1[6], P1[7], P5[4]...
trn1 v16.4h, v5.4h, v7.4h // P2[4], P2[5], P2[6], P2[7], P6[4]...
trn2 v17.4h, v18.4h, v20.4h // P3[4], P3[5], P3[6], P3[7], P7[4]...
trn2 v5.4h, v5.4h, v7.4h // P4[4], P4[5], P4[6], P4[7], P8[4]...
trn1 v7.2s, v6.2s, v3.2s // P1
trn1 v18.2s, v19.2s, v16.2s // P2
trn2 v3.2s, v6.2s, v3.2s // P5
trn2 v6.2s, v19.2s, v16.2s // P6
trn1 v16.2s, v2.2s, v17.2s // P3
trn2 v2.2s, v2.2s, v17.2s // P7
ushll v7.8h, v7.8b, #1 // 2*P1
trn1 v17.2s, v1.2s, v5.2s // P4
ushll v19.8h, v3.8b, #1 // 2*P5
trn2 v1.2s, v1.2s, v5.2s // P8
uxtl v5.8h, v18.8b // P2
uxtl v6.8h, v6.8b // P6
uxtl v18.8h, v16.8b // P3
mls v7.8h, v5.8h, v0.h[1] // 2*P1-5*P2
uxtl v2.8h, v2.8b // P7
ushll v5.8h, v16.8b, #1 // 2*P3
mls v19.8h, v6.8h, v0.h[1] // 2*P5-5*P6
uxtl v16.8h, v17.8b // P4
uxtl v1.8h, v1.8b // P8
mla v19.8h, v2.8h, v0.h[1] // 2*P5-5*P6+5*P7
uxtl v2.8h, v3.8b // P5
mla v7.8h, v18.8h, v0.h[1] // 2*P1-5*P2+5*P3
sub v3.8h, v16.8h, v2.8h // P4-P5
mls v5.8h, v16.8h, v0.h[1] // 2*P3-5*P4
mls v19.8h, v1.8h, v0.h[0] // 2*P5-5*P6+5*P7-2*P8
abs v1.8h, v3.8h
sshr v3.8h, v3.8h, #8 // clip_sign
mls v7.8h, v16.8h, v0.h[0] // 2*P1-5*P2+5*P3-2*P4
sshr v1.8h, v1.8h, #1 // clip
mla v5.8h, v2.8h, v0.h[1] // 2*P3-5*P4+5*P5
srshr v17.8h, v19.8h, #3
mls v5.8h, v6.8h, v0.h[0] // 2*P3-5*P4+5*P5-2*P6
cmeq v6.8h, v1.8h, #0 // test clip == 0
srshr v7.8h, v7.8h, #3
abs v17.8h, v17.8h // a2
abs v7.8h, v7.8h // a1
srshr v5.8h, v5.8h, #3
cmhs v18.8h, v7.8h, v17.8h // test a1 >= a2
abs v19.8h, v5.8h // a0
sshr v5.8h, v5.8h, #8 // a0_sign
bsl v18.16b, v17.16b, v7.16b // a3
cmhs v4.8h, v19.8h, v4.8h // test a0 >= pq
sub v3.8h, v3.8h, v5.8h // clip_sign - a0_sign
uqsub v5.8h, v19.8h, v18.8h // a0 >= a3 ? a0-a3 : 0 (a0 > a3 in all cases where filtering is enabled, so makes more sense to subtract this way round than the opposite and then taking the abs)
cmhs v7.8h, v18.8h, v19.8h // test a3 >= a0
orr v4.16b, v6.16b, v4.16b // test clip == 0 || a0 >= pq
mul v0.8h, v5.8h, v0.h[1] // a0 >= a3 ? 5*(a0-a3) : 0
orr v5.16b, v4.16b, v7.16b // test clip == 0 || a0 >= pq || a3 >= a0
mov w2, v5.s[1] // move to gp reg
ushr v0.8h, v0.8h, #3 // a0 >= a3 ? (5*(a0-a3))>>3 : 0
mov w3, v5.s[3]
cmhs v5.8h, v0.8h, v1.8h
and w5, w2, w3
bsl v5.16b, v1.16b, v0.16b // FFMIN(d, clip)
tbnz w5, #0, 2f // none of the 8 pixel pairs should be updated in this case
bic v0.16b, v5.16b, v4.16b // set each d to zero if it should not be filtered because clip == 0 || a0 >= pq (a3 > a0 case already zeroed by saturating sub)
mla v2.8h, v0.8h, v3.8h // invert d depending on clip_sign & a0_sign, or zero it if they match, and accumulate into P5
mls v16.8h, v0.8h, v3.8h // invert d depending on clip_sign & a0_sign, or zero it if they match, and accumulate into P4
sqxtun v1.8b, v2.8h
sqxtun v0.8b, v16.8h
tbnz w2, #0, 1f // none of the first 4 pixel pairs should be updated if so
st2 {v0.b, v1.b}[0], [x0], x1
st2 {v0.b, v1.b}[1], [x0], x1
st2 {v0.b, v1.b}[2], [x0], x1
st2 {v0.b, v1.b}[3], [x0]
1: tbnz w3, #0, 2f // none of the second 4 pixel pairs should be updated if so
st2 {v0.b, v1.b}[4], [x4], x1
st2 {v0.b, v1.b}[5], [x4], x1
st2 {v0.b, v1.b}[6], [x4], x1
st2 {v0.b, v1.b}[7], [x4]
2: ret
endfunc
// VC-1 in-loop deblocking filter for 16 pixel pairs at boundary of vertically-neighbouring blocks
// On entry:
// x0 -> top-left pel of lower block
// x1 = row stride, bytes
// w2 = PQUANT bitstream parameter
function ff_vc1_v_loop_filter16_neon, export=1
sub x3, x0, w1, sxtw #2
ldr d0, .Lcoeffs
ld1 {v1.16b}, [x0], x1 // P5
movi v2.2d, #0x0000ffff00000000
ld1 {v3.16b}, [x3], x1 // P1
ld1 {v4.16b}, [x3], x1 // P2
ld1 {v5.16b}, [x0], x1 // P6
ld1 {v6.16b}, [x3], x1 // P3
ld1 {v7.16b}, [x0], x1 // P7
ushll v16.8h, v1.8b, #1 // 2*P5[0..7]
ushll v17.8h, v3.8b, #1 // 2*P1[0..7]
ld1 {v18.16b}, [x3] // P4
uxtl v19.8h, v4.8b // P2[0..7]
ld1 {v20.16b}, [x0] // P8
uxtl v21.8h, v5.8b // P6[0..7]
dup v22.8h, w2 // pq
ushll2 v3.8h, v3.16b, #1 // 2*P1[8..15]
mls v17.8h, v19.8h, v0.h[1] // 2*P1[0..7]-5*P2[0..7]
ushll2 v19.8h, v1.16b, #1 // 2*P5[8..15]
uxtl2 v4.8h, v4.16b // P2[8..15]
mls v16.8h, v21.8h, v0.h[1] // 2*P5[0..7]-5*P6[0..7]
uxtl2 v5.8h, v5.16b // P6[8..15]
uxtl v23.8h, v6.8b // P3[0..7]
uxtl v24.8h, v7.8b // P7[0..7]
mls v3.8h, v4.8h, v0.h[1] // 2*P1[8..15]-5*P2[8..15]
ushll v4.8h, v6.8b, #1 // 2*P3[0..7]
uxtl v25.8h, v18.8b // P4[0..7]
mls v19.8h, v5.8h, v0.h[1] // 2*P5[8..15]-5*P6[8..15]
uxtl2 v26.8h, v6.16b // P3[8..15]
mla v17.8h, v23.8h, v0.h[1] // 2*P1[0..7]-5*P2[0..7]+5*P3[0..7]
uxtl2 v7.8h, v7.16b // P7[8..15]
ushll2 v6.8h, v6.16b, #1 // 2*P3[8..15]
mla v16.8h, v24.8h, v0.h[1] // 2*P5[0..7]-5*P6[0..7]+5*P7[0..7]
uxtl2 v18.8h, v18.16b // P4[8..15]
uxtl v23.8h, v20.8b // P8[0..7]
mls v4.8h, v25.8h, v0.h[1] // 2*P3[0..7]-5*P4[0..7]
uxtl v24.8h, v1.8b // P5[0..7]
uxtl2 v20.8h, v20.16b // P8[8..15]
mla v3.8h, v26.8h, v0.h[1] // 2*P1[8..15]-5*P2[8..15]+5*P3[8..15]
uxtl2 v1.8h, v1.16b // P5[8..15]
sub v26.8h, v25.8h, v24.8h // P4[0..7]-P5[0..7]
mla v19.8h, v7.8h, v0.h[1] // 2*P5[8..15]-5*P6[8..15]+5*P7[8..15]
sub v7.8h, v18.8h, v1.8h // P4[8..15]-P5[8..15]
mls v6.8h, v18.8h, v0.h[1] // 2*P3[8..15]-5*P4[8..15]
abs v27.8h, v26.8h
sshr v26.8h, v26.8h, #8 // clip_sign[0..7]
mls v17.8h, v25.8h, v0.h[0] // 2*P1[0..7]-5*P2[0..7]+5*P3[0..7]-2*P4[0..7]
abs v28.8h, v7.8h
sshr v27.8h, v27.8h, #1 // clip[0..7]
mls v16.8h, v23.8h, v0.h[0] // 2*P5[0..7]-5*P6[0..7]+5*P7[0..7]-2*P8[0..7]
sshr v7.8h, v7.8h, #8 // clip_sign[8..15]
sshr v23.8h, v28.8h, #1 // clip[8..15]
mla v4.8h, v24.8h, v0.h[1] // 2*P3[0..7]-5*P4[0..7]+5*P5[0..7]
cmeq v28.8h, v27.8h, #0 // test clip[0..7] == 0
srshr v17.8h, v17.8h, #3
mls v3.8h, v18.8h, v0.h[0] // 2*P1[8..15]-5*P2[8..15]+5*P3[8..15]-2*P4[8..15]
cmeq v29.8h, v23.8h, #0 // test clip[8..15] == 0
srshr v16.8h, v16.8h, #3
mls v19.8h, v20.8h, v0.h[0] // 2*P5[8..15]-5*P6[8..15]+5*P7[8..15]-2*P8[8..15]
abs v17.8h, v17.8h // a1[0..7]
mla v6.8h, v1.8h, v0.h[1] // 2*P3[8..15]-5*P4[8..15]+5*P5[8..15]
srshr v3.8h, v3.8h, #3
mls v4.8h, v21.8h, v0.h[0] // 2*P3[0..7]-5*P4[0..7]+5*P5[0..7]-2*P6[0..7]
abs v16.8h, v16.8h // a2[0..7]
srshr v19.8h, v19.8h, #3
mls v6.8h, v5.8h, v0.h[0] // 2*P3[8..15]-5*P4[8..15]+5*P5[8..15]-2*P6[8..15]
cmhs v5.8h, v17.8h, v16.8h // test a1[0..7] >= a2[0..7]
abs v3.8h, v3.8h // a1[8..15]
srshr v4.8h, v4.8h, #3
abs v19.8h, v19.8h // a2[8..15]
bsl v5.16b, v16.16b, v17.16b // a3[0..7]
srshr v6.8h, v6.8h, #3
cmhs v16.8h, v3.8h, v19.8h // test a1[8..15] >= a2[8.15]
abs v17.8h, v4.8h // a0[0..7]
sshr v4.8h, v4.8h, #8 // a0_sign[0..7]
bsl v16.16b, v19.16b, v3.16b // a3[8..15]
uqsub v3.8h, v17.8h, v5.8h // a0[0..7] >= a3[0..7] ? a0[0..7]-a3[0..7] : 0 (a0 > a3 in all cases where filtering is enabled, so makes more sense to subtract this way round than the opposite and then taking the abs)
abs v19.8h, v6.8h // a0[8..15]
cmhs v20.8h, v17.8h, v22.8h // test a0[0..7] >= pq
cmhs v5.8h, v5.8h, v17.8h // test a3[0..7] >= a0[0..7]
sub v4.8h, v26.8h, v4.8h // clip_sign[0..7] - a0_sign[0..7]
sshr v6.8h, v6.8h, #8 // a0_sign[8..15]
mul v3.8h, v3.8h, v0.h[1] // a0[0..7] >= a3[0..7] ? 5*(a0[0..7]-a3[0..7]) : 0
uqsub v17.8h, v19.8h, v16.8h // a0[8..15] >= a3[8..15] ? a0[8..15]-a3[8..15] : 0 (a0 > a3 in all cases where filtering is enabled, so makes more sense to subtract this way round than the opposite and then taking the abs)
orr v20.16b, v28.16b, v20.16b // test clip[0..7] == 0 || a0[0..7] >= pq
cmhs v21.8h, v19.8h, v22.8h // test a0[8..15] >= pq
cmhs v16.8h, v16.8h, v19.8h // test a3[8..15] >= a0[8..15]
mul v0.8h, v17.8h, v0.h[1] // a0[8..15] >= a3[8..15] ? 5*(a0[8..15]-a3[8..15]) : 0
sub v6.8h, v7.8h, v6.8h // clip_sign[8..15] - a0_sign[8..15]
orr v5.16b, v20.16b, v5.16b // test clip[0..7] == 0 || a0[0..7] >= pq || a3[0..7] >= a0[0..7]
ushr v3.8h, v3.8h, #3 // a0[0..7] >= a3[0..7] ? (5*(a0[0..7]-a3[0..7]))>>3 : 0
orr v7.16b, v29.16b, v21.16b // test clip[8..15] == 0 || a0[8..15] >= pq
cmtst v17.2d, v5.2d, v2.2d // if 2nd of each group of is not filtered, then none of the others in the group should be either
mov w0, v5.s[1] // move to gp reg
cmhs v19.8h, v3.8h, v27.8h
ushr v0.8h, v0.8h, #3 // a0[8..15] >= a3[8..15] ? (5*(a0[8..15]-a3[8..15]))>>3 : 0
mov w2, v5.s[3]
orr v5.16b, v7.16b, v16.16b // test clip[8..15] == 0 || a0[8..15] >= pq || a3[8..15] >= a0[8..15]
orr v16.16b, v20.16b, v17.16b
bsl v19.16b, v27.16b, v3.16b // FFMIN(d[0..7], clip[0..7])
cmtst v2.2d, v5.2d, v2.2d
cmhs v3.8h, v0.8h, v23.8h
mov w4, v5.s[1]
mov w5, v5.s[3]
and w0, w0, w2
bic v5.16b, v19.16b, v16.16b // set each d[0..7] to zero if it should not be filtered because clip[0..7] == 0 || a0[0..7] >= pq (a3 > a0 case already zeroed by saturating sub)
orr v2.16b, v7.16b, v2.16b
bsl v3.16b, v23.16b, v0.16b // FFMIN(d[8..15], clip[8..15])
mls v25.8h, v5.8h, v4.8h // invert d[0..7] depending on clip_sign[0..7] & a0_sign[0..7], or zero it if they match, and accumulate into P4[0..7]
and w2, w4, w5
bic v0.16b, v3.16b, v2.16b // set each d[8..15] to zero if it should not be filtered because clip[8..15] == 0 || a0[8..15] >= pq (a3 > a0 case already zeroed by saturating sub)
mla v24.8h, v5.8h, v4.8h // invert d[0..7] depending on clip_sign[0..7] & a0_sign[0..7], or zero it if they match, and accumulate into P5[0..7]
and w0, w0, w2
mls v18.8h, v0.8h, v6.8h // invert d[8..15] depending on clip_sign[8..15] & a0_sign[8..15], or zero it if they match, and accumulate into P4[8..15]
sqxtun v2.8b, v25.8h
tbnz w0, #0, 1f // none of the 16 pixel pairs should be updated in this case
mla v1.8h, v0.8h, v6.8h // invert d[8..15] depending on clip_sign[8..15] & a0_sign[8..15], or zero it if they match, and accumulate into P5[8..15]
sqxtun v0.8b, v24.8h
sqxtun2 v2.16b, v18.8h
sqxtun2 v0.16b, v1.8h
st1 {v2.16b}, [x3], x1
st1 {v0.16b}, [x3]
1: ret
endfunc
// VC-1 in-loop deblocking filter for 16 pixel pairs at boundary of horizontally-neighbouring blocks
// On entry:
// x0 -> top-left pel of right block
// x1 = row stride, bytes
// w2 = PQUANT bitstream parameter
function ff_vc1_h_loop_filter16_neon, export=1
sub x3, x0, #4 // where to start reading
ldr d0, .Lcoeffs
ld1 {v1.8b}, [x3], x1 // P1[0], P2[0]...
sub x0, x0, #1 // where to start writing
ld1 {v2.8b}, [x3], x1
add x4, x0, x1, lsl #3
ld1 {v3.8b}, [x3], x1
add x5, x0, x1, lsl #2
ld1 {v4.8b}, [x3], x1
add x6, x4, x1, lsl #2
ld1 {v5.8b}, [x3], x1
ld1 {v6.8b}, [x3], x1
ld1 {v7.8b}, [x3], x1
trn1 v16.8b, v1.8b, v2.8b // P1[0], P1[1], P3[0]...
ld1 {v17.8b}, [x3], x1
trn2 v1.8b, v1.8b, v2.8b // P2[0], P2[1], P4[0]...
ld1 {v2.8b}, [x3], x1
trn1 v18.8b, v3.8b, v4.8b // P1[2], P1[3], P3[2]...
ld1 {v19.8b}, [x3], x1
trn2 v3.8b, v3.8b, v4.8b // P2[2], P2[3], P4[2]...
ld1 {v4.8b}, [x3], x1
trn1 v20.8b, v5.8b, v6.8b // P1[4], P1[5], P3[4]...
ld1 {v21.8b}, [x3], x1
trn2 v5.8b, v5.8b, v6.8b // P2[4], P2[5], P4[4]...
ld1 {v6.8b}, [x3], x1
trn1 v22.8b, v7.8b, v17.8b // P1[6], P1[7], P3[6]...
ld1 {v23.8b}, [x3], x1
trn2 v7.8b, v7.8b, v17.8b // P2[6], P2[7], P4[6]...
ld1 {v17.8b}, [x3], x1
trn1 v24.8b, v2.8b, v19.8b // P1[8], P1[9], P3[8]...
ld1 {v25.8b}, [x3]
trn2 v2.8b, v2.8b, v19.8b // P2[8], P2[9], P4[8]...
trn1 v19.4h, v16.4h, v18.4h // P1[0], P1[1], P1[2], P1[3], P5[0]...
trn1 v26.8b, v4.8b, v21.8b // P1[10], P1[11], P3[10]...
trn2 v4.8b, v4.8b, v21.8b // P2[10], P2[11], P4[10]...
trn1 v21.4h, v1.4h, v3.4h // P2[0], P2[1], P2[2], P2[3], P6[0]...
trn1 v27.4h, v20.4h, v22.4h // P1[4], P1[5], P1[6], P1[7], P5[4]...
trn1 v28.8b, v6.8b, v23.8b // P1[12], P1[13], P3[12]...
trn2 v6.8b, v6.8b, v23.8b // P2[12], P2[13], P4[12]...
trn1 v23.4h, v5.4h, v7.4h // P2[4], P2[5], P2[6], P2[7], P6[4]...
trn1 v29.4h, v24.4h, v26.4h // P1[8], P1[9], P1[10], P1[11], P5[8]...
trn1 v30.8b, v17.8b, v25.8b // P1[14], P1[15], P3[14]...
trn2 v17.8b, v17.8b, v25.8b // P2[14], P2[15], P4[14]...
trn1 v25.4h, v2.4h, v4.4h // P2[8], P2[9], P2[10], P2[11], P6[8]...
trn1 v31.2s, v19.2s, v27.2s // P1[0..7]
trn2 v19.2s, v19.2s, v27.2s // P5[0..7]
trn1 v27.2s, v21.2s, v23.2s // P2[0..7]
trn2 v21.2s, v21.2s, v23.2s // P6[0..7]
trn1 v23.4h, v28.4h, v30.4h // P1[12], P1[13], P1[14], P1[15], P5[12]...
trn2 v16.4h, v16.4h, v18.4h // P3[0], P3[1], P3[2], P3[3], P7[0]...
trn1 v18.4h, v6.4h, v17.4h // P2[12], P2[13], P2[14], P2[15], P6[12]...
trn2 v20.4h, v20.4h, v22.4h // P3[4], P3[5], P3[6], P3[7], P7[4]...
trn2 v22.4h, v24.4h, v26.4h // P3[8], P3[9], P3[10], P3[11], P7[8]...
trn1 v24.2s, v29.2s, v23.2s // P1[8..15]
trn2 v23.2s, v29.2s, v23.2s // P5[8..15]
trn1 v26.2s, v25.2s, v18.2s // P2[8..15]
trn2 v18.2s, v25.2s, v18.2s // P6[8..15]
trn2 v25.4h, v28.4h, v30.4h // P3[12], P3[13], P3[14], P3[15], P7[12]...
trn2 v1.4h, v1.4h, v3.4h // P4[0], P4[1], P4[2], P4[3], P8[0]...
trn2 v3.4h, v5.4h, v7.4h // P4[4], P4[5], P4[6], P4[7], P8[4]...
trn2 v2.4h, v2.4h, v4.4h // P4[8], P4[9], P4[10], P4[11], P8[8]...
trn2 v4.4h, v6.4h, v17.4h // P4[12], P4[13], P4[14], P4[15], P8[12]...
ushll v5.8h, v31.8b, #1 // 2*P1[0..7]
ushll v6.8h, v19.8b, #1 // 2*P5[0..7]
trn1 v7.2s, v16.2s, v20.2s // P3[0..7]
uxtl v17.8h, v27.8b // P2[0..7]
trn2 v16.2s, v16.2s, v20.2s // P7[0..7]
uxtl v20.8h, v21.8b // P6[0..7]
trn1 v21.2s, v22.2s, v25.2s // P3[8..15]
ushll v24.8h, v24.8b, #1 // 2*P1[8..15]
trn2 v22.2s, v22.2s, v25.2s // P7[8..15]
ushll v25.8h, v23.8b, #1 // 2*P5[8..15]
trn1 v27.2s, v1.2s, v3.2s // P4[0..7]
uxtl v26.8h, v26.8b // P2[8..15]
mls v5.8h, v17.8h, v0.h[1] // 2*P1[0..7]-5*P2[0..7]
uxtl v17.8h, v18.8b // P6[8..15]
mls v6.8h, v20.8h, v0.h[1] // 2*P5[0..7]-5*P6[0..7]
trn1 v18.2s, v2.2s, v4.2s // P4[8..15]
uxtl v28.8h, v7.8b // P3[0..7]
mls v24.8h, v26.8h, v0.h[1] // 2*P1[8..15]-5*P2[8..15]
uxtl v16.8h, v16.8b // P7[0..7]
uxtl v26.8h, v21.8b // P3[8..15]
mls v25.8h, v17.8h, v0.h[1] // 2*P5[8..15]-5*P6[8..15]
uxtl v22.8h, v22.8b // P7[8..15]
ushll v7.8h, v7.8b, #1 // 2*P3[0..7]
uxtl v27.8h, v27.8b // P4[0..7]
trn2 v1.2s, v1.2s, v3.2s // P8[0..7]
ushll v3.8h, v21.8b, #1 // 2*P3[8..15]
trn2 v2.2s, v2.2s, v4.2s // P8[8..15]
uxtl v4.8h, v18.8b // P4[8..15]
mla v5.8h, v28.8h, v0.h[1] // 2*P1[0..7]-5*P2[0..7]+5*P3[0..7]
uxtl v1.8h, v1.8b // P8[0..7]
mla v6.8h, v16.8h, v0.h[1] // 2*P5[0..7]-5*P6[0..7]+5*P7[0..7]
uxtl v2.8h, v2.8b // P8[8..15]
uxtl v16.8h, v19.8b // P5[0..7]
mla v24.8h, v26.8h, v0.h[1] // 2*P1[8..15]-5*P2[8..15]+5*P3[8..15]
uxtl v18.8h, v23.8b // P5[8..15]
dup v19.8h, w2 // pq
mla v25.8h, v22.8h, v0.h[1] // 2*P5[8..15]-5*P6[8..15]+5*P7[8..15]
sub v21.8h, v27.8h, v16.8h // P4[0..7]-P5[0..7]
sub v22.8h, v4.8h, v18.8h // P4[8..15]-P5[8..15]
mls v7.8h, v27.8h, v0.h[1] // 2*P3[0..7]-5*P4[0..7]
abs v23.8h, v21.8h
mls v3.8h, v4.8h, v0.h[1] // 2*P3[8..15]-5*P4[8..15]
abs v26.8h, v22.8h
sshr v21.8h, v21.8h, #8 // clip_sign[0..7]
mls v5.8h, v27.8h, v0.h[0] // 2*P1[0..7]-5*P2[0..7]+5*P3[0..7]-2*P4[0..7]
sshr v23.8h, v23.8h, #1 // clip[0..7]
sshr v26.8h, v26.8h, #1 // clip[8..15]
mls v6.8h, v1.8h, v0.h[0] // 2*P5[0..7]-5*P6[0..7]+5*P7[0..7]-2*P8[0..7]
sshr v1.8h, v22.8h, #8 // clip_sign[8..15]
cmeq v22.8h, v23.8h, #0 // test clip[0..7] == 0
mls v24.8h, v4.8h, v0.h[0] // 2*P1[8..15]-5*P2[8..15]+5*P3[8..15]-2*P4[8..15]
cmeq v28.8h, v26.8h, #0 // test clip[8..15] == 0
srshr v5.8h, v5.8h, #3
mls v25.8h, v2.8h, v0.h[0] // 2*P5[8..15]-5*P6[8..15]+5*P7[8..15]-2*P8[8..15]
srshr v2.8h, v6.8h, #3
mla v7.8h, v16.8h, v0.h[1] // 2*P3[0..7]-5*P4[0..7]+5*P5[0..7]
srshr v6.8h, v24.8h, #3
mla v3.8h, v18.8h, v0.h[1] // 2*P3[8..15]-5*P4[8..15]+5*P5[8..15]
abs v5.8h, v5.8h // a1[0..7]
srshr v24.8h, v25.8h, #3
mls v3.8h, v17.8h, v0.h[0] // 2*P3[8..15]-5*P4[8..15]+5*P5[8..15]-2*P6[8..15]
abs v2.8h, v2.8h // a2[0..7]
abs v6.8h, v6.8h // a1[8..15]
mls v7.8h, v20.8h, v0.h[0] // 2*P3[0..7]-5*P4[0..7]+5*P5[0..7]-2*P6[0..7]
abs v17.8h, v24.8h // a2[8..15]
cmhs v20.8h, v5.8h, v2.8h // test a1[0..7] >= a2[0..7]
srshr v3.8h, v3.8h, #3
cmhs v24.8h, v6.8h, v17.8h // test a1[8..15] >= a2[8.15]
srshr v7.8h, v7.8h, #3
bsl v20.16b, v2.16b, v5.16b // a3[0..7]
abs v2.8h, v3.8h // a0[8..15]
sshr v3.8h, v3.8h, #8 // a0_sign[8..15]
bsl v24.16b, v17.16b, v6.16b // a3[8..15]
abs v5.8h, v7.8h // a0[0..7]
sshr v6.8h, v7.8h, #8 // a0_sign[0..7]
cmhs v7.8h, v2.8h, v19.8h // test a0[8..15] >= pq
sub v1.8h, v1.8h, v3.8h // clip_sign[8..15] - a0_sign[8..15]
uqsub v3.8h, v2.8h, v24.8h // a0[8..15] >= a3[8..15] ? a0[8..15]-a3[8..15] : 0 (a0 > a3 in all cases where filtering is enabled, so makes more sense to subtract this way round than the opposite and then taking the abs)
cmhs v2.8h, v24.8h, v2.8h // test a3[8..15] >= a0[8..15]
uqsub v17.8h, v5.8h, v20.8h // a0[0..7] >= a3[0..7] ? a0[0..7]-a3[0..7] : 0 (a0 > a3 in all cases where filtering is enabled, so makes more sense to subtract this way round than the opposite and then taking the abs)
cmhs v19.8h, v5.8h, v19.8h // test a0[0..7] >= pq
orr v7.16b, v28.16b, v7.16b // test clip[8..15] == 0 || a0[8..15] >= pq
sub v6.8h, v21.8h, v6.8h // clip_sign[0..7] - a0_sign[0..7]
mul v3.8h, v3.8h, v0.h[1] // a0[8..15] >= a3[8..15] ? 5*(a0[8..15]-a3[8..15]) : 0
cmhs v5.8h, v20.8h, v5.8h // test a3[0..7] >= a0[0..7]
orr v19.16b, v22.16b, v19.16b // test clip[0..7] == 0 || a0[0..7] >= pq
mul v0.8h, v17.8h, v0.h[1] // a0[0..7] >= a3[0..7] ? 5*(a0[0..7]-a3[0..7]) : 0
orr v2.16b, v7.16b, v2.16b // test clip[8..15] == 0 || a0[8..15] >= pq || a3[8..15] >= a0[8..15]
orr v5.16b, v19.16b, v5.16b // test clip[0..7] == 0 || a0[0..7] >= pq || a3[0..7] >= a0[0..7]
ushr v3.8h, v3.8h, #3 // a0[8..15] >= a3[8..15] ? (5*(a0[8..15]-a3[8..15]))>>3 : 0
mov w7, v2.s[1]
mov w8, v2.s[3]
ushr v0.8h, v0.8h, #3 // a0[0..7] >= a3[0..7] ? (5*(a0[0..7]-a3[0..7]))>>3 : 0
mov w2, v5.s[1] // move to gp reg
cmhs v2.8h, v3.8h, v26.8h
mov w3, v5.s[3]
cmhs v5.8h, v0.8h, v23.8h
bsl v2.16b, v26.16b, v3.16b // FFMIN(d[8..15], clip[8..15])
and w9, w7, w8
bsl v5.16b, v23.16b, v0.16b // FFMIN(d[0..7], clip[0..7])
and w10, w2, w3
bic v0.16b, v2.16b, v7.16b // set each d[8..15] to zero if it should not be filtered because clip[8..15] == 0 || a0[8..15] >= pq (a3 > a0 case already zeroed by saturating sub)
and w9, w10, w9
bic v2.16b, v5.16b, v19.16b // set each d[0..7] to zero if it should not be filtered because clip[0..7] == 0 || a0[0..7] >= pq (a3 > a0 case already zeroed by saturating sub)
mls v4.8h, v0.8h, v1.8h // invert d[8..15] depending on clip_sign[8..15] & a0_sign[8..15], or zero it if they match, and accumulate into P4
tbnz w9, #0, 4f // none of the 16 pixel pairs should be updated in this case
mls v27.8h, v2.8h, v6.8h // invert d[0..7] depending on clip_sign[0..7] & a0_sign[0..7], or zero it if they match, and accumulate into P4
mla v16.8h, v2.8h, v6.8h // invert d[0..7] depending on clip_sign[0..7] & a0_sign[0..7], or zero it if they match, and accumulate into P5
sqxtun v2.8b, v4.8h
mla v18.8h, v0.8h, v1.8h // invert d[8..15] depending on clip_sign[8..15] & a0_sign[8..15], or zero it if they match, and accumulate into P5
sqxtun v0.8b, v27.8h
sqxtun v1.8b, v16.8h
sqxtun v3.8b, v18.8h
tbnz w2, #0, 1f
st2 {v0.b, v1.b}[0], [x0], x1
st2 {v0.b, v1.b}[1], [x0], x1
st2 {v0.b, v1.b}[2], [x0], x1
st2 {v0.b, v1.b}[3], [x0]
1: tbnz w3, #0, 2f
st2 {v0.b, v1.b}[4], [x5], x1
st2 {v0.b, v1.b}[5], [x5], x1
st2 {v0.b, v1.b}[6], [x5], x1
st2 {v0.b, v1.b}[7], [x5]
2: tbnz w7, #0, 3f
st2 {v2.b, v3.b}[0], [x4], x1
st2 {v2.b, v3.b}[1], [x4], x1
st2 {v2.b, v3.b}[2], [x4], x1
st2 {v2.b, v3.b}[3], [x4]
3: tbnz w8, #0, 4f
st2 {v2.b, v3.b}[4], [x6], x1
st2 {v2.b, v3.b}[5], [x6], x1
st2 {v2.b, v3.b}[6], [x6], x1
st2 {v2.b, v3.b}[7], [x6]
4: ret
endfunc
// Copy at most the specified number of bytes from source to destination buffer,
// stopping at a multiple of 32 bytes, none of which are the start of an escape sequence
// On entry:
// x0 -> source buffer
// w1 = max number of bytes to copy
// x2 -> destination buffer, optimally 8-byte aligned
// On exit:
// w0 = number of bytes not copied
function ff_vc1_unescape_buffer_helper_neon, export=1
// Offset by 80 to screen out cases that are too short for us to handle,
// and also make it easy to test for loop termination, or to determine
// whether we need an odd number of half-iterations of the loop.
subs w1, w1, #80
b.mi 90f
// Set up useful constants
movi v20.4s, #3, lsl #24
movi v21.4s, #3, lsl #16
tst w1, #32
b.ne 1f
ld1 {v0.16b, v1.16b, v2.16b}, [x0], #48
ext v25.16b, v0.16b, v1.16b, #1
ext v26.16b, v0.16b, v1.16b, #2
ext v27.16b, v0.16b, v1.16b, #3
ext v29.16b, v1.16b, v2.16b, #1
ext v30.16b, v1.16b, v2.16b, #2
ext v31.16b, v1.16b, v2.16b, #3
bic v24.16b, v0.16b, v20.16b
bic v25.16b, v25.16b, v20.16b
bic v26.16b, v26.16b, v20.16b
bic v27.16b, v27.16b, v20.16b
bic v28.16b, v1.16b, v20.16b
bic v29.16b, v29.16b, v20.16b
bic v30.16b, v30.16b, v20.16b
bic v31.16b, v31.16b, v20.16b
eor v24.16b, v24.16b, v21.16b
eor v25.16b, v25.16b, v21.16b
eor v26.16b, v26.16b, v21.16b
eor v27.16b, v27.16b, v21.16b
eor v28.16b, v28.16b, v21.16b
eor v29.16b, v29.16b, v21.16b
eor v30.16b, v30.16b, v21.16b
eor v31.16b, v31.16b, v21.16b
cmeq v24.4s, v24.4s, #0
cmeq v25.4s, v25.4s, #0
cmeq v26.4s, v26.4s, #0
cmeq v27.4s, v27.4s, #0
add w1, w1, #32
b 3f
1: ld1 {v3.16b, v4.16b, v5.16b}, [x0], #48
ext v25.16b, v3.16b, v4.16b, #1
ext v26.16b, v3.16b, v4.16b, #2
ext v27.16b, v3.16b, v4.16b, #3
ext v29.16b, v4.16b, v5.16b, #1
ext v30.16b, v4.16b, v5.16b, #2
ext v31.16b, v4.16b, v5.16b, #3
bic v24.16b, v3.16b, v20.16b
bic v25.16b, v25.16b, v20.16b
bic v26.16b, v26.16b, v20.16b
bic v27.16b, v27.16b, v20.16b
bic v28.16b, v4.16b, v20.16b
bic v29.16b, v29.16b, v20.16b
bic v30.16b, v30.16b, v20.16b
bic v31.16b, v31.16b, v20.16b
eor v24.16b, v24.16b, v21.16b
eor v25.16b, v25.16b, v21.16b
eor v26.16b, v26.16b, v21.16b
eor v27.16b, v27.16b, v21.16b
eor v28.16b, v28.16b, v21.16b
eor v29.16b, v29.16b, v21.16b
eor v30.16b, v30.16b, v21.16b
eor v31.16b, v31.16b, v21.16b
cmeq v24.4s, v24.4s, #0
cmeq v25.4s, v25.4s, #0
cmeq v26.4s, v26.4s, #0
cmeq v27.4s, v27.4s, #0
// Drop through...
2: mov v0.16b, v5.16b
ld1 {v1.16b, v2.16b}, [x0], #32
cmeq v28.4s, v28.4s, #0
cmeq v29.4s, v29.4s, #0
cmeq v30.4s, v30.4s, #0
cmeq v31.4s, v31.4s, #0
orr v24.16b, v24.16b, v25.16b
orr v26.16b, v26.16b, v27.16b
orr v28.16b, v28.16b, v29.16b
orr v30.16b, v30.16b, v31.16b
ext v25.16b, v0.16b, v1.16b, #1
orr v22.16b, v24.16b, v26.16b
ext v26.16b, v0.16b, v1.16b, #2
ext v27.16b, v0.16b, v1.16b, #3
ext v29.16b, v1.16b, v2.16b, #1
orr v23.16b, v28.16b, v30.16b
ext v30.16b, v1.16b, v2.16b, #2
ext v31.16b, v1.16b, v2.16b, #3
bic v24.16b, v0.16b, v20.16b
bic v25.16b, v25.16b, v20.16b
bic v26.16b, v26.16b, v20.16b
orr v22.16b, v22.16b, v23.16b
bic v27.16b, v27.16b, v20.16b
bic v28.16b, v1.16b, v20.16b
bic v29.16b, v29.16b, v20.16b
bic v30.16b, v30.16b, v20.16b
bic v31.16b, v31.16b, v20.16b
addv s22, v22.4s
eor v24.16b, v24.16b, v21.16b
eor v25.16b, v25.16b, v21.16b
eor v26.16b, v26.16b, v21.16b
eor v27.16b, v27.16b, v21.16b
eor v28.16b, v28.16b, v21.16b
mov w3, v22.s[0]
eor v29.16b, v29.16b, v21.16b
eor v30.16b, v30.16b, v21.16b
eor v31.16b, v31.16b, v21.16b
cmeq v24.4s, v24.4s, #0
cmeq v25.4s, v25.4s, #0
cmeq v26.4s, v26.4s, #0
cmeq v27.4s, v27.4s, #0
cbnz w3, 90f
st1 {v3.16b, v4.16b}, [x2], #32
3: mov v3.16b, v2.16b
ld1 {v4.16b, v5.16b}, [x0], #32
cmeq v28.4s, v28.4s, #0
cmeq v29.4s, v29.4s, #0
cmeq v30.4s, v30.4s, #0
cmeq v31.4s, v31.4s, #0
orr v24.16b, v24.16b, v25.16b
orr v26.16b, v26.16b, v27.16b
orr v28.16b, v28.16b, v29.16b
orr v30.16b, v30.16b, v31.16b
ext v25.16b, v3.16b, v4.16b, #1
orr v22.16b, v24.16b, v26.16b
ext v26.16b, v3.16b, v4.16b, #2
ext v27.16b, v3.16b, v4.16b, #3
ext v29.16b, v4.16b, v5.16b, #1
orr v23.16b, v28.16b, v30.16b
ext v30.16b, v4.16b, v5.16b, #2
ext v31.16b, v4.16b, v5.16b, #3
bic v24.16b, v3.16b, v20.16b
bic v25.16b, v25.16b, v20.16b
bic v26.16b, v26.16b, v20.16b
orr v22.16b, v22.16b, v23.16b
bic v27.16b, v27.16b, v20.16b
bic v28.16b, v4.16b, v20.16b
bic v29.16b, v29.16b, v20.16b
bic v30.16b, v30.16b, v20.16b
bic v31.16b, v31.16b, v20.16b
addv s22, v22.4s
eor v24.16b, v24.16b, v21.16b
eor v25.16b, v25.16b, v21.16b
eor v26.16b, v26.16b, v21.16b
eor v27.16b, v27.16b, v21.16b
eor v28.16b, v28.16b, v21.16b
mov w3, v22.s[0]
eor v29.16b, v29.16b, v21.16b
eor v30.16b, v30.16b, v21.16b
eor v31.16b, v31.16b, v21.16b
cmeq v24.4s, v24.4s, #0
cmeq v25.4s, v25.4s, #0
cmeq v26.4s, v26.4s, #0
cmeq v27.4s, v27.4s, #0
cbnz w3, 91f
st1 {v0.16b, v1.16b}, [x2], #32
subs w1, w1, #64
b.pl 2b
90: add w0, w1, #80
ret
91: sub w1, w1, #32
b 90b
endfunc