mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2024-12-27 18:02:11 +00:00
7175544c0b
apply_bdof_8_8x16_c: 5776.5 apply_bdof_8_8x16_avx2: 396.2 apply_bdof_8_16x8_c: 5722.0 apply_bdof_8_16x8_avx2: 216.0 apply_bdof_8_16x16_c: 11213.2 apply_bdof_8_16x16_avx2: 434.5 apply_bdof_10_8x16_c: 5657.7 apply_bdof_10_8x16_avx2: 1096.0 apply_bdof_10_16x8_c: 5531.7 apply_bdof_10_16x8_avx2: 212.5 apply_bdof_10_16x16_c: 11043.7 apply_bdof_10_16x16_avx2: 1252.7 apply_bdof_12_8x16_c: 5680.0 apply_bdof_12_8x16_avx2: 1096.5 apply_bdof_12_16x8_c: 5646.2 apply_bdof_12_16x8_avx2: 624.5 apply_bdof_12_16x16_c: 11076.0 apply_bdof_12_16x16_avx2: 1241.5
482 lines
22 KiB
C
482 lines
22 KiB
C
/*
|
|
* Copyright (c) 2023-2024 Nuo Mi
|
|
* Copyright (c) 2023-2024 Wu Jianhua
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with FFmpeg; if not, write to the Free Software Foundation, Inc.,
|
|
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*/
|
|
|
|
#include <string.h>
|
|
|
|
#include "checkasm.h"
|
|
#include "libavcodec/vvc/ctu.h"
|
|
#include "libavcodec/vvc/data.h"
|
|
#include "libavcodec/vvc/dsp.h"
|
|
|
|
#include "libavutil/common.h"
|
|
#include "libavutil/intreadwrite.h"
|
|
#include "libavutil/mem_internal.h"
|
|
|
|
static const uint32_t pixel_mask[] = { 0xffffffff, 0x03ff03ff, 0x0fff0fff, 0x3fff3fff, 0xffffffff };
|
|
static const int sizes[] = { 2, 4, 8, 16, 32, 64, 128 };
|
|
|
|
#define SIZEOF_PIXEL ((bit_depth + 7) / 8)
|
|
#define PIXEL_STRIDE (MAX_CTU_SIZE * 2)
|
|
#define EXTRA_BEFORE 3
|
|
#define EXTRA_AFTER 4
|
|
#define SRC_EXTRA (EXTRA_BEFORE + EXTRA_AFTER) * 2
|
|
#define SRC_BUF_SIZE (PIXEL_STRIDE + SRC_EXTRA) * (PIXEL_STRIDE + SRC_EXTRA)
|
|
#define DST_BUF_SIZE (MAX_CTU_SIZE * MAX_CTU_SIZE * 2)
|
|
#define SRC_OFFSET ((PIXEL_STRIDE + EXTRA_BEFORE * 2) * EXTRA_BEFORE)
|
|
|
|
#define randomize_buffers(buf0, buf1, size, mask) \
|
|
do { \
|
|
int k; \
|
|
for (k = 0; k < size; k += 4 / sizeof(*buf0)) { \
|
|
uint32_t r = rnd() & mask; \
|
|
AV_WN32A(buf0 + k, r); \
|
|
AV_WN32A(buf1 + k, r); \
|
|
} \
|
|
} while (0)
|
|
|
|
#define randomize_pixels(buf0, buf1, size) \
|
|
do { \
|
|
uint32_t mask = pixel_mask[(bit_depth - 8) >> 1]; \
|
|
randomize_buffers(buf0, buf1, size, mask); \
|
|
} while (0)
|
|
|
|
#define randomize_avg_src(buf0, buf1, size) \
|
|
do { \
|
|
uint32_t mask = 0x3fff3fff; \
|
|
randomize_buffers(buf0, buf1, size, mask); \
|
|
} while (0)
|
|
|
|
#define randomize_prof_src(buf0, buf1, size) \
|
|
do { \
|
|
const int shift = 14 - bit_depth; \
|
|
const int mask16 = 0x3fff >> shift << shift; \
|
|
uint32_t mask = (mask16 << 16) | mask16; \
|
|
randomize_buffers(buf0, buf1, size, mask); \
|
|
} while (0)
|
|
|
|
static void check_put_vvc_luma(void)
|
|
{
|
|
LOCAL_ALIGNED_32(int16_t, dst0, [DST_BUF_SIZE / 2]);
|
|
LOCAL_ALIGNED_32(int16_t, dst1, [DST_BUF_SIZE / 2]);
|
|
LOCAL_ALIGNED_32(uint8_t, src0, [SRC_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, src1, [SRC_BUF_SIZE]);
|
|
VVCDSPContext c;
|
|
|
|
declare_func(void, int16_t *dst, const uint8_t *src, const ptrdiff_t src_stride,
|
|
const int height, const int8_t *hf, const int8_t *vf, const int width);
|
|
|
|
for (int bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
|
|
randomize_pixels(src0, src1, SRC_BUF_SIZE);
|
|
ff_vvc_dsp_init(&c, bit_depth);
|
|
for (int i = 0; i < 2; i++) {
|
|
for (int j = 0; j < 2; j++) {
|
|
for (int h = 4; h <= MAX_CTU_SIZE; h *= 2) {
|
|
for (int w = 4; w <= MAX_CTU_SIZE; w *= 2) {
|
|
const int idx = av_log2(w) - 1;
|
|
const int mx = rnd() % 16;
|
|
const int my = rnd() % 16;
|
|
const int8_t *hf = ff_vvc_inter_luma_filters[rnd() % 3][mx];
|
|
const int8_t *vf = ff_vvc_inter_luma_filters[rnd() % 3][my];
|
|
const char *type;
|
|
switch ((j << 1) | i) {
|
|
case 0: type = "put_luma_pixels"; break; // 0 0
|
|
case 1: type = "put_luma_h"; break; // 0 1
|
|
case 2: type = "put_luma_v"; break; // 1 0
|
|
case 3: type = "put_luma_hv"; break; // 1 1
|
|
}
|
|
if (check_func(c.inter.put[LUMA][idx][j][i], "%s_%d_%dx%d", type, bit_depth, w, h)) {
|
|
memset(dst0, 0, DST_BUF_SIZE);
|
|
memset(dst1, 0, DST_BUF_SIZE);
|
|
call_ref(dst0, src0 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
call_new(dst1, src1 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
if (memcmp(dst0, dst1, DST_BUF_SIZE))
|
|
fail();
|
|
if (w == h)
|
|
bench_new(dst1, src1 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
report("put_luma");
|
|
}
|
|
|
|
static void check_put_vvc_luma_uni(void)
|
|
{
|
|
LOCAL_ALIGNED_32(uint8_t, dst0, [DST_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, dst1, [DST_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, src0, [SRC_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, src1, [SRC_BUF_SIZE]);
|
|
|
|
VVCDSPContext c;
|
|
declare_func(void, uint8_t *dst, ptrdiff_t dststride,
|
|
const uint8_t *src, ptrdiff_t srcstride, int height,
|
|
const int8_t *hf, const int8_t *vf, int width);
|
|
|
|
for (int bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
|
|
ff_vvc_dsp_init(&c, bit_depth);
|
|
randomize_pixels(src0, src1, SRC_BUF_SIZE);
|
|
for (int i = 0; i < 2; i++) {
|
|
for (int j = 0; j < 2; j++) {
|
|
for (int h = 4; h <= MAX_CTU_SIZE; h *= 2) {
|
|
for (int w = 4; w <= MAX_CTU_SIZE; w *= 2) {
|
|
const int idx = av_log2(w) - 1;
|
|
const int mx = rnd() % VVC_INTER_LUMA_FACTS;
|
|
const int my = rnd() % VVC_INTER_LUMA_FACTS;
|
|
const int8_t *hf = ff_vvc_inter_luma_filters[rnd() % VVC_INTER_LUMA_FILTER_TYPES][mx];
|
|
const int8_t *vf = ff_vvc_inter_luma_filters[rnd() % VVC_INTER_LUMA_FILTER_TYPES][my];
|
|
const char *type;
|
|
|
|
switch ((j << 1) | i) {
|
|
case 0: type = "put_uni_pixels"; break; // 0 0
|
|
case 1: type = "put_uni_h"; break; // 0 1
|
|
case 2: type = "put_uni_v"; break; // 1 0
|
|
case 3: type = "put_uni_hv"; break; // 1 1
|
|
}
|
|
|
|
if (check_func(c.inter.put_uni[LUMA][idx][j][i], "%s_luma_%d_%dx%d", type, bit_depth, w, h)) {
|
|
memset(dst0, 0, DST_BUF_SIZE);
|
|
memset(dst1, 0, DST_BUF_SIZE);
|
|
call_ref(dst0, PIXEL_STRIDE, src0 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
call_new(dst1, PIXEL_STRIDE, src1 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
if (memcmp(dst0, dst1, DST_BUF_SIZE))
|
|
fail();
|
|
if (w == h)
|
|
bench_new(dst1, PIXEL_STRIDE, src1 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
report("put_uni_luma");
|
|
}
|
|
|
|
static void check_put_vvc_chroma(void)
|
|
{
|
|
LOCAL_ALIGNED_32(int16_t, dst0, [DST_BUF_SIZE / 2]);
|
|
LOCAL_ALIGNED_32(int16_t, dst1, [DST_BUF_SIZE / 2]);
|
|
LOCAL_ALIGNED_32(uint8_t, src0, [SRC_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, src1, [SRC_BUF_SIZE]);
|
|
VVCDSPContext c;
|
|
|
|
declare_func(void, int16_t *dst, const uint8_t *src, const ptrdiff_t src_stride,
|
|
const int height, const int8_t *hf, const int8_t *vf, const int width);
|
|
|
|
for (int bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
|
|
randomize_pixels(src0, src1, SRC_BUF_SIZE);
|
|
ff_vvc_dsp_init(&c, bit_depth);
|
|
for (int i = 0; i < 2; i++) {
|
|
for (int j = 0; j < 2; j++) {
|
|
for (int h = 2; h <= MAX_CTU_SIZE; h *= 2) {
|
|
for (int w = 2; w <= MAX_CTU_SIZE; w *= 2) {
|
|
const int idx = av_log2(w) - 1;
|
|
const int mx = rnd() % VVC_INTER_CHROMA_FACTS;
|
|
const int my = rnd() % VVC_INTER_CHROMA_FACTS;
|
|
const int8_t *hf = ff_vvc_inter_chroma_filters[rnd() % VVC_INTER_CHROMA_FILTER_TYPES][mx];
|
|
const int8_t *vf = ff_vvc_inter_chroma_filters[rnd() % VVC_INTER_CHROMA_FILTER_TYPES][my];
|
|
const char *type;
|
|
switch ((j << 1) | i) {
|
|
case 0: type = "put_chroma_pixels"; break; // 0 0
|
|
case 1: type = "put_chroma_h"; break; // 0 1
|
|
case 2: type = "put_chroma_v"; break; // 1 0
|
|
case 3: type = "put_chroma_hv"; break; // 1 1
|
|
}
|
|
if (check_func(c.inter.put[CHROMA][idx][j][i], "%s_%d_%dx%d", type, bit_depth, w, h)) {
|
|
memset(dst0, 0, DST_BUF_SIZE);
|
|
memset(dst1, 0, DST_BUF_SIZE);
|
|
call_ref(dst0, src0 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
call_new(dst1, src1 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
if (memcmp(dst0, dst1, DST_BUF_SIZE))
|
|
fail();
|
|
if (w == h)
|
|
bench_new(dst1, src1 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
report("put_chroma");
|
|
}
|
|
|
|
static void check_put_vvc_chroma_uni(void)
|
|
{
|
|
LOCAL_ALIGNED_32(uint8_t, dst0, [DST_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, dst1, [DST_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, src0, [SRC_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, src1, [SRC_BUF_SIZE]);
|
|
|
|
VVCDSPContext c;
|
|
declare_func(void, uint8_t *dst, ptrdiff_t dststride,
|
|
const uint8_t *src, ptrdiff_t srcstride, int height,
|
|
const int8_t *hf, const int8_t *vf, int width);
|
|
|
|
for (int bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
|
|
ff_vvc_dsp_init(&c, bit_depth);
|
|
randomize_pixels(src0, src1, SRC_BUF_SIZE);
|
|
for (int i = 0; i < 2; i++) {
|
|
for (int j = 0; j < 2; j++) {
|
|
for (int h = 4; h <= MAX_CTU_SIZE; h *= 2) {
|
|
for (int w = 4; w <= MAX_CTU_SIZE; w *= 2) {
|
|
const int idx = av_log2(w) - 1;
|
|
const int mx = rnd() % VVC_INTER_CHROMA_FACTS;
|
|
const int my = rnd() % VVC_INTER_CHROMA_FACTS;
|
|
const int8_t *hf = ff_vvc_inter_chroma_filters[rnd() % VVC_INTER_CHROMA_FILTER_TYPES][mx];
|
|
const int8_t *vf = ff_vvc_inter_chroma_filters[rnd() % VVC_INTER_CHROMA_FILTER_TYPES][my];
|
|
const char *type;
|
|
|
|
switch ((j << 1) | i) {
|
|
case 0: type = "put_uni_pixels"; break; // 0 0
|
|
case 1: type = "put_uni_h"; break; // 0 1
|
|
case 2: type = "put_uni_v"; break; // 1 0
|
|
case 3: type = "put_uni_hv"; break; // 1 1
|
|
}
|
|
|
|
if (check_func(c.inter.put_uni[CHROMA][idx][j][i], "%s_chroma_%d_%dx%d", type, bit_depth, w, h)) {
|
|
memset(dst0, 0, DST_BUF_SIZE);
|
|
memset(dst1, 0, DST_BUF_SIZE);
|
|
call_ref(dst0, PIXEL_STRIDE, src0 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
call_new(dst1, PIXEL_STRIDE, src1 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
if (memcmp(dst0, dst1, DST_BUF_SIZE))
|
|
fail();
|
|
if (w == h)
|
|
bench_new(dst1, PIXEL_STRIDE, src1 + SRC_OFFSET, PIXEL_STRIDE, h, hf, vf, w);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
report("put_uni_chroma");
|
|
}
|
|
|
|
#define AVG_SRC_BUF_SIZE (MAX_CTU_SIZE * MAX_CTU_SIZE)
|
|
#define AVG_DST_BUF_SIZE (MAX_PB_SIZE * MAX_PB_SIZE * 2)
|
|
|
|
static void check_avg(void)
|
|
{
|
|
LOCAL_ALIGNED_32(int16_t, src00, [AVG_SRC_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(int16_t, src01, [AVG_SRC_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(int16_t, src10, [AVG_SRC_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(int16_t, src11, [AVG_SRC_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, dst0, [AVG_DST_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, dst1, [AVG_DST_BUF_SIZE]);
|
|
VVCDSPContext c;
|
|
|
|
for (int bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
|
|
randomize_avg_src((uint8_t*)src00, (uint8_t*)src10, AVG_SRC_BUF_SIZE * sizeof(int16_t));
|
|
randomize_avg_src((uint8_t*)src01, (uint8_t*)src11, AVG_SRC_BUF_SIZE * sizeof(int16_t));
|
|
ff_vvc_dsp_init(&c, bit_depth);
|
|
for (int h = 2; h <= MAX_CTU_SIZE; h *= 2) {
|
|
for (int w = 2; w <= MAX_CTU_SIZE; w *= 2) {
|
|
{
|
|
declare_func(void, uint8_t *dst, ptrdiff_t dst_stride,
|
|
const int16_t *src0, const int16_t *src1, int width, int height);
|
|
if (check_func(c.inter.avg, "avg_%d_%dx%d", bit_depth, w, h)) {
|
|
memset(dst0, 0, AVG_DST_BUF_SIZE);
|
|
memset(dst1, 0, AVG_DST_BUF_SIZE);
|
|
call_ref(dst0, MAX_CTU_SIZE * SIZEOF_PIXEL, src00, src01, w, h);
|
|
call_new(dst1, MAX_CTU_SIZE * SIZEOF_PIXEL, src10, src11, w, h);
|
|
if (memcmp(dst0, dst1, DST_BUF_SIZE))
|
|
fail();
|
|
if (w == h)
|
|
bench_new(dst0, MAX_CTU_SIZE * SIZEOF_PIXEL, src00, src01, w, h);
|
|
}
|
|
}
|
|
{
|
|
declare_func(void, uint8_t *dst, ptrdiff_t dst_stride,
|
|
const int16_t *src0, const int16_t *src1, int width, int height,
|
|
int denom, int w0, int w1, int o0, int o1);
|
|
{
|
|
const int denom = rnd() % 8;
|
|
const int w0 = rnd() % 256 - 128;
|
|
const int w1 = rnd() % 256 - 128;
|
|
const int o0 = rnd() % 256 - 128;
|
|
const int o1 = rnd() % 256 - 128;
|
|
if (check_func(c.inter.w_avg, "w_avg_%d_%dx%d", bit_depth, w, h)) {
|
|
memset(dst0, 0, AVG_DST_BUF_SIZE);
|
|
memset(dst1, 0, AVG_DST_BUF_SIZE);
|
|
|
|
call_ref(dst0, MAX_CTU_SIZE * SIZEOF_PIXEL, src00, src01, w, h, denom, w0, w1, o0, o1);
|
|
call_new(dst1, MAX_CTU_SIZE * SIZEOF_PIXEL, src10, src11, w, h, denom, w0, w1, o0, o1);
|
|
if (memcmp(dst0, dst1, DST_BUF_SIZE))
|
|
fail();
|
|
if (w == h)
|
|
bench_new(dst0, MAX_CTU_SIZE * SIZEOF_PIXEL, src00, src01, w, h, denom, w0, w1, o0, o1);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
report("avg");
|
|
}
|
|
|
|
#define SR_RANGE 2
|
|
static void check_dmvr(void)
|
|
{
|
|
LOCAL_ALIGNED_32(uint16_t, dst0, [DST_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint16_t, dst1, [DST_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, src0, [SRC_BUF_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, src1, [SRC_BUF_SIZE]);
|
|
const int dst_stride = MAX_PB_SIZE * sizeof(int16_t);
|
|
|
|
VVCDSPContext c;
|
|
declare_func(void, int16_t *dst, const uint8_t *src, ptrdiff_t src_stride, int height,
|
|
intptr_t mx, intptr_t my, int width);
|
|
|
|
for (int bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
|
|
ff_vvc_dsp_init(&c, bit_depth);
|
|
randomize_pixels(src0, src1, SRC_BUF_SIZE);
|
|
for (int i = 0; i < 2; i++) {
|
|
for (int j = 0; j < 2; j++) {
|
|
for (int h = 8; h <= 16; h *= 2) {
|
|
for (int w = 8; w <= 16; w *= 2) {
|
|
const int pred_w = w + 2 * SR_RANGE;
|
|
const int pred_h = h + 2 * SR_RANGE;
|
|
const int mx = rnd() % VVC_INTER_LUMA_DMVR_FACTS;
|
|
const int my = rnd() % VVC_INTER_LUMA_DMVR_FACTS;
|
|
const char *type;
|
|
|
|
if (w * h < 128)
|
|
continue;
|
|
|
|
switch ((j << 1) | i) {
|
|
case 0: type = "dmvr"; break; // 0 0
|
|
case 1: type = "dmvr_h"; break; // 0 1
|
|
case 2: type = "dmvr_v"; break; // 1 0
|
|
case 3: type = "dmvr_hv"; break; // 1 1
|
|
}
|
|
|
|
if (check_func(c.inter.dmvr[j][i], "%s_%d_%dx%d", type, bit_depth, pred_w, pred_h)) {
|
|
memset(dst0, 0, DST_BUF_SIZE);
|
|
memset(dst1, 0, DST_BUF_SIZE);
|
|
call_ref(dst0, src0 + SRC_OFFSET, PIXEL_STRIDE, pred_h, mx, my, pred_w);
|
|
call_new(dst1, src1 + SRC_OFFSET, PIXEL_STRIDE, pred_h, mx, my, pred_w);
|
|
for (int k = 0; k < pred_h; k++) {
|
|
if (memcmp(dst0 + k * dst_stride, dst1 + k * dst_stride, pred_w * sizeof(int16_t))) {
|
|
fail();
|
|
break;
|
|
}
|
|
}
|
|
|
|
bench_new(dst1, src1 + SRC_OFFSET, PIXEL_STRIDE, pred_h, mx, my, pred_w);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
report("dmvr");
|
|
}
|
|
|
|
#define BDOF_BLOCK_SIZE 16
|
|
#define BDOF_SRC_SIZE (MAX_PB_SIZE* (BDOF_BLOCK_SIZE + 2))
|
|
#define BDOF_SRC_OFFSET (MAX_PB_SIZE + 1)
|
|
#define BDOF_DST_SIZE (BDOF_BLOCK_SIZE * BDOF_BLOCK_SIZE * 2)
|
|
static void check_bdof(void)
|
|
{
|
|
LOCAL_ALIGNED_32(uint8_t, dst0, [BDOF_DST_SIZE]);
|
|
LOCAL_ALIGNED_32(uint8_t, dst1, [BDOF_DST_SIZE]);
|
|
LOCAL_ALIGNED_32(uint16_t, src00, [BDOF_SRC_SIZE]);
|
|
LOCAL_ALIGNED_32(uint16_t, src01, [BDOF_SRC_SIZE]);
|
|
LOCAL_ALIGNED_32(uint16_t, src10, [BDOF_SRC_SIZE]);
|
|
LOCAL_ALIGNED_32(uint16_t, src11, [BDOF_SRC_SIZE]);
|
|
|
|
VVCDSPContext c;
|
|
declare_func(void, uint8_t *dst, ptrdiff_t dst_stride, const int16_t *src0, const int16_t *src1, int block_w, int block_h);
|
|
|
|
for (int bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
|
|
const int dst_stride = BDOF_BLOCK_SIZE * SIZEOF_PIXEL;
|
|
|
|
ff_vvc_dsp_init(&c, bit_depth);
|
|
randomize_prof_src(src00, src10, BDOF_SRC_SIZE);
|
|
randomize_prof_src(src01, src11, BDOF_SRC_SIZE);
|
|
for (int h = 8; h <= 16; h *= 2) {
|
|
for (int w = 8; w <= 16; w *= 2) {
|
|
if (w * h < 128)
|
|
continue;
|
|
if (check_func(c.inter.apply_bdof, "apply_bdof_%d_%dx%d", bit_depth, w, h)) {
|
|
memset(dst0, 0, BDOF_DST_SIZE);
|
|
memset(dst1, 0, BDOF_DST_SIZE);
|
|
call_ref(dst0, dst_stride, src00 + BDOF_SRC_OFFSET, src01 + BDOF_SRC_OFFSET, w, h);
|
|
call_new(dst1, dst_stride, src10 + BDOF_SRC_OFFSET, src11 + BDOF_SRC_OFFSET, w, h);
|
|
if (memcmp(dst0, dst1, BDOF_DST_SIZE))
|
|
fail();
|
|
bench_new(dst0, dst_stride, src00 + BDOF_SRC_OFFSET, src01 + BDOF_SRC_OFFSET, w, h);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
report("apply_bdof");
|
|
}
|
|
|
|
static void check_vvc_sad(void)
|
|
{
|
|
const int bit_depth = 10;
|
|
VVCDSPContext c;
|
|
LOCAL_ALIGNED_32(uint16_t, src0, [MAX_CTU_SIZE * MAX_CTU_SIZE * 4]);
|
|
LOCAL_ALIGNED_32(uint16_t, src1, [MAX_CTU_SIZE * MAX_CTU_SIZE * 4]);
|
|
declare_func(int, const int16_t *src0, const int16_t *src1, int dx, int dy, int block_w, int block_h);
|
|
|
|
ff_vvc_dsp_init(&c, bit_depth);
|
|
randomize_pixels(src0, src1, MAX_CTU_SIZE * MAX_CTU_SIZE * 4);
|
|
for (int h = 8; h <= 16; h *= 2) {
|
|
for (int w = 8; w <= 16; w *= 2) {
|
|
for(int offy = 0; offy <= 4; offy++) {
|
|
for(int offx = 0; offx <= 4; offx++) {
|
|
if (w * h < 128)
|
|
continue;
|
|
|
|
if (check_func(c.inter.sad, "sad_%dx%d", w, h)) {
|
|
int result0;
|
|
int result1;
|
|
|
|
result0 = call_ref(src0 + PIXEL_STRIDE * 2 + 2, src1 + PIXEL_STRIDE * 2 + 2, offx, offy, w, h);
|
|
result1 = call_new(src0 + PIXEL_STRIDE * 2 + 2, src1 + PIXEL_STRIDE * 2 + 2, offx, offy, w, h);
|
|
|
|
if (result1 != result0)
|
|
fail();
|
|
if(offx == 0 && offy == 0)
|
|
bench_new(src0 + PIXEL_STRIDE * 2 + 2, src1 + PIXEL_STRIDE * 2 + 2, offx, offy, w, h);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
report("sad");
|
|
}
|
|
|
|
void checkasm_check_vvc_mc(void)
|
|
{
|
|
check_dmvr();
|
|
check_bdof();
|
|
check_vvc_sad();
|
|
check_put_vvc_luma();
|
|
check_put_vvc_luma_uni();
|
|
check_put_vvc_chroma();
|
|
check_put_vvc_chroma_uni();
|
|
check_avg();
|
|
}
|