1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
|
/*
* Copyright (c) 2023-2024 Nuo Mi <nuomi2021@gmail.com>
* Copyright (c) 2023-2024 Wu Jianhua <toqsxw@outlook.com>
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License along
* with FFmpeg; if not, write to the Free Software Foundation, Inc.,
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
*/
#include <string.h>
#include "checkasm.h"
#include "libavcodec/vvc/ctu.h"
#include "libavcodec/vvc/data.h"
#include "libavcodec/vvc/dsp.h"
#include "libavutil/common.h"
#include "libavutil/intreadwrite.h"
#include "libavutil/mem_internal.h"
static const uint32_t pixel_mask[3] = { 0xffffffff, 0x03ff03ff, 0x0fff0fff };
#define SIZEOF_PIXEL ((bit_depth + 7) / 8)
#define SRC_PIXEL_STRIDE (MAX_CTU_SIZE + 2 * ALF_PADDING_SIZE)
#define DST_PIXEL_STRIDE (SRC_PIXEL_STRIDE + 4)
#define SRC_BUF_SIZE (SRC_PIXEL_STRIDE * (MAX_CTU_SIZE + 3 * 2) * 2) //+3 * 2 for top and bottom row, *2 for high bit depth
#define DST_BUF_SIZE (DST_PIXEL_STRIDE * (MAX_CTU_SIZE + 3 * 2) * 2)
#define LUMA_PARAMS_SIZE (MAX_CTU_SIZE * MAX_CTU_SIZE / ALF_BLOCK_SIZE / ALF_BLOCK_SIZE * ALF_NUM_COEFF_LUMA)
#define randomize_buffers(buf0, buf1, size) \
do { \
uint32_t mask = pixel_mask[(bit_depth - 8) >> 1]; \
int k; \
for (k = 0; k < size; k += 4) { \
uint32_t r = rnd() & mask; \
AV_WN32A(buf0 + k, r); \
AV_WN32A(buf1 + k, r); \
} \
} while (0)
#define randomize_buffers2(buf, size, filter) \
do { \
int k; \
if (filter) { \
for (k = 0; k < size; k++) { \
int8_t r = rnd(); \
buf[k] = r; \
} \
} else { \
for (k = 0; k < size; k++) { \
int r = rnd() % FF_ARRAY_ELEMS(clip_set); \
buf[k] = clip_set[r]; \
} \
} \
} while (0)
static int get_alf_vb_pos(const int h, const int vb_pos_above)
{
if (h == MAX_CTU_SIZE)
return MAX_CTU_SIZE - vb_pos_above;
// If h < MAX_CTU_SIZE and picture virtual boundaries are involved, ALF virtual boundaries can either be within or outside this ALF block.
return ((rnd() & 1) ? h : MAX_CTU_SIZE) - vb_pos_above;
}
static void check_alf_filter(VVCDSPContext *c, const int bit_depth)
{
LOCAL_ALIGNED_32(uint8_t, dst0, [DST_BUF_SIZE]);
LOCAL_ALIGNED_32(uint8_t, dst1, [DST_BUF_SIZE]);
LOCAL_ALIGNED_32(uint8_t, src0, [SRC_BUF_SIZE]);
LOCAL_ALIGNED_32(uint8_t, src1, [SRC_BUF_SIZE]);
int16_t filter[LUMA_PARAMS_SIZE];
int16_t clip[LUMA_PARAMS_SIZE];
const int16_t clip_set[] = {
1 << bit_depth, 1 << (bit_depth - 3), 1 << (bit_depth - 5), 1 << (bit_depth - 7)
};
ptrdiff_t src_stride = SRC_PIXEL_STRIDE * SIZEOF_PIXEL;
ptrdiff_t dst_stride = DST_PIXEL_STRIDE * SIZEOF_PIXEL;
int offset = (3 * SRC_PIXEL_STRIDE + 3) * SIZEOF_PIXEL;
declare_func(void, uint8_t *dst, ptrdiff_t dst_stride, const uint8_t *src, ptrdiff_t src_stride,
int width, int height, const int16_t *filter, const int16_t *clip, const int vb_pos);
randomize_buffers(src0, src1, SRC_BUF_SIZE);
randomize_buffers2(filter, LUMA_PARAMS_SIZE, 1);
randomize_buffers2(clip, LUMA_PARAMS_SIZE, 0);
for (int h = 4; h <= MAX_CTU_SIZE; h += 4) {
for (int w = 4; w <= MAX_CTU_SIZE; w += 4) {
//Both picture size and virtual boundaries are 8-aligned. For luma, we only need to check 8-aligned sizes.
if (!(w % 8) && !(h % 8)) {
if (check_func(c->alf.filter[LUMA], "vvc_alf_filter_luma_%dx%d_%d", w, h, bit_depth)) {
const int vb_pos = get_alf_vb_pos(h, ALF_VB_POS_ABOVE_LUMA);
memset(dst0, 0, DST_BUF_SIZE);
memset(dst1, 0, DST_BUF_SIZE);
call_ref(dst0, dst_stride, src0 + offset, src_stride, w, h, filter, clip, vb_pos);
call_new(dst1, dst_stride, src1 + offset, src_stride, w, h, filter, clip, vb_pos);
for (int i = 0; i < (h + 1); i++) {
if (memcmp(dst0 + i * dst_stride, dst1 + i * dst_stride, (w + 1) * SIZEOF_PIXEL))
fail();
}
// Bench only square sizes, and ones with dimensions being a power of two.
if (w == h && (w & (w - 1)) == 0)
bench_new(dst1, dst_stride, src1 + offset, src_stride, w, h, filter, clip, vb_pos);
}
}
//For chroma, once it exceeds 64, it's not a 4:2:0 format, so we only need to check 8-aligned sizes as well.
if ((w <= 64 || !(w % 8)) && (h <= 64 || !(h % 8))) {
if (check_func(c->alf.filter[CHROMA], "vvc_alf_filter_chroma_%dx%d_%d", w, h, bit_depth)) {
const int vb_pos = get_alf_vb_pos(h, ALF_VB_POS_ABOVE_CHROMA);
memset(dst0, 0, DST_BUF_SIZE);
memset(dst1, 0, DST_BUF_SIZE);
call_ref(dst0, dst_stride, src0 + offset, src_stride, w, h, filter, clip, vb_pos);
call_new(dst1, dst_stride, src1 + offset, src_stride, w, h, filter, clip, vb_pos);
for (int i = 0; i < (h + 1); i++) {
if (memcmp(dst0 + i * dst_stride, dst1 + i * dst_stride, (w + 1) * SIZEOF_PIXEL))
fail();
}
if (w == h && (w & (w - 1)) == 0)
bench_new(dst1, dst_stride, src1 + offset, src_stride, w, h, filter, clip, vb_pos);
}
}
}
}
}
static void check_alf_classify(VVCDSPContext *c, const int bit_depth)
{
LOCAL_ALIGNED_32(int, class_idx0, [SRC_BUF_SIZE]);
LOCAL_ALIGNED_32(int, transpose_idx0, [SRC_BUF_SIZE]);
LOCAL_ALIGNED_32(int, class_idx1, [SRC_BUF_SIZE]);
LOCAL_ALIGNED_32(int, transpose_idx1, [SRC_BUF_SIZE]);
LOCAL_ALIGNED_32(uint8_t, src0, [SRC_BUF_SIZE]);
LOCAL_ALIGNED_32(uint8_t, src1, [SRC_BUF_SIZE]);
LOCAL_ALIGNED_32(int32_t, alf_gradient_tmp, [ALF_GRADIENT_SIZE * ALF_GRADIENT_SIZE * ALF_NUM_DIR]);
ptrdiff_t stride = SRC_PIXEL_STRIDE * SIZEOF_PIXEL;
int offset = (3 * SRC_PIXEL_STRIDE + 3) * SIZEOF_PIXEL;
declare_func(void, int *class_idx, int *transpose_idx,
const uint8_t *src, ptrdiff_t src_stride, int width, int height, int vb_pos, int *gradient_tmp);
randomize_buffers(src0, src1, SRC_BUF_SIZE);
//Both picture size and virtual boundaries are 8-aligned. Classify is luma only, we only need to check 8-aligned sizes.
for (int h = 8; h <= MAX_CTU_SIZE; h += 8) {
for (int w = 8; w <= MAX_CTU_SIZE; w += 8) {
const int id_size = w * h / ALF_BLOCK_SIZE / ALF_BLOCK_SIZE * sizeof(int);
const int vb_pos = get_alf_vb_pos(h, ALF_VB_POS_ABOVE_LUMA);
if (check_func(c->alf.classify, "vvc_alf_classify_%dx%d_%d", w, h, bit_depth)) {
memset(class_idx0, 0, id_size);
memset(class_idx1, 0, id_size);
memset(transpose_idx0, 0, id_size);
memset(transpose_idx1, 0, id_size);
call_ref(class_idx0, transpose_idx0, src0 + offset, stride, w, h, vb_pos, alf_gradient_tmp);
call_new(class_idx1, transpose_idx1, src1 + offset, stride, w, h, vb_pos, alf_gradient_tmp);
if (memcmp(class_idx0, class_idx1, id_size))
fail();
if (memcmp(transpose_idx0, transpose_idx1, id_size))
fail();
// Bench only square sizes, and ones with dimensions being a power of two.
if (w == h && (w & (w - 1)) == 0)
bench_new(class_idx1, transpose_idx1, src1 + offset, stride, w, h, vb_pos, alf_gradient_tmp);
}
}
}
}
void checkasm_check_vvc_alf(void)
{
int bit_depth;
VVCDSPContext h;
for (bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
ff_vvc_dsp_init(&h, bit_depth);
check_alf_filter(&h, bit_depth);
}
report("alf_filter");
for (bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
ff_vvc_dsp_init(&h, bit_depth);
check_alf_classify(&h, bit_depth);
}
report("alf_classify");
}
|