diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 19:33:14 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 19:33:14 +0000 |
commit | 36d22d82aa202bb199967e9512281e9a53db42c9 (patch) | |
tree | 105e8c98ddea1c1e4784a60a5a6410fa416be2de /third_party/dav1d/tests/checkasm/refmvs.c | |
parent | Initial commit. (diff) | |
download | firefox-esr-upstream.tar.xz firefox-esr-upstream.zip |
Adding upstream version 115.7.0esr.upstream/115.7.0esrupstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to '')
-rw-r--r-- | third_party/dav1d/tests/checkasm/refmvs.c | 167 |
1 files changed, 167 insertions, 0 deletions
diff --git a/third_party/dav1d/tests/checkasm/refmvs.c b/third_party/dav1d/tests/checkasm/refmvs.c new file mode 100644 index 0000000000..f21c81f85a --- /dev/null +++ b/third_party/dav1d/tests/checkasm/refmvs.c @@ -0,0 +1,167 @@ +/* + * Copyright © 2021, VideoLAN and dav1d authors + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, this + * list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above copyright notice, + * this list of conditions and the following disclaimer in the documentation + * and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND + * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR + * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES + * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; + * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND + * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS + * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include "tests/checkasm/checkasm.h" +#include "src/refmvs.h" + +#include <stdio.h> + +static inline int gen_mv(const int total_bits, int spel_bits) { + int bits = rnd() & ((1 << spel_bits) - 1); + do { + bits |= (rnd() & 1) << spel_bits; + } while (rnd() & 1 && ++spel_bits < total_bits); + // the do/while makes it relatively more likely to be close to zero (fpel) + // than far away + return rnd() & 1 ? -bits : bits; +} + +static void check_save_tmvs(const Dav1dRefmvsDSPContext *const c) { + refmvs_block *rr[31]; + refmvs_block r[31 * 256]; + ALIGN_STK_64(refmvs_temporal_block, c_rp, 128 * 16,); + ALIGN_STK_64(refmvs_temporal_block, a_rp, 128 * 16,); + uint8_t ref_sign[7]; + + for (int i = 0; i < 31; i++) + rr[i] = &r[i * 256]; + + declare_func(void, refmvs_temporal_block *rp, const ptrdiff_t stride, + refmvs_block *const *const rr, const uint8_t *const ref_sign, + int col_end8, int row_end8, int col_start8, int row_start8); + + if (check_func(c->save_tmvs, "save_tmvs")) { + const int row_start8 = rnd() & 7; + const int row_end8 = 8 + (rnd() & 7); + const int col_start8 = rnd() & 31; + const int col_end8 = 96 + (rnd() & 31); + + for (int i = 0; i < 7; i++) + ref_sign[i] = rnd() & 1; + + for (int i = row_start8; i < row_end8; i++) + for (int j = col_start8; j < col_end8;) { + int bs = rnd() % N_BS_SIZES; + while (j + ((dav1d_block_dimensions[bs][0] + 1) >> 1) > col_end8) + bs++; + rr[i * 2][j * 2 + 1] = (refmvs_block) { + .mv.mv[0].x = gen_mv(14, 10), + .mv.mv[0].y = gen_mv(14, 10), + .mv.mv[1].x = gen_mv(14, 10), + .mv.mv[1].y = gen_mv(14, 10), + .ref.ref = { (rnd() % 9) - 1, (rnd() % 9) - 1 }, + .bs = bs + }; + for (int k = 0; k < (dav1d_block_dimensions[bs][0] + 1) >> 1; k++, j++) { + c_rp[i * 128 + j].mv.n = 0xdeadbeef; + c_rp[i * 128 + j].ref = 0xdd; + } + } + + call_ref(c_rp + row_start8 * 128, 128, rr, ref_sign, + col_end8, row_end8, col_start8, row_start8); + call_new(a_rp + row_start8 * 128, 128, rr, ref_sign, + col_end8, row_end8, col_start8, row_start8); + for (int i = row_start8; i < row_end8; i++) + for (int j = col_start8; j < col_end8; j++) + if (c_rp[i * 128 + j].mv.n != a_rp[i * 128 + j].mv.n || + c_rp[i * 128 + j].ref != a_rp[i * 128 + j].ref) + { + if (fail()) { + fprintf(stderr, "[%d][%d] c_rp.mv.x = 0x%x a_rp.mv.x = 0x%x\n", + i, j, c_rp[i * 128 + j].mv.x, a_rp[i * 128 + j].mv.x); + fprintf(stderr, "[%d][%d] c_rp.mv.y = 0x%x a_rp.mv.y = 0x%x\n", + i, j, c_rp[i * 128 + j].mv.y, a_rp[i * 128 + j].mv.y); + fprintf(stderr, "[%d][%d] c_rp.ref = %u a_rp.ref = %u\n", + i, j, c_rp[i * 128 + j].ref, a_rp[i * 128 + j].ref); + } + } + + for (int bs = BS_4x4; bs < N_BS_SIZES; bs++) { + const int bw8 = (dav1d_block_dimensions[bs][0] + 1) >> 1; + for (int i = 0; i < 16; i++) + for (int j = 0; j < 128; j += bw8) { + rr[i * 2][j * 2 + 1].ref.ref[0] = (rnd() % 9) - 1; + rr[i * 2][j * 2 + 1].ref.ref[1] = (rnd() % 9) - 1; + rr[i * 2][j * 2 + 1].bs = bs; + } + bench_new(alternate(c_rp, a_rp), 128, rr, ref_sign, 128, 16, 0, 0); + } + } + + report("save_tmvs"); +} + +static void check_splat_mv(const Dav1dRefmvsDSPContext *const c) { + ALIGN_STK_64(refmvs_block, c_buf, 32 * 32,); + ALIGN_STK_64(refmvs_block, a_buf, 32 * 32,); + refmvs_block *c_dst[32]; + refmvs_block *a_dst[32]; + const size_t stride = 32 * sizeof(refmvs_block); + + for (int i = 0; i < 32; i++) { + c_dst[i] = c_buf + 32 * i; + a_dst[i] = a_buf + 32 * i; + } + + declare_func(void, refmvs_block **rr, const refmvs_block *rmv, + int bx4, int bw4, int bh4); + + for (int w = 1; w <= 32; w *= 2) { + if (check_func(c->splat_mv, "splat_mv_w%d", w)) { + const int h_min = imax(w / 4, 1); + const int h_max = imin(w * 4, 32); + const int w_uint32 = w * sizeof(refmvs_block) / sizeof(uint32_t); + for (int h = h_min; h <= h_max; h *= 2) { + const int offset = (int) ((unsigned) w * rnd()) & 31; + union { + refmvs_block rmv; + uint32_t u32[3]; + } ALIGN(tmp, 16); + tmp.u32[0] = rnd(); + tmp.u32[1] = rnd(); + tmp.u32[2] = rnd(); + + call_ref(c_dst, &tmp.rmv, offset, w, h); + call_new(a_dst, &tmp.rmv, offset, w, h); + checkasm_check(uint32_t, (uint32_t*)(c_buf + offset), stride, + (uint32_t*)(a_buf + offset), stride, + w_uint32, h, "dst"); + + bench_new(a_dst, &tmp.rmv, 0, w, h); + } + } + } + report("splat_mv"); +} + +void checkasm_check_refmvs(void) { + Dav1dRefmvsDSPContext c; + dav1d_refmvs_dsp_init(&c); + + check_save_tmvs(&c); + check_splat_mv(&c); +} |