/* * Copyright (c) 2020, Alliance for Open Media. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include #include "config/aom_dsp_rtcd.h" static INLINE void copy_128(const uint8_t *src, uint8_t *dst) { __m256i s[4]; s[0] = _mm256_loadu_si256((__m256i *)(src + 0 * 32)); s[1] = _mm256_loadu_si256((__m256i *)(src + 1 * 32)); s[2] = _mm256_loadu_si256((__m256i *)(src + 2 * 32)); s[3] = _mm256_loadu_si256((__m256i *)(src + 3 * 32)); _mm256_storeu_si256((__m256i *)(dst + 0 * 32), s[0]); _mm256_storeu_si256((__m256i *)(dst + 1 * 32), s[1]); _mm256_storeu_si256((__m256i *)(dst + 2 * 32), s[2]); _mm256_storeu_si256((__m256i *)(dst + 3 * 32), s[3]); } void aom_convolve_copy_avx2(const uint8_t *src, ptrdiff_t src_stride, uint8_t *dst, ptrdiff_t dst_stride, int w, int h) { if (w >= 16) { assert(!((intptr_t)dst % 16)); assert(!(dst_stride % 16)); } if (w == 2) { do { memmove(dst, src, 2 * sizeof(*src)); src += src_stride; dst += dst_stride; memmove(dst, src, 2 * sizeof(*src)); src += src_stride; dst += dst_stride; h -= 2; } while (h); } else if (w == 4) { do { memmove(dst, src, 4 * sizeof(*src)); src += src_stride; dst += dst_stride; memmove(dst, src, 4 * sizeof(*src)); src += src_stride; dst += dst_stride; h -= 2; } while (h); } else if (w == 8) { do { __m128i s[2]; s[0] = _mm_loadl_epi64((__m128i *)src); src += src_stride; s[1] = _mm_loadl_epi64((__m128i *)src); src += src_stride; _mm_storel_epi64((__m128i *)dst, s[0]); dst += dst_stride; _mm_storel_epi64((__m128i *)dst, s[1]); dst += dst_stride; h -= 2; } while (h); } else if (w == 16) { do { __m128i s[2]; s[0] = _mm_loadu_si128((__m128i *)src); src += src_stride; s[1] = _mm_loadu_si128((__m128i *)src); src += src_stride; _mm_store_si128((__m128i *)dst, s[0]); dst += dst_stride; _mm_store_si128((__m128i *)dst, s[1]); dst += dst_stride; h -= 2; } while (h); } else if (w == 32) { do { __m256i s[2]; s[0] = _mm256_loadu_si256((__m256i *)src); src += src_stride; s[1] = _mm256_loadu_si256((__m256i *)src); src += src_stride; _mm256_storeu_si256((__m256i *)dst, s[0]); dst += dst_stride; _mm256_storeu_si256((__m256i *)dst, s[1]); dst += dst_stride; h -= 2; } while (h); } else if (w == 64) { do { __m256i s[4]; s[0] = _mm256_loadu_si256((__m256i *)(src + 0 * 32)); s[1] = _mm256_loadu_si256((__m256i *)(src + 1 * 32)); src += src_stride; s[2] = _mm256_loadu_si256((__m256i *)(src + 0 * 32)); s[3] = _mm256_loadu_si256((__m256i *)(src + 1 * 32)); src += src_stride; _mm256_storeu_si256((__m256i *)(dst + 0 * 32), s[0]); _mm256_storeu_si256((__m256i *)(dst + 1 * 32), s[1]); dst += dst_stride; _mm256_storeu_si256((__m256i *)(dst + 0 * 32), s[2]); _mm256_storeu_si256((__m256i *)(dst + 1 * 32), s[3]); dst += dst_stride; h -= 2; } while (h); } else { do { copy_128(src, dst); src += src_stride; dst += dst_stride; copy_128(src, dst); src += src_stride; dst += dst_stride; h -= 2; } while (h); } } #if CONFIG_AV1_HIGHBITDEPTH static INLINE void highbd_copy_64(const uint16_t *src, uint16_t *dst) { __m256i s[4]; s[0] = _mm256_loadu_si256((__m256i *)(src + 0 * 16)); s[1] = _mm256_loadu_si256((__m256i *)(src + 1 * 16)); s[2] = _mm256_loadu_si256((__m256i *)(src + 2 * 16)); s[3] = _mm256_loadu_si256((__m256i *)(src + 3 * 16)); _mm256_storeu_si256((__m256i *)(dst + 0 * 16), s[0]); _mm256_storeu_si256((__m256i *)(dst + 1 * 16), s[1]); _mm256_storeu_si256((__m256i *)(dst + 2 * 16), s[2]); _mm256_storeu_si256((__m256i *)(dst + 3 * 16), s[3]); } static INLINE void highbd_copy_128(const uint16_t *src, uint16_t *dst) { __m256i s[8]; s[0] = _mm256_loadu_si256((__m256i *)(src + 0 * 16)); s[1] = _mm256_loadu_si256((__m256i *)(src + 1 * 16)); s[2] = _mm256_loadu_si256((__m256i *)(src + 2 * 16)); s[3] = _mm256_loadu_si256((__m256i *)(src + 3 * 16)); s[4] = _mm256_loadu_si256((__m256i *)(src + 4 * 16)); s[5] = _mm256_loadu_si256((__m256i *)(src + 5 * 16)); s[6] = _mm256_loadu_si256((__m256i *)(src + 6 * 16)); s[7] = _mm256_loadu_si256((__m256i *)(src + 7 * 16)); _mm256_storeu_si256((__m256i *)(dst + 0 * 16), s[0]); _mm256_storeu_si256((__m256i *)(dst + 1 * 16), s[1]); _mm256_storeu_si256((__m256i *)(dst + 2 * 16), s[2]); _mm256_storeu_si256((__m256i *)(dst + 3 * 16), s[3]); _mm256_storeu_si256((__m256i *)(dst + 4 * 16), s[4]); _mm256_storeu_si256((__m256i *)(dst + 5 * 16), s[5]); _mm256_storeu_si256((__m256i *)(dst + 6 * 16), s[6]); _mm256_storeu_si256((__m256i *)(dst + 7 * 16), s[7]); } void aom_highbd_convolve_copy_avx2(const uint16_t *src, ptrdiff_t src_stride, uint16_t *dst, ptrdiff_t dst_stride, int w, int h) { if (w >= 16) { assert(!((intptr_t)dst % 16)); assert(!(dst_stride % 16)); } if (w == 2) { do { memmove(dst, src, 2 * sizeof(*src)); src += src_stride; dst += dst_stride; memmove(dst, src, 2 * sizeof(*src)); src += src_stride; dst += dst_stride; h -= 2; } while (h); } else if (w == 4) { do { __m128i s[2]; s[0] = _mm_loadl_epi64((__m128i *)src); src += src_stride; s[1] = _mm_loadl_epi64((__m128i *)src); src += src_stride; _mm_storel_epi64((__m128i *)dst, s[0]); dst += dst_stride; _mm_storel_epi64((__m128i *)dst, s[1]); dst += dst_stride; h -= 2; } while (h); } else if (w == 8) { do { __m128i s[2]; s[0] = _mm_loadu_si128((__m128i *)src); src += src_stride; s[1] = _mm_loadu_si128((__m128i *)src); src += src_stride; _mm_store_si128((__m128i *)dst, s[0]); dst += dst_stride; _mm_store_si128((__m128i *)dst, s[1]); dst += dst_stride; h -= 2; } while (h); } else if (w == 16) { do { __m256i s[2]; s[0] = _mm256_loadu_si256((__m256i *)src); src += src_stride; s[1] = _mm256_loadu_si256((__m256i *)src); src += src_stride; _mm256_storeu_si256((__m256i *)dst, s[0]); dst += dst_stride; _mm256_storeu_si256((__m256i *)dst, s[1]); dst += dst_stride; h -= 2; } while (h); } else if (w == 32) { do { __m256i s[4]; s[0] = _mm256_loadu_si256((__m256i *)(src + 0 * 16)); s[1] = _mm256_loadu_si256((__m256i *)(src + 1 * 16)); src += src_stride; s[2] = _mm256_loadu_si256((__m256i *)(src + 0 * 16)); s[3] = _mm256_loadu_si256((__m256i *)(src + 1 * 16)); src += src_stride; _mm256_storeu_si256((__m256i *)(dst + 0 * 16), s[0]); _mm256_storeu_si256((__m256i *)(dst + 1 * 16), s[1]); dst += dst_stride; _mm256_storeu_si256((__m256i *)(dst + 0 * 16), s[2]); _mm256_storeu_si256((__m256i *)(dst + 1 * 16), s[3]); dst += dst_stride; h -= 2; } while (h); } else if (w == 64) { do { highbd_copy_64(src, dst); src += src_stride; dst += dst_stride; highbd_copy_64(src, dst); src += src_stride; dst += dst_stride; h -= 2; } while (h); } else { assert(w == 128); do { highbd_copy_128(src, dst); src += src_stride; dst += dst_stride; highbd_copy_128(src, dst); src += src_stride; dst += dst_stride; h -= 2; } while (h); } } #endif // CONFIG_AV1_HIGHBITDEPTH