diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-28 14:29:10 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-28 14:29:10 +0000 |
commit | 2aa4a82499d4becd2284cdb482213d541b8804dd (patch) | |
tree | b80bf8bf13c3766139fbacc530efd0dd9d54394c /media/libjpeg/simd/loongson/jccolext-mmi.c | |
parent | Initial commit. (diff) | |
download | firefox-2aa4a82499d4becd2284cdb482213d541b8804dd.tar.xz firefox-2aa4a82499d4becd2284cdb482213d541b8804dd.zip |
Adding upstream version 86.0.1.upstream/86.0.1upstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'media/libjpeg/simd/loongson/jccolext-mmi.c')
-rw-r--r-- | media/libjpeg/simd/loongson/jccolext-mmi.c | 483 |
1 files changed, 483 insertions, 0 deletions
diff --git a/media/libjpeg/simd/loongson/jccolext-mmi.c b/media/libjpeg/simd/loongson/jccolext-mmi.c new file mode 100644 index 0000000000..6cdeb5e09a --- /dev/null +++ b/media/libjpeg/simd/loongson/jccolext-mmi.c @@ -0,0 +1,483 @@ +/* + * Loongson MMI optimizations for libjpeg-turbo + * + * Copyright 2009 Pierre Ossman <ossman@cendio.se> for Cendio AB + * Copyright (C) 2014-2015, 2019, D. R. Commander. All Rights Reserved. + * Copyright (C) 2016-2018, Loongson Technology Corporation Limited, BeiJing. + * All Rights Reserved. + * Authors: ZhuChen <zhuchen@loongson.cn> + * SunZhangzhi <sunzhangzhi-cq@loongson.cn> + * CaiWanwei <caiwanwei@loongson.cn> + * ZhangLixia <zhanglixia-hf@loongson.cn> + * + * Based on the x86 SIMD extension for IJG JPEG library + * Copyright (C) 1999-2006, MIYASAKA Masaru. + * + * This software is provided 'as-is', without any express or implied + * warranty. In no event will the authors be held liable for any damages + * arising from the use of this software. + * + * Permission is granted to anyone to use this software for any purpose, + * including commercial applications, and to alter it and redistribute it + * freely, subject to the following restrictions: + * + * 1. The origin of this software must not be misrepresented; you must not + * claim that you wrote the original software. If you use this software + * in a product, an acknowledgment in the product documentation would be + * appreciated but is not required. + * 2. Altered source versions must be plainly marked as such, and must not be + * misrepresented as being the original software. + * 3. This notice may not be removed or altered from any source distribution. + */ + +/* This file is included by jccolor-mmi.c */ + + +#if RGB_RED == 0 +#define mmA mm0 +#define mmB mm1 +#elif RGB_GREEN == 0 +#define mmA mm2 +#define mmB mm3 +#elif RGB_BLUE == 0 +#define mmA mm4 +#define mmB mm5 +#else +#define mmA mm6 +#define mmB mm7 +#endif + +#if RGB_RED == 1 +#define mmC mm0 +#define mmD mm1 +#elif RGB_GREEN == 1 +#define mmC mm2 +#define mmD mm3 +#elif RGB_BLUE == 1 +#define mmC mm4 +#define mmD mm5 +#else +#define mmC mm6 +#define mmD mm7 +#endif + +#if RGB_RED == 2 +#define mmE mm0 +#define mmF mm1 +#elif RGB_GREEN == 2 +#define mmE mm2 +#define mmF mm3 +#elif RGB_BLUE == 2 +#define mmE mm4 +#define mmF mm5 +#else +#define mmE mm6 +#define mmF mm7 +#endif + +#if RGB_RED == 3 +#define mmG mm0 +#define mmH mm1 +#elif RGB_GREEN == 3 +#define mmG mm2 +#define mmH mm3 +#elif RGB_BLUE == 3 +#define mmG mm4 +#define mmH mm5 +#else +#define mmG mm6 +#define mmH mm7 +#endif + + +void jsimd_rgb_ycc_convert_mmi(JDIMENSION image_width, JSAMPARRAY input_buf, + JSAMPIMAGE output_buf, JDIMENSION output_row, + int num_rows) +{ + JSAMPROW inptr, outptr0, outptr1, outptr2; + int num_cols, col; + __m64 mm0, mm1, mm2, mm3, mm4, mm5, mm6, mm7; + __m64 wk[7]; + __m64 Y_BG, Cb_RG, Cr_BG; + + while (--num_rows >= 0) { + inptr = *input_buf++; + outptr0 = output_buf[0][output_row]; + outptr1 = output_buf[1][output_row]; + outptr2 = output_buf[2][output_row]; + output_row++; + + for (num_cols = image_width; num_cols > 0; num_cols -= 8, + outptr0 += 8, outptr1 += 8, outptr2 += 8) { + +#if RGB_PIXELSIZE == 3 + + if (num_cols < 8) { + col = num_cols * 3; + asm(".set noreorder\r\n" + + "li $8, 1\r\n" + "move $9, %3\r\n" + "and $10, $9, $8\r\n" + "beqz $10, 1f\r\n" + "nop \r\n" + "subu $9, $9, 1\r\n" + "xor $12, $12, $12\r\n" + "move $13, %5\r\n" + "dadd $13, $13, $9\r\n" + "lbu $12, 0($13)\r\n" + + "1: \r\n" + "li $8, 2\r\n" + "and $10, $9, $8\r\n" + "beqz $10, 2f\r\n" + "nop \r\n" + "subu $9, $9, 2\r\n" + "xor $11, $11, $11\r\n" + "move $13, %5\r\n" + "dadd $13, $13, $9\r\n" + "lhu $11, 0($13)\r\n" + "sll $12, $12, 16\r\n" + "or $12, $12, $11\r\n" + + "2: \r\n" + "dmtc1 $12, %0\r\n" + "li $8, 4\r\n" + "and $10, $9, $8\r\n" + "beqz $10, 3f\r\n" + "nop \r\n" + "subu $9, $9, 4\r\n" + "move $13, %5\r\n" + "dadd $13, $13, $9\r\n" + "lwu $14, 0($13)\r\n" + "dmtc1 $14, %1\r\n" + "dsll32 $12, $12, 0\r\n" + "or $12, $12, $14\r\n" + "dmtc1 $12, %0\r\n" + + "3: \r\n" + "li $8, 8\r\n" + "and $10, $9, $8\r\n" + "beqz $10, 4f\r\n" + "nop \r\n" + "mov.s %1, %0\r\n" + "ldc1 %0, 0(%5)\r\n" + "li $9, 8\r\n" + "j 5f\r\n" + "nop \r\n" + + "4: \r\n" + "li $8, 16\r\n" + "and $10, $9, $8\r\n" + "beqz $10, 5f\r\n" + "nop \r\n" + "mov.s %2, %0\r\n" + "ldc1 %0, 0(%5)\r\n" + "ldc1 %1, 8(%5)\r\n" + + "5: \r\n" + "nop \r\n" + ".set reorder\r\n" + + : "=f" (mmA), "=f" (mmG), "=f" (mmF) + : "r" (col), "r" (num_rows), "r" (inptr) + : "$f0", "$f2", "$f4", "$8", "$9", "$10", "$11", "$12", "$13", + "$14", "memory" + ); + } else { + if (!(((long)inptr) & 7)) { + mmA = _mm_load_si64((__m64 *)&inptr[0]); + mmG = _mm_load_si64((__m64 *)&inptr[8]); + mmF = _mm_load_si64((__m64 *)&inptr[16]); + } else { + mmA = _mm_loadu_si64((__m64 *)&inptr[0]); + mmG = _mm_loadu_si64((__m64 *)&inptr[8]); + mmF = _mm_loadu_si64((__m64 *)&inptr[16]); + } + inptr += RGB_PIXELSIZE * 8; + } + mmD = mmA; + mmA = _mm_slli_si64(mmA, 4 * BYTE_BIT); + mmD = _mm_srli_si64(mmD, 4 * BYTE_BIT); + + mmA = _mm_unpackhi_pi8(mmA, mmG); + mmG = _mm_slli_si64(mmG, 4 * BYTE_BIT); + + mmD = _mm_unpacklo_pi8(mmD, mmF); + mmG = _mm_unpackhi_pi8(mmG, mmF); + + mmE = mmA; + mmA = _mm_slli_si64(mmA, 4 * BYTE_BIT); + mmE = _mm_srli_si64(mmE, 4 * BYTE_BIT); + + mmA = _mm_unpackhi_pi8(mmA, mmD); + mmD = _mm_slli_si64(mmD, 4 * BYTE_BIT); + + mmE = _mm_unpacklo_pi8(mmE, mmG); + mmD = _mm_unpackhi_pi8(mmD, mmG); + mmC = mmA; + mmA = _mm_loadlo_pi8_f(mmA); + mmC = _mm_loadhi_pi8_f(mmC); + + mmB = mmE; + mmE = _mm_loadlo_pi8_f(mmE); + mmB = _mm_loadhi_pi8_f(mmB); + + mmF = mmD; + mmD = _mm_loadlo_pi8_f(mmD); + mmF = _mm_loadhi_pi8_f(mmF); + +#else /* RGB_PIXELSIZE == 4 */ + + if (num_cols < 8) { + col = num_cols; + asm(".set noreorder\r\n" + + "li $8, 1\r\n" + "move $9, %4\r\n" + "and $10, $9, $8\r\n" + "beqz $10, 1f\r\n" + "nop \r\n" + "subu $9, $9, 1\r\n" + "dsll $11, $9, 2\r\n" + "move $13, %5\r\n" + "daddu $13, $13, $11\r\n" + "lwc1 %0, 0($13)\r\n" + + "1: \r\n" + "li $8, 2\r\n" + "and $10, $9, $8\r\n" + "beqz $10, 2f\r\n" + "nop \r\n" + "subu $9, $9, 2\r\n" + "dsll $11, $9, 2\r\n" + "move $13, %5\r\n" + "daddu $13, $13, $11\r\n" + "mov.s %1, %0\r\n" + "ldc1 %0, 0($13)\r\n" + + "2: \r\n" + "li $8, 4\r\n" + "and $10, $9, $8\r\n" + "beqz $10, 3f\r\n" + "nop \r\n" + "mov.s %2, %0\r\n" + "mov.s %3, %1\r\n" + "ldc1 %0, 0(%5)\r\n" + "ldc1 %1, 8(%5)\r\n" + + "3: \r\n" + "nop \r\n" + ".set reorder\r\n" + + : "=f" (mmA), "=f" (mmF), "=f" (mmD), "=f" (mmC) + : "r" (col), "r" (inptr) + : "$f0", "$f2", "$8", "$9", "$10", "$11", "$13", "memory" + ); + } else { + if (!(((long)inptr) & 7)) { + mmA = _mm_load_si64((__m64 *)&inptr[0]); + mmF = _mm_load_si64((__m64 *)&inptr[8]); + mmD = _mm_load_si64((__m64 *)&inptr[16]); + mmC = _mm_load_si64((__m64 *)&inptr[24]); + } else { + mmA = _mm_loadu_si64((__m64 *)&inptr[0]); + mmF = _mm_loadu_si64((__m64 *)&inptr[8]); + mmD = _mm_loadu_si64((__m64 *)&inptr[16]); + mmC = _mm_loadu_si64((__m64 *)&inptr[24]); + } + inptr += RGB_PIXELSIZE * 8; + } + mmB = mmA; + mmA = _mm_unpacklo_pi8(mmA, mmF); + mmB = _mm_unpackhi_pi8(mmB, mmF); + + mmG = mmD; + mmD = _mm_unpacklo_pi8(mmD, mmC); + mmG = _mm_unpackhi_pi8(mmG, mmC); + + mmE = mmA; + mmA = _mm_unpacklo_pi16(mmA, mmD); + mmE = _mm_unpackhi_pi16(mmE, mmD); + + mmH = mmB; + mmB = _mm_unpacklo_pi16(mmB, mmG); + mmH = _mm_unpackhi_pi16(mmH, mmG); + + mmC = mmA; + mmA = _mm_loadlo_pi8_f(mmA); + mmC = _mm_loadhi_pi8_f(mmC); + + mmD = mmB; + mmB = _mm_loadlo_pi8_f(mmB); + mmD = _mm_loadhi_pi8_f(mmD); + + mmG = mmE; + mmE = _mm_loadlo_pi8_f(mmE); + mmG = _mm_loadhi_pi8_f(mmG); + + mmF = mmH; + mmF = _mm_unpacklo_pi8(mmF, mmH); + mmH = _mm_unpackhi_pi8(mmH, mmH); + mmF = _mm_srli_pi16(mmF, BYTE_BIT); + mmH = _mm_srli_pi16(mmH, BYTE_BIT); + +#endif + + wk[0] = mm0; + wk[1] = mm1; + wk[2] = mm4; + wk[3] = mm5; + + mm6 = mm1; + mm1 = _mm_unpacklo_pi16(mm1, mm3); + mm6 = _mm_unpackhi_pi16(mm6, mm3); + mm7 = mm1; + mm4 = mm6; + mm1 = _mm_madd_pi16(mm1, PW_F0299_F0337); + mm6 = _mm_madd_pi16(mm6, PW_F0299_F0337); + mm7 = _mm_madd_pi16(mm7, PW_MF016_MF033); + mm4 = _mm_madd_pi16(mm4, PW_MF016_MF033); + + wk[4] = mm1; + wk[5] = mm6; + + mm1 = _mm_loadlo_pi16_f(mm5); + mm6 = _mm_loadhi_pi16_f(mm5); + mm1 = _mm_srli_pi32(mm1, 1); + mm6 = _mm_srli_pi32(mm6, 1); + + mm5 = PD_ONEHALFM1_CJ; + mm7 = _mm_add_pi32(mm7, mm1); + mm4 = _mm_add_pi32(mm4, mm6); + mm7 = _mm_add_pi32(mm7, mm5); + mm4 = _mm_add_pi32(mm4, mm5); + mm7 = _mm_srli_pi32(mm7, SCALEBITS); + mm4 = _mm_srli_pi32(mm4, SCALEBITS); + mm7 = _mm_packs_pi32(mm7, mm4); + + mm1 = wk[2]; + mm6 = mm0; + mm0 = _mm_unpacklo_pi16(mm0, mm2); + mm6 = _mm_unpackhi_pi16(mm6, mm2); + mm5 = mm0; + mm4 = mm6; + mm0 = _mm_madd_pi16(mm0, PW_F0299_F0337); + mm6 = _mm_madd_pi16(mm6, PW_F0299_F0337); + mm5 = _mm_madd_pi16(mm5, PW_MF016_MF033); + mm4 = _mm_madd_pi16(mm4, PW_MF016_MF033); + + wk[6] = mm0; + wk[7] = mm6; + mm0 = _mm_loadlo_pi16_f(mm1); + mm6 = _mm_loadhi_pi16_f(mm1); + mm0 = _mm_srli_pi32(mm0, 1); + mm6 = _mm_srli_pi32(mm6, 1); + + mm1 = PD_ONEHALFM1_CJ; + mm5 = _mm_add_pi32(mm5, mm0); + mm4 = _mm_add_pi32(mm4, mm6); + mm5 = _mm_add_pi32(mm5, mm1); + mm4 = _mm_add_pi32(mm4, mm1); + mm5 = _mm_srli_pi32(mm5, SCALEBITS); + mm4 = _mm_srli_pi32(mm4, SCALEBITS); + mm5 = _mm_packs_pi32(mm5, mm4); + + mm7 = _mm_slli_pi16(mm7, BYTE_BIT); + mm5 = _mm_or_si64(mm5, mm7); + Cb_RG = mm5; + + mm0 = wk[3]; + mm6 = wk[2]; + mm1 = wk[1]; + + mm4 = mm0; + mm0 = _mm_unpacklo_pi16(mm0, mm3); + mm4 = _mm_unpackhi_pi16(mm4, mm3); + mm7 = mm0; + mm5 = mm4; + mm0 = _mm_madd_pi16(mm0, PW_F0114_F0250); + mm4 = _mm_madd_pi16(mm4, PW_F0114_F0250); + mm7 = _mm_madd_pi16(mm7, PW_MF008_MF041); + mm5 = _mm_madd_pi16(mm5, PW_MF008_MF041); + + mm3 = PD_ONEHALF; + mm0 = _mm_add_pi32(mm0, wk[4]); + mm4 = _mm_add_pi32(mm4, wk[5]); + mm0 = _mm_add_pi32(mm0, mm3); + mm4 = _mm_add_pi32(mm4, mm3); + mm0 = _mm_srli_pi32(mm0, SCALEBITS); + mm4 = _mm_srli_pi32(mm4, SCALEBITS); + mm0 = _mm_packs_pi32(mm0, mm4); + + mm3 = _mm_loadlo_pi16_f(mm1); + mm4 = _mm_loadhi_pi16_f(mm1); + mm3 = _mm_srli_pi32(mm3, 1); + mm4 = _mm_srli_pi32(mm4, 1); + + mm1 = PD_ONEHALFM1_CJ; + mm7 = _mm_add_pi32(mm7, mm3); + mm5 = _mm_add_pi32(mm5, mm4); + mm7 = _mm_add_pi32(mm7, mm1); + mm5 = _mm_add_pi32(mm5, mm1); + mm7 = _mm_srli_pi32(mm7, SCALEBITS); + mm5 = _mm_srli_pi32(mm5, SCALEBITS); + mm7 = _mm_packs_pi32(mm7, mm5); + + mm3 = wk[0]; + mm4 = mm6; + mm6 = _mm_unpacklo_pi16(mm6, mm2); + mm4 = _mm_unpackhi_pi16(mm4, mm2); + mm1 = mm6; + mm5 = mm4; + mm6 = _mm_madd_pi16(mm6, PW_F0114_F0250); + mm4 = _mm_madd_pi16(mm4, PW_F0114_F0250); + mm1 = _mm_madd_pi16(mm1, PW_MF008_MF041); + mm5 = _mm_madd_pi16(mm5, PW_MF008_MF041); + + mm2 = PD_ONEHALF; + mm6 = _mm_add_pi32(mm6, wk[6]); + mm4 = _mm_add_pi32(mm4, wk[7]); + mm6 = _mm_add_pi32(mm6, mm2); + mm4 = _mm_add_pi32(mm4, mm2); + mm6 = _mm_srli_pi32(mm6, SCALEBITS); + mm4 = _mm_srli_pi32(mm4, SCALEBITS); + mm6 = _mm_packs_pi32(mm6, mm4); + + mm0 = _mm_slli_pi16(mm0, BYTE_BIT); + mm6 = _mm_or_si64(mm6, mm0); + Y_BG = mm6; + + mm2 = _mm_loadlo_pi16_f(mm3); + mm4 = _mm_loadhi_pi16_f(mm3); + mm2 = _mm_srli_pi32(mm2, 1); + mm4 = _mm_srli_pi32(mm4, 1); + + mm0 = PD_ONEHALFM1_CJ; + mm1 = _mm_add_pi32(mm1, mm2); + mm5 = _mm_add_pi32(mm5, mm4); + mm1 = _mm_add_pi32(mm1, mm0); + mm5 = _mm_add_pi32(mm5, mm0); + mm1 = _mm_srli_pi32(mm1, SCALEBITS); + mm5 = _mm_srli_pi32(mm5, SCALEBITS); + mm1 = _mm_packs_pi32(mm1, mm5); + + mm7 = _mm_slli_pi16(mm7, BYTE_BIT); + mm1 = _mm_or_si64(mm1, mm7); + Cr_BG = mm1; + + _mm_store_si64((__m64 *)&outptr0[0], Y_BG); + _mm_store_si64((__m64 *)&outptr1[0], Cb_RG); + _mm_store_si64((__m64 *)&outptr2[0], Cr_BG); + } + } +} + +#undef mmA +#undef mmB +#undef mmC +#undef mmD +#undef mmE +#undef mmF +#undef mmG +#undef mmH |