/* * Loongson MMI optimizations for libjpeg-turbo * * Copyright 2009 Pierre Ossman for Cendio AB * Copyright (C) 2015, 2019, D. R. Commander. All Rights Reserved. * Copyright (C) 2016-2018, Loongson Technology Corporation Limited, BeiJing. * All Rights Reserved. * Authors: ZhangLixia * * Based on the x86 SIMD extension for IJG JPEG library * Copyright (C) 1999-2006, MIYASAKA Masaru. * * This software is provided 'as-is', without any express or implied * warranty. In no event will the authors be held liable for any damages * arising from the use of this software. * * Permission is granted to anyone to use this software for any purpose, * including commercial applications, and to alter it and redistribute it * freely, subject to the following restrictions: * * 1. The origin of this software must not be misrepresented; you must not * claim that you wrote the original software. If you use this software * in a product, an acknowledgment in the product documentation would be * appreciated but is not required. * 2. Altered source versions must be plainly marked as such, and must not be * misrepresented as being the original software. * 3. This notice may not be removed or altered from any source distribution. */ /* This file is included by jdmerge-mmi.c */ #if RGB_RED == 0 #define mmA re #define mmB ro #elif RGB_GREEN == 0 #define mmA ge #define mmB go #elif RGB_BLUE == 0 #define mmA be #define mmB bo #else #define mmA xe #define mmB xo #endif #if RGB_RED == 1 #define mmC re #define mmD ro #elif RGB_GREEN == 1 #define mmC ge #define mmD go #elif RGB_BLUE == 1 #define mmC be #define mmD bo #else #define mmC xe #define mmD xo #endif #if RGB_RED == 2 #define mmE re #define mmF ro #elif RGB_GREEN == 2 #define mmE ge #define mmF go #elif RGB_BLUE == 2 #define mmE be #define mmF bo #else #define mmE xe #define mmF xo #endif #if RGB_RED == 3 #define mmG re #define mmH ro #elif RGB_GREEN == 3 #define mmG ge #define mmH go #elif RGB_BLUE == 3 #define mmG be #define mmH bo #else #define mmG xe #define mmH xo #endif void jsimd_h2v1_merged_upsample_mmi(JDIMENSION output_width, JSAMPIMAGE input_buf, JDIMENSION in_row_group_ctr, JSAMPARRAY output_buf) { JSAMPROW outptr, inptr0, inptr1, inptr2; int num_cols, col; __m64 ythise, ythiso, ythis, ynexte, ynexto, ynext, yl, y; __m64 cbl, cbl2, cbh, cbh2, cb, crl, crl2, crh, crh2, cr; __m64 rle, rlo, rl, rhe, rho, rh, re, ro; __m64 ga, gb, gle, glo, gl, gc, gd, ghe, gho, gh, ge, go; __m64 ble, blo, bl, bhe, bho, bh, be, bo, xe = 0.0, xo = 0.0; __m64 decenter, mask, zero = 0.0; #if RGB_PIXELSIZE == 4 __m64 mm8, mm9; #endif inptr0 = input_buf[0][in_row_group_ctr]; inptr1 = input_buf[1][in_row_group_ctr]; inptr2 = input_buf[2][in_row_group_ctr]; outptr = output_buf[0]; for (num_cols = output_width >> 1; num_cols > 0; num_cols -= 8, inptr0 += 16, inptr1 += 8, inptr2 += 8) { cb = _mm_load_si64((__m64 *)inptr1); cr = _mm_load_si64((__m64 *)inptr2); ythis = _mm_load_si64((__m64 *)inptr0); ynext = _mm_load_si64((__m64 *)inptr0 + 1); mask = decenter = 0.0; mask = _mm_cmpeq_pi16(mask, mask); decenter = _mm_cmpeq_pi16(decenter, decenter); mask = _mm_srli_pi16(mask, BYTE_BIT); /* {0xFF 0x00 0xFF 0x00 ..} */ decenter = _mm_slli_pi16(decenter, 7); /* {0xFF80 0xFF80 0xFF80 0xFF80} */ cbl = _mm_unpacklo_pi8(cb, zero); /* Cb(0123) */ cbh = _mm_unpackhi_pi8(cb, zero); /* Cb(4567) */ crl = _mm_unpacklo_pi8(cr, zero); /* Cr(0123) */ crh = _mm_unpackhi_pi8(cr, zero); /* Cr(4567) */ cbl = _mm_add_pi16(cbl, decenter); cbh = _mm_add_pi16(cbh, decenter); crl = _mm_add_pi16(crl, decenter); crh = _mm_add_pi16(crh, decenter); /* (Original) * R = Y + 1.40200 * Cr * G = Y - 0.34414 * Cb - 0.71414 * Cr * B = Y + 1.77200 * Cb * * (This implementation) * R = Y + 0.40200 * Cr + Cr * G = Y - 0.34414 * Cb + 0.28586 * Cr - Cr * B = Y - 0.22800 * Cb + Cb + Cb */ cbl2 = _mm_add_pi16(cbl, cbl); /* 2*CbL */ cbh2 = _mm_add_pi16(cbh, cbh); /* 2*CbH */ crl2 = _mm_add_pi16(crl, crl); /* 2*CrL */ crh2 = _mm_add_pi16(crh, crh); /* 2*CrH */ bl = _mm_mulhi_pi16(cbl2, PW_MF0228); /* (2*CbL * -FIX(0.22800) */ bh = _mm_mulhi_pi16(cbh2, PW_MF0228); /* (2*CbH * -FIX(0.22800) */ rl = _mm_mulhi_pi16(crl2, PW_F0402); /* (2*CrL * FIX(0.40200)) */ rh = _mm_mulhi_pi16(crh2, PW_F0402); /* (2*CrH * FIX(0.40200)) */ bl = _mm_add_pi16(bl, PW_ONE); bh = _mm_add_pi16(bh, PW_ONE); bl = _mm_srai_pi16(bl, 1); /* (CbL * -FIX(0.22800)) */ bh = _mm_srai_pi16(bh, 1); /* (CbH * -FIX(0.22800)) */ rl = _mm_add_pi16(rl, PW_ONE); rh = _mm_add_pi16(rh, PW_ONE); rl = _mm_srai_pi16(rl, 1); /* (CrL * FIX(0.40200)) */ rh = _mm_srai_pi16(rh, 1); /* (CrH * FIX(0.40200)) */ bl = _mm_add_pi16(bl, cbl); bh = _mm_add_pi16(bh, cbh); bl = _mm_add_pi16(bl, cbl); /* (CbL * FIX(1.77200))=(B-Y)L */ bh = _mm_add_pi16(bh, cbh); /* (CbH * FIX(1.77200))=(B-Y)H */ rl = _mm_add_pi16(rl, crl); /* (CrL * FIX(1.40200))=(R-Y)L */ rh = _mm_add_pi16(rh, crh); /* (CrH * FIX(1.40200))=(R-Y)H */ ga = _mm_unpacklo_pi16(cbl, crl); gb = _mm_unpackhi_pi16(cbl, crl); ga = _mm_madd_pi16(ga, PW_MF0344_F0285); gb = _mm_madd_pi16(gb, PW_MF0344_F0285); gc = _mm_unpacklo_pi16(cbh, crh); gd = _mm_unpackhi_pi16(cbh, crh); gc = _mm_madd_pi16(gc, PW_MF0344_F0285); gd = _mm_madd_pi16(gd, PW_MF0344_F0285); ga = _mm_add_pi32(ga, PD_ONEHALF); gb = _mm_add_pi32(gb, PD_ONEHALF); ga = _mm_srai_pi32(ga, SCALEBITS); gb = _mm_srai_pi32(gb, SCALEBITS); gc = _mm_add_pi32(gc, PD_ONEHALF); gd = _mm_add_pi32(gd, PD_ONEHALF); gc = _mm_srai_pi32(gc, SCALEBITS); gd = _mm_srai_pi32(gd, SCALEBITS); gl = _mm_packs_pi32(ga, gb); /* CbL*-FIX(0.344)+CrL*FIX(0.285) */ gh = _mm_packs_pi32(gc, gd); /* CbH*-FIX(0.344)+CrH*FIX(0.285) */ gl = _mm_sub_pi16(gl, crl); /* CbL*-FIX(0.344)+CrL*-FIX(0.714)=(G-Y)L */ gh = _mm_sub_pi16(gh, crh); /* CbH*-FIX(0.344)+CrH*-FIX(0.714)=(G-Y)H */ ythise = _mm_and_si64(mask, ythis); /* Y(0246) */ ythiso = _mm_srli_pi16(ythis, BYTE_BIT); /* Y(1357) */ ynexte = _mm_and_si64(mask, ynext); /* Y(8ACE) */ ynexto = _mm_srli_pi16(ynext, BYTE_BIT); /* Y(9BDF) */ rle = _mm_add_pi16(rl, ythise); /* (R0 R2 R4 R6) */ rlo = _mm_add_pi16(rl, ythiso); /* (R1 R3 R5 R7) */ rhe = _mm_add_pi16(rh, ynexte); /* (R8 RA RC RE) */ rho = _mm_add_pi16(rh, ynexto); /* (R9 RB RD RF) */ re = _mm_packs_pu16(rle, rhe); /* (R0 R2 R4 R6 R8 RA RC RE) */ ro = _mm_packs_pu16(rlo, rho); /* (R1 R3 R5 R7 R9 RB RD RF) */ gle = _mm_add_pi16(gl, ythise); /* (G0 G2 G4 G6) */ glo = _mm_add_pi16(gl, ythiso); /* (G1 G3 G5 G7) */ ghe = _mm_add_pi16(gh, ynexte); /* (G8 GA GC GE) */ gho = _mm_add_pi16(gh, ynexto); /* (G9 GB GD GF) */ ge = _mm_packs_pu16(gle, ghe); /* (G0 G2 G4 G6 G8 GA GC GE) */ go = _mm_packs_pu16(glo, gho); /* (G1 G3 G5 G7 G9 GB GD GF) */ ble = _mm_add_pi16(bl, ythise); /* (B0 B2 B4 B6) */ blo = _mm_add_pi16(bl, ythiso); /* (B1 B3 B5 B7) */ bhe = _mm_add_pi16(bh, ynexte); /* (B8 BA BC BE) */ bho = _mm_add_pi16(bh, ynexto); /* (B9 BB BD BF) */ be = _mm_packs_pu16(ble, bhe); /* (B0 B2 B4 B6 B8 BA BC BE) */ bo = _mm_packs_pu16(blo, bho); /* (B1 B3 B5 B7 B9 BB BD BF) */ #if RGB_PIXELSIZE == 3 /* mmA=(00 02 04 06 08 0A 0C 0E), mmB=(01 03 05 07 09 0B 0D 0F) */ /* mmC=(10 12 14 16 18 1A 1C 1E), mmD=(11 13 15 17 19 1B 1D 1F) */ /* mmE=(20 22 24 26 28 2A 2C 2E), mmF=(21 23 25 27 29 2B 2D 2F) */ mmG = _mm_unpacklo_pi8(mmA, mmC); /* (00 10 02 12 04 14 06 16) */ mmA = _mm_unpackhi_pi8(mmA, mmC); /* (08 18 0A 1A 0C 1C 0E 1E) */ mmH = _mm_unpacklo_pi8(mmE, mmB); /* (20 01 22 03 24 05 26 07) */ mmE = _mm_unpackhi_pi8(mmE, mmB); /* (28 09 2A 0B 2C 0D 2E 0F) */ mmC = _mm_unpacklo_pi8(mmD, mmF); /* (11 21 13 23 15 25 17 27) */ mmD = _mm_unpackhi_pi8(mmD, mmF); /* (19 29 1B 2B 1D 2D 1F 2F) */ mmB = _mm_unpacklo_pi16(mmG, mmA); /* (00 10 08 18 02 12 0A 1A) */ mmA = _mm_unpackhi_pi16(mmG, mmA); /* (04 14 0C 1C 06 16 0E 1E) */ mmF = _mm_unpacklo_pi16(mmH, mmE); /* (20 01 28 09 22 03 2A 0B) */ mmE = _mm_unpackhi_pi16(mmH, mmE); /* (24 05 2C 0D 26 07 2E 0F) */ mmH = _mm_unpacklo_pi16(mmC, mmD); /* (11 21 19 29 13 23 1B 2B) */ mmG = _mm_unpackhi_pi16(mmC, mmD); /* (15 25 1D 2D 17 27 1F 2F) */ mmC = _mm_unpacklo_pi16(mmB, mmF); /* (00 10 20 01 08 18 28 09) */ mmB = _mm_srli_si64(mmB, 4 * BYTE_BIT); mmB = _mm_unpacklo_pi16(mmH, mmB); /* (11 21 02 12 19 29 0A 1A) */ mmD = _mm_unpackhi_pi16(mmF, mmH); /* (22 03 13 23 2A 0B 1B 2B) */ mmF = _mm_unpacklo_pi16(mmA, mmE); /* (04 14 24 05 0C 1C 2C 0D) */ mmA = _mm_srli_si64(mmA, 4 * BYTE_BIT); mmH = _mm_unpacklo_pi16(mmG, mmA); /* (15 25 06 16 1D 2D 0E 1E) */ mmG = _mm_unpackhi_pi16(mmE, mmG); /* (26 07 17 27 2E 0F 1F 2F) */ mmA = _mm_unpacklo_pi32(mmC, mmB); /* (00 10 20 01 11 21 02 12) */ mmE = _mm_unpackhi_pi32(mmC, mmB); /* (08 18 28 09 19 29 0A 1A) */ mmB = _mm_unpacklo_pi32(mmD, mmF); /* (22 03 13 23 04 14 24 05) */ mmF = _mm_unpackhi_pi32(mmD, mmF); /* (2A 0B 1B 2B 0C 1C 2C 0D) */ mmC = _mm_unpacklo_pi32(mmH, mmG); /* (15 25 06 16 26 07 17 27) */ mmG = _mm_unpackhi_pi32(mmH, mmG); /* (1D 2D 0E 1E 2E 0F 1F 2F) */ if (num_cols >= 8) { if (!(((long)outptr) & 7)) { _mm_store_si64((__m64 *)outptr, mmA); _mm_store_si64((__m64 *)(outptr + 8), mmB); _mm_store_si64((__m64 *)(outptr + 16), mmC); _mm_store_si64((__m64 *)(outptr + 24), mmE); _mm_store_si64((__m64 *)(outptr + 32), mmF); _mm_store_si64((__m64 *)(outptr + 40), mmG); } else { _mm_storeu_si64((__m64 *)outptr, mmA); _mm_storeu_si64((__m64 *)(outptr + 8), mmB); _mm_storeu_si64((__m64 *)(outptr + 16), mmC); _mm_storeu_si64((__m64 *)(outptr + 24), mmE); _mm_storeu_si64((__m64 *)(outptr + 32), mmF); _mm_storeu_si64((__m64 *)(outptr + 40), mmG); } outptr += RGB_PIXELSIZE * 16; } else { if (output_width & 1) col = num_cols * 6 + 3; else col = num_cols * 6; asm(".set noreorder\r\n" /* st24 */ "li $8, 24\r\n" "move $9, %7\r\n" "mov.s $f4, %1\r\n" "mov.s $f6, %2\r\n" "mov.s $f8, %3\r\n" "move $10, %8\r\n" "bltu $9, $8, 1f\r\n" "nop \r\n" "gssdlc1 $f4, 7($10)\r\n" "gssdrc1 $f4, 0($10)\r\n" "gssdlc1 $f6, 7+8($10)\r\n" "gssdrc1 $f6, 8($10)\r\n" "gssdlc1 $f8, 7+16($10)\r\n" "gssdrc1 $f8, 16($10)\r\n" "mov.s $f4, %4\r\n" "mov.s $f6, %5\r\n" "mov.s $f8, %6\r\n" "subu $9, $9, 24\r\n" PTR_ADDU "$10, $10, 24\r\n" "1: \r\n" "li $8, 16\r\n" /* st16 */ "bltu $9, $8, 2f\r\n" "nop \r\n" "gssdlc1 $f4, 7($10)\r\n" "gssdrc1 $f4, 0($10)\r\n" "gssdlc1 $f6, 7+8($10)\r\n" "gssdrc1 $f6, 8($10)\r\n" "mov.s $f4, $f8\r\n" "subu $9, $9, 16\r\n" PTR_ADDU "$10, $10, 16\r\n" "2: \r\n" "li $8, 8\r\n" /* st8 */ "bltu $9, $8, 3f\r\n" "nop \r\n" "gssdlc1 $f4, 7($10)\r\n" "gssdrc1 $f4, 0($10)\r\n" "mov.s $f4, $f6\r\n" "subu $9, $9, 8\r\n" PTR_ADDU "$10, $10, 8\r\n" "3: \r\n" "li $8, 4\r\n" /* st4 */ "mfc1 $11, $f4\r\n" "bltu $9, $8, 4f\r\n" "nop \r\n" "swl $11, 3($10)\r\n" "swr $11, 0($10)\r\n" "li $8, 32\r\n" "mtc1 $8, $f6\r\n" "dsrl $f4, $f4, $f6\r\n" "mfc1 $11, $f4\r\n" "subu $9, $9, 4\r\n" PTR_ADDU "$10, $10, 4\r\n" "4: \r\n" "li $8, 2\r\n" /* st2 */ "bltu $9, $8, 5f\r\n" "nop \r\n" "ush $11, 0($10)\r\n" "srl $11, 16\r\n" "subu $9, $9, 2\r\n" PTR_ADDU "$10, $10, 2\r\n" "5: \r\n" "li $8, 1\r\n" /* st1 */ "bltu $9, $8, 6f\r\n" "nop \r\n" "sb $11, 0($10)\r\n" "6: \r\n" "nop \r\n" /* end */ : "=m" (*outptr) : "f" (mmA), "f" (mmB), "f" (mmC), "f" (mmE), "f" (mmF), "f" (mmG), "r" (col), "r" (outptr) : "$f4", "$f6", "$f8", "$8", "$9", "$10", "$11", "memory" ); } #else /* RGB_PIXELSIZE == 4 */ #ifdef RGBX_FILLER_0XFF xe = _mm_cmpeq_pi8(xe, xe); xo = _mm_cmpeq_pi8(xo, xo); #else xe = _mm_xor_si64(xe, xe); xo = _mm_xor_si64(xo, xo); #endif /* mmA=(00 02 04 06 08 0A 0C 0E), mmB=(01 03 05 07 09 0B 0D 0F) */ /* mmC=(10 12 14 16 18 1A 1C 1E), mmD=(11 13 15 17 19 1B 1D 1F) */ /* mmE=(20 22 24 26 28 2A 2C 2E), mmF=(21 23 25 27 29 2B 2D 2F) */ /* mmG=(30 32 34 36 38 3A 3C 3E), mmH=(31 33 35 37 39 3B 3D 3F) */ mm8 = _mm_unpacklo_pi8(mmA, mmC); /* (00 10 02 12 04 14 06 16) */ mm9 = _mm_unpackhi_pi8(mmA, mmC); /* (08 18 0A 1A 0C 1C 0E 1E) */ mmA = _mm_unpacklo_pi8(mmE, mmG); /* (20 30 22 32 24 34 26 36) */ mmE = _mm_unpackhi_pi8(mmE, mmG); /* (28 38 2A 3A 2C 3C 2E 3E) */ mmG = _mm_unpacklo_pi8(mmB, mmD); /* (01 11 03 13 05 15 07 17) */ mmB = _mm_unpackhi_pi8(mmB, mmD); /* (09 19 0B 1B 0D 1D 0F 1F) */ mmD = _mm_unpacklo_pi8(mmF, mmH); /* (21 31 23 33 25 35 27 37) */ mmF = _mm_unpackhi_pi8(mmF, mmH); /* (29 39 2B 3B 2D 3D 2F 3F) */ mmH = _mm_unpacklo_pi16(mm8, mmA); /* (00 10 20 30 02 12 22 32) */ mm8 = _mm_unpackhi_pi16(mm8, mmA); /* (04 14 24 34 06 16 26 36) */ mmA = _mm_unpacklo_pi16(mmG, mmD); /* (01 11 21 31 03 13 23 33) */ mmD = _mm_unpackhi_pi16(mmG, mmD); /* (05 15 25 35 07 17 27 37) */ mmG = _mm_unpackhi_pi16(mm9, mmE); /* (0C 1C 2C 3C 0E 1E 2E 3E) */ mm9 = _mm_unpacklo_pi16(mm9, mmE); /* (08 18 28 38 0A 1A 2A 3A) */ mmE = _mm_unpacklo_pi16(mmB, mmF); /* (09 19 29 39 0B 1B 2B 3B) */ mmF = _mm_unpackhi_pi16(mmB, mmF); /* (0D 1D 2D 3D 0F 1F 2F 3F) */ mmB = _mm_unpackhi_pi32(mmH, mmA); /* (02 12 22 32 03 13 23 33) */ mmA = _mm_unpacklo_pi32(mmH, mmA); /* (00 10 20 30 01 11 21 31) */ mmC = _mm_unpacklo_pi32(mm8, mmD); /* (04 14 24 34 05 15 25 35) */ mmD = _mm_unpackhi_pi32(mm8, mmD); /* (06 16 26 36 07 17 27 37) */ mmH = _mm_unpackhi_pi32(mmG, mmF); /* (0E 1E 2E 3E 0F 1F 2F 3F) */ mmG = _mm_unpacklo_pi32(mmG, mmF); /* (0C 1C 2C 3C 0D 1D 2D 3D) */ mmF = _mm_unpackhi_pi32(mm9, mmE); /* (0A 1A 2A 3A 0B 1B 2B 3B) */ mmE = _mm_unpacklo_pi32(mm9, mmE); /* (08 18 28 38 09 19 29 39) */ if (num_cols >= 8) { if (!(((long)outptr) & 7)) { _mm_store_si64((__m64 *)outptr, mmA); _mm_store_si64((__m64 *)(outptr + 8), mmB); _mm_store_si64((__m64 *)(outptr + 16), mmC); _mm_store_si64((__m64 *)(outptr + 24), mmD); _mm_store_si64((__m64 *)(outptr + 32), mmE); _mm_store_si64((__m64 *)(outptr + 40), mmF); _mm_store_si64((__m64 *)(outptr + 48), mmG); _mm_store_si64((__m64 *)(outptr + 56), mmH); } else { _mm_storeu_si64((__m64 *)outptr, mmA); _mm_storeu_si64((__m64 *)(outptr + 8), mmB); _mm_storeu_si64((__m64 *)(outptr + 16), mmC); _mm_storeu_si64((__m64 *)(outptr + 24), mmD); _mm_storeu_si64((__m64 *)(outptr + 32), mmE); _mm_storeu_si64((__m64 *)(outptr + 40), mmF); _mm_storeu_si64((__m64 *)(outptr + 48), mmG); _mm_storeu_si64((__m64 *)(outptr + 56), mmH); } outptr += RGB_PIXELSIZE * 16; } else { if (output_width & 1) col = num_cols * 2 + 1; else col = num_cols * 2; asm(".set noreorder\r\n" /* st32 */ "li $8, 8\r\n" "move $9, %10\r\n" "move $10, %11\r\n" "mov.s $f4, %2\r\n" "mov.s $f6, %3\r\n" "mov.s $f8, %4\r\n" "mov.s $f10, %5\r\n" "bltu $9, $8, 1f\r\n" "nop \r\n" "gssdlc1 $f4, 7($10)\r\n" "gssdrc1 $f4, 0($10)\r\n" "gssdlc1 $f6, 7+8($10)\r\n" "gssdrc1 $f6, 8($10)\r\n" "gssdlc1 $f8, 7+16($10)\r\n" "gssdrc1 $f8, 16($10)\r\n" "gssdlc1 $f10, 7+24($10)\r\n" "gssdrc1 $f10, 24($10)\r\n" "mov.s $f4, %6\r\n" "mov.s $f6, %7\r\n" "mov.s $f8, %8\r\n" "mov.s $f10, %9\r\n" "subu $9, $9, 8\r\n" PTR_ADDU "$10, $10, 32\r\n" "1: \r\n" "li $8, 4\r\n" /* st16 */ "bltu $9, $8, 2f\r\n" "nop \r\n" "gssdlc1 $f4, 7($10)\r\n" "gssdrc1 $f4, 0($10)\r\n" "gssdlc1 $f6, 7+8($10)\r\n" "gssdrc1 $f6, 8($10)\r\n" "mov.s $f4, $f8\r\n" "mov.s $f6, $f10\r\n" "subu $9, $9, 4\r\n" PTR_ADDU "$10, $10, 16\r\n" "2: \r\n" "li $8, 2\r\n" /* st8 */ "bltu $9, $8, 3f\r\n" "nop \r\n" "gssdlc1 $f4, 7($10)\r\n" "gssdrc1 $f4, 0($10)\r\n" "mov.s $f4, $f6\r\n" "subu $9, $9, 2\r\n" PTR_ADDU "$10, $10, 8\r\n" "3: \r\n" "li $8, 1\r\n" /* st4 */ "bltu $9, $8, 4f\r\n" "nop \r\n" "gsswlc1 $f4, 3($10)\r\n" "gsswrc1 $f4, 0($10)\r\n" "4: \r\n" "li %1, 0\r\n" /* end */ : "=m" (*outptr), "=r" (col) : "f" (mmA), "f" (mmB), "f" (mmC), "f" (mmD), "f" (mmE), "f" (mmF), "f" (mmG), "f" (mmH), "r" (col), "r" (outptr) : "$f4", "$f6", "$f8", "$f10", "$8", "$9", "$10", "memory" ); } #endif } if (!((output_width >> 1) & 7)) { if (output_width & 1) { cb = _mm_load_si64((__m64 *)inptr1); cr = _mm_load_si64((__m64 *)inptr2); y = _mm_load_si64((__m64 *)inptr0); decenter = 0.0; decenter = _mm_cmpeq_pi16(decenter, decenter); decenter = _mm_slli_pi16(decenter, 7); /* {0xFF80 0xFF80 0xFF80 0xFF80} */ cbl = _mm_unpacklo_pi8(cb, zero); /* Cb(0123) */ crl = _mm_unpacklo_pi8(cr, zero); /* Cr(0123) */ cbl = _mm_add_pi16(cbl, decenter); crl = _mm_add_pi16(crl, decenter); cbl2 = _mm_add_pi16(cbl, cbl); /* 2*CbL */ crl2 = _mm_add_pi16(crl, crl); /* 2*CrL */ bl = _mm_mulhi_pi16(cbl2, PW_MF0228); /* (2*CbL * -FIX(0.22800) */ rl = _mm_mulhi_pi16(crl2, PW_F0402); /* (2*CrL * FIX(0.40200)) */ bl = _mm_add_pi16(bl, PW_ONE); bl = _mm_srai_pi16(bl, 1); /* (CbL * -FIX(0.22800)) */ rl = _mm_add_pi16(rl, PW_ONE); rl = _mm_srai_pi16(rl, 1); /* (CrL * FIX(0.40200)) */ bl = _mm_add_pi16(bl, cbl); bl = _mm_add_pi16(bl, cbl); /* (CbL * FIX(1.77200))=(B-Y)L */ rl = _mm_add_pi16(rl, crl); /* (CrL * FIX(1.40200))=(R-Y)L */ gl = _mm_unpacklo_pi16(cbl, crl); gl = _mm_madd_pi16(gl, PW_MF0344_F0285); gl = _mm_add_pi32(gl, PD_ONEHALF); gl = _mm_srai_pi32(gl, SCALEBITS); gl = _mm_packs_pi32(gl, zero); /* CbL*-FIX(0.344)+CrL*FIX(0.285) */ gl = _mm_sub_pi16(gl, crl); /* CbL*-FIX(0.344)+CrL*-FIX(0.714)=(G-Y)L */ yl = _mm_unpacklo_pi8(y, zero); /* Y(0123) */ rl = _mm_add_pi16(rl, yl); /* (R0 R1 R2 R3) */ gl = _mm_add_pi16(gl, yl); /* (G0 G1 G2 G3) */ bl = _mm_add_pi16(bl, yl); /* (B0 B1 B2 B3) */ re = _mm_packs_pu16(rl, rl); ge = _mm_packs_pu16(gl, gl); be = _mm_packs_pu16(bl, bl); #if RGB_PIXELSIZE == 3 mmA = _mm_unpacklo_pi8(mmA, mmC); mmA = _mm_unpacklo_pi16(mmA, mmE); asm(".set noreorder\r\n" "move $8, %2\r\n" "mov.s $f4, %1\r\n" "mfc1 $9, $f4\r\n" "ush $9, 0($8)\r\n" "srl $9, 16\r\n" "sb $9, 2($8)\r\n" : "=m" (*outptr) : "f" (mmA), "r" (outptr) : "$f4", "$8", "$9", "memory" ); #else /* RGB_PIXELSIZE == 4 */ #ifdef RGBX_FILLER_0XFF xe = _mm_cmpeq_pi8(xe, xe); #else xe = _mm_xor_si64(xe, xe); #endif mmA = _mm_unpacklo_pi8(mmA, mmC); mmE = _mm_unpacklo_pi8(mmE, mmG); mmA = _mm_unpacklo_pi16(mmA, mmE); asm(".set noreorder\r\n" "move $8, %2\r\n" "mov.s $f4, %1\r\n" "gsswlc1 $f4, 3($8)\r\n" "gsswrc1 $f4, 0($8)\r\n" : "=m" (*outptr) : "f" (mmA), "r" (outptr) : "$f4", "$8", "memory" ); #endif } } } void jsimd_h2v2_merged_upsample_mmi(JDIMENSION output_width, JSAMPIMAGE input_buf, JDIMENSION in_row_group_ctr, JSAMPARRAY output_buf) { JSAMPROW inptr, outptr; inptr = input_buf[0][in_row_group_ctr]; outptr = output_buf[0]; input_buf[0][in_row_group_ctr] = input_buf[0][in_row_group_ctr * 2]; jsimd_h2v1_merged_upsample_mmi(output_width, input_buf, in_row_group_ctr, output_buf); input_buf[0][in_row_group_ctr] = input_buf[0][in_row_group_ctr * 2 + 1]; output_buf[0] = output_buf[1]; jsimd_h2v1_merged_upsample_mmi(output_width, input_buf, in_row_group_ctr, output_buf); input_buf[0][in_row_group_ctr] = inptr; output_buf[0] = outptr; } #undef mmA #undef mmB #undef mmC #undef mmD #undef mmE #undef mmF #undef mmG #undef mmH