summaryrefslogtreecommitdiffstats
path: root/libfreerdp/codec/rfx_neon.c
diff options
context:
space:
mode:
Diffstat (limited to 'libfreerdp/codec/rfx_neon.c')
-rw-r--r--libfreerdp/codec/rfx_neon.c536
1 files changed, 536 insertions, 0 deletions
diff --git a/libfreerdp/codec/rfx_neon.c b/libfreerdp/codec/rfx_neon.c
new file mode 100644
index 0000000..f723efd
--- /dev/null
+++ b/libfreerdp/codec/rfx_neon.c
@@ -0,0 +1,536 @@
+/*
+ FreeRDP: A Remote Desktop Protocol Implementation
+ RemoteFX Codec Library - NEON Optimizations
+
+ Copyright 2011 Martin Fleisz <martin.fleisz@thincast.com>
+
+ Licensed under the Apache License, Version 2.0 (the "License");
+ you may not use this file except in compliance with the License.
+ You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
+*/
+
+#include <freerdp/config.h>
+
+#if defined(WITH_NEON)
+
+#include <stdio.h>
+#include <stdlib.h>
+#include <string.h>
+#include <arm_neon.h>
+#include <winpr/sysinfo.h>
+
+#include "rfx_types.h"
+#include "rfx_neon.h"
+
+/* rfx_decode_YCbCr_to_RGB_NEON code now resides in the primitives library. */
+
+static __inline void __attribute__((__gnu_inline__, __always_inline__, __artificial__))
+rfx_quantization_decode_block_NEON(INT16* buffer, const size_t buffer_size, const UINT32 factor)
+{
+ int16x8_t quantFactors = vdupq_n_s16(factor);
+ int16x8_t* buf = (int16x8_t*)buffer;
+ int16x8_t* buf_end = (int16x8_t*)(buffer + buffer_size);
+
+ do
+ {
+ int16x8_t val = vld1q_s16((INT16*)buf);
+ val = vshlq_s16(val, quantFactors);
+ vst1q_s16((INT16*)buf, val);
+ buf++;
+ } while (buf < buf_end);
+}
+
+static void rfx_quantization_decode_NEON(INT16* buffer, const UINT32* WINPR_RESTRICT quantVals)
+{
+ WINPR_ASSERT(buffer);
+ WINPR_ASSERT(quantVals);
+
+ rfx_quantization_decode_block_NEON(&buffer[0], 1024, quantVals[8] - 1); /* HL1 */
+ rfx_quantization_decode_block_NEON(&buffer[1024], 1024, quantVals[7] - 1); /* LH1 */
+ rfx_quantization_decode_block_NEON(&buffer[2048], 1024, quantVals[9] - 1); /* HH1 */
+ rfx_quantization_decode_block_NEON(&buffer[3072], 256, quantVals[5] - 1); /* HL2 */
+ rfx_quantization_decode_block_NEON(&buffer[3328], 256, quantVals[4] - 1); /* LH2 */
+ rfx_quantization_decode_block_NEON(&buffer[3584], 256, quantVals[6] - 1); /* HH2 */
+ rfx_quantization_decode_block_NEON(&buffer[3840], 64, quantVals[2] - 1); /* HL3 */
+ rfx_quantization_decode_block_NEON(&buffer[3904], 64, quantVals[1] - 1); /* LH3 */
+ rfx_quantization_decode_block_NEON(&buffer[3968], 64, quantVals[3] - 1); /* HH3 */
+ rfx_quantization_decode_block_NEON(&buffer[4032], 64, quantVals[0] - 1); /* LL3 */
+}
+
+static __inline void __attribute__((__gnu_inline__, __always_inline__, __artificial__))
+rfx_dwt_2d_decode_block_horiz_NEON(INT16* WINPR_RESTRICT l, INT16* WINPR_RESTRICT h,
+ INT16* WINPR_RESTRICT dst, size_t subband_width)
+{
+ INT16* l_ptr = l;
+ INT16* h_ptr = h;
+ INT16* dst_ptr = dst;
+
+ for (size_t y = 0; y < subband_width; y++)
+ {
+ /* Even coefficients */
+ for (size_t n = 0; n < subband_width; n += 8)
+ {
+ // dst[2n] = l[n] - ((h[n-1] + h[n] + 1) >> 1);
+ int16x8_t l_n = vld1q_s16(l_ptr);
+ int16x8_t h_n = vld1q_s16(h_ptr);
+ int16x8_t h_n_m = vld1q_s16(h_ptr - 1);
+
+ if (n == 0)
+ {
+ int16_t first = vgetq_lane_s16(h_n_m, 1);
+ h_n_m = vsetq_lane_s16(first, h_n_m, 0);
+ }
+
+ int16x8_t tmp_n = vaddq_s16(h_n, h_n_m);
+ tmp_n = vaddq_s16(tmp_n, vdupq_n_s16(1));
+ tmp_n = vshrq_n_s16(tmp_n, 1);
+ int16x8_t dst_n = vsubq_s16(l_n, tmp_n);
+ vst1q_s16(l_ptr, dst_n);
+ l_ptr += 8;
+ h_ptr += 8;
+ }
+
+ l_ptr -= subband_width;
+ h_ptr -= subband_width;
+
+ /* Odd coefficients */
+ for (size_t n = 0; n < subband_width; n += 8)
+ {
+ // dst[2n + 1] = (h[n] << 1) + ((dst[2n] + dst[2n + 2]) >> 1);
+ int16x8_t h_n = vld1q_s16(h_ptr);
+ h_n = vshlq_n_s16(h_n, 1);
+ int16x8x2_t dst_n;
+ dst_n.val[0] = vld1q_s16(l_ptr);
+ int16x8_t dst_n_p = vld1q_s16(l_ptr + 1);
+
+ if (n == subband_width - 8)
+ {
+ int16_t last = vgetq_lane_s16(dst_n_p, 6);
+ dst_n_p = vsetq_lane_s16(last, dst_n_p, 7);
+ }
+
+ dst_n.val[1] = vaddq_s16(dst_n_p, dst_n.val[0]);
+ dst_n.val[1] = vshrq_n_s16(dst_n.val[1], 1);
+ dst_n.val[1] = vaddq_s16(dst_n.val[1], h_n);
+ vst2q_s16(dst_ptr, dst_n);
+ l_ptr += 8;
+ h_ptr += 8;
+ dst_ptr += 16;
+ }
+ }
+}
+
+static __inline void __attribute__((__gnu_inline__, __always_inline__, __artificial__))
+rfx_dwt_2d_decode_block_vert_NEON(INT16* WINPR_RESTRICT l, INT16* WINPR_RESTRICT h,
+ INT16* WINPR_RESTRICT dst, size_t subband_width)
+{
+ INT16* l_ptr = l;
+ INT16* h_ptr = h;
+ INT16* dst_ptr = dst;
+ const size_t total_width = subband_width + subband_width;
+
+ /* Even coefficients */
+ for (size_t n = 0; n < subband_width; n++)
+ {
+ for (size_t x = 0; x < total_width; x += 8)
+ {
+ // dst[2n] = l[n] - ((h[n-1] + h[n] + 1) >> 1);
+ int16x8_t l_n = vld1q_s16(l_ptr);
+ int16x8_t h_n = vld1q_s16(h_ptr);
+ int16x8_t tmp_n = vaddq_s16(h_n, vdupq_n_s16(1));
+
+ if (n == 0)
+ tmp_n = vaddq_s16(tmp_n, h_n);
+ else
+ {
+ int16x8_t h_n_m = vld1q_s16((h_ptr - total_width));
+ tmp_n = vaddq_s16(tmp_n, h_n_m);
+ }
+
+ tmp_n = vshrq_n_s16(tmp_n, 1);
+ int16x8_t dst_n = vsubq_s16(l_n, tmp_n);
+ vst1q_s16(dst_ptr, dst_n);
+ l_ptr += 8;
+ h_ptr += 8;
+ dst_ptr += 8;
+ }
+
+ dst_ptr += total_width;
+ }
+
+ h_ptr = h;
+ dst_ptr = dst + total_width;
+
+ /* Odd coefficients */
+ for (size_t n = 0; n < subband_width; n++)
+ {
+ for (size_t x = 0; x < total_width; x += 8)
+ {
+ // dst[2n + 1] = (h[n] << 1) + ((dst[2n] + dst[2n + 2]) >> 1);
+ int16x8_t h_n = vld1q_s16(h_ptr);
+ int16x8_t dst_n_m = vld1q_s16(dst_ptr - total_width);
+ h_n = vshlq_n_s16(h_n, 1);
+ int16x8_t tmp_n = dst_n_m;
+
+ if (n == subband_width - 1)
+ tmp_n = vaddq_s16(tmp_n, dst_n_m);
+ else
+ {
+ int16x8_t dst_n_p = vld1q_s16((dst_ptr + total_width));
+ tmp_n = vaddq_s16(tmp_n, dst_n_p);
+ }
+
+ tmp_n = vshrq_n_s16(tmp_n, 1);
+ int16x8_t dst_n = vaddq_s16(tmp_n, h_n);
+ vst1q_s16(dst_ptr, dst_n);
+ h_ptr += 8;
+ dst_ptr += 8;
+ }
+
+ dst_ptr += total_width;
+ }
+}
+
+static __inline void __attribute__((__gnu_inline__, __always_inline__, __artificial__))
+rfx_dwt_2d_decode_block_NEON(INT16* WINPR_RESTRICT buffer, INT16* WINPR_RESTRICT idwt,
+ size_t subband_width)
+{
+ INT16 *hl, *lh, *hh, *ll;
+ INT16 *l_dst, *h_dst;
+ /* Inverse DWT in horizontal direction, results in 2 sub-bands in L, H order in tmp buffer idwt.
+ */
+ /* The 4 sub-bands are stored in HL(0), LH(1), HH(2), LL(3) order. */
+ /* The lower part L uses LL(3) and HL(0). */
+ /* The higher part H uses LH(1) and HH(2). */
+ ll = buffer + subband_width * subband_width * 3;
+ hl = buffer;
+ l_dst = idwt;
+ rfx_dwt_2d_decode_block_horiz_NEON(ll, hl, l_dst, subband_width);
+ lh = buffer + subband_width * subband_width;
+ hh = buffer + subband_width * subband_width * 2;
+ h_dst = idwt + subband_width * subband_width * 2;
+ rfx_dwt_2d_decode_block_horiz_NEON(lh, hh, h_dst, subband_width);
+ /* Inverse DWT in vertical direction, results are stored in original buffer. */
+ rfx_dwt_2d_decode_block_vert_NEON(l_dst, h_dst, buffer, subband_width);
+}
+
+static void rfx_dwt_2d_decode_NEON(INT16* buffer, INT16* dwt_buffer)
+{
+ rfx_dwt_2d_decode_block_NEON(buffer + 3840, dwt_buffer, 8);
+ rfx_dwt_2d_decode_block_NEON(buffer + 3072, dwt_buffer, 16);
+ rfx_dwt_2d_decode_block_NEON(buffer, dwt_buffer, 32);
+}
+
+static INLINE void rfx_idwt_extrapolate_horiz_neon(INT16* restrict pLowBand, size_t nLowStep,
+ const INT16* restrict pHighBand,
+ size_t nHighStep, INT16* restrict pDstBand,
+ size_t nDstStep, size_t nLowCount,
+ size_t nHighCount, size_t nDstCount)
+{
+ WINPR_ASSERT(pLowBand);
+ WINPR_ASSERT(pHighBand);
+ WINPR_ASSERT(pDstBand);
+
+ INT16* l_ptr = pLowBand;
+ const INT16* h_ptr = pHighBand;
+ INT16* dst_ptr = pDstBand;
+ size_t batchSize = (nLowCount + nHighCount) >> 1;
+
+ for (size_t y = 0; y < nDstCount; y++)
+ {
+ /* Even coefficients */
+ size_t n = 0;
+ for (; n < batchSize; n += 8)
+ {
+ // dst[2n] = l[n] - ((h[n-1] + h[n] + 1) >> 1);
+ int16x8_t l_n = vld1q_s16(l_ptr);
+ int16x8_t h_n = vld1q_s16(h_ptr);
+ int16x8_t h_n_m = vld1q_s16(h_ptr - 1);
+
+ if (n == 0)
+ {
+ int16_t first = vgetq_lane_s16(h_n_m, 1);
+ h_n_m = vsetq_lane_s16(first, h_n_m, 0);
+ }
+ else if (n == 24)
+ h_n = vsetq_lane_s16(0, h_n, 7);
+
+ int16x8_t tmp_n = vaddq_s16(h_n, h_n_m);
+ tmp_n = vaddq_s16(tmp_n, vdupq_n_s16(1));
+ tmp_n = vshrq_n_s16(tmp_n, 1);
+ int16x8_t dst_n = vsubq_s16(l_n, tmp_n);
+ vst1q_s16(l_ptr, dst_n);
+ l_ptr += 8;
+ h_ptr += 8;
+ }
+ if (n < 32)
+ *l_ptr -= *(h_ptr - 1);
+
+ l_ptr -= batchSize;
+ h_ptr -= batchSize;
+
+ /* Odd coefficients */
+ n = 0;
+ for (; n < batchSize; n += 8)
+ {
+ // dst[2n + 1] = (h[n] << 1) + ((dst[2n] + dst[2n + 2]) >> 1);
+ int16x8_t h_n = vld1q_s16(h_ptr);
+ h_n = vshlq_n_s16(h_n, 1);
+ int16x8x2_t dst_n;
+ dst_n.val[0] = vld1q_s16(l_ptr);
+ int16x8_t dst_n_p = vld1q_s16(l_ptr + 1);
+
+ if (n == 24)
+ h_n = vsetq_lane_s16(0, h_n, 7);
+
+ dst_n.val[1] = vaddq_s16(dst_n_p, dst_n.val[0]);
+ dst_n.val[1] = vshrq_n_s16(dst_n.val[1], 1);
+ dst_n.val[1] = vaddq_s16(dst_n.val[1], h_n);
+ vst2q_s16(dst_ptr, dst_n);
+ l_ptr += 8;
+ h_ptr += 8;
+ dst_ptr += 16;
+ }
+ if (n == 32)
+ {
+ h_ptr -= 1;
+ l_ptr += 1;
+ }
+ else
+ {
+ *dst_ptr = *l_ptr;
+ l_ptr += 1;
+ dst_ptr += 1;
+ }
+ }
+}
+
+static INLINE void rfx_idwt_extrapolate_vert_neon(const INT16* restrict pLowBand, size_t nLowStep,
+ const INT16* restrict pHighBand, size_t nHighStep,
+ INT16* restrict pDstBand, size_t nDstStep,
+ size_t nLowCount, size_t nHighCount,
+ size_t nDstCount)
+{
+ WINPR_ASSERT(pLowBand);
+ WINPR_ASSERT(pHighBand);
+ WINPR_ASSERT(pDstBand);
+
+ const INT16* l_ptr = pLowBand;
+ const INT16* h_ptr = pHighBand;
+ INT16* dst_ptr = pDstBand;
+ size_t batchSize = (nDstCount >> 3) << 3;
+ size_t forceBandSize = (nLowCount + nHighCount) >> 1;
+
+ /* Even coefficients */
+ for (size_t n = 0; n < forceBandSize; n++)
+ {
+ for (size_t x = 0; x < batchSize; x += 8)
+ {
+ // dst[2n] = l[n] - ((h[n-1] + h[n] + 1) >> 1);
+ int16x8_t l_n = vld1q_s16(l_ptr);
+ int16x8_t h_n = vld1q_s16((n == 31) ? (h_ptr - nHighStep) : h_ptr);
+ int16x8_t tmp_n = vaddq_s16(h_n, vdupq_n_s16(1));
+
+ if (n == 0)
+ tmp_n = vaddq_s16(tmp_n, h_n);
+ else if (n < 31)
+ {
+ int16x8_t h_n_m = vld1q_s16((h_ptr - nHighStep));
+ tmp_n = vaddq_s16(tmp_n, h_n_m);
+ }
+
+ tmp_n = vshrq_n_s16(tmp_n, 1);
+ int16x8_t dst_n = vsubq_s16(l_n, tmp_n);
+ vst1q_s16(dst_ptr, dst_n);
+ l_ptr += 8;
+ h_ptr += 8;
+ dst_ptr += 8;
+ }
+
+ if (nDstCount > batchSize)
+ {
+ int16_t h_n = (n == 31) ? *(h_ptr - nHighStep) : *h_ptr;
+ int16_t tmp_n = h_n + 1;
+ if (n == 0)
+ tmp_n += h_n;
+ else if (n < 31)
+ tmp_n += *(h_ptr - nHighStep);
+ tmp_n >>= 1;
+ *dst_ptr = *l_ptr - tmp_n;
+ l_ptr += 1;
+ h_ptr += 1;
+ dst_ptr += 1;
+ }
+
+ dst_ptr += nDstStep;
+ }
+
+ if (forceBandSize < 32)
+ {
+ for (size_t x = 0; x < batchSize; x += 8)
+ {
+ int16x8_t l_n = vld1q_s16(l_ptr);
+ int16x8_t h_n = vld1q_s16(h_ptr - nHighStep);
+ int16x8_t tmp_n = vsubq_s16(l_n, h_n);
+ vst1q_s16(dst_ptr, tmp_n);
+ l_ptr += 8;
+ h_ptr += 8;
+ dst_ptr += 8;
+ }
+
+ if (nDstCount > batchSize)
+ {
+ *dst_ptr = *l_ptr - *(h_ptr - nHighStep);
+ l_ptr += 1;
+ h_ptr += 1;
+ dst_ptr += 1;
+ }
+ }
+
+ h_ptr = pHighBand;
+ dst_ptr = pDstBand + nDstStep;
+
+ /* Odd coefficients */
+ for (size_t n = 0; n < forceBandSize; n++)
+ {
+ for (size_t x = 0; x < batchSize; x += 8)
+ {
+ // dst[2n + 1] = (h[n] << 1) + ((dst[2n] + dst[2n + 2]) >> 1);
+ int16x8_t tmp_n = vld1q_s16(dst_ptr - nDstStep);
+ if (n == 31)
+ {
+ int16x8_t dst_n_p = vld1q_s16(l_ptr);
+ l_ptr += 8;
+ tmp_n = vaddq_s16(tmp_n, dst_n_p);
+ tmp_n = vshrq_n_s16(tmp_n, 1);
+ }
+ else
+ {
+ int16x8_t dst_n_p = vld1q_s16(dst_ptr + nDstStep);
+ tmp_n = vaddq_s16(tmp_n, dst_n_p);
+ tmp_n = vshrq_n_s16(tmp_n, 1);
+ int16x8_t h_n = vld1q_s16(h_ptr);
+ h_n = vshlq_n_s16(h_n, 1);
+ tmp_n = vaddq_s16(tmp_n, h_n);
+ }
+ vst1q_s16(dst_ptr, tmp_n);
+ h_ptr += 8;
+ dst_ptr += 8;
+ }
+
+ if (nDstCount > batchSize)
+ {
+ int16_t tmp_n = *(dst_ptr - nDstStep);
+ if (n == 31)
+ {
+ int16_t dst_n_p = *l_ptr;
+ l_ptr += 1;
+ tmp_n += dst_n_p;
+ tmp_n >>= 1;
+ }
+ else
+ {
+ int16_t dst_n_p = *(dst_ptr + nDstStep);
+ tmp_n += dst_n_p;
+ tmp_n >>= 1;
+ int16_t h_n = *h_ptr;
+ h_n <<= 1;
+ tmp_n += h_n;
+ }
+ *dst_ptr = tmp_n;
+ h_ptr += 1;
+ dst_ptr += 1;
+ }
+
+ dst_ptr += nDstStep;
+ }
+}
+
+static INLINE size_t prfx_get_band_l_count(size_t level)
+{
+ return (64 >> level) + 1;
+}
+
+static INLINE size_t prfx_get_band_h_count(size_t level)
+{
+ if (level == 1)
+ return (64 >> 1) - 1;
+ else
+ return (64 + (1 << (level - 1))) >> level;
+}
+
+static INLINE void rfx_dwt_2d_decode_extrapolate_block_neon(INT16* buffer, INT16* temp,
+ size_t level)
+{
+ size_t nDstStepX;
+ size_t nDstStepY;
+ INT16 *HL, *LH;
+ INT16 *HH, *LL;
+ INT16 *L, *H, *LLx;
+
+ const size_t nBandL = prfx_get_band_l_count(level);
+ const size_t nBandH = prfx_get_band_h_count(level);
+ size_t offset = 0;
+
+ WINPR_ASSERT(buffer);
+ WINPR_ASSERT(temp);
+
+ HL = &buffer[offset];
+ offset += (nBandH * nBandL);
+ LH = &buffer[offset];
+ offset += (nBandL * nBandH);
+ HH = &buffer[offset];
+ offset += (nBandH * nBandH);
+ LL = &buffer[offset];
+ nDstStepX = (nBandL + nBandH);
+ nDstStepY = (nBandL + nBandH);
+ offset = 0;
+ L = &temp[offset];
+ offset += (nBandL * nDstStepX);
+ H = &temp[offset];
+ LLx = &buffer[0];
+
+ /* horizontal (LL + HL -> L) */
+ rfx_idwt_extrapolate_horiz_neon(LL, nBandL, HL, nBandH, L, nDstStepX, nBandL, nBandH, nBandL);
+
+ /* horizontal (LH + HH -> H) */
+ rfx_idwt_extrapolate_horiz_neon(LH, nBandL, HH, nBandH, H, nDstStepX, nBandL, nBandH, nBandH);
+
+ /* vertical (L + H -> LL) */
+ rfx_idwt_extrapolate_vert_neon(L, nDstStepX, H, nDstStepX, LLx, nDstStepY, nBandL, nBandH,
+ nBandL + nBandH);
+}
+
+static void rfx_dwt_2d_extrapolate_decode_neon(INT16* buffer, INT16* temp)
+{
+ WINPR_ASSERT(buffer);
+ WINPR_ASSERT(temp);
+ rfx_dwt_2d_decode_extrapolate_block_neon(&buffer[3807], temp, 3);
+ rfx_dwt_2d_decode_extrapolate_block_neon(&buffer[3007], temp, 2);
+ rfx_dwt_2d_decode_extrapolate_block_neon(&buffer[0], temp, 1);
+}
+
+void rfx_init_neon(RFX_CONTEXT* context)
+{
+ if (IsProcessorFeaturePresent(PF_ARM_NEON_INSTRUCTIONS_AVAILABLE))
+ {
+ DEBUG_RFX("Using NEON optimizations");
+ PROFILER_RENAME(context->priv->prof_rfx_ycbcr_to_rgb, "rfx_decode_YCbCr_to_RGB_NEON");
+ PROFILER_RENAME(context->priv->prof_rfx_quantization_decode,
+ "rfx_quantization_decode_NEON");
+ PROFILER_RENAME(context->priv->prof_rfx_dwt_2d_decode, "rfx_dwt_2d_decode_NEON");
+ context->quantization_decode = rfx_quantization_decode_NEON;
+ context->dwt_2d_decode = rfx_dwt_2d_decode_NEON;
+ context->dwt_2d_extrapolate_decode = rfx_dwt_2d_extrapolate_decode_neon;
+ }
+}
+
+#endif // WITH_NEON