summaryrefslogtreecommitdiffstats
path: root/spa/plugins/audiomixer/mix-ops.h
diff options
context:
space:
mode:
Diffstat (limited to '')
-rw-r--r--spa/plugins/audiomixer/mix-ops.h169
1 files changed, 169 insertions, 0 deletions
diff --git a/spa/plugins/audiomixer/mix-ops.h b/spa/plugins/audiomixer/mix-ops.h
new file mode 100644
index 0000000..11e88dc
--- /dev/null
+++ b/spa/plugins/audiomixer/mix-ops.h
@@ -0,0 +1,169 @@
+/* Spa
+ *
+ * Copyright © 2019 Wim Taymans
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
+ * DEALINGS IN THE SOFTWARE.
+ */
+
+#include <spa/utils/defs.h>
+
+typedef struct {
+#if __BYTE_ORDER == __LITTLE_ENDIAN
+ uint8_t v3;
+ uint8_t v2;
+ uint8_t v1;
+#else
+ uint8_t v1;
+ uint8_t v2;
+ uint8_t v3;
+#endif
+} __attribute__ ((packed)) uint24_t;
+
+typedef struct {
+#if __BYTE_ORDER == __LITTLE_ENDIAN
+ uint8_t v3;
+ uint8_t v2;
+ int8_t v1;
+#else
+ int8_t v1;
+ uint8_t v2;
+ uint8_t v3;
+#endif
+} __attribute__ ((packed)) int24_t;
+
+static inline uint32_t u24_to_u32(uint24_t src)
+{
+ return ((uint32_t)src.v1 << 16) | ((uint32_t)src.v2 << 8) | (uint32_t)src.v3;
+}
+
+#define U32_TO_U24(s) (uint24_t) { .v1 = (uint8_t)(((uint32_t)s) >> 16), \
+ .v2 = (uint8_t)(((uint32_t)s) >> 8), .v3 = (uint8_t)((uint32_t)s) }
+
+static inline uint24_t u32_to_u24(uint32_t src)
+{
+ return U32_TO_U24(src);
+}
+
+static inline int32_t s24_to_s32(int24_t src)
+{
+ return ((int32_t)src.v1 << 16) | ((uint32_t)src.v2 << 8) | (uint32_t)src.v3;
+}
+
+#define S32_TO_S24(s) (int24_t) { .v1 = (int8_t)(((int32_t)s) >> 16), \
+ .v2 = (uint8_t)(((uint32_t)s) >> 8), .v3 = (uint8_t)((uint32_t)s) }
+
+static inline int24_t s32_to_s24(int32_t src)
+{
+ return S32_TO_S24(src);
+}
+
+
+#define S8_MIN -128
+#define S8_MAX 127
+#define S8_ACCUM(a,b) ((a) + (int16_t)(b))
+#define S8_CLAMP(a) (int8_t)(SPA_CLAMP((a), S8_MIN, S8_MAX))
+#define U8_OFFS 128
+#define U8_ACCUM(a,b) ((a) + ((int16_t)(b) - U8_OFFS))
+#define U8_CLAMP(a) (uint8_t)(SPA_CLAMP((a), S8_MIN, S8_MAX) + U8_OFFS)
+
+#define S16_MIN -32768
+#define S16_MAX 32767
+#define S16_ACCUM(a,b) ((a) + (int32_t)(b))
+#define S16_CLAMP(a) (int16_t)(SPA_CLAMP((a), S16_MIN, S16_MAX))
+#define U16_OFFS 32768
+#define U16_ACCUM(a,b) ((a) + ((int32_t)(b) - U16_OFFS))
+#define U16_CLAMP(a) (uint16_t)(SPA_CLAMP((a), S16_MIN, S16_MAX) + U16_OFFS)
+
+#define S24_32_MIN -8388608
+#define S24_32_MAX 8388607
+#define S24_32_ACCUM(a,b) ((a) + (int32_t)(b))
+#define S24_32_CLAMP(a) (int32_t)(SPA_CLAMP((a), S24_32_MIN, S24_32_MAX))
+#define U24_32_OFFS 8388608
+#define U24_32_ACCUM(a,b) ((a) + ((int32_t)(b) - U24_32_OFFS))
+#define U24_32_CLAMP(a) (uint32_t)(SPA_CLAMP((a), S24_32_MIN, S24_32_MAX) + U24_32_OFFS)
+
+#define S24_ACCUM(a,b) S24_32_ACCUM(a, s24_to_s32(b))
+#define S24_CLAMP(a) s32_to_s24(S24_32_CLAMP(a))
+#define U24_ACCUM(a,b) U24_32_ACCUM(a, u24_to_u32(b))
+#define U24_CLAMP(a) u32_to_u24(U24_32_CLAMP(a))
+
+#define S32_MIN -2147483648
+#define S32_MAX 2147483647
+#define S32_ACCUM(a,b) ((a) + (int64_t)(b))
+#define S32_CLAMP(a) (int32_t)(SPA_CLAMP((a), S32_MIN, S32_MAX))
+#define U32_OFFS 2147483648
+#define U32_ACCUM(a,b) ((a) + ((int64_t)(b) - U32_OFFS))
+#define U32_CLAMP(a) (uint32_t)(SPA_CLAMP((a), S32_MIN, S32_MAX) + U32_OFFS)
+
+#define F32_ACCUM(a,b) ((a) + (b))
+#define F32_CLAMP(a) (a)
+#define F64_ACCUM(a,b) ((a) + (b))
+#define F64_CLAMP(a) (a)
+
+struct mix_ops {
+ uint32_t fmt;
+ uint32_t n_channels;
+ uint32_t cpu_flags;
+
+ void (*clear) (struct mix_ops *ops, void * SPA_RESTRICT dst, uint32_t n_samples);
+ void (*process) (struct mix_ops *ops,
+ void * SPA_RESTRICT dst,
+ const void * SPA_RESTRICT src[], uint32_t n_src,
+ uint32_t n_samples);
+ void (*free) (struct mix_ops *ops);
+
+ const void *priv;
+};
+
+int mix_ops_init(struct mix_ops *ops);
+
+#define mix_ops_clear(ops,...) (ops)->clear(ops, __VA_ARGS__)
+#define mix_ops_process(ops,...) (ops)->process(ops, __VA_ARGS__)
+#define mix_ops_free(ops) (ops)->free(ops)
+
+#define DEFINE_FUNCTION(name,arch) \
+void mix_##name##_##arch(struct mix_ops *ops, void * SPA_RESTRICT dst, \
+ const void * SPA_RESTRICT src[], uint32_t n_src, \
+ uint32_t n_samples) \
+
+#define MIX_OPS_MAX_ALIGN 32
+
+DEFINE_FUNCTION(s8, c);
+DEFINE_FUNCTION(u8, c);
+DEFINE_FUNCTION(s16, c);
+DEFINE_FUNCTION(u16, c);
+DEFINE_FUNCTION(s24, c);
+DEFINE_FUNCTION(u24, c);
+DEFINE_FUNCTION(s32, c);
+DEFINE_FUNCTION(u32, c);
+DEFINE_FUNCTION(s24_32, c);
+DEFINE_FUNCTION(u24_32, c);
+DEFINE_FUNCTION(f32, c);
+DEFINE_FUNCTION(f64, c);
+
+#if defined(HAVE_SSE)
+DEFINE_FUNCTION(f32, sse);
+#endif
+#if defined(HAVE_SSE2)
+DEFINE_FUNCTION(f64, sse2);
+#endif
+#if defined(HAVE_AVX)
+DEFINE_FUNCTION(f32, avx);
+#endif