summaryrefslogtreecommitdiffstats
path: root/third_party/xsimd/include/xsimd/arch/xsimd_fma3_avx.hpp
diff options
context:
space:
mode:
Diffstat (limited to 'third_party/xsimd/include/xsimd/arch/xsimd_fma3_avx.hpp')
-rw-r--r--third_party/xsimd/include/xsimd/arch/xsimd_fma3_avx.hpp80
1 files changed, 80 insertions, 0 deletions
diff --git a/third_party/xsimd/include/xsimd/arch/xsimd_fma3_avx.hpp b/third_party/xsimd/include/xsimd/arch/xsimd_fma3_avx.hpp
new file mode 100644
index 0000000000..64e9ed65d1
--- /dev/null
+++ b/third_party/xsimd/include/xsimd/arch/xsimd_fma3_avx.hpp
@@ -0,0 +1,80 @@
+/***************************************************************************
+ * Copyright (c) Johan Mabille, Sylvain Corlay, Wolf Vollprecht and *
+ * Martin Renou *
+ * Copyright (c) QuantStack *
+ * Copyright (c) Serge Guelton *
+ * *
+ * Distributed under the terms of the BSD 3-Clause License. *
+ * *
+ * The full license is in the file LICENSE, distributed with this software. *
+ ****************************************************************************/
+
+#ifndef XSIMD_FMA3_AVX_HPP
+#define XSIMD_FMA3_AVX_HPP
+
+#include "../types/xsimd_fma3_avx_register.hpp"
+
+namespace xsimd
+{
+
+ namespace kernel
+ {
+ using namespace types;
+
+ // fnma
+ template <class A>
+ inline batch<float, A> fnma(batch<float, A> const& x, batch<float, A> const& y, batch<float, A> const& z, requires_arch<fma3<avx>>) noexcept
+ {
+ return _mm256_fnmadd_ps(x, y, z);
+ }
+
+ template <class A>
+ inline batch<double, A> fnma(batch<double, A> const& x, batch<double, A> const& y, batch<double, A> const& z, requires_arch<fma3<avx>>) noexcept
+ {
+ return _mm256_fnmadd_pd(x, y, z);
+ }
+
+ // fnms
+ template <class A>
+ inline batch<float, A> fnms(batch<float, A> const& x, batch<float, A> const& y, batch<float, A> const& z, requires_arch<fma3<avx>>) noexcept
+ {
+ return _mm256_fnmsub_ps(x, y, z);
+ }
+
+ template <class A>
+ inline batch<double, A> fnms(batch<double, A> const& x, batch<double, A> const& y, batch<double, A> const& z, requires_arch<fma3<avx>>) noexcept
+ {
+ return _mm256_fnmsub_pd(x, y, z);
+ }
+
+ // fma
+ template <class A>
+ inline batch<float, A> fma(batch<float, A> const& x, batch<float, A> const& y, batch<float, A> const& z, requires_arch<fma3<avx>>) noexcept
+ {
+ return _mm256_fmadd_ps(x, y, z);
+ }
+
+ template <class A>
+ inline batch<double, A> fma(batch<double, A> const& x, batch<double, A> const& y, batch<double, A> const& z, requires_arch<fma3<avx>>) noexcept
+ {
+ return _mm256_fmadd_pd(x, y, z);
+ }
+
+ // fms
+ template <class A>
+ inline batch<float, A> fms(batch<float, A> const& x, batch<float, A> const& y, batch<float, A> const& z, requires_arch<fma3<avx>>) noexcept
+ {
+ return _mm256_fmsub_ps(x, y, z);
+ }
+
+ template <class A>
+ inline batch<double, A> fms(batch<double, A> const& x, batch<double, A> const& y, batch<double, A> const& z, requires_arch<fma3<avx>>) noexcept
+ {
+ return _mm256_fmsub_pd(x, y, z);
+ }
+
+ }
+
+}
+
+#endif