1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
|
/***************************************************************************
* Copyright (c) Johan Mabille, Sylvain Corlay, Wolf Vollprecht and *
* Martin Renou *
* Copyright (c) QuantStack *
* Copyright (c) Serge Guelton *
* *
* Distributed under the terms of the BSD 3-Clause License. *
* *
* The full license is in the file LICENSE, distributed with this software. *
****************************************************************************/
#ifndef XSIMD_GENERIC_LOGICAL_HPP
#define XSIMD_GENERIC_LOGICAL_HPP
#include "./xsimd_generic_details.hpp"
namespace xsimd
{
namespace kernel
{
using namespace types;
// from mask
template <class A, class T>
inline batch_bool<T, A> from_mask(batch_bool<T, A> const&, uint64_t mask, requires_arch<generic>) noexcept
{
alignas(A::alignment()) bool buffer[batch_bool<T, A>::size];
// This is inefficient but should never be called. It's just a
// temporary implementation until arm support is added.
for (size_t i = 0; i < batch_bool<T, A>::size; ++i)
buffer[i] = mask & (1ull << i);
return batch_bool<T, A>::load_aligned(buffer);
}
// ge
template <class A, class T>
inline batch_bool<T, A> ge(batch<T, A> const& self, batch<T, A> const& other, requires_arch<generic>) noexcept
{
return other <= self;
}
// gt
template <class A, class T>
inline batch_bool<T, A> gt(batch<T, A> const& self, batch<T, A> const& other, requires_arch<generic>) noexcept
{
return other < self;
}
// is_even
template <class A, class T>
inline batch_bool<T, A> is_even(batch<T, A> const& self, requires_arch<generic>) noexcept
{
return is_flint(self * T(0.5));
}
// is_flint
template <class A, class T>
inline batch_bool<T, A> is_flint(batch<T, A> const& self, requires_arch<generic>) noexcept
{
auto frac = select(isnan(self - self), constants::nan<batch<T, A>>(), self - trunc(self));
return frac == T(0.);
}
// is_odd
template <class A, class T>
inline batch_bool<T, A> is_odd(batch<T, A> const& self, requires_arch<generic>) noexcept
{
return is_even(self - T(1.));
}
// isinf
template <class A, class T, class = typename std::enable_if<std::is_integral<T>::value, void>::type>
inline batch_bool<T, A> isinf(batch<T, A> const&, requires_arch<generic>) noexcept
{
return batch_bool<T, A>(false);
}
template <class A>
inline batch_bool<float, A> isinf(batch<float, A> const& self, requires_arch<generic>) noexcept
{
return abs(self) == std::numeric_limits<float>::infinity();
}
template <class A>
inline batch_bool<double, A> isinf(batch<double, A> const& self, requires_arch<generic>) noexcept
{
return abs(self) == std::numeric_limits<double>::infinity();
}
// isfinite
template <class A, class T, class = typename std::enable_if<std::is_integral<T>::value, void>::type>
inline batch_bool<T, A> isfinite(batch<T, A> const&, requires_arch<generic>) noexcept
{
return batch_bool<T, A>(true);
}
template <class A>
inline batch_bool<float, A> isfinite(batch<float, A> const& self, requires_arch<generic>) noexcept
{
return (self - self) == 0.f;
}
template <class A>
inline batch_bool<double, A> isfinite(batch<double, A> const& self, requires_arch<generic>) noexcept
{
return (self - self) == 0.;
}
// isnan
template <class A, class T, class = typename std::enable_if<std::is_integral<T>::value, void>::type>
inline batch_bool<T, A> isnan(batch<T, A> const&, requires_arch<generic>) noexcept
{
return batch_bool<T, A>(false);
}
// le
template <class A, class T, class = typename std::enable_if<std::is_integral<T>::value, void>::type>
inline batch_bool<T, A> le(batch<T, A> const& self, batch<T, A> const& other, requires_arch<generic>) noexcept
{
return (self < other) || (self == other);
}
// neq
template <class A, class T>
inline batch_bool<T, A> neq(batch<T, A> const& self, batch<T, A> const& other, requires_arch<generic>) noexcept
{
return !(other == self);
}
// logical_and
template <class A, class T>
inline batch<T, A> logical_and(batch<T, A> const& self, batch<T, A> const& other, requires_arch<generic>) noexcept
{
return detail::apply([](T x, T y) noexcept
{ return x && y; },
self, other);
}
// logical_or
template <class A, class T>
inline batch<T, A> logical_or(batch<T, A> const& self, batch<T, A> const& other, requires_arch<generic>) noexcept
{
return detail::apply([](T x, T y) noexcept
{ return x || y; },
self, other);
}
// mask
template <class A, class T>
inline uint64_t mask(batch_bool<T, A> const& self, requires_arch<generic>) noexcept
{
alignas(A::alignment()) bool buffer[batch_bool<T, A>::size];
self.store_aligned(buffer);
// This is inefficient but should never be called. It's just a
// temporary implementation until arm support is added.
uint64_t res = 0;
for (size_t i = 0; i < batch_bool<T, A>::size; ++i)
if (buffer[i])
res |= 1ul << i;
return res;
}
}
}
#endif
|