//! LLVM bit manipulation intrinsics. #[rustfmt::skip] use crate::*; #[allow(improper_ctypes, dead_code)] extern "C" { #[link_name = "llvm.ctlz.v2i8"] fn ctlz_u8x2(x: u8x2, is_zero_undef: bool) -> u8x2; #[link_name = "llvm.ctlz.v4i8"] fn ctlz_u8x4(x: u8x4, is_zero_undef: bool) -> u8x4; #[link_name = "llvm.ctlz.v8i8"] fn ctlz_u8x8(x: u8x8, is_zero_undef: bool) -> u8x8; #[link_name = "llvm.ctlz.v16i8"] fn ctlz_u8x16(x: u8x16, is_zero_undef: bool) -> u8x16; #[link_name = "llvm.ctlz.v32i8"] fn ctlz_u8x32(x: u8x32, is_zero_undef: bool) -> u8x32; #[link_name = "llvm.ctlz.v64i8"] fn ctlz_u8x64(x: u8x64, is_zero_undef: bool) -> u8x64; #[link_name = "llvm.ctlz.v2i16"] fn ctlz_u16x2(x: u16x2, is_zero_undef: bool) -> u16x2; #[link_name = "llvm.ctlz.v4i16"] fn ctlz_u16x4(x: u16x4, is_zero_undef: bool) -> u16x4; #[link_name = "llvm.ctlz.v8i16"] fn ctlz_u16x8(x: u16x8, is_zero_undef: bool) -> u16x8; #[link_name = "llvm.ctlz.v16i16"] fn ctlz_u16x16(x: u16x16, is_zero_undef: bool) -> u16x16; #[link_name = "llvm.ctlz.v32i16"] fn ctlz_u16x32(x: u16x32, is_zero_undef: bool) -> u16x32; #[link_name = "llvm.ctlz.v2i32"] fn ctlz_u32x2(x: u32x2, is_zero_undef: bool) -> u32x2; #[link_name = "llvm.ctlz.v4i32"] fn ctlz_u32x4(x: u32x4, is_zero_undef: bool) -> u32x4; #[link_name = "llvm.ctlz.v8i32"] fn ctlz_u32x8(x: u32x8, is_zero_undef: bool) -> u32x8; #[link_name = "llvm.ctlz.v16i32"] fn ctlz_u32x16(x: u32x16, is_zero_undef: bool) -> u32x16; #[link_name = "llvm.ctlz.v2i64"] fn ctlz_u64x2(x: u64x2, is_zero_undef: bool) -> u64x2; #[link_name = "llvm.ctlz.v4i64"] fn ctlz_u64x4(x: u64x4, is_zero_undef: bool) -> u64x4; #[link_name = "llvm.ctlz.v8i64"] fn ctlz_u64x8(x: u64x8, is_zero_undef: bool) -> u64x8; #[link_name = "llvm.ctlz.v1i128"] fn ctlz_u128x1(x: u128x1, is_zero_undef: bool) -> u128x1; #[link_name = "llvm.ctlz.v2i128"] fn ctlz_u128x2(x: u128x2, is_zero_undef: bool) -> u128x2; #[link_name = "llvm.ctlz.v4i128"] fn ctlz_u128x4(x: u128x4, is_zero_undef: bool) -> u128x4; #[link_name = "llvm.cttz.v2i8"] fn cttz_u8x2(x: u8x2, is_zero_undef: bool) -> u8x2; #[link_name = "llvm.cttz.v4i8"] fn cttz_u8x4(x: u8x4, is_zero_undef: bool) -> u8x4; #[link_name = "llvm.cttz.v8i8"] fn cttz_u8x8(x: u8x8, is_zero_undef: bool) -> u8x8; #[link_name = "llvm.cttz.v16i8"] fn cttz_u8x16(x: u8x16, is_zero_undef: bool) -> u8x16; #[link_name = "llvm.cttz.v32i8"] fn cttz_u8x32(x: u8x32, is_zero_undef: bool) -> u8x32; #[link_name = "llvm.cttz.v64i8"] fn cttz_u8x64(x: u8x64, is_zero_undef: bool) -> u8x64; #[link_name = "llvm.cttz.v2i16"] fn cttz_u16x2(x: u16x2, is_zero_undef: bool) -> u16x2; #[link_name = "llvm.cttz.v4i16"] fn cttz_u16x4(x: u16x4, is_zero_undef: bool) -> u16x4; #[link_name = "llvm.cttz.v8i16"] fn cttz_u16x8(x: u16x8, is_zero_undef: bool) -> u16x8; #[link_name = "llvm.cttz.v16i16"] fn cttz_u16x16(x: u16x16, is_zero_undef: bool) -> u16x16; #[link_name = "llvm.cttz.v32i16"] fn cttz_u16x32(x: u16x32, is_zero_undef: bool) -> u16x32; #[link_name = "llvm.cttz.v2i32"] fn cttz_u32x2(x: u32x2, is_zero_undef: bool) -> u32x2; #[link_name = "llvm.cttz.v4i32"] fn cttz_u32x4(x: u32x4, is_zero_undef: bool) -> u32x4; #[link_name = "llvm.cttz.v8i32"] fn cttz_u32x8(x: u32x8, is_zero_undef: bool) -> u32x8; #[link_name = "llvm.cttz.v16i32"] fn cttz_u32x16(x: u32x16, is_zero_undef: bool) -> u32x16; #[link_name = "llvm.cttz.v2i64"] fn cttz_u64x2(x: u64x2, is_zero_undef: bool) -> u64x2; #[link_name = "llvm.cttz.v4i64"] fn cttz_u64x4(x: u64x4, is_zero_undef: bool) -> u64x4; #[link_name = "llvm.cttz.v8i64"] fn cttz_u64x8(x: u64x8, is_zero_undef: bool) -> u64x8; #[link_name = "llvm.cttz.v1i128"] fn cttz_u128x1(x: u128x1, is_zero_undef: bool) -> u128x1; #[link_name = "llvm.cttz.v2i128"] fn cttz_u128x2(x: u128x2, is_zero_undef: bool) -> u128x2; #[link_name = "llvm.cttz.v4i128"] fn cttz_u128x4(x: u128x4, is_zero_undef: bool) -> u128x4; #[link_name = "llvm.ctpop.v2i8"] fn ctpop_u8x2(x: u8x2) -> u8x2; #[link_name = "llvm.ctpop.v4i8"] fn ctpop_u8x4(x: u8x4) -> u8x4; #[link_name = "llvm.ctpop.v8i8"] fn ctpop_u8x8(x: u8x8) -> u8x8; #[link_name = "llvm.ctpop.v16i8"] fn ctpop_u8x16(x: u8x16) -> u8x16; #[link_name = "llvm.ctpop.v32i8"] fn ctpop_u8x32(x: u8x32) -> u8x32; #[link_name = "llvm.ctpop.v64i8"] fn ctpop_u8x64(x: u8x64) -> u8x64; #[link_name = "llvm.ctpop.v2i16"] fn ctpop_u16x2(x: u16x2) -> u16x2; #[link_name = "llvm.ctpop.v4i16"] fn ctpop_u16x4(x: u16x4) -> u16x4; #[link_name = "llvm.ctpop.v8i16"] fn ctpop_u16x8(x: u16x8) -> u16x8; #[link_name = "llvm.ctpop.v16i16"] fn ctpop_u16x16(x: u16x16) -> u16x16; #[link_name = "llvm.ctpop.v32i16"] fn ctpop_u16x32(x: u16x32) -> u16x32; #[link_name = "llvm.ctpop.v2i32"] fn ctpop_u32x2(x: u32x2) -> u32x2; #[link_name = "llvm.ctpop.v4i32"] fn ctpop_u32x4(x: u32x4) -> u32x4; #[link_name = "llvm.ctpop.v8i32"] fn ctpop_u32x8(x: u32x8) -> u32x8; #[link_name = "llvm.ctpop.v16i32"] fn ctpop_u32x16(x: u32x16) -> u32x16; #[link_name = "llvm.ctpop.v2i64"] fn ctpop_u64x2(x: u64x2) -> u64x2; #[link_name = "llvm.ctpop.v4i64"] fn ctpop_u64x4(x: u64x4) -> u64x4; #[link_name = "llvm.ctpop.v8i64"] fn ctpop_u64x8(x: u64x8) -> u64x8; #[link_name = "llvm.ctpop.v1i128"] fn ctpop_u128x1(x: u128x1) -> u128x1; #[link_name = "llvm.ctpop.v2i128"] fn ctpop_u128x2(x: u128x2) -> u128x2; #[link_name = "llvm.ctpop.v4i128"] fn ctpop_u128x4(x: u128x4) -> u128x4; } crate trait BitManip { fn ctpop(self) -> Self; fn ctlz(self) -> Self; fn cttz(self) -> Self; } macro_rules! impl_bit_manip { (inner: $ty:ident, $scalar:ty, $uty:ident, $ctpop:ident, $ctlz:ident, $cttz:ident) => { // FIXME: several LLVM intrinsics break on s390x https://github.com/rust-lang-nursery/packed_simd/issues/192 #[cfg(target_arch = "s390x")] impl_bit_manip! { scalar: $ty, $scalar } #[cfg(not(target_arch = "s390x"))] impl BitManip for $ty { #[inline] fn ctpop(self) -> Self { let y: $uty = self.cast(); unsafe { $ctpop(y).cast() } } #[inline] fn ctlz(self) -> Self { let y: $uty = self.cast(); // the ctxx intrinsics need compile-time constant // `is_zero_undef` unsafe { $ctlz(y, false).cast() } } #[inline] fn cttz(self) -> Self { let y: $uty = self.cast(); unsafe { $cttz(y, false).cast() } } } }; (sized_inner: $ty:ident, $scalar:ty, $uty:ident) => { #[cfg(target_arch = "s390x")] impl_bit_manip! { scalar: $ty, $scalar } #[cfg(not(target_arch = "s390x"))] impl BitManip for $ty { #[inline] fn ctpop(self) -> Self { let y: $uty = self.cast(); $uty::ctpop(y).cast() } #[inline] fn ctlz(self) -> Self { let y: $uty = self.cast(); $uty::ctlz(y).cast() } #[inline] fn cttz(self) -> Self { let y: $uty = self.cast(); $uty::cttz(y).cast() } } }; (scalar: $ty:ident, $scalar:ty) => { impl BitManip for $ty { #[inline] fn ctpop(self) -> Self { let mut ones = self; for i in 0..Self::lanes() { ones = ones .replace(i, self.extract(i).count_ones() as $scalar); } ones } #[inline] fn ctlz(self) -> Self { let mut lz = self; for i in 0..Self::lanes() { lz = lz.replace( i, self.extract(i).leading_zeros() as $scalar, ); } lz } #[inline] fn cttz(self) -> Self { let mut tz = self; for i in 0..Self::lanes() { tz = tz.replace( i, self.extract(i).trailing_zeros() as $scalar, ); } tz } } }; ($uty:ident, $uscalar:ty, $ity:ident, $iscalar:ty, $ctpop:ident, $ctlz:ident, $cttz:ident) => { impl_bit_manip! { inner: $uty, $uscalar, $uty, $ctpop, $ctlz, $cttz } impl_bit_manip! { inner: $ity, $iscalar, $uty, $ctpop, $ctlz, $cttz } }; (sized: $usize:ident, $uscalar:ty, $isize:ident, $iscalar:ty, $ty:ident) => { impl_bit_manip! { sized_inner: $usize, $uscalar, $ty } impl_bit_manip! { sized_inner: $isize, $iscalar, $ty } }; } impl_bit_manip! { u8x2 , u8, i8x2, i8, ctpop_u8x2, ctlz_u8x2, cttz_u8x2 } impl_bit_manip! { u8x4 , u8, i8x4, i8, ctpop_u8x4, ctlz_u8x4, cttz_u8x4 } #[cfg(not(target_arch = "aarch64"))] // see below impl_bit_manip! { u8x8 , u8, i8x8, i8, ctpop_u8x8, ctlz_u8x8, cttz_u8x8 } impl_bit_manip! { u8x16 , u8, i8x16, i8, ctpop_u8x16, ctlz_u8x16, cttz_u8x16 } impl_bit_manip! { u8x32 , u8, i8x32, i8, ctpop_u8x32, ctlz_u8x32, cttz_u8x32 } impl_bit_manip! { u8x64 , u8, i8x64, i8, ctpop_u8x64, ctlz_u8x64, cttz_u8x64 } impl_bit_manip! { u16x2 , u16, i16x2, i16, ctpop_u16x2, ctlz_u16x2, cttz_u16x2 } impl_bit_manip! { u16x4 , u16, i16x4, i16, ctpop_u16x4, ctlz_u16x4, cttz_u16x4 } impl_bit_manip! { u16x8 , u16, i16x8, i16, ctpop_u16x8, ctlz_u16x8, cttz_u16x8 } impl_bit_manip! { u16x16 , u16, i16x16, i16, ctpop_u16x16, ctlz_u16x16, cttz_u16x16 } impl_bit_manip! { u16x32 , u16, i16x32, i16, ctpop_u16x32, ctlz_u16x32, cttz_u16x32 } impl_bit_manip! { u32x2 , u32, i32x2, i32, ctpop_u32x2, ctlz_u32x2, cttz_u32x2 } impl_bit_manip! { u32x4 , u32, i32x4, i32, ctpop_u32x4, ctlz_u32x4, cttz_u32x4 } impl_bit_manip! { u32x8 , u32, i32x8, i32, ctpop_u32x8, ctlz_u32x8, cttz_u32x8 } impl_bit_manip! { u32x16 , u32, i32x16, i32, ctpop_u32x16, ctlz_u32x16, cttz_u32x16 } impl_bit_manip! { u64x2 , u64, i64x2, i64, ctpop_u64x2, ctlz_u64x2, cttz_u64x2 } impl_bit_manip! { u64x4 , u64, i64x4, i64, ctpop_u64x4, ctlz_u64x4, cttz_u64x4 } impl_bit_manip! { u64x8 , u64, i64x8, i64, ctpop_u64x8, ctlz_u64x8, cttz_u64x8 } impl_bit_manip! { u128x1 , u128, i128x1, i128, ctpop_u128x1, ctlz_u128x1, cttz_u128x1 } impl_bit_manip! { u128x2 , u128, i128x2, i128, ctpop_u128x2, ctlz_u128x2, cttz_u128x2 } impl_bit_manip! { u128x4 , u128, i128x4, i128, ctpop_u128x4, ctlz_u128x4, cttz_u128x4 } #[cfg(target_arch = "aarch64")] impl BitManip for u8x8 { #[inline] fn ctpop(self) -> Self { let y: u8x8 = self.cast(); unsafe { ctpop_u8x8(y).cast() } } #[inline] fn ctlz(self) -> Self { let y: u8x8 = self.cast(); unsafe { ctlz_u8x8(y, false).cast() } } #[inline] fn cttz(self) -> Self { // FIXME: LLVM cttz.v8i8 broken on aarch64 https://github.com/rust-lang-nursery/packed_simd/issues/191 // OPTIMIZE: adapt the algorithm used for v8i16/etc to Rust's aarch64 // intrinsics let mut tz = self; for i in 0..Self::lanes() { tz = tz.replace(i, self.extract(i).trailing_zeros() as u8); } tz } } #[cfg(target_arch = "aarch64")] impl BitManip for i8x8 { #[inline] fn ctpop(self) -> Self { let y: u8x8 = self.cast(); unsafe { ctpop_u8x8(y).cast() } } #[inline] fn ctlz(self) -> Self { let y: u8x8 = self.cast(); unsafe { ctlz_u8x8(y, false).cast() } } #[inline] fn cttz(self) -> Self { // FIXME: LLVM cttz.v8i8 broken on aarch64 https://github.com/rust-lang-nursery/packed_simd/issues/191 // OPTIMIZE: adapt the algorithm used for v8i16/etc to Rust's aarch64 // intrinsics let mut tz = self; for i in 0..Self::lanes() { tz = tz.replace(i, self.extract(i).trailing_zeros() as i8); } tz } } cfg_if! { if #[cfg(target_pointer_width = "8")] { impl_bit_manip! { sized: usizex2, usize, isizex2, isize, u8x2 } impl_bit_manip! { sized: usizex4, usize, isizex4, isize, u8x4 } impl_bit_manip! { sized: usizex8, usize, isizex8, isize, u8x8 } } else if #[cfg(target_pointer_width = "16")] { impl_bit_manip! { sized: usizex2, usize, isizex2, isize, u16x2 } impl_bit_manip! { sized: usizex4, usize, isizex4, isize, u16x4 } impl_bit_manip! { sized: usizex8, usize, isizex8, isize, u16x8 } } else if #[cfg(target_pointer_width = "32")] { impl_bit_manip! { sized: usizex2, usize, isizex2, isize, u32x2 } impl_bit_manip! { sized: usizex4, usize, isizex4, isize, u32x4 } impl_bit_manip! { sized: usizex8, usize, isizex8, isize, u32x8 } } else if #[cfg(target_pointer_width = "64")] { impl_bit_manip! { sized: usizex2, usize, isizex2, isize, u64x2 } impl_bit_manip! { sized: usizex4, usize, isizex4, isize, u64x4 } impl_bit_manip! { sized: usizex8, usize, isizex8, isize, u64x8 } } else { compile_error!("unsupported target_pointer_width"); } }