1 #![allow(unused_imports)]
2 use super::MaskElement
;
3 use crate::simd
::intrinsics
;
4 use crate::simd
::{LaneCount, Simd, SupportedLaneCount, ToBitMask}
;
5 use core
::marker
::PhantomData
;
7 /// A mask where each lane is represented by a single bit.
9 pub struct Mask
<T
, const LANES
: usize>(
10 <LaneCount
<LANES
> as SupportedLaneCount
>::BitMask
,
15 LaneCount
<LANES
>: SupportedLaneCount
;
17 impl<T
, const LANES
: usize> Copy
for Mask
<T
, LANES
>
20 LaneCount
<LANES
>: SupportedLaneCount
,
24 impl<T
, const LANES
: usize> Clone
for Mask
<T
, LANES
>
27 LaneCount
<LANES
>: SupportedLaneCount
,
29 fn clone(&self) -> Self {
34 impl<T
, const LANES
: usize> PartialEq
for Mask
<T
, LANES
>
37 LaneCount
<LANES
>: SupportedLaneCount
,
39 fn eq(&self, other
: &Self) -> bool
{
40 self.0.as_ref() == other
.0.as_ref()
44 impl<T
, const LANES
: usize> PartialOrd
for Mask
<T
, LANES
>
47 LaneCount
<LANES
>: SupportedLaneCount
,
49 fn partial_cmp(&self, other
: &Self) -> Option
<core
::cmp
::Ordering
> {
50 self.0.as_ref().partial_cmp(other
.0.as_ref())
54 impl<T
, const LANES
: usize> Eq
for Mask
<T
, LANES
>
57 LaneCount
<LANES
>: SupportedLaneCount
,
61 impl<T
, const LANES
: usize> Ord
for Mask
<T
, LANES
>
64 LaneCount
<LANES
>: SupportedLaneCount
,
66 fn cmp(&self, other
: &Self) -> core
::cmp
::Ordering
{
67 self.0.as_ref().cmp(other
.0.as_ref())
71 impl<T
, const LANES
: usize> Mask
<T
, LANES
>
74 LaneCount
<LANES
>: SupportedLaneCount
,
77 #[must_use = "method returns a new mask and does not mutate the original value"]
78 pub fn splat(value
: bool
) -> Self {
79 let mut mask
= <LaneCount
<LANES
> as SupportedLaneCount
>::BitMask
::default();
81 mask
.as_mut().fill(u8::MAX
)
83 mask
.as_mut().fill(u8::MIN
)
86 *mask
.as_mut().last_mut().unwrap() &= u8::MAX
>> (8 - LANES
% 8);
88 Self(mask
, PhantomData
)
92 #[must_use = "method returns a new bool and does not mutate the original value"]
93 pub unsafe fn test_unchecked(&self, lane
: usize) -> bool
{
94 (self.0.as_ref()[lane
/ 8] >> (lane
% 8)) & 0x1 > 0
98 pub unsafe fn set_unchecked(&mut self, lane
: usize, value
: bool
) {
100 self.0.as_mut()[lane
/ 8] ^
= ((value ^
self.test_unchecked(lane
)) as u8) << (lane
% 8)
105 #[must_use = "method returns a new vector and does not mutate the original value"]
106 pub fn to_int(self) -> Simd
<T
, LANES
> {
108 intrinsics
::simd_select_bitmask(self.0, Simd
::splat(T
::TRUE
), Simd
::splat(T
::FALSE
))
113 #[must_use = "method returns a new mask and does not mutate the original value"]
114 pub unsafe fn from_int_unchecked(value
: Simd
<T
, LANES
>) -> Self {
115 unsafe { Self(intrinsics::simd_bitmask(value), PhantomData) }
119 pub fn to_bitmask_integer
<U
>(self) -> U
121 super::Mask
<T
, LANES
>: ToBitMask
<BitMask
= U
>,
123 // Safety: these are the same types
124 unsafe { core::mem::transmute_copy(&self.0) }
128 pub fn from_bitmask_integer
<U
>(bitmask
: U
) -> Self
130 super::Mask
<T
, LANES
>: ToBitMask
<BitMask
= U
>,
132 // Safety: these are the same types
133 unsafe { Self(core::mem::transmute_copy(&bitmask), PhantomData) }
137 #[must_use = "method returns a new mask and does not mutate the original value"]
138 pub fn convert
<U
>(self) -> Mask
<U
, LANES
>
142 // Safety: bitmask layout does not depend on the element width
143 unsafe { core::mem::transmute_copy(&self) }
147 #[must_use = "method returns a new bool and does not mutate the original value"]
148 pub fn any(self) -> bool
{
149 self != Self::splat(false)
153 #[must_use = "method returns a new bool and does not mutate the original value"]
154 pub fn all(self) -> bool
{
155 self == Self::splat(true)
159 impl<T
, const LANES
: usize> core
::ops
::BitAnd
for Mask
<T
, LANES
>
162 LaneCount
<LANES
>: SupportedLaneCount
,
163 <LaneCount
<LANES
> as SupportedLaneCount
>::BitMask
: AsRef
<[u8]> + AsMut
<[u8]>,
167 #[must_use = "method returns a new mask and does not mutate the original value"]
168 fn bitand(mut self, rhs
: Self) -> Self {
169 for (l
, r
) in self.0.as_mut().iter_mut().zip(rhs
.0.as_ref().iter()) {
176 impl<T
, const LANES
: usize> core
::ops
::BitOr
for Mask
<T
, LANES
>
179 LaneCount
<LANES
>: SupportedLaneCount
,
180 <LaneCount
<LANES
> as SupportedLaneCount
>::BitMask
: AsRef
<[u8]> + AsMut
<[u8]>,
184 #[must_use = "method returns a new mask and does not mutate the original value"]
185 fn bitor(mut self, rhs
: Self) -> Self {
186 for (l
, r
) in self.0.as_mut().iter_mut().zip(rhs
.0.as_ref().iter()) {
193 impl<T
, const LANES
: usize> core
::ops
::BitXor
for Mask
<T
, LANES
>
196 LaneCount
<LANES
>: SupportedLaneCount
,
200 #[must_use = "method returns a new mask and does not mutate the original value"]
201 fn bitxor(mut self, rhs
: Self) -> Self::Output
{
202 for (l
, r
) in self.0.as_mut().iter_mut().zip(rhs
.0.as_ref().iter()) {
209 impl<T
, const LANES
: usize> core
::ops
::Not
for Mask
<T
, LANES
>
212 LaneCount
<LANES
>: SupportedLaneCount
,
216 #[must_use = "method returns a new mask and does not mutate the original value"]
217 fn not(mut self) -> Self::Output
{
218 for x
in self.0.as_mut() {
222 *self.0.as_mut().last_mut().unwrap() &= u8::MAX
>> (8 - LANES
% 8);