diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 19:33:14 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 19:33:14 +0000 |
commit | 36d22d82aa202bb199967e9512281e9a53db42c9 (patch) | |
tree | 105e8c98ddea1c1e4784a60a5a6410fa416be2de /third_party/rust/hashlink/src | |
parent | Initial commit. (diff) | |
download | firefox-esr-36d22d82aa202bb199967e9512281e9a53db42c9.tar.xz firefox-esr-36d22d82aa202bb199967e9512281e9a53db42c9.zip |
Adding upstream version 115.7.0esr.upstream/115.7.0esrupstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to '')
-rw-r--r-- | third_party/rust/hashlink/src/lib.rs | 9 | ||||
-rw-r--r-- | third_party/rust/hashlink/src/linked_hash_map.rs | 2179 | ||||
-rw-r--r-- | third_party/rust/hashlink/src/linked_hash_set.rs | 766 | ||||
-rw-r--r-- | third_party/rust/hashlink/src/lru_cache.rs | 292 | ||||
-rw-r--r-- | third_party/rust/hashlink/src/serde.rs | 161 |
5 files changed, 3407 insertions, 0 deletions
diff --git a/third_party/rust/hashlink/src/lib.rs b/third_party/rust/hashlink/src/lib.rs new file mode 100644 index 0000000000..55bdcd2ef7 --- /dev/null +++ b/third_party/rust/hashlink/src/lib.rs @@ -0,0 +1,9 @@ +pub mod linked_hash_map; +pub mod linked_hash_set; +pub mod lru_cache; +#[cfg(feature = "serde_impl")] +pub mod serde; + +pub use linked_hash_map::LinkedHashMap; +pub use linked_hash_set::LinkedHashSet; +pub use lru_cache::LruCache; diff --git a/third_party/rust/hashlink/src/linked_hash_map.rs b/third_party/rust/hashlink/src/linked_hash_map.rs new file mode 100644 index 0000000000..b27c98b82b --- /dev/null +++ b/third_party/rust/hashlink/src/linked_hash_map.rs @@ -0,0 +1,2179 @@ +use std::{ + alloc::Layout, + borrow::Borrow, + cmp::Ordering, + fmt, + hash::{BuildHasher, Hash, Hasher}, + iter::FromIterator, + marker::PhantomData, + mem::{self, MaybeUninit}, + ops::{Index, IndexMut}, + ptr::{self, NonNull}, +}; + +use hashbrown::{hash_map, HashMap}; + +pub enum TryReserveError { + CapacityOverflow, + AllocError { layout: Layout }, +} + +/// A version of `HashMap` that has a user controllable order for its entries. +/// +/// It achieves this by keeping its entries in an internal linked list and using a `HashMap` to +/// point at nodes in this linked list. +/// +/// The order of entries defaults to "insertion order", but the user can also modify the order of +/// existing entries by manually moving them to the front or back. +/// +/// There are two kinds of methods that modify the order of the internal list: +/// +/// * Methods that have names like `to_front` and `to_back` will unsurprisingly move an existing +/// entry to the front or back +/// * Methods that have the word `insert` will insert a new entry ot the back of the list, and if +/// that method might replace an entry, that method will *also move that existing entry to the +/// back*. +pub struct LinkedHashMap<K, V, S = hash_map::DefaultHashBuilder> { + map: HashMap<NonNull<Node<K, V>>, (), NullHasher>, + // We need to keep any custom hash builder outside of the HashMap so we can access it alongside + // the entry API without mutable aliasing. + hash_builder: S, + // Circular linked list of nodes. If `values` is non-null, it will point to a "guard node" + // which will never have an initialized key or value, `values.prev` will contain the last key / + // value in the list, `values.next` will contain the first key / value in the list. + values: Option<NonNull<Node<K, V>>>, + // *Singly* linked list of free nodes. The `prev` pointers in the free list should be assumed + // invalid. + free: Option<NonNull<Node<K, V>>>, +} + +impl<K, V> LinkedHashMap<K, V> { + #[inline] + pub fn new() -> Self { + Self { + hash_builder: hash_map::DefaultHashBuilder::default(), + map: HashMap::with_hasher(NullHasher), + values: None, + free: None, + } + } + + #[inline] + pub fn with_capacity(capacity: usize) -> Self { + Self { + hash_builder: hash_map::DefaultHashBuilder::default(), + map: HashMap::with_capacity_and_hasher(capacity, NullHasher), + values: None, + free: None, + } + } +} + +impl<K, V, S> LinkedHashMap<K, V, S> { + #[inline] + pub fn with_hasher(hash_builder: S) -> Self { + Self { + hash_builder, + map: HashMap::with_hasher(NullHasher), + values: None, + free: None, + } + } + + #[inline] + pub fn with_capacity_and_hasher(capacity: usize, hash_builder: S) -> Self { + Self { + hash_builder, + map: HashMap::with_capacity_and_hasher(capacity, NullHasher), + values: None, + free: None, + } + } + + #[inline] + pub fn reserve(&mut self, additional: usize) { + self.map.reserve(additional); + } + + #[inline] + pub fn try_reserve(&mut self, additional: usize) -> Result<(), TryReserveError> { + self.map.try_reserve(additional).map_err(|e| match e { + hashbrown::TryReserveError::CapacityOverflow => TryReserveError::CapacityOverflow, + hashbrown::TryReserveError::AllocError { layout } => { + TryReserveError::AllocError { layout } + } + }) + } + + #[inline] + pub fn len(&self) -> usize { + self.map.len() + } + + #[inline] + pub fn is_empty(&self) -> bool { + self.len() == 0 + } + + #[inline] + pub fn clear(&mut self) { + self.map.clear(); + if let Some(mut values) = self.values { + unsafe { + drop_value_nodes(values); + values.as_mut().links.value = ValueLinks { + prev: values, + next: values, + }; + } + } + } + + #[inline] + pub fn iter(&self) -> Iter<K, V> { + let (head, tail) = if let Some(values) = self.values { + unsafe { + let ValueLinks { next, prev } = values.as_ref().links.value; + (next.as_ptr(), prev.as_ptr()) + } + } else { + (ptr::null_mut(), ptr::null_mut()) + }; + + Iter { + head, + tail, + remaining: self.len(), + marker: PhantomData, + } + } + + #[inline] + pub fn iter_mut(&mut self) -> IterMut<K, V> { + let (head, tail) = if let Some(values) = self.values { + unsafe { + let ValueLinks { next, prev } = values.as_ref().links.value; + (Some(next), Some(prev)) + } + } else { + (None, None) + }; + + IterMut { + head, + tail, + remaining: self.len(), + marker: PhantomData, + } + } + + #[inline] + pub fn drain(&mut self) -> Drain<'_, K, V> { + unsafe { + let (head, tail) = if let Some(mut values) = self.values { + let ValueLinks { next, prev } = values.as_ref().links.value; + values.as_mut().links.value = ValueLinks { + next: values, + prev: values, + }; + (Some(next), Some(prev)) + } else { + (None, None) + }; + let len = self.len(); + + self.map.clear(); + + Drain { + free: (&mut self.free).into(), + head, + tail, + remaining: len, + marker: PhantomData, + } + } + } + + #[inline] + pub fn keys(&self) -> Keys<K, V> { + Keys { inner: self.iter() } + } + + #[inline] + pub fn values(&self) -> Values<K, V> { + Values { inner: self.iter() } + } + + #[inline] + pub fn values_mut(&mut self) -> ValuesMut<K, V> { + ValuesMut { + inner: self.iter_mut(), + } + } + + #[inline] + pub fn front(&self) -> Option<(&K, &V)> { + if self.is_empty() { + return None; + } + unsafe { + let front = (*self.values.as_ptr()).links.value.next.as_ptr(); + let (key, value) = (*front).entry_ref(); + Some((key, value)) + } + } + + #[inline] + pub fn back(&self) -> Option<(&K, &V)> { + if self.is_empty() { + return None; + } + unsafe { + let back = &*(*self.values.as_ptr()).links.value.prev.as_ptr(); + let (key, value) = (*back).entry_ref(); + Some((key, value)) + } + } + + #[inline] + pub fn retain<F>(&mut self, mut f: F) + where + F: FnMut(&K, &mut V) -> bool, + { + let free = self.free; + let mut drop_filtered_values = DropFilteredValues { + free: &mut self.free, + cur_free: free, + }; + + self.map.retain(|&node, _| unsafe { + let (k, v) = (*node.as_ptr()).entry_mut(); + if f(k, v) { + true + } else { + drop_filtered_values.drop_later(node); + false + } + }); + } + + #[inline] + pub fn hasher(&self) -> &S { + &self.hash_builder + } + + #[inline] + pub fn capacity(&self) -> usize { + self.map.capacity() + } +} + +impl<K, V, S> LinkedHashMap<K, V, S> +where + K: Eq + Hash, + S: BuildHasher, +{ + #[inline] + pub fn entry(&mut self, key: K) -> Entry<'_, K, V, S> { + match self.raw_entry_mut().from_key(&key) { + RawEntryMut::Occupied(occupied) => Entry::Occupied(OccupiedEntry { + key, + raw_entry: occupied, + }), + RawEntryMut::Vacant(vacant) => Entry::Vacant(VacantEntry { + key, + raw_entry: vacant, + }), + } + } + + #[inline] + pub fn get<Q>(&self, k: &Q) -> Option<&V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.raw_entry().from_key(k).map(|(_, v)| v) + } + + #[inline] + pub fn get_key_value<Q>(&self, k: &Q) -> Option<(&K, &V)> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.raw_entry().from_key(k) + } + + #[inline] + pub fn contains_key<Q>(&self, k: &Q) -> bool + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.get(k).is_some() + } + + #[inline] + pub fn get_mut<Q>(&mut self, k: &Q) -> Option<&mut V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + match self.raw_entry_mut().from_key(k) { + RawEntryMut::Occupied(occupied) => Some(occupied.into_mut()), + RawEntryMut::Vacant(_) => None, + } + } + + /// Inserts the given key / value pair at the *back* of the internal linked list. + /// + /// Returns the previously set value, if one existed prior to this call. After this call, + /// calling `LinkedHashMap::back` will return a reference to this key / value pair. + #[inline] + pub fn insert(&mut self, k: K, v: V) -> Option<V> { + match self.raw_entry_mut().from_key(&k) { + RawEntryMut::Occupied(mut occupied) => { + occupied.to_back(); + Some(occupied.replace_value(v)) + } + RawEntryMut::Vacant(vacant) => { + vacant.insert(k, v); + None + } + } + } + + /// If the given key is not in this map, inserts the key / value pair at the *back* of the + /// internal linked list and returns `None`, otherwise, replaces the existing value with the + /// given value *without* moving the entry in the internal linked list and returns the previous + /// value. + #[inline] + pub fn replace(&mut self, k: K, v: V) -> Option<V> { + match self.raw_entry_mut().from_key(&k) { + RawEntryMut::Occupied(mut occupied) => Some(occupied.replace_value(v)), + RawEntryMut::Vacant(vacant) => { + vacant.insert(k, v); + None + } + } + } + + #[inline] + pub fn remove<Q>(&mut self, k: &Q) -> Option<V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + match self.raw_entry_mut().from_key(&k) { + RawEntryMut::Occupied(occupied) => Some(occupied.remove()), + RawEntryMut::Vacant(_) => None, + } + } + + #[inline] + pub fn remove_entry<Q>(&mut self, k: &Q) -> Option<(K, V)> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + match self.raw_entry_mut().from_key(&k) { + RawEntryMut::Occupied(occupied) => Some(occupied.remove_entry()), + RawEntryMut::Vacant(_) => None, + } + } + + #[inline] + pub fn pop_front(&mut self) -> Option<(K, V)> { + if self.is_empty() { + return None; + } + unsafe { + let front = (*self.values.as_ptr()).links.value.next; + match self.map.raw_entry_mut().from_hash( + hash_key(&self.hash_builder, front.as_ref().key_ref()), + |k| (*k).as_ref().key_ref().eq(front.as_ref().key_ref()), + ) { + hash_map::RawEntryMut::Occupied(occupied) => { + Some(remove_node(&mut self.free, occupied.remove_entry().0)) + } + hash_map::RawEntryMut::Vacant(_) => None, + } + } + } + + #[inline] + pub fn pop_back(&mut self) -> Option<(K, V)> { + if self.is_empty() { + return None; + } + unsafe { + let back = (*self.values.as_ptr()).links.value.prev; + match self + .map + .raw_entry_mut() + .from_hash(hash_key(&self.hash_builder, back.as_ref().key_ref()), |k| { + (*k).as_ref().key_ref().eq(back.as_ref().key_ref()) + }) { + hash_map::RawEntryMut::Occupied(occupied) => { + Some(remove_node(&mut self.free, occupied.remove_entry().0)) + } + hash_map::RawEntryMut::Vacant(_) => None, + } + } + } + + /// If an entry with this key exists, move it to the front of the list and return a reference to + /// the value. + #[inline] + pub fn to_front<Q>(&mut self, k: &Q) -> Option<&mut V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + match self.raw_entry_mut().from_key(k) { + RawEntryMut::Occupied(mut occupied) => { + occupied.to_front(); + Some(occupied.into_mut()) + } + RawEntryMut::Vacant(_) => None, + } + } + + /// If an entry with this key exists, move it to the back of the list and return a reference to + /// the value. + #[inline] + pub fn to_back<Q>(&mut self, k: &Q) -> Option<&mut V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + match self.raw_entry_mut().from_key(k) { + RawEntryMut::Occupied(mut occupied) => { + occupied.to_back(); + Some(occupied.into_mut()) + } + RawEntryMut::Vacant(_) => None, + } + } + + #[inline] + pub fn shrink_to_fit(&mut self) { + unsafe { + let len = self.map.len(); + if len != self.map.capacity() { + self.map = HashMap::with_hasher(NullHasher); + self.map.reserve(len); + + if let Some(guard) = self.values { + let mut cur = guard.as_ref().links.value.next; + while cur != guard { + let hash = hash_key(&self.hash_builder, cur.as_ref().key_ref()); + match self + .map + .raw_entry_mut() + .from_hash(hash, |k| (*k).as_ref().key_ref().eq(cur.as_ref().key_ref())) + { + hash_map::RawEntryMut::Occupied(_) => unreachable!(), + hash_map::RawEntryMut::Vacant(vacant) => { + let hash_builder = &self.hash_builder; + vacant.insert_with_hasher(hash, cur, (), |k| { + hash_key(hash_builder, (*k).as_ref().key_ref()) + }); + } + } + cur = cur.as_ref().links.value.next; + } + } + } + + drop_free_nodes(self.free); + self.free = None; + } + } + + pub fn retain_with_order<F>(&mut self, mut f: F) + where + F: FnMut(&K, &mut V) -> bool, + { + let free = self.free; + let mut drop_filtered_values = DropFilteredValues { + free: &mut self.free, + cur_free: free, + }; + + if let Some(values) = self.values { + unsafe { + let mut cur = values.as_ref().links.value.next; + while cur != values { + let next = cur.as_ref().links.value.next; + let filter = { + let (k, v) = (*cur.as_ptr()).entry_mut(); + !f(k, v) + }; + if filter { + let k = (*cur.as_ptr()).key_ref(); + let hash = hash_key(&self.hash_builder, k); + match self + .map + .raw_entry_mut() + .from_hash(hash, |o| (*o).as_ref().key_ref().eq(k)) + { + hash_map::RawEntryMut::Occupied(entry) => { + entry.remove(); + drop_filtered_values.drop_later(cur); + } + hash_map::RawEntryMut::Vacant(_) => unreachable!(), + } + } + cur = next; + } + } + } + } +} + +impl<K, V, S> LinkedHashMap<K, V, S> +where + S: BuildHasher, +{ + #[inline] + pub fn raw_entry(&self) -> RawEntryBuilder<'_, K, V, S> { + RawEntryBuilder { + hash_builder: &self.hash_builder, + entry: self.map.raw_entry(), + } + } + + #[inline] + pub fn raw_entry_mut(&mut self) -> RawEntryBuilderMut<'_, K, V, S> { + RawEntryBuilderMut { + hash_builder: &self.hash_builder, + values: &mut self.values, + free: &mut self.free, + entry: self.map.raw_entry_mut(), + } + } +} + +impl<K, V, S> Default for LinkedHashMap<K, V, S> +where + S: Default, +{ + #[inline] + fn default() -> Self { + Self::with_hasher(S::default()) + } +} + +impl<K: Hash + Eq, V, S: BuildHasher + Default> FromIterator<(K, V)> for LinkedHashMap<K, V, S> { + #[inline] + fn from_iter<I: IntoIterator<Item = (K, V)>>(iter: I) -> Self { + let iter = iter.into_iter(); + let mut map = Self::with_capacity_and_hasher(iter.size_hint().0, S::default()); + map.extend(iter); + map + } +} + +impl<K, V, S> fmt::Debug for LinkedHashMap<K, V, S> +where + K: fmt::Debug, + V: fmt::Debug, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + f.debug_map().entries(self).finish() + } +} + +impl<K: Hash + Eq, V: PartialEq, S: BuildHasher> PartialEq for LinkedHashMap<K, V, S> { + #[inline] + fn eq(&self, other: &Self) -> bool { + self.len() == other.len() && self.iter().eq(other) + } +} + +impl<K: Hash + Eq, V: Eq, S: BuildHasher> Eq for LinkedHashMap<K, V, S> {} + +impl<K: Hash + Eq + PartialOrd, V: PartialOrd, S: BuildHasher> PartialOrd + for LinkedHashMap<K, V, S> +{ + #[inline] + fn partial_cmp(&self, other: &Self) -> Option<Ordering> { + self.iter().partial_cmp(other) + } + + #[inline] + fn lt(&self, other: &Self) -> bool { + self.iter().lt(other) + } + + #[inline] + fn le(&self, other: &Self) -> bool { + self.iter().le(other) + } + + #[inline] + fn ge(&self, other: &Self) -> bool { + self.iter().ge(other) + } + + #[inline] + fn gt(&self, other: &Self) -> bool { + self.iter().gt(other) + } +} + +impl<K: Hash + Eq + Ord, V: Ord, S: BuildHasher> Ord for LinkedHashMap<K, V, S> { + #[inline] + fn cmp(&self, other: &Self) -> Ordering { + self.iter().cmp(other) + } +} + +impl<K: Hash + Eq, V: Hash, S: BuildHasher> Hash for LinkedHashMap<K, V, S> { + #[inline] + fn hash<H: Hasher>(&self, h: &mut H) { + for e in self.iter() { + e.hash(h); + } + } +} + +impl<K, V, S> Drop for LinkedHashMap<K, V, S> { + #[inline] + fn drop(&mut self) { + unsafe { + if let Some(values) = self.values { + drop_value_nodes(values); + let _ = Box::from_raw(values.as_ptr()); + } + drop_free_nodes(self.free); + } + } +} + +unsafe impl<K: Send, V: Send, S: Send> Send for LinkedHashMap<K, V, S> {} +unsafe impl<K: Sync, V: Sync, S: Sync> Sync for LinkedHashMap<K, V, S> {} + +impl<'a, K, V, S, Q> Index<&'a Q> for LinkedHashMap<K, V, S> +where + K: Hash + Eq + Borrow<Q>, + S: BuildHasher, + Q: Eq + Hash + ?Sized, +{ + type Output = V; + + #[inline] + fn index(&self, index: &'a Q) -> &V { + self.get(index).expect("no entry found for key") + } +} + +impl<'a, K, V, S, Q> IndexMut<&'a Q> for LinkedHashMap<K, V, S> +where + K: Hash + Eq + Borrow<Q>, + S: BuildHasher, + Q: Eq + Hash + ?Sized, +{ + #[inline] + fn index_mut(&mut self, index: &'a Q) -> &mut V { + self.get_mut(index).expect("no entry found for key") + } +} + +impl<K: Hash + Eq + Clone, V: Clone, S: BuildHasher + Clone> Clone for LinkedHashMap<K, V, S> { + #[inline] + fn clone(&self) -> Self { + let mut map = Self::with_hasher(self.hash_builder.clone()); + map.extend(self.iter().map(|(k, v)| (k.clone(), v.clone()))); + map + } +} + +impl<K: Hash + Eq, V, S: BuildHasher> Extend<(K, V)> for LinkedHashMap<K, V, S> { + #[inline] + fn extend<I: IntoIterator<Item = (K, V)>>(&mut self, iter: I) { + for (k, v) in iter { + self.insert(k, v); + } + } +} + +impl<'a, K, V, S> Extend<(&'a K, &'a V)> for LinkedHashMap<K, V, S> +where + K: 'a + Hash + Eq + Copy, + V: 'a + Copy, + S: BuildHasher, +{ + #[inline] + fn extend<I: IntoIterator<Item = (&'a K, &'a V)>>(&mut self, iter: I) { + for (&k, &v) in iter { + self.insert(k, v); + } + } +} + +pub enum Entry<'a, K, V, S> { + Occupied(OccupiedEntry<'a, K, V>), + Vacant(VacantEntry<'a, K, V, S>), +} + +impl<K: fmt::Debug, V: fmt::Debug, S> fmt::Debug for Entry<'_, K, V, S> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + match *self { + Entry::Vacant(ref v) => f.debug_tuple("Entry").field(v).finish(), + Entry::Occupied(ref o) => f.debug_tuple("Entry").field(o).finish(), + } + } +} + +impl<'a, K, V, S> Entry<'a, K, V, S> { + /// If this entry is vacant, inserts a new entry with the given value and returns a reference to + /// it. + /// + /// If this entry is occupied, this method *moves the occupied entry to the back of the internal + /// linked list* and returns a reference to the existing value. + #[inline] + pub fn or_insert(self, default: V) -> &'a mut V + where + K: Hash, + S: BuildHasher, + { + match self { + Entry::Occupied(mut entry) => { + entry.to_back(); + entry.into_mut() + } + Entry::Vacant(entry) => entry.insert(default), + } + } + + /// Similar to `Entry::or_insert`, but accepts a function to construct a new value if this entry + /// is vacant. + #[inline] + pub fn or_insert_with<F: FnOnce() -> V>(self, default: F) -> &'a mut V + where + K: Hash, + S: BuildHasher, + { + match self { + Entry::Occupied(mut entry) => { + entry.to_back(); + entry.into_mut() + } + Entry::Vacant(entry) => entry.insert(default()), + } + } + + #[inline] + pub fn key(&self) -> &K { + match *self { + Entry::Occupied(ref entry) => entry.key(), + Entry::Vacant(ref entry) => entry.key(), + } + } + + #[inline] + pub fn and_modify<F>(self, f: F) -> Self + where + F: FnOnce(&mut V), + { + match self { + Entry::Occupied(mut entry) => { + f(entry.get_mut()); + Entry::Occupied(entry) + } + Entry::Vacant(entry) => Entry::Vacant(entry), + } + } +} + +pub struct OccupiedEntry<'a, K, V> { + key: K, + raw_entry: RawOccupiedEntryMut<'a, K, V>, +} + +impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for OccupiedEntry<'_, K, V> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_struct("OccupiedEntry") + .field("key", self.key()) + .field("value", self.get()) + .finish() + } +} + +impl<'a, K, V> OccupiedEntry<'a, K, V> { + #[inline] + pub fn key(&self) -> &K { + self.raw_entry.key() + } + + #[inline] + pub fn remove_entry(self) -> (K, V) { + self.raw_entry.remove_entry() + } + + #[inline] + pub fn get(&self) -> &V { + self.raw_entry.get() + } + + #[inline] + pub fn get_mut(&mut self) -> &mut V { + self.raw_entry.get_mut() + } + + #[inline] + pub fn into_mut(self) -> &'a mut V { + self.raw_entry.into_mut() + } + + #[inline] + pub fn to_back(&mut self) { + self.raw_entry.to_back() + } + + #[inline] + pub fn to_front(&mut self) { + self.raw_entry.to_front() + } + + /// Replaces this entry's value with the provided value. + /// + /// Similarly to `LinkedHashMap::insert`, this moves the existing entry to the back of the + /// internal linked list. + #[inline] + pub fn insert(&mut self, value: V) -> V { + self.raw_entry.to_back(); + self.raw_entry.replace_value(value) + } + + #[inline] + pub fn remove(self) -> V { + self.raw_entry.remove() + } + + /// Similar to `OccupiedEntry::replace_entry`, but *does* move the entry to the back of the + /// internal linked list. + #[inline] + pub fn insert_entry(mut self, value: V) -> (K, V) { + self.raw_entry.to_back(); + self.replace_entry(value) + } + + /// Replaces the entry's key with the key provided to `LinkedHashMap::entry`, and replaces the + /// entry's value with the given `value` parameter. + /// + /// Does *not* move the entry to the back of the internal linked list. + pub fn replace_entry(mut self, value: V) -> (K, V) { + let old_key = mem::replace(self.raw_entry.key_mut(), self.key); + let old_value = mem::replace(self.raw_entry.get_mut(), value); + (old_key, old_value) + } + + /// Replaces this entry's key with the key provided to `LinkedHashMap::entry`. + /// + /// Does *not* move the entry to the back of the internal linked list. + #[inline] + pub fn replace_key(mut self) -> K { + mem::replace(self.raw_entry.key_mut(), self.key) + } +} + +pub struct VacantEntry<'a, K, V, S> { + key: K, + raw_entry: RawVacantEntryMut<'a, K, V, S>, +} + +impl<K: fmt::Debug, V, S> fmt::Debug for VacantEntry<'_, K, V, S> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_tuple("VacantEntry").field(self.key()).finish() + } +} + +impl<'a, K, V, S> VacantEntry<'a, K, V, S> { + #[inline] + pub fn key(&self) -> &K { + &self.key + } + + #[inline] + pub fn into_key(self) -> K { + self.key + } + + /// Insert's the key for this vacant entry paired with the given value as a new entry at the + /// *back* of the internal linked list. + #[inline] + pub fn insert(self, value: V) -> &'a mut V + where + K: Hash, + S: BuildHasher, + { + self.raw_entry.insert(self.key, value).1 + } +} + +pub struct RawEntryBuilder<'a, K, V, S> { + hash_builder: &'a S, + entry: hash_map::RawEntryBuilder<'a, NonNull<Node<K, V>>, (), NullHasher>, +} + +impl<'a, K, V, S> RawEntryBuilder<'a, K, V, S> +where + S: BuildHasher, +{ + #[inline] + pub fn from_key<Q>(self, k: &Q) -> Option<(&'a K, &'a V)> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + let hash = hash_key(self.hash_builder, k); + self.from_key_hashed_nocheck(hash, k) + } + + #[inline] + pub fn from_key_hashed_nocheck<Q>(self, hash: u64, k: &Q) -> Option<(&'a K, &'a V)> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.from_hash(hash, move |o| k.eq(o.borrow())) + } + + #[inline] + pub fn from_hash( + self, + hash: u64, + mut is_match: impl FnMut(&K) -> bool, + ) -> Option<(&'a K, &'a V)> { + unsafe { + let node = *self + .entry + .from_hash(hash, move |k| is_match((*k).as_ref().key_ref()))? + .0; + + let (key, value) = (*node.as_ptr()).entry_ref(); + Some((key, value)) + } + } +} + +unsafe impl<'a, K, V, S> Send for RawEntryBuilder<'a, K, V, S> +where + K: Send, + V: Send, + S: Send, +{ +} + +unsafe impl<'a, K, V, S> Sync for RawEntryBuilder<'a, K, V, S> +where + K: Sync, + V: Sync, + S: Sync, +{ +} + +pub struct RawEntryBuilderMut<'a, K, V, S> { + hash_builder: &'a S, + values: &'a mut Option<NonNull<Node<K, V>>>, + free: &'a mut Option<NonNull<Node<K, V>>>, + entry: hash_map::RawEntryBuilderMut<'a, NonNull<Node<K, V>>, (), NullHasher>, +} + +impl<'a, K, V, S> RawEntryBuilderMut<'a, K, V, S> +where + S: BuildHasher, +{ + #[inline] + pub fn from_key<Q>(self, k: &Q) -> RawEntryMut<'a, K, V, S> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + let hash = hash_key(self.hash_builder, k); + self.from_key_hashed_nocheck(hash, k) + } + + #[inline] + pub fn from_key_hashed_nocheck<Q>(self, hash: u64, k: &Q) -> RawEntryMut<'a, K, V, S> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.from_hash(hash, move |o| k.eq(o.borrow())) + } + + #[inline] + pub fn from_hash( + self, + hash: u64, + mut is_match: impl FnMut(&K) -> bool, + ) -> RawEntryMut<'a, K, V, S> { + let entry = self + .entry + .from_hash(hash, move |k| is_match(unsafe { (*k).as_ref().key_ref() })); + + match entry { + hash_map::RawEntryMut::Occupied(occupied) => { + RawEntryMut::Occupied(RawOccupiedEntryMut { + free: self.free, + values: self.values, + entry: occupied, + }) + } + hash_map::RawEntryMut::Vacant(vacant) => RawEntryMut::Vacant(RawVacantEntryMut { + hash_builder: self.hash_builder, + values: self.values, + free: self.free, + entry: vacant, + }), + } + } +} + +unsafe impl<'a, K, V, S> Send for RawEntryBuilderMut<'a, K, V, S> +where + K: Send, + V: Send, + S: Send, +{ +} + +unsafe impl<'a, K, V, S> Sync for RawEntryBuilderMut<'a, K, V, S> +where + K: Sync, + V: Sync, + S: Sync, +{ +} + +pub enum RawEntryMut<'a, K, V, S> { + Occupied(RawOccupiedEntryMut<'a, K, V>), + Vacant(RawVacantEntryMut<'a, K, V, S>), +} + +impl<'a, K, V, S> RawEntryMut<'a, K, V, S> { + /// Similarly to `Entry::or_insert`, if this entry is occupied, it will move the existing entry + /// to the back of the internal linked list. + #[inline] + pub fn or_insert(self, default_key: K, default_val: V) -> (&'a mut K, &'a mut V) + where + K: Hash, + S: BuildHasher, + { + match self { + RawEntryMut::Occupied(mut entry) => { + entry.to_back(); + entry.into_key_value() + } + RawEntryMut::Vacant(entry) => entry.insert(default_key, default_val), + } + } + + /// Similarly to `Entry::or_insert_with`, if this entry is occupied, it will move the existing + /// entry to the back of the internal linked list. + #[inline] + pub fn or_insert_with<F>(self, default: F) -> (&'a mut K, &'a mut V) + where + F: FnOnce() -> (K, V), + K: Hash, + S: BuildHasher, + { + match self { + RawEntryMut::Occupied(mut entry) => { + entry.to_back(); + entry.into_key_value() + } + RawEntryMut::Vacant(entry) => { + let (k, v) = default(); + entry.insert(k, v) + } + } + } + + #[inline] + pub fn and_modify<F>(self, f: F) -> Self + where + F: FnOnce(&mut K, &mut V), + { + match self { + RawEntryMut::Occupied(mut entry) => { + { + let (k, v) = entry.get_key_value_mut(); + f(k, v); + } + RawEntryMut::Occupied(entry) + } + RawEntryMut::Vacant(entry) => RawEntryMut::Vacant(entry), + } + } +} + +pub struct RawOccupiedEntryMut<'a, K, V> { + free: &'a mut Option<NonNull<Node<K, V>>>, + values: &'a mut Option<NonNull<Node<K, V>>>, + entry: hash_map::RawOccupiedEntryMut<'a, NonNull<Node<K, V>>, (), NullHasher>, +} + +impl<'a, K, V> RawOccupiedEntryMut<'a, K, V> { + #[inline] + pub fn key(&self) -> &K { + self.get_key_value().0 + } + + #[inline] + pub fn key_mut(&mut self) -> &mut K { + self.get_key_value_mut().0 + } + + #[inline] + pub fn into_key(self) -> &'a mut K { + self.into_key_value().0 + } + + #[inline] + pub fn get(&self) -> &V { + self.get_key_value().1 + } + + #[inline] + pub fn get_mut(&mut self) -> &mut V { + self.get_key_value_mut().1 + } + + #[inline] + pub fn into_mut(self) -> &'a mut V { + self.into_key_value().1 + } + + #[inline] + pub fn get_key_value(&self) -> (&K, &V) { + unsafe { + let node = *self.entry.key(); + let (key, value) = (*node.as_ptr()).entry_ref(); + (key, value) + } + } + + #[inline] + pub fn get_key_value_mut(&mut self) -> (&mut K, &mut V) { + unsafe { + let node = *self.entry.key_mut(); + let (key, value) = (*node.as_ptr()).entry_mut(); + (key, value) + } + } + + #[inline] + pub fn into_key_value(self) -> (&'a mut K, &'a mut V) { + unsafe { + let node = *self.entry.into_key(); + let (key, value) = (*node.as_ptr()).entry_mut(); + (key, value) + } + } + + #[inline] + pub fn to_back(&mut self) { + unsafe { + let node = *self.entry.key_mut(); + detach_node(node); + attach_before(node, NonNull::new_unchecked(self.values.as_ptr())); + } + } + + #[inline] + pub fn to_front(&mut self) { + unsafe { + let node = *self.entry.key_mut(); + detach_node(node); + attach_before(node, (*self.values.as_ptr()).links.value.next); + } + } + + #[inline] + pub fn replace_value(&mut self, value: V) -> V { + unsafe { + let mut node = *self.entry.key_mut(); + mem::replace(&mut node.as_mut().entry_mut().1, value) + } + } + + #[inline] + pub fn replace_key(&mut self, key: K) -> K { + unsafe { + let mut node = *self.entry.key_mut(); + mem::replace(&mut node.as_mut().entry_mut().0, key) + } + } + + #[inline] + pub fn remove(self) -> V { + self.remove_entry().1 + } + + #[inline] + pub fn remove_entry(self) -> (K, V) { + let node = self.entry.remove_entry().0; + unsafe { remove_node(self.free, node) } + } +} + +pub struct RawVacantEntryMut<'a, K, V, S> { + hash_builder: &'a S, + values: &'a mut Option<NonNull<Node<K, V>>>, + free: &'a mut Option<NonNull<Node<K, V>>>, + entry: hash_map::RawVacantEntryMut<'a, NonNull<Node<K, V>>, (), NullHasher>, +} + +impl<'a, K, V, S> RawVacantEntryMut<'a, K, V, S> { + #[inline] + pub fn insert(self, key: K, value: V) -> (&'a mut K, &'a mut V) + where + K: Hash, + S: BuildHasher, + { + let hash = hash_key(self.hash_builder, &key); + self.insert_hashed_nocheck(hash, key, value) + } + + #[inline] + pub fn insert_hashed_nocheck(self, hash: u64, key: K, value: V) -> (&'a mut K, &'a mut V) + where + K: Hash, + S: BuildHasher, + { + let hash_builder = self.hash_builder; + self.insert_with_hasher(hash, key, value, |k| hash_key(hash_builder, k)) + } + + #[inline] + pub fn insert_with_hasher( + self, + hash: u64, + key: K, + value: V, + hasher: impl Fn(&K) -> u64, + ) -> (&'a mut K, &'a mut V) + where + S: BuildHasher, + { + unsafe { + ensure_guard_node(self.values); + let mut new_node = allocate_node(self.free); + new_node.as_mut().put_entry((key, value)); + attach_before(new_node, NonNull::new_unchecked(self.values.as_ptr())); + + let node = *self + .entry + .insert_with_hasher(hash, new_node, (), move |k| hasher((*k).as_ref().key_ref())) + .0; + + let (key, value) = (*node.as_ptr()).entry_mut(); + (key, value) + } + } +} + +impl<K, V, S> fmt::Debug for RawEntryBuilderMut<'_, K, V, S> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_struct("RawEntryBuilder").finish() + } +} + +impl<K: fmt::Debug, V: fmt::Debug, S> fmt::Debug for RawEntryMut<'_, K, V, S> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + match *self { + RawEntryMut::Vacant(ref v) => f.debug_tuple("RawEntry").field(v).finish(), + RawEntryMut::Occupied(ref o) => f.debug_tuple("RawEntry").field(o).finish(), + } + } +} + +impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for RawOccupiedEntryMut<'_, K, V> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_struct("RawOccupiedEntryMut") + .field("key", self.key()) + .field("value", self.get()) + .finish() + } +} + +impl<K, V, S> fmt::Debug for RawVacantEntryMut<'_, K, V, S> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_struct("RawVacantEntryMut").finish() + } +} + +impl<K, V, S> fmt::Debug for RawEntryBuilder<'_, K, V, S> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_struct("RawEntryBuilder").finish() + } +} + +unsafe impl<'a, K, V> Send for RawOccupiedEntryMut<'a, K, V> +where + K: Send, + V: Send, +{ +} + +unsafe impl<'a, K, V> Sync for RawOccupiedEntryMut<'a, K, V> +where + K: Sync, + V: Sync, +{ +} + +unsafe impl<'a, K, V, S> Send for RawVacantEntryMut<'a, K, V, S> +where + K: Send, + V: Send, + S: Send, +{ +} + +unsafe impl<'a, K, V, S> Sync for RawVacantEntryMut<'a, K, V, S> +where + K: Sync, + V: Sync, + S: Sync, +{ +} + +pub struct Iter<'a, K, V> { + head: *const Node<K, V>, + tail: *const Node<K, V>, + remaining: usize, + marker: PhantomData<(&'a K, &'a V)>, +} + +pub struct IterMut<'a, K, V> { + head: Option<NonNull<Node<K, V>>>, + tail: Option<NonNull<Node<K, V>>>, + remaining: usize, + marker: PhantomData<(&'a K, &'a mut V)>, +} + +pub struct IntoIter<K, V> { + head: Option<NonNull<Node<K, V>>>, + tail: Option<NonNull<Node<K, V>>>, + remaining: usize, + marker: PhantomData<(K, V)>, +} + +pub struct Drain<'a, K, V> { + free: NonNull<Option<NonNull<Node<K, V>>>>, + head: Option<NonNull<Node<K, V>>>, + tail: Option<NonNull<Node<K, V>>>, + remaining: usize, + // We want `Drain` to be covariant + marker: PhantomData<(K, V, &'a LinkedHashMap<K, V>)>, +} + +impl<K, V> IterMut<'_, K, V> { + #[inline] + pub(crate) fn iter(&self) -> Iter<'_, K, V> { + Iter { + head: self.head.as_ptr(), + tail: self.tail.as_ptr(), + remaining: self.remaining, + marker: PhantomData, + } + } +} + +impl<K, V> IntoIter<K, V> { + #[inline] + pub(crate) fn iter(&self) -> Iter<'_, K, V> { + Iter { + head: self.head.as_ptr(), + tail: self.tail.as_ptr(), + remaining: self.remaining, + marker: PhantomData, + } + } +} + +impl<K, V> Drain<'_, K, V> { + #[inline] + pub(crate) fn iter(&self) -> Iter<'_, K, V> { + Iter { + head: self.head.as_ptr(), + tail: self.tail.as_ptr(), + remaining: self.remaining, + marker: PhantomData, + } + } +} + +unsafe impl<'a, K, V> Send for Iter<'a, K, V> +where + K: Send, + V: Send, +{ +} + +unsafe impl<'a, K, V> Send for IterMut<'a, K, V> +where + K: Send, + V: Send, +{ +} + +unsafe impl<K, V> Send for IntoIter<K, V> +where + K: Send, + V: Send, +{ +} + +unsafe impl<'a, K, V> Send for Drain<'a, K, V> +where + K: Send, + V: Send, +{ +} + +unsafe impl<'a, K, V> Sync for Iter<'a, K, V> +where + K: Sync, + V: Sync, +{ +} + +unsafe impl<'a, K, V> Sync for IterMut<'a, K, V> +where + K: Sync, + V: Sync, +{ +} + +unsafe impl<K, V> Sync for IntoIter<K, V> +where + K: Sync, + V: Sync, +{ +} + +unsafe impl<'a, K, V> Sync for Drain<'a, K, V> +where + K: Sync, + V: Sync, +{ +} + +impl<'a, K, V> Clone for Iter<'a, K, V> { + #[inline] + fn clone(&self) -> Self { + Iter { ..*self } + } +} + +impl<K: fmt::Debug, V: fmt::Debug> fmt::Debug for Iter<'_, K, V> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.clone()).finish() + } +} + +impl<K, V> fmt::Debug for IterMut<'_, K, V> +where + K: fmt::Debug, + V: fmt::Debug, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.iter()).finish() + } +} + +impl<K, V> fmt::Debug for IntoIter<K, V> +where + K: fmt::Debug, + V: fmt::Debug, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.iter()).finish() + } +} + +impl<K, V> fmt::Debug for Drain<'_, K, V> +where + K: fmt::Debug, + V: fmt::Debug, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.iter()).finish() + } +} + +impl<'a, K, V> Iterator for Iter<'a, K, V> { + type Item = (&'a K, &'a V); + + #[inline] + fn next(&mut self) -> Option<(&'a K, &'a V)> { + if self.remaining == 0 { + None + } else { + self.remaining -= 1; + unsafe { + let (key, value) = (*self.head).entry_ref(); + self.head = (*self.head).links.value.next.as_ptr(); + Some((key, value)) + } + } + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + (self.remaining, Some(self.remaining)) + } +} + +impl<'a, K, V> Iterator for IterMut<'a, K, V> { + type Item = (&'a K, &'a mut V); + + #[inline] + fn next(&mut self) -> Option<(&'a K, &'a mut V)> { + if self.remaining == 0 { + None + } else { + self.remaining -= 1; + unsafe { + let head = self.head.as_ptr(); + let (key, value) = (*head).entry_mut(); + self.head = Some((*head).links.value.next); + Some((key, value)) + } + } + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + (self.remaining, Some(self.remaining)) + } +} + +impl<K, V> Iterator for IntoIter<K, V> { + type Item = (K, V); + + #[inline] + fn next(&mut self) -> Option<(K, V)> { + if self.remaining == 0 { + return None; + } + self.remaining -= 1; + unsafe { + let head = self.head.as_ptr(); + self.head = Some((*head).links.value.next); + let mut e = Box::from_raw(head); + Some(e.take_entry()) + } + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + (self.remaining, Some(self.remaining)) + } +} + +impl<'a, K, V> Iterator for Drain<'a, K, V> { + type Item = (K, V); + + #[inline] + fn next(&mut self) -> Option<(K, V)> { + if self.remaining == 0 { + return None; + } + self.remaining -= 1; + unsafe { + let mut head = NonNull::new_unchecked(self.head.as_ptr()); + self.head = Some(head.as_ref().links.value.next); + let entry = head.as_mut().take_entry(); + push_free(self.free.as_mut(), head); + Some(entry) + } + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + (self.remaining, Some(self.remaining)) + } +} + +impl<'a, K, V> DoubleEndedIterator for Iter<'a, K, V> { + #[inline] + fn next_back(&mut self) -> Option<(&'a K, &'a V)> { + if self.remaining == 0 { + None + } else { + self.remaining -= 1; + unsafe { + let tail = self.tail; + self.tail = (*tail).links.value.prev.as_ptr(); + let (key, value) = (*tail).entry_ref(); + Some((key, value)) + } + } + } +} + +impl<'a, K, V> DoubleEndedIterator for IterMut<'a, K, V> { + #[inline] + fn next_back(&mut self) -> Option<(&'a K, &'a mut V)> { + if self.remaining == 0 { + None + } else { + self.remaining -= 1; + unsafe { + let tail = self.tail.as_ptr(); + self.tail = Some((*tail).links.value.prev); + let (key, value) = (*tail).entry_mut(); + Some((key, value)) + } + } + } +} + +impl<K, V> DoubleEndedIterator for IntoIter<K, V> { + #[inline] + fn next_back(&mut self) -> Option<(K, V)> { + if self.remaining == 0 { + return None; + } + self.remaining -= 1; + unsafe { + let mut e = *Box::from_raw(self.tail.as_ptr()); + self.tail = Some(e.links.value.prev); + Some(e.take_entry()) + } + } +} + +impl<'a, K, V> DoubleEndedIterator for Drain<'a, K, V> { + #[inline] + fn next_back(&mut self) -> Option<(K, V)> { + if self.remaining == 0 { + return None; + } + self.remaining -= 1; + unsafe { + let mut tail = NonNull::new_unchecked(self.tail.as_ptr()); + self.tail = Some(tail.as_ref().links.value.prev); + let entry = tail.as_mut().take_entry(); + push_free(&mut *self.free.as_ptr(), tail); + Some(entry) + } + } +} + +impl<'a, K, V> ExactSizeIterator for Iter<'a, K, V> {} + +impl<'a, K, V> ExactSizeIterator for IterMut<'a, K, V> {} + +impl<K, V> ExactSizeIterator for IntoIter<K, V> {} + +impl<K, V> Drop for IntoIter<K, V> { + #[inline] + fn drop(&mut self) { + for _ in 0..self.remaining { + unsafe { + let tail = self.tail.as_ptr(); + self.tail = Some((*tail).links.value.prev); + (*tail).take_entry(); + let _ = Box::from_raw(tail); + } + } + } +} + +impl<'a, K, V> Drop for Drain<'a, K, V> { + #[inline] + fn drop(&mut self) { + for _ in 0..self.remaining { + unsafe { + let mut tail = NonNull::new_unchecked(self.tail.as_ptr()); + self.tail = Some(tail.as_ref().links.value.prev); + tail.as_mut().take_entry(); + push_free(&mut *self.free.as_ptr(), tail); + } + } + } +} + +pub struct Keys<'a, K, V> { + inner: Iter<'a, K, V>, +} + +impl<K: fmt::Debug, V> fmt::Debug for Keys<'_, K, V> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.clone()).finish() + } +} + +impl<'a, K, V> Clone for Keys<'a, K, V> { + #[inline] + fn clone(&self) -> Keys<'a, K, V> { + Keys { + inner: self.inner.clone(), + } + } +} + +impl<'a, K, V> Iterator for Keys<'a, K, V> { + type Item = &'a K; + + #[inline] + fn next(&mut self) -> Option<&'a K> { + self.inner.next().map(|e| e.0) + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + self.inner.size_hint() + } +} + +impl<'a, K, V> DoubleEndedIterator for Keys<'a, K, V> { + #[inline] + fn next_back(&mut self) -> Option<&'a K> { + self.inner.next_back().map(|e| e.0) + } +} + +impl<'a, K, V> ExactSizeIterator for Keys<'a, K, V> { + #[inline] + fn len(&self) -> usize { + self.inner.len() + } +} + +pub struct Values<'a, K, V> { + inner: Iter<'a, K, V>, +} + +impl<K, V> Clone for Values<'_, K, V> { + #[inline] + fn clone(&self) -> Self { + Values { + inner: self.inner.clone(), + } + } +} + +impl<K, V: fmt::Debug> fmt::Debug for Values<'_, K, V> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.clone()).finish() + } +} + +impl<'a, K, V> Iterator for Values<'a, K, V> { + type Item = &'a V; + + #[inline] + fn next(&mut self) -> Option<&'a V> { + self.inner.next().map(|e| e.1) + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + self.inner.size_hint() + } +} + +impl<'a, K, V> DoubleEndedIterator for Values<'a, K, V> { + #[inline] + fn next_back(&mut self) -> Option<&'a V> { + self.inner.next_back().map(|e| e.1) + } +} + +impl<'a, K, V> ExactSizeIterator for Values<'a, K, V> { + #[inline] + fn len(&self) -> usize { + self.inner.len() + } +} + +pub struct ValuesMut<'a, K, V> { + inner: IterMut<'a, K, V>, +} + +impl<K, V> fmt::Debug for ValuesMut<'_, K, V> +where + K: fmt::Debug, + V: fmt::Debug, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.inner.iter()).finish() + } +} + +impl<'a, K, V> Iterator for ValuesMut<'a, K, V> { + type Item = &'a mut V; + + #[inline] + fn next(&mut self) -> Option<&'a mut V> { + self.inner.next().map(|e| e.1) + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + self.inner.size_hint() + } +} + +impl<'a, K, V> DoubleEndedIterator for ValuesMut<'a, K, V> { + #[inline] + fn next_back(&mut self) -> Option<&'a mut V> { + self.inner.next_back().map(|e| e.1) + } +} + +impl<'a, K, V> ExactSizeIterator for ValuesMut<'a, K, V> { + #[inline] + fn len(&self) -> usize { + self.inner.len() + } +} + +impl<'a, K, V, S> IntoIterator for &'a LinkedHashMap<K, V, S> { + type Item = (&'a K, &'a V); + type IntoIter = Iter<'a, K, V>; + + #[inline] + fn into_iter(self) -> Iter<'a, K, V> { + self.iter() + } +} + +impl<'a, K, V, S> IntoIterator for &'a mut LinkedHashMap<K, V, S> { + type Item = (&'a K, &'a mut V); + type IntoIter = IterMut<'a, K, V>; + + #[inline] + fn into_iter(self) -> IterMut<'a, K, V> { + self.iter_mut() + } +} + +impl<K, V, S> IntoIterator for LinkedHashMap<K, V, S> { + type Item = (K, V); + type IntoIter = IntoIter<K, V>; + + #[inline] + fn into_iter(mut self) -> IntoIter<K, V> { + unsafe { + let (head, tail) = if let Some(values) = self.values { + let ValueLinks { + next: head, + prev: tail, + } = values.as_ref().links.value; + + let _ = Box::from_raw(self.values.as_ptr()); + self.values = None; + + (Some(head), Some(tail)) + } else { + (None, None) + }; + let len = self.len(); + + drop_free_nodes(self.free); + self.free = None; + + self.map.clear(); + + IntoIter { + head, + tail, + remaining: len, + marker: PhantomData, + } + } + } +} + +// A ZST that asserts that the inner HashMap will not do its own key hashing +struct NullHasher; + +impl BuildHasher for NullHasher { + type Hasher = Self; + + #[inline] + fn build_hasher(&self) -> Self { + Self + } +} + +impl Hasher for NullHasher { + #[inline] + fn write(&mut self, _bytes: &[u8]) { + unreachable!("inner map should not be using its built-in hasher") + } + + #[inline] + fn finish(&self) -> u64 { + unreachable!("inner map should not be using its built-in hasher") + } +} + +struct ValueLinks<K, V> { + next: NonNull<Node<K, V>>, + prev: NonNull<Node<K, V>>, +} + +impl<K, V> Clone for ValueLinks<K, V> { + #[inline] + fn clone(&self) -> Self { + ValueLinks { + next: self.next, + prev: self.prev, + } + } +} + +impl<K, V> Copy for ValueLinks<K, V> {} + +struct FreeLink<K, V> { + next: Option<NonNull<Node<K, V>>>, +} + +impl<K, V> Clone for FreeLink<K, V> { + #[inline] + fn clone(&self) -> Self { + FreeLink { next: self.next } + } +} + +impl<K, V> Copy for FreeLink<K, V> {} + +union Links<K, V> { + value: ValueLinks<K, V>, + free: FreeLink<K, V>, +} + +struct Node<K, V> { + entry: MaybeUninit<(K, V)>, + links: Links<K, V>, +} + +impl<K, V> Node<K, V> { + #[inline] + unsafe fn put_entry(&mut self, entry: (K, V)) { + self.entry.as_mut_ptr().write(entry) + } + + #[inline] + unsafe fn entry_ref(&self) -> &(K, V) { + &*self.entry.as_ptr() + } + + #[inline] + unsafe fn key_ref(&self) -> &K { + &(*self.entry.as_ptr()).0 + } + + #[inline] + unsafe fn entry_mut(&mut self) -> &mut (K, V) { + &mut *self.entry.as_mut_ptr() + } + + #[inline] + unsafe fn take_entry(&mut self) -> (K, V) { + self.entry.as_ptr().read() + } +} + +trait OptNonNullExt<T> { + fn as_ptr(self) -> *mut T; +} + +impl<T> OptNonNullExt<T> for Option<NonNull<T>> { + #[inline] + fn as_ptr(self) -> *mut T { + match self { + Some(ptr) => ptr.as_ptr(), + None => ptr::null_mut(), + } + } +} + +// Allocate a circular list guard node if not present. +#[inline] +unsafe fn ensure_guard_node<K, V>(head: &mut Option<NonNull<Node<K, V>>>) { + if head.is_none() { + let mut p = NonNull::new_unchecked(Box::into_raw(Box::new(Node { + entry: MaybeUninit::uninit(), + links: Links { + value: ValueLinks { + next: NonNull::dangling(), + prev: NonNull::dangling(), + }, + }, + }))); + p.as_mut().links.value = ValueLinks { next: p, prev: p }; + *head = Some(p); + } +} + +// Attach the `to_attach` node to the existing circular list *before* `node`. +#[inline] +unsafe fn attach_before<K, V>(mut to_attach: NonNull<Node<K, V>>, mut node: NonNull<Node<K, V>>) { + to_attach.as_mut().links.value = ValueLinks { + prev: node.as_ref().links.value.prev, + next: node, + }; + node.as_mut().links.value.prev = to_attach; + (*to_attach.as_mut().links.value.prev.as_ptr()) + .links + .value + .next = to_attach; +} + +#[inline] +unsafe fn detach_node<K, V>(mut node: NonNull<Node<K, V>>) { + node.as_mut().links.value.prev.as_mut().links.value.next = node.as_ref().links.value.next; + node.as_mut().links.value.next.as_mut().links.value.prev = node.as_ref().links.value.prev; +} + +#[inline] +unsafe fn push_free<K, V>( + free_list: &mut Option<NonNull<Node<K, V>>>, + mut node: NonNull<Node<K, V>>, +) { + node.as_mut().links.free.next = *free_list; + *free_list = Some(node); +} + +#[inline] +unsafe fn pop_free<K, V>( + free_list: &mut Option<NonNull<Node<K, V>>>, +) -> Option<NonNull<Node<K, V>>> { + if let Some(free) = *free_list { + *free_list = free.as_ref().links.free.next; + Some(free) + } else { + None + } +} + +#[inline] +unsafe fn allocate_node<K, V>(free_list: &mut Option<NonNull<Node<K, V>>>) -> NonNull<Node<K, V>> { + if let Some(mut free) = pop_free(free_list) { + free.as_mut().links.value = ValueLinks { + next: NonNull::dangling(), + prev: NonNull::dangling(), + }; + free + } else { + NonNull::new_unchecked(Box::into_raw(Box::new(Node { + entry: MaybeUninit::uninit(), + links: Links { + value: ValueLinks { + next: NonNull::dangling(), + prev: NonNull::dangling(), + }, + }, + }))) + } +} + +// Given node is assumed to be the guard node and is *not* dropped. +#[inline] +unsafe fn drop_value_nodes<K, V>(guard: NonNull<Node<K, V>>) { + let mut cur = guard.as_ref().links.value.prev; + while cur != guard { + let prev = cur.as_ref().links.value.prev; + cur.as_mut().take_entry(); + let _ = Box::from_raw(cur.as_ptr()); + cur = prev; + } +} + +// Drops all linked free nodes starting with the given node. Free nodes are only non-circular +// singly linked, and should have uninitialized keys / values. +#[inline] +unsafe fn drop_free_nodes<K, V>(mut free: Option<NonNull<Node<K, V>>>) { + while let Some(some_free) = free { + let next_free = some_free.as_ref().links.free.next; + let _ = Box::from_raw(some_free.as_ptr()); + free = next_free; + } +} + +#[inline] +unsafe fn remove_node<K, V>( + free_list: &mut Option<NonNull<Node<K, V>>>, + mut node: NonNull<Node<K, V>>, +) -> (K, V) { + detach_node(node); + push_free(free_list, node); + node.as_mut().take_entry() +} + +#[inline] +fn hash_key<S, Q>(s: &S, k: &Q) -> u64 +where + S: BuildHasher, + Q: Hash + ?Sized, +{ + let mut hasher = s.build_hasher(); + k.hash(&mut hasher); + hasher.finish() +} + +// We do not drop the key and value when a value is filtered from the map during the call to +// `retain`. We need to be very careful not to have a live `HashMap` entry pointing to +// either a dangling `Node` or a `Node` with dropped keys / values. Since the key and value +// types may panic on drop, they may short-circuit the entry in the map actually being +// removed. Instead, we push the removed nodes onto the free list eagerly, then try and +// drop the keys and values for any newly freed nodes *after* `HashMap::retain` has +// completely finished. +struct DropFilteredValues<'a, K, V> { + free: &'a mut Option<NonNull<Node<K, V>>>, + cur_free: Option<NonNull<Node<K, V>>>, +} + +impl<'a, K, V> DropFilteredValues<'a, K, V> { + #[inline] + fn drop_later(&mut self, node: NonNull<Node<K, V>>) { + unsafe { + detach_node(node); + push_free(&mut self.cur_free, node); + } + } +} + +impl<'a, K, V> Drop for DropFilteredValues<'a, K, V> { + fn drop(&mut self) { + unsafe { + let end_free = self.cur_free; + while self.cur_free != *self.free { + let cur_free = self.cur_free.as_ptr(); + (*cur_free).take_entry(); + self.cur_free = (*cur_free).links.free.next; + } + *self.free = end_free; + } + } +} diff --git a/third_party/rust/hashlink/src/linked_hash_set.rs b/third_party/rust/hashlink/src/linked_hash_set.rs new file mode 100644 index 0000000000..5a89875d47 --- /dev/null +++ b/third_party/rust/hashlink/src/linked_hash_set.rs @@ -0,0 +1,766 @@ +use std::{ + borrow::Borrow, + fmt, + hash::{BuildHasher, Hash, Hasher}, + iter::{Chain, FromIterator}, + ops::{BitAnd, BitOr, BitXor, Sub}, +}; + +use hashbrown::hash_map::DefaultHashBuilder; + +use crate::linked_hash_map::{self, LinkedHashMap, TryReserveError}; + +pub struct LinkedHashSet<T, S = DefaultHashBuilder> { + map: LinkedHashMap<T, (), S>, +} + +impl<T: Hash + Eq> LinkedHashSet<T, DefaultHashBuilder> { + #[inline] + pub fn new() -> LinkedHashSet<T, DefaultHashBuilder> { + LinkedHashSet { + map: LinkedHashMap::new(), + } + } + + #[inline] + pub fn with_capacity(capacity: usize) -> LinkedHashSet<T, DefaultHashBuilder> { + LinkedHashSet { + map: LinkedHashMap::with_capacity(capacity), + } + } +} + +impl<T, S> LinkedHashSet<T, S> { + #[inline] + pub fn capacity(&self) -> usize { + self.map.capacity() + } + + #[inline] + pub fn iter(&self) -> Iter<'_, T> { + Iter { + iter: self.map.keys(), + } + } + + #[inline] + pub fn len(&self) -> usize { + self.map.len() + } + + #[inline] + pub fn is_empty(&self) -> bool { + self.map.is_empty() + } + + #[inline] + pub fn drain(&mut self) -> Drain<T> { + Drain { + iter: self.map.drain(), + } + } + + #[inline] + pub fn clear(&mut self) { + self.map.clear() + } + + #[inline] + pub fn retain<F>(&mut self, mut f: F) + where + F: FnMut(&T) -> bool, + { + self.map.retain(|k, _| f(k)); + } +} + +impl<T, S> LinkedHashSet<T, S> +where + T: Eq + Hash, + S: BuildHasher, +{ + #[inline] + pub fn with_hasher(hasher: S) -> LinkedHashSet<T, S> { + LinkedHashSet { + map: LinkedHashMap::with_hasher(hasher), + } + } + + #[inline] + pub fn with_capacity_and_hasher(capacity: usize, hasher: S) -> LinkedHashSet<T, S> { + LinkedHashSet { + map: LinkedHashMap::with_capacity_and_hasher(capacity, hasher), + } + } + + #[inline] + pub fn hasher(&self) -> &S { + self.map.hasher() + } + + #[inline] + pub fn reserve(&mut self, additional: usize) { + self.map.reserve(additional) + } + + #[inline] + pub fn try_reserve(&mut self, additional: usize) -> Result<(), TryReserveError> { + self.map.try_reserve(additional) + } + + #[inline] + pub fn shrink_to_fit(&mut self) { + self.map.shrink_to_fit() + } + + #[inline] + pub fn difference<'a>(&'a self, other: &'a LinkedHashSet<T, S>) -> Difference<'a, T, S> { + Difference { + iter: self.iter(), + other, + } + } + + #[inline] + pub fn symmetric_difference<'a>( + &'a self, + other: &'a LinkedHashSet<T, S>, + ) -> SymmetricDifference<'a, T, S> { + SymmetricDifference { + iter: self.difference(other).chain(other.difference(self)), + } + } + + #[inline] + pub fn intersection<'a>(&'a self, other: &'a LinkedHashSet<T, S>) -> Intersection<'a, T, S> { + Intersection { + iter: self.iter(), + other, + } + } + + #[inline] + pub fn union<'a>(&'a self, other: &'a LinkedHashSet<T, S>) -> Union<'a, T, S> { + Union { + iter: self.iter().chain(other.difference(self)), + } + } + + #[inline] + pub fn contains<Q: ?Sized>(&self, value: &Q) -> bool + where + T: Borrow<Q>, + Q: Hash + Eq, + { + self.map.contains_key(value) + } + + #[inline] + pub fn get<Q: ?Sized>(&self, value: &Q) -> Option<&T> + where + T: Borrow<Q>, + Q: Hash + Eq, + { + self.map.raw_entry().from_key(value).map(|p| p.0) + } + + #[inline] + pub fn get_or_insert(&mut self, value: T) -> &T { + self.map + .raw_entry_mut() + .from_key(&value) + .or_insert(value, ()) + .0 + } + + #[inline] + pub fn get_or_insert_with<Q: ?Sized, F>(&mut self, value: &Q, f: F) -> &T + where + T: Borrow<Q>, + Q: Hash + Eq, + F: FnOnce(&Q) -> T, + { + self.map + .raw_entry_mut() + .from_key(value) + .or_insert_with(|| (f(value), ())) + .0 + } + + #[inline] + pub fn is_disjoint(&self, other: &LinkedHashSet<T, S>) -> bool { + self.iter().all(|v| !other.contains(v)) + } + + #[inline] + pub fn is_subset(&self, other: &LinkedHashSet<T, S>) -> bool { + self.iter().all(|v| other.contains(v)) + } + + #[inline] + pub fn is_superset(&self, other: &LinkedHashSet<T, S>) -> bool { + other.is_subset(self) + } + + /// Inserts the given value into the set. + /// + /// If the set did not have this value present, inserts it at the *back* of the internal linked + /// list and returns true, otherwise it moves the existing value to the *back* of the internal + /// linked list and returns false. + #[inline] + pub fn insert(&mut self, value: T) -> bool { + self.map.insert(value, ()).is_none() + } + + /// Adds the given value to the set, replacing the existing value. + /// + /// If a previous value existed, returns the replaced value. In this case, the value's position + /// in the internal linked list is *not* changed. + #[inline] + pub fn replace(&mut self, value: T) -> Option<T> { + match self.map.entry(value) { + linked_hash_map::Entry::Occupied(occupied) => Some(occupied.replace_key()), + linked_hash_map::Entry::Vacant(vacant) => { + vacant.insert(()); + None + } + } + } + + #[inline] + pub fn remove<Q: ?Sized>(&mut self, value: &Q) -> bool + where + T: Borrow<Q>, + Q: Hash + Eq, + { + self.map.remove(value).is_some() + } + + #[inline] + pub fn take<Q: ?Sized>(&mut self, value: &Q) -> Option<T> + where + T: Borrow<Q>, + Q: Hash + Eq, + { + match self.map.raw_entry_mut().from_key(value) { + linked_hash_map::RawEntryMut::Occupied(occupied) => Some(occupied.remove_entry().0), + linked_hash_map::RawEntryMut::Vacant(_) => None, + } + } + + #[inline] + pub fn front(&self) -> Option<&T> { + self.map.front().map(|(k, _)| k) + } + + #[inline] + pub fn pop_front(&mut self) -> Option<T> { + self.map.pop_front().map(|(k, _)| k) + } + + #[inline] + pub fn back(&self) -> Option<&T> { + self.map.back().map(|(k, _)| k) + } + + #[inline] + pub fn pop_back(&mut self) -> Option<T> { + self.map.pop_back().map(|(k, _)| k) + } + + #[inline] + pub fn to_front<Q: ?Sized>(&mut self, value: &Q) -> bool + where + T: Borrow<Q>, + Q: Hash + Eq, + { + match self.map.raw_entry_mut().from_key(value) { + linked_hash_map::RawEntryMut::Occupied(mut occupied) => { + occupied.to_front(); + true + } + linked_hash_map::RawEntryMut::Vacant(_) => false, + } + } + + #[inline] + pub fn to_back<Q: ?Sized>(&mut self, value: &Q) -> bool + where + T: Borrow<Q>, + Q: Hash + Eq, + { + match self.map.raw_entry_mut().from_key(value) { + linked_hash_map::RawEntryMut::Occupied(mut occupied) => { + occupied.to_back(); + true + } + linked_hash_map::RawEntryMut::Vacant(_) => false, + } + } + + #[inline] + pub fn retain_with_order<F>(&mut self, mut f: F) + where + F: FnMut(&T) -> bool, + { + self.map.retain_with_order(|k, _| f(k)); + } +} + +impl<T: Hash + Eq + Clone, S: BuildHasher + Clone> Clone for LinkedHashSet<T, S> { + #[inline] + fn clone(&self) -> Self { + let map = self.map.clone(); + Self { map } + } +} + +impl<T, S> PartialEq for LinkedHashSet<T, S> +where + T: Eq + Hash, + S: BuildHasher, +{ + #[inline] + fn eq(&self, other: &Self) -> bool { + self.len() == other.len() && self.iter().eq(other) + } +} + +impl<T, S> Hash for LinkedHashSet<T, S> +where + T: Eq + Hash, + S: BuildHasher, +{ + #[inline] + fn hash<H: Hasher>(&self, state: &mut H) { + for e in self { + e.hash(state); + } + } +} + +impl<T, S> Eq for LinkedHashSet<T, S> +where + T: Eq + Hash, + S: BuildHasher, +{ +} + +impl<T, S> fmt::Debug for LinkedHashSet<T, S> +where + T: fmt::Debug, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_set().entries(self.iter()).finish() + } +} + +impl<T, S> FromIterator<T> for LinkedHashSet<T, S> +where + T: Eq + Hash, + S: BuildHasher + Default, +{ + #[inline] + fn from_iter<I: IntoIterator<Item = T>>(iter: I) -> LinkedHashSet<T, S> { + let mut set = LinkedHashSet::with_hasher(Default::default()); + set.extend(iter); + set + } +} + +impl<T, S> Extend<T> for LinkedHashSet<T, S> +where + T: Eq + Hash, + S: BuildHasher, +{ + #[inline] + fn extend<I: IntoIterator<Item = T>>(&mut self, iter: I) { + self.map.extend(iter.into_iter().map(|k| (k, ()))); + } +} + +impl<'a, T, S> Extend<&'a T> for LinkedHashSet<T, S> +where + T: 'a + Eq + Hash + Copy, + S: BuildHasher, +{ + #[inline] + fn extend<I: IntoIterator<Item = &'a T>>(&mut self, iter: I) { + self.extend(iter.into_iter().cloned()); + } +} + +impl<T, S> Default for LinkedHashSet<T, S> +where + S: Default, +{ + #[inline] + fn default() -> LinkedHashSet<T, S> { + LinkedHashSet { + map: LinkedHashMap::default(), + } + } +} + +impl<'a, 'b, T, S> BitOr<&'b LinkedHashSet<T, S>> for &'a LinkedHashSet<T, S> +where + T: Eq + Hash + Clone, + S: BuildHasher + Default, +{ + type Output = LinkedHashSet<T, S>; + + #[inline] + fn bitor(self, rhs: &LinkedHashSet<T, S>) -> LinkedHashSet<T, S> { + self.union(rhs).cloned().collect() + } +} + +impl<'a, 'b, T, S> BitAnd<&'b LinkedHashSet<T, S>> for &'a LinkedHashSet<T, S> +where + T: Eq + Hash + Clone, + S: BuildHasher + Default, +{ + type Output = LinkedHashSet<T, S>; + + #[inline] + fn bitand(self, rhs: &LinkedHashSet<T, S>) -> LinkedHashSet<T, S> { + self.intersection(rhs).cloned().collect() + } +} + +impl<'a, 'b, T, S> BitXor<&'b LinkedHashSet<T, S>> for &'a LinkedHashSet<T, S> +where + T: Eq + Hash + Clone, + S: BuildHasher + Default, +{ + type Output = LinkedHashSet<T, S>; + + #[inline] + fn bitxor(self, rhs: &LinkedHashSet<T, S>) -> LinkedHashSet<T, S> { + self.symmetric_difference(rhs).cloned().collect() + } +} + +impl<'a, 'b, T, S> Sub<&'b LinkedHashSet<T, S>> for &'a LinkedHashSet<T, S> +where + T: Eq + Hash + Clone, + S: BuildHasher + Default, +{ + type Output = LinkedHashSet<T, S>; + + #[inline] + fn sub(self, rhs: &LinkedHashSet<T, S>) -> LinkedHashSet<T, S> { + self.difference(rhs).cloned().collect() + } +} + +pub struct Iter<'a, K> { + iter: linked_hash_map::Keys<'a, K, ()>, +} + +pub struct IntoIter<K> { + iter: linked_hash_map::IntoIter<K, ()>, +} + +pub struct Drain<'a, K: 'a> { + iter: linked_hash_map::Drain<'a, K, ()>, +} + +pub struct Intersection<'a, T, S> { + iter: Iter<'a, T>, + other: &'a LinkedHashSet<T, S>, +} + +pub struct Difference<'a, T, S> { + iter: Iter<'a, T>, + other: &'a LinkedHashSet<T, S>, +} + +pub struct SymmetricDifference<'a, T, S> { + iter: Chain<Difference<'a, T, S>, Difference<'a, T, S>>, +} + +pub struct Union<'a, T, S> { + iter: Chain<Iter<'a, T>, Difference<'a, T, S>>, +} + +impl<'a, T, S> IntoIterator for &'a LinkedHashSet<T, S> { + type Item = &'a T; + type IntoIter = Iter<'a, T>; + + #[inline] + fn into_iter(self) -> Iter<'a, T> { + self.iter() + } +} + +impl<T, S> IntoIterator for LinkedHashSet<T, S> { + type Item = T; + type IntoIter = IntoIter<T>; + + #[inline] + fn into_iter(self) -> IntoIter<T> { + IntoIter { + iter: self.map.into_iter(), + } + } +} + +impl<'a, K> Clone for Iter<'a, K> { + #[inline] + fn clone(&self) -> Iter<'a, K> { + Iter { + iter: self.iter.clone(), + } + } +} +impl<'a, K> Iterator for Iter<'a, K> { + type Item = &'a K; + + #[inline] + fn next(&mut self) -> Option<&'a K> { + self.iter.next() + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + self.iter.size_hint() + } +} + +impl<'a, K> ExactSizeIterator for Iter<'a, K> {} + +impl<'a, T> DoubleEndedIterator for Iter<'a, T> { + #[inline] + fn next_back(&mut self) -> Option<&'a T> { + self.iter.next_back() + } +} + +impl<'a, K: fmt::Debug> fmt::Debug for Iter<'a, K> { + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.clone()).finish() + } +} + +impl<K> Iterator for IntoIter<K> { + type Item = K; + + #[inline] + fn next(&mut self) -> Option<K> { + self.iter.next().map(|(k, _)| k) + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + self.iter.size_hint() + } +} + +impl<K> ExactSizeIterator for IntoIter<K> {} + +impl<K> DoubleEndedIterator for IntoIter<K> { + #[inline] + fn next_back(&mut self) -> Option<K> { + self.iter.next_back().map(|(k, _)| k) + } +} + +impl<'a, K> Iterator for Drain<'a, K> { + type Item = K; + + #[inline] + fn next(&mut self) -> Option<K> { + self.iter.next().map(|(k, _)| k) + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + self.iter.size_hint() + } +} + +impl<'a, K> DoubleEndedIterator for Drain<'a, K> { + #[inline] + fn next_back(&mut self) -> Option<K> { + self.iter.next_back().map(|(k, _)| k) + } +} + +impl<'a, K> ExactSizeIterator for Drain<'a, K> {} + +impl<'a, T, S> Clone for Intersection<'a, T, S> { + #[inline] + fn clone(&self) -> Intersection<'a, T, S> { + Intersection { + iter: self.iter.clone(), + ..*self + } + } +} + +impl<'a, T, S> Iterator for Intersection<'a, T, S> +where + T: Eq + Hash, + S: BuildHasher, +{ + type Item = &'a T; + + #[inline] + fn next(&mut self) -> Option<&'a T> { + loop { + match self.iter.next() { + None => return None, + Some(elt) => { + if self.other.contains(elt) { + return Some(elt); + } + } + } + } + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + let (_, upper) = self.iter.size_hint(); + (0, upper) + } +} + +impl<'a, T, S> fmt::Debug for Intersection<'a, T, S> +where + T: fmt::Debug + Eq + Hash, + S: BuildHasher, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.clone()).finish() + } +} + +impl<'a, T, S> Clone for Difference<'a, T, S> { + #[inline] + fn clone(&self) -> Difference<'a, T, S> { + Difference { + iter: self.iter.clone(), + ..*self + } + } +} + +impl<'a, T, S> Iterator for Difference<'a, T, S> +where + T: Eq + Hash, + S: BuildHasher, +{ + type Item = &'a T; + + #[inline] + fn next(&mut self) -> Option<&'a T> { + loop { + match self.iter.next() { + None => return None, + Some(elt) => { + if !self.other.contains(elt) { + return Some(elt); + } + } + } + } + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + let (_, upper) = self.iter.size_hint(); + (0, upper) + } +} + +impl<'a, T, S> fmt::Debug for Difference<'a, T, S> +where + T: fmt::Debug + Eq + Hash, + S: BuildHasher, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.clone()).finish() + } +} + +impl<'a, T, S> Clone for SymmetricDifference<'a, T, S> { + #[inline] + fn clone(&self) -> SymmetricDifference<'a, T, S> { + SymmetricDifference { + iter: self.iter.clone(), + } + } +} + +impl<'a, T, S> Iterator for SymmetricDifference<'a, T, S> +where + T: Eq + Hash, + S: BuildHasher, +{ + type Item = &'a T; + + #[inline] + fn next(&mut self) -> Option<&'a T> { + self.iter.next() + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + self.iter.size_hint() + } +} + +impl<'a, T, S> fmt::Debug for SymmetricDifference<'a, T, S> +where + T: fmt::Debug + Eq + Hash, + S: BuildHasher, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.clone()).finish() + } +} + +impl<'a, T, S> Clone for Union<'a, T, S> { + #[inline] + fn clone(&self) -> Union<'a, T, S> { + Union { + iter: self.iter.clone(), + } + } +} + +impl<'a, T, S> fmt::Debug for Union<'a, T, S> +where + T: fmt::Debug + Eq + Hash, + S: BuildHasher, +{ + #[inline] + fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { + f.debug_list().entries(self.clone()).finish() + } +} + +impl<'a, T, S> Iterator for Union<'a, T, S> +where + T: Eq + Hash, + S: BuildHasher, +{ + type Item = &'a T; + + #[inline] + fn next(&mut self) -> Option<&'a T> { + self.iter.next() + } + + #[inline] + fn size_hint(&self) -> (usize, Option<usize>) { + self.iter.size_hint() + } +} diff --git a/third_party/rust/hashlink/src/lru_cache.rs b/third_party/rust/hashlink/src/lru_cache.rs new file mode 100644 index 0000000000..9e5740ea60 --- /dev/null +++ b/third_party/rust/hashlink/src/lru_cache.rs @@ -0,0 +1,292 @@ +use std::{ + borrow::Borrow, + fmt, + hash::{BuildHasher, Hash}, + usize, +}; + +use hashbrown::hash_map; + +use crate::linked_hash_map::{self, LinkedHashMap}; + +pub use crate::linked_hash_map::{ + Drain, Entry, IntoIter, Iter, IterMut, OccupiedEntry, RawEntryBuilder, RawEntryBuilderMut, + RawOccupiedEntryMut, RawVacantEntryMut, VacantEntry, +}; + +pub struct LruCache<K, V, S = hash_map::DefaultHashBuilder> { + map: LinkedHashMap<K, V, S>, + max_size: usize, +} + +impl<K: Eq + Hash, V> LruCache<K, V> { + #[inline] + pub fn new(capacity: usize) -> Self { + LruCache { + map: LinkedHashMap::new(), + max_size: capacity, + } + } + + /// Create a new unbounded `LruCache` that does not automatically evict entries. + /// + /// A simple convenience method that is equivalent to `LruCache::new(usize::MAX)` + #[inline] + pub fn new_unbounded() -> Self { + LruCache::new(usize::MAX) + } +} + +impl<K, V, S> LruCache<K, V, S> { + #[inline] + pub fn with_hasher(capacity: usize, hash_builder: S) -> Self { + LruCache { + map: LinkedHashMap::with_hasher(hash_builder), + max_size: capacity, + } + } + + #[inline] + pub fn capacity(&self) -> usize { + self.max_size + } + + #[inline] + pub fn len(&self) -> usize { + self.map.len() + } + + #[inline] + pub fn is_empty(&self) -> bool { + self.map.is_empty() + } + + #[inline] + pub fn clear(&mut self) { + self.map.clear(); + } + + #[inline] + pub fn iter(&self) -> Iter<K, V> { + self.map.iter() + } + + #[inline] + pub fn iter_mut(&mut self) -> IterMut<K, V> { + self.map.iter_mut() + } + + #[inline] + pub fn drain(&mut self) -> Drain<K, V> { + self.map.drain() + } +} + +impl<K: Eq + Hash, V, S> LruCache<K, V, S> +where + S: BuildHasher, +{ + #[inline] + pub fn contains_key<Q>(&mut self, key: &Q) -> bool + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.get_mut(key).is_some() + } + + /// Insert a new value into the `LruCache`. + /// + /// If necessary, will remove the value at the front of the LRU list to make room. + #[inline] + pub fn insert(&mut self, k: K, v: V) -> Option<V> { + let old_val = self.map.insert(k, v); + if self.len() > self.capacity() { + self.remove_lru(); + } + old_val + } + + /// Get the value for the given key, *without* marking the value as recently used and moving it + /// to the back of the LRU list. + #[inline] + pub fn peek<Q>(&self, k: &Q) -> Option<&V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.map.get(k) + } + + /// Get the value for the given key mutably, *without* marking the value as recently used and + /// moving it to the back of the LRU list. + #[inline] + pub fn peek_mut<Q>(&mut self, k: &Q) -> Option<&mut V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.map.get_mut(k) + } + + /// Retrieve the given key, marking it as recently used and moving it to the back of the LRU + /// list. + #[inline] + pub fn get<Q>(&mut self, k: &Q) -> Option<&V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.get_mut(k).map(|v| &*v) + } + + /// Retrieve the given key, marking it as recently used and moving it to the back of the LRU + /// list. + #[inline] + pub fn get_mut<Q>(&mut self, k: &Q) -> Option<&mut V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + match self.map.raw_entry_mut().from_key(k) { + linked_hash_map::RawEntryMut::Occupied(mut occupied) => { + occupied.to_back(); + Some(occupied.into_mut()) + } + linked_hash_map::RawEntryMut::Vacant(_) => None, + } + } + + /// If the returned entry is vacant, it will always have room to insert a single value. By + /// using the entry API, you can exceed the configured capacity by 1. + /// + /// The returned entry is not automatically moved to the back of the LRU list. By calling + /// `Entry::to_back` / `Entry::to_front` you can manually control the position of this entry in + /// the LRU list. + #[inline] + pub fn entry(&mut self, key: K) -> Entry<'_, K, V, S> { + if self.len() > self.capacity() { + self.remove_lru(); + } + self.map.entry(key) + } + + /// The constructed raw entry is never automatically moved to the back of the LRU list. By + /// calling `Entry::to_back` / `Entry::to_front` you can manually control the position of this + /// entry in the LRU list. + #[inline] + pub fn raw_entry(&self) -> RawEntryBuilder<'_, K, V, S> { + self.map.raw_entry() + } + + /// If the constructed raw entry is vacant, it will always have room to insert a single value. + /// By using the raw entry API, you can exceed the configured capacity by 1. + /// + /// The constructed raw entry is never automatically moved to the back of the LRU list. By + /// calling `Entry::to_back` / `Entry::to_front` you can manually control the position of this + /// entry in the LRU list. + #[inline] + pub fn raw_entry_mut(&mut self) -> RawEntryBuilderMut<'_, K, V, S> { + if self.len() > self.capacity() { + self.remove_lru(); + } + self.map.raw_entry_mut() + } + + #[inline] + pub fn remove<Q>(&mut self, k: &Q) -> Option<V> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.map.remove(k) + } + + #[inline] + pub fn remove_entry<Q>(&mut self, k: &Q) -> Option<(K, V)> + where + K: Borrow<Q>, + Q: Hash + Eq + ?Sized, + { + self.map.remove_entry(k) + } + + /// Set the new cache capacity for the `LruCache`. + /// + /// If there are more entries in the `LruCache` than the new capacity will allow, they are + /// removed. + #[inline] + pub fn set_capacity(&mut self, capacity: usize) { + for _ in capacity..self.len() { + self.remove_lru(); + } + self.max_size = capacity; + } + + /// Remove the least recently used entry and return it. + /// + /// If the `LruCache` is empty this will return None. + #[inline] + pub fn remove_lru(&mut self) -> Option<(K, V)> { + self.map.pop_front() + } +} + +impl<K: Hash + Eq + Clone, V: Clone, S: BuildHasher + Clone> Clone for LruCache<K, V, S> { + #[inline] + fn clone(&self) -> Self { + LruCache { + map: self.map.clone(), + max_size: self.max_size, + } + } +} + +impl<K: Eq + Hash, V, S: BuildHasher> Extend<(K, V)> for LruCache<K, V, S> { + #[inline] + fn extend<I: IntoIterator<Item = (K, V)>>(&mut self, iter: I) { + for (k, v) in iter { + self.insert(k, v); + } + } +} + +impl<K, V, S> IntoIterator for LruCache<K, V, S> { + type Item = (K, V); + type IntoIter = IntoIter<K, V>; + + #[inline] + fn into_iter(self) -> IntoIter<K, V> { + self.map.into_iter() + } +} + +impl<'a, K, V, S> IntoIterator for &'a LruCache<K, V, S> { + type Item = (&'a K, &'a V); + type IntoIter = Iter<'a, K, V>; + + #[inline] + fn into_iter(self) -> Iter<'a, K, V> { + self.iter() + } +} + +impl<'a, K, V, S> IntoIterator for &'a mut LruCache<K, V, S> { + type Item = (&'a K, &'a mut V); + type IntoIter = IterMut<'a, K, V>; + + #[inline] + fn into_iter(self) -> IterMut<'a, K, V> { + self.iter_mut() + } +} + +impl<K, V, S> fmt::Debug for LruCache<K, V, S> +where + K: fmt::Debug, + V: fmt::Debug, +{ + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + f.debug_map().entries(self.iter().rev()).finish() + } +} diff --git a/third_party/rust/hashlink/src/serde.rs b/third_party/rust/hashlink/src/serde.rs new file mode 100644 index 0000000000..f44ebb3a65 --- /dev/null +++ b/third_party/rust/hashlink/src/serde.rs @@ -0,0 +1,161 @@ +use std::{ + fmt::{self, Formatter}, + hash::{BuildHasher, Hash}, + marker::PhantomData, +}; + +use serde::{ + de::{MapAccess, SeqAccess, Visitor}, + ser::{SerializeMap, SerializeSeq}, + Deserialize, Deserializer, Serialize, Serializer, +}; + +use crate::{LinkedHashMap, LinkedHashSet}; + +// LinkedHashMap impls + +impl<K, V, S> Serialize for LinkedHashMap<K, V, S> +where + K: Serialize + Eq + Hash, + V: Serialize, + S: BuildHasher, +{ + #[inline] + fn serialize<T: Serializer>(&self, serializer: T) -> Result<T::Ok, T::Error> { + let mut map_serializer = serializer.serialize_map(Some(self.len()))?; + for (k, v) in self { + map_serializer.serialize_key(k)?; + map_serializer.serialize_value(v)?; + } + map_serializer.end() + } +} + +impl<'de, K, V, S> Deserialize<'de> for LinkedHashMap<K, V, S> +where + K: Deserialize<'de> + Eq + Hash, + V: Deserialize<'de>, + S: BuildHasher + Default, +{ + fn deserialize<D: Deserializer<'de>>(deserializer: D) -> Result<Self, D::Error> { + #[derive(Debug)] + pub struct LinkedHashMapVisitor<K, V, S> { + marker: PhantomData<LinkedHashMap<K, V, S>>, + } + + impl<K, V, S> LinkedHashMapVisitor<K, V, S> { + fn new() -> Self { + LinkedHashMapVisitor { + marker: PhantomData, + } + } + } + + impl<K, V, S> Default for LinkedHashMapVisitor<K, V, S> { + fn default() -> Self { + Self::new() + } + } + + impl<'de, K, V, S> Visitor<'de> for LinkedHashMapVisitor<K, V, S> + where + K: Deserialize<'de> + Eq + Hash, + V: Deserialize<'de>, + S: BuildHasher + Default, + { + type Value = LinkedHashMap<K, V, S>; + + fn expecting(&self, formatter: &mut Formatter) -> fmt::Result { + write!(formatter, "a map") + } + + #[inline] + fn visit_map<M: MapAccess<'de>>(self, mut map: M) -> Result<Self::Value, M::Error> { + let mut values = LinkedHashMap::with_capacity_and_hasher( + map.size_hint().unwrap_or(0), + S::default(), + ); + + while let Some((k, v)) = map.next_entry()? { + values.insert(k, v); + } + + Ok(values) + } + } + + deserializer.deserialize_map(LinkedHashMapVisitor::default()) + } +} + +// LinkedHashSet impls + +impl<T, S> Serialize for LinkedHashSet<T, S> +where + T: Serialize + Eq + Hash, + S: BuildHasher, +{ + #[inline] + fn serialize<U: Serializer>(&self, serializer: U) -> Result<U::Ok, U::Error> { + let mut seq_serializer = serializer.serialize_seq(Some(self.len()))?; + for v in self { + seq_serializer.serialize_element(v)?; + } + seq_serializer.end() + } +} + +impl<'de, T, S> Deserialize<'de> for LinkedHashSet<T, S> +where + T: Deserialize<'de> + Eq + Hash, + S: BuildHasher + Default, +{ + fn deserialize<D: Deserializer<'de>>(deserializer: D) -> Result<Self, D::Error> { + #[derive(Debug)] + pub struct LinkedHashSetVisitor<T, S> { + marker: PhantomData<LinkedHashSet<T, S>>, + } + + impl<T, S> LinkedHashSetVisitor<T, S> { + fn new() -> Self { + LinkedHashSetVisitor { + marker: PhantomData, + } + } + } + + impl<T, S> Default for LinkedHashSetVisitor<T, S> { + fn default() -> Self { + Self::new() + } + } + + impl<'de, T, S> Visitor<'de> for LinkedHashSetVisitor<T, S> + where + T: Deserialize<'de> + Eq + Hash, + S: BuildHasher + Default, + { + type Value = LinkedHashSet<T, S>; + + fn expecting(&self, formatter: &mut Formatter) -> fmt::Result { + write!(formatter, "a sequence") + } + + #[inline] + fn visit_seq<SA: SeqAccess<'de>>(self, mut seq: SA) -> Result<Self::Value, SA::Error> { + let mut values = LinkedHashSet::with_capacity_and_hasher( + seq.size_hint().unwrap_or(0), + S::default(), + ); + + while let Some(v) = seq.next_element()? { + values.insert(v); + } + + Ok(values) + } + } + + deserializer.deserialize_seq(LinkedHashSetVisitor::default()) + } +} |