hashbrown/raw/
mod.rs

1use crate::alloc::alloc::{handle_alloc_error, Layout};
2use crate::scopeguard::{guard, ScopeGuard};
3use crate::TryReserveError;
4use core::iter::FusedIterator;
5use core::marker::PhantomData;
6use core::mem;
7use core::mem::MaybeUninit;
8use core::ptr::NonNull;
9use core::{hint, ptr};
10
11cfg_if! {
12    // Use the SSE2 implementation if possible: it allows us to scan 16 buckets
13    // at once instead of 8. We don't bother with AVX since it would require
14    // runtime dispatch and wouldn't gain us much anyways: the probability of
15    // finding a match drops off drastically after the first few buckets.
16    //
17    // I attempted an implementation on ARM using NEON instructions, but it
18    // turns out that most NEON instructions have multi-cycle latency, which in
19    // the end outweighs any gains over the generic implementation.
20    if #[cfg(all(
21        target_feature = "sse2",
22        any(target_arch = "x86", target_arch = "x86_64"),
23        not(miri),
24    ))] {
25        mod sse2;
26        use sse2 as imp;
27    } else if #[cfg(all(
28        target_arch = "aarch64",
29        target_feature = "neon",
30        // NEON intrinsics are currently broken on big-endian targets.
31        // See https://github.com/rust-lang/stdarch/issues/1484.
32        target_endian = "little",
33        not(miri),
34    ))] {
35        mod neon;
36        use neon as imp;
37    } else {
38        mod generic;
39        use generic as imp;
40    }
41}
42
43mod alloc;
44pub(crate) use self::alloc::{do_alloc, Allocator, Global};
45
46mod bitmask;
47
48use self::bitmask::BitMaskIter;
49use self::imp::Group;
50
51// Branch prediction hint. This is currently only available on nightly but it
52// consistently improves performance by 10-15%.
53#[cfg(not(feature = "nightly"))]
54use core::convert::identity as likely;
55#[cfg(not(feature = "nightly"))]
56use core::convert::identity as unlikely;
57#[cfg(feature = "nightly")]
58use core::intrinsics::{likely, unlikely};
59
60// FIXME: use strict provenance functions once they are stable.
61// Implement it with a transmute for now.
62#[inline(always)]
63#[allow(clippy::useless_transmute)] // clippy is wrong, cast and transmute are different here
64fn invalid_mut<T>(addr: usize) -> *mut T {
65    unsafe { core::mem::transmute(addr) }
66}
67
68#[inline]
69unsafe fn offset_from<T>(to: *const T, from: *const T) -> usize {
70    to.offset_from(from) as usize
71}
72
73/// Whether memory allocation errors should return an error or abort.
74#[derive(Copy, Clone)]
75enum Fallibility {
76    Fallible,
77    Infallible,
78}
79
80impl Fallibility {
81    /// Error to return on capacity overflow.
82    #[cfg_attr(feature = "inline-more", inline)]
83    fn capacity_overflow(self) -> TryReserveError {
84        match self {
85            Fallibility::Fallible => TryReserveError::CapacityOverflow,
86            Fallibility::Infallible => panic!("Hash table capacity overflow"),
87        }
88    }
89
90    /// Error to return on allocation error.
91    #[cfg_attr(feature = "inline-more", inline)]
92    fn alloc_err(self, layout: Layout) -> TryReserveError {
93        match self {
94            Fallibility::Fallible => TryReserveError::AllocError { layout },
95            Fallibility::Infallible => handle_alloc_error(layout),
96        }
97    }
98}
99
100trait SizedTypeProperties: Sized {
101    const IS_ZERO_SIZED: bool = mem::size_of::<Self>() == 0;
102    const NEEDS_DROP: bool = mem::needs_drop::<Self>();
103}
104
105impl<T> SizedTypeProperties for T {}
106
107/// Control byte value for an empty bucket.
108const EMPTY: u8 = 0b1111_1111;
109
110/// Control byte value for a deleted bucket.
111const DELETED: u8 = 0b1000_0000;
112
113/// Checks whether a control byte represents a full bucket (top bit is clear).
114#[inline]
115fn is_full(ctrl: u8) -> bool {
116    ctrl & 0x80 == 0
117}
118
119/// Checks whether a control byte represents a special value (top bit is set).
120#[inline]
121fn is_special(ctrl: u8) -> bool {
122    ctrl & 0x80 != 0
123}
124
125/// Checks whether a special control value is EMPTY (just check 1 bit).
126#[inline]
127fn special_is_empty(ctrl: u8) -> bool {
128    debug_assert!(is_special(ctrl));
129    ctrl & 0x01 != 0
130}
131
132/// Primary hash function, used to select the initial bucket to probe from.
133#[inline]
134#[allow(clippy::cast_possible_truncation)]
135fn h1(hash: u64) -> usize {
136    // On 32-bit platforms we simply ignore the higher hash bits.
137    hash as usize
138}
139
140// Constant for h2 function that grabing the top 7 bits of the hash.
141const MIN_HASH_LEN: usize = if mem::size_of::<usize>() < mem::size_of::<u64>() {
142    mem::size_of::<usize>()
143} else {
144    mem::size_of::<u64>()
145};
146
147/// Secondary hash function, saved in the low 7 bits of the control byte.
148#[inline]
149#[allow(clippy::cast_possible_truncation)]
150fn h2(hash: u64) -> u8 {
151    // Grab the top 7 bits of the hash. While the hash is normally a full 64-bit
152    // value, some hash functions (such as FxHash) produce a usize result
153    // instead, which means that the top 32 bits are 0 on 32-bit platforms.
154    // So we use MIN_HASH_LEN constant to handle this.
155    let top7 = hash >> (MIN_HASH_LEN * 8 - 7);
156    (top7 & 0x7f) as u8 // truncation
157}
158
159/// Probe sequence based on triangular numbers, which is guaranteed (since our
160/// table size is a power of two) to visit every group of elements exactly once.
161///
162/// A triangular probe has us jump by 1 more group every time. So first we
163/// jump by 1 group (meaning we just continue our linear scan), then 2 groups
164/// (skipping over 1 group), then 3 groups (skipping over 2 groups), and so on.
165///
166/// Proof that the probe will visit every group in the table:
167/// <https://fgiesen.wordpress.com/2015/02/22/triangular-numbers-mod-2n/>
168struct ProbeSeq {
169    pos: usize,
170    stride: usize,
171}
172
173impl ProbeSeq {
174    #[inline]
175    fn move_next(&mut self, bucket_mask: usize) {
176        // We should have found an empty bucket by now and ended the probe.
177        debug_assert!(
178            self.stride <= bucket_mask,
179            "Went past end of probe sequence"
180        );
181
182        self.stride += Group::WIDTH;
183        self.pos += self.stride;
184        self.pos &= bucket_mask;
185    }
186}
187
188/// Returns the number of buckets needed to hold the given number of items,
189/// taking the maximum load factor into account.
190///
191/// Returns `None` if an overflow occurs.
192// Workaround for emscripten bug emscripten-core/emscripten-fastcomp#258
193#[cfg_attr(target_os = "emscripten", inline(never))]
194#[cfg_attr(not(target_os = "emscripten"), inline)]
195fn capacity_to_buckets(cap: usize) -> Option<usize> {
196    debug_assert_ne!(cap, 0);
197
198    // For small tables we require at least 1 empty bucket so that lookups are
199    // guaranteed to terminate if an element doesn't exist in the table.
200    if cap < 8 {
201        // We don't bother with a table size of 2 buckets since that can only
202        // hold a single element. Instead we skip directly to a 4 bucket table
203        // which can hold 3 elements.
204        return Some(if cap < 4 { 4 } else { 8 });
205    }
206
207    // Otherwise require 1/8 buckets to be empty (87.5% load)
208    //
209    // Be careful when modifying this, calculate_layout relies on the
210    // overflow check here.
211    let adjusted_cap = cap.checked_mul(8)? / 7;
212
213    // Any overflows will have been caught by the checked_mul. Also, any
214    // rounding errors from the division above will be cleaned up by
215    // next_power_of_two (which can't overflow because of the previous division).
216    Some(adjusted_cap.next_power_of_two())
217}
218
219/// Returns the maximum effective capacity for the given bucket mask, taking
220/// the maximum load factor into account.
221#[inline]
222fn bucket_mask_to_capacity(bucket_mask: usize) -> usize {
223    if bucket_mask < 8 {
224        // For tables with 1/2/4/8 buckets, we always reserve one empty slot.
225        // Keep in mind that the bucket mask is one less than the bucket count.
226        bucket_mask
227    } else {
228        // For larger tables we reserve 12.5% of the slots as empty.
229        ((bucket_mask + 1) / 8) * 7
230    }
231}
232
233/// Helper which allows the max calculation for ctrl_align to be statically computed for each T
234/// while keeping the rest of `calculate_layout_for` independent of `T`
235#[derive(Copy, Clone)]
236struct TableLayout {
237    size: usize,
238    ctrl_align: usize,
239}
240
241impl TableLayout {
242    #[inline]
243    const fn new<T>() -> Self {
244        let layout = Layout::new::<T>();
245        Self {
246            size: layout.size(),
247            ctrl_align: if layout.align() > Group::WIDTH {
248                layout.align()
249            } else {
250                Group::WIDTH
251            },
252        }
253    }
254
255    #[inline]
256    fn calculate_layout_for(self, buckets: usize) -> Option<(Layout, usize)> {
257        debug_assert!(buckets.is_power_of_two());
258
259        let TableLayout { size, ctrl_align } = self;
260        // Manual layout calculation since Layout methods are not yet stable.
261        let ctrl_offset =
262            size.checked_mul(buckets)?.checked_add(ctrl_align - 1)? & !(ctrl_align - 1);
263        let len = ctrl_offset.checked_add(buckets + Group::WIDTH)?;
264
265        // We need an additional check to ensure that the allocation doesn't
266        // exceed `isize::MAX` (https://github.com/rust-lang/rust/pull/95295).
267        if len > isize::MAX as usize - (ctrl_align - 1) {
268            return None;
269        }
270
271        Some((
272            unsafe { Layout::from_size_align_unchecked(len, ctrl_align) },
273            ctrl_offset,
274        ))
275    }
276}
277
278/// A reference to an empty bucket into which an can be inserted.
279pub struct InsertSlot {
280    index: usize,
281}
282
283/// A reference to a hash table bucket containing a `T`.
284///
285/// This is usually just a pointer to the element itself. However if the element
286/// is a ZST, then we instead track the index of the element in the table so
287/// that `erase` works properly.
288pub struct Bucket<T> {
289    // Actually it is pointer to next element than element itself
290    // this is needed to maintain pointer arithmetic invariants
291    // keeping direct pointer to element introduces difficulty.
292    // Using `NonNull` for variance and niche layout
293    ptr: NonNull<T>,
294}
295
296// This Send impl is needed for rayon support. This is safe since Bucket is
297// never exposed in a public API.
298unsafe impl<T> Send for Bucket<T> {}
299
300impl<T> Clone for Bucket<T> {
301    #[inline]
302    fn clone(&self) -> Self {
303        Self { ptr: self.ptr }
304    }
305}
306
307impl<T> Bucket<T> {
308    /// Creates a [`Bucket`] that contain pointer to the data.
309    /// The pointer calculation is performed by calculating the
310    /// offset from given `base` pointer (convenience for
311    /// `base.as_ptr().sub(index)`).
312    ///
313    /// `index` is in units of `T`; e.g., an `index` of 3 represents a pointer
314    /// offset of `3 * size_of::<T>()` bytes.
315    ///
316    /// If the `T` is a ZST, then we instead track the index of the element
317    /// in the table so that `erase` works properly (return
318    /// `NonNull::new_unchecked((index + 1) as *mut T)`)
319    ///
320    /// # Safety
321    ///
322    /// If `mem::size_of::<T>() != 0`, then the safety rules are directly derived
323    /// from the safety rules for [`<*mut T>::sub`] method of `*mut T` and the safety
324    /// rules of [`NonNull::new_unchecked`] function.
325    ///
326    /// Thus, in order to uphold the safety contracts for the [`<*mut T>::sub`] method
327    /// and [`NonNull::new_unchecked`] function, as well as for the correct
328    /// logic of the work of this crate, the following rules are necessary and
329    /// sufficient:
330    ///
331    /// * the `base` pointer must not be `dangling` and must points to the
332    ///   end of the first `value element` from the `data part` of the table, i.e.
333    ///   must be the pointer that returned by [`RawTable::data_end`] or by
334    ///   [`RawTableInner::data_end<T>`];
335    ///
336    /// * `index` must not be greater than `RawTableInner.bucket_mask`, i.e.
337    ///   `index <= RawTableInner.bucket_mask` or, in other words, `(index + 1)`
338    ///   must be no greater than the number returned by the function
339    ///   [`RawTable::buckets`] or [`RawTableInner::buckets`].
340    ///
341    /// If `mem::size_of::<T>() == 0`, then the only requirement is that the
342    /// `index` must not be greater than `RawTableInner.bucket_mask`, i.e.
343    /// `index <= RawTableInner.bucket_mask` or, in other words, `(index + 1)`
344    /// must be no greater than the number returned by the function
345    /// [`RawTable::buckets`] or [`RawTableInner::buckets`].
346    ///
347    /// [`Bucket`]: crate::raw::Bucket
348    /// [`<*mut T>::sub`]: https://doc.rust-lang.org/core/primitive.pointer.html#method.sub-1
349    /// [`NonNull::new_unchecked`]: https://doc.rust-lang.org/stable/std/ptr/struct.NonNull.html#method.new_unchecked
350    /// [`RawTable::data_end`]: crate::raw::RawTable::data_end
351    /// [`RawTableInner::data_end<T>`]: RawTableInner::data_end<T>
352    /// [`RawTable::buckets`]: crate::raw::RawTable::buckets
353    /// [`RawTableInner::buckets`]: RawTableInner::buckets
354    #[inline]
355    unsafe fn from_base_index(base: NonNull<T>, index: usize) -> Self {
356        // If mem::size_of::<T>() != 0 then return a pointer to an `element` in
357        // the data part of the table (we start counting from "0", so that
358        // in the expression T[last], the "last" index actually one less than the
359        // "buckets" number in the table, i.e. "last = RawTableInner.bucket_mask"):
360        //
361        //                   `from_base_index(base, 1).as_ptr()` returns a pointer that
362        //                   points here in the data part of the table
363        //                   (to the start of T1)
364        //                        |
365        //                        |        `base: NonNull<T>` must point here
366        //                        |         (to the end of T0 or to the start of C0)
367        //                        v         v
368        // [Padding], Tlast, ..., |T1|, T0, |C0, C1, ..., Clast
369        //                           ^
370        //                           `from_base_index(base, 1)` returns a pointer
371        //                           that points here in the data part of the table
372        //                           (to the end of T1)
373        //
374        // where: T0...Tlast - our stored data; C0...Clast - control bytes
375        // or metadata for data.
376        let ptr = if T::IS_ZERO_SIZED {
377            // won't overflow because index must be less than length (bucket_mask)
378            // and bucket_mask is guaranteed to be less than `isize::MAX`
379            // (see TableLayout::calculate_layout_for method)
380            invalid_mut(index + 1)
381        } else {
382            base.as_ptr().sub(index)
383        };
384        Self {
385            ptr: NonNull::new_unchecked(ptr),
386        }
387    }
388
389    /// Calculates the index of a [`Bucket`] as distance between two pointers
390    /// (convenience for `base.as_ptr().offset_from(self.ptr.as_ptr()) as usize`).
391    /// The returned value is in units of T: the distance in bytes divided by
392    /// [`core::mem::size_of::<T>()`].
393    ///
394    /// If the `T` is a ZST, then we return the index of the element in
395    /// the table so that `erase` works properly (return `self.ptr.as_ptr() as usize - 1`).
396    ///
397    /// This function is the inverse of [`from_base_index`].
398    ///
399    /// # Safety
400    ///
401    /// If `mem::size_of::<T>() != 0`, then the safety rules are directly derived
402    /// from the safety rules for [`<*const T>::offset_from`] method of `*const T`.
403    ///
404    /// Thus, in order to uphold the safety contracts for [`<*const T>::offset_from`]
405    /// method, as well as for the correct logic of the work of this crate, the
406    /// following rules are necessary and sufficient:
407    ///
408    /// * `base` contained pointer must not be `dangling` and must point to the
409    ///   end of the first `element` from the `data part` of the table, i.e.
410    ///   must be a pointer that returns by [`RawTable::data_end`] or by
411    ///   [`RawTableInner::data_end<T>`];
412    ///
413    /// * `self` also must not contain dangling pointer;
414    ///
415    /// * both `self` and `base` must be created from the same [`RawTable`]
416    ///   (or [`RawTableInner`]).
417    ///
418    /// If `mem::size_of::<T>() == 0`, this function is always safe.
419    ///
420    /// [`Bucket`]: crate::raw::Bucket
421    /// [`from_base_index`]: crate::raw::Bucket::from_base_index
422    /// [`RawTable::data_end`]: crate::raw::RawTable::data_end
423    /// [`RawTableInner::data_end<T>`]: RawTableInner::data_end<T>
424    /// [`RawTable`]: crate::raw::RawTable
425    /// [`RawTableInner`]: RawTableInner
426    /// [`<*const T>::offset_from`]: https://doc.rust-lang.org/nightly/core/primitive.pointer.html#method.offset_from
427    #[inline]
428    unsafe fn to_base_index(&self, base: NonNull<T>) -> usize {
429        // If mem::size_of::<T>() != 0 then return an index under which we used to store the
430        // `element` in the data part of the table (we start counting from "0", so
431        // that in the expression T[last], the "last" index actually is one less than the
432        // "buckets" number in the table, i.e. "last = RawTableInner.bucket_mask").
433        // For example for 5th element in table calculation is performed like this:
434        //
435        //                        mem::size_of::<T>()
436        //                          |
437        //                          |         `self = from_base_index(base, 5)` that returns pointer
438        //                          |         that points here in tha data part of the table
439        //                          |         (to the end of T5)
440        //                          |           |                    `base: NonNull<T>` must point here
441        //                          v           |                    (to the end of T0 or to the start of C0)
442        //                        /???\         v                      v
443        // [Padding], Tlast, ..., |T10|, ..., T5|, T4, T3, T2, T1, T0, |C0, C1, C2, C3, C4, C5, ..., C10, ..., Clast
444        //                                      \__________  __________/
445        //                                                 \/
446        //                                     `bucket.to_base_index(base)` = 5
447        //                                     (base.as_ptr() as usize - self.ptr.as_ptr() as usize) / mem::size_of::<T>()
448        //
449        // where: T0...Tlast - our stored data; C0...Clast - control bytes or metadata for data.
450        if T::IS_ZERO_SIZED {
451            // this can not be UB
452            self.ptr.as_ptr() as usize - 1
453        } else {
454            offset_from(base.as_ptr(), self.ptr.as_ptr())
455        }
456    }
457
458    /// Acquires the underlying raw pointer `*mut T` to `data`.
459    ///
460    /// # Note
461    ///
462    /// If `T` is not [`Copy`], do not use `*mut T` methods that can cause calling the
463    /// destructor of `T` (for example the [`<*mut T>::drop_in_place`] method), because
464    /// for properly dropping the data we also need to clear `data` control bytes. If we
465    /// drop data, but do not clear `data control byte` it leads to double drop when
466    /// [`RawTable`] goes out of scope.
467    ///
468    /// If you modify an already initialized `value`, so [`Hash`] and [`Eq`] on the new
469    /// `T` value and its borrowed form *must* match those for the old `T` value, as the map
470    /// will not re-evaluate where the new value should go, meaning the value may become
471    /// "lost" if their location does not reflect their state.
472    ///
473    /// [`RawTable`]: crate::raw::RawTable
474    /// [`<*mut T>::drop_in_place`]: https://doc.rust-lang.org/core/primitive.pointer.html#method.drop_in_place
475    /// [`Hash`]: https://doc.rust-lang.org/core/hash/trait.Hash.html
476    /// [`Eq`]: https://doc.rust-lang.org/core/cmp/trait.Eq.html
477    ///
478    /// # Examples
479    ///
480    /// ```
481    /// # #[cfg(feature = "raw")]
482    /// # fn test() {
483    /// use core::hash::{BuildHasher, Hash};
484    /// use hashbrown::raw::{Bucket, RawTable};
485    ///
486    /// type NewHashBuilder = core::hash::BuildHasherDefault<ahash::AHasher>;
487    ///
488    /// fn make_hash<K: Hash + ?Sized, S: BuildHasher>(hash_builder: &S, key: &K) -> u64 {
489    ///     use core::hash::Hasher;
490    ///     let mut state = hash_builder.build_hasher();
491    ///     key.hash(&mut state);
492    ///     state.finish()
493    /// }
494    ///
495    /// let hash_builder = NewHashBuilder::default();
496    /// let mut table = RawTable::new();
497    ///
498    /// let value = ("a", 100);
499    /// let hash = make_hash(&hash_builder, &value.0);
500    ///
501    /// table.insert(hash, value.clone(), |val| make_hash(&hash_builder, &val.0));
502    ///
503    /// let bucket: Bucket<(&str, i32)> = table.find(hash, |(k1, _)| k1 == &value.0).unwrap();
504    ///
505    /// assert_eq!(unsafe { &*bucket.as_ptr() }, &("a", 100));
506    /// # }
507    /// # fn main() {
508    /// #     #[cfg(feature = "raw")]
509    /// #     test()
510    /// # }
511    /// ```
512    #[inline]
513    pub fn as_ptr(&self) -> *mut T {
514        if T::IS_ZERO_SIZED {
515            // Just return an arbitrary ZST pointer which is properly aligned
516            // invalid pointer is good enough for ZST
517            invalid_mut(mem::align_of::<T>())
518        } else {
519            unsafe { self.ptr.as_ptr().sub(1) }
520        }
521    }
522
523    /// Create a new [`Bucket`] that is offset from the `self` by the given
524    /// `offset`. The pointer calculation is performed by calculating the
525    /// offset from `self` pointer (convenience for `self.ptr.as_ptr().sub(offset)`).
526    /// This function is used for iterators.
527    ///
528    /// `offset` is in units of `T`; e.g., a `offset` of 3 represents a pointer
529    /// offset of `3 * size_of::<T>()` bytes.
530    ///
531    /// # Safety
532    ///
533    /// If `mem::size_of::<T>() != 0`, then the safety rules are directly derived
534    /// from the safety rules for [`<*mut T>::sub`] method of `*mut T` and safety
535    /// rules of [`NonNull::new_unchecked`] function.
536    ///
537    /// Thus, in order to uphold the safety contracts for [`<*mut T>::sub`] method
538    /// and [`NonNull::new_unchecked`] function, as well as for the correct
539    /// logic of the work of this crate, the following rules are necessary and
540    /// sufficient:
541    ///
542    /// * `self` contained pointer must not be `dangling`;
543    ///
544    /// * `self.to_base_index() + ofset` must not be greater than `RawTableInner.bucket_mask`,
545    ///   i.e. `(self.to_base_index() + ofset) <= RawTableInner.bucket_mask` or, in other
546    ///   words, `self.to_base_index() + ofset + 1` must be no greater than the number returned
547    ///   by the function [`RawTable::buckets`] or [`RawTableInner::buckets`].
548    ///
549    /// If `mem::size_of::<T>() == 0`, then the only requirement is that the
550    /// `self.to_base_index() + ofset` must not be greater than `RawTableInner.bucket_mask`,
551    /// i.e. `(self.to_base_index() + ofset) <= RawTableInner.bucket_mask` or, in other words,
552    /// `self.to_base_index() + ofset + 1` must be no greater than the number returned by the
553    /// function [`RawTable::buckets`] or [`RawTableInner::buckets`].
554    ///
555    /// [`Bucket`]: crate::raw::Bucket
556    /// [`<*mut T>::sub`]: https://doc.rust-lang.org/core/primitive.pointer.html#method.sub-1
557    /// [`NonNull::new_unchecked`]: https://doc.rust-lang.org/stable/std/ptr/struct.NonNull.html#method.new_unchecked
558    /// [`RawTable::buckets`]: crate::raw::RawTable::buckets
559    /// [`RawTableInner::buckets`]: RawTableInner::buckets
560    #[inline]
561    unsafe fn next_n(&self, offset: usize) -> Self {
562        let ptr = if T::IS_ZERO_SIZED {
563            // invalid pointer is good enough for ZST
564            invalid_mut(self.ptr.as_ptr() as usize + offset)
565        } else {
566            self.ptr.as_ptr().sub(offset)
567        };
568        Self {
569            ptr: NonNull::new_unchecked(ptr),
570        }
571    }
572
573    /// Executes the destructor (if any) of the pointed-to `data`.
574    ///
575    /// # Safety
576    ///
577    /// See [`ptr::drop_in_place`] for safety concerns.
578    ///
579    /// You should use [`RawTable::erase`] instead of this function,
580    /// or be careful with calling this function directly, because for
581    /// properly dropping the data we need also clear `data` control bytes.
582    /// If we drop data, but do not erase `data control byte` it leads to
583    /// double drop when [`RawTable`] goes out of scope.
584    ///
585    /// [`ptr::drop_in_place`]: https://doc.rust-lang.org/core/ptr/fn.drop_in_place.html
586    /// [`RawTable`]: crate::raw::RawTable
587    /// [`RawTable::erase`]: crate::raw::RawTable::erase
588    #[cfg_attr(feature = "inline-more", inline)]
589    pub(crate) unsafe fn drop(&self) {
590        self.as_ptr().drop_in_place();
591    }
592
593    /// Reads the `value` from `self` without moving it. This leaves the
594    /// memory in `self` unchanged.
595    ///
596    /// # Safety
597    ///
598    /// See [`ptr::read`] for safety concerns.
599    ///
600    /// You should use [`RawTable::remove`] instead of this function,
601    /// or be careful with calling this function directly, because compiler
602    /// calls its destructor when readed `value` goes out of scope. It
603    /// can cause double dropping when [`RawTable`] goes out of scope,
604    /// because of not erased `data control byte`.
605    ///
606    /// [`ptr::read`]: https://doc.rust-lang.org/core/ptr/fn.read.html
607    /// [`RawTable`]: crate::raw::RawTable
608    /// [`RawTable::remove`]: crate::raw::RawTable::remove
609    #[inline]
610    pub(crate) unsafe fn read(&self) -> T {
611        self.as_ptr().read()
612    }
613
614    /// Overwrites a memory location with the given `value` without reading
615    /// or dropping the old value (like [`ptr::write`] function).
616    ///
617    /// # Safety
618    ///
619    /// See [`ptr::write`] for safety concerns.
620    ///
621    /// # Note
622    ///
623    /// [`Hash`] and [`Eq`] on the new `T` value and its borrowed form *must* match
624    /// those for the old `T` value, as the map will not re-evaluate where the new
625    /// value should go, meaning the value may become "lost" if their location
626    /// does not reflect their state.
627    ///
628    /// [`ptr::write`]: https://doc.rust-lang.org/core/ptr/fn.write.html
629    /// [`Hash`]: https://doc.rust-lang.org/core/hash/trait.Hash.html
630    /// [`Eq`]: https://doc.rust-lang.org/core/cmp/trait.Eq.html
631    #[inline]
632    pub(crate) unsafe fn write(&self, val: T) {
633        self.as_ptr().write(val);
634    }
635
636    /// Returns a shared immutable reference to the `value`.
637    ///
638    /// # Safety
639    ///
640    /// See [`NonNull::as_ref`] for safety concerns.
641    ///
642    /// [`NonNull::as_ref`]: https://doc.rust-lang.org/core/ptr/struct.NonNull.html#method.as_ref
643    ///
644    /// # Examples
645    ///
646    /// ```
647    /// # #[cfg(feature = "raw")]
648    /// # fn test() {
649    /// use core::hash::{BuildHasher, Hash};
650    /// use hashbrown::raw::{Bucket, RawTable};
651    ///
652    /// type NewHashBuilder = core::hash::BuildHasherDefault<ahash::AHasher>;
653    ///
654    /// fn make_hash<K: Hash + ?Sized, S: BuildHasher>(hash_builder: &S, key: &K) -> u64 {
655    ///     use core::hash::Hasher;
656    ///     let mut state = hash_builder.build_hasher();
657    ///     key.hash(&mut state);
658    ///     state.finish()
659    /// }
660    ///
661    /// let hash_builder = NewHashBuilder::default();
662    /// let mut table = RawTable::new();
663    ///
664    /// let value: (&str, String) = ("A pony", "is a small horse".to_owned());
665    /// let hash = make_hash(&hash_builder, &value.0);
666    ///
667    /// table.insert(hash, value.clone(), |val| make_hash(&hash_builder, &val.0));
668    ///
669    /// let bucket: Bucket<(&str, String)> = table.find(hash, |(k, _)| k == &value.0).unwrap();
670    ///
671    /// assert_eq!(
672    ///     unsafe { bucket.as_ref() },
673    ///     &("A pony", "is a small horse".to_owned())
674    /// );
675    /// # }
676    /// # fn main() {
677    /// #     #[cfg(feature = "raw")]
678    /// #     test()
679    /// # }
680    /// ```
681    #[inline]
682    pub unsafe fn as_ref<'a>(&self) -> &'a T {
683        &*self.as_ptr()
684    }
685
686    /// Returns a unique mutable reference to the `value`.
687    ///
688    /// # Safety
689    ///
690    /// See [`NonNull::as_mut`] for safety concerns.
691    ///
692    /// # Note
693    ///
694    /// [`Hash`] and [`Eq`] on the new `T` value and its borrowed form *must* match
695    /// those for the old `T` value, as the map will not re-evaluate where the new
696    /// value should go, meaning the value may become "lost" if their location
697    /// does not reflect their state.
698    ///
699    /// [`NonNull::as_mut`]: https://doc.rust-lang.org/core/ptr/struct.NonNull.html#method.as_mut
700    /// [`Hash`]: https://doc.rust-lang.org/core/hash/trait.Hash.html
701    /// [`Eq`]: https://doc.rust-lang.org/core/cmp/trait.Eq.html
702    ///
703    /// # Examples
704    ///
705    /// ```
706    /// # #[cfg(feature = "raw")]
707    /// # fn test() {
708    /// use core::hash::{BuildHasher, Hash};
709    /// use hashbrown::raw::{Bucket, RawTable};
710    ///
711    /// type NewHashBuilder = core::hash::BuildHasherDefault<ahash::AHasher>;
712    ///
713    /// fn make_hash<K: Hash + ?Sized, S: BuildHasher>(hash_builder: &S, key: &K) -> u64 {
714    ///     use core::hash::Hasher;
715    ///     let mut state = hash_builder.build_hasher();
716    ///     key.hash(&mut state);
717    ///     state.finish()
718    /// }
719    ///
720    /// let hash_builder = NewHashBuilder::default();
721    /// let mut table = RawTable::new();
722    ///
723    /// let value: (&str, String) = ("A pony", "is a small horse".to_owned());
724    /// let hash = make_hash(&hash_builder, &value.0);
725    ///
726    /// table.insert(hash, value.clone(), |val| make_hash(&hash_builder, &val.0));
727    ///
728    /// let bucket: Bucket<(&str, String)> = table.find(hash, |(k, _)| k == &value.0).unwrap();
729    ///
730    /// unsafe {
731    ///     bucket
732    ///         .as_mut()
733    ///         .1
734    ///         .push_str(" less than 147 cm at the withers")
735    /// };
736    /// assert_eq!(
737    ///     unsafe { bucket.as_ref() },
738    ///     &(
739    ///         "A pony",
740    ///         "is a small horse less than 147 cm at the withers".to_owned()
741    ///     )
742    /// );
743    /// # }
744    /// # fn main() {
745    /// #     #[cfg(feature = "raw")]
746    /// #     test()
747    /// # }
748    /// ```
749    #[inline]
750    pub unsafe fn as_mut<'a>(&self) -> &'a mut T {
751        &mut *self.as_ptr()
752    }
753
754    /// Copies `size_of<T>` bytes from `other` to `self`. The source
755    /// and destination may *not* overlap.
756    ///
757    /// # Safety
758    ///
759    /// See [`ptr::copy_nonoverlapping`] for safety concerns.
760    ///
761    /// Like [`read`], `copy_nonoverlapping` creates a bitwise copy of `T`, regardless of
762    /// whether `T` is [`Copy`]. If `T` is not [`Copy`], using *both* the values
763    /// in the region beginning at `*self` and the region beginning at `*other` can
764    /// [violate memory safety].
765    ///
766    /// # Note
767    ///
768    /// [`Hash`] and [`Eq`] on the new `T` value and its borrowed form *must* match
769    /// those for the old `T` value, as the map will not re-evaluate where the new
770    /// value should go, meaning the value may become "lost" if their location
771    /// does not reflect their state.
772    ///
773    /// [`ptr::copy_nonoverlapping`]: https://doc.rust-lang.org/core/ptr/fn.copy_nonoverlapping.html
774    /// [`read`]: https://doc.rust-lang.org/core/ptr/fn.read.html
775    /// [violate memory safety]: https://doc.rust-lang.org/std/ptr/fn.read.html#ownership-of-the-returned-value
776    /// [`Hash`]: https://doc.rust-lang.org/core/hash/trait.Hash.html
777    /// [`Eq`]: https://doc.rust-lang.org/core/cmp/trait.Eq.html
778    #[cfg(feature = "raw")]
779    #[inline]
780    pub unsafe fn copy_from_nonoverlapping(&self, other: &Self) {
781        self.as_ptr().copy_from_nonoverlapping(other.as_ptr(), 1);
782    }
783}
784
785/// A raw hash table with an unsafe API.
786pub struct RawTable<T, A: Allocator = Global> {
787    table: RawTableInner,
788    alloc: A,
789    // Tell dropck that we own instances of T.
790    marker: PhantomData<T>,
791}
792
793/// Non-generic part of `RawTable` which allows functions to be instantiated only once regardless
794/// of how many different key-value types are used.
795struct RawTableInner {
796    // Mask to get an index from a hash value. The value is one less than the
797    // number of buckets in the table.
798    bucket_mask: usize,
799
800    // [Padding], T1, T2, ..., Tlast, C1, C2, ...
801    //                                ^ points here
802    ctrl: NonNull<u8>,
803
804    // Number of elements that can be inserted before we need to grow the table
805    growth_left: usize,
806
807    // Number of elements in the table, only really used by len()
808    items: usize,
809}
810
811impl<T> RawTable<T, Global> {
812    /// Creates a new empty hash table without allocating any memory.
813    ///
814    /// In effect this returns a table with exactly 1 bucket. However we can
815    /// leave the data pointer dangling since that bucket is never written to
816    /// due to our load factor forcing us to always have at least 1 free bucket.
817    #[inline]
818    pub const fn new() -> Self {
819        Self {
820            table: RawTableInner::NEW,
821            alloc: Global,
822            marker: PhantomData,
823        }
824    }
825
826    /// Attempts to allocate a new hash table with at least enough capacity
827    /// for inserting the given number of elements without reallocating.
828    #[cfg(feature = "raw")]
829    pub fn try_with_capacity(capacity: usize) -> Result<Self, TryReserveError> {
830        Self::try_with_capacity_in(capacity, Global)
831    }
832
833    /// Allocates a new hash table with at least enough capacity for inserting
834    /// the given number of elements without reallocating.
835    pub fn with_capacity(capacity: usize) -> Self {
836        Self::with_capacity_in(capacity, Global)
837    }
838}
839
840impl<T, A: Allocator> RawTable<T, A> {
841    const TABLE_LAYOUT: TableLayout = TableLayout::new::<T>();
842
843    /// Creates a new empty hash table without allocating any memory, using the
844    /// given allocator.
845    ///
846    /// In effect this returns a table with exactly 1 bucket. However we can
847    /// leave the data pointer dangling since that bucket is never written to
848    /// due to our load factor forcing us to always have at least 1 free bucket.
849    #[inline]
850    pub const fn new_in(alloc: A) -> Self {
851        Self {
852            table: RawTableInner::NEW,
853            alloc,
854            marker: PhantomData,
855        }
856    }
857
858    /// Allocates a new hash table with the given number of buckets.
859    ///
860    /// The control bytes are left uninitialized.
861    #[cfg_attr(feature = "inline-more", inline)]
862    unsafe fn new_uninitialized(
863        alloc: A,
864        buckets: usize,
865        fallibility: Fallibility,
866    ) -> Result<Self, TryReserveError> {
867        debug_assert!(buckets.is_power_of_two());
868
869        Ok(Self {
870            table: RawTableInner::new_uninitialized(
871                &alloc,
872                Self::TABLE_LAYOUT,
873                buckets,
874                fallibility,
875            )?,
876            alloc,
877            marker: PhantomData,
878        })
879    }
880
881    /// Attempts to allocate a new hash table using the given allocator, with at least enough
882    /// capacity for inserting the given number of elements without reallocating.
883    #[cfg(feature = "raw")]
884    pub fn try_with_capacity_in(capacity: usize, alloc: A) -> Result<Self, TryReserveError> {
885        Ok(Self {
886            table: RawTableInner::fallible_with_capacity(
887                &alloc,
888                Self::TABLE_LAYOUT,
889                capacity,
890                Fallibility::Fallible,
891            )?,
892            alloc,
893            marker: PhantomData,
894        })
895    }
896
897    /// Allocates a new hash table using the given allocator, with at least enough capacity for
898    /// inserting the given number of elements without reallocating.
899    pub fn with_capacity_in(capacity: usize, alloc: A) -> Self {
900        Self {
901            table: RawTableInner::with_capacity(&alloc, Self::TABLE_LAYOUT, capacity),
902            alloc,
903            marker: PhantomData,
904        }
905    }
906
907    /// Returns a reference to the underlying allocator.
908    #[inline]
909    pub fn allocator(&self) -> &A {
910        &self.alloc
911    }
912
913    /// Returns pointer to one past last `data` element in the table as viewed from
914    /// the start point of the allocation.
915    ///
916    /// The caller must ensure that the `RawTable` outlives the returned [`NonNull<T>`],
917    /// otherwise using it may result in [`undefined behavior`].
918    ///
919    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
920    #[inline]
921    pub fn data_end(&self) -> NonNull<T> {
922        //                        `self.table.ctrl.cast()` returns pointer that
923        //                        points here (to the end of `T0`)
924        //                          ∨
925        // [Pad], T_n, ..., T1, T0, |CT0, CT1, ..., CT_n|, CTa_0, CTa_1, ..., CTa_m
926        //                           \________  ________/
927        //                                    \/
928        //       `n = buckets - 1`, i.e. `RawTable::buckets() - 1`
929        //
930        // where: T0...T_n  - our stored data;
931        //        CT0...CT_n - control bytes or metadata for `data`.
932        //        CTa_0...CTa_m - additional control bytes, where `m = Group::WIDTH - 1` (so that the search
933        //                        with loading `Group` bytes from the heap works properly, even if the result
934        //                        of `h1(hash) & self.bucket_mask` is equal to `self.bucket_mask`). See also
935        //                        `RawTableInner::set_ctrl` function.
936        //
937        // P.S. `h1(hash) & self.bucket_mask` is the same as `hash as usize % self.buckets()` because the number
938        // of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
939        self.table.ctrl.cast()
940    }
941
942    /// Returns pointer to start of data table.
943    #[inline]
944    #[cfg(any(feature = "raw", feature = "nightly"))]
945    pub unsafe fn data_start(&self) -> NonNull<T> {
946        NonNull::new_unchecked(self.data_end().as_ptr().wrapping_sub(self.buckets()))
947    }
948
949    /// Return the information about memory allocated by the table.
950    ///
951    /// `RawTable` allocates single memory block to store both data and metadata.
952    /// This function returns allocation size and alignment and the beginning of the area.
953    /// These are the arguments which will be passed to `dealloc` when the table is dropped.
954    ///
955    /// This function might be useful for memory profiling.
956    #[inline]
957    #[cfg(feature = "raw")]
958    pub fn allocation_info(&self) -> (NonNull<u8>, Layout) {
959        // SAFETY: We use the same `table_layout` that was used to allocate
960        // this table.
961        unsafe { self.table.allocation_info_or_zero(Self::TABLE_LAYOUT) }
962    }
963
964    /// Returns the index of a bucket from a `Bucket`.
965    #[inline]
966    pub unsafe fn bucket_index(&self, bucket: &Bucket<T>) -> usize {
967        bucket.to_base_index(self.data_end())
968    }
969
970    /// Returns a pointer to an element in the table.
971    ///
972    /// The caller must ensure that the `RawTable` outlives the returned [`Bucket<T>`],
973    /// otherwise using it may result in [`undefined behavior`].
974    ///
975    /// # Safety
976    ///
977    /// If `mem::size_of::<T>() != 0`, then the caller of this function must observe the
978    /// following safety rules:
979    ///
980    /// * The table must already be allocated;
981    ///
982    /// * The `index` must not be greater than the number returned by the [`RawTable::buckets`]
983    ///   function, i.e. `(index + 1) <= self.buckets()`.
984    ///
985    /// It is safe to call this function with index of zero (`index == 0`) on a table that has
986    /// not been allocated, but using the returned [`Bucket`] results in [`undefined behavior`].
987    ///
988    /// If `mem::size_of::<T>() == 0`, then the only requirement is that the `index` must
989    /// not be greater than the number returned by the [`RawTable::buckets`] function, i.e.
990    /// `(index + 1) <= self.buckets()`.
991    ///
992    /// [`RawTable::buckets`]: RawTable::buckets
993    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
994    #[inline]
995    pub unsafe fn bucket(&self, index: usize) -> Bucket<T> {
996        // If mem::size_of::<T>() != 0 then return a pointer to the `element` in the `data part` of the table
997        // (we start counting from "0", so that in the expression T[n], the "n" index actually one less than
998        // the "buckets" number of our `RawTable`, i.e. "n = RawTable::buckets() - 1"):
999        //
1000        //           `table.bucket(3).as_ptr()` returns a pointer that points here in the `data`
1001        //           part of the `RawTable`, i.e. to the start of T3 (see `Bucket::as_ptr`)
1002        //                  |
1003        //                  |               `base = self.data_end()` points here
1004        //                  |               (to the start of CT0 or to the end of T0)
1005        //                  v                 v
1006        // [Pad], T_n, ..., |T3|, T2, T1, T0, |CT0, CT1, CT2, CT3, ..., CT_n, CTa_0, CTa_1, ..., CTa_m
1007        //                     ^                                              \__________  __________/
1008        //        `table.bucket(3)` returns a pointer that points                        \/
1009        //         here in the `data` part of the `RawTable` (to              additional control bytes
1010        //         the end of T3)                                              `m = Group::WIDTH - 1`
1011        //
1012        // where: T0...T_n  - our stored data;
1013        //        CT0...CT_n - control bytes or metadata for `data`;
1014        //        CTa_0...CTa_m - additional control bytes (so that the search with loading `Group` bytes from
1015        //                        the heap works properly, even if the result of `h1(hash) & self.table.bucket_mask`
1016        //                        is equal to `self.table.bucket_mask`). See also `RawTableInner::set_ctrl` function.
1017        //
1018        // P.S. `h1(hash) & self.table.bucket_mask` is the same as `hash as usize % self.buckets()` because the number
1019        // of buckets is a power of two, and `self.table.bucket_mask = self.buckets() - 1`.
1020        debug_assert_ne!(self.table.bucket_mask, 0);
1021        debug_assert!(index < self.buckets());
1022        Bucket::from_base_index(self.data_end(), index)
1023    }
1024
1025    /// Erases an element from the table without dropping it.
1026    #[cfg_attr(feature = "inline-more", inline)]
1027    unsafe fn erase_no_drop(&mut self, item: &Bucket<T>) {
1028        let index = self.bucket_index(item);
1029        self.table.erase(index);
1030    }
1031
1032    /// Erases an element from the table, dropping it in place.
1033    #[cfg_attr(feature = "inline-more", inline)]
1034    #[allow(clippy::needless_pass_by_value)]
1035    pub unsafe fn erase(&mut self, item: Bucket<T>) {
1036        // Erase the element from the table first since drop might panic.
1037        self.erase_no_drop(&item);
1038        item.drop();
1039    }
1040
1041    /// Finds and erases an element from the table, dropping it in place.
1042    /// Returns true if an element was found.
1043    #[cfg(feature = "raw")]
1044    #[cfg_attr(feature = "inline-more", inline)]
1045    pub fn erase_entry(&mut self, hash: u64, eq: impl FnMut(&T) -> bool) -> bool {
1046        // Avoid `Option::map` because it bloats LLVM IR.
1047        if let Some(bucket) = self.find(hash, eq) {
1048            unsafe {
1049                self.erase(bucket);
1050            }
1051            true
1052        } else {
1053            false
1054        }
1055    }
1056
1057    /// Removes an element from the table, returning it.
1058    ///
1059    /// This also returns an `InsertSlot` pointing to the newly free bucket.
1060    #[cfg_attr(feature = "inline-more", inline)]
1061    #[allow(clippy::needless_pass_by_value)]
1062    pub unsafe fn remove(&mut self, item: Bucket<T>) -> (T, InsertSlot) {
1063        self.erase_no_drop(&item);
1064        (
1065            item.read(),
1066            InsertSlot {
1067                index: self.bucket_index(&item),
1068            },
1069        )
1070    }
1071
1072    /// Finds and removes an element from the table, returning it.
1073    #[cfg_attr(feature = "inline-more", inline)]
1074    pub fn remove_entry(&mut self, hash: u64, eq: impl FnMut(&T) -> bool) -> Option<T> {
1075        // Avoid `Option::map` because it bloats LLVM IR.
1076        match self.find(hash, eq) {
1077            Some(bucket) => Some(unsafe { self.remove(bucket).0 }),
1078            None => None,
1079        }
1080    }
1081
1082    /// Marks all table buckets as empty without dropping their contents.
1083    #[cfg_attr(feature = "inline-more", inline)]
1084    pub fn clear_no_drop(&mut self) {
1085        self.table.clear_no_drop();
1086    }
1087
1088    /// Removes all elements from the table without freeing the backing memory.
1089    #[cfg_attr(feature = "inline-more", inline)]
1090    pub fn clear(&mut self) {
1091        if self.is_empty() {
1092            // Special case empty table to avoid surprising O(capacity) time.
1093            return;
1094        }
1095        // Ensure that the table is reset even if one of the drops panic
1096        let mut self_ = guard(self, |self_| self_.clear_no_drop());
1097        unsafe {
1098            // SAFETY: ScopeGuard sets to zero the `items` field of the table
1099            // even in case of panic during the dropping of the elements so
1100            // that there will be no double drop of the elements.
1101            self_.table.drop_elements::<T>();
1102        }
1103    }
1104
1105    /// Shrinks the table to fit `max(self.len(), min_size)` elements.
1106    #[cfg_attr(feature = "inline-more", inline)]
1107    pub fn shrink_to(&mut self, min_size: usize, hasher: impl Fn(&T) -> u64) {
1108        // Calculate the minimal number of elements that we need to reserve
1109        // space for.
1110        let min_size = usize::max(self.table.items, min_size);
1111        if min_size == 0 {
1112            let mut old_inner = mem::replace(&mut self.table, RawTableInner::NEW);
1113            unsafe {
1114                // SAFETY:
1115                // 1. We call the function only once;
1116                // 2. We know for sure that `alloc` and `table_layout` matches the [`Allocator`]
1117                //    and [`TableLayout`] that were used to allocate this table.
1118                // 3. If any elements' drop function panics, then there will only be a memory leak,
1119                //    because we have replaced the inner table with a new one.
1120                old_inner.drop_inner_table::<T, _>(&self.alloc, Self::TABLE_LAYOUT);
1121            }
1122            return;
1123        }
1124
1125        // Calculate the number of buckets that we need for this number of
1126        // elements. If the calculation overflows then the requested bucket
1127        // count must be larger than what we have right and nothing needs to be
1128        // done.
1129        let min_buckets = match capacity_to_buckets(min_size) {
1130            Some(buckets) => buckets,
1131            None => return,
1132        };
1133
1134        // If we have more buckets than we need, shrink the table.
1135        if min_buckets < self.buckets() {
1136            // Fast path if the table is empty
1137            if self.table.items == 0 {
1138                let new_inner =
1139                    RawTableInner::with_capacity(&self.alloc, Self::TABLE_LAYOUT, min_size);
1140                let mut old_inner = mem::replace(&mut self.table, new_inner);
1141                unsafe {
1142                    // SAFETY:
1143                    // 1. We call the function only once;
1144                    // 2. We know for sure that `alloc` and `table_layout` matches the [`Allocator`]
1145                    //    and [`TableLayout`] that were used to allocate this table.
1146                    // 3. If any elements' drop function panics, then there will only be a memory leak,
1147                    //    because we have replaced the inner table with a new one.
1148                    old_inner.drop_inner_table::<T, _>(&self.alloc, Self::TABLE_LAYOUT);
1149                }
1150            } else {
1151                // Avoid `Result::unwrap_or_else` because it bloats LLVM IR.
1152                unsafe {
1153                    // SAFETY:
1154                    // 1. We know for sure that `min_size >= self.table.items`.
1155                    // 2. The [`RawTableInner`] must already have properly initialized control bytes since
1156                    //    we will never expose RawTable::new_uninitialized in a public API.
1157                    if self
1158                        .resize(min_size, hasher, Fallibility::Infallible)
1159                        .is_err()
1160                    {
1161                        // SAFETY: The result of calling the `resize` function cannot be an error
1162                        // because `fallibility == Fallibility::Infallible.
1163                        hint::unreachable_unchecked()
1164                    }
1165                }
1166            }
1167        }
1168    }
1169
1170    /// Ensures that at least `additional` items can be inserted into the table
1171    /// without reallocation.
1172    #[cfg_attr(feature = "inline-more", inline)]
1173    pub fn reserve(&mut self, additional: usize, hasher: impl Fn(&T) -> u64) {
1174        if unlikely(additional > self.table.growth_left) {
1175            // Avoid `Result::unwrap_or_else` because it bloats LLVM IR.
1176            unsafe {
1177                // SAFETY: The [`RawTableInner`] must already have properly initialized control
1178                // bytes since we will never expose RawTable::new_uninitialized in a public API.
1179                if self
1180                    .reserve_rehash(additional, hasher, Fallibility::Infallible)
1181                    .is_err()
1182                {
1183                    // SAFETY: All allocation errors will be caught inside `RawTableInner::reserve_rehash`.
1184                    hint::unreachable_unchecked()
1185                }
1186            }
1187        }
1188    }
1189
1190    /// Tries to ensure that at least `additional` items can be inserted into
1191    /// the table without reallocation.
1192    #[cfg_attr(feature = "inline-more", inline)]
1193    pub fn try_reserve(
1194        &mut self,
1195        additional: usize,
1196        hasher: impl Fn(&T) -> u64,
1197    ) -> Result<(), TryReserveError> {
1198        if additional > self.table.growth_left {
1199            // SAFETY: The [`RawTableInner`] must already have properly initialized control
1200            // bytes since we will never expose RawTable::new_uninitialized in a public API.
1201            unsafe { self.reserve_rehash(additional, hasher, Fallibility::Fallible) }
1202        } else {
1203            Ok(())
1204        }
1205    }
1206
1207    /// Out-of-line slow path for `reserve` and `try_reserve`.
1208    ///
1209    /// # Safety
1210    ///
1211    /// The [`RawTableInner`] must have properly initialized control bytes,
1212    /// otherwise calling this function results in [`undefined behavior`]
1213    ///
1214    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
1215    #[cold]
1216    #[inline(never)]
1217    unsafe fn reserve_rehash(
1218        &mut self,
1219        additional: usize,
1220        hasher: impl Fn(&T) -> u64,
1221        fallibility: Fallibility,
1222    ) -> Result<(), TryReserveError> {
1223        unsafe {
1224            // SAFETY:
1225            // 1. We know for sure that `alloc` and `layout` matches the [`Allocator`] and
1226            //    [`TableLayout`] that were used to allocate this table.
1227            // 2. The `drop` function is the actual drop function of the elements stored in
1228            //    the table.
1229            // 3. The caller ensures that the control bytes of the `RawTableInner`
1230            //    are already initialized.
1231            self.table.reserve_rehash_inner(
1232                &self.alloc,
1233                additional,
1234                &|table, index| hasher(table.bucket::<T>(index).as_ref()),
1235                fallibility,
1236                Self::TABLE_LAYOUT,
1237                if T::NEEDS_DROP {
1238                    Some(mem::transmute(ptr::drop_in_place::<T> as unsafe fn(*mut T)))
1239                } else {
1240                    None
1241                },
1242            )
1243        }
1244    }
1245
1246    /// Allocates a new table of a different size and moves the contents of the
1247    /// current table into it.
1248    ///
1249    /// # Safety
1250    ///
1251    /// The [`RawTableInner`] must have properly initialized control bytes,
1252    /// otherwise calling this function results in [`undefined behavior`]
1253    ///
1254    /// The caller of this function must ensure that `capacity >= self.table.items`
1255    /// otherwise:
1256    ///
1257    /// * If `self.table.items != 0`, calling of this function with `capacity`
1258    ///   equal to 0 (`capacity == 0`) results in [`undefined behavior`].
1259    ///
1260    /// * If `capacity_to_buckets(capacity) < Group::WIDTH` and
1261    ///   `self.table.items > capacity_to_buckets(capacity)`
1262    ///   calling this function results in [`undefined behavior`].
1263    ///
1264    /// * If `capacity_to_buckets(capacity) >= Group::WIDTH` and
1265    ///   `self.table.items > capacity_to_buckets(capacity)`
1266    ///   calling this function are never return (will go into an
1267    ///   infinite loop).
1268    ///
1269    /// See [`RawTableInner::find_insert_slot`] for more information.
1270    ///
1271    /// [`RawTableInner::find_insert_slot`]: RawTableInner::find_insert_slot
1272    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
1273    unsafe fn resize(
1274        &mut self,
1275        capacity: usize,
1276        hasher: impl Fn(&T) -> u64,
1277        fallibility: Fallibility,
1278    ) -> Result<(), TryReserveError> {
1279        // SAFETY:
1280        // 1. The caller of this function guarantees that `capacity >= self.table.items`.
1281        // 2. We know for sure that `alloc` and `layout` matches the [`Allocator`] and
1282        //    [`TableLayout`] that were used to allocate this table.
1283        // 3. The caller ensures that the control bytes of the `RawTableInner`
1284        //    are already initialized.
1285        self.table.resize_inner(
1286            &self.alloc,
1287            capacity,
1288            &|table, index| hasher(table.bucket::<T>(index).as_ref()),
1289            fallibility,
1290            Self::TABLE_LAYOUT,
1291        )
1292    }
1293
1294    /// Inserts a new element into the table, and returns its raw bucket.
1295    ///
1296    /// This does not check if the given element already exists in the table.
1297    #[cfg_attr(feature = "inline-more", inline)]
1298    pub fn insert(&mut self, hash: u64, value: T, hasher: impl Fn(&T) -> u64) -> Bucket<T> {
1299        unsafe {
1300            // SAFETY:
1301            // 1. The [`RawTableInner`] must already have properly initialized control bytes since
1302            //    we will never expose `RawTable::new_uninitialized` in a public API.
1303            //
1304            // 2. We reserve additional space (if necessary) right after calling this function.
1305            let mut slot = self.table.find_insert_slot(hash);
1306
1307            // We can avoid growing the table once we have reached our load factor if we are replacing
1308            // a tombstone. This works since the number of EMPTY slots does not change in this case.
1309            //
1310            // SAFETY: The function is guaranteed to return [`InsertSlot`] that contains an index
1311            // in the range `0..=self.buckets()`.
1312            let old_ctrl = *self.table.ctrl(slot.index);
1313            if unlikely(self.table.growth_left == 0 && special_is_empty(old_ctrl)) {
1314                self.reserve(1, hasher);
1315                // SAFETY: We know for sure that `RawTableInner` has control bytes
1316                // initialized and that there is extra space in the table.
1317                slot = self.table.find_insert_slot(hash);
1318            }
1319
1320            self.insert_in_slot(hash, slot, value)
1321        }
1322    }
1323
1324    /// Attempts to insert a new element without growing the table and return its raw bucket.
1325    ///
1326    /// Returns an `Err` containing the given element if inserting it would require growing the
1327    /// table.
1328    ///
1329    /// This does not check if the given element already exists in the table.
1330    #[cfg(feature = "raw")]
1331    #[cfg_attr(feature = "inline-more", inline)]
1332    pub fn try_insert_no_grow(&mut self, hash: u64, value: T) -> Result<Bucket<T>, T> {
1333        unsafe {
1334            match self.table.prepare_insert_no_grow(hash) {
1335                Ok(index) => {
1336                    let bucket = self.bucket(index);
1337                    bucket.write(value);
1338                    Ok(bucket)
1339                }
1340                Err(()) => Err(value),
1341            }
1342        }
1343    }
1344
1345    /// Inserts a new element into the table, and returns a mutable reference to it.
1346    ///
1347    /// This does not check if the given element already exists in the table.
1348    #[cfg_attr(feature = "inline-more", inline)]
1349    pub fn insert_entry(&mut self, hash: u64, value: T, hasher: impl Fn(&T) -> u64) -> &mut T {
1350        unsafe { self.insert(hash, value, hasher).as_mut() }
1351    }
1352
1353    /// Inserts a new element into the table, without growing the table.
1354    ///
1355    /// There must be enough space in the table to insert the new element.
1356    ///
1357    /// This does not check if the given element already exists in the table.
1358    #[cfg_attr(feature = "inline-more", inline)]
1359    #[cfg(any(feature = "raw", feature = "rustc-internal-api"))]
1360    pub unsafe fn insert_no_grow(&mut self, hash: u64, value: T) -> Bucket<T> {
1361        let (index, old_ctrl) = self.table.prepare_insert_slot(hash);
1362        let bucket = self.table.bucket(index);
1363
1364        // If we are replacing a DELETED entry then we don't need to update
1365        // the load counter.
1366        self.table.growth_left -= special_is_empty(old_ctrl) as usize;
1367
1368        bucket.write(value);
1369        self.table.items += 1;
1370        bucket
1371    }
1372
1373    /// Temporary removes a bucket, applying the given function to the removed
1374    /// element and optionally put back the returned value in the same bucket.
1375    ///
1376    /// Returns `true` if the bucket still contains an element
1377    ///
1378    /// This does not check if the given bucket is actually occupied.
1379    #[cfg_attr(feature = "inline-more", inline)]
1380    pub unsafe fn replace_bucket_with<F>(&mut self, bucket: Bucket<T>, f: F) -> bool
1381    where
1382        F: FnOnce(T) -> Option<T>,
1383    {
1384        let index = self.bucket_index(&bucket);
1385        let old_ctrl = *self.table.ctrl(index);
1386        debug_assert!(self.is_bucket_full(index));
1387        let old_growth_left = self.table.growth_left;
1388        let item = self.remove(bucket).0;
1389        if let Some(new_item) = f(item) {
1390            self.table.growth_left = old_growth_left;
1391            self.table.set_ctrl(index, old_ctrl);
1392            self.table.items += 1;
1393            self.bucket(index).write(new_item);
1394            true
1395        } else {
1396            false
1397        }
1398    }
1399
1400    /// Searches for an element in the table. If the element is not found,
1401    /// returns `Err` with the position of a slot where an element with the
1402    /// same hash could be inserted.
1403    ///
1404    /// This function may resize the table if additional space is required for
1405    /// inserting an element.
1406    #[inline]
1407    pub fn find_or_find_insert_slot(
1408        &mut self,
1409        hash: u64,
1410        mut eq: impl FnMut(&T) -> bool,
1411        hasher: impl Fn(&T) -> u64,
1412    ) -> Result<Bucket<T>, InsertSlot> {
1413        self.reserve(1, hasher);
1414
1415        unsafe {
1416            // SAFETY:
1417            // 1. We know for sure that there is at least one empty `bucket` in the table.
1418            // 2. The [`RawTableInner`] must already have properly initialized control bytes since we will
1419            //    never expose `RawTable::new_uninitialized` in a public API.
1420            // 3. The `find_or_find_insert_slot_inner` function returns the `index` of only the full bucket,
1421            //    which is in the range `0..self.buckets()` (since there is at least one empty `bucket` in
1422            //    the table), so calling `self.bucket(index)` and `Bucket::as_ref` is safe.
1423            match self
1424                .table
1425                .find_or_find_insert_slot_inner(hash, &mut |index| eq(self.bucket(index).as_ref()))
1426            {
1427                // SAFETY: See explanation above.
1428                Ok(index) => Ok(self.bucket(index)),
1429                Err(slot) => Err(slot),
1430            }
1431        }
1432    }
1433
1434    /// Inserts a new element into the table in the given slot, and returns its
1435    /// raw bucket.
1436    ///
1437    /// # Safety
1438    ///
1439    /// `slot` must point to a slot previously returned by
1440    /// `find_or_find_insert_slot`, and no mutation of the table must have
1441    /// occurred since that call.
1442    #[inline]
1443    pub unsafe fn insert_in_slot(&mut self, hash: u64, slot: InsertSlot, value: T) -> Bucket<T> {
1444        let old_ctrl = *self.table.ctrl(slot.index);
1445        self.table.record_item_insert_at(slot.index, old_ctrl, hash);
1446
1447        let bucket = self.bucket(slot.index);
1448        bucket.write(value);
1449        bucket
1450    }
1451
1452    /// Searches for an element in the table.
1453    #[inline]
1454    pub fn find(&self, hash: u64, mut eq: impl FnMut(&T) -> bool) -> Option<Bucket<T>> {
1455        unsafe {
1456            // SAFETY:
1457            // 1. The [`RawTableInner`] must already have properly initialized control bytes since we
1458            //    will never expose `RawTable::new_uninitialized` in a public API.
1459            // 1. The `find_inner` function returns the `index` of only the full bucket, which is in
1460            //    the range `0..self.buckets()`, so calling `self.bucket(index)` and `Bucket::as_ref`
1461            //    is safe.
1462            let result = self
1463                .table
1464                .find_inner(hash, &mut |index| eq(self.bucket(index).as_ref()));
1465
1466            // Avoid `Option::map` because it bloats LLVM IR.
1467            match result {
1468                // SAFETY: See explanation above.
1469                Some(index) => Some(self.bucket(index)),
1470                None => None,
1471            }
1472        }
1473    }
1474
1475    /// Gets a reference to an element in the table.
1476    #[inline]
1477    pub fn get(&self, hash: u64, eq: impl FnMut(&T) -> bool) -> Option<&T> {
1478        // Avoid `Option::map` because it bloats LLVM IR.
1479        match self.find(hash, eq) {
1480            Some(bucket) => Some(unsafe { bucket.as_ref() }),
1481            None => None,
1482        }
1483    }
1484
1485    /// Gets a mutable reference to an element in the table.
1486    #[inline]
1487    pub fn get_mut(&mut self, hash: u64, eq: impl FnMut(&T) -> bool) -> Option<&mut T> {
1488        // Avoid `Option::map` because it bloats LLVM IR.
1489        match self.find(hash, eq) {
1490            Some(bucket) => Some(unsafe { bucket.as_mut() }),
1491            None => None,
1492        }
1493    }
1494
1495    /// Attempts to get mutable references to `N` entries in the table at once.
1496    ///
1497    /// Returns an array of length `N` with the results of each query.
1498    ///
1499    /// At most one mutable reference will be returned to any entry. `None` will be returned if any
1500    /// of the hashes are duplicates. `None` will be returned if the hash is not found.
1501    ///
1502    /// The `eq` argument should be a closure such that `eq(i, k)` returns true if `k` is equal to
1503    /// the `i`th key to be looked up.
1504    pub fn get_many_mut<const N: usize>(
1505        &mut self,
1506        hashes: [u64; N],
1507        eq: impl FnMut(usize, &T) -> bool,
1508    ) -> Option<[&'_ mut T; N]> {
1509        unsafe {
1510            let ptrs = self.get_many_mut_pointers(hashes, eq)?;
1511
1512            for (i, &cur) in ptrs.iter().enumerate() {
1513                if ptrs[..i].iter().any(|&prev| ptr::eq::<T>(prev, cur)) {
1514                    return None;
1515                }
1516            }
1517            // All bucket are distinct from all previous buckets so we're clear to return the result
1518            // of the lookup.
1519
1520            // TODO use `MaybeUninit::array_assume_init` here instead once that's stable.
1521            Some(mem::transmute_copy(&ptrs))
1522        }
1523    }
1524
1525    pub unsafe fn get_many_unchecked_mut<const N: usize>(
1526        &mut self,
1527        hashes: [u64; N],
1528        eq: impl FnMut(usize, &T) -> bool,
1529    ) -> Option<[&'_ mut T; N]> {
1530        let ptrs = self.get_many_mut_pointers(hashes, eq)?;
1531        Some(mem::transmute_copy(&ptrs))
1532    }
1533
1534    unsafe fn get_many_mut_pointers<const N: usize>(
1535        &mut self,
1536        hashes: [u64; N],
1537        mut eq: impl FnMut(usize, &T) -> bool,
1538    ) -> Option<[*mut T; N]> {
1539        // TODO use `MaybeUninit::uninit_array` here instead once that's stable.
1540        let mut outs: MaybeUninit<[*mut T; N]> = MaybeUninit::uninit();
1541        let outs_ptr = outs.as_mut_ptr();
1542
1543        for (i, &hash) in hashes.iter().enumerate() {
1544            let cur = self.find(hash, |k| eq(i, k))?;
1545            *(*outs_ptr).get_unchecked_mut(i) = cur.as_mut();
1546        }
1547
1548        // TODO use `MaybeUninit::array_assume_init` here instead once that's stable.
1549        Some(outs.assume_init())
1550    }
1551
1552    /// Returns the number of elements the map can hold without reallocating.
1553    ///
1554    /// This number is a lower bound; the table might be able to hold
1555    /// more, but is guaranteed to be able to hold at least this many.
1556    #[inline]
1557    pub fn capacity(&self) -> usize {
1558        self.table.items + self.table.growth_left
1559    }
1560
1561    /// Returns the number of elements in the table.
1562    #[inline]
1563    pub fn len(&self) -> usize {
1564        self.table.items
1565    }
1566
1567    /// Returns `true` if the table contains no elements.
1568    #[inline]
1569    pub fn is_empty(&self) -> bool {
1570        self.len() == 0
1571    }
1572
1573    /// Returns the number of buckets in the table.
1574    #[inline]
1575    pub fn buckets(&self) -> usize {
1576        self.table.bucket_mask + 1
1577    }
1578
1579    /// Checks whether the bucket at `index` is full.
1580    ///
1581    /// # Safety
1582    ///
1583    /// The caller must ensure `index` is less than the number of buckets.
1584    #[inline]
1585    pub unsafe fn is_bucket_full(&self, index: usize) -> bool {
1586        self.table.is_bucket_full(index)
1587    }
1588
1589    /// Returns an iterator over every element in the table. It is up to
1590    /// the caller to ensure that the `RawTable` outlives the `RawIter`.
1591    /// Because we cannot make the `next` method unsafe on the `RawIter`
1592    /// struct, we have to make the `iter` method unsafe.
1593    #[inline]
1594    pub unsafe fn iter(&self) -> RawIter<T> {
1595        // SAFETY:
1596        // 1. The caller must uphold the safety contract for `iter` method.
1597        // 2. The [`RawTableInner`] must already have properly initialized control bytes since
1598        //    we will never expose RawTable::new_uninitialized in a public API.
1599        self.table.iter()
1600    }
1601
1602    /// Returns an iterator over occupied buckets that could match a given hash.
1603    ///
1604    /// `RawTable` only stores 7 bits of the hash value, so this iterator may
1605    /// return items that have a hash value different than the one provided. You
1606    /// should always validate the returned values before using them.
1607    ///
1608    /// It is up to the caller to ensure that the `RawTable` outlives the
1609    /// `RawIterHash`. Because we cannot make the `next` method unsafe on the
1610    /// `RawIterHash` struct, we have to make the `iter_hash` method unsafe.
1611    #[cfg_attr(feature = "inline-more", inline)]
1612    #[cfg(feature = "raw")]
1613    pub unsafe fn iter_hash(&self, hash: u64) -> RawIterHash<T> {
1614        RawIterHash::new(self, hash)
1615    }
1616
1617    /// Returns an iterator which removes all elements from the table without
1618    /// freeing the memory.
1619    #[cfg_attr(feature = "inline-more", inline)]
1620    pub fn drain(&mut self) -> RawDrain<'_, T, A> {
1621        unsafe {
1622            let iter = self.iter();
1623            self.drain_iter_from(iter)
1624        }
1625    }
1626
1627    /// Returns an iterator which removes all elements from the table without
1628    /// freeing the memory.
1629    ///
1630    /// Iteration starts at the provided iterator's current location.
1631    ///
1632    /// It is up to the caller to ensure that the iterator is valid for this
1633    /// `RawTable` and covers all items that remain in the table.
1634    #[cfg_attr(feature = "inline-more", inline)]
1635    pub unsafe fn drain_iter_from(&mut self, iter: RawIter<T>) -> RawDrain<'_, T, A> {
1636        debug_assert_eq!(iter.len(), self.len());
1637        RawDrain {
1638            iter,
1639            table: mem::replace(&mut self.table, RawTableInner::NEW),
1640            orig_table: NonNull::from(&mut self.table),
1641            marker: PhantomData,
1642        }
1643    }
1644
1645    /// Returns an iterator which consumes all elements from the table.
1646    ///
1647    /// Iteration starts at the provided iterator's current location.
1648    ///
1649    /// It is up to the caller to ensure that the iterator is valid for this
1650    /// `RawTable` and covers all items that remain in the table.
1651    pub unsafe fn into_iter_from(self, iter: RawIter<T>) -> RawIntoIter<T, A> {
1652        debug_assert_eq!(iter.len(), self.len());
1653
1654        let allocation = self.into_allocation();
1655        RawIntoIter {
1656            iter,
1657            allocation,
1658            marker: PhantomData,
1659        }
1660    }
1661
1662    /// Converts the table into a raw allocation. The contents of the table
1663    /// should be dropped using a `RawIter` before freeing the allocation.
1664    #[cfg_attr(feature = "inline-more", inline)]
1665    pub(crate) fn into_allocation(self) -> Option<(NonNull<u8>, Layout, A)> {
1666        let alloc = if self.table.is_empty_singleton() {
1667            None
1668        } else {
1669            // Avoid `Option::unwrap_or_else` because it bloats LLVM IR.
1670            let (layout, ctrl_offset) =
1671                match Self::TABLE_LAYOUT.calculate_layout_for(self.table.buckets()) {
1672                    Some(lco) => lco,
1673                    None => unsafe { hint::unreachable_unchecked() },
1674                };
1675            Some((
1676                unsafe { NonNull::new_unchecked(self.table.ctrl.as_ptr().sub(ctrl_offset)) },
1677                layout,
1678                unsafe { ptr::read(&self.alloc) },
1679            ))
1680        };
1681        mem::forget(self);
1682        alloc
1683    }
1684}
1685
1686unsafe impl<T, A: Allocator> Send for RawTable<T, A>
1687where
1688    T: Send,
1689    A: Send,
1690{
1691}
1692unsafe impl<T, A: Allocator> Sync for RawTable<T, A>
1693where
1694    T: Sync,
1695    A: Sync,
1696{
1697}
1698
1699impl RawTableInner {
1700    const NEW: Self = RawTableInner::new();
1701
1702    /// Creates a new empty hash table without allocating any memory.
1703    ///
1704    /// In effect this returns a table with exactly 1 bucket. However we can
1705    /// leave the data pointer dangling since that bucket is never accessed
1706    /// due to our load factor forcing us to always have at least 1 free bucket.
1707    #[inline]
1708    const fn new() -> Self {
1709        Self {
1710            // Be careful to cast the entire slice to a raw pointer.
1711            ctrl: unsafe { NonNull::new_unchecked(Group::static_empty() as *const _ as *mut u8) },
1712            bucket_mask: 0,
1713            items: 0,
1714            growth_left: 0,
1715        }
1716    }
1717}
1718
1719impl RawTableInner {
1720    /// Allocates a new [`RawTableInner`] with the given number of buckets.
1721    /// The control bytes and buckets are left uninitialized.
1722    ///
1723    /// # Safety
1724    ///
1725    /// The caller of this function must ensure that the `buckets` is power of two
1726    /// and also initialize all control bytes of the length `self.bucket_mask + 1 +
1727    /// Group::WIDTH` with the [`EMPTY`] bytes.
1728    ///
1729    /// See also [`Allocator`] API for other safety concerns.
1730    ///
1731    /// [`Allocator`]: https://doc.rust-lang.org/alloc/alloc/trait.Allocator.html
1732    #[cfg_attr(feature = "inline-more", inline)]
1733    unsafe fn new_uninitialized<A>(
1734        alloc: &A,
1735        table_layout: TableLayout,
1736        buckets: usize,
1737        fallibility: Fallibility,
1738    ) -> Result<Self, TryReserveError>
1739    where
1740        A: Allocator,
1741    {
1742        debug_assert!(buckets.is_power_of_two());
1743
1744        // Avoid `Option::ok_or_else` because it bloats LLVM IR.
1745        let (layout, ctrl_offset) = match table_layout.calculate_layout_for(buckets) {
1746            Some(lco) => lco,
1747            None => return Err(fallibility.capacity_overflow()),
1748        };
1749
1750        let ptr: NonNull<u8> = match do_alloc(alloc, layout) {
1751            Ok(block) => block.cast(),
1752            Err(_) => return Err(fallibility.alloc_err(layout)),
1753        };
1754
1755        // SAFETY: null pointer will be caught in above check
1756        let ctrl = NonNull::new_unchecked(ptr.as_ptr().add(ctrl_offset));
1757        Ok(Self {
1758            ctrl,
1759            bucket_mask: buckets - 1,
1760            items: 0,
1761            growth_left: bucket_mask_to_capacity(buckets - 1),
1762        })
1763    }
1764
1765    /// Attempts to allocate a new [`RawTableInner`] with at least enough
1766    /// capacity for inserting the given number of elements without reallocating.
1767    ///
1768    /// All the control bytes are initialized with the [`EMPTY`] bytes.
1769    #[inline]
1770    fn fallible_with_capacity<A>(
1771        alloc: &A,
1772        table_layout: TableLayout,
1773        capacity: usize,
1774        fallibility: Fallibility,
1775    ) -> Result<Self, TryReserveError>
1776    where
1777        A: Allocator,
1778    {
1779        if capacity == 0 {
1780            Ok(Self::NEW)
1781        } else {
1782            // SAFETY: We checked that we could successfully allocate the new table, and then
1783            // initialized all control bytes with the constant `EMPTY` byte.
1784            unsafe {
1785                let buckets =
1786                    capacity_to_buckets(capacity).ok_or_else(|| fallibility.capacity_overflow())?;
1787
1788                let result = Self::new_uninitialized(alloc, table_layout, buckets, fallibility)?;
1789                // SAFETY: We checked that the table is allocated and therefore the table already has
1790                // `self.bucket_mask + 1 + Group::WIDTH` number of control bytes (see TableLayout::calculate_layout_for)
1791                // so writing `self.num_ctrl_bytes() == bucket_mask + 1 + Group::WIDTH` bytes is safe.
1792                result.ctrl(0).write_bytes(EMPTY, result.num_ctrl_bytes());
1793
1794                Ok(result)
1795            }
1796        }
1797    }
1798
1799    /// Allocates a new [`RawTableInner`] with at least enough capacity for inserting
1800    /// the given number of elements without reallocating.
1801    ///
1802    /// Panics if the new capacity exceeds [`isize::MAX`] bytes and [`abort`] the program
1803    /// in case of allocation error. Use [`fallible_with_capacity`] instead if you want to
1804    /// handle memory allocation failure.
1805    ///
1806    /// All the control bytes are initialized with the [`EMPTY`] bytes.
1807    ///
1808    /// [`fallible_with_capacity`]: RawTableInner::fallible_with_capacity
1809    /// [`abort`]: https://doc.rust-lang.org/alloc/alloc/fn.handle_alloc_error.html
1810    fn with_capacity<A>(alloc: &A, table_layout: TableLayout, capacity: usize) -> Self
1811    where
1812        A: Allocator,
1813    {
1814        // Avoid `Result::unwrap_or_else` because it bloats LLVM IR.
1815        match Self::fallible_with_capacity(alloc, table_layout, capacity, Fallibility::Infallible) {
1816            Ok(table_inner) => table_inner,
1817            // SAFETY: All allocation errors will be caught inside `RawTableInner::new_uninitialized`.
1818            Err(_) => unsafe { hint::unreachable_unchecked() },
1819        }
1820    }
1821
1822    /// Fixes up an insertion slot returned by the [`RawTableInner::find_insert_slot_in_group`] method.
1823    ///
1824    /// In tables smaller than the group width (`self.buckets() < Group::WIDTH`), trailing control
1825    /// bytes outside the range of the table are filled with [`EMPTY`] entries. These will unfortunately
1826    /// trigger a match of [`RawTableInner::find_insert_slot_in_group`] function. This is because
1827    /// the `Some(bit)` returned by `group.match_empty_or_deleted().lowest_set_bit()` after masking
1828    /// (`(probe_seq.pos + bit) & self.bucket_mask`) may point to a full bucket that is already occupied.
1829    /// We detect this situation here and perform a second scan starting at the beginning of the table.
1830    /// This second scan is guaranteed to find an empty slot (due to the load factor) before hitting the
1831    /// trailing control bytes (containing [`EMPTY`] bytes).
1832    ///
1833    /// If this function is called correctly, it is guaranteed to return [`InsertSlot`] with an
1834    /// index of an empty or deleted bucket in the range `0..self.buckets()` (see `Warning` and
1835    /// `Safety`).
1836    ///
1837    /// # Warning
1838    ///
1839    /// The table must have at least 1 empty or deleted `bucket`, otherwise if the table is less than
1840    /// the group width (`self.buckets() < Group::WIDTH`) this function returns an index outside of the
1841    /// table indices range `0..self.buckets()` (`0..=self.bucket_mask`). Attempt to write data at that
1842    /// index will cause immediate [`undefined behavior`].
1843    ///
1844    /// # Safety
1845    ///
1846    /// The safety rules are directly derived from the safety rules for [`RawTableInner::ctrl`] method.
1847    /// Thus, in order to uphold those safety contracts, as well as for the correct logic of the work
1848    /// of this crate, the following rules are necessary and sufficient:
1849    ///
1850    /// * The [`RawTableInner`] must have properly initialized control bytes otherwise calling this
1851    ///   function results in [`undefined behavior`].
1852    ///
1853    /// * This function must only be used on insertion slots found by [`RawTableInner::find_insert_slot_in_group`]
1854    ///   (after the `find_insert_slot_in_group` function, but before insertion into the table).
1855    ///
1856    /// * The `index` must not be greater than the `self.bucket_mask`, i.e. `(index + 1) <= self.buckets()`
1857    ///   (this one is provided by the [`RawTableInner::find_insert_slot_in_group`] function).
1858    ///
1859    /// Calling this function with an index not provided by [`RawTableInner::find_insert_slot_in_group`]
1860    /// may result in [`undefined behavior`] even if the index satisfies the safety rules of the
1861    /// [`RawTableInner::ctrl`] function (`index < self.bucket_mask + 1 + Group::WIDTH`).
1862    ///
1863    /// [`RawTableInner::ctrl`]: RawTableInner::ctrl
1864    /// [`RawTableInner::find_insert_slot_in_group`]: RawTableInner::find_insert_slot_in_group
1865    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
1866    #[inline]
1867    unsafe fn fix_insert_slot(&self, mut index: usize) -> InsertSlot {
1868        // SAFETY: The caller of this function ensures that `index` is in the range `0..=self.bucket_mask`.
1869        if unlikely(self.is_bucket_full(index)) {
1870            debug_assert!(self.bucket_mask < Group::WIDTH);
1871            // SAFETY:
1872            //
1873            // * Since the caller of this function ensures that the control bytes are properly
1874            //   initialized and `ptr = self.ctrl(0)` points to the start of the array of control
1875            //   bytes, therefore: `ctrl` is valid for reads, properly aligned to `Group::WIDTH`
1876            //   and points to the properly initialized control bytes (see also
1877            //   `TableLayout::calculate_layout_for` and `ptr::read`);
1878            //
1879            // * Because the caller of this function ensures that the index was provided by the
1880            //   `self.find_insert_slot_in_group()` function, so for for tables larger than the
1881            //   group width (self.buckets() >= Group::WIDTH), we will never end up in the given
1882            //   branch, since `(probe_seq.pos + bit) & self.bucket_mask` in `find_insert_slot_in_group`
1883            //   cannot return a full bucket index. For tables smaller than the group width, calling
1884            //   the `unwrap_unchecked` function is also safe, as the trailing control bytes outside
1885            //   the range of the table are filled with EMPTY bytes (and we know for sure that there
1886            //   is at least one FULL bucket), so this second scan either finds an empty slot (due to
1887            //   the load factor) or hits the trailing control bytes (containing EMPTY).
1888            index = Group::load_aligned(self.ctrl(0))
1889                .match_empty_or_deleted()
1890                .lowest_set_bit()
1891                .unwrap_unchecked();
1892        }
1893        InsertSlot { index }
1894    }
1895
1896    /// Finds the position to insert something in a group.
1897    ///
1898    /// **This may have false positives and must be fixed up with `fix_insert_slot`
1899    /// before it's used.**
1900    ///
1901    /// The function is guaranteed to return the index of an empty or deleted [`Bucket`]
1902    /// in the range `0..self.buckets()` (`0..=self.bucket_mask`).
1903    #[inline]
1904    fn find_insert_slot_in_group(&self, group: &Group, probe_seq: &ProbeSeq) -> Option<usize> {
1905        let bit = group.match_empty_or_deleted().lowest_set_bit();
1906
1907        if likely(bit.is_some()) {
1908            // This is the same as `(probe_seq.pos + bit) % self.buckets()` because the number
1909            // of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
1910            Some((probe_seq.pos + bit.unwrap()) & self.bucket_mask)
1911        } else {
1912            None
1913        }
1914    }
1915
1916    /// Searches for an element in the table, or a potential slot where that element could
1917    /// be inserted (an empty or deleted [`Bucket`] index).
1918    ///
1919    /// This uses dynamic dispatch to reduce the amount of code generated, but that is
1920    /// eliminated by LLVM optimizations.
1921    ///
1922    /// This function does not make any changes to the `data` part of the table, or any
1923    /// changes to the `items` or `growth_left` field of the table.
1924    ///
1925    /// The table must have at least 1 empty or deleted `bucket`, otherwise, if the
1926    /// `eq: &mut dyn FnMut(usize) -> bool` function does not return `true`, this function
1927    /// will never return (will go into an infinite loop) for tables larger than the group
1928    /// width, or return an index outside of the table indices range if the table is less
1929    /// than the group width.
1930    ///
1931    /// This function is guaranteed to provide the `eq: &mut dyn FnMut(usize) -> bool`
1932    /// function with only `FULL` buckets' indices and return the `index` of the found
1933    /// element (as `Ok(index)`). If the element is not found and there is at least 1
1934    /// empty or deleted [`Bucket`] in the table, the function is guaranteed to return
1935    /// [InsertSlot] with an index in the range `0..self.buckets()`, but in any case,
1936    /// if this function returns [`InsertSlot`], it will contain an index in the range
1937    /// `0..=self.buckets()`.
1938    ///
1939    /// # Safety
1940    ///
1941    /// The [`RawTableInner`] must have properly initialized control bytes otherwise calling
1942    /// this function results in [`undefined behavior`].
1943    ///
1944    /// Attempt to write data at the [`InsertSlot`] returned by this function when the table is
1945    /// less than the group width and if there was not at least one empty or deleted bucket in
1946    /// the table will cause immediate [`undefined behavior`]. This is because in this case the
1947    /// function will return `self.bucket_mask + 1` as an index due to the trailing [`EMPTY]
1948    /// control bytes outside the table range.
1949    ///
1950    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
1951    #[inline]
1952    unsafe fn find_or_find_insert_slot_inner(
1953        &self,
1954        hash: u64,
1955        eq: &mut dyn FnMut(usize) -> bool,
1956    ) -> Result<usize, InsertSlot> {
1957        let mut insert_slot = None;
1958
1959        let h2_hash = h2(hash);
1960        let mut probe_seq = self.probe_seq(hash);
1961
1962        loop {
1963            // SAFETY:
1964            // * Caller of this function ensures that the control bytes are properly initialized.
1965            //
1966            // * `ProbeSeq.pos` cannot be greater than `self.bucket_mask = self.buckets() - 1`
1967            //   of the table due to masking with `self.bucket_mask` and also because mumber of
1968            //   buckets is a power of two (see `self.probe_seq` function).
1969            //
1970            // * Even if `ProbeSeq.pos` returns `position == self.bucket_mask`, it is safe to
1971            //   call `Group::load` due to the extended control bytes range, which is
1972            //  `self.bucket_mask + 1 + Group::WIDTH` (in fact, this means that the last control
1973            //   byte will never be read for the allocated table);
1974            //
1975            // * Also, even if `RawTableInner` is not already allocated, `ProbeSeq.pos` will
1976            //   always return "0" (zero), so Group::load will read unaligned `Group::static_empty()`
1977            //   bytes, which is safe (see RawTableInner::new).
1978            let group = unsafe { Group::load(self.ctrl(probe_seq.pos)) };
1979
1980            for bit in group.match_byte(h2_hash) {
1981                let index = (probe_seq.pos + bit) & self.bucket_mask;
1982
1983                if likely(eq(index)) {
1984                    return Ok(index);
1985                }
1986            }
1987
1988            // We didn't find the element we were looking for in the group, try to get an
1989            // insertion slot from the group if we don't have one yet.
1990            if likely(insert_slot.is_none()) {
1991                insert_slot = self.find_insert_slot_in_group(&group, &probe_seq);
1992            }
1993
1994            // Only stop the search if the group contains at least one empty element.
1995            // Otherwise, the element that we are looking for might be in a following group.
1996            if likely(group.match_empty().any_bit_set()) {
1997                // We must have found a insert slot by now, since the current group contains at
1998                // least one. For tables smaller than the group width, there will still be an
1999                // empty element in the current (and only) group due to the load factor.
2000                unsafe {
2001                    // SAFETY:
2002                    // * Caller of this function ensures that the control bytes are properly initialized.
2003                    //
2004                    // * We use this function with the slot / index found by `self.find_insert_slot_in_group`
2005                    return Err(self.fix_insert_slot(insert_slot.unwrap_unchecked()));
2006                }
2007            }
2008
2009            probe_seq.move_next(self.bucket_mask);
2010        }
2011    }
2012
2013    /// Searches for an empty or deleted bucket which is suitable for inserting a new
2014    /// element and sets the hash for that slot. Returns an index of that slot and the
2015    /// old control byte stored in the found index.
2016    ///
2017    /// This function does not check if the given element exists in the table. Also,
2018    /// this function does not check if there is enough space in the table to insert
2019    /// a new element. Caller of the funtion must make ensure that the table has at
2020    /// least 1 empty or deleted `bucket`, otherwise this function will never return
2021    /// (will go into an infinite loop) for tables larger than the group width, or
2022    /// return an index outside of the table indices range if the table is less than
2023    /// the group width.
2024    ///
2025    /// If there is at least 1 empty or deleted `bucket` in the table, the function is
2026    /// guaranteed to return an `index` in the range `0..self.buckets()`, but in any case,
2027    /// if this function returns an `index` it will be in the range `0..=self.buckets()`.
2028    ///
2029    /// This function does not make any changes to the `data` parts of the table,
2030    /// or any changes to the `items` or `growth_left` field of the table.
2031    ///
2032    /// # Safety
2033    ///
2034    /// The safety rules are directly derived from the safety rules for the
2035    /// [`RawTableInner::set_ctrl_h2`] and [`RawTableInner::find_insert_slot`] methods.
2036    /// Thus, in order to uphold the safety contracts for that methods, as well as for
2037    /// the correct logic of the work of this crate, you must observe the following rules
2038    /// when calling this function:
2039    ///
2040    /// * The [`RawTableInner`] has already been allocated and has properly initialized
2041    ///   control bytes otherwise calling this function results in [`undefined behavior`].
2042    ///
2043    /// * The caller of this function must ensure that the "data" parts of the table
2044    ///   will have an entry in the returned index (matching the given hash) right
2045    ///   after calling this function.
2046    ///
2047    /// Attempt to write data at the `index` returned by this function when the table is
2048    /// less than the group width and if there was not at least one empty or deleted bucket in
2049    /// the table will cause immediate [`undefined behavior`]. This is because in this case the
2050    /// function will return `self.bucket_mask + 1` as an index due to the trailing [`EMPTY]
2051    /// control bytes outside the table range.
2052    ///
2053    /// The caller must independently increase the `items` field of the table, and also,
2054    /// if the old control byte was [`EMPTY`], then decrease the table's `growth_left`
2055    /// field, and do not change it if the old control byte was [`DELETED`].
2056    ///
2057    /// See also [`Bucket::as_ptr`] method, for more information about of properly removing
2058    /// or saving `element` from / into the [`RawTable`] / [`RawTableInner`].
2059    ///
2060    /// [`Bucket::as_ptr`]: Bucket::as_ptr
2061    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2062    /// [`RawTableInner::ctrl`]: RawTableInner::ctrl
2063    /// [`RawTableInner::set_ctrl_h2`]: RawTableInner::set_ctrl_h2
2064    /// [`RawTableInner::find_insert_slot`]: RawTableInner::find_insert_slot
2065    #[inline]
2066    unsafe fn prepare_insert_slot(&mut self, hash: u64) -> (usize, u8) {
2067        // SAFETY: Caller of this function ensures that the control bytes are properly initialized.
2068        let index: usize = self.find_insert_slot(hash).index;
2069        // SAFETY:
2070        // 1. The `find_insert_slot` function either returns an `index` less than or
2071        //    equal to `self.buckets() = self.bucket_mask + 1` of the table, or never
2072        //    returns if it cannot find an empty or deleted slot.
2073        // 2. The caller of this function guarantees that the table has already been
2074        //    allocated
2075        let old_ctrl = *self.ctrl(index);
2076        self.set_ctrl_h2(index, hash);
2077        (index, old_ctrl)
2078    }
2079
2080    /// Searches for an empty or deleted bucket which is suitable for inserting
2081    /// a new element, returning the `index` for the new [`Bucket`].
2082    ///
2083    /// This function does not make any changes to the `data` part of the table, or any
2084    /// changes to the `items` or `growth_left` field of the table.
2085    ///
2086    /// The table must have at least 1 empty or deleted `bucket`, otherwise this function
2087    /// will never return (will go into an infinite loop) for tables larger than the group
2088    /// width, or return an index outside of the table indices range if the table is less
2089    /// than the group width.
2090    ///
2091    /// If there is at least 1 empty or deleted `bucket` in the table, the function is
2092    /// guaranteed to return [`InsertSlot`] with an index in the range `0..self.buckets()`,
2093    /// but in any case, if this function returns [`InsertSlot`], it will contain an index
2094    /// in the range `0..=self.buckets()`.
2095    ///
2096    /// # Safety
2097    ///
2098    /// The [`RawTableInner`] must have properly initialized control bytes otherwise calling
2099    /// this function results in [`undefined behavior`].
2100    ///
2101    /// Attempt to write data at the [`InsertSlot`] returned by this function when the table is
2102    /// less than the group width and if there was not at least one empty or deleted bucket in
2103    /// the table will cause immediate [`undefined behavior`]. This is because in this case the
2104    /// function will return `self.bucket_mask + 1` as an index due to the trailing [`EMPTY]
2105    /// control bytes outside the table range.
2106    ///
2107    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2108    #[inline]
2109    unsafe fn find_insert_slot(&self, hash: u64) -> InsertSlot {
2110        let mut probe_seq = self.probe_seq(hash);
2111        loop {
2112            // SAFETY:
2113            // * Caller of this function ensures that the control bytes are properly initialized.
2114            //
2115            // * `ProbeSeq.pos` cannot be greater than `self.bucket_mask = self.buckets() - 1`
2116            //   of the table due to masking with `self.bucket_mask` and also because mumber of
2117            //   buckets is a power of two (see `self.probe_seq` function).
2118            //
2119            // * Even if `ProbeSeq.pos` returns `position == self.bucket_mask`, it is safe to
2120            //   call `Group::load` due to the extended control bytes range, which is
2121            //  `self.bucket_mask + 1 + Group::WIDTH` (in fact, this means that the last control
2122            //   byte will never be read for the allocated table);
2123            //
2124            // * Also, even if `RawTableInner` is not already allocated, `ProbeSeq.pos` will
2125            //   always return "0" (zero), so Group::load will read unaligned `Group::static_empty()`
2126            //   bytes, which is safe (see RawTableInner::new).
2127            let group = unsafe { Group::load(self.ctrl(probe_seq.pos)) };
2128
2129            let index = self.find_insert_slot_in_group(&group, &probe_seq);
2130            if likely(index.is_some()) {
2131                // SAFETY:
2132                // * Caller of this function ensures that the control bytes are properly initialized.
2133                //
2134                // * We use this function with the slot / index found by `self.find_insert_slot_in_group`
2135                unsafe {
2136                    return self.fix_insert_slot(index.unwrap_unchecked());
2137                }
2138            }
2139            probe_seq.move_next(self.bucket_mask);
2140        }
2141    }
2142
2143    /// Searches for an element in a table, returning the `index` of the found element.
2144    /// This uses dynamic dispatch to reduce the amount of code generated, but it is
2145    /// eliminated by LLVM optimizations.
2146    ///
2147    /// This function does not make any changes to the `data` part of the table, or any
2148    /// changes to the `items` or `growth_left` field of the table.
2149    ///
2150    /// The table must have at least 1 empty `bucket`, otherwise, if the
2151    /// `eq: &mut dyn FnMut(usize) -> bool` function does not return `true`,
2152    /// this function will also never return (will go into an infinite loop).
2153    ///
2154    /// This function is guaranteed to provide the `eq: &mut dyn FnMut(usize) -> bool`
2155    /// function with only `FULL` buckets' indices and return the `index` of the found
2156    /// element as `Some(index)`, so the index will always be in the range
2157    /// `0..self.buckets()`.
2158    ///
2159    /// # Safety
2160    ///
2161    /// The [`RawTableInner`] must have properly initialized control bytes otherwise calling
2162    /// this function results in [`undefined behavior`].
2163    ///
2164    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2165    #[inline(always)]
2166    unsafe fn find_inner(&self, hash: u64, eq: &mut dyn FnMut(usize) -> bool) -> Option<usize> {
2167        let h2_hash = h2(hash);
2168        let mut probe_seq = self.probe_seq(hash);
2169
2170        loop {
2171            // SAFETY:
2172            // * Caller of this function ensures that the control bytes are properly initialized.
2173            //
2174            // * `ProbeSeq.pos` cannot be greater than `self.bucket_mask = self.buckets() - 1`
2175            //   of the table due to masking with `self.bucket_mask`.
2176            //
2177            // * Even if `ProbeSeq.pos` returns `position == self.bucket_mask`, it is safe to
2178            //   call `Group::load` due to the extended control bytes range, which is
2179            //  `self.bucket_mask + 1 + Group::WIDTH` (in fact, this means that the last control
2180            //   byte will never be read for the allocated table);
2181            //
2182            // * Also, even if `RawTableInner` is not already allocated, `ProbeSeq.pos` will
2183            //   always return "0" (zero), so Group::load will read unaligned `Group::static_empty()`
2184            //   bytes, which is safe (see RawTableInner::new_in).
2185            let group = unsafe { Group::load(self.ctrl(probe_seq.pos)) };
2186
2187            for bit in group.match_byte(h2_hash) {
2188                // This is the same as `(probe_seq.pos + bit) % self.buckets()` because the number
2189                // of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
2190                let index = (probe_seq.pos + bit) & self.bucket_mask;
2191
2192                if likely(eq(index)) {
2193                    return Some(index);
2194                }
2195            }
2196
2197            if likely(group.match_empty().any_bit_set()) {
2198                return None;
2199            }
2200
2201            probe_seq.move_next(self.bucket_mask);
2202        }
2203    }
2204
2205    /// Prepares for rehashing data in place (that is, without allocating new memory).
2206    /// Converts all full index `control bytes` to `DELETED` and all `DELETED` control
2207    /// bytes to `EMPTY`, i.e. performs the following conversion:
2208    ///
2209    /// - `EMPTY` control bytes   -> `EMPTY`;
2210    /// - `DELETED` control bytes -> `EMPTY`;
2211    /// - `FULL` control bytes    -> `DELETED`.
2212    ///
2213    /// This function does not make any changes to the `data` parts of the table,
2214    /// or any changes to the `items` or `growth_left` field of the table.
2215    ///
2216    /// # Safety
2217    ///
2218    /// You must observe the following safety rules when calling this function:
2219    ///
2220    /// * The [`RawTableInner`] has already been allocated;
2221    ///
2222    /// * The caller of this function must convert the `DELETED` bytes back to `FULL`
2223    ///   bytes when re-inserting them into their ideal position (which was impossible
2224    ///   to do during the first insert due to tombstones). If the caller does not do
2225    ///   this, then calling this function may result in a memory leak.
2226    ///
2227    /// * The [`RawTableInner`] must have properly initialized control bytes otherwise
2228    ///   calling this function results in [`undefined behavior`].
2229    ///
2230    /// Calling this function on a table that has not been allocated results in
2231    /// [`undefined behavior`].
2232    ///
2233    /// See also [`Bucket::as_ptr`] method, for more information about of properly removing
2234    /// or saving `data element` from / into the [`RawTable`] / [`RawTableInner`].
2235    ///
2236    /// [`Bucket::as_ptr`]: Bucket::as_ptr
2237    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2238    #[allow(clippy::mut_mut)]
2239    #[inline]
2240    unsafe fn prepare_rehash_in_place(&mut self) {
2241        // Bulk convert all full control bytes to DELETED, and all DELETED control bytes to EMPTY.
2242        // This effectively frees up all buckets containing a DELETED entry.
2243        //
2244        // SAFETY:
2245        // 1. `i` is guaranteed to be within bounds since we are iterating from zero to `buckets - 1`;
2246        // 2. Even if `i` will be `i == self.bucket_mask`, it is safe to call `Group::load_aligned`
2247        //    due to the extended control bytes range, which is `self.bucket_mask + 1 + Group::WIDTH`;
2248        // 3. The caller of this function guarantees that [`RawTableInner`] has already been allocated;
2249        // 4. We can use `Group::load_aligned` and `Group::store_aligned` here since we start from 0
2250        //    and go to the end with a step equal to `Group::WIDTH` (see TableLayout::calculate_layout_for).
2251        for i in (0..self.buckets()).step_by(Group::WIDTH) {
2252            let group = Group::load_aligned(self.ctrl(i));
2253            let group = group.convert_special_to_empty_and_full_to_deleted();
2254            group.store_aligned(self.ctrl(i));
2255        }
2256
2257        // Fix up the trailing control bytes. See the comments in set_ctrl
2258        // for the handling of tables smaller than the group width.
2259        //
2260        // SAFETY: The caller of this function guarantees that [`RawTableInner`]
2261        // has already been allocated
2262        if unlikely(self.buckets() < Group::WIDTH) {
2263            // SAFETY: We have `self.bucket_mask + 1 + Group::WIDTH` number of control bytes,
2264            // so copying `self.buckets() == self.bucket_mask + 1` bytes with offset equal to
2265            // `Group::WIDTH` is safe
2266            self.ctrl(0)
2267                .copy_to(self.ctrl(Group::WIDTH), self.buckets());
2268        } else {
2269            // SAFETY: We have `self.bucket_mask + 1 + Group::WIDTH` number of
2270            // control bytes,so copying `Group::WIDTH` bytes with offset equal
2271            // to `self.buckets() == self.bucket_mask + 1` is safe
2272            self.ctrl(0)
2273                .copy_to(self.ctrl(self.buckets()), Group::WIDTH);
2274        }
2275    }
2276
2277    /// Returns an iterator over every element in the table.
2278    ///
2279    /// # Safety
2280    ///
2281    /// If any of the following conditions are violated, the result
2282    /// is [`undefined behavior`]:
2283    ///
2284    /// * The caller has to ensure that the `RawTableInner` outlives the
2285    ///   `RawIter`. Because we cannot make the `next` method unsafe on
2286    ///   the `RawIter` struct, we have to make the `iter` method unsafe.
2287    ///
2288    /// * The [`RawTableInner`] must have properly initialized control bytes.
2289    ///
2290    /// The type `T` must be the actual type of the elements stored in the table,
2291    /// otherwise using the returned [`RawIter`] results in [`undefined behavior`].
2292    ///
2293    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2294    #[inline]
2295    unsafe fn iter<T>(&self) -> RawIter<T> {
2296        // SAFETY:
2297        // 1. Since the caller of this function ensures that the control bytes
2298        //    are properly initialized and `self.data_end()` points to the start
2299        //    of the array of control bytes, therefore: `ctrl` is valid for reads,
2300        //    properly aligned to `Group::WIDTH` and points to the properly initialized
2301        //    control bytes.
2302        // 2. `data` bucket index in the table is equal to the `ctrl` index (i.e.
2303        //    equal to zero).
2304        // 3. We pass the exact value of buckets of the table to the function.
2305        //
2306        //                         `ctrl` points here (to the start
2307        //                         of the first control byte `CT0`)
2308        //                          ∨
2309        // [Pad], T_n, ..., T1, T0, |CT0, CT1, ..., CT_n|, CTa_0, CTa_1, ..., CTa_m
2310        //                           \________  ________/
2311        //                                    \/
2312        //       `n = buckets - 1`, i.e. `RawTableInner::buckets() - 1`
2313        //
2314        // where: T0...T_n  - our stored data;
2315        //        CT0...CT_n - control bytes or metadata for `data`.
2316        //        CTa_0...CTa_m - additional control bytes, where `m = Group::WIDTH - 1` (so that the search
2317        //                        with loading `Group` bytes from the heap works properly, even if the result
2318        //                        of `h1(hash) & self.bucket_mask` is equal to `self.bucket_mask`). See also
2319        //                        `RawTableInner::set_ctrl` function.
2320        //
2321        // P.S. `h1(hash) & self.bucket_mask` is the same as `hash as usize % self.buckets()` because the number
2322        // of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
2323        let data = Bucket::from_base_index(self.data_end(), 0);
2324        RawIter {
2325            // SAFETY: See explanation above
2326            iter: RawIterRange::new(self.ctrl.as_ptr(), data, self.buckets()),
2327            items: self.items,
2328        }
2329    }
2330
2331    /// Executes the destructors (if any) of the values stored in the table.
2332    ///
2333    /// # Note
2334    ///
2335    /// This function does not erase the control bytes of the table and does
2336    /// not make any changes to the `items` or `growth_left` fields of the
2337    /// table. If necessary, the caller of this function must manually set
2338    /// up these table fields, for example using the [`clear_no_drop`] function.
2339    ///
2340    /// Be careful during calling this function, because drop function of
2341    /// the elements can panic, and this can leave table in an inconsistent
2342    /// state.
2343    ///
2344    /// # Safety
2345    ///
2346    /// The type `T` must be the actual type of the elements stored in the table,
2347    /// otherwise calling this function may result in [`undefined behavior`].
2348    ///
2349    /// If `T` is a type that should be dropped and **the table is not empty**,
2350    /// calling this function more than once results in [`undefined behavior`].
2351    ///
2352    /// If `T` is not [`Copy`], attempting to use values stored in the table after
2353    /// calling this function may result in [`undefined behavior`].
2354    ///
2355    /// It is safe to call this function on a table that has not been allocated,
2356    /// on a table with uninitialized control bytes, and on a table with no actual
2357    /// data but with `Full` control bytes if `self.items == 0`.
2358    ///
2359    /// See also [`Bucket::drop`] / [`Bucket::as_ptr`] methods, for more information
2360    /// about of properly removing or saving `element` from / into the [`RawTable`] /
2361    /// [`RawTableInner`].
2362    ///
2363    /// [`Bucket::drop`]: Bucket::drop
2364    /// [`Bucket::as_ptr`]: Bucket::as_ptr
2365    /// [`clear_no_drop`]: RawTableInner::clear_no_drop
2366    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2367    unsafe fn drop_elements<T>(&mut self) {
2368        // Check that `self.items != 0`. Protects against the possibility
2369        // of creating an iterator on an table with uninitialized control bytes.
2370        if T::NEEDS_DROP && self.items != 0 {
2371            // SAFETY: We know for sure that RawTableInner will outlive the
2372            // returned `RawIter` iterator, and the caller of this function
2373            // must uphold the safety contract for `drop_elements` method.
2374            for item in self.iter::<T>() {
2375                // SAFETY: The caller must uphold the safety contract for
2376                // `drop_elements` method.
2377                item.drop();
2378            }
2379        }
2380    }
2381
2382    /// Executes the destructors (if any) of the values stored in the table and than
2383    /// deallocates the table.
2384    ///
2385    /// # Note
2386    ///
2387    /// Calling this function automatically makes invalid (dangling) all instances of
2388    /// buckets ([`Bucket`]) and makes invalid (dangling) the `ctrl` field of the table.
2389    ///
2390    /// This function does not make any changes to the `bucket_mask`, `items` or `growth_left`
2391    /// fields of the table. If necessary, the caller of this function must manually set
2392    /// up these table fields.
2393    ///
2394    /// # Safety
2395    ///
2396    /// If any of the following conditions are violated, the result is [`undefined behavior`]:
2397    ///
2398    /// * Calling this function more than once;
2399    ///
2400    /// * The type `T` must be the actual type of the elements stored in the table.
2401    ///
2402    /// * The `alloc` must be the same [`Allocator`] as the `Allocator` that was used
2403    ///   to allocate this table.
2404    ///
2405    /// * The `table_layout` must be the same [`TableLayout`] as the `TableLayout` that
2406    ///   was used to allocate this table.
2407    ///
2408    /// The caller of this function should pay attention to the possibility of the
2409    /// elements' drop function panicking, because this:
2410    ///
2411    ///    * May leave the table in an inconsistent state;
2412    ///
2413    ///    * Memory is never deallocated, so a memory leak may occur.
2414    ///
2415    /// Attempt to use the `ctrl` field of the table (dereference) after calling this
2416    /// function results in [`undefined behavior`].
2417    ///
2418    /// It is safe to call this function on a table that has not been allocated,
2419    /// on a table with uninitialized control bytes, and on a table with no actual
2420    /// data but with `Full` control bytes if `self.items == 0`.
2421    ///
2422    /// See also [`RawTableInner::drop_elements`] or [`RawTableInner::free_buckets`]
2423    /// for more  information.
2424    ///
2425    /// [`RawTableInner::drop_elements`]: RawTableInner::drop_elements
2426    /// [`RawTableInner::free_buckets`]: RawTableInner::free_buckets
2427    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2428    unsafe fn drop_inner_table<T, A: Allocator>(&mut self, alloc: &A, table_layout: TableLayout) {
2429        if !self.is_empty_singleton() {
2430            unsafe {
2431                // SAFETY: The caller must uphold the safety contract for `drop_inner_table` method.
2432                self.drop_elements::<T>();
2433                // SAFETY:
2434                // 1. We have checked that our table is allocated.
2435                // 2. The caller must uphold the safety contract for `drop_inner_table` method.
2436                self.free_buckets(alloc, table_layout);
2437            }
2438        }
2439    }
2440
2441    /// Returns a pointer to an element in the table (convenience for
2442    /// `Bucket::from_base_index(self.data_end::<T>(), index)`).
2443    ///
2444    /// The caller must ensure that the `RawTableInner` outlives the returned [`Bucket<T>`],
2445    /// otherwise using it may result in [`undefined behavior`].
2446    ///
2447    /// # Safety
2448    ///
2449    /// If `mem::size_of::<T>() != 0`, then the safety rules are directly derived from the
2450    /// safety rules of the [`Bucket::from_base_index`] function. Therefore, when calling
2451    /// this function, the following safety rules must be observed:
2452    ///
2453    /// * The table must already be allocated;
2454    ///
2455    /// * The `index` must not be greater than the number returned by the [`RawTableInner::buckets`]
2456    ///   function, i.e. `(index + 1) <= self.buckets()`.
2457    ///
2458    /// * The type `T` must be the actual type of the elements stored in the table, otherwise
2459    ///   using the returned [`Bucket`] may result in [`undefined behavior`].
2460    ///
2461    /// It is safe to call this function with index of zero (`index == 0`) on a table that has
2462    /// not been allocated, but using the returned [`Bucket`] results in [`undefined behavior`].
2463    ///
2464    /// If `mem::size_of::<T>() == 0`, then the only requirement is that the `index` must
2465    /// not be greater than the number returned by the [`RawTable::buckets`] function, i.e.
2466    /// `(index + 1) <= self.buckets()`.
2467    ///
2468    /// ```none
2469    /// If mem::size_of::<T>() != 0 then return a pointer to the `element` in the `data part` of the table
2470    /// (we start counting from "0", so that in the expression T[n], the "n" index actually one less than
2471    /// the "buckets" number of our `RawTableInner`, i.e. "n = RawTableInner::buckets() - 1"):
2472    ///
2473    ///           `table.bucket(3).as_ptr()` returns a pointer that points here in the `data`
2474    ///           part of the `RawTableInner`, i.e. to the start of T3 (see [`Bucket::as_ptr`])
2475    ///                  |
2476    ///                  |               `base = table.data_end::<T>()` points here
2477    ///                  |               (to the start of CT0 or to the end of T0)
2478    ///                  v                 v
2479    /// [Pad], T_n, ..., |T3|, T2, T1, T0, |CT0, CT1, CT2, CT3, ..., CT_n, CTa_0, CTa_1, ..., CTa_m
2480    ///                     ^                                              \__________  __________/
2481    ///        `table.bucket(3)` returns a pointer that points                        \/
2482    ///         here in the `data` part of the `RawTableInner`             additional control bytes
2483    ///         (to the end of T3)                                          `m = Group::WIDTH - 1`
2484    ///
2485    /// where: T0...T_n  - our stored data;
2486    ///        CT0...CT_n - control bytes or metadata for `data`;
2487    ///        CTa_0...CTa_m - additional control bytes (so that the search with loading `Group` bytes from
2488    ///                        the heap works properly, even if the result of `h1(hash) & self.bucket_mask`
2489    ///                        is equal to `self.bucket_mask`). See also `RawTableInner::set_ctrl` function.
2490    ///
2491    /// P.S. `h1(hash) & self.bucket_mask` is the same as `hash as usize % self.buckets()` because the number
2492    /// of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
2493    /// ```
2494    ///
2495    /// [`Bucket::from_base_index`]: Bucket::from_base_index
2496    /// [`RawTableInner::buckets`]: RawTableInner::buckets
2497    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2498    #[inline]
2499    unsafe fn bucket<T>(&self, index: usize) -> Bucket<T> {
2500        debug_assert_ne!(self.bucket_mask, 0);
2501        debug_assert!(index < self.buckets());
2502        Bucket::from_base_index(self.data_end(), index)
2503    }
2504
2505    /// Returns a raw `*mut u8` pointer to the start of the `data` element in the table
2506    /// (convenience for `self.data_end::<u8>().as_ptr().sub((index + 1) * size_of)`).
2507    ///
2508    /// The caller must ensure that the `RawTableInner` outlives the returned `*mut u8`,
2509    /// otherwise using it may result in [`undefined behavior`].
2510    ///
2511    /// # Safety
2512    ///
2513    /// If any of the following conditions are violated, the result is [`undefined behavior`]:
2514    ///
2515    /// * The table must already be allocated;
2516    ///
2517    /// * The `index` must not be greater than the number returned by the [`RawTableInner::buckets`]
2518    ///   function, i.e. `(index + 1) <= self.buckets()`;
2519    ///
2520    /// * The `size_of` must be equal to the size of the elements stored in the table;
2521    ///
2522    /// ```none
2523    /// If mem::size_of::<T>() != 0 then return a pointer to the `element` in the `data part` of the table
2524    /// (we start counting from "0", so that in the expression T[n], the "n" index actually one less than
2525    /// the "buckets" number of our `RawTableInner`, i.e. "n = RawTableInner::buckets() - 1"):
2526    ///
2527    ///           `table.bucket_ptr(3, mem::size_of::<T>())` returns a pointer that points here in the
2528    ///           `data` part of the `RawTableInner`, i.e. to the start of T3
2529    ///                  |
2530    ///                  |               `base = table.data_end::<u8>()` points here
2531    ///                  |               (to the start of CT0 or to the end of T0)
2532    ///                  v                 v
2533    /// [Pad], T_n, ..., |T3|, T2, T1, T0, |CT0, CT1, CT2, CT3, ..., CT_n, CTa_0, CTa_1, ..., CTa_m
2534    ///                                                                    \__________  __________/
2535    ///                                                                               \/
2536    ///                                                                    additional control bytes
2537    ///                                                                     `m = Group::WIDTH - 1`
2538    ///
2539    /// where: T0...T_n  - our stored data;
2540    ///        CT0...CT_n - control bytes or metadata for `data`;
2541    ///        CTa_0...CTa_m - additional control bytes (so that the search with loading `Group` bytes from
2542    ///                        the heap works properly, even if the result of `h1(hash) & self.bucket_mask`
2543    ///                        is equal to `self.bucket_mask`). See also `RawTableInner::set_ctrl` function.
2544    ///
2545    /// P.S. `h1(hash) & self.bucket_mask` is the same as `hash as usize % self.buckets()` because the number
2546    /// of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
2547    /// ```
2548    ///
2549    /// [`RawTableInner::buckets`]: RawTableInner::buckets
2550    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2551    #[inline]
2552    unsafe fn bucket_ptr(&self, index: usize, size_of: usize) -> *mut u8 {
2553        debug_assert_ne!(self.bucket_mask, 0);
2554        debug_assert!(index < self.buckets());
2555        let base: *mut u8 = self.data_end().as_ptr();
2556        base.sub((index + 1) * size_of)
2557    }
2558
2559    /// Returns pointer to one past last `data` element in the table as viewed from
2560    /// the start point of the allocation (convenience for `self.ctrl.cast()`).
2561    ///
2562    /// This function actually returns a pointer to the end of the `data element` at
2563    /// index "0" (zero).
2564    ///
2565    /// The caller must ensure that the `RawTableInner` outlives the returned [`NonNull<T>`],
2566    /// otherwise using it may result in [`undefined behavior`].
2567    ///
2568    /// # Note
2569    ///
2570    /// The type `T` must be the actual type of the elements stored in the table, otherwise
2571    /// using the returned [`NonNull<T>`] may result in [`undefined behavior`].
2572    ///
2573    /// ```none
2574    ///                        `table.data_end::<T>()` returns pointer that points here
2575    ///                        (to the end of `T0`)
2576    ///                          ∨
2577    /// [Pad], T_n, ..., T1, T0, |CT0, CT1, ..., CT_n|, CTa_0, CTa_1, ..., CTa_m
2578    ///                           \________  ________/
2579    ///                                    \/
2580    ///       `n = buckets - 1`, i.e. `RawTableInner::buckets() - 1`
2581    ///
2582    /// where: T0...T_n  - our stored data;
2583    ///        CT0...CT_n - control bytes or metadata for `data`.
2584    ///        CTa_0...CTa_m - additional control bytes, where `m = Group::WIDTH - 1` (so that the search
2585    ///                        with loading `Group` bytes from the heap works properly, even if the result
2586    ///                        of `h1(hash) & self.bucket_mask` is equal to `self.bucket_mask`). See also
2587    ///                        `RawTableInner::set_ctrl` function.
2588    ///
2589    /// P.S. `h1(hash) & self.bucket_mask` is the same as `hash as usize % self.buckets()` because the number
2590    /// of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
2591    /// ```
2592    ///
2593    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2594    #[inline]
2595    fn data_end<T>(&self) -> NonNull<T> {
2596        self.ctrl.cast()
2597    }
2598
2599    /// Returns an iterator-like object for a probe sequence on the table.
2600    ///
2601    /// This iterator never terminates, but is guaranteed to visit each bucket
2602    /// group exactly once. The loop using `probe_seq` must terminate upon
2603    /// reaching a group containing an empty bucket.
2604    #[inline]
2605    fn probe_seq(&self, hash: u64) -> ProbeSeq {
2606        ProbeSeq {
2607            // This is the same as `hash as usize % self.buckets()` because the number
2608            // of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
2609            pos: h1(hash) & self.bucket_mask,
2610            stride: 0,
2611        }
2612    }
2613
2614    /// Returns the index of a bucket for which a value must be inserted if there is enough rooom
2615    /// in the table, otherwise returns error
2616    #[cfg(feature = "raw")]
2617    #[inline]
2618    unsafe fn prepare_insert_no_grow(&mut self, hash: u64) -> Result<usize, ()> {
2619        let index = self.find_insert_slot(hash).index;
2620        let old_ctrl = *self.ctrl(index);
2621        if unlikely(self.growth_left == 0 && special_is_empty(old_ctrl)) {
2622            Err(())
2623        } else {
2624            self.record_item_insert_at(index, old_ctrl, hash);
2625            Ok(index)
2626        }
2627    }
2628
2629    #[inline]
2630    unsafe fn record_item_insert_at(&mut self, index: usize, old_ctrl: u8, hash: u64) {
2631        self.growth_left -= usize::from(special_is_empty(old_ctrl));
2632        self.set_ctrl_h2(index, hash);
2633        self.items += 1;
2634    }
2635
2636    #[inline]
2637    fn is_in_same_group(&self, i: usize, new_i: usize, hash: u64) -> bool {
2638        let probe_seq_pos = self.probe_seq(hash).pos;
2639        let probe_index =
2640            |pos: usize| (pos.wrapping_sub(probe_seq_pos) & self.bucket_mask) / Group::WIDTH;
2641        probe_index(i) == probe_index(new_i)
2642    }
2643
2644    /// Sets a control byte to the hash, and possibly also the replicated control byte at
2645    /// the end of the array.
2646    ///
2647    /// This function does not make any changes to the `data` parts of the table,
2648    /// or any changes to the `items` or `growth_left` field of the table.
2649    ///
2650    /// # Safety
2651    ///
2652    /// The safety rules are directly derived from the safety rules for [`RawTableInner::set_ctrl`]
2653    /// method. Thus, in order to uphold the safety contracts for the method, you must observe the
2654    /// following rules when calling this function:
2655    ///
2656    /// * The [`RawTableInner`] has already been allocated;
2657    ///
2658    /// * The `index` must not be greater than the `RawTableInner.bucket_mask`, i.e.
2659    ///   `index <= RawTableInner.bucket_mask` or, in other words, `(index + 1)` must
2660    ///   be no greater than the number returned by the function [`RawTableInner::buckets`].
2661    ///
2662    /// Calling this function on a table that has not been allocated results in [`undefined behavior`].
2663    ///
2664    /// See also [`Bucket::as_ptr`] method, for more information about of properly removing
2665    /// or saving `data element` from / into the [`RawTable`] / [`RawTableInner`].
2666    ///
2667    /// [`RawTableInner::set_ctrl`]: RawTableInner::set_ctrl
2668    /// [`RawTableInner::buckets`]: RawTableInner::buckets
2669    /// [`Bucket::as_ptr`]: Bucket::as_ptr
2670    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2671    #[inline]
2672    unsafe fn set_ctrl_h2(&mut self, index: usize, hash: u64) {
2673        // SAFETY: The caller must uphold the safety rules for the [`RawTableInner::set_ctrl_h2`]
2674        self.set_ctrl(index, h2(hash));
2675    }
2676
2677    /// Replaces the hash in the control byte at the given index with the provided one,
2678    /// and possibly also replicates the new control byte at the end of the array of control
2679    /// bytes, returning the old control byte.
2680    ///
2681    /// This function does not make any changes to the `data` parts of the table,
2682    /// or any changes to the `items` or `growth_left` field of the table.
2683    ///
2684    /// # Safety
2685    ///
2686    /// The safety rules are directly derived from the safety rules for [`RawTableInner::set_ctrl_h2`]
2687    /// and [`RawTableInner::ctrl`] methods. Thus, in order to uphold the safety contracts for both
2688    /// methods, you must observe the following rules when calling this function:
2689    ///
2690    /// * The [`RawTableInner`] has already been allocated;
2691    ///
2692    /// * The `index` must not be greater than the `RawTableInner.bucket_mask`, i.e.
2693    ///   `index <= RawTableInner.bucket_mask` or, in other words, `(index + 1)` must
2694    ///   be no greater than the number returned by the function [`RawTableInner::buckets`].
2695    ///
2696    /// Calling this function on a table that has not been allocated results in [`undefined behavior`].
2697    ///
2698    /// See also [`Bucket::as_ptr`] method, for more information about of properly removing
2699    /// or saving `data element` from / into the [`RawTable`] / [`RawTableInner`].
2700    ///
2701    /// [`RawTableInner::set_ctrl_h2`]: RawTableInner::set_ctrl_h2
2702    /// [`RawTableInner::buckets`]: RawTableInner::buckets
2703    /// [`Bucket::as_ptr`]: Bucket::as_ptr
2704    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2705    #[inline]
2706    unsafe fn replace_ctrl_h2(&mut self, index: usize, hash: u64) -> u8 {
2707        // SAFETY: The caller must uphold the safety rules for the [`RawTableInner::replace_ctrl_h2`]
2708        let prev_ctrl = *self.ctrl(index);
2709        self.set_ctrl_h2(index, hash);
2710        prev_ctrl
2711    }
2712
2713    /// Sets a control byte, and possibly also the replicated control byte at
2714    /// the end of the array.
2715    ///
2716    /// This function does not make any changes to the `data` parts of the table,
2717    /// or any changes to the `items` or `growth_left` field of the table.
2718    ///
2719    /// # Safety
2720    ///
2721    /// You must observe the following safety rules when calling this function:
2722    ///
2723    /// * The [`RawTableInner`] has already been allocated;
2724    ///
2725    /// * The `index` must not be greater than the `RawTableInner.bucket_mask`, i.e.
2726    ///   `index <= RawTableInner.bucket_mask` or, in other words, `(index + 1)` must
2727    ///   be no greater than the number returned by the function [`RawTableInner::buckets`].
2728    ///
2729    /// Calling this function on a table that has not been allocated results in [`undefined behavior`].
2730    ///
2731    /// See also [`Bucket::as_ptr`] method, for more information about of properly removing
2732    /// or saving `data element` from / into the [`RawTable`] / [`RawTableInner`].
2733    ///
2734    /// [`RawTableInner::buckets`]: RawTableInner::buckets
2735    /// [`Bucket::as_ptr`]: Bucket::as_ptr
2736    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2737    #[inline]
2738    unsafe fn set_ctrl(&mut self, index: usize, ctrl: u8) {
2739        // Replicate the first Group::WIDTH control bytes at the end of
2740        // the array without using a branch. If the tables smaller than
2741        // the group width (self.buckets() < Group::WIDTH),
2742        // `index2 = Group::WIDTH + index`, otherwise `index2` is:
2743        //
2744        // - If index >= Group::WIDTH then index == index2.
2745        // - Otherwise index2 == self.bucket_mask + 1 + index.
2746        //
2747        // The very last replicated control byte is never actually read because
2748        // we mask the initial index for unaligned loads, but we write it
2749        // anyways because it makes the set_ctrl implementation simpler.
2750        //
2751        // If there are fewer buckets than Group::WIDTH then this code will
2752        // replicate the buckets at the end of the trailing group. For example
2753        // with 2 buckets and a group size of 4, the control bytes will look
2754        // like this:
2755        //
2756        //     Real    |             Replicated
2757        // ---------------------------------------------
2758        // | [A] | [B] | [EMPTY] | [EMPTY] | [A] | [B] |
2759        // ---------------------------------------------
2760
2761        // This is the same as `(index.wrapping_sub(Group::WIDTH)) % self.buckets() + Group::WIDTH`
2762        // because the number of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
2763        let index2 = ((index.wrapping_sub(Group::WIDTH)) & self.bucket_mask) + Group::WIDTH;
2764
2765        // SAFETY: The caller must uphold the safety rules for the [`RawTableInner::set_ctrl`]
2766        *self.ctrl(index) = ctrl;
2767        *self.ctrl(index2) = ctrl;
2768    }
2769
2770    /// Returns a pointer to a control byte.
2771    ///
2772    /// # Safety
2773    ///
2774    /// For the allocated [`RawTableInner`], the result is [`Undefined Behavior`],
2775    /// if the `index` is greater than the `self.bucket_mask + 1 + Group::WIDTH`.
2776    /// In that case, calling this function with `index == self.bucket_mask + 1 + Group::WIDTH`
2777    /// will return a pointer to the end of the allocated table and it is useless on its own.
2778    ///
2779    /// Calling this function with `index >= self.bucket_mask + 1 + Group::WIDTH` on a
2780    /// table that has not been allocated results in [`Undefined Behavior`].
2781    ///
2782    /// So to satisfy both requirements you should always follow the rule that
2783    /// `index < self.bucket_mask + 1 + Group::WIDTH`
2784    ///
2785    /// Calling this function on [`RawTableInner`] that are not already allocated is safe
2786    /// for read-only purpose.
2787    ///
2788    /// See also [`Bucket::as_ptr()`] method, for more information about of properly removing
2789    /// or saving `data element` from / into the [`RawTable`] / [`RawTableInner`].
2790    ///
2791    /// [`Bucket::as_ptr()`]: Bucket::as_ptr()
2792    /// [`Undefined Behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2793    #[inline]
2794    unsafe fn ctrl(&self, index: usize) -> *mut u8 {
2795        debug_assert!(index < self.num_ctrl_bytes());
2796        // SAFETY: The caller must uphold the safety rules for the [`RawTableInner::ctrl`]
2797        self.ctrl.as_ptr().add(index)
2798    }
2799
2800    #[inline]
2801    fn buckets(&self) -> usize {
2802        self.bucket_mask + 1
2803    }
2804
2805    /// Checks whether the bucket at `index` is full.
2806    ///
2807    /// # Safety
2808    ///
2809    /// The caller must ensure `index` is less than the number of buckets.
2810    #[inline]
2811    unsafe fn is_bucket_full(&self, index: usize) -> bool {
2812        debug_assert!(index < self.buckets());
2813        is_full(*self.ctrl(index))
2814    }
2815
2816    #[inline]
2817    fn num_ctrl_bytes(&self) -> usize {
2818        self.bucket_mask + 1 + Group::WIDTH
2819    }
2820
2821    #[inline]
2822    fn is_empty_singleton(&self) -> bool {
2823        self.bucket_mask == 0
2824    }
2825
2826    /// Attempts to allocate a new hash table with at least enough capacity
2827    /// for inserting the given number of elements without reallocating,
2828    /// and return it inside ScopeGuard to protect against panic in the hash
2829    /// function.
2830    ///
2831    /// # Note
2832    ///
2833    /// It is recommended (but not required):
2834    ///
2835    /// * That the new table's `capacity` be greater than or equal to `self.items`.
2836    ///
2837    /// * The `alloc` is the same [`Allocator`] as the `Allocator` used
2838    ///   to allocate this table.
2839    ///
2840    /// * The `table_layout` is the same [`TableLayout`] as the `TableLayout` used
2841    ///   to allocate this table.
2842    ///
2843    /// If `table_layout` does not match the `TableLayout` that was used to allocate
2844    /// this table, then using `mem::swap` with the `self` and the new table returned
2845    /// by this function results in [`undefined behavior`].
2846    ///
2847    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2848    #[allow(clippy::mut_mut)]
2849    #[inline]
2850    fn prepare_resize<'a, A>(
2851        &self,
2852        alloc: &'a A,
2853        table_layout: TableLayout,
2854        capacity: usize,
2855        fallibility: Fallibility,
2856    ) -> Result<crate::scopeguard::ScopeGuard<Self, impl FnMut(&mut Self) + 'a>, TryReserveError>
2857    where
2858        A: Allocator,
2859    {
2860        debug_assert!(self.items <= capacity);
2861
2862        // Allocate and initialize the new table.
2863        let new_table =
2864            RawTableInner::fallible_with_capacity(alloc, table_layout, capacity, fallibility)?;
2865
2866        // The hash function may panic, in which case we simply free the new
2867        // table without dropping any elements that may have been copied into
2868        // it.
2869        //
2870        // This guard is also used to free the old table on success, see
2871        // the comment at the bottom of this function.
2872        Ok(guard(new_table, move |self_| {
2873            if !self_.is_empty_singleton() {
2874                // SAFETY:
2875                // 1. We have checked that our table is allocated.
2876                // 2. We know for sure that the `alloc` and `table_layout` matches the
2877                //    [`Allocator`] and [`TableLayout`] used to allocate this table.
2878                unsafe { self_.free_buckets(alloc, table_layout) };
2879            }
2880        }))
2881    }
2882
2883    /// Reserves or rehashes to make room for `additional` more elements.
2884    ///
2885    /// This uses dynamic dispatch to reduce the amount of
2886    /// code generated, but it is eliminated by LLVM optimizations when inlined.
2887    ///
2888    /// # Safety
2889    ///
2890    /// If any of the following conditions are violated, the result is
2891    /// [`undefined behavior`]:
2892    ///
2893    /// * The `alloc` must be the same [`Allocator`] as the `Allocator` used
2894    ///   to allocate this table.
2895    ///
2896    /// * The `layout` must be the same [`TableLayout`] as the `TableLayout`
2897    ///   used to allocate this table.
2898    ///
2899    /// * The `drop` function (`fn(*mut u8)`) must be the actual drop function of
2900    ///   the elements stored in the table.
2901    ///
2902    /// * The [`RawTableInner`] must have properly initialized control bytes.
2903    ///
2904    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
2905    #[allow(clippy::inline_always)]
2906    #[inline(always)]
2907    unsafe fn reserve_rehash_inner<A>(
2908        &mut self,
2909        alloc: &A,
2910        additional: usize,
2911        hasher: &dyn Fn(&mut Self, usize) -> u64,
2912        fallibility: Fallibility,
2913        layout: TableLayout,
2914        drop: Option<fn(*mut u8)>,
2915    ) -> Result<(), TryReserveError>
2916    where
2917        A: Allocator,
2918    {
2919        // Avoid `Option::ok_or_else` because it bloats LLVM IR.
2920        let new_items = match self.items.checked_add(additional) {
2921            Some(new_items) => new_items,
2922            None => return Err(fallibility.capacity_overflow()),
2923        };
2924        let full_capacity = bucket_mask_to_capacity(self.bucket_mask);
2925        if new_items <= full_capacity / 2 {
2926            // Rehash in-place without re-allocating if we have plenty of spare
2927            // capacity that is locked up due to DELETED entries.
2928
2929            // SAFETY:
2930            // 1. We know for sure that `[`RawTableInner`]` has already been allocated
2931            //    (since new_items <= full_capacity / 2);
2932            // 2. The caller ensures that `drop` function is the actual drop function of
2933            //    the elements stored in the table.
2934            // 3. The caller ensures that `layout` matches the [`TableLayout`] that was
2935            //    used to allocate this table.
2936            // 4. The caller ensures that the control bytes of the `RawTableInner`
2937            //    are already initialized.
2938            self.rehash_in_place(hasher, layout.size, drop);
2939            Ok(())
2940        } else {
2941            // Otherwise, conservatively resize to at least the next size up
2942            // to avoid churning deletes into frequent rehashes.
2943            //
2944            // SAFETY:
2945            // 1. We know for sure that `capacity >= self.items`.
2946            // 2. The caller ensures that `alloc` and `layout` matches the [`Allocator`] and
2947            //    [`TableLayout`] that were used to allocate this table.
2948            // 3. The caller ensures that the control bytes of the `RawTableInner`
2949            //    are already initialized.
2950            self.resize_inner(
2951                alloc,
2952                usize::max(new_items, full_capacity + 1),
2953                hasher,
2954                fallibility,
2955                layout,
2956            )
2957        }
2958    }
2959
2960    /// Returns an iterator over full buckets indices in the table.
2961    ///
2962    /// # Safety
2963    ///
2964    /// Behavior is undefined if any of the following conditions are violated:
2965    ///
2966    /// * The caller has to ensure that the `RawTableInner` outlives the
2967    ///   `FullBucketsIndices`. Because we cannot make the `next` method
2968    ///   unsafe on the `FullBucketsIndices` struct, we have to make the
2969    ///   `full_buckets_indices` method unsafe.
2970    ///
2971    /// * The [`RawTableInner`] must have properly initialized control bytes.
2972    #[inline(always)]
2973    unsafe fn full_buckets_indices(&self) -> FullBucketsIndices {
2974        // SAFETY:
2975        // 1. Since the caller of this function ensures that the control bytes
2976        //    are properly initialized and `self.ctrl(0)` points to the start
2977        //    of the array of control bytes, therefore: `ctrl` is valid for reads,
2978        //    properly aligned to `Group::WIDTH` and points to the properly initialized
2979        //    control bytes.
2980        // 2. The value of `items` is equal to the amount of data (values) added
2981        //    to the table.
2982        //
2983        //                         `ctrl` points here (to the start
2984        //                         of the first control byte `CT0`)
2985        //                          ∨
2986        // [Pad], T_n, ..., T1, T0, |CT0, CT1, ..., CT_n|, Group::WIDTH
2987        //                           \________  ________/
2988        //                                    \/
2989        //       `n = buckets - 1`, i.e. `RawTableInner::buckets() - 1`
2990        //
2991        // where: T0...T_n  - our stored data;
2992        //        CT0...CT_n - control bytes or metadata for `data`.
2993        let ctrl = NonNull::new_unchecked(self.ctrl(0));
2994
2995        FullBucketsIndices {
2996            // Load the first group
2997            // SAFETY: See explanation above.
2998            current_group: Group::load_aligned(ctrl.as_ptr()).match_full().into_iter(),
2999            group_first_index: 0,
3000            ctrl,
3001            items: self.items,
3002        }
3003    }
3004
3005    /// Allocates a new table of a different size and moves the contents of the
3006    /// current table into it.
3007    ///
3008    /// This uses dynamic dispatch to reduce the amount of
3009    /// code generated, but it is eliminated by LLVM optimizations when inlined.
3010    ///
3011    /// # Safety
3012    ///
3013    /// If any of the following conditions are violated, the result is
3014    /// [`undefined behavior`]:
3015    ///
3016    /// * The `alloc` must be the same [`Allocator`] as the `Allocator` used
3017    ///   to allocate this table;
3018    ///
3019    /// * The `layout` must be the same [`TableLayout`] as the `TableLayout`
3020    ///   used to allocate this table;
3021    ///
3022    /// * The [`RawTableInner`] must have properly initialized control bytes.
3023    ///
3024    /// The caller of this function must ensure that `capacity >= self.items`
3025    /// otherwise:
3026    ///
3027    /// * If `self.items != 0`, calling of this function with `capacity == 0`
3028    ///   results in [`undefined behavior`].
3029    ///
3030    /// * If `capacity_to_buckets(capacity) < Group::WIDTH` and
3031    ///   `self.items > capacity_to_buckets(capacity)` calling this function
3032    ///   results in [`undefined behavior`].
3033    ///
3034    /// * If `capacity_to_buckets(capacity) >= Group::WIDTH` and
3035    ///   `self.items > capacity_to_buckets(capacity)` calling this function
3036    ///   are never return (will go into an infinite loop).
3037    ///
3038    /// Note: It is recommended (but not required) that the new table's `capacity`
3039    /// be greater than or equal to `self.items`. In case if `capacity <= self.items`
3040    /// this function can never return. See [`RawTableInner::find_insert_slot`] for
3041    /// more information.
3042    ///
3043    /// [`RawTableInner::find_insert_slot`]: RawTableInner::find_insert_slot
3044    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
3045    #[allow(clippy::inline_always)]
3046    #[inline(always)]
3047    unsafe fn resize_inner<A>(
3048        &mut self,
3049        alloc: &A,
3050        capacity: usize,
3051        hasher: &dyn Fn(&mut Self, usize) -> u64,
3052        fallibility: Fallibility,
3053        layout: TableLayout,
3054    ) -> Result<(), TryReserveError>
3055    where
3056        A: Allocator,
3057    {
3058        // SAFETY: We know for sure that `alloc` and `layout` matches the [`Allocator`] and [`TableLayout`]
3059        // that were used to allocate this table.
3060        let mut new_table = self.prepare_resize(alloc, layout, capacity, fallibility)?;
3061
3062        // SAFETY: We know for sure that RawTableInner will outlive the
3063        // returned `FullBucketsIndices` iterator, and the caller of this
3064        // function ensures that the control bytes are properly initialized.
3065        for full_byte_index in self.full_buckets_indices() {
3066            // This may panic.
3067            let hash = hasher(self, full_byte_index);
3068
3069            // SAFETY:
3070            // We can use a simpler version of insert() here since:
3071            // 1. There are no DELETED entries.
3072            // 2. We know there is enough space in the table.
3073            // 3. All elements are unique.
3074            // 4. The caller of this function guarantees that `capacity > 0`
3075            //    so `new_table` must already have some allocated memory.
3076            // 5. We set `growth_left` and `items` fields of the new table
3077            //    after the loop.
3078            // 6. We insert into the table, at the returned index, the data
3079            //    matching the given hash immediately after calling this function.
3080            let (new_index, _) = new_table.prepare_insert_slot(hash);
3081
3082            // SAFETY:
3083            //
3084            // * `src` is valid for reads of `layout.size` bytes, since the
3085            //   table is alive and the `full_byte_index` is guaranteed to be
3086            //   within bounds (see `FullBucketsIndices::next_impl`);
3087            //
3088            // * `dst` is valid for writes of `layout.size` bytes, since the
3089            //   caller ensures that `table_layout` matches the [`TableLayout`]
3090            //   that was used to allocate old table and we have the `new_index`
3091            //   returned by `prepare_insert_slot`.
3092            //
3093            // * Both `src` and `dst` are properly aligned.
3094            //
3095            // * Both `src` and `dst` point to different region of memory.
3096            ptr::copy_nonoverlapping(
3097                self.bucket_ptr(full_byte_index, layout.size),
3098                new_table.bucket_ptr(new_index, layout.size),
3099                layout.size,
3100            );
3101        }
3102
3103        // The hash function didn't panic, so we can safely set the
3104        // `growth_left` and `items` fields of the new table.
3105        new_table.growth_left -= self.items;
3106        new_table.items = self.items;
3107
3108        // We successfully copied all elements without panicking. Now replace
3109        // self with the new table. The old table will have its memory freed but
3110        // the items will not be dropped (since they have been moved into the
3111        // new table).
3112        // SAFETY: The caller ensures that `table_layout` matches the [`TableLayout`]
3113        // that was used to allocate this table.
3114        mem::swap(self, &mut new_table);
3115
3116        Ok(())
3117    }
3118
3119    /// Rehashes the contents of the table in place (i.e. without changing the
3120    /// allocation).
3121    ///
3122    /// If `hasher` panics then some the table's contents may be lost.
3123    ///
3124    /// This uses dynamic dispatch to reduce the amount of
3125    /// code generated, but it is eliminated by LLVM optimizations when inlined.
3126    ///
3127    /// # Safety
3128    ///
3129    /// If any of the following conditions are violated, the result is [`undefined behavior`]:
3130    ///
3131    /// * The `size_of` must be equal to the size of the elements stored in the table;
3132    ///
3133    /// * The `drop` function (`fn(*mut u8)`) must be the actual drop function of
3134    ///   the elements stored in the table.
3135    ///
3136    /// * The [`RawTableInner`] has already been allocated;
3137    ///
3138    /// * The [`RawTableInner`] must have properly initialized control bytes.
3139    ///
3140    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
3141    #[allow(clippy::inline_always)]
3142    #[cfg_attr(feature = "inline-more", inline(always))]
3143    #[cfg_attr(not(feature = "inline-more"), inline)]
3144    unsafe fn rehash_in_place(
3145        &mut self,
3146        hasher: &dyn Fn(&mut Self, usize) -> u64,
3147        size_of: usize,
3148        drop: Option<fn(*mut u8)>,
3149    ) {
3150        // If the hash function panics then properly clean up any elements
3151        // that we haven't rehashed yet. We unfortunately can't preserve the
3152        // element since we lost their hash and have no way of recovering it
3153        // without risking another panic.
3154        self.prepare_rehash_in_place();
3155
3156        let mut guard = guard(self, move |self_| {
3157            if let Some(drop) = drop {
3158                for i in 0..self_.buckets() {
3159                    if *self_.ctrl(i) == DELETED {
3160                        self_.set_ctrl(i, EMPTY);
3161                        drop(self_.bucket_ptr(i, size_of));
3162                        self_.items -= 1;
3163                    }
3164                }
3165            }
3166            self_.growth_left = bucket_mask_to_capacity(self_.bucket_mask) - self_.items;
3167        });
3168
3169        // At this point, DELETED elements are elements that we haven't
3170        // rehashed yet. Find them and re-insert them at their ideal
3171        // position.
3172        'outer: for i in 0..guard.buckets() {
3173            if *guard.ctrl(i) != DELETED {
3174                continue;
3175            }
3176
3177            let i_p = guard.bucket_ptr(i, size_of);
3178
3179            'inner: loop {
3180                // Hash the current item
3181                let hash = hasher(*guard, i);
3182
3183                // Search for a suitable place to put it
3184                //
3185                // SAFETY: Caller of this function ensures that the control bytes
3186                // are properly initialized.
3187                let new_i = guard.find_insert_slot(hash).index;
3188
3189                // Probing works by scanning through all of the control
3190                // bytes in groups, which may not be aligned to the group
3191                // size. If both the new and old position fall within the
3192                // same unaligned group, then there is no benefit in moving
3193                // it and we can just continue to the next item.
3194                if likely(guard.is_in_same_group(i, new_i, hash)) {
3195                    guard.set_ctrl_h2(i, hash);
3196                    continue 'outer;
3197                }
3198
3199                let new_i_p = guard.bucket_ptr(new_i, size_of);
3200
3201                // We are moving the current item to a new position. Write
3202                // our H2 to the control byte of the new position.
3203                let prev_ctrl = guard.replace_ctrl_h2(new_i, hash);
3204                if prev_ctrl == EMPTY {
3205                    guard.set_ctrl(i, EMPTY);
3206                    // If the target slot is empty, simply move the current
3207                    // element into the new slot and clear the old control
3208                    // byte.
3209                    ptr::copy_nonoverlapping(i_p, new_i_p, size_of);
3210                    continue 'outer;
3211                } else {
3212                    // If the target slot is occupied, swap the two elements
3213                    // and then continue processing the element that we just
3214                    // swapped into the old slot.
3215                    debug_assert_eq!(prev_ctrl, DELETED);
3216                    ptr::swap_nonoverlapping(i_p, new_i_p, size_of);
3217                    continue 'inner;
3218                }
3219            }
3220        }
3221
3222        guard.growth_left = bucket_mask_to_capacity(guard.bucket_mask) - guard.items;
3223
3224        mem::forget(guard);
3225    }
3226
3227    /// Deallocates the table without dropping any entries.
3228    ///
3229    /// # Note
3230    ///
3231    /// This function must be called only after [`drop_elements`](RawTableInner::drop_elements),
3232    /// else it can lead to leaking of memory. Also calling this function automatically
3233    /// makes invalid (dangling) all instances of buckets ([`Bucket`]) and makes invalid
3234    /// (dangling) the `ctrl` field of the table.
3235    ///
3236    /// # Safety
3237    ///
3238    /// If any of the following conditions are violated, the result is [`Undefined Behavior`]:
3239    ///
3240    /// * The [`RawTableInner`] has already been allocated;
3241    ///
3242    /// * The `alloc` must be the same [`Allocator`] as the `Allocator` that was used
3243    ///   to allocate this table.
3244    ///
3245    /// * The `table_layout` must be the same [`TableLayout`] as the `TableLayout` that was used
3246    ///   to allocate this table.
3247    ///
3248    /// See also [`GlobalAlloc::dealloc`] or [`Allocator::deallocate`] for more  information.
3249    ///
3250    /// [`Undefined Behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
3251    /// [`GlobalAlloc::dealloc`]: https://doc.rust-lang.org/alloc/alloc/trait.GlobalAlloc.html#tymethod.dealloc
3252    /// [`Allocator::deallocate`]: https://doc.rust-lang.org/alloc/alloc/trait.Allocator.html#tymethod.deallocate
3253    #[inline]
3254    unsafe fn free_buckets<A>(&mut self, alloc: &A, table_layout: TableLayout)
3255    where
3256        A: Allocator,
3257    {
3258        // SAFETY: The caller must uphold the safety contract for `free_buckets`
3259        // method.
3260        let (ptr, layout) = self.allocation_info(table_layout);
3261        alloc.deallocate(ptr, layout);
3262    }
3263
3264    /// Returns a pointer to the allocated memory and the layout that was used to
3265    /// allocate the table.
3266    ///
3267    /// # Safety
3268    ///
3269    /// Caller of this function must observe the following safety rules:
3270    ///
3271    /// * The [`RawTableInner`] has already been allocated, otherwise
3272    ///   calling this function results in [`undefined behavior`]
3273    ///
3274    /// * The `table_layout` must be the same [`TableLayout`] as the `TableLayout`
3275    ///   that was used to allocate this table. Failure to comply with this condition
3276    ///   may result in [`undefined behavior`].
3277    ///
3278    /// See also [`GlobalAlloc::dealloc`] or [`Allocator::deallocate`] for more  information.
3279    ///
3280    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
3281    /// [`GlobalAlloc::dealloc`]: https://doc.rust-lang.org/alloc/alloc/trait.GlobalAlloc.html#tymethod.dealloc
3282    /// [`Allocator::deallocate`]: https://doc.rust-lang.org/alloc/alloc/trait.Allocator.html#tymethod.deallocate
3283    #[inline]
3284    unsafe fn allocation_info(&self, table_layout: TableLayout) -> (NonNull<u8>, Layout) {
3285        debug_assert!(
3286            !self.is_empty_singleton(),
3287            "this function can only be called on non-empty tables"
3288        );
3289
3290        // Avoid `Option::unwrap_or_else` because it bloats LLVM IR.
3291        let (layout, ctrl_offset) = match table_layout.calculate_layout_for(self.buckets()) {
3292            Some(lco) => lco,
3293            None => unsafe { hint::unreachable_unchecked() },
3294        };
3295        (
3296            // SAFETY: The caller must uphold the safety contract for `allocation_info` method.
3297            unsafe { NonNull::new_unchecked(self.ctrl.as_ptr().sub(ctrl_offset)) },
3298            layout,
3299        )
3300    }
3301
3302    /// Returns a pointer to the allocated memory and the layout that was used to
3303    /// allocate the table. If [`RawTableInner`] has not been allocated, this
3304    /// function return `dangling` pointer and `()` (unit) layout.
3305    ///
3306    /// # Safety
3307    ///
3308    /// The `table_layout` must be the same [`TableLayout`] as the `TableLayout`
3309    /// that was used to allocate this table. Failure to comply with this condition
3310    /// may result in [`undefined behavior`].
3311    ///
3312    /// See also [`GlobalAlloc::dealloc`] or [`Allocator::deallocate`] for more  information.
3313    ///
3314    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
3315    /// [`GlobalAlloc::dealloc`]: https://doc.rust-lang.org/alloc/alloc/trait.GlobalAlloc.html#tymethod.dealloc
3316    /// [`Allocator::deallocate`]: https://doc.rust-lang.org/alloc/alloc/trait.Allocator.html#tymethod.deallocate
3317    #[cfg(feature = "raw")]
3318    unsafe fn allocation_info_or_zero(&self, table_layout: TableLayout) -> (NonNull<u8>, Layout) {
3319        if self.is_empty_singleton() {
3320            (NonNull::dangling(), Layout::new::<()>())
3321        } else {
3322            // SAFETY:
3323            // 1. We have checked that our table is allocated.
3324            // 2. The caller ensures that `table_layout` matches the [`TableLayout`]
3325            // that was used to allocate this table.
3326            unsafe { self.allocation_info(table_layout) }
3327        }
3328    }
3329
3330    /// Marks all table buckets as empty without dropping their contents.
3331    #[inline]
3332    fn clear_no_drop(&mut self) {
3333        if !self.is_empty_singleton() {
3334            unsafe {
3335                self.ctrl(0).write_bytes(EMPTY, self.num_ctrl_bytes());
3336            }
3337        }
3338        self.items = 0;
3339        self.growth_left = bucket_mask_to_capacity(self.bucket_mask);
3340    }
3341
3342    /// Erases the [`Bucket`]'s control byte at the given index so that it does not
3343    /// triggered as full, decreases the `items` of the table and, if it can be done,
3344    /// increases `self.growth_left`.
3345    ///
3346    /// This function does not actually erase / drop the [`Bucket`] itself, i.e. it
3347    /// does not make any changes to the `data` parts of the table. The caller of this
3348    /// function must take care to properly drop the `data`, otherwise calling this
3349    /// function may result in a memory leak.
3350    ///
3351    /// # Safety
3352    ///
3353    /// You must observe the following safety rules when calling this function:
3354    ///
3355    /// * The [`RawTableInner`] has already been allocated;
3356    ///
3357    /// * It must be the full control byte at the given position;
3358    ///
3359    /// * The `index` must not be greater than the `RawTableInner.bucket_mask`, i.e.
3360    ///   `index <= RawTableInner.bucket_mask` or, in other words, `(index + 1)` must
3361    ///   be no greater than the number returned by the function [`RawTableInner::buckets`].
3362    ///
3363    /// Calling this function on a table that has not been allocated results in [`undefined behavior`].
3364    ///
3365    /// Calling this function on a table with no elements is unspecified, but calling subsequent
3366    /// functions is likely to result in [`undefined behavior`] due to overflow subtraction
3367    /// (`self.items -= 1 cause overflow when self.items == 0`).
3368    ///
3369    /// See also [`Bucket::as_ptr`] method, for more information about of properly removing
3370    /// or saving `data element` from / into the [`RawTable`] / [`RawTableInner`].
3371    ///
3372    /// [`RawTableInner::buckets`]: RawTableInner::buckets
3373    /// [`Bucket::as_ptr`]: Bucket::as_ptr
3374    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
3375    #[inline]
3376    unsafe fn erase(&mut self, index: usize) {
3377        debug_assert!(self.is_bucket_full(index));
3378
3379        // This is the same as `index.wrapping_sub(Group::WIDTH) % self.buckets()` because
3380        // the number of buckets is a power of two, and `self.bucket_mask = self.buckets() - 1`.
3381        let index_before = index.wrapping_sub(Group::WIDTH) & self.bucket_mask;
3382        // SAFETY:
3383        // - The caller must uphold the safety contract for `erase` method;
3384        // - `index_before` is guaranteed to be in range due to masking with `self.bucket_mask`
3385        let empty_before = Group::load(self.ctrl(index_before)).match_empty();
3386        let empty_after = Group::load(self.ctrl(index)).match_empty();
3387
3388        // Inserting and searching in the map is performed by two key functions:
3389        //
3390        // - The `find_insert_slot` function that looks up the index of any `EMPTY` or `DELETED`
3391        //   slot in a group to be able to insert. If it doesn't find an `EMPTY` or `DELETED`
3392        //   slot immediately in the first group, it jumps to the next `Group` looking for it,
3393        //   and so on until it has gone through all the groups in the control bytes.
3394        //
3395        // - The `find_inner` function that looks for the index of the desired element by looking
3396        //   at all the `FULL` bytes in the group. If it did not find the element right away, and
3397        //   there is no `EMPTY` byte in the group, then this means that the `find_insert_slot`
3398        //   function may have found a suitable slot in the next group. Therefore, `find_inner`
3399        //   jumps further, and if it does not find the desired element and again there is no `EMPTY`
3400        //   byte, then it jumps further, and so on. The search stops only if `find_inner` function
3401        //   finds the desired element or hits an `EMPTY` slot/byte.
3402        //
3403        // Accordingly, this leads to two consequences:
3404        //
3405        // - The map must have `EMPTY` slots (bytes);
3406        //
3407        // - You can't just mark the byte to be erased as `EMPTY`, because otherwise the `find_inner`
3408        //   function may stumble upon an `EMPTY` byte before finding the desired element and stop
3409        //   searching.
3410        //
3411        // Thus it is necessary to check all bytes after and before the erased element. If we are in
3412        // a contiguous `Group` of `FULL` or `DELETED` bytes (the number of `FULL` or `DELETED` bytes
3413        // before and after is greater than or equal to `Group::WIDTH`), then we must mark our byte as
3414        // `DELETED` in order for the `find_inner` function to go further. On the other hand, if there
3415        // is at least one `EMPTY` slot in the `Group`, then the `find_inner` function will still stumble
3416        // upon an `EMPTY` byte, so we can safely mark our erased byte as `EMPTY` as well.
3417        //
3418        // Finally, since `index_before == (index.wrapping_sub(Group::WIDTH) & self.bucket_mask) == index`
3419        // and given all of the above, tables smaller than the group width (self.buckets() < Group::WIDTH)
3420        // cannot have `DELETED` bytes.
3421        //
3422        // Note that in this context `leading_zeros` refers to the bytes at the end of a group, while
3423        // `trailing_zeros` refers to the bytes at the beginning of a group.
3424        let ctrl = if empty_before.leading_zeros() + empty_after.trailing_zeros() >= Group::WIDTH {
3425            DELETED
3426        } else {
3427            self.growth_left += 1;
3428            EMPTY
3429        };
3430        // SAFETY: the caller must uphold the safety contract for `erase` method.
3431        self.set_ctrl(index, ctrl);
3432        self.items -= 1;
3433    }
3434}
3435
3436impl<T: Clone, A: Allocator + Clone> Clone for RawTable<T, A> {
3437    fn clone(&self) -> Self {
3438        if self.table.is_empty_singleton() {
3439            Self::new_in(self.alloc.clone())
3440        } else {
3441            unsafe {
3442                // Avoid `Result::ok_or_else` because it bloats LLVM IR.
3443                //
3444                // SAFETY: This is safe as we are taking the size of an already allocated table
3445                // and therefore сapacity overflow cannot occur, `self.table.buckets()` is power
3446                // of two and all allocator errors will be caught inside `RawTableInner::new_uninitialized`.
3447                let mut new_table = match Self::new_uninitialized(
3448                    self.alloc.clone(),
3449                    self.table.buckets(),
3450                    Fallibility::Infallible,
3451                ) {
3452                    Ok(table) => table,
3453                    Err(_) => hint::unreachable_unchecked(),
3454                };
3455
3456                // Cloning elements may fail (the clone function may panic). But we don't
3457                // need to worry about uninitialized control bits, since:
3458                // 1. The number of items (elements) in the table is zero, which means that
3459                //    the control bits will not be readed by Drop function.
3460                // 2. The `clone_from_spec` method will first copy all control bits from
3461                //    `self` (thus initializing them). But this will not affect the `Drop`
3462                //    function, since the `clone_from_spec` function sets `items` only after
3463                //    successfully clonning all elements.
3464                new_table.clone_from_spec(self);
3465                new_table
3466            }
3467        }
3468    }
3469
3470    fn clone_from(&mut self, source: &Self) {
3471        if source.table.is_empty_singleton() {
3472            let mut old_inner = mem::replace(&mut self.table, RawTableInner::NEW);
3473            unsafe {
3474                // SAFETY:
3475                // 1. We call the function only once;
3476                // 2. We know for sure that `alloc` and `table_layout` matches the [`Allocator`]
3477                //    and [`TableLayout`] that were used to allocate this table.
3478                // 3. If any elements' drop function panics, then there will only be a memory leak,
3479                //    because we have replaced the inner table with a new one.
3480                old_inner.drop_inner_table::<T, _>(&self.alloc, Self::TABLE_LAYOUT);
3481            }
3482        } else {
3483            unsafe {
3484                // Make sure that if any panics occurs, we clear the table and
3485                // leave it in an empty state.
3486                let mut self_ = guard(self, |self_| {
3487                    self_.clear_no_drop();
3488                });
3489
3490                // First, drop all our elements without clearing the control
3491                // bytes. If this panics then the scope guard will clear the
3492                // table, leaking any elements that were not dropped yet.
3493                //
3494                // This leak is unavoidable: we can't try dropping more elements
3495                // since this could lead to another panic and abort the process.
3496                //
3497                // SAFETY: If something gets wrong we clear our table right after
3498                // dropping the elements, so there is no double drop, since `items`
3499                // will be equal to zero.
3500                self_.table.drop_elements::<T>();
3501
3502                // If necessary, resize our table to match the source.
3503                if self_.buckets() != source.buckets() {
3504                    let new_inner = match RawTableInner::new_uninitialized(
3505                        &self_.alloc,
3506                        Self::TABLE_LAYOUT,
3507                        source.buckets(),
3508                        Fallibility::Infallible,
3509                    ) {
3510                        Ok(table) => table,
3511                        Err(_) => hint::unreachable_unchecked(),
3512                    };
3513                    // Replace the old inner with new uninitialized one. It's ok, since if something gets
3514                    // wrong `ScopeGuard` will initialize all control bytes and leave empty table.
3515                    let mut old_inner = mem::replace(&mut self_.table, new_inner);
3516                    if !old_inner.is_empty_singleton() {
3517                        // SAFETY:
3518                        // 1. We have checked that our table is allocated.
3519                        // 2. We know for sure that `alloc` and `table_layout` matches
3520                        // the [`Allocator`] and [`TableLayout`] that were used to allocate this table.
3521                        old_inner.free_buckets(&self_.alloc, Self::TABLE_LAYOUT);
3522                    }
3523                }
3524
3525                // Cloning elements may fail (the clone function may panic), but the `ScopeGuard`
3526                // inside the `clone_from_impl` function will take care of that, dropping all
3527                // cloned elements if necessary. Our `ScopeGuard` will clear the table.
3528                self_.clone_from_spec(source);
3529
3530                // Disarm the scope guard if cloning was successful.
3531                ScopeGuard::into_inner(self_);
3532            }
3533        }
3534    }
3535}
3536
3537/// Specialization of `clone_from` for `Copy` types
3538trait RawTableClone {
3539    unsafe fn clone_from_spec(&mut self, source: &Self);
3540}
3541impl<T: Clone, A: Allocator + Clone> RawTableClone for RawTable<T, A> {
3542    default_fn! {
3543        #[cfg_attr(feature = "inline-more", inline)]
3544        unsafe fn clone_from_spec(&mut self, source: &Self) {
3545            self.clone_from_impl(source);
3546        }
3547    }
3548}
3549#[cfg(feature = "nightly")]
3550impl<T: Copy, A: Allocator + Clone> RawTableClone for RawTable<T, A> {
3551    #[cfg_attr(feature = "inline-more", inline)]
3552    unsafe fn clone_from_spec(&mut self, source: &Self) {
3553        source
3554            .table
3555            .ctrl(0)
3556            .copy_to_nonoverlapping(self.table.ctrl(0), self.table.num_ctrl_bytes());
3557        source
3558            .data_start()
3559            .as_ptr()
3560            .copy_to_nonoverlapping(self.data_start().as_ptr(), self.table.buckets());
3561
3562        self.table.items = source.table.items;
3563        self.table.growth_left = source.table.growth_left;
3564    }
3565}
3566
3567impl<T: Clone, A: Allocator + Clone> RawTable<T, A> {
3568    /// Common code for clone and clone_from. Assumes:
3569    /// - `self.buckets() == source.buckets()`.
3570    /// - Any existing elements have been dropped.
3571    /// - The control bytes are not initialized yet.
3572    #[cfg_attr(feature = "inline-more", inline)]
3573    unsafe fn clone_from_impl(&mut self, source: &Self) {
3574        // Copy the control bytes unchanged. We do this in a single pass
3575        source
3576            .table
3577            .ctrl(0)
3578            .copy_to_nonoverlapping(self.table.ctrl(0), self.table.num_ctrl_bytes());
3579
3580        // The cloning of elements may panic, in which case we need
3581        // to make sure we drop only the elements that have been
3582        // cloned so far.
3583        let mut guard = guard((0, &mut *self), |(index, self_)| {
3584            if T::NEEDS_DROP {
3585                for i in 0..*index {
3586                    if self_.is_bucket_full(i) {
3587                        self_.bucket(i).drop();
3588                    }
3589                }
3590            }
3591        });
3592
3593        for from in source.iter() {
3594            let index = source.bucket_index(&from);
3595            let to = guard.1.bucket(index);
3596            to.write(from.as_ref().clone());
3597
3598            // Update the index in case we need to unwind.
3599            guard.0 = index + 1;
3600        }
3601
3602        // Successfully cloned all items, no need to clean up.
3603        mem::forget(guard);
3604
3605        self.table.items = source.table.items;
3606        self.table.growth_left = source.table.growth_left;
3607    }
3608
3609    /// Variant of `clone_from` to use when a hasher is available.
3610    #[cfg(feature = "raw")]
3611    pub fn clone_from_with_hasher(&mut self, source: &Self, hasher: impl Fn(&T) -> u64) {
3612        // If we have enough capacity in the table, just clear it and insert
3613        // elements one by one. We don't do this if we have the same number of
3614        // buckets as the source since we can just copy the contents directly
3615        // in that case.
3616        if self.table.buckets() != source.table.buckets()
3617            && bucket_mask_to_capacity(self.table.bucket_mask) >= source.len()
3618        {
3619            self.clear();
3620
3621            let mut guard_self = guard(&mut *self, |self_| {
3622                // Clear the partially copied table if a panic occurs, otherwise
3623                // items and growth_left will be out of sync with the contents
3624                // of the table.
3625                self_.clear();
3626            });
3627
3628            unsafe {
3629                for item in source.iter() {
3630                    // This may panic.
3631                    let item = item.as_ref().clone();
3632                    let hash = hasher(&item);
3633
3634                    // We can use a simpler version of insert() here since:
3635                    // - there are no DELETED entries.
3636                    // - we know there is enough space in the table.
3637                    // - all elements are unique.
3638                    let (index, _) = guard_self.table.prepare_insert_slot(hash);
3639                    guard_self.bucket(index).write(item);
3640                }
3641            }
3642
3643            // Successfully cloned all items, no need to clean up.
3644            mem::forget(guard_self);
3645
3646            self.table.items = source.table.items;
3647            self.table.growth_left -= source.table.items;
3648        } else {
3649            self.clone_from(source);
3650        }
3651    }
3652}
3653
3654impl<T, A: Allocator + Default> Default for RawTable<T, A> {
3655    #[inline]
3656    fn default() -> Self {
3657        Self::new_in(Default::default())
3658    }
3659}
3660
3661#[cfg(feature = "nightly")]
3662unsafe impl<#[may_dangle] T, A: Allocator> Drop for RawTable<T, A> {
3663    #[cfg_attr(feature = "inline-more", inline)]
3664    fn drop(&mut self) {
3665        unsafe {
3666            // SAFETY:
3667            // 1. We call the function only once;
3668            // 2. We know for sure that `alloc` and `table_layout` matches the [`Allocator`]
3669            //    and [`TableLayout`] that were used to allocate this table.
3670            // 3. If the drop function of any elements fails, then only a memory leak will occur,
3671            //    and we don't care because we are inside the `Drop` function of the `RawTable`,
3672            //    so there won't be any table left in an inconsistent state.
3673            self.table
3674                .drop_inner_table::<T, _>(&self.alloc, Self::TABLE_LAYOUT);
3675        }
3676    }
3677}
3678#[cfg(not(feature = "nightly"))]
3679impl<T, A: Allocator> Drop for RawTable<T, A> {
3680    #[cfg_attr(feature = "inline-more", inline)]
3681    fn drop(&mut self) {
3682        unsafe {
3683            // SAFETY:
3684            // 1. We call the function only once;
3685            // 2. We know for sure that `alloc` and `table_layout` matches the [`Allocator`]
3686            //    and [`TableLayout`] that were used to allocate this table.
3687            // 3. If the drop function of any elements fails, then only a memory leak will occur,
3688            //    and we don't care because we are inside the `Drop` function of the `RawTable`,
3689            //    so there won't be any table left in an inconsistent state.
3690            self.table
3691                .drop_inner_table::<T, _>(&self.alloc, Self::TABLE_LAYOUT);
3692        }
3693    }
3694}
3695
3696impl<T, A: Allocator> IntoIterator for RawTable<T, A> {
3697    type Item = T;
3698    type IntoIter = RawIntoIter<T, A>;
3699
3700    #[cfg_attr(feature = "inline-more", inline)]
3701    fn into_iter(self) -> RawIntoIter<T, A> {
3702        unsafe {
3703            let iter = self.iter();
3704            self.into_iter_from(iter)
3705        }
3706    }
3707}
3708
3709/// Iterator over a sub-range of a table. Unlike `RawIter` this iterator does
3710/// not track an item count.
3711pub(crate) struct RawIterRange<T> {
3712    // Mask of full buckets in the current group. Bits are cleared from this
3713    // mask as each element is processed.
3714    current_group: BitMaskIter,
3715
3716    // Pointer to the buckets for the current group.
3717    data: Bucket<T>,
3718
3719    // Pointer to the next group of control bytes,
3720    // Must be aligned to the group size.
3721    next_ctrl: *const u8,
3722
3723    // Pointer one past the last control byte of this range.
3724    end: *const u8,
3725}
3726
3727impl<T> RawIterRange<T> {
3728    /// Returns a `RawIterRange` covering a subset of a table.
3729    ///
3730    /// # Safety
3731    ///
3732    /// If any of the following conditions are violated, the result is
3733    /// [`undefined behavior`]:
3734    ///
3735    /// * `ctrl` must be [valid] for reads, i.e. table outlives the `RawIterRange`;
3736    ///
3737    /// * `ctrl` must be properly aligned to the group size (Group::WIDTH);
3738    ///
3739    /// * `ctrl` must point to the array of properly initialized control bytes;
3740    ///
3741    /// * `data` must be the [`Bucket`] at the `ctrl` index in the table;
3742    ///
3743    /// * the value of `len` must be less than or equal to the number of table buckets,
3744    ///   and the returned value of `ctrl.as_ptr().add(len).offset_from(ctrl.as_ptr())`
3745    ///   must be positive.
3746    ///
3747    /// * The `ctrl.add(len)` pointer must be either in bounds or one
3748    ///   byte past the end of the same [allocated table].
3749    ///
3750    /// * The `len` must be a power of two.
3751    ///
3752    /// [valid]: https://doc.rust-lang.org/std/ptr/index.html#safety
3753    /// [`undefined behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
3754    #[cfg_attr(feature = "inline-more", inline)]
3755    unsafe fn new(ctrl: *const u8, data: Bucket<T>, len: usize) -> Self {
3756        debug_assert_ne!(len, 0);
3757        debug_assert_eq!(ctrl as usize % Group::WIDTH, 0);
3758        // SAFETY: The caller must uphold the safety rules for the [`RawIterRange::new`]
3759        let end = ctrl.add(len);
3760
3761        // Load the first group and advance ctrl to point to the next group
3762        // SAFETY: The caller must uphold the safety rules for the [`RawIterRange::new`]
3763        let current_group = Group::load_aligned(ctrl).match_full();
3764        let next_ctrl = ctrl.add(Group::WIDTH);
3765
3766        Self {
3767            current_group: current_group.into_iter(),
3768            data,
3769            next_ctrl,
3770            end,
3771        }
3772    }
3773
3774    /// Splits a `RawIterRange` into two halves.
3775    ///
3776    /// Returns `None` if the remaining range is smaller than or equal to the
3777    /// group width.
3778    #[cfg_attr(feature = "inline-more", inline)]
3779    #[cfg(feature = "rayon")]
3780    pub(crate) fn split(mut self) -> (Self, Option<RawIterRange<T>>) {
3781        unsafe {
3782            if self.end <= self.next_ctrl {
3783                // Nothing to split if the group that we are current processing
3784                // is the last one.
3785                (self, None)
3786            } else {
3787                // len is the remaining number of elements after the group that
3788                // we are currently processing. It must be a multiple of the
3789                // group size (small tables are caught by the check above).
3790                let len = offset_from(self.end, self.next_ctrl);
3791                debug_assert_eq!(len % Group::WIDTH, 0);
3792
3793                // Split the remaining elements into two halves, but round the
3794                // midpoint down in case there is an odd number of groups
3795                // remaining. This ensures that:
3796                // - The tail is at least 1 group long.
3797                // - The split is roughly even considering we still have the
3798                //   current group to process.
3799                let mid = (len / 2) & !(Group::WIDTH - 1);
3800
3801                let tail = Self::new(
3802                    self.next_ctrl.add(mid),
3803                    self.data.next_n(Group::WIDTH).next_n(mid),
3804                    len - mid,
3805                );
3806                debug_assert_eq!(
3807                    self.data.next_n(Group::WIDTH).next_n(mid).ptr,
3808                    tail.data.ptr
3809                );
3810                debug_assert_eq!(self.end, tail.end);
3811                self.end = self.next_ctrl.add(mid);
3812                debug_assert_eq!(self.end.add(Group::WIDTH), tail.next_ctrl);
3813                (self, Some(tail))
3814            }
3815        }
3816    }
3817
3818    /// # Safety
3819    /// If DO_CHECK_PTR_RANGE is false, caller must ensure that we never try to iterate
3820    /// after yielding all elements.
3821    #[cfg_attr(feature = "inline-more", inline)]
3822    unsafe fn next_impl<const DO_CHECK_PTR_RANGE: bool>(&mut self) -> Option<Bucket<T>> {
3823        loop {
3824            if let Some(index) = self.current_group.next() {
3825                return Some(self.data.next_n(index));
3826            }
3827
3828            if DO_CHECK_PTR_RANGE && self.next_ctrl >= self.end {
3829                return None;
3830            }
3831
3832            // We might read past self.end up to the next group boundary,
3833            // but this is fine because it only occurs on tables smaller
3834            // than the group size where the trailing control bytes are all
3835            // EMPTY. On larger tables self.end is guaranteed to be aligned
3836            // to the group size (since tables are power-of-two sized).
3837            self.current_group = Group::load_aligned(self.next_ctrl).match_full().into_iter();
3838            self.data = self.data.next_n(Group::WIDTH);
3839            self.next_ctrl = self.next_ctrl.add(Group::WIDTH);
3840        }
3841    }
3842
3843    /// Folds every element into an accumulator by applying an operation,
3844    /// returning the final result.
3845    ///
3846    /// `fold_impl()` takes three arguments: the number of items remaining in
3847    /// the iterator, an initial value, and a closure with two arguments: an
3848    /// 'accumulator', and an element. The closure returns the value that the
3849    /// accumulator should have for the next iteration.
3850    ///
3851    /// The initial value is the value the accumulator will have on the first call.
3852    ///
3853    /// After applying this closure to every element of the iterator, `fold_impl()`
3854    /// returns the accumulator.
3855    ///
3856    /// # Safety
3857    ///
3858    /// If any of the following conditions are violated, the result is
3859    /// [`Undefined Behavior`]:
3860    ///
3861    /// * The [`RawTableInner`] / [`RawTable`] must be alive and not moved,
3862    ///   i.e. table outlives the `RawIterRange`;
3863    ///
3864    /// * The provided `n` value must match the actual number of items
3865    ///   in the table.
3866    ///
3867    /// [`Undefined Behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
3868    #[allow(clippy::while_let_on_iterator)]
3869    #[cfg_attr(feature = "inline-more", inline)]
3870    unsafe fn fold_impl<F, B>(mut self, mut n: usize, mut acc: B, mut f: F) -> B
3871    where
3872        F: FnMut(B, Bucket<T>) -> B,
3873    {
3874        loop {
3875            while let Some(index) = self.current_group.next() {
3876                // The returned `index` will always be in the range `0..Group::WIDTH`,
3877                // so that calling `self.data.next_n(index)` is safe (see detailed explanation below).
3878                debug_assert!(n != 0);
3879                let bucket = self.data.next_n(index);
3880                acc = f(acc, bucket);
3881                n -= 1;
3882            }
3883
3884            if n == 0 {
3885                return acc;
3886            }
3887
3888            // SAFETY: The caller of this function ensures that:
3889            //
3890            // 1. The provided `n` value matches the actual number of items in the table;
3891            // 2. The table is alive and did not moved.
3892            //
3893            // Taking the above into account, we always stay within the bounds, because:
3894            //
3895            // 1. For tables smaller than the group width (self.buckets() <= Group::WIDTH),
3896            //    we will never end up in the given branch, since we should have already
3897            //    yielded all the elements of the table.
3898            //
3899            // 2. For tables larger than the group width. The number of buckets is a
3900            //    power of two (2 ^ n), Group::WIDTH is also power of two (2 ^ k). Since
3901            //    `(2 ^ n) > (2 ^ k)`, than `(2 ^ n) % (2 ^ k) = 0`. As we start from the
3902            //    start of the array of control bytes, and never try to iterate after
3903            //    getting all the elements, the last `self.current_group` will read bytes
3904            //    from the `self.buckets() - Group::WIDTH` index.  We know also that
3905            //    `self.current_group.next()` will always retun indices within the range
3906            //    `0..Group::WIDTH`.
3907            //
3908            //    Knowing all of the above and taking into account that we are synchronizing
3909            //    the `self.data` index with the index we used to read the `self.current_group`,
3910            //    the subsequent `self.data.next_n(index)` will always return a bucket with
3911            //    an index number less than `self.buckets()`.
3912            //
3913            //    The last `self.next_ctrl`, whose index would be `self.buckets()`, will never
3914            //    actually be read, since we should have already yielded all the elements of
3915            //    the table.
3916            self.current_group = Group::load_aligned(self.next_ctrl).match_full().into_iter();
3917            self.data = self.data.next_n(Group::WIDTH);
3918            self.next_ctrl = self.next_ctrl.add(Group::WIDTH);
3919        }
3920    }
3921}
3922
3923// We make raw iterators unconditionally Send and Sync, and let the PhantomData
3924// in the actual iterator implementations determine the real Send/Sync bounds.
3925unsafe impl<T> Send for RawIterRange<T> {}
3926unsafe impl<T> Sync for RawIterRange<T> {}
3927
3928impl<T> Clone for RawIterRange<T> {
3929    #[cfg_attr(feature = "inline-more", inline)]
3930    fn clone(&self) -> Self {
3931        Self {
3932            data: self.data.clone(),
3933            next_ctrl: self.next_ctrl,
3934            current_group: self.current_group,
3935            end: self.end,
3936        }
3937    }
3938}
3939
3940impl<T> Iterator for RawIterRange<T> {
3941    type Item = Bucket<T>;
3942
3943    #[cfg_attr(feature = "inline-more", inline)]
3944    fn next(&mut self) -> Option<Bucket<T>> {
3945        unsafe {
3946            // SAFETY: We set checker flag to true.
3947            self.next_impl::<true>()
3948        }
3949    }
3950
3951    #[inline]
3952    fn size_hint(&self) -> (usize, Option<usize>) {
3953        // We don't have an item count, so just guess based on the range size.
3954        let remaining_buckets = if self.end > self.next_ctrl {
3955            unsafe { offset_from(self.end, self.next_ctrl) }
3956        } else {
3957            0
3958        };
3959
3960        // Add a group width to include the group we are currently processing.
3961        (0, Some(Group::WIDTH + remaining_buckets))
3962    }
3963}
3964
3965impl<T> FusedIterator for RawIterRange<T> {}
3966
3967/// Iterator which returns a raw pointer to every full bucket in the table.
3968///
3969/// For maximum flexibility this iterator is not bound by a lifetime, but you
3970/// must observe several rules when using it:
3971/// - You must not free the hash table while iterating (including via growing/shrinking).
3972/// - It is fine to erase a bucket that has been yielded by the iterator.
3973/// - Erasing a bucket that has not yet been yielded by the iterator may still
3974///   result in the iterator yielding that bucket (unless `reflect_remove` is called).
3975/// - It is unspecified whether an element inserted after the iterator was
3976///   created will be yielded by that iterator (unless `reflect_insert` is called).
3977/// - The order in which the iterator yields bucket is unspecified and may
3978///   change in the future.
3979pub struct RawIter<T> {
3980    pub(crate) iter: RawIterRange<T>,
3981    items: usize,
3982}
3983
3984impl<T> RawIter<T> {
3985    /// Refresh the iterator so that it reflects a removal from the given bucket.
3986    ///
3987    /// For the iterator to remain valid, this method must be called once
3988    /// for each removed bucket before `next` is called again.
3989    ///
3990    /// This method should be called _before_ the removal is made. It is not necessary to call this
3991    /// method if you are removing an item that this iterator yielded in the past.
3992    #[cfg(feature = "raw")]
3993    pub unsafe fn reflect_remove(&mut self, b: &Bucket<T>) {
3994        self.reflect_toggle_full(b, false);
3995    }
3996
3997    /// Refresh the iterator so that it reflects an insertion into the given bucket.
3998    ///
3999    /// For the iterator to remain valid, this method must be called once
4000    /// for each insert before `next` is called again.
4001    ///
4002    /// This method does not guarantee that an insertion of a bucket with a greater
4003    /// index than the last one yielded will be reflected in the iterator.
4004    ///
4005    /// This method should be called _after_ the given insert is made.
4006    #[cfg(feature = "raw")]
4007    pub unsafe fn reflect_insert(&mut self, b: &Bucket<T>) {
4008        self.reflect_toggle_full(b, true);
4009    }
4010
4011    /// Refresh the iterator so that it reflects a change to the state of the given bucket.
4012    #[cfg(feature = "raw")]
4013    unsafe fn reflect_toggle_full(&mut self, b: &Bucket<T>, is_insert: bool) {
4014        if b.as_ptr() > self.iter.data.as_ptr() {
4015            // The iterator has already passed the bucket's group.
4016            // So the toggle isn't relevant to this iterator.
4017            return;
4018        }
4019
4020        if self.iter.next_ctrl < self.iter.end
4021            && b.as_ptr() <= self.iter.data.next_n(Group::WIDTH).as_ptr()
4022        {
4023            // The iterator has not yet reached the bucket's group.
4024            // We don't need to reload anything, but we do need to adjust the item count.
4025
4026            if cfg!(debug_assertions) {
4027                // Double-check that the user isn't lying to us by checking the bucket state.
4028                // To do that, we need to find its control byte. We know that self.iter.data is
4029                // at self.iter.next_ctrl - Group::WIDTH, so we work from there:
4030                let offset = offset_from(self.iter.data.as_ptr(), b.as_ptr());
4031                let ctrl = self.iter.next_ctrl.sub(Group::WIDTH).add(offset);
4032                // This method should be called _before_ a removal, or _after_ an insert,
4033                // so in both cases the ctrl byte should indicate that the bucket is full.
4034                assert!(is_full(*ctrl));
4035            }
4036
4037            if is_insert {
4038                self.items += 1;
4039            } else {
4040                self.items -= 1;
4041            }
4042
4043            return;
4044        }
4045
4046        // The iterator is at the bucket group that the toggled bucket is in.
4047        // We need to do two things:
4048        //
4049        //  - Determine if the iterator already yielded the toggled bucket.
4050        //    If it did, we're done.
4051        //  - Otherwise, update the iterator cached group so that it won't
4052        //    yield a to-be-removed bucket, or _will_ yield a to-be-added bucket.
4053        //    We'll also need to update the item count accordingly.
4054        if let Some(index) = self.iter.current_group.0.lowest_set_bit() {
4055            let next_bucket = self.iter.data.next_n(index);
4056            if b.as_ptr() > next_bucket.as_ptr() {
4057                // The toggled bucket is "before" the bucket the iterator would yield next. We
4058                // therefore don't need to do anything --- the iterator has already passed the
4059                // bucket in question.
4060                //
4061                // The item count must already be correct, since a removal or insert "prior" to
4062                // the iterator's position wouldn't affect the item count.
4063            } else {
4064                // The removed bucket is an upcoming bucket. We need to make sure it does _not_
4065                // get yielded, and also that it's no longer included in the item count.
4066                //
4067                // NOTE: We can't just reload the group here, both since that might reflect
4068                // inserts we've already passed, and because that might inadvertently unset the
4069                // bits for _other_ removals. If we do that, we'd have to also decrement the
4070                // item count for those other bits that we unset. But the presumably subsequent
4071                // call to reflect for those buckets might _also_ decrement the item count.
4072                // Instead, we _just_ flip the bit for the particular bucket the caller asked
4073                // us to reflect.
4074                let our_bit = offset_from(self.iter.data.as_ptr(), b.as_ptr());
4075                let was_full = self.iter.current_group.flip(our_bit);
4076                debug_assert_ne!(was_full, is_insert);
4077
4078                if is_insert {
4079                    self.items += 1;
4080                } else {
4081                    self.items -= 1;
4082                }
4083
4084                if cfg!(debug_assertions) {
4085                    if b.as_ptr() == next_bucket.as_ptr() {
4086                        // The removed bucket should no longer be next
4087                        debug_assert_ne!(self.iter.current_group.0.lowest_set_bit(), Some(index));
4088                    } else {
4089                        // We should not have changed what bucket comes next.
4090                        debug_assert_eq!(self.iter.current_group.0.lowest_set_bit(), Some(index));
4091                    }
4092                }
4093            }
4094        } else {
4095            // We must have already iterated past the removed item.
4096        }
4097    }
4098
4099    unsafe fn drop_elements(&mut self) {
4100        if T::NEEDS_DROP && self.items != 0 {
4101            for item in self {
4102                item.drop();
4103            }
4104        }
4105    }
4106}
4107
4108impl<T> Clone for RawIter<T> {
4109    #[cfg_attr(feature = "inline-more", inline)]
4110    fn clone(&self) -> Self {
4111        Self {
4112            iter: self.iter.clone(),
4113            items: self.items,
4114        }
4115    }
4116}
4117
4118impl<T> Iterator for RawIter<T> {
4119    type Item = Bucket<T>;
4120
4121    #[cfg_attr(feature = "inline-more", inline)]
4122    fn next(&mut self) -> Option<Bucket<T>> {
4123        // Inner iterator iterates over buckets
4124        // so it can do unnecessary work if we already yielded all items.
4125        if self.items == 0 {
4126            return None;
4127        }
4128
4129        let nxt = unsafe {
4130            // SAFETY: We check number of items to yield using `items` field.
4131            self.iter.next_impl::<false>()
4132        };
4133
4134        debug_assert!(nxt.is_some());
4135        self.items -= 1;
4136
4137        nxt
4138    }
4139
4140    #[inline]
4141    fn size_hint(&self) -> (usize, Option<usize>) {
4142        (self.items, Some(self.items))
4143    }
4144
4145    #[inline]
4146    fn fold<B, F>(self, init: B, f: F) -> B
4147    where
4148        Self: Sized,
4149        F: FnMut(B, Self::Item) -> B,
4150    {
4151        unsafe { self.iter.fold_impl(self.items, init, f) }
4152    }
4153}
4154
4155impl<T> ExactSizeIterator for RawIter<T> {}
4156impl<T> FusedIterator for RawIter<T> {}
4157
4158/// Iterator which returns an index of every full bucket in the table.
4159///
4160/// For maximum flexibility this iterator is not bound by a lifetime, but you
4161/// must observe several rules when using it:
4162/// - You must not free the hash table while iterating (including via growing/shrinking).
4163/// - It is fine to erase a bucket that has been yielded by the iterator.
4164/// - Erasing a bucket that has not yet been yielded by the iterator may still
4165///   result in the iterator yielding index of that bucket.
4166/// - It is unspecified whether an element inserted after the iterator was
4167///   created will be yielded by that iterator.
4168/// - The order in which the iterator yields indices of the buckets is unspecified
4169///   and may change in the future.
4170pub(crate) struct FullBucketsIndices {
4171    // Mask of full buckets in the current group. Bits are cleared from this
4172    // mask as each element is processed.
4173    current_group: BitMaskIter,
4174
4175    // Initial value of the bytes' indices of the current group (relative
4176    // to the start of the control bytes).
4177    group_first_index: usize,
4178
4179    // Pointer to the current group of control bytes,
4180    // Must be aligned to the group size (Group::WIDTH).
4181    ctrl: NonNull<u8>,
4182
4183    // Number of elements in the table.
4184    items: usize,
4185}
4186
4187impl FullBucketsIndices {
4188    /// Advances the iterator and returns the next value.
4189    ///
4190    /// # Safety
4191    ///
4192    /// If any of the following conditions are violated, the result is
4193    /// [`Undefined Behavior`]:
4194    ///
4195    /// * The [`RawTableInner`] / [`RawTable`] must be alive and not moved,
4196    ///   i.e. table outlives the `FullBucketsIndices`;
4197    ///
4198    /// * It never tries to iterate after getting all elements.
4199    ///
4200    /// [`Undefined Behavior`]: https://doc.rust-lang.org/reference/behavior-considered-undefined.html
4201    #[inline(always)]
4202    unsafe fn next_impl(&mut self) -> Option<usize> {
4203        loop {
4204            if let Some(index) = self.current_group.next() {
4205                // The returned `self.group_first_index + index` will always
4206                // be in the range `0..self.buckets()`. See explanation below.
4207                return Some(self.group_first_index + index);
4208            }
4209
4210            // SAFETY: The caller of this function ensures that:
4211            //
4212            // 1. It never tries to iterate after getting all the elements;
4213            // 2. The table is alive and did not moved;
4214            // 3. The first `self.ctrl` pointed to the start of the array of control bytes.
4215            //
4216            // Taking the above into account, we always stay within the bounds, because:
4217            //
4218            // 1. For tables smaller than the group width (self.buckets() <= Group::WIDTH),
4219            //    we will never end up in the given branch, since we should have already
4220            //    yielded all the elements of the table.
4221            //
4222            // 2. For tables larger than the group width. The number of buckets is a
4223            //    power of two (2 ^ n), Group::WIDTH is also power of two (2 ^ k). Since
4224            //    `(2 ^ n) > (2 ^ k)`, than `(2 ^ n) % (2 ^ k) = 0`. As we start from the
4225            //    the start of the array of control bytes, and never try to iterate after
4226            //    getting all the elements, the last `self.ctrl` will be equal to
4227            //    the `self.buckets() - Group::WIDTH`, so `self.current_group.next()`
4228            //    will always contains indices within the range `0..Group::WIDTH`,
4229            //    and subsequent `self.group_first_index + index` will always return a
4230            //    number less than `self.buckets()`.
4231            self.ctrl = NonNull::new_unchecked(self.ctrl.as_ptr().add(Group::WIDTH));
4232
4233            // SAFETY: See explanation above.
4234            self.current_group = Group::load_aligned(self.ctrl.as_ptr())
4235                .match_full()
4236                .into_iter();
4237            self.group_first_index += Group::WIDTH;
4238        }
4239    }
4240}
4241
4242impl Iterator for FullBucketsIndices {
4243    type Item = usize;
4244
4245    /// Advances the iterator and returns the next value. It is up to
4246    /// the caller to ensure that the `RawTable` outlives the `FullBucketsIndices`,
4247    /// because we cannot make the `next` method unsafe.
4248    #[inline(always)]
4249    fn next(&mut self) -> Option<usize> {
4250        // Return if we already yielded all items.
4251        if self.items == 0 {
4252            return None;
4253        }
4254
4255        let nxt = unsafe {
4256            // SAFETY:
4257            // 1. We check number of items to yield using `items` field.
4258            // 2. The caller ensures that the table is alive and has not moved.
4259            self.next_impl()
4260        };
4261
4262        debug_assert!(nxt.is_some());
4263        self.items -= 1;
4264
4265        nxt
4266    }
4267
4268    #[inline(always)]
4269    fn size_hint(&self) -> (usize, Option<usize>) {
4270        (self.items, Some(self.items))
4271    }
4272}
4273
4274impl ExactSizeIterator for FullBucketsIndices {}
4275impl FusedIterator for FullBucketsIndices {}
4276
4277/// Iterator which consumes a table and returns elements.
4278pub struct RawIntoIter<T, A: Allocator = Global> {
4279    iter: RawIter<T>,
4280    allocation: Option<(NonNull<u8>, Layout, A)>,
4281    marker: PhantomData<T>,
4282}
4283
4284impl<T, A: Allocator> RawIntoIter<T, A> {
4285    #[cfg_attr(feature = "inline-more", inline)]
4286    pub fn iter(&self) -> RawIter<T> {
4287        self.iter.clone()
4288    }
4289}
4290
4291unsafe impl<T, A: Allocator> Send for RawIntoIter<T, A>
4292where
4293    T: Send,
4294    A: Send,
4295{
4296}
4297unsafe impl<T, A: Allocator> Sync for RawIntoIter<T, A>
4298where
4299    T: Sync,
4300    A: Sync,
4301{
4302}
4303
4304#[cfg(feature = "nightly")]
4305unsafe impl<#[may_dangle] T, A: Allocator> Drop for RawIntoIter<T, A> {
4306    #[cfg_attr(feature = "inline-more", inline)]
4307    fn drop(&mut self) {
4308        unsafe {
4309            // Drop all remaining elements
4310            self.iter.drop_elements();
4311
4312            // Free the table
4313            if let Some((ptr, layout, ref alloc)) = self.allocation {
4314                alloc.deallocate(ptr, layout);
4315            }
4316        }
4317    }
4318}
4319#[cfg(not(feature = "nightly"))]
4320impl<T, A: Allocator> Drop for RawIntoIter<T, A> {
4321    #[cfg_attr(feature = "inline-more", inline)]
4322    fn drop(&mut self) {
4323        unsafe {
4324            // Drop all remaining elements
4325            self.iter.drop_elements();
4326
4327            // Free the table
4328            if let Some((ptr, layout, ref alloc)) = self.allocation {
4329                alloc.deallocate(ptr, layout);
4330            }
4331        }
4332    }
4333}
4334
4335impl<T, A: Allocator> Iterator for RawIntoIter<T, A> {
4336    type Item = T;
4337
4338    #[cfg_attr(feature = "inline-more", inline)]
4339    fn next(&mut self) -> Option<T> {
4340        unsafe { Some(self.iter.next()?.read()) }
4341    }
4342
4343    #[inline]
4344    fn size_hint(&self) -> (usize, Option<usize>) {
4345        self.iter.size_hint()
4346    }
4347}
4348
4349impl<T, A: Allocator> ExactSizeIterator for RawIntoIter<T, A> {}
4350impl<T, A: Allocator> FusedIterator for RawIntoIter<T, A> {}
4351
4352/// Iterator which consumes elements without freeing the table storage.
4353pub struct RawDrain<'a, T, A: Allocator = Global> {
4354    iter: RawIter<T>,
4355
4356    // The table is moved into the iterator for the duration of the drain. This
4357    // ensures that an empty table is left if the drain iterator is leaked
4358    // without dropping.
4359    table: RawTableInner,
4360    orig_table: NonNull<RawTableInner>,
4361
4362    // We don't use a &'a mut RawTable<T> because we want RawDrain to be
4363    // covariant over T.
4364    marker: PhantomData<&'a RawTable<T, A>>,
4365}
4366
4367impl<T, A: Allocator> RawDrain<'_, T, A> {
4368    #[cfg_attr(feature = "inline-more", inline)]
4369    pub fn iter(&self) -> RawIter<T> {
4370        self.iter.clone()
4371    }
4372}
4373
4374unsafe impl<T, A: Allocator> Send for RawDrain<'_, T, A>
4375where
4376    T: Send,
4377    A: Send,
4378{
4379}
4380unsafe impl<T, A: Allocator> Sync for RawDrain<'_, T, A>
4381where
4382    T: Sync,
4383    A: Sync,
4384{
4385}
4386
4387impl<T, A: Allocator> Drop for RawDrain<'_, T, A> {
4388    #[cfg_attr(feature = "inline-more", inline)]
4389    fn drop(&mut self) {
4390        unsafe {
4391            // Drop all remaining elements. Note that this may panic.
4392            self.iter.drop_elements();
4393
4394            // Reset the contents of the table now that all elements have been
4395            // dropped.
4396            self.table.clear_no_drop();
4397
4398            // Move the now empty table back to its original location.
4399            self.orig_table
4400                .as_ptr()
4401                .copy_from_nonoverlapping(&self.table, 1);
4402        }
4403    }
4404}
4405
4406impl<T, A: Allocator> Iterator for RawDrain<'_, T, A> {
4407    type Item = T;
4408
4409    #[cfg_attr(feature = "inline-more", inline)]
4410    fn next(&mut self) -> Option<T> {
4411        unsafe {
4412            let item = self.iter.next()?;
4413            Some(item.read())
4414        }
4415    }
4416
4417    #[inline]
4418    fn size_hint(&self) -> (usize, Option<usize>) {
4419        self.iter.size_hint()
4420    }
4421}
4422
4423impl<T, A: Allocator> ExactSizeIterator for RawDrain<'_, T, A> {}
4424impl<T, A: Allocator> FusedIterator for RawDrain<'_, T, A> {}
4425
4426/// Iterator over occupied buckets that could match a given hash.
4427///
4428/// `RawTable` only stores 7 bits of the hash value, so this iterator may return
4429/// items that have a hash value different than the one provided. You should
4430/// always validate the returned values before using them.
4431///
4432/// For maximum flexibility this iterator is not bound by a lifetime, but you
4433/// must observe several rules when using it:
4434/// - You must not free the hash table while iterating (including via growing/shrinking).
4435/// - It is fine to erase a bucket that has been yielded by the iterator.
4436/// - Erasing a bucket that has not yet been yielded by the iterator may still
4437///   result in the iterator yielding that bucket.
4438/// - It is unspecified whether an element inserted after the iterator was
4439///   created will be yielded by that iterator.
4440/// - The order in which the iterator yields buckets is unspecified and may
4441///   change in the future.
4442pub struct RawIterHash<T> {
4443    inner: RawIterHashInner,
4444    _marker: PhantomData<T>,
4445}
4446
4447struct RawIterHashInner {
4448    // See `RawTableInner`'s corresponding fields for details.
4449    // We can't store a `*const RawTableInner` as it would get
4450    // invalidated by the user calling `&mut` methods on `RawTable`.
4451    bucket_mask: usize,
4452    ctrl: NonNull<u8>,
4453
4454    // The top 7 bits of the hash.
4455    h2_hash: u8,
4456
4457    // The sequence of groups to probe in the search.
4458    probe_seq: ProbeSeq,
4459
4460    group: Group,
4461
4462    // The elements within the group with a matching h2-hash.
4463    bitmask: BitMaskIter,
4464}
4465
4466impl<T> RawIterHash<T> {
4467    #[cfg_attr(feature = "inline-more", inline)]
4468    #[cfg(feature = "raw")]
4469    unsafe fn new<A: Allocator>(table: &RawTable<T, A>, hash: u64) -> Self {
4470        RawIterHash {
4471            inner: RawIterHashInner::new(&table.table, hash),
4472            _marker: PhantomData,
4473        }
4474    }
4475}
4476impl RawIterHashInner {
4477    #[cfg_attr(feature = "inline-more", inline)]
4478    #[cfg(feature = "raw")]
4479    unsafe fn new(table: &RawTableInner, hash: u64) -> Self {
4480        let h2_hash = h2(hash);
4481        let probe_seq = table.probe_seq(hash);
4482        let group = Group::load(table.ctrl(probe_seq.pos));
4483        let bitmask = group.match_byte(h2_hash).into_iter();
4484
4485        RawIterHashInner {
4486            bucket_mask: table.bucket_mask,
4487            ctrl: table.ctrl,
4488            h2_hash,
4489            probe_seq,
4490            group,
4491            bitmask,
4492        }
4493    }
4494}
4495
4496impl<T> Iterator for RawIterHash<T> {
4497    type Item = Bucket<T>;
4498
4499    fn next(&mut self) -> Option<Bucket<T>> {
4500        unsafe {
4501            match self.inner.next() {
4502                Some(index) => {
4503                    // Can't use `RawTable::bucket` here as we don't have
4504                    // an actual `RawTable` reference to use.
4505                    debug_assert!(index <= self.inner.bucket_mask);
4506                    let bucket = Bucket::from_base_index(self.inner.ctrl.cast(), index);
4507                    Some(bucket)
4508                }
4509                None => None,
4510            }
4511        }
4512    }
4513}
4514
4515impl Iterator for RawIterHashInner {
4516    type Item = usize;
4517
4518    fn next(&mut self) -> Option<Self::Item> {
4519        unsafe {
4520            loop {
4521                if let Some(bit) = self.bitmask.next() {
4522                    let index = (self.probe_seq.pos + bit) & self.bucket_mask;
4523                    return Some(index);
4524                }
4525                if likely(self.group.match_empty().any_bit_set()) {
4526                    return None;
4527                }
4528                self.probe_seq.move_next(self.bucket_mask);
4529
4530                // Can't use `RawTableInner::ctrl` here as we don't have
4531                // an actual `RawTableInner` reference to use.
4532                let index = self.probe_seq.pos;
4533                debug_assert!(index < self.bucket_mask + 1 + Group::WIDTH);
4534                let group_ctrl = self.ctrl.as_ptr().add(index);
4535
4536                self.group = Group::load(group_ctrl);
4537                self.bitmask = self.group.match_byte(self.h2_hash).into_iter();
4538            }
4539        }
4540    }
4541}
4542
4543pub(crate) struct RawExtractIf<'a, T, A: Allocator> {
4544    pub iter: RawIter<T>,
4545    pub table: &'a mut RawTable<T, A>,
4546}
4547
4548impl<T, A: Allocator> RawExtractIf<'_, T, A> {
4549    #[cfg_attr(feature = "inline-more", inline)]
4550    pub(crate) fn next<F>(&mut self, mut f: F) -> Option<T>
4551    where
4552        F: FnMut(&mut T) -> bool,
4553    {
4554        unsafe {
4555            for item in &mut self.iter {
4556                if f(item.as_mut()) {
4557                    return Some(self.table.remove(item).0);
4558                }
4559            }
4560        }
4561        None
4562    }
4563}
4564
4565#[cfg(test)]
4566mod test_map {
4567    use super::*;
4568
4569    fn rehash_in_place<T>(table: &mut RawTable<T>, hasher: impl Fn(&T) -> u64) {
4570        unsafe {
4571            table.table.rehash_in_place(
4572                &|table, index| hasher(table.bucket::<T>(index).as_ref()),
4573                mem::size_of::<T>(),
4574                if mem::needs_drop::<T>() {
4575                    Some(mem::transmute(ptr::drop_in_place::<T> as unsafe fn(*mut T)))
4576                } else {
4577                    None
4578                },
4579            );
4580        }
4581    }
4582
4583    #[test]
4584    fn rehash() {
4585        let mut table = RawTable::new();
4586        let hasher = |i: &u64| *i;
4587        for i in 0..100 {
4588            table.insert(i, i, hasher);
4589        }
4590
4591        for i in 0..100 {
4592            unsafe {
4593                assert_eq!(table.find(i, |x| *x == i).map(|b| b.read()), Some(i));
4594            }
4595            assert!(table.find(i + 100, |x| *x == i + 100).is_none());
4596        }
4597
4598        rehash_in_place(&mut table, hasher);
4599
4600        for i in 0..100 {
4601            unsafe {
4602                assert_eq!(table.find(i, |x| *x == i).map(|b| b.read()), Some(i));
4603            }
4604            assert!(table.find(i + 100, |x| *x == i + 100).is_none());
4605        }
4606    }
4607
4608    /// CHECKING THAT WE ARE NOT TRYING TO READ THE MEMORY OF
4609    /// AN UNINITIALIZED TABLE DURING THE DROP
4610    #[test]
4611    fn test_drop_uninitialized() {
4612        use ::alloc::vec::Vec;
4613
4614        let table = unsafe {
4615            // SAFETY: The `buckets` is power of two and we're not
4616            // trying to actually use the returned RawTable.
4617            RawTable::<(u64, Vec<i32>)>::new_uninitialized(Global, 8, Fallibility::Infallible)
4618                .unwrap()
4619        };
4620        drop(table);
4621    }
4622
4623    /// CHECKING THAT WE DON'T TRY TO DROP DATA IF THE `ITEMS`
4624    /// ARE ZERO, EVEN IF WE HAVE `FULL` CONTROL BYTES.
4625    #[test]
4626    fn test_drop_zero_items() {
4627        use ::alloc::vec::Vec;
4628        unsafe {
4629            // SAFETY: The `buckets` is power of two and we're not
4630            // trying to actually use the returned RawTable.
4631            let table =
4632                RawTable::<(u64, Vec<i32>)>::new_uninitialized(Global, 8, Fallibility::Infallible)
4633                    .unwrap();
4634
4635            // WE SIMULATE, AS IT WERE, A FULL TABLE.
4636
4637            // SAFETY: We checked that the table is allocated and therefore the table already has
4638            // `self.bucket_mask + 1 + Group::WIDTH` number of control bytes (see TableLayout::calculate_layout_for)
4639            // so writing `table.table.num_ctrl_bytes() == bucket_mask + 1 + Group::WIDTH` bytes is safe.
4640            table
4641                .table
4642                .ctrl(0)
4643                .write_bytes(EMPTY, table.table.num_ctrl_bytes());
4644
4645            // SAFETY: table.capacity() is guaranteed to be smaller than table.buckets()
4646            table.table.ctrl(0).write_bytes(0, table.capacity());
4647
4648            // Fix up the trailing control bytes. See the comments in set_ctrl
4649            // for the handling of tables smaller than the group width.
4650            if table.buckets() < Group::WIDTH {
4651                // SAFETY: We have `self.bucket_mask + 1 + Group::WIDTH` number of control bytes,
4652                // so copying `self.buckets() == self.bucket_mask + 1` bytes with offset equal to
4653                // `Group::WIDTH` is safe
4654                table
4655                    .table
4656                    .ctrl(0)
4657                    .copy_to(table.table.ctrl(Group::WIDTH), table.table.buckets());
4658            } else {
4659                // SAFETY: We have `self.bucket_mask + 1 + Group::WIDTH` number of
4660                // control bytes,so copying `Group::WIDTH` bytes with offset equal
4661                // to `self.buckets() == self.bucket_mask + 1` is safe
4662                table
4663                    .table
4664                    .ctrl(0)
4665                    .copy_to(table.table.ctrl(table.table.buckets()), Group::WIDTH);
4666            }
4667            drop(table);
4668        }
4669    }
4670
4671    /// CHECKING THAT WE DON'T TRY TO DROP DATA IF THE `ITEMS`
4672    /// ARE ZERO, EVEN IF WE HAVE `FULL` CONTROL BYTES.
4673    #[test]
4674    fn test_catch_panic_clone_from() {
4675        use ::alloc::sync::Arc;
4676        use ::alloc::vec::Vec;
4677        use allocator_api2::alloc::{AllocError, Allocator, Global};
4678        use core::sync::atomic::{AtomicI8, Ordering};
4679        use std::thread;
4680
4681        struct MyAllocInner {
4682            drop_count: Arc<AtomicI8>,
4683        }
4684
4685        #[derive(Clone)]
4686        struct MyAlloc {
4687            _inner: Arc<MyAllocInner>,
4688        }
4689
4690        impl Drop for MyAllocInner {
4691            fn drop(&mut self) {
4692                println!("MyAlloc freed.");
4693                self.drop_count.fetch_sub(1, Ordering::SeqCst);
4694            }
4695        }
4696
4697        unsafe impl Allocator for MyAlloc {
4698            fn allocate(&self, layout: Layout) -> std::result::Result<NonNull<[u8]>, AllocError> {
4699                let g = Global;
4700                g.allocate(layout)
4701            }
4702
4703            unsafe fn deallocate(&self, ptr: NonNull<u8>, layout: Layout) {
4704                let g = Global;
4705                g.deallocate(ptr, layout)
4706            }
4707        }
4708
4709        const DISARMED: bool = false;
4710        const ARMED: bool = true;
4711
4712        struct CheckedCloneDrop {
4713            panic_in_clone: bool,
4714            dropped: bool,
4715            need_drop: Vec<u64>,
4716        }
4717
4718        impl Clone for CheckedCloneDrop {
4719            fn clone(&self) -> Self {
4720                if self.panic_in_clone {
4721                    panic!("panic in clone")
4722                }
4723                Self {
4724                    panic_in_clone: self.panic_in_clone,
4725                    dropped: self.dropped,
4726                    need_drop: self.need_drop.clone(),
4727                }
4728            }
4729        }
4730
4731        impl Drop for CheckedCloneDrop {
4732            fn drop(&mut self) {
4733                if self.dropped {
4734                    panic!("double drop");
4735                }
4736                self.dropped = true;
4737            }
4738        }
4739
4740        let dropped: Arc<AtomicI8> = Arc::new(AtomicI8::new(2));
4741
4742        let mut table = RawTable::new_in(MyAlloc {
4743            _inner: Arc::new(MyAllocInner {
4744                drop_count: dropped.clone(),
4745            }),
4746        });
4747
4748        for (idx, panic_in_clone) in core::iter::repeat(DISARMED).take(7).enumerate() {
4749            let idx = idx as u64;
4750            table.insert(
4751                idx,
4752                (
4753                    idx,
4754                    CheckedCloneDrop {
4755                        panic_in_clone,
4756                        dropped: false,
4757                        need_drop: vec![idx],
4758                    },
4759                ),
4760                |(k, _)| *k,
4761            );
4762        }
4763
4764        assert_eq!(table.len(), 7);
4765
4766        thread::scope(|s| {
4767            let result = s.spawn(|| {
4768                let armed_flags = [
4769                    DISARMED, DISARMED, ARMED, DISARMED, DISARMED, DISARMED, DISARMED,
4770                ];
4771                let mut scope_table = RawTable::new_in(MyAlloc {
4772                    _inner: Arc::new(MyAllocInner {
4773                        drop_count: dropped.clone(),
4774                    }),
4775                });
4776                for (idx, &panic_in_clone) in armed_flags.iter().enumerate() {
4777                    let idx = idx as u64;
4778                    scope_table.insert(
4779                        idx,
4780                        (
4781                            idx,
4782                            CheckedCloneDrop {
4783                                panic_in_clone,
4784                                dropped: false,
4785                                need_drop: vec![idx + 100],
4786                            },
4787                        ),
4788                        |(k, _)| *k,
4789                    );
4790                }
4791                table.clone_from(&scope_table);
4792            });
4793            assert!(result.join().is_err());
4794        });
4795
4796        // Let's check that all iterators work fine and do not return elements
4797        // (especially `RawIterRange`, which does not depend on the number of
4798        // elements in the table, but looks directly at the control bytes)
4799        //
4800        // SAFETY: We know for sure that `RawTable` will outlive
4801        // the returned `RawIter / RawIterRange` iterator.
4802        assert_eq!(table.len(), 0);
4803        assert_eq!(unsafe { table.iter().count() }, 0);
4804        assert_eq!(unsafe { table.iter().iter.count() }, 0);
4805
4806        for idx in 0..table.buckets() {
4807            let idx = idx as u64;
4808            assert!(
4809                table.find(idx, |(k, _)| *k == idx).is_none(),
4810                "Index: {idx}"
4811            );
4812        }
4813
4814        // All allocator clones should already be dropped.
4815        assert_eq!(dropped.load(Ordering::SeqCst), 1);
4816    }
4817}