zerocopy/util/
mod.rs

1// Copyright 2023 The Fuchsia Authors
2//
3// Licensed under a BSD-style license <LICENSE-BSD>, Apache License, Version 2.0
4// <LICENSE-APACHE or https://www.apache.org/licenses/LICENSE-2.0>, or the MIT
5// license <LICENSE-MIT or https://opensource.org/licenses/MIT>, at your option.
6// This file may not be copied, modified, or distributed except according to
7// those terms.
8
9#[macro_use]
10pub(crate) mod macros;
11
12#[doc(hidden)]
13pub mod macro_util;
14
15use core::{
16    marker::PhantomData,
17    mem::{self, ManuallyDrop},
18    num::NonZeroUsize,
19    ptr::NonNull,
20};
21
22use super::*;
23
24/// Like [`PhantomData`], but [`Send`] and [`Sync`] regardless of whether the
25/// wrapped `T` is.
26pub(crate) struct SendSyncPhantomData<T: ?Sized>(PhantomData<T>);
27
28// SAFETY: `SendSyncPhantomData` does not enable any behavior which isn't sound
29// to be called from multiple threads.
30unsafe impl<T: ?Sized> Send for SendSyncPhantomData<T> {}
31// SAFETY: `SendSyncPhantomData` does not enable any behavior which isn't sound
32// to be called from multiple threads.
33unsafe impl<T: ?Sized> Sync for SendSyncPhantomData<T> {}
34
35impl<T: ?Sized> Default for SendSyncPhantomData<T> {
36    fn default() -> SendSyncPhantomData<T> {
37        SendSyncPhantomData(PhantomData)
38    }
39}
40
41impl<T: ?Sized> PartialEq for SendSyncPhantomData<T> {
42    fn eq(&self, _other: &Self) -> bool {
43        true
44    }
45}
46
47impl<T: ?Sized> Eq for SendSyncPhantomData<T> {}
48
49impl<T: ?Sized> Clone for SendSyncPhantomData<T> {
50    fn clone(&self) -> Self {
51        SendSyncPhantomData(PhantomData)
52    }
53}
54
55#[cfg(miri)]
56extern "Rust" {
57    /// Miri-provided intrinsic that marks the pointer `ptr` as aligned to
58    /// `align`.
59    ///
60    /// This intrinsic is used to inform Miri's symbolic alignment checker that
61    /// a pointer is aligned, even if Miri cannot statically deduce that fact.
62    /// This is often required when performing raw pointer arithmetic or casts
63    /// where the alignment is guaranteed by runtime checks or invariants that
64    /// Miri is not aware of.
65    pub(crate) fn miri_promise_symbolic_alignment(ptr: *const (), align: usize);
66}
67
68pub(crate) trait AsAddress {
69    fn addr(self) -> usize;
70}
71
72impl<T: ?Sized> AsAddress for &T {
73    #[inline(always)]
74    fn addr(self) -> usize {
75        let ptr: *const T = self;
76        AsAddress::addr(ptr)
77    }
78}
79
80impl<T: ?Sized> AsAddress for &mut T {
81    #[inline(always)]
82    fn addr(self) -> usize {
83        let ptr: *const T = self;
84        AsAddress::addr(ptr)
85    }
86}
87
88impl<T: ?Sized> AsAddress for NonNull<T> {
89    #[inline(always)]
90    fn addr(self) -> usize {
91        AsAddress::addr(self.as_ptr())
92    }
93}
94
95impl<T: ?Sized> AsAddress for *const T {
96    #[inline(always)]
97    fn addr(self) -> usize {
98        // FIXME(#181), FIXME(https://github.com/rust-lang/rust/issues/95228):
99        // Use `.addr()` instead of `as usize` once it's stable, and get rid of
100        // this `allow`. Currently, `as usize` is the only way to accomplish
101        // this.
102        #[allow(clippy::as_conversions)]
103        #[cfg_attr(
104            __ZEROCOPY_INTERNAL_USE_ONLY_NIGHTLY_FEATURES_IN_TESTS,
105            allow(lossy_provenance_casts)
106        )]
107        return self.cast::<()>() as usize;
108    }
109}
110
111impl<T: ?Sized> AsAddress for *mut T {
112    #[inline(always)]
113    fn addr(self) -> usize {
114        let ptr: *const T = self;
115        AsAddress::addr(ptr)
116    }
117}
118
119/// Validates that `t` is aligned to `align_of::<U>()`.
120#[inline(always)]
121pub(crate) fn validate_aligned_to<T: AsAddress, U>(t: T) -> Result<(), AlignmentError<(), U>> {
122    // `mem::align_of::<U>()` is guaranteed to return a non-zero value, which in
123    // turn guarantees that this mod operation will not panic.
124    #[allow(clippy::arithmetic_side_effects)]
125    let remainder = t.addr() % mem::align_of::<U>();
126    if remainder == 0 {
127        Ok(())
128    } else {
129        // SAFETY: We just confirmed that `t.addr() % align_of::<U>() != 0`.
130        // That's only possible if `align_of::<U>() > 1`.
131        Err(unsafe { AlignmentError::new_unchecked(()) })
132    }
133}
134
135/// Returns the bytes needed to pad `len` to the next multiple of `align`.
136///
137/// This function assumes that align is a power of two; there are no guarantees
138/// on the answer it gives if this is not the case.
139#[cfg_attr(
140    kani,
141    kani::requires(len <= isize::MAX as usize),
142    kani::requires(align.is_power_of_two()),
143    kani::ensures(|&p| (len + p) % align.get() == 0),
144    // Ensures that we add the minimum required padding.
145    kani::ensures(|&p| p < align.get()),
146)]
147pub(crate) const fn padding_needed_for(len: usize, align: NonZeroUsize) -> usize {
148    #[cfg(kani)]
149    #[kani::proof_for_contract(padding_needed_for)]
150    fn proof() {
151        padding_needed_for(kani::any(), kani::any());
152    }
153
154    // Abstractly, we want to compute:
155    //   align - (len % align).
156    // Handling the case where len%align is 0.
157    // Because align is a power of two, len % align = len & (align-1).
158    // Guaranteed not to underflow as align is nonzero.
159    #[allow(clippy::arithmetic_side_effects)]
160    let mask = align.get() - 1;
161
162    // To efficiently subtract this value from align, we can use the bitwise
163    // complement.
164    // Note that ((!len) & (align-1)) gives us a number that with (len &
165    // (align-1)) sums to align-1. So subtracting 1 from x before taking the
166    // complement subtracts `len` from `align`. Some quick inspection of
167    // cases shows that this also handles the case where `len % align = 0`
168    // correctly too: len-1 % align then equals align-1, so the complement mod
169    // align will be 0, as desired.
170    //
171    // The following reasoning can be verified quickly by an SMT solver
172    // supporting the theory of bitvectors:
173    // ```smtlib
174    // ; Naive implementation of padding
175    // (define-fun padding1 (
176    //     (len (_ BitVec 32))
177    //     (align (_ BitVec 32))) (_ BitVec 32)
178    //    (ite
179    //      (= (_ bv0 32) (bvand len (bvsub align (_ bv1 32))))
180    //      (_ bv0 32)
181    //      (bvsub align (bvand len (bvsub align (_ bv1 32))))))
182    //
183    // ; The implementation below
184    // (define-fun padding2 (
185    //     (len (_ BitVec 32))
186    //     (align (_ BitVec 32))) (_ BitVec 32)
187    // (bvand (bvnot (bvsub len (_ bv1 32))) (bvsub align (_ bv1 32))))
188    //
189    // (define-fun is-power-of-two ((x (_ BitVec 32))) Bool
190    //   (= (_ bv0 32) (bvand x (bvsub x (_ bv1 32)))))
191    //
192    // (declare-const len (_ BitVec 32))
193    // (declare-const align (_ BitVec 32))
194    // ; Search for a case where align is a power of two and padding2 disagrees
195    // ; with padding1
196    // (assert (and (is-power-of-two align)
197    //              (not (= (padding1 len align) (padding2 len align)))))
198    // (simplify (padding1 (_ bv300 32) (_ bv32 32))) ; 20
199    // (simplify (padding2 (_ bv300 32) (_ bv32 32))) ; 20
200    // (simplify (padding1 (_ bv322 32) (_ bv32 32))) ; 30
201    // (simplify (padding2 (_ bv322 32) (_ bv32 32))) ; 30
202    // (simplify (padding1 (_ bv8 32) (_ bv8 32)))    ; 0
203    // (simplify (padding2 (_ bv8 32) (_ bv8 32)))    ; 0
204    // (check-sat) ; unsat, also works for 64-bit bitvectors
205    // ```
206    !(len.wrapping_sub(1)) & mask
207}
208
209/// Rounds `n` down to the largest value `m` such that `m <= n` and `m % align
210/// == 0`.
211///
212/// # Panics
213///
214/// May panic if `align` is not a power of two. Even if it doesn't panic in this
215/// case, it will produce nonsense results.
216#[inline(always)]
217#[cfg_attr(
218    kani,
219    kani::requires(align.is_power_of_two()),
220    kani::ensures(|&m| m <= n && m % align.get() == 0),
221    // Guarantees that `m` is the *largest* value such that `m % align == 0`.
222    kani::ensures(|&m| {
223        // If this `checked_add` fails, then the next multiple would wrap
224        // around, which trivially satisfies the "largest value" requirement.
225        m.checked_add(align.get()).map(|next_mul| next_mul > n).unwrap_or(true)
226    })
227)]
228pub(crate) const fn round_down_to_next_multiple_of_alignment(
229    n: usize,
230    align: NonZeroUsize,
231) -> usize {
232    #[cfg(kani)]
233    #[kani::proof_for_contract(round_down_to_next_multiple_of_alignment)]
234    fn proof() {
235        round_down_to_next_multiple_of_alignment(kani::any(), kani::any());
236    }
237
238    let align = align.get();
239    #[cfg(not(no_zerocopy_panic_in_const_and_vec_try_reserve_1_57_0))]
240    debug_assert!(align.is_power_of_two());
241
242    // Subtraction can't underflow because `align.get() >= 1`.
243    #[allow(clippy::arithmetic_side_effects)]
244    let mask = !(align - 1);
245    n & mask
246}
247
248pub(crate) const fn max(a: NonZeroUsize, b: NonZeroUsize) -> NonZeroUsize {
249    if a.get() < b.get() {
250        b
251    } else {
252        a
253    }
254}
255
256pub(crate) const fn min(a: NonZeroUsize, b: NonZeroUsize) -> NonZeroUsize {
257    if a.get() > b.get() {
258        b
259    } else {
260        a
261    }
262}
263
264/// Copies `src` into the prefix of `dst`.
265///
266/// # Safety
267///
268/// The caller guarantees that `src.len() <= dst.len()`.
269#[inline(always)]
270pub(crate) unsafe fn copy_unchecked(src: &[u8], dst: &mut [u8]) {
271    debug_assert!(src.len() <= dst.len());
272    // SAFETY: This invocation satisfies the safety contract of
273    // copy_nonoverlapping [1]:
274    // - `src.as_ptr()` is trivially valid for reads of `src.len()` bytes
275    // - `dst.as_ptr()` is valid for writes of `src.len()` bytes, because the
276    //   caller has promised that `src.len() <= dst.len()`
277    // - `src` and `dst` are, trivially, properly aligned
278    // - the region of memory beginning at `src` with a size of `src.len()`
279    //   bytes does not overlap with the region of memory beginning at `dst`
280    //   with the same size, because `dst` is derived from an exclusive
281    //   reference.
282    unsafe {
283        core::ptr::copy_nonoverlapping(src.as_ptr(), dst.as_mut_ptr(), src.len());
284    };
285}
286
287/// Unsafely transmutes the given `src` into a type `Dst`.
288///
289/// # Safety
290///
291/// The value `src` must be a valid instance of `Dst`.
292#[inline(always)]
293pub(crate) const unsafe fn transmute_unchecked<Src, Dst>(src: Src) -> Dst {
294    static_assert!(Src, Dst => core::mem::size_of::<Src>() == core::mem::size_of::<Dst>());
295
296    #[repr(C)]
297    union Transmute<Src, Dst> {
298        src: ManuallyDrop<Src>,
299        dst: ManuallyDrop<Dst>,
300    }
301
302    // SAFETY: Since `Transmute<Src, Dst>` is `#[repr(C)]`, its `src` and `dst`
303    // fields both start at the same offset and the types of those fields are
304    // transparent wrappers around `Src` and `Dst` [1]. Consequently,
305    // initializing `Transmute` with with `src` and then reading out `dst` is
306    // equivalent to transmuting from `Src` to `Dst` [2]. Transmuting from `src`
307    // to `Dst` is valid because — by contract on the caller — `src` is a valid
308    // instance of `Dst`.
309    //
310    // [1] Per https://doc.rust-lang.org/1.82.0/std/mem/struct.ManuallyDrop.html:
311    //
312    //     `ManuallyDrop<T>` is guaranteed to have the same layout and bit
313    //     validity as `T`, and is subject to the same layout optimizations as
314    //     `T`.
315    //
316    // [2] Per https://doc.rust-lang.org/1.82.0/reference/items/unions.html#reading-and-writing-union-fields:
317    //
318    //     Effectively, writing to and then reading from a union with the C
319    //     representation is analogous to a transmute from the type used for
320    //     writing to the type used for reading.
321    unsafe { ManuallyDrop::into_inner(Transmute { src: ManuallyDrop::new(src) }.dst) }
322}
323
324/// Uses `allocate` to create a `Box<T>`.
325///
326/// # Errors
327///
328/// Returns an error on allocation failure. Allocation failure is guaranteed
329/// never to cause a panic or an abort.
330///
331/// # Safety
332///
333/// `allocate` must be either `alloc::alloc::alloc` or
334/// `alloc::alloc::alloc_zeroed`. The referent of the box returned by `new_box`
335/// has the same bit-validity as the referent of the pointer returned by the
336/// given `allocate` and sufficient size to store `T` with `meta`.
337#[must_use = "has no side effects (other than allocation)"]
338#[cfg(feature = "alloc")]
339#[inline]
340pub(crate) unsafe fn new_box<T>(
341    meta: T::PointerMetadata,
342    allocate: unsafe fn(core::alloc::Layout) -> *mut u8,
343) -> Result<alloc::boxed::Box<T>, AllocError>
344where
345    T: ?Sized + crate::KnownLayout,
346{
347    let size = match T::size_for_metadata(meta) {
348        Some(size) => size,
349        None => return Err(AllocError),
350    };
351
352    let align = T::LAYOUT.align.get();
353    // On stable Rust versions <= 1.64.0, `Layout::from_size_align` has a bug in
354    // which sufficiently-large allocations (those which, when rounded up to the
355    // alignment, overflow `isize`) are not rejected, which can cause undefined
356    // behavior. See #64 for details.
357    //
358    // FIXME(#67): Once our MSRV is > 1.64.0, remove this assertion.
359    #[allow(clippy::as_conversions)]
360    let max_alloc = (isize::MAX as usize).saturating_sub(align);
361    if size > max_alloc {
362        return Err(AllocError);
363    }
364
365    // FIXME(https://github.com/rust-lang/rust/issues/55724): Use
366    // `Layout::repeat` once it's stabilized.
367    let layout = Layout::from_size_align(size, align).or(Err(AllocError))?;
368
369    let ptr = if layout.size() != 0 {
370        // SAFETY: By contract on the caller, `allocate` is either
371        // `alloc::alloc::alloc` or `alloc::alloc::alloc_zeroed`. The above
372        // check ensures their shared safety precondition: that the supplied
373        // layout is not zero-sized type [1].
374        //
375        // [1] Per https://doc.rust-lang.org/1.81.0/std/alloc/trait.GlobalAlloc.html#tymethod.alloc:
376        //
377        //     This function is unsafe because undefined behavior can result if
378        //     the caller does not ensure that layout has non-zero size.
379        let ptr = unsafe { allocate(layout) };
380        match NonNull::new(ptr) {
381            Some(ptr) => ptr,
382            None => return Err(AllocError),
383        }
384    } else {
385        let align = T::LAYOUT.align.get();
386
387        // We use `transmute` instead of an `as` cast since Miri (with strict
388        // provenance enabled) notices and complains that an `as` cast creates a
389        // pointer with no provenance. Miri isn't smart enough to realize that
390        // we're only executing this branch when we're constructing a zero-sized
391        // `Box`, which doesn't require provenance.
392        //
393        // SAFETY: any initialized bit sequence is a bit-valid `*mut u8`. All
394        // bits of a `usize` are initialized.
395        //
396        // `#[allow(unknown_lints)]` is for `integer_to_ptr_transmutes`
397        #[allow(unknown_lints)]
398        #[allow(clippy::useless_transmute, integer_to_ptr_transmutes)]
399        let dangling = unsafe { mem::transmute::<usize, *mut u8>(align) };
400        // SAFETY: `dangling` is constructed from `T::LAYOUT.align`, which is a
401        // `NonZeroUsize`, which is guaranteed to be non-zero.
402        //
403        // `Box<[T]>` does not allocate when `T` is zero-sized or when `len` is
404        // zero, but it does require a non-null dangling pointer for its
405        // allocation.
406        //
407        // FIXME(https://github.com/rust-lang/rust/issues/95228): Use
408        // `std::ptr::without_provenance` once it's stable. That may optimize
409        // better. As written, Rust may assume that this consumes "exposed"
410        // provenance, and thus Rust may have to assume that this may consume
411        // provenance from any pointer whose provenance has been exposed.
412        unsafe { NonNull::new_unchecked(dangling) }
413    };
414
415    let ptr = T::raw_from_ptr_len(ptr, meta);
416
417    // FIXME(#429): Add a "SAFETY" comment and remove this `allow`. Make sure to
418    // include a justification that `ptr.as_ptr()` is validly-aligned in the ZST
419    // case (in which we manually construct a dangling pointer) and to justify
420    // why `Box` is safe to drop (it's because `allocate` uses the system
421    // allocator).
422    #[allow(clippy::undocumented_unsafe_blocks)]
423    Ok(unsafe { alloc::boxed::Box::from_raw(ptr.as_ptr()) })
424}
425
426mod len_of {
427    use super::*;
428
429    /// A witness type for metadata of a valid instance of `&T`.
430    pub(crate) struct MetadataOf<T: ?Sized + KnownLayout> {
431        /// # Safety
432        ///
433        /// The size of an instance of `&T` with the given metadata is not
434        /// larger than `isize::MAX`.
435        meta: T::PointerMetadata,
436        _p: PhantomData<T>,
437    }
438
439    impl<T: ?Sized + KnownLayout> Copy for MetadataOf<T> {}
440    impl<T: ?Sized + KnownLayout> Clone for MetadataOf<T> {
441        fn clone(&self) -> Self {
442            *self
443        }
444    }
445
446    impl<T: ?Sized> MetadataOf<T>
447    where
448        T: KnownLayout,
449    {
450        /// Returns `None` if `meta` is greater than `t`'s metadata.
451        #[inline(always)]
452        pub(crate) fn new_in_bounds(t: &T, meta: usize) -> Option<Self>
453        where
454            T: KnownLayout<PointerMetadata = usize>,
455        {
456            if meta <= Ptr::from_ref(t).len() {
457                // SAFETY: We have checked that `meta` is not greater than `t`'s
458                // metadata, which, by invariant on `&T`, addresses no more than
459                // `isize::MAX` bytes [1][2].
460                //
461                // [1] Per https://doc.rust-lang.org/1.85.0/std/primitive.reference.html#safety:
462                //
463                //    For all types, `T: ?Sized`, and for all `t: &T` or `t:
464                //    &mut T`, when such values cross an API boundary, the
465                //    following invariants must generally be upheld:
466                //
467                //    * `t` is non-null
468                //    * `t` is aligned to `align_of_val(t)`
469                //    * if `size_of_val(t) > 0`, then `t` is dereferenceable for
470                //      `size_of_val(t)` many bytes
471                //
472                //    If `t` points at address `a`, being "dereferenceable" for
473                //    N bytes means that the memory range `[a, a + N)` is all
474                //    contained within a single allocated object.
475                //
476                // [2] Per https://doc.rust-lang.org/1.85.0/std/ptr/index.html#allocated-object:
477                //
478                //    For any allocated object with `base` address, `size`, and
479                //    a set of `addresses`, the following are guaranteed:
480                //    - For all addresses `a` in `addresses`, `a` is in the
481                //      range `base .. (base + size)` (note that this requires
482                //      `a < base + size`, not `a <= base + size`)
483                //    - `base` is not equal to [`null()`] (i.e., the address
484                //      with the numerical value 0)
485                //    - `base + size <= usize::MAX`
486                //    - `size <= isize::MAX`
487                Some(unsafe { Self::new_unchecked(meta) })
488            } else {
489                None
490            }
491        }
492
493        /// # Safety
494        ///
495        /// The size of an instance of `&T` with the given metadata is not
496        /// larger than `isize::MAX`.
497        pub(crate) unsafe fn new_unchecked(meta: T::PointerMetadata) -> Self {
498            // SAFETY: The caller has promised that the size of an instance of
499            // `&T` with the given metadata is not larger than `isize::MAX`.
500            Self { meta, _p: PhantomData }
501        }
502
503        pub(crate) fn get(&self) -> T::PointerMetadata
504        where
505            T::PointerMetadata: Copy,
506        {
507            self.meta
508        }
509
510        #[inline]
511        pub(crate) fn padding_needed_for(&self) -> usize
512        where
513            T: KnownLayout<PointerMetadata = usize>,
514        {
515            let trailing_slice_layout = crate::trailing_slice_layout::<T>();
516
517            // FIXME(#67): Remove this allow. See NumExt for more details.
518            #[allow(
519                unstable_name_collisions,
520                clippy::incompatible_msrv,
521                clippy::multiple_unsafe_ops_per_block
522            )]
523            // SAFETY: By invariant on `self`, a `&T` with metadata `self.meta`
524            // describes an object of size `<= isize::MAX`. This computes the
525            // size of such a `&T` without any trailing padding, and so neither
526            // the multiplication nor the addition will overflow.
527            let unpadded_size = unsafe {
528                let trailing_size = self.meta.unchecked_mul(trailing_slice_layout.elem_size);
529                trailing_size.unchecked_add(trailing_slice_layout.offset)
530            };
531
532            util::padding_needed_for(unpadded_size, T::LAYOUT.align)
533        }
534
535        #[inline(always)]
536        pub(crate) fn validate_cast_and_convert_metadata(
537            addr: usize,
538            bytes_len: MetadataOf<[u8]>,
539            cast_type: CastType,
540            meta: Option<T::PointerMetadata>,
541        ) -> Result<(MetadataOf<T>, MetadataOf<[u8]>), MetadataCastError> {
542            let layout = match meta {
543                None => T::LAYOUT,
544                // This can return `None` if the metadata describes an object
545                // which can't fit in an `isize`.
546                Some(meta) => {
547                    let size = match T::size_for_metadata(meta) {
548                        Some(size) => size,
549                        None => return Err(MetadataCastError::Size),
550                    };
551                    DstLayout {
552                        align: T::LAYOUT.align,
553                        size_info: crate::SizeInfo::Sized { size },
554                        statically_shallow_unpadded: false,
555                    }
556                }
557            };
558            // Lemma 0: By contract on `validate_cast_and_convert_metadata`, if
559            // the result is `Ok(..)`, then a `&T` with `elems` trailing slice
560            // elements is no larger in size than `bytes_len.get()`.
561            let (elems, split_at) =
562                layout.validate_cast_and_convert_metadata(addr, bytes_len.get(), cast_type)?;
563            let elems = T::PointerMetadata::from_elem_count(elems);
564
565            // For a slice DST type, if `meta` is `Some(elems)`, then we
566            // synthesize `layout` to describe a sized type whose size is equal
567            // to the size of the instance that we are asked to cast. For sized
568            // types, `validate_cast_and_convert_metadata` returns `elems == 0`.
569            // Thus, in this case, we need to use the `elems` passed by the
570            // caller, not the one returned by
571            // `validate_cast_and_convert_metadata`.
572            //
573            // Lemma 1: A `&T` with `elems` trailing slice elements is no larger
574            // in size than `bytes_len.get()`. Proof:
575            // - If `meta` is `None`, then `elems` satisfies this condition by
576            //   Lemma 0.
577            // - If `meta` is `Some(meta)`, then `layout` describes an object
578            //   whose size is equal to the size of an `&T` with `meta`
579            //   metadata. By Lemma 0, that size is not larger than
580            //   `bytes_len.get()`.
581            //
582            // Lemma 2: A `&T` with `elems` trailing slice elements is no larger
583            // than `isize::MAX` bytes. Proof: By Lemma 1, a `&T` with metadata
584            // `elems` is not larger in size than `bytes_len.get()`. By
585            // invariant on `MetadataOf<[u8]>`, a `&[u8]` with metadata
586            // `bytes_len` is not larger than `isize::MAX`. Because
587            // `size_of::<u8>()` is `1`, a `&[u8]` with metadata `bytes_len` has
588            // size `bytes_len.get()` bytes. Therefore, a `&T` with metadata
589            // `elems` has size not larger than `isize::MAX`.
590            let elems = meta.unwrap_or(elems);
591
592            // SAFETY: See Lemma 2.
593            let elems = unsafe { MetadataOf::new_unchecked(elems) };
594
595            // SAFETY: Let `size` be the size of a `&T` with metadata `elems`.
596            // By post-condition on `validate_cast_and_convert_metadata`, one of
597            // the following conditions holds:
598            // - `split_at == size`, in which case, by Lemma 2, `split_at <=
599            //   isize::MAX`. Since `size_of::<u8>() == 1`, a `[u8]` with
600            //   `split_at` elems has size not larger than `isize::MAX`.
601            // - `split_at == bytes_len - size`. Since `bytes_len:
602            //   MetadataOf<u8>`, and since `size` is non-negative, `split_at`
603            //   addresses no more bytes than `bytes_len` does. Since
604            //   `bytes_len: MetadataOf<u8>`, `bytes_len` describes a `[u8]`
605            //   which has no more than `isize::MAX` bytes, and thus so does
606            //   `split_at`.
607            let split_at = unsafe { MetadataOf::<[u8]>::new_unchecked(split_at) };
608            Ok((elems, split_at))
609        }
610    }
611}
612
613pub(crate) use len_of::MetadataOf;
614
615/// Since we support multiple versions of Rust, there are often features which
616/// have been stabilized in the most recent stable release which do not yet
617/// exist (stably) on our MSRV. This module provides polyfills for those
618/// features so that we can write more "modern" code, and just remove the
619/// polyfill once our MSRV supports the corresponding feature. Without this,
620/// we'd have to write worse/more verbose code and leave FIXME comments
621/// sprinkled throughout the codebase to update to the new pattern once it's
622/// stabilized.
623///
624/// Each trait is imported as `_` at the crate root; each polyfill should "just
625/// work" at usage sites.
626pub(crate) mod polyfills {
627    use core::ptr::{self, NonNull};
628
629    // A polyfill for `NonNull::slice_from_raw_parts` that we can use before our
630    // MSRV is 1.70, when that function was stabilized.
631    //
632    // The `#[allow(unused)]` is necessary because, on sufficiently recent
633    // toolchain versions, `ptr.slice_from_raw_parts()` resolves to the inherent
634    // method rather than to this trait, and so this trait is considered unused.
635    //
636    // FIXME(#67): Once our MSRV is 1.70, remove this.
637    #[allow(unused)]
638    pub(crate) trait NonNullExt<T> {
639        fn slice_from_raw_parts(data: Self, len: usize) -> NonNull<[T]>;
640    }
641
642    impl<T> NonNullExt<T> for NonNull<T> {
643        // NOTE on coverage: this will never be tested in nightly since it's a
644        // polyfill for a feature which has been stabilized on our nightly
645        // toolchain.
646        #[cfg_attr(
647            all(coverage_nightly, __ZEROCOPY_INTERNAL_USE_ONLY_NIGHTLY_FEATURES_IN_TESTS),
648            coverage(off)
649        )]
650        #[inline(always)]
651        fn slice_from_raw_parts(data: Self, len: usize) -> NonNull<[T]> {
652            let ptr = ptr::slice_from_raw_parts_mut(data.as_ptr(), len);
653            // SAFETY: `ptr` is converted from `data`, which is non-null.
654            unsafe { NonNull::new_unchecked(ptr) }
655        }
656    }
657
658    // A polyfill for `Self::unchecked_sub` that we can use until methods like
659    // `usize::unchecked_sub` is stabilized.
660    //
661    // The `#[allow(unused)]` is necessary because, on sufficiently recent
662    // toolchain versions, `ptr.slice_from_raw_parts()` resolves to the inherent
663    // method rather than to this trait, and so this trait is considered unused.
664    //
665    // FIXME(#67): Once our MSRV is high enough, remove this.
666    #[allow(unused)]
667    pub(crate) trait NumExt {
668        /// Add without checking for overflow.
669        ///
670        /// # Safety
671        ///
672        /// The caller promises that the addition will not overflow.
673        unsafe fn unchecked_add(self, rhs: Self) -> Self;
674
675        /// Subtract without checking for underflow.
676        ///
677        /// # Safety
678        ///
679        /// The caller promises that the subtraction will not underflow.
680        unsafe fn unchecked_sub(self, rhs: Self) -> Self;
681
682        /// Multiply without checking for overflow.
683        ///
684        /// # Safety
685        ///
686        /// The caller promises that the multiplication will not overflow.
687        unsafe fn unchecked_mul(self, rhs: Self) -> Self;
688    }
689
690    // NOTE on coverage: these will never be tested in nightly since they're
691    // polyfills for a feature which has been stabilized on our nightly
692    // toolchain.
693    impl NumExt for usize {
694        #[cfg_attr(
695            all(coverage_nightly, __ZEROCOPY_INTERNAL_USE_ONLY_NIGHTLY_FEATURES_IN_TESTS),
696            coverage(off)
697        )]
698        #[inline(always)]
699        unsafe fn unchecked_add(self, rhs: usize) -> usize {
700            match self.checked_add(rhs) {
701                Some(x) => x,
702                None => {
703                    // SAFETY: The caller promises that the addition will not
704                    // underflow.
705                    unsafe { core::hint::unreachable_unchecked() }
706                }
707            }
708        }
709
710        #[cfg_attr(
711            all(coverage_nightly, __ZEROCOPY_INTERNAL_USE_ONLY_NIGHTLY_FEATURES_IN_TESTS),
712            coverage(off)
713        )]
714        #[inline(always)]
715        unsafe fn unchecked_sub(self, rhs: usize) -> usize {
716            match self.checked_sub(rhs) {
717                Some(x) => x,
718                None => {
719                    // SAFETY: The caller promises that the subtraction will not
720                    // underflow.
721                    unsafe { core::hint::unreachable_unchecked() }
722                }
723            }
724        }
725
726        #[cfg_attr(
727            all(coverage_nightly, __ZEROCOPY_INTERNAL_USE_ONLY_NIGHTLY_FEATURES_IN_TESTS),
728            coverage(off)
729        )]
730        #[inline(always)]
731        unsafe fn unchecked_mul(self, rhs: usize) -> usize {
732            match self.checked_mul(rhs) {
733                Some(x) => x,
734                None => {
735                    // SAFETY: The caller promises that the multiplication will
736                    // not overflow.
737                    unsafe { core::hint::unreachable_unchecked() }
738                }
739            }
740        }
741    }
742}
743
744#[cfg(test)]
745pub(crate) mod testutil {
746    use crate::*;
747
748    /// A `T` which is aligned to at least `align_of::<A>()`.
749    #[derive(Default)]
750    pub(crate) struct Align<T, A> {
751        pub(crate) t: T,
752        _a: [A; 0],
753    }
754
755    impl<T: Default, A> Align<T, A> {
756        pub(crate) fn set_default(&mut self) {
757            self.t = T::default();
758        }
759    }
760
761    impl<T, A> Align<T, A> {
762        pub(crate) const fn new(t: T) -> Align<T, A> {
763            Align { t, _a: [] }
764        }
765    }
766
767    /// A `T` which is guaranteed not to satisfy `align_of::<A>()`.
768    ///
769    /// It must be the case that `align_of::<T>() < align_of::<A>()` in order
770    /// for this type to work properly.
771    #[repr(C)]
772    pub(crate) struct ForceUnalign<T: Unaligned, A> {
773        // The outer struct is aligned to `A`, and, thanks to `repr(C)`, `t` is
774        // placed at the minimum offset that guarantees its alignment. If
775        // `align_of::<T>() < align_of::<A>()`, then that offset will be
776        // guaranteed *not* to satisfy `align_of::<A>()`.
777        //
778        // Note that we need `T: Unaligned` in order to guarantee that there is
779        // no padding between `_u` and `t`.
780        _u: u8,
781        pub(crate) t: T,
782        _a: [A; 0],
783    }
784
785    impl<T: Unaligned, A> ForceUnalign<T, A> {
786        pub(crate) fn new(t: T) -> ForceUnalign<T, A> {
787            ForceUnalign { _u: 0, t, _a: [] }
788        }
789    }
790    // A `u64` with alignment 8.
791    //
792    // Though `u64` has alignment 8 on some platforms, it's not guaranteed. By
793    // contrast, `AU64` is guaranteed to have alignment 8 on all platforms.
794    #[derive(
795        KnownLayout,
796        Immutable,
797        FromBytes,
798        IntoBytes,
799        Eq,
800        PartialEq,
801        Ord,
802        PartialOrd,
803        Default,
804        Debug,
805        Copy,
806        Clone,
807    )]
808    #[repr(C, align(8))]
809    pub(crate) struct AU64(pub(crate) u64);
810
811    impl AU64 {
812        // Converts this `AU64` to bytes using this platform's endianness.
813        pub(crate) fn to_bytes(self) -> [u8; 8] {
814            crate::transmute!(self)
815        }
816    }
817
818    impl Display for AU64 {
819        #[cfg_attr(
820            all(coverage_nightly, __ZEROCOPY_INTERNAL_USE_ONLY_NIGHTLY_FEATURES_IN_TESTS),
821            coverage(off)
822        )]
823        fn fmt(&self, f: &mut Formatter<'_>) -> fmt::Result {
824            Display::fmt(&self.0, f)
825        }
826    }
827}
828
829#[cfg(test)]
830mod tests {
831    use super::*;
832
833    #[test]
834    fn test_round_down_to_next_multiple_of_alignment() {
835        fn alt_impl(n: usize, align: NonZeroUsize) -> usize {
836            let mul = n / align.get();
837            mul * align.get()
838        }
839
840        for align in [1, 2, 4, 8, 16] {
841            for n in 0..256 {
842                let align = NonZeroUsize::new(align).unwrap();
843                let want = alt_impl(n, align);
844                let got = round_down_to_next_multiple_of_alignment(n, align);
845                assert_eq!(got, want, "round_down_to_next_multiple_of_alignment({}, {})", n, align);
846            }
847        }
848    }
849
850    #[rustversion::since(1.57.0)]
851    #[test]
852    #[should_panic]
853    fn test_round_down_to_next_multiple_of_alignment_zerocopy_panic_in_const_and_vec_try_reserve() {
854        round_down_to_next_multiple_of_alignment(0, NonZeroUsize::new(3).unwrap());
855    }
856}