musli/alloc/
stack.rs

1#[cfg(test)]
2mod tests;
3
4use core::cell::UnsafeCell;
5use core::marker::PhantomData;
6use core::mem::{align_of, forget, replace, size_of, MaybeUninit};
7use core::num::NonZeroU16;
8use core::ops::{Deref, DerefMut};
9use core::ptr;
10
11use super::{Allocator, RawVec};
12
13// We keep max bytes to 2^31, since that ensures that addition between two
14// magnitutes never overflow.
15const MAX_BYTES: usize = i32::MAX as usize;
16
17/// A no-std compatible slice-based allocator that can be used with the `musli`
18/// crate.
19///
20/// It is geared towards handling few allocations, but they can be arbitrarily
21/// large. It is optimized to work best when allocations are short lived and
22/// "merged back" into one previously allocated region through
23/// `Buffer::extend`.
24///
25/// It's also optimized to write to one allocation "at a time". So once an
26/// allocation has been grown once, it will be put in a region where it is
27/// unlikely to need to be moved again, usually the last region which has access
28/// to the remainder of the provided buffer.
29///
30/// For the moment, this allocator only supports 65535 unique allocations, which
31/// is fine for use with the `musli` crate, but might be a limitation for other
32/// use-cases.
33///
34/// ## Examples
35///
36/// ```
37/// use musli::alloc::{ArrayBuffer, Slice, Vec};
38///
39/// let mut buf = ArrayBuffer::new();
40/// let alloc = Slice::new(&mut buf);
41///
42/// let mut a = Vec::new_in(&alloc);
43/// let mut b = Vec::new_in(&alloc);
44///
45/// b.write(b"He11o");
46/// a.write(b.as_slice());
47///
48/// assert_eq!(a.as_slice(), b"He11o");
49/// assert_eq!(a.len(), 5);
50///
51/// a.write(b" W0rld");
52///
53/// assert_eq!(a.as_slice(), b"He11o W0rld");
54/// assert_eq!(a.len(), 11);
55///
56/// let mut c = Vec::new_in(&alloc);
57/// c.write(b"!");
58/// a.write(c.as_slice());
59///
60/// assert_eq!(a.as_slice(), b"He11o W0rld!");
61/// assert_eq!(a.len(), 12);
62/// ```
63///
64/// ## Design
65///
66/// The allocator takes a buffer of contiguous memory. This is dynamically
67/// diviced into two parts:
68///
69/// * One part which grows upwards from the base, constituting the memory being
70///   allocated.
71/// * Its metadata growing downward from the end of the buffer, containing
72///   headers for all allocated region.
73///
74/// By designing the allocator so that the memory allocated and its metadata is
75/// separate, neighbouring regions can efficiently be merged as they are written
76/// or freed.
77///
78/// Each allocation is sparse, meaning it does not try to over-allocate memory.
79/// This ensures that subsequent regions with initialized memory can be merged
80/// efficiently, but degrades performance for many small writes performed across
81/// multiple allocations concurrently.
82///
83/// Below is an illustration of this, where `a` and `b` are two allocations
84/// where we write one byte at a time to each. Here `x` below indicates an
85/// occupied `gap` in memory regions.
86///
87/// ```text
88/// a
89/// ab
90/// # a moved to end
91/// xbaa
92/// # b moved to 0
93/// bbaa
94/// # aa not moved
95/// bbaaa
96/// # bb moved to end
97/// xxaaabbb
98/// # aaa moved to 0
99/// aaaaxbbb
100/// # bbb not moved
101/// aaaaxbbbb
102/// # aaaa not moved
103/// aaaaabbbb
104/// # bbbbb not moved
105/// aaaaabbbbb
106/// # aaaaa moved to end
107/// xxxxxbbbbbaaaaaa
108/// # bbbbb moved to 0
109/// bbbbbbxxxxaaaaaa
110/// ```
111pub struct Slice<'a> {
112    // This must be an unsafe cell, since it's mutably accessed through an
113    // immutable pointers. We simply make sure that those accesses do not
114    // clobber each other, which we can do since the API is restricted through
115    // the `RawVec` trait.
116    internal: UnsafeCell<Internal>,
117    // The underlying vector being borrowed.
118    _marker: PhantomData<&'a mut [MaybeUninit<u8>]>,
119}
120
121impl<'a> Slice<'a> {
122    /// Construct a new slice allocator.
123    ///
124    /// See [type-level documentation][Slice] for more information.
125    ///
126    /// # Panics
127    ///
128    /// This panics if called with a buffer larger than `2^31` bytes.
129    pub fn new(buffer: &'a mut [MaybeUninit<u8>]) -> Self {
130        let size = buffer.len();
131        assert!(
132            size <= MAX_BYTES,
133            "Buffer of {size} bytes is larger than the maximum {MAX_BYTES}"
134        );
135
136        let mut data = Range::new(buffer.as_mut_ptr_range());
137
138        // The region of allocated headers grows downwards from the end of the
139        // buffer, so in order to ensure headers are aligned, we simply align
140        // the end pointer of the buffer preemptively here. Then we don't have
141        // to worry about it.
142        let align = data.end.align_offset(align_of::<Header>());
143
144        if align != 0 {
145            let sub = align_of::<Header>() - align;
146
147            if sub <= size {
148                // SAFETY: We've ensured that the adjustment is less than the
149                // size of the buffer.
150                unsafe {
151                    data.end = data.end.sub(sub);
152                }
153            } else {
154                data.end = data.start;
155            }
156        }
157
158        Self {
159            internal: UnsafeCell::new(Internal {
160                free_head: None,
161                tail: None,
162                occupied: None,
163                full: data,
164                free: data,
165            }),
166            _marker: PhantomData,
167        }
168    }
169}
170
171impl Allocator for Slice<'_> {
172    type RawVec<'this, T> = SliceBuf<'this, T> where Self: 'this, T: 'this;
173
174    #[inline]
175    fn new_raw_vec<'a, T>(&'a self) -> Self::RawVec<'a, T>
176    where
177        T: 'a,
178    {
179        // SAFETY: We have exclusive access to the internal state, and it's only
180        // held for the duration of this call.
181        let region = if size_of::<T>() == 0 {
182            None
183        } else {
184            unsafe {
185                (*self.internal.get())
186                    .alloc(0, align_of::<T>())
187                    .map(|r| r.id)
188            }
189        };
190
191        SliceBuf {
192            region,
193            internal: &self.internal,
194            _marker: PhantomData,
195        }
196    }
197}
198
199/// A slice allocated buffer.
200///
201/// See [`Slice`].
202pub struct SliceBuf<'a, T> {
203    region: Option<HeaderId>,
204    internal: &'a UnsafeCell<Internal>,
205    _marker: PhantomData<T>,
206}
207
208impl<'a, T> RawVec<T> for SliceBuf<'a, T> {
209    #[inline]
210    fn resize(&mut self, len: usize, additional: usize) -> bool {
211        if additional == 0 || size_of::<T>() == 0 {
212            return true;
213        }
214
215        let Some(region_id) = self.region else {
216            return false;
217        };
218
219        let Some(len) = len.checked_mul(size_of::<T>()) else {
220            return false;
221        };
222
223        let Some(additional) = additional.checked_mul(size_of::<T>()) else {
224            return false;
225        };
226
227        let Some(requested) = len.checked_add(additional) else {
228            return false;
229        };
230
231        if requested > MAX_BYTES {
232            return false;
233        }
234
235        // SAFETY: Due to invariants in the Buffer trait we know that these
236        // cannot be used incorrectly.
237        unsafe {
238            let i = &mut *self.internal.get();
239
240            let region = i.region(region_id);
241
242            // Region can already fit in the requested bytes.
243            if region.capacity() >= requested {
244                return true;
245            };
246
247            let Some(region) = i.realloc(region_id, len, requested, align_of::<T>()) else {
248                return false;
249            };
250
251            self.region = Some(region.id);
252            true
253        }
254    }
255
256    #[inline]
257    fn as_ptr(&self) -> *const T {
258        let Some(region) = self.region else {
259            return ptr::NonNull::dangling().as_ptr();
260        };
261
262        unsafe {
263            let i = &*self.internal.get();
264            let this = i.header(region);
265            this.range.start.cast_const().cast()
266        }
267    }
268
269    #[inline]
270    fn as_mut_ptr(&mut self) -> *mut T {
271        let Some(region) = self.region else {
272            return ptr::NonNull::dangling().as_ptr();
273        };
274
275        unsafe {
276            let i = &*self.internal.get();
277            let this = i.header(region);
278            this.range.start.cast()
279        }
280    }
281
282    #[inline]
283    fn try_merge<B>(&mut self, this_len: usize, buf: B, other_len: usize) -> Result<(), B>
284    where
285        B: RawVec<T>,
286    {
287        let Some(region) = self.region else {
288            return Err(buf);
289        };
290
291        let this_len = this_len * size_of::<T>();
292        let other_len = other_len * size_of::<T>();
293
294        // NB: Placing this here to make miri happy, since accessing the
295        // slice will mean mutably accessing the internal state.
296        let other_ptr = buf.as_ptr().cast();
297
298        unsafe {
299            let i = &mut *self.internal.get();
300            let mut this = i.region(region);
301
302            debug_assert!(this.capacity() >= this_len);
303
304            // If this region immediately follows the other region, we can
305            // optimize the write by simply growing the current region and
306            // de-allocating the second since the only conclusion is that
307            // they share the same allocator.
308            if !ptr::eq(this.range.end.cast_const(), other_ptr) {
309                return Err(buf);
310            }
311
312            let Some(next) = this.next else {
313                return Err(buf);
314            };
315
316            // Prevent the other buffer from being dropped, since we're
317            // taking care of the allocation in here directly instead.
318            forget(buf);
319
320            let next = i.region(next);
321
322            let to = this.range.start.wrapping_add(this_len);
323
324            // Data needs to be shuffle back to the end of the initialized
325            // region.
326            if this.range.end != to {
327                this.range.end.copy_to(to, other_len);
328            }
329
330            let old = i.free_region(next);
331            this.range.end = old.range.end;
332            Ok(())
333        }
334    }
335}
336
337impl<T> Drop for SliceBuf<'_, T> {
338    fn drop(&mut self) {
339        if let Some(region) = self.region.take() {
340            // SAFETY: We have exclusive access to the internal state.
341            unsafe {
342                (*self.internal.get()).free(region);
343            }
344        }
345    }
346}
347
348struct Region {
349    id: HeaderId,
350    ptr: *mut Header,
351}
352
353impl Deref for Region {
354    type Target = Header;
355
356    #[inline]
357    fn deref(&self) -> &Self::Target {
358        // SAFETY: Construction of the region is unsafe, so the caller must
359        // ensure that it's used correctly after that.
360        unsafe { &*self.ptr }
361    }
362}
363
364impl DerefMut for Region {
365    #[inline]
366    fn deref_mut(&mut self) -> &mut Self::Target {
367        // SAFETY: Construction of the region is unsafe, so the caller must
368        // ensure that it's used correctly after that.
369        unsafe { &mut *self.ptr }
370    }
371}
372
373/// The identifier of a region.
374#[derive(Debug, Clone, Copy, PartialEq, Eq)]
375#[cfg_attr(test, derive(PartialOrd, Ord, Hash))]
376#[repr(transparent)]
377struct HeaderId(NonZeroU16);
378
379impl HeaderId {
380    #[cfg(test)]
381    const unsafe fn new_unchecked(value: u16) -> Self {
382        Self(NonZeroU16::new_unchecked(value))
383    }
384
385    /// Create a new region identifier.
386    ///
387    /// # Safety
388    ///
389    /// The given value must be non-zero.
390    #[inline]
391    fn new(value: isize) -> Option<Self> {
392        Some(Self(NonZeroU16::new(u16::try_from(value).ok()?)?))
393    }
394
395    /// Get the value of the region identifier.
396    #[inline]
397    fn get(self) -> usize {
398        self.0.get() as usize
399    }
400}
401
402#[derive(Debug, Clone, Copy, PartialEq, Eq)]
403struct Range {
404    start: *mut MaybeUninit<u8>,
405    end: *mut MaybeUninit<u8>,
406}
407
408impl Range {
409    fn new(range: core::ops::Range<*mut MaybeUninit<u8>>) -> Self {
410        Self {
411            start: range.start,
412            end: range.end,
413        }
414    }
415
416    fn head(self) -> Range {
417        Self {
418            start: self.start,
419            end: self.start,
420        }
421    }
422}
423
424struct Internal {
425    // The first free region.
426    free_head: Option<HeaderId>,
427    // Pointer to the tail region.
428    tail: Option<HeaderId>,
429    // The occupied header region
430    occupied: Option<HeaderId>,
431    // The full range used by the allocator.
432    full: Range,
433    // The free range available to the allocator.
434    free: Range,
435}
436
437impl Internal {
438    // Return the number of allocates bytes.
439    #[cfg(test)]
440    #[inline]
441    fn bytes(&self) -> usize {
442        // SAFETY: It is guaranteed that free_end >= free_start inside of the provided region.
443        unsafe { self.free.start.byte_offset_from(self.full.start) as usize }
444    }
445
446    #[cfg(test)]
447    #[inline]
448    fn headers(&self) -> usize {
449        unsafe {
450            self.full
451                .end
452                .cast::<Header>()
453                .offset_from(self.free.end.cast()) as usize
454        }
455    }
456
457    // Return the number of remaining bytes.
458    #[inline]
459    fn remaining(&self) -> usize {
460        // SAFETY: It is guaranteed that free_end >= free_start inside of the provided region.
461        unsafe { self.free.end.byte_offset_from(self.free.start) as usize }
462    }
463
464    /// Get the header pointer corresponding to the given id.
465    #[inline]
466    fn header(&self, at: HeaderId) -> &Header {
467        // SAFETY: Once we've coerced to `&self`, then we guarantee that we can
468        // get a header immutably.
469        unsafe { &*self.full.end.cast::<Header>().wrapping_sub(at.get()) }
470    }
471
472    /// Get the mutable header pointer corresponding to the given id.
473    #[inline]
474    fn header_mut(&mut self, at: HeaderId) -> *mut Header {
475        self.full.end.cast::<Header>().wrapping_sub(at.get())
476    }
477
478    /// Get the mutable region corresponding to the given id.
479    #[inline]
480    fn region(&mut self, id: HeaderId) -> Region {
481        Region {
482            id,
483            ptr: self.header_mut(id),
484        }
485    }
486
487    unsafe fn unlink(&mut self, header: &Header) {
488        if let Some(next) = header.next {
489            (*self.header_mut(next)).prev = header.prev;
490        } else {
491            self.tail = header.prev;
492        }
493
494        if let Some(prev) = header.prev {
495            (*self.header_mut(prev)).next = header.next;
496        }
497    }
498
499    unsafe fn replace_back(&mut self, region: &mut Region) {
500        let prev = region.prev.take();
501        let next = region.next.take();
502
503        if let Some(prev) = prev {
504            (*self.header_mut(prev)).next = next;
505        }
506
507        if let Some(next) = next {
508            (*self.header_mut(next)).prev = prev;
509        }
510
511        self.push_back(region);
512    }
513
514    unsafe fn push_back(&mut self, region: &mut Region) {
515        if let Some(tail) = self.tail.replace(region.id) {
516            region.prev = Some(tail);
517            (*self.header_mut(tail)).next = Some(region.id);
518        }
519    }
520
521    /// Free a region.
522    unsafe fn free_region(&mut self, region: Region) -> Header {
523        self.unlink(&region);
524
525        region.ptr.replace(Header {
526            range: self.full.head(),
527            next: self.free_head.replace(region.id),
528            prev: None,
529        })
530    }
531
532    /// Allocate a new header.
533    ///
534    /// # Safety
535    ///
536    /// The caller msut ensure that there is enough space for the region and
537    /// that the end pointer has been aligned to the requirements of `Header`.
538    unsafe fn alloc_header(&mut self, end: *mut MaybeUninit<u8>) -> Option<Region> {
539        if let Some(region) = self.free_head.take() {
540            let mut region = self.region(region);
541
542            region.range.start = self.free.start;
543            region.range.end = end;
544
545            return Some(region);
546        }
547
548        debug_assert_eq!(
549            self.free.end.align_offset(align_of::<Header>()),
550            0,
551            "End pointer should be aligned to header"
552        );
553
554        let ptr = self.free.end.cast::<Header>().wrapping_sub(1);
555
556        if ptr < self.free.start.cast() || ptr >= self.free.end.cast() {
557            return None;
558        }
559
560        let id = HeaderId::new(self.full.end.cast::<Header>().offset_from(ptr))?;
561
562        ptr.write(Header {
563            range: Range::new(self.free.start..end),
564            prev: None,
565            next: None,
566        });
567
568        self.free.end = ptr.cast();
569        Some(Region { id, ptr })
570    }
571
572    /// Allocate a region.
573    ///
574    /// # Safety
575    ///
576    /// The caller must ensure that `this` is exclusively available.
577    unsafe fn alloc(&mut self, requested: usize, align: usize) -> Option<Region> {
578        if let Some(occupied) = self.occupied {
579            let region = self.region(occupied);
580
581            if region.capacity() >= requested && region.is_aligned(align) {
582                self.occupied = None;
583                return Some(region);
584            }
585        }
586
587        self.align(align)?;
588
589        if self.remaining() < requested {
590            return None;
591        }
592
593        let end = self.free.start.wrapping_add(requested);
594
595        let mut region = self.alloc_header(end)?;
596
597        self.free.start = end;
598        debug_assert!(self.free.start <= self.free.end);
599        self.push_back(&mut region);
600        Some(region)
601    }
602
603    /// Align the free region by the specified alignment.
604    ///
605    /// This might require either expanding the tail region, or introducing an
606    /// occupied region which matches the number of bytes needed to fulfill the
607    /// specified alignment.
608    unsafe fn align(&mut self, align: usize) -> Option<()> {
609        let align = self.free.start.align_offset(align);
610
611        if align == 0 {
612            return Some(());
613        }
614
615        if self.remaining() < align {
616            return None;
617        }
618
619        let aligned_start = self.free.start.wrapping_add(align);
620
621        if let Some(tail) = self.tail {
622            // Simply expand the tail region to fill the gap created.
623            self.region(tail).range.end = aligned_start;
624        } else {
625            // We need to construct a new occupied header to fill in the gap
626            // which we just aligned from since there is no previous region to
627            // expand.
628            let mut region = self.alloc_header(aligned_start)?;
629
630            self.push_back(&mut region);
631        }
632
633        self.free.start = aligned_start;
634        Some(())
635    }
636
637    unsafe fn free(&mut self, region: HeaderId) {
638        let region = self.region(region);
639
640        // Just free up the last region in the slab.
641        if region.next.is_none() {
642            self.free_tail(region);
643            return;
644        }
645
646        // If there is no previous region, then mark this region as occupy.
647        let Some(prev) = region.prev else {
648            debug_assert!(
649                self.occupied.is_none(),
650                "There can only be one occupied region"
651            );
652
653            self.occupied = Some(region.id);
654            return;
655        };
656
657        let mut prev = self.region(prev);
658
659        // Move allocation to the previous region.
660        let region = self.free_region(region);
661
662        prev.range.end = region.range.end;
663
664        // The current header being freed is the last in the list.
665        if region.next.is_none() {
666            self.free.start = region.range.start;
667        }
668    }
669
670    /// Free the tail starting at the `current` region.
671    unsafe fn free_tail(&mut self, current: Region) {
672        debug_assert_eq!(self.tail, Some(current.id));
673
674        let current = self.free_region(current);
675        debug_assert_eq!(current.next, None);
676
677        self.free.start = match current.prev {
678            // The prior region is occupied, so we can free that as well.
679            Some(prev) if self.occupied == Some(prev) => {
680                self.occupied = None;
681                let prev = self.region(prev);
682                self.free_region(prev).range.start
683            }
684            _ => current.range.start,
685        };
686    }
687
688    unsafe fn reserve(&mut self, additional: usize, align: usize) -> Option<*mut MaybeUninit<u8>> {
689        self.align(align)?;
690
691        let free_start = self.free.start.wrapping_add(additional);
692
693        if free_start > self.free.end || free_start < self.free.start {
694            return None;
695        }
696
697        Some(free_start)
698    }
699
700    unsafe fn realloc(
701        &mut self,
702        from: HeaderId,
703        len: usize,
704        requested: usize,
705        align: usize,
706    ) -> Option<Region> {
707        let mut from = self.region(from);
708
709        // This is the last region in the slab, so we can just expand it.
710        if from.next.is_none() {
711            // Before we call realloc, we check the capacity of the current
712            // region. So we know that it is <= requested.
713            let additional = requested - from.capacity();
714            self.free.start = self.reserve(additional, align)?;
715            from.range.end = from.range.end.add(additional);
716            return Some(from);
717        }
718
719        // There is no data allocated in the current region, so we can simply
720        // re-link it to the end of the chain of allocation.
721        if from.range.start == from.range.end {
722            let free_start = self.reserve(requested, align)?;
723            from.range.start = replace(&mut self.free.start, free_start);
724            from.range.end = free_start;
725            self.replace_back(&mut from);
726            return Some(from);
727        }
728
729        // Try to merge with a preceeding region, if the requested memory can
730        // fit in it.
731        'bail: {
732            // Check if the immediate prior region can fit the requested allocation.
733            let Some(prev) = from.prev else {
734                break 'bail;
735            };
736
737            if self.occupied != Some(prev) {
738                break 'bail;
739            }
740
741            let mut prev = self.region(prev);
742
743            if prev.capacity() + from.capacity() < requested {
744                break 'bail;
745            }
746
747            if !prev.is_aligned(align) {
748                break 'bail;
749            }
750
751            let from = self.free_region(from);
752
753            from.range.start.copy_to(prev.range.start, len);
754            prev.range.end = from.range.end;
755            self.occupied = None;
756            return Some(prev);
757        }
758
759        let to = self.alloc(requested, align)?;
760        from.range.start.copy_to_nonoverlapping(to.range.start, len);
761        self.free(from.id);
762        Some(to)
763    }
764}
765
766/// The header of a region.
767#[derive(Debug, Clone, Copy, PartialEq, Eq)]
768struct Header {
769    // The range of the allocated region.
770    range: Range,
771    // The previous region.
772    prev: Option<HeaderId>,
773    // The next region.
774    next: Option<HeaderId>,
775}
776
777impl Header {
778    #[inline]
779    fn capacity(&self) -> usize {
780        // SAFETY: Both pointers are defined within the region.
781        unsafe { self.range.end.byte_offset_from(self.range.start) as usize }
782    }
783
784    /// Test if region is aligned to `align`.
785    #[inline]
786    fn is_aligned(&self, align: usize) -> bool {
787        self.range.start.align_offset(align) == 0
788    }
789}