| //===-- Implementation header for a block of memory -------------*- C++ -*-===// |
| // |
| // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| // See https://llvm.org/LICENSE.txt for license information. |
| // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| // |
| //===----------------------------------------------------------------------===// |
| |
| #ifndef LLVM_LIBC_SRC___SUPPORT_BLOCK_H |
| #define LLVM_LIBC_SRC___SUPPORT_BLOCK_H |
| |
| #include "src/__support/CPP/algorithm.h" |
| #include "src/__support/CPP/cstddef.h" |
| #include "src/__support/CPP/limits.h" |
| #include "src/__support/CPP/new.h" |
| #include "src/__support/CPP/optional.h" |
| #include "src/__support/CPP/span.h" |
| #include "src/__support/CPP/type_traits.h" |
| #include "src/__support/libc_assert.h" |
| #include "src/__support/macros/config.h" |
| #include "src/__support/math_extras.h" |
| |
| #include <stdint.h> |
| |
| namespace LIBC_NAMESPACE_DECL { |
| |
| /// Returns the value rounded down to the nearest multiple of alignment. |
| LIBC_INLINE constexpr size_t align_down(size_t value, size_t alignment) { |
| // Note this shouldn't overflow since the result will always be <= value. |
| return (value / alignment) * alignment; |
| } |
| |
| /// Returns the value rounded up to the nearest multiple of alignment. May wrap |
| /// around. |
| LIBC_INLINE constexpr size_t align_up(size_t value, size_t alignment) { |
| return align_down(value + alignment - 1, alignment); |
| } |
| |
| using ByteSpan = cpp::span<LIBC_NAMESPACE::cpp::byte>; |
| using cpp::optional; |
| |
| /// Memory region with links to adjacent blocks. |
| /// |
| /// The blocks store their offsets to the previous and next blocks. The latter |
| /// is also the block's size. |
| /// |
| /// All blocks have their usable space aligned to some multiple of max_align_t. |
| /// This also implies that block outer sizes are aligned to max_align_t. |
| /// |
| /// As an example, the diagram below represents two contiguous `Block`s. The |
| /// indices indicate byte offsets: |
| /// |
| /// @code{.unparsed} |
| /// Block 1: |
| /// +---------------------+--------------+ |
| /// | Header | Usable space | |
| /// +----------+----------+--------------+ |
| /// | prev | next | | |
| /// | 0......3 | 4......7 | 8........227 | |
| /// | 00000000 | 00000230 | <app data> | |
| /// +----------+----------+--------------+ |
| /// Block 2: |
| /// +---------------------+--------------+ |
| /// | Header | Usable space | |
| /// +----------+----------+--------------+ |
| /// | prev | next | | |
| /// | 0......3 | 4......7 | 8........827 | |
| /// | 00000230 | 00000830 | f7f7....f7f7 | |
| /// +----------+----------+--------------+ |
| /// @endcode |
| /// |
| /// As a space optimization, when a block is allocated, it consumes the prev |
| /// field of the following block: |
| /// |
| /// Block 1 (used): |
| /// +---------------------+--------------+ |
| /// | Header | Usable space | |
| /// +----------+----------+--------------+ |
| /// | prev | next | | |
| /// | 0......3 | 4......7 | 8........230 | |
| /// | 00000000 | 00000230 | <app data> | |
| /// +----------+----------+--------------+ |
| /// Block 2: |
| /// +---------------------+--------------+ |
| /// | B1 | Header | Usable space | |
| /// +----------+----------+--------------+ |
| /// | | next | | |
| /// | 0......3 | 4......7 | 8........827 | |
| /// | xxxxxxxx | 00000830 | f7f7....f7f7 | |
| /// +----------+----------+--------------+ |
| /// |
| /// The next offset of a block matches the previous offset of its next block. |
| /// The first block in a list is denoted by having a previous offset of `0`. |
| class Block { |
| // Masks for the contents of the next_ field. |
| static constexpr size_t PREV_FREE_MASK = 1 << 0; |
| static constexpr size_t LAST_MASK = 1 << 1; |
| static constexpr size_t SIZE_MASK = ~(PREV_FREE_MASK | LAST_MASK); |
| |
| public: |
| // No copy or move. |
| Block(const Block &other) = delete; |
| Block &operator=(const Block &other) = delete; |
| |
| /// Initializes a given memory region into a first block and a sentinel last |
| /// block. Returns the first block, which has its usable space aligned to |
| /// max_align_t. |
| static optional<Block *> init(ByteSpan region); |
| |
| /// @returns A pointer to a `Block`, given a pointer to the start of the |
| /// usable space inside the block. |
| /// |
| /// This is the inverse of `usable_space()`. |
| /// |
| /// @warning This method does not do any checking; passing a random |
| /// pointer will return a non-null pointer. |
| LIBC_INLINE static Block *from_usable_space(void *usable_space) { |
| auto *bytes = reinterpret_cast<cpp::byte *>(usable_space); |
| return reinterpret_cast<Block *>(bytes - sizeof(Block)); |
| } |
| LIBC_INLINE static const Block *from_usable_space(const void *usable_space) { |
| const auto *bytes = reinterpret_cast<const cpp::byte *>(usable_space); |
| return reinterpret_cast<const Block *>(bytes - sizeof(Block)); |
| } |
| |
| /// @returns The total size of the block in bytes, including the header. |
| LIBC_INLINE size_t outer_size() const { return next_ & SIZE_MASK; } |
| |
| LIBC_INLINE static size_t outer_size(size_t inner_size) { |
| // The usable region includes the prev_ field of the next block. |
| return inner_size - sizeof(prev_) + sizeof(Block); |
| } |
| |
| /// @returns The number of usable bytes inside the block were it to be |
| /// allocated. |
| LIBC_INLINE size_t inner_size() const { |
| if (!next()) |
| return 0; |
| return inner_size(outer_size()); |
| } |
| |
| /// @returns The number of usable bytes inside a block with the given outer |
| /// size were it to be allocated. |
| LIBC_INLINE static size_t inner_size(size_t outer_size) { |
| // The usable region includes the prev_ field of the next block. |
| return inner_size_free(outer_size) + sizeof(prev_); |
| } |
| |
| /// @returns The number of usable bytes inside the block if it remains free. |
| LIBC_INLINE size_t inner_size_free() const { |
| if (!next()) |
| return 0; |
| return inner_size_free(outer_size()); |
| } |
| |
| /// @returns The number of usable bytes inside a block with the given outer |
| /// size if it remains free. |
| LIBC_INLINE static size_t inner_size_free(size_t outer_size) { |
| return outer_size - sizeof(Block); |
| } |
| |
| /// @returns A pointer to the usable space inside this block. |
| /// |
| /// Aligned to some multiple of max_align_t. |
| LIBC_INLINE cpp::byte *usable_space() { |
| auto *s = reinterpret_cast<cpp::byte *>(this) + sizeof(Block); |
| LIBC_ASSERT(reinterpret_cast<uintptr_t>(s) % alignof(max_align_t) == 0 && |
| "usable space must be aligned to a multiple of max_align_t"); |
| return s; |
| } |
| LIBC_INLINE const cpp::byte *usable_space() const { |
| const auto *s = reinterpret_cast<const cpp::byte *>(this) + sizeof(Block); |
| LIBC_ASSERT(reinterpret_cast<uintptr_t>(s) % alignof(max_align_t) == 0 && |
| "usable space must be aligned to a multiple of max_align_t"); |
| return s; |
| } |
| |
| // @returns The region of memory the block manages, including the header. |
| LIBC_INLINE ByteSpan region() { |
| return {reinterpret_cast<cpp::byte *>(this), outer_size()}; |
| } |
| |
| /// Attempts to split this block. |
| /// |
| /// If successful, the block will have an inner size of at least |
| /// `new_inner_size`. The remaining space will be returned as a new block, |
| /// with usable space aligned to `usable_space_alignment`. Note that the prev_ |
| /// field of the next block counts as part of the inner size of the block. |
| /// `usable_space_alignment` must be a multiple of max_align_t. |
| optional<Block *> split(size_t new_inner_size, |
| size_t usable_space_alignment = alignof(max_align_t)); |
| |
| /// Merges this block with the one that comes after it. |
| bool merge_next(); |
| |
| /// @returns The block immediately after this one, or a null pointer if this |
| /// is the last block. |
| LIBC_INLINE Block *next() const { |
| if (next_ & LAST_MASK) |
| return nullptr; |
| return reinterpret_cast<Block *>(reinterpret_cast<uintptr_t>(this) + |
| outer_size()); |
| } |
| |
| /// @returns The free block immediately before this one, otherwise nullptr. |
| LIBC_INLINE Block *prev_free() const { |
| if (!(next_ & PREV_FREE_MASK)) |
| return nullptr; |
| return reinterpret_cast<Block *>(reinterpret_cast<uintptr_t>(this) - prev_); |
| } |
| |
| /// @returns Whether the block is unavailable for allocation. |
| LIBC_INLINE bool used() const { return !next() || !next()->prev_free(); } |
| |
| /// Marks this block as in use. |
| LIBC_INLINE void mark_used() { |
| LIBC_ASSERT(next() && "last block is always considered used"); |
| next()->next_ &= ~PREV_FREE_MASK; |
| } |
| |
| /// Marks this block as free. |
| LIBC_INLINE void mark_free() { |
| LIBC_ASSERT(next() && "last block is always considered used"); |
| next()->next_ |= PREV_FREE_MASK; |
| // The next block's prev_ field becomes alive, as it is no longer part of |
| // this block's used space. |
| *new (&next()->prev_) size_t = outer_size(); |
| } |
| |
| /// Marks this block as the last one in the chain. Makes next() return |
| /// nullptr. |
| LIBC_INLINE void mark_last() { next_ |= LAST_MASK; } |
| |
| LIBC_INLINE Block(size_t outer_size) : next_(outer_size) { |
| LIBC_ASSERT(outer_size % alignof(max_align_t) == 0 && |
| "block sizes must be aligned"); |
| LIBC_ASSERT(is_usable_space_aligned(alignof(max_align_t)) && |
| "usable space must be aligned to a multiple of max_align_t"); |
| } |
| |
| LIBC_INLINE bool is_usable_space_aligned(size_t alignment) const { |
| return reinterpret_cast<uintptr_t>(usable_space()) % alignment == 0; |
| } |
| |
| // Returns the minimum inner size necessary for a block of that size to |
| // always be able to allocate at the given size and alignment. |
| // |
| // Returns 0 if there is no such size. |
| LIBC_INLINE static size_t min_size_for_allocation(size_t alignment, |
| size_t size) { |
| LIBC_ASSERT(alignment >= alignof(max_align_t) && |
| alignment % alignof(max_align_t) == 0 && |
| "alignment must be multiple of max_align_t"); |
| |
| if (alignment == alignof(max_align_t)) |
| return size; |
| |
| // We must create a new block inside this one (splitting). This requires a |
| // block header in addition to the requested size. |
| if (add_overflow(size, sizeof(Block), size)) |
| return 0; |
| |
| // Beyond that, padding space may need to remain in this block to ensure |
| // that the usable space of the next block is aligned. |
| // |
| // Consider a position P of some lesser alignment, L, with maximal distance |
| // to the next position of some greater alignment, G, where G is a multiple |
| // of L. P must be one L unit past a G-aligned point. If it were one L-unit |
| // earlier, its distance would be zero. If it were one L-unit later, its |
| // distance would not be maximal. If it were not some integral number of L |
| // units away, it would not be L-aligned. |
| // |
| // So the maximum distance would be G - L. As a special case, if L is 1 |
| // (unaligned), the max distance is G - 1. |
| // |
| // This block's usable space is aligned to max_align_t >= Block. With zero |
| // padding, the next block's usable space is sizeof(Block) past it, which is |
| // a point aligned to Block. Thus the max padding needed is alignment - |
| // alignof(Block). |
| if (add_overflow(size, alignment - alignof(Block), size)) |
| return 0; |
| return size; |
| } |
| |
| // This is the return type for `allocate` which can split one block into up to |
| // three blocks. |
| struct BlockInfo { |
| // This is the newly aligned block. It will have the alignment requested by |
| // a call to `allocate` and at most `size`. |
| Block *block; |
| |
| // If the usable_space in the new block was not aligned according to the |
| // `alignment` parameter, we will need to split into this block and the |
| // `block` to ensure `block` is properly aligned. In this case, `prev` will |
| // be a pointer to this new "padding" block. `prev` will be nullptr if no |
| // new block was created or we were able to merge the block before the |
| // original block with the "padding" block. |
| Block *prev; |
| |
| // This is the remainder of the next block after splitting the `block` |
| // according to `size`. This can happen if there's enough space after the |
| // `block`. |
| Block *next; |
| }; |
| |
| // Divide a block into up to 3 blocks according to `BlockInfo`. Behavior is |
| // undefined if allocation is not possible for the given size and alignment. |
| static BlockInfo allocate(Block *block, size_t alignment, size_t size); |
| |
| // These two functions may wrap around. |
| LIBC_INLINE static uintptr_t next_possible_block_start( |
| uintptr_t ptr, size_t usable_space_alignment = alignof(max_align_t)) { |
| return align_up(ptr + sizeof(Block), usable_space_alignment) - |
| sizeof(Block); |
| } |
| LIBC_INLINE static uintptr_t prev_possible_block_start( |
| uintptr_t ptr, size_t usable_space_alignment = alignof(max_align_t)) { |
| return align_down(ptr, usable_space_alignment) - sizeof(Block); |
| } |
| |
| private: |
| /// Construct a block to represent a span of bytes. Overwrites only enough |
| /// memory for the block header; the rest of the span is left alone. |
| LIBC_INLINE static Block *as_block(ByteSpan bytes) { |
| LIBC_ASSERT(reinterpret_cast<uintptr_t>(bytes.data()) % alignof(Block) == |
| 0 && |
| "block start must be suitably aligned"); |
| return ::new (bytes.data()) Block(bytes.size()); |
| } |
| |
| /// Offset from this block to the previous block. 0 if this is the first |
| /// block. This field is only alive when the previous block is free; |
| /// otherwise, its memory is reused as part of the previous block's usable |
| /// space. |
| size_t prev_ = 0; |
| |
| /// Offset from this block to the next block. Valid even if this is the last |
| /// block, since it equals the size of the block. |
| size_t next_ = 0; |
| |
| /// Information about the current state of the block is stored in the two low |
| /// order bits of the next_ value. These are guaranteed free by a minimum |
| /// alignment (and thus, alignment of the size) of 4. The lowest bit is the |
| /// `prev_free` flag, and the other bit is the `last` flag. |
| /// |
| /// * If the `prev_free` flag is set, the block isn't the first and the |
| /// previous block is free. |
| /// * If the `last` flag is set, the block is the sentinel last block. It is |
| /// summarily considered used and has no next block. |
| |
| public: |
| /// Only for testing. |
| static constexpr size_t PREV_FIELD_SIZE = sizeof(prev_); |
| }; |
| |
| static_assert(alignof(max_align_t) >= 4, |
| "at least 2 bits must be available in block sizes for flags"); |
| |
| LIBC_INLINE |
| optional<Block *> Block::init(ByteSpan region) { |
| if (!region.data()) |
| return {}; |
| |
| uintptr_t start = reinterpret_cast<uintptr_t>(region.data()); |
| uintptr_t end = start + region.size(); |
| if (end < start) |
| return {}; |
| |
| uintptr_t block_start = next_possible_block_start(start); |
| if (block_start < start) |
| return {}; |
| |
| uintptr_t last_start = prev_possible_block_start(end); |
| if (last_start >= end) |
| return {}; |
| |
| if (block_start + sizeof(Block) > last_start) |
| return {}; |
| |
| auto *last_start_ptr = reinterpret_cast<cpp::byte *>(last_start); |
| Block *block = |
| as_block({reinterpret_cast<cpp::byte *>(block_start), last_start_ptr}); |
| Block *last = as_block({last_start_ptr, sizeof(Block)}); |
| block->mark_free(); |
| last->mark_last(); |
| return block; |
| } |
| |
| LIBC_INLINE |
| Block::BlockInfo Block::allocate(Block *block, size_t alignment, size_t size) { |
| LIBC_ASSERT(alignment % alignof(max_align_t) == 0 && |
| "alignment must be a multiple of max_align_t"); |
| |
| BlockInfo info{block, /*prev=*/nullptr, /*next=*/nullptr}; |
| |
| if (!info.block->is_usable_space_aligned(alignment)) { |
| Block *original = info.block; |
| // The padding block has no minimum size requirement. |
| optional<Block *> maybe_aligned_block = original->split(0, alignment); |
| LIBC_ASSERT(maybe_aligned_block.has_value() && |
| "it should always be possible to split for alignment"); |
| |
| if (Block *prev = original->prev_free()) { |
| // If there is a free block before this, we can merge the current one with |
| // the newly created one. |
| prev->merge_next(); |
| } else { |
| info.prev = original; |
| } |
| |
| Block *aligned_block = *maybe_aligned_block; |
| LIBC_ASSERT(aligned_block->is_usable_space_aligned(alignment) && |
| "The aligned block isn't aligned somehow."); |
| info.block = aligned_block; |
| } |
| |
| // Now get a block for the requested size. |
| if (optional<Block *> next = info.block->split(size)) |
| info.next = *next; |
| |
| return info; |
| } |
| |
| LIBC_INLINE |
| optional<Block *> Block::split(size_t new_inner_size, |
| size_t usable_space_alignment) { |
| LIBC_ASSERT(usable_space_alignment % alignof(max_align_t) == 0 && |
| "alignment must be a multiple of max_align_t"); |
| if (used()) |
| return {}; |
| |
| // Compute the minimum outer size that produces a block of at least |
| // `new_inner_size`. |
| size_t min_outer_size = outer_size(cpp::max(new_inner_size, sizeof(prev_))); |
| |
| uintptr_t start = reinterpret_cast<uintptr_t>(this); |
| uintptr_t next_block_start = |
| next_possible_block_start(start + min_outer_size, usable_space_alignment); |
| if (next_block_start < start) |
| return {}; |
| size_t new_outer_size = next_block_start - start; |
| LIBC_ASSERT(new_outer_size % alignof(max_align_t) == 0 && |
| "new size must be aligned to max_align_t"); |
| |
| if (outer_size() < new_outer_size || |
| outer_size() - new_outer_size < sizeof(Block)) |
| return {}; |
| |
| ByteSpan new_region = region().subspan(new_outer_size); |
| next_ &= ~SIZE_MASK; |
| next_ |= new_outer_size; |
| |
| Block *new_block = as_block(new_region); |
| mark_free(); // Free status for this block is now stored in new_block. |
| new_block->next()->prev_ = new_region.size(); |
| |
| LIBC_ASSERT(new_block->is_usable_space_aligned(usable_space_alignment) && |
| "usable space must have requested alignment"); |
| return new_block; |
| } |
| |
| LIBC_INLINE |
| bool Block::merge_next() { |
| if (used() || next()->used()) |
| return false; |
| size_t new_size = outer_size() + next()->outer_size(); |
| next_ &= ~SIZE_MASK; |
| next_ |= new_size; |
| next()->prev_ = new_size; |
| return true; |
| } |
| |
| } // namespace LIBC_NAMESPACE_DECL |
| |
| #endif // LLVM_LIBC_SRC___SUPPORT_BLOCK_H |