mirror of
https://github.com/LadybirdBrowser/ladybird.git
synced 2024-11-11 01:06:01 +03:00
860a3bbce3
This may seem like a no-op change, however it shrinks down the Kernel by a bit: .text -432 .unmap_after_init -60 .data -480 .debug_info -673 .debug_aranges 8 .debug_ranges -232 .debug_line -558 .debug_str -308 .debug_frame -40 With '= default', the compiler can do more inlining, hence the savings. I intentionally omitted some opportunities for '= default', because they would increase the Kernel size.
398 lines
14 KiB
C++
398 lines
14 KiB
C++
/*
|
|
* Copyright (c) 2020, The SerenityOS developers.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* 1. Redistributions of source code must retain the above copyright notice, this
|
|
* list of conditions and the following disclaimer.
|
|
*
|
|
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
|
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
|
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#pragma once
|
|
|
|
#include <AK/Bitmap.h>
|
|
#include <AK/ScopeGuard.h>
|
|
#include <AK/TemporaryChange.h>
|
|
#include <AK/Vector.h>
|
|
#include <AK/kmalloc.h>
|
|
|
|
namespace Kernel {
|
|
|
|
template<size_t CHUNK_SIZE, unsigned HEAP_SCRUB_BYTE_ALLOC = 0, unsigned HEAP_SCRUB_BYTE_FREE = 0>
|
|
class Heap {
|
|
AK_MAKE_NONCOPYABLE(Heap);
|
|
|
|
struct AllocationHeader {
|
|
size_t allocation_size_in_chunks;
|
|
u8 data[0];
|
|
};
|
|
|
|
static size_t calculate_chunks(size_t memory_size)
|
|
{
|
|
return (sizeof(u8) * memory_size) / (sizeof(u8) * CHUNK_SIZE + 1);
|
|
}
|
|
|
|
public:
|
|
Heap(u8* memory, size_t memory_size)
|
|
: m_total_chunks(calculate_chunks(memory_size))
|
|
, m_chunks(memory)
|
|
, m_bitmap(Bitmap::wrap(memory + m_total_chunks * CHUNK_SIZE, m_total_chunks))
|
|
{
|
|
// To keep the alignment of the memory passed in, place the bitmap
|
|
// at the end of the memory block.
|
|
VERIFY(m_total_chunks * CHUNK_SIZE + (m_total_chunks + 7) / 8 <= memory_size);
|
|
}
|
|
~Heap() = default;
|
|
|
|
static size_t calculate_memory_for_bytes(size_t bytes)
|
|
{
|
|
size_t needed_chunks = (sizeof(AllocationHeader) + bytes + CHUNK_SIZE - 1) / CHUNK_SIZE;
|
|
return needed_chunks * CHUNK_SIZE + (needed_chunks + 7) / 8;
|
|
}
|
|
|
|
void* allocate(size_t size)
|
|
{
|
|
// We need space for the AllocationHeader at the head of the block.
|
|
size_t real_size = size + sizeof(AllocationHeader);
|
|
size_t chunks_needed = (real_size + CHUNK_SIZE - 1) / CHUNK_SIZE;
|
|
|
|
if (chunks_needed > free_chunks())
|
|
return nullptr;
|
|
|
|
Optional<size_t> first_chunk;
|
|
|
|
// Choose the right politic for allocation.
|
|
constexpr u32 best_fit_threshold = 128;
|
|
if (chunks_needed < best_fit_threshold) {
|
|
first_chunk = m_bitmap.find_first_fit(chunks_needed);
|
|
} else {
|
|
first_chunk = m_bitmap.find_best_fit(chunks_needed);
|
|
}
|
|
|
|
if (!first_chunk.has_value())
|
|
return nullptr;
|
|
|
|
auto* a = (AllocationHeader*)(m_chunks + (first_chunk.value() * CHUNK_SIZE));
|
|
u8* ptr = a->data;
|
|
a->allocation_size_in_chunks = chunks_needed;
|
|
|
|
m_bitmap.set_range(first_chunk.value(), chunks_needed, true);
|
|
|
|
m_allocated_chunks += chunks_needed;
|
|
if constexpr (HEAP_SCRUB_BYTE_ALLOC != 0) {
|
|
__builtin_memset(ptr, HEAP_SCRUB_BYTE_ALLOC, (chunks_needed * CHUNK_SIZE) - sizeof(AllocationHeader));
|
|
}
|
|
return ptr;
|
|
}
|
|
|
|
void deallocate(void* ptr)
|
|
{
|
|
if (!ptr)
|
|
return;
|
|
auto* a = (AllocationHeader*)((((u8*)ptr) - sizeof(AllocationHeader)));
|
|
VERIFY((u8*)a >= m_chunks && (u8*)ptr < m_chunks + m_total_chunks * CHUNK_SIZE);
|
|
VERIFY((u8*)a + a->allocation_size_in_chunks * CHUNK_SIZE <= m_chunks + m_total_chunks * CHUNK_SIZE);
|
|
FlatPtr start = ((FlatPtr)a - (FlatPtr)m_chunks) / CHUNK_SIZE;
|
|
|
|
m_bitmap.set_range(start, a->allocation_size_in_chunks, false);
|
|
|
|
VERIFY(m_allocated_chunks >= a->allocation_size_in_chunks);
|
|
m_allocated_chunks -= a->allocation_size_in_chunks;
|
|
|
|
if constexpr (HEAP_SCRUB_BYTE_FREE != 0) {
|
|
__builtin_memset(a, HEAP_SCRUB_BYTE_FREE, a->allocation_size_in_chunks * CHUNK_SIZE);
|
|
}
|
|
}
|
|
|
|
template<typename MainHeap>
|
|
void* reallocate(void* ptr, size_t new_size, MainHeap& h)
|
|
{
|
|
if (!ptr)
|
|
return h.allocate(new_size);
|
|
|
|
auto* a = (AllocationHeader*)((((u8*)ptr) - sizeof(AllocationHeader)));
|
|
VERIFY((u8*)a >= m_chunks && (u8*)ptr < m_chunks + m_total_chunks * CHUNK_SIZE);
|
|
VERIFY((u8*)a + a->allocation_size_in_chunks * CHUNK_SIZE <= m_chunks + m_total_chunks * CHUNK_SIZE);
|
|
|
|
size_t old_size = a->allocation_size_in_chunks * CHUNK_SIZE;
|
|
|
|
if (old_size == new_size)
|
|
return ptr;
|
|
|
|
auto* new_ptr = h.allocate(new_size);
|
|
if (new_ptr)
|
|
__builtin_memcpy(new_ptr, ptr, min(old_size, new_size));
|
|
deallocate(ptr);
|
|
return new_ptr;
|
|
}
|
|
|
|
void* reallocate(void* ptr, size_t new_size)
|
|
{
|
|
return reallocate(ptr, new_size, *this);
|
|
}
|
|
|
|
bool contains(const void* ptr) const
|
|
{
|
|
const auto* a = (const AllocationHeader*)((((const u8*)ptr) - sizeof(AllocationHeader)));
|
|
if ((const u8*)a < m_chunks)
|
|
return false;
|
|
if ((const u8*)ptr >= m_chunks + m_total_chunks * CHUNK_SIZE)
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
u8* memory() const { return m_chunks; }
|
|
|
|
size_t total_chunks() const { return m_total_chunks; }
|
|
size_t total_bytes() const { return m_total_chunks * CHUNK_SIZE; }
|
|
size_t free_chunks() const { return m_total_chunks - m_allocated_chunks; };
|
|
size_t free_bytes() const { return free_chunks() * CHUNK_SIZE; }
|
|
size_t allocated_chunks() const { return m_allocated_chunks; }
|
|
size_t allocated_bytes() const { return m_allocated_chunks * CHUNK_SIZE; }
|
|
|
|
private:
|
|
size_t m_total_chunks { 0 };
|
|
size_t m_allocated_chunks { 0 };
|
|
u8* m_chunks { nullptr };
|
|
Bitmap m_bitmap;
|
|
};
|
|
|
|
template<typename ExpandHeap>
|
|
struct ExpandableHeapTraits {
|
|
static bool add_memory(ExpandHeap& expand, size_t allocation_request)
|
|
{
|
|
return expand.add_memory(allocation_request);
|
|
}
|
|
|
|
static bool remove_memory(ExpandHeap& expand, void* memory)
|
|
{
|
|
return expand.remove_memory(memory);
|
|
}
|
|
};
|
|
|
|
struct DefaultExpandHeap {
|
|
bool add_memory(size_t)
|
|
{
|
|
// Requires explicit implementation
|
|
return false;
|
|
}
|
|
|
|
bool remove_memory(void*)
|
|
{
|
|
return false;
|
|
}
|
|
};
|
|
|
|
template<size_t CHUNK_SIZE, unsigned HEAP_SCRUB_BYTE_ALLOC = 0, unsigned HEAP_SCRUB_BYTE_FREE = 0, typename ExpandHeap = DefaultExpandHeap>
|
|
class ExpandableHeap {
|
|
AK_MAKE_NONCOPYABLE(ExpandableHeap);
|
|
AK_MAKE_NONMOVABLE(ExpandableHeap);
|
|
|
|
public:
|
|
typedef ExpandHeap ExpandHeapType;
|
|
typedef Heap<CHUNK_SIZE, HEAP_SCRUB_BYTE_ALLOC, HEAP_SCRUB_BYTE_FREE> HeapType;
|
|
|
|
struct SubHeap {
|
|
HeapType heap;
|
|
SubHeap* next { nullptr };
|
|
size_t memory_size { 0 };
|
|
|
|
template<typename... Args>
|
|
SubHeap(size_t memory_size, Args&&... args)
|
|
: heap(forward<Args>(args)...)
|
|
, memory_size(memory_size)
|
|
{
|
|
}
|
|
};
|
|
|
|
ExpandableHeap(u8* memory, size_t memory_size, const ExpandHeapType& expand = ExpandHeapType())
|
|
: m_heaps(memory_size, memory, memory_size)
|
|
, m_expand(expand)
|
|
{
|
|
}
|
|
~ExpandableHeap()
|
|
{
|
|
// We don't own the main heap, only remove memory that we added previously
|
|
SubHeap* next;
|
|
for (auto* heap = m_heaps.next; heap; heap = next) {
|
|
next = heap->next;
|
|
|
|
heap->~SubHeap();
|
|
ExpandableHeapTraits<ExpandHeap>::remove_memory(m_expand, (void*)heap);
|
|
}
|
|
}
|
|
|
|
static size_t calculate_memory_for_bytes(size_t bytes)
|
|
{
|
|
return sizeof(SubHeap) + HeapType::calculate_memory_for_bytes(bytes);
|
|
}
|
|
|
|
bool expand_memory(size_t size)
|
|
{
|
|
if (m_expanding)
|
|
return false;
|
|
|
|
// Allocating more memory itself may trigger allocations and deallocations
|
|
// on this heap. We need to prevent recursive expansion. We also disable
|
|
// removing memory while trying to expand the heap.
|
|
TemporaryChange change(m_expanding, true);
|
|
return ExpandableHeapTraits<ExpandHeap>::add_memory(m_expand, size);
|
|
}
|
|
|
|
void* allocate(size_t size)
|
|
{
|
|
int attempt = 0;
|
|
do {
|
|
for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) {
|
|
if (void* ptr = subheap->heap.allocate(size))
|
|
return ptr;
|
|
}
|
|
|
|
// We need to loop because we won't know how much memory was added.
|
|
// Even though we make a best guess how much memory needs to be added,
|
|
// it doesn't guarantee that enough will be available after adding it.
|
|
// This is especially true for the kmalloc heap, where adding memory
|
|
// requires several other objects to be allocated just to be able to
|
|
// expand the heap.
|
|
|
|
// To avoid an infinite expansion loop, limit to two attempts
|
|
if (attempt++ >= 2)
|
|
break;
|
|
} while (expand_memory(size));
|
|
return nullptr;
|
|
}
|
|
|
|
void deallocate(void* ptr)
|
|
{
|
|
if (!ptr)
|
|
return;
|
|
for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) {
|
|
if (subheap->heap.contains(ptr)) {
|
|
subheap->heap.deallocate(ptr);
|
|
if (subheap->heap.allocated_chunks() == 0 && subheap != &m_heaps && !m_expanding) {
|
|
// remove_memory expects the memory to be unused and
|
|
// may deallocate the memory. We need to therefore first
|
|
// unlink the subheap and destroy it. If remove_memory
|
|
// ends up not not removing the memory, we'll initialize
|
|
// a new subheap and re-add it.
|
|
// We need to remove the subheap before calling remove_memory
|
|
// because it's possible that remove_memory itself could
|
|
// cause a memory allocation that we don't want to end up
|
|
// potentially being made in the subheap we're about to remove.
|
|
{
|
|
auto* subheap2 = m_heaps.next;
|
|
auto** subheap_link = &m_heaps.next;
|
|
while (subheap2 != subheap) {
|
|
subheap_link = &subheap2->next;
|
|
subheap2 = subheap2->next;
|
|
}
|
|
*subheap_link = subheap->next;
|
|
}
|
|
|
|
auto memory_size = subheap->memory_size;
|
|
subheap->~SubHeap();
|
|
|
|
if (!ExpandableHeapTraits<ExpandHeap>::remove_memory(m_expand, subheap)) {
|
|
// Removal of the subheap was rejected, add it back in and
|
|
// re-initialize with a clean subheap.
|
|
add_subheap(subheap, memory_size);
|
|
}
|
|
}
|
|
return;
|
|
}
|
|
}
|
|
VERIFY_NOT_REACHED();
|
|
}
|
|
|
|
void* reallocate(void* ptr, size_t new_size)
|
|
{
|
|
if (!ptr)
|
|
return allocate(new_size);
|
|
for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) {
|
|
if (subheap->heap.contains(ptr))
|
|
return subheap->heap.reallocate(ptr, new_size, *this);
|
|
}
|
|
VERIFY_NOT_REACHED();
|
|
}
|
|
|
|
HeapType& add_subheap(void* memory, size_t memory_size)
|
|
{
|
|
VERIFY(memory_size > sizeof(SubHeap));
|
|
|
|
// Place the SubHeap structure at the beginning of the new memory block
|
|
memory_size -= sizeof(SubHeap);
|
|
SubHeap* new_heap = (SubHeap*)memory;
|
|
new (new_heap) SubHeap(memory_size, (u8*)(new_heap + 1), memory_size);
|
|
|
|
// Add the subheap to the list (but leave the main heap where it is)
|
|
SubHeap* next_heap = m_heaps.next;
|
|
SubHeap** next_heap_link = &m_heaps.next;
|
|
while (next_heap) {
|
|
if (new_heap->heap.memory() < next_heap->heap.memory())
|
|
break;
|
|
next_heap_link = &next_heap->next;
|
|
next_heap = next_heap->next;
|
|
}
|
|
new_heap->next = *next_heap_link;
|
|
*next_heap_link = new_heap;
|
|
return new_heap->heap;
|
|
}
|
|
|
|
bool contains(const void* ptr) const
|
|
{
|
|
for (auto* subheap = &m_heaps; subheap; subheap = subheap->next) {
|
|
if (subheap->heap.contains(ptr))
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
size_t total_chunks() const
|
|
{
|
|
size_t total = 0;
|
|
for (auto* subheap = &m_heaps; subheap; subheap = subheap->next)
|
|
total += subheap->heap.total_chunks();
|
|
return total;
|
|
}
|
|
size_t total_bytes() const { return total_chunks() * CHUNK_SIZE; }
|
|
size_t free_chunks() const
|
|
{
|
|
size_t total = 0;
|
|
for (auto* subheap = &m_heaps; subheap; subheap = subheap->next)
|
|
total += subheap->heap.free_chunks();
|
|
return total;
|
|
}
|
|
size_t free_bytes() const { return free_chunks() * CHUNK_SIZE; }
|
|
size_t allocated_chunks() const
|
|
{
|
|
size_t total = 0;
|
|
for (auto* subheap = &m_heaps; subheap; subheap = subheap->next)
|
|
total += subheap->heap.allocated_chunks();
|
|
return total;
|
|
}
|
|
size_t allocated_bytes() const { return allocated_chunks() * CHUNK_SIZE; }
|
|
|
|
private:
|
|
SubHeap m_heaps;
|
|
ExpandHeap m_expand;
|
|
bool m_expanding { false };
|
|
};
|
|
|
|
}
|