mirror of
https://github.com/LadybirdBrowser/ladybird.git
synced 2025-04-21 20:15:17 +00:00
LibC: Statically allocate allocators
These allocators take up 660 bytes, combined. Let's not waste two physical pages for them in each process :^)
This commit is contained in:
parent
bf8dacf0c1
commit
e0ecfc0c92
Notes:
sideshowbarker
2024-07-19 09:15:04 +09:00
Author: https://github.com/bugaevc Commit: https://github.com/SerenityOS/serenity/commit/e0ecfc0c92a Pull-request: https://github.com/SerenityOS/serenity/pull/1240
1 changed files with 23 additions and 13 deletions
|
@ -131,16 +131,31 @@ struct BigAllocator {
|
|||
Vector<BigAllocationBlock*, number_of_big_blocks_to_keep_around_per_size_class> blocks;
|
||||
};
|
||||
|
||||
// Allocators will be mmapped in __malloc_init
|
||||
Allocator* g_allocators = nullptr;
|
||||
BigAllocator* g_big_allocators = nullptr;
|
||||
// Allocators will be initialized in __malloc_init.
|
||||
// We can not rely on global constructors to initialize them,
|
||||
// because they must be initialized before other global constructors
|
||||
// are run. Similarly, we can not allow global destructors to destruct
|
||||
// them. We could have used AK::NeverDestoyed to prevent the latter,
|
||||
// but it would have not helped with the former.
|
||||
static u8 g_allocators_storage[sizeof(Allocator) * num_size_classes];
|
||||
static u8 g_big_allocators_storage[sizeof(BigAllocator)];
|
||||
|
||||
static inline Allocator (&allocators())[num_size_classes]
|
||||
{
|
||||
return reinterpret_cast<Allocator(&)[num_size_classes]>(g_allocators_storage);
|
||||
}
|
||||
|
||||
static inline BigAllocator (&big_allocators())[1]
|
||||
{
|
||||
return reinterpret_cast<BigAllocator(&)[1]>(g_big_allocators_storage);
|
||||
}
|
||||
|
||||
static Allocator* allocator_for_size(size_t size, size_t& good_size)
|
||||
{
|
||||
for (int i = 0; size_classes[i]; ++i) {
|
||||
if (size <= size_classes[i]) {
|
||||
good_size = size_classes[i];
|
||||
return &g_allocators[i];
|
||||
return &allocators()[i];
|
||||
}
|
||||
}
|
||||
good_size = PAGE_ROUND_UP(size);
|
||||
|
@ -150,7 +165,7 @@ static Allocator* allocator_for_size(size_t size, size_t& good_size)
|
|||
static BigAllocator* big_allocator_for_size(size_t size)
|
||||
{
|
||||
if (size == 65536)
|
||||
return &g_big_allocators[0];
|
||||
return &big_allocators()[0];
|
||||
return nullptr;
|
||||
}
|
||||
|
||||
|
@ -414,16 +429,11 @@ void __malloc_init()
|
|||
if (getenv("LIBC_PROFILE_MALLOC"))
|
||||
s_profiling = true;
|
||||
|
||||
g_allocators = (Allocator*)mmap_with_name(nullptr, sizeof(Allocator) * num_size_classes, PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_PRIVATE, 0, 0, "LibC Allocators");
|
||||
for (size_t i = 0; i < num_size_classes; ++i) {
|
||||
new (&g_allocators[i]) Allocator();
|
||||
g_allocators[i].size = size_classes[i];
|
||||
new (&allocators()[i]) Allocator();
|
||||
allocators()[i].size = size_classes[i];
|
||||
}
|
||||
|
||||
g_big_allocators = (BigAllocator*)mmap_with_name(nullptr, sizeof(BigAllocator), PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_PRIVATE, 0, 0, "LibC BigAllocators");
|
||||
new (g_big_allocators)(BigAllocator);
|
||||
|
||||
// We could mprotect the mmaps here with atexit, but, since this method is called in _start before
|
||||
// _init and __init_array entries, our mprotect method would always be the last thing run before _exit.
|
||||
new (&big_allocators()[0])(BigAllocator);
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Add table
Reference in a new issue