rulimine/stage23/mm/pmm.s2.c

454 lines
13 KiB
C
Raw Normal View History

2020-06-05 18:51:33 +03:00
#include <stddef.h>
#include <stdint.h>
#include <stdbool.h>
2020-09-20 13:03:44 +03:00
#include <mm/pmm.h>
2020-09-18 21:02:10 +03:00
#include <sys/e820.h>
2020-06-05 18:51:33 +03:00
#include <lib/blib.h>
2020-09-20 13:03:44 +03:00
#include <lib/libc.h>
2020-06-05 21:57:09 +03:00
#include <lib/print.h>
2021-03-02 12:23:43 +03:00
#if defined (uefi)
# include <efi.h>
#endif
2020-06-05 18:51:33 +03:00
#define PAGE_SIZE 4096
#define MEMMAP_BASE ((size_t)0x100000)
#define MEMMAP_MAX_ENTRIES 256
#if defined (bios)
extern symbol bss_end;
static size_t bump_allocator_base = (size_t)bss_end;
static size_t bump_allocator_limit = 0x70000;
#endif
#if defined (uefi)
static size_t bump_allocator_base = 0x100000;
static size_t bump_allocator_limit = 0x100000;
#endif
void *conv_mem_alloc(size_t count) {
return conv_mem_alloc_aligned(count, 4);
}
void *conv_mem_alloc_aligned(size_t count, size_t alignment) {
size_t new_base = ALIGN_UP(bump_allocator_base, alignment);
void *ret = (void *)new_base;
new_base += count;
if (new_base >= bump_allocator_limit)
panic("Memory allocation failed");
bump_allocator_base = new_base;
// Zero out allocated space
memset(ret, 0, count);
return ret;
}
struct e820_entry_t memmap[MEMMAP_MAX_ENTRIES];
size_t memmap_entries = 0;
2020-06-05 18:51:33 +03:00
2020-06-05 21:57:09 +03:00
static const char *memmap_type(uint32_t type) {
switch (type) {
case MEMMAP_USABLE:
2020-06-05 21:57:09 +03:00
return "Usable RAM";
case MEMMAP_RESERVED:
2020-06-05 21:57:09 +03:00
return "Reserved";
case MEMMAP_ACPI_RECLAIMABLE:
2020-06-05 21:57:09 +03:00
return "ACPI reclaimable";
case MEMMAP_ACPI_NVS:
2020-06-05 21:57:09 +03:00
return "ACPI NVS";
case MEMMAP_BAD_MEMORY:
2020-06-05 21:57:09 +03:00
return "Bad memory";
case MEMMAP_BOOTLOADER_RECLAIMABLE:
return "Bootloader reclaimable";
case MEMMAP_KERNEL_AND_MODULES:
2020-06-05 21:57:09 +03:00
return "Kernel/Modules";
2021-03-05 08:10:58 +03:00
case MEMMAP_EFI_RECLAIMABLE:
return "EFI reclaimable";
2020-06-05 21:57:09 +03:00
default:
return "???";
}
}
void print_memmap(struct e820_entry_t *mm, size_t size) {
for (size_t i = 0; i < size; i++) {
2020-06-05 22:34:11 +03:00
print("[%X -> %X] : %X <%s>\n",
2020-06-05 21:57:09 +03:00
mm[i].base,
mm[i].base + mm[i].length,
mm[i].length,
memmap_type(mm[i].type));
}
}
static bool align_entry(uint64_t *base, uint64_t *length) {
2020-06-06 17:52:21 +03:00
if (*length < PAGE_SIZE)
return false;
2020-06-05 22:34:11 +03:00
2020-06-06 17:52:21 +03:00
uint64_t orig_base = *base;
2020-06-05 22:34:11 +03:00
2020-06-06 17:52:21 +03:00
*base = ALIGN_UP(*base, PAGE_SIZE);
2020-06-05 18:51:33 +03:00
2020-06-06 17:52:21 +03:00
*length -= (*base - orig_base);
*length = ALIGN_DOWN(*length, PAGE_SIZE);
2020-06-05 18:51:33 +03:00
if (!length)
return false;
2020-06-05 18:51:33 +03:00
uint64_t top = *base + *length;
if (*base < MEMMAP_BASE) {
if (top > MEMMAP_BASE) {
*length -= MEMMAP_BASE - *base;
*base = MEMMAP_BASE;
} else {
return false;
2020-06-05 18:51:33 +03:00
}
}
return true;
2020-06-05 18:51:33 +03:00
}
static void sanitise_entries(bool align_entries) {
2020-06-05 18:51:33 +03:00
for (size_t i = 0; i < memmap_entries; i++) {
if (memmap[i].type != MEMMAP_USABLE)
2020-06-05 18:51:33 +03:00
continue;
2020-06-06 17:52:21 +03:00
// Check if the entry overlaps other entries
for (size_t j = 0; j < memmap_entries; j++) {
if (j == i)
continue;
uint64_t base = memmap[i].base;
uint64_t length = memmap[i].length;
uint64_t top = base + length;
uint64_t res_base = memmap[j].base;
uint64_t res_length = memmap[j].length;
uint64_t res_top = res_base + res_length;
// TODO actually handle splitting off usable chunks
if ( (res_base >= base && res_base < top)
&& (res_top >= base && res_top < top) ) {
panic("A non-usable e820 entry is inside a usable section.");
}
if (res_base >= base && res_base < top) {
top = res_base;
}
if (res_top >= base && res_top < top) {
base = res_top;
}
memmap[i].base = base;
memmap[i].length = top - base;
}
if (!memmap[i].length
|| (align_entries && !align_entry(&memmap[i].base, &memmap[i].length))) {
2020-06-05 18:51:33 +03:00
// Eradicate from memmap
for (size_t j = i; j < memmap_entries - 1; j++) {
memmap[j] = memmap[j+1];
}
memmap_entries--;
2020-06-06 17:52:21 +03:00
i--;
2020-06-05 18:51:33 +03:00
}
}
2020-06-05 20:09:57 +03:00
// Sort the entries
for (size_t p = 0; p < memmap_entries - 1; p++) {
uint64_t min = memmap[p].base;
size_t min_index = p;
for (size_t i = p; i < memmap_entries; i++) {
if (memmap[i].base < min) {
min = memmap[i].base;
min_index = i;
}
}
struct e820_entry_t min_e = memmap[min_index];
memmap[min_index] = memmap[p];
memmap[p] = min_e;
}
// Merge contiguous bootloader-reclaimable entries
for (size_t i = 0; i < memmap_entries - 1; i++) {
if (memmap[i].type != MEMMAP_BOOTLOADER_RECLAIMABLE)
continue;
if (memmap[i+1].type == MEMMAP_BOOTLOADER_RECLAIMABLE
&& memmap[i+1].base == memmap[i].base + memmap[i].length) {
memmap[i].length += memmap[i+1].length;
// Eradicate from memmap
for (size_t j = i+1; j < memmap_entries - 1; j++) {
memmap[j] = memmap[j+1];
}
memmap_entries--;
i--;
}
}
// Align bootloader-reclaimable entries
if (align_entries) {
for (size_t i = 0; i < memmap_entries; i++) {
if (memmap[i].type != MEMMAP_BOOTLOADER_RECLAIMABLE)
continue;
if (!align_entry(&memmap[i].base, &memmap[i].length)) {
// Eradicate from memmap
for (size_t j = i; j < memmap_entries - 1; j++) {
memmap[j] = memmap[j+1];
}
memmap_entries--;
i--;
}
}
}
2020-06-06 17:52:21 +03:00
}
static bool allocations_disallowed = true;
2020-06-06 17:52:21 +03:00
struct e820_entry_t *get_memmap(size_t *entries) {
sanitise_entries(true);
2020-06-05 20:09:57 +03:00
2020-06-05 18:51:33 +03:00
*entries = memmap_entries;
2020-06-05 21:57:09 +03:00
allocations_disallowed = true;
2020-06-05 18:51:33 +03:00
return memmap;
}
2021-03-02 12:23:43 +03:00
#if defined (bios)
2020-06-05 18:51:33 +03:00
void init_memmap(void) {
for (size_t i = 0; i < e820_entries; i++) {
if (memmap_entries == MEMMAP_MAX_ENTRIES) {
panic("Memory map exhausted.");
}
2020-06-06 17:52:21 +03:00
memmap[memmap_entries++] = e820_map[i];
2020-06-05 18:51:33 +03:00
}
2020-06-06 17:52:21 +03:00
sanitise_entries(false);
allocations_disallowed = false;
2020-06-05 18:51:33 +03:00
}
2021-03-02 12:23:43 +03:00
#endif
2020-06-05 18:51:33 +03:00
2021-03-05 08:10:58 +03:00
#if defined (uefi)
void init_memmap(void) {
EFI_STATUS status;
2021-03-05 08:10:58 +03:00
EFI_MEMORY_DESCRIPTOR tmp_mmap[1];
UINTN mmap_size = sizeof(tmp_mmap);
UINTN mmap_key = 0, desc_size = 0, desc_ver = 0;
2021-03-05 08:10:58 +03:00
status = uefi_call_wrapper(gBS->GetMemoryMap, 5,
&mmap_size, tmp_mmap, &mmap_key, &desc_size, &desc_ver);
2021-03-05 08:10:58 +03:00
EFI_MEMORY_DESCRIPTOR *efi_mmap;
2021-03-05 08:10:58 +03:00
mmap_size += 4096;
2020-09-20 13:03:44 +03:00
2021-03-05 08:10:58 +03:00
status = uefi_call_wrapper(gBS->AllocatePool, 3,
EfiLoaderData, mmap_size, &efi_mmap);
2020-09-20 13:03:44 +03:00
2021-03-05 08:10:58 +03:00
status = uefi_call_wrapper(gBS->GetMemoryMap, 5,
&mmap_size, efi_mmap, &mmap_key, &desc_size, &desc_ver);
2021-03-05 08:10:58 +03:00
size_t entry_count = mmap_size / desc_size;
2021-03-02 12:23:43 +03:00
for (size_t i = 0; i < entry_count; i++) {
EFI_MEMORY_DESCRIPTOR *entry = (void *)efi_mmap + i * desc_size;
uint32_t our_type;
switch (entry->Type) {
case EfiReservedMemoryType:
case EfiRuntimeServicesCode:
case EfiRuntimeServicesData:
case EfiUnusableMemory:
case EfiMemoryMappedIO:
case EfiMemoryMappedIOPortSpace:
case EfiPalCode:
default:
our_type = MEMMAP_RESERVED; break;
case EfiBootServicesCode:
case EfiBootServicesData:
case EfiLoaderCode:
case EfiLoaderData:
our_type = MEMMAP_BOOTLOADER_RECLAIMABLE; break;
case EfiACPIReclaimMemory:
our_type = MEMMAP_ACPI_RECLAIMABLE; break;
case EfiACPIMemoryNVS:
our_type = MEMMAP_ACPI_NVS; break;
case EfiConventionalMemory:
our_type = MEMMAP_USABLE; break;
}
memmap[memmap_entries].type = our_type;
memmap[memmap_entries].base = entry->PhysicalStart;
memmap[memmap_entries].length = entry->NumberOfPages * 4096;
if (our_type == MEMMAP_USABLE
&& entry->PhysicalStart < bump_allocator_base) {
bump_allocator_base = entry->PhysicalStart;
bump_allocator_limit =
entry->PhysicalStart + entry->NumberOfPages * 4096;
if (bump_allocator_limit > 0x100000)
bump_allocator_limit = 0x100000;
}
memmap_entries++;
}
sanitise_entries(false);
allocations_disallowed = false;
2021-03-05 08:10:58 +03:00
// Let's leave 64MiB to the firmware
ext_mem_alloc_aligned_type(65536, 4096, MEMMAP_EFI_RECLAIMABLE);
// Now own all the usable entries
for (size_t i = 0; i < memmap_entries; i++) {
if (memmap[i].type != MEMMAP_USABLE)
continue;
EFI_PHYSICAL_ADDRESS base = memmap[i].base;
status = uefi_call_wrapper(gBS->AllocatePages, 4,
AllocateAddress, EfiLoaderData, memmap[i].length / 4096, &base);
if (status)
panic("AllocatePages %x", status);
}
2021-03-07 06:31:06 +03:00
memmap_alloc_range(bump_allocator_base,
bump_allocator_limit - bump_allocator_base,
MEMMAP_RESERVED, true, true);
print("pmm: Conventional mem allocator base: %X\n", bump_allocator_base);
print("pmm: Conventional mem allocator limit: %X\n", bump_allocator_limit);
2021-03-05 08:10:58 +03:00
}
#endif
void *ext_mem_alloc(size_t count) {
return ext_mem_alloc_type(count, MEMMAP_BOOTLOADER_RECLAIMABLE);
}
void *ext_mem_alloc_aligned(size_t count, size_t alignment) {
return ext_mem_alloc_aligned_type(count, alignment, MEMMAP_BOOTLOADER_RECLAIMABLE);
}
2021-03-05 08:10:58 +03:00
void *ext_mem_alloc_type(size_t count, uint32_t type) {
return ext_mem_alloc_aligned_type(count, 4, type);
}
// Allocate memory top down, hopefully without bumping into kernel or modules
2021-03-02 12:23:43 +03:00
void *ext_mem_alloc_aligned_type(size_t count, size_t alignment, uint32_t type) {
2021-03-05 08:10:58 +03:00
if (allocations_disallowed)
panic("Extended memory allocations disallowed");
2021-03-02 12:23:43 +03:00
2021-03-05 08:10:58 +03:00
for (int i = memmap_entries - 1; i >= 0; i--) {
if (memmap[i].type != 1)
continue;
int64_t entry_base = (int64_t)(memmap[i].base);
int64_t entry_top = (int64_t)(memmap[i].base + memmap[i].length);
2021-03-04 00:38:28 +03:00
2021-03-05 08:10:58 +03:00
// Let's make sure the entry is not > 4GiB
if (entry_base >= 0x100000000 || entry_top >= 0x100000000) {
// Theoretically there could be an entry which crosses the 4GiB
// boundary, but realistically this does not happen as far as I
// have seen. Let's just discard the entry.
continue;
}
2021-03-02 12:23:43 +03:00
2021-03-05 08:10:58 +03:00
int64_t alloc_base = ALIGN_DOWN(entry_top - (int64_t)count, alignment);
2021-03-04 00:38:28 +03:00
2021-03-05 08:10:58 +03:00
// This entry is too small for us.
if (alloc_base < entry_base)
continue;
2021-03-05 08:10:58 +03:00
// We now reserve the range we need.
int64_t aligned_length = entry_top - alloc_base;
memmap_alloc_range((uint64_t)alloc_base, (uint64_t)aligned_length, type, true, true);
void *ret = (void *)(size_t)alloc_base;
// Zero out allocated space
memset(ret, 0, count);
2021-03-02 12:23:43 +03:00
2021-03-05 08:10:58 +03:00
sanitise_entries(false);
2021-03-04 00:38:28 +03:00
2021-03-05 08:10:58 +03:00
return ret;
}
panic("High memory allocator: Out of memory");
2021-03-02 12:23:43 +03:00
}
2020-09-02 06:33:22 +03:00
2021-01-27 18:16:36 +03:00
bool memmap_alloc_range(uint64_t base, uint64_t length, uint32_t type, bool free_only, bool do_panic) {
2020-06-05 18:51:33 +03:00
uint64_t top = base + length;
2021-03-07 06:31:06 +03:00
#if defined (bios)
if (base < 0x100000) {
2021-01-27 18:16:36 +03:00
if (do_panic) {
// We don't do allocations below 1 MiB
panic("Attempt to allocate memory below 1 MiB (%X-%X)",
base, base + length);
} else {
return false;
}
}
2021-03-07 06:31:06 +03:00
#endif
2020-06-05 18:51:33 +03:00
for (size_t i = 0; i < memmap_entries; i++) {
if (free_only && memmap[i].type != 1)
2020-06-05 18:51:33 +03:00
continue;
uint64_t entry_base = memmap[i].base;
uint64_t entry_top = memmap[i].base + memmap[i].length;
if (base >= entry_base && base < entry_top &&
top >= entry_base && top <= entry_top) {
2020-06-05 18:51:33 +03:00
memmap[i].length = base - entry_base;
if (memmap[i].length == 0) {
// Eradicate from memmap
for (size_t j = i; j < memmap_entries - 1; j++) {
memmap[j] = memmap[j+1];
}
memmap_entries--;
}
if (memmap_entries >= MEMMAP_MAX_ENTRIES) {
panic("Memory map exhausted.");
}
struct e820_entry_t *target = &memmap[memmap_entries];
target->length = entry_top - top;
if (target->length != 0) {
target->base = top;
target->type = 1;
memmap_entries++;
}
if (memmap_entries >= MEMMAP_MAX_ENTRIES) {
panic("Memory map exhausted.");
}
target = &memmap[memmap_entries++];
2020-08-11 19:00:51 +03:00
target->type = type;
2020-06-05 18:51:33 +03:00
target->base = base;
target->length = length;
2021-01-27 18:16:36 +03:00
return true;
2020-06-05 18:51:33 +03:00
}
}
2021-01-27 18:16:36 +03:00
if (do_panic)
panic("Out of memory");
return false;
2020-06-05 21:57:09 +03:00
}