From 0836a62fbda7777f34de1738ae1dbeda1ab7a906 Mon Sep 17 00:00:00 2001 From: mintsuki Date: Sun, 26 Jun 2022 04:23:13 +0200 Subject: [PATCH] pmm: Minor overhaul to try and fix #185 --- common/mm/pmm.s2.c | 81 ++++++++++++++++++++++------------------------ 1 file changed, 39 insertions(+), 42 deletions(-) diff --git a/common/mm/pmm.s2.c b/common/mm/pmm.s2.c index 51924723..8ccb7e9b 100644 --- a/common/mm/pmm.s2.c +++ b/common/mm/pmm.s2.c @@ -121,6 +121,8 @@ static bool align_entry(uint64_t *base, uint64_t *length) { static bool sanitiser_keep_first_page = false; +#define MEMMAP_DROP_LATER ((uint32_t)-1) + static void sanitise_entries(struct e820_entry_t *m, size_t *_count, bool align_entries) { size_t count = *_count; @@ -141,10 +143,18 @@ static void sanitise_entries(struct e820_entry_t *m, size_t *_count, bool align_ uint64_t res_length = m[j].length; uint64_t res_top = res_base + res_length; - // TODO actually handle splitting off usable chunks if ( (res_base >= base && res_base < top) && (res_top >= base && res_top < top) ) { - panic(false, "A non-usable e820 entry is inside a usable section."); + // Drop the entry entirely if usable + if (m[j].type == MEMMAP_USABLE) { + m[j].type = MEMMAP_DROP_LATER; + } + if (m[j].type == MEMMAP_DROP_LATER) { + continue; + } + + // TODO actually handle splitting off usable chunks + panic(false, "A non-usable memory map entry is inside a usable section."); } if (res_base >= base && res_base < top) { @@ -167,6 +177,16 @@ static void sanitise_entries(struct e820_entry_t *m, size_t *_count, bool align_ } } + // Collect "drop later" entries + for (size_t i = 0; i < count; i++) { + if (m[i].type != MEMMAP_DROP_LATER) { + continue; + } + + m[i] = m[count - 1]; + count--; i--; + } + // Remove 0 length usable entries and usable entries below 0x1000 for (size_t i = 0; i < count; i++) { if (m[i].type != MEMMAP_USABLE) @@ -642,7 +662,7 @@ void pmm_free(void *ptr, size_t count) { count = ALIGN_UP(count, 4096); if (allocations_disallowed) panic(false, "Memory allocations disallowed"); - memmap_alloc_range((uintptr_t)ptr, count, MEMMAP_USABLE, false, false, false, false); + memmap_alloc_range((uintptr_t)ptr, count, MEMMAP_USABLE, false, false, false, true); } void *ext_mem_alloc(size_t count) { @@ -736,9 +756,7 @@ static bool pmm_new_entry(uint64_t base, uint64_t length, uint32_t type) { // Full overlap if (base <= entry_base && top >= entry_top) { // Remove overlapped entry - for (size_t j = i + 1; j < memmap_entries; j++) { - memmap[j - 1] = memmap[j]; - } + memmap[i] = memmap[memmap_entries - 1]; memmap_entries--; i--; continue; @@ -787,8 +805,6 @@ static bool pmm_new_entry(uint64_t base, uint64_t length, uint32_t type) { target->base = base; target->length = length; - sanitise_entries(memmap, &memmap_entries, false); - return true; } @@ -796,6 +812,10 @@ bool memmap_alloc_range(uint64_t base, uint64_t length, uint32_t type, bool free if (length == 0) return true; + if (simulation && new_entry) { + return true; + } + uint64_t top = base + length; for (size_t i = 0; i < memmap_entries; i++) { @@ -804,52 +824,29 @@ bool memmap_alloc_range(uint64_t base, uint64_t length, uint32_t type, bool free uint64_t entry_base = memmap[i].base; uint64_t entry_top = memmap[i].base + memmap[i].length; - uint32_t entry_type = memmap[i].type; if (base >= entry_base && base < entry_top && top <= entry_top) { if (simulation) return true; - struct e820_entry_t *target; - - memmap[i].length -= entry_top - base; - - if (memmap[i].length == 0) { - target = &memmap[i]; - } else { - if (memmap_entries >= memmap_max_entries) - panic(false, "Memory map exhausted."); - - target = &memmap[memmap_entries++]; + if (pmm_new_entry(base, length, type) == true) { + goto success; } - - target->type = type; - target->base = base; - target->length = length; - - if (top < entry_top) { - if (memmap_entries >= memmap_max_entries) - panic(false, "Memory map exhausted."); - - target = &memmap[memmap_entries++]; - - target->type = entry_type; - target->base = top; - target->length = entry_top - top; - } - - sanitise_entries(memmap, &memmap_entries, false); - - return true; } } if (!new_entry && do_panic) panic(false, "Memory allocation failure."); - if (new_entry) { - return pmm_new_entry(base, length, type); + if (!new_entry) { + return false; } - return false; + if (pmm_new_entry(base, length, type) == false) { + return false; + } + +success: + sanitise_entries(memmap, &memmap_entries, false); + return true; }