#include #include "kutil/assert.h" #include "kutil/memory_manager.h" #include "console.h" #include "log.h" #include "page_manager.h" page_manager g_page_manager; kutil::memory_manager g_kernel_memory_manager; static addr_t pt_to_phys(page_table *pt) { return reinterpret_cast(pt) - page_manager::page_offset; } static page_table * pt_from_phys(addr_t p) { return reinterpret_cast((p + page_manager::page_offset) & ~0xfffull); } struct free_page_header { free_page_header *next; size_t count; }; void mm_grow_callback(void *next, size_t length) { kassert(length % page_manager::page_size == 0, "Heap manager requested a fractional page."); size_t pages = length / page_manager::page_size; log::info(logs::memory, "Heap manager growing heap by %d pages.", pages); g_page_manager.map_pages(reinterpret_cast(next), pages); } size_t page_block::length(page_block *list) { size_t i = 0; for (page_block *b = list; b; b = b->next) ++i; return i; } page_block * page_block::append(page_block *list, page_block *extra) { if (list == nullptr) return extra; else if (extra == nullptr) return list; page_block *cur = list; while (cur->next) cur = cur->next; cur->next = extra; return list; } page_block * page_block::insert(page_block *list, page_block *block) { if (list == nullptr) return block; else if (block == nullptr) return list; page_block *cur = list; page_block *prev = nullptr; while (cur && page_block::compare(block, cur) > 0) { prev = cur; cur = cur->next; } block->next = cur; if (prev) { prev->next = block; return list; } return block; } int page_block::compare(const page_block *lhs, const page_block *rhs) { if (lhs->virtual_address < rhs->virtual_address) return -1; else if (lhs->virtual_address > rhs->virtual_address) return 1; if (lhs->physical_address < rhs->physical_address) return -1; else if (lhs->physical_address > rhs->physical_address) return 1; return 0; } page_block * page_block::consolidate(page_block *list) { page_block *freed = nullptr; page_block *cur = list; while (cur) { page_block *next = cur->next; if (next && cur->flags == next->flags && cur->physical_end() == next->physical_address && (!cur->has_flag(page_block_flags::mapped) || cur->virtual_end() == next->virtual_address)) { cur->count += next->count; cur->next = next->next; next->zero(freed); freed = next; continue; } cur = cur->next; } return freed; } void page_block::dump(page_block *list, const char *name, bool show_unmapped) { log::info(logs::memory, "Block list %s:", name); int count = 0; for (page_block *cur = list; cur; cur = cur->next) { count += 1; if (!(show_unmapped || cur->has_flag(page_block_flags::mapped))) continue; if (cur->virtual_address) { page_table_indices start{cur->virtual_address}; log::info(logs::memory, " %016lx %08x [%6d] %016lx (%d,%d,%d,%d)", cur->physical_address, cur->flags, cur->count, cur->virtual_address, start[0], start[1], start[2], start[3]); } else { page_table_indices start{cur->virtual_address}; log::info(logs::memory, " %016lx %08x [%6d]", cur->physical_address, cur->flags, cur->count); } } log::info(logs::memory, " Total: %d", count); } void page_block::zero(page_block *set_next) { physical_address = 0; virtual_address = 0; count = 0; flags = page_block_flags::free; next = set_next; } void page_block::copy(page_block *other) { physical_address = other->physical_address; virtual_address = other->virtual_address; count = other->count; flags = other->flags; next = other->next; } page_manager::page_manager() : m_free(nullptr), m_used(nullptr), m_block_cache(nullptr), m_page_cache(nullptr) { kassert(this == &g_page_manager, "Attempt to create another page_manager."); } void page_manager::init( page_block *free, page_block *used, page_block *block_cache) { m_free = free; m_used = used; m_block_cache = block_cache; // For now we're ignoring that we've got the scratch pages // allocated, full of page_block structs. Eventually hand // control of that to a slab allocator. consolidate_blocks(); // Initialize the kernel memory manager addr_t end = 0; for (page_block *b = m_used; b; b = b->next) { if (b->virtual_address < page_offset) { end = b->virtual_end(); } else { break; } } new (&g_kernel_memory_manager) kutil::memory_manager( reinterpret_cast(end), mm_grow_callback); kutil::setup::set_heap(&g_kernel_memory_manager); m_kernel_pml4 = get_pml4(); } page_table * page_manager::create_process_map() { page_table *table = get_table_page(); kutil::memset(table, 0, page_size); table->entries[510] = m_kernel_pml4->entries[510]; table->entries[511] = m_kernel_pml4->entries[511]; // Create the initial user stack map_pages( initial_stack - (initial_stack_pages * page_size), initial_stack_pages, true, // This is the ring3 stack, user = true table); return table; } void page_manager::delete_process_map(page_table *table) { // TODO: recurse table entries and mark them free unmap_pages(table, 1); } void page_manager::map_offset_pointer(void **pointer, size_t length) { addr_t *p = reinterpret_cast(pointer); addr_t v = *p + page_offset; addr_t c = ((length - 1) / page_size) + 1; // TODO: cleanly search/split this as a block out of used/free if possible page_block *block = get_block(); // TODO: page-align block->physical_address = *p; block->virtual_address = v; block->count = c; block->flags = page_block_flags::used | page_block_flags::mapped | page_block_flags::mmio; m_used = page_block::insert(m_used, block); page_table *pml4 = get_pml4(); page_in(pml4, *p, v, c); *p = v; } void page_manager::dump_blocks(bool used_only) { page_block::dump(m_used, "used", true); if (!used_only) page_block::dump(m_free, "free", true); } void page_manager::dump_pml4(page_table *pml4, int max_index) { if (pml4 == nullptr) pml4 = get_pml4(); pml4->dump(4, max_index); } page_block * page_manager::get_block() { page_block *block = m_block_cache; if (block) { m_block_cache = block->next; block->next = 0; return block; } else { kassert(0, "NYI: page_manager::get_block() needed to allocate."); return nullptr; } } void page_manager::free_blocks(page_block *block) { if (!block) return; page_block *cur = block; while (cur) { page_block *next = cur->next; cur->zero(cur->next ? cur->next : m_block_cache); cur = next; } m_block_cache = block; } page_table * page_manager::get_table_page() { if (!m_page_cache) { addr_t phys = 0; size_t n = pop_pages(32, &phys); addr_t virt = phys + page_offset; page_block *block = get_block(); block->physical_address = phys; block->virtual_address = virt; block->count = n; page_block::insert(m_used, block); page_in(get_pml4(), phys, virt, n); m_page_cache = reinterpret_cast(virt); // The last one needs to be null, so do n-1 addr_t end = virt + (n-1) * page_size; while (virt < end) { reinterpret_cast(virt)->next = reinterpret_cast(virt + page_size); virt += page_size; } reinterpret_cast(virt)->next = nullptr; log::info(logs::memory, "Mappd %d new page table pages at %lx", n, phys); } free_page_header *page = m_page_cache; m_page_cache = page->next; return reinterpret_cast(page); } void page_manager::free_table_pages(void *pages, size_t count) { addr_t start = reinterpret_cast(pages); for (size_t i = 0; i < count; ++i) { addr_t addr = start + (i * page_size); free_page_header *header = reinterpret_cast(addr); header->count = 1; header->next = m_page_cache; m_page_cache = header; } } void page_manager::consolidate_blocks() { m_block_cache = page_block::append(m_block_cache, page_block::consolidate(m_free)); m_block_cache = page_block::append(m_block_cache, page_block::consolidate(m_used)); } void * page_manager::map_pages(addr_t address, size_t count, bool user, page_table *pml4) { void *ret = reinterpret_cast(address); if (!pml4) pml4 = get_pml4(); while (count) { kassert(m_free, "page_manager::map_pages ran out of free pages!"); addr_t phys = 0; size_t n = pop_pages(count, &phys); page_block *block = get_block(); block->physical_address = phys; block->virtual_address = address; block->count = n; block->flags = page_block_flags::used | page_block_flags::mapped; page_block::insert(m_used, block); log::debug(logs::memory, "Paging in %d pages at p:%016lx to v:%016lx into %016lx table", n, phys, address, pml4); page_in(pml4, phys, address, n, user); address += n * page_size; count -= n; } return ret; } void * page_manager::map_offset_pages(size_t count) { page_table *pml4 = get_pml4(); page_block *free = m_free; page_block *prev = nullptr; while (free) { if (free->count < count) { prev = free; free = free->next; continue; } page_block *used = get_block(); used->count = count; used->physical_address = free->physical_address; used->virtual_address = used->physical_address + page_offset; used->flags = page_block_flags::used | page_block_flags::mapped; page_block::insert(m_used, used); free->physical_address += count * page_size; free->count -= count; if (free->count == 0) { if (prev) prev->next = free->next; else m_free = free->next; free->zero(m_block_cache); m_block_cache = free; } log::debug(logs::memory, "Got request for offset map %016lx [%d]", used->virtual_address, count); page_in(pml4, used->physical_address, used->virtual_address, count); return reinterpret_cast(used->virtual_address); } return nullptr; } void page_manager::unmap_pages(void* address, size_t count) { addr_t addr = reinterpret_cast(address); page_block **prev = &m_used; page_block *cur = m_used; while (cur && !cur->contains(addr)) { prev = &cur->next; cur = cur->next; } kassert(cur, "Couldn't find existing mapped pages to unmap"); size_t size = page_size * count; addr_t end = addr + size; while (cur && cur->contains(addr)) { size_t leading = addr - cur->virtual_address; size_t trailing = end > cur->virtual_end() ? 0 : (cur->virtual_end() - end); if (leading) { size_t pages = leading / page_size; page_block *lead_block = get_block(); lead_block->copy(cur); lead_block->next = cur; lead_block->count = pages; cur->count -= pages; cur->physical_address += leading; cur->virtual_address += leading; *prev = lead_block; prev = &lead_block->next; } if (trailing) { size_t pages = trailing / page_size; page_block *trail_block = get_block(); trail_block->copy(cur); trail_block->next = cur->next; trail_block->count = pages; trail_block->physical_address += size; trail_block->virtual_address += size; cur->count -= pages; cur->next = trail_block; } addr += cur->count * page_size; page_block *next = cur->next; *prev = cur->next; cur->next = nullptr; cur->virtual_address = 0; cur->flags = cur->flags & ~(page_block_flags::used | page_block_flags::mapped); m_free = page_block::insert(m_free, cur); cur = next; } } void page_manager::check_needs_page(page_table *table, unsigned index, bool user) { if ((table->entries[index] & 0x1) == 1) return; page_table *new_table = get_table_page(); for (int i=0; i<512; ++i) new_table->entries[i] = 0; table->entries[index] = pt_to_phys(new_table) | (user ? 0xf : 0xb); } void page_manager::page_in(page_table *pml4, addr_t phys_addr, addr_t virt_addr, size_t count, bool user) { page_table_indices idx{virt_addr}; page_table *tables[4] = {pml4, nullptr, nullptr, nullptr}; uint64_t flags = user ? 0x00f: // writethru, user, write, present 0x10b; // global, writethru, write, present for (; idx[0] < 512; idx[0] += 1) { check_needs_page(tables[0], idx[0], user); tables[1] = tables[0]->get(idx[0]); for (; idx[1] < 512; idx[1] += 1, idx[2] = 0, idx[3] = 0) { check_needs_page(tables[1], idx[1], user); tables[2] = tables[1]->get(idx[1]); for (; idx[2] < 512; idx[2] += 1, idx[3] = 0) { if (idx[3] == 0 && count >= 512 && tables[2]->get(idx[2]) == nullptr) { // Do a 2MiB page instead tables[2]->entries[idx[2]] = phys_addr | flags | 0x80; phys_addr += page_size * 512; count -= 512; if (count == 0) return; continue; } check_needs_page(tables[2], idx[2], user); tables[3] = tables[2]->get(idx[2]); for (; idx[3] < 512; idx[3] += 1) { tables[3]->entries[idx[3]] = phys_addr | flags; phys_addr += page_size; if (--count == 0) return; } } } } kassert(0, "Ran to end of page_in"); } void page_manager::page_out(page_table *pml4, addr_t virt_addr, size_t count) { page_table_indices idx{virt_addr}; page_table *tables[4] = {pml4, nullptr, nullptr, nullptr}; for (; idx[0] < 512; idx[0] += 1) { tables[1] = reinterpret_cast( tables[0]->entries[idx[0]] & ~0xfffull); for (; idx[1] < 512; idx[1] += 1) { tables[2] = reinterpret_cast( tables[1]->entries[idx[1]] & ~0xfffull); for (; idx[2] < 512; idx[2] += 1) { tables[3] = reinterpret_cast( tables[2]->entries[idx[2]] & ~0xfffull); for (; idx[3] < 512; idx[3] += 1) { tables[3]->entries[idx[3]] = 0; if (--count == 0) return; } } } } kassert(0, "Ran to end of page_out"); } size_t page_manager::pop_pages(size_t count, addr_t *address) { kassert(m_free, "page_manager::pop_pages ran out of free pages!"); unsigned n = std::min(count, static_cast(m_free->count)); *address = m_free->physical_address; m_free->physical_address += n * page_size; m_free->count -= n; if (m_free->count == 0) { page_block *block = m_free; m_free = m_free->next; block->zero(m_block_cache); m_block_cache = block; } return n; } void page_table::dump(int level, int max_index, uint64_t offset) { console *cons = console::get(); cons->printf("\nLevel %d page table @ %lx (off %lx):\n", level, this, offset); for (int i=0; i<512; ++i) { uint64_t ent = entries[i]; if (ent == 0) continue; if ((ent & 0x1) == 0) { cons->printf(" %3d: %016lx NOT PRESENT\n", i, ent); continue; } if ((level == 2 || level == 3) && (ent & 0x80) == 0x80) { cons->printf(" %3d: %016lx -> Large page at %016lx\n", i, ent, ent & ~0xfffull); continue; } else if (level == 1) { cons->printf(" %3d: %016lx -> Page at %016lx\n", i, ent, ent & ~0xfffull); } else { cons->printf(" %3d: %016lx -> Level %d table at %016lx\n", i, ent, level - 1, (ent & ~0xfffull) + offset); continue; } } if (--level > 0) { for (int i=0; i<=max_index; ++i) { uint64_t ent = entries[i]; if ((ent & 0x1) == 0) continue; if ((ent & 0x80)) continue; page_table *next = reinterpret_cast((ent & ~0xffful) + offset); next->dump(level, 511, offset); } } }