#include "frame_allocator.h" #include "log.h" #include "objects/process.h" #include "objects/thread.h" #include "objects/vm_area.h" #include "vm_space.h" int vm_space::area::compare(const vm_space::area &o) const { if (base > o.base) return 1; else if (base < o.base) return -1; else return 0; } bool vm_space::area::operator==(const vm_space::area &o) const { return o.base == base && o.area == area; } vm_space::vm_space(page_table *p) : m_kernel(true), m_pml4(p) {} vm_space::vm_space() : m_kernel(false) { m_pml4 = page_table::get_table_page(); page_table *kpml4 = kernel_space().m_pml4; kutil::memset(m_pml4, 0, memory::frame_size/2); for (unsigned i = memory::pml4e_kernel; i < memory::table_entries; ++i) m_pml4->entries[i] = kpml4->entries[i]; } vm_space::~vm_space() { for (auto &a : m_areas) a.area->remove_from(this); kassert(!is_kernel(), "Kernel vm_space destructor!"); vm_space &kernel = kernel_space(); if (active()) kernel.activate(); // All VMAs have been removed by now, so just // free all remaining pages and tables m_pml4->free(page_table::level::pml4); } vm_space & vm_space::kernel_space() { return process::kernel_process().space(); } bool vm_space::add(uintptr_t base, vm_area *area) { //TODO: check for collisions m_areas.sorted_insert({base, area}); return true; } bool vm_space::remove(vm_area *area) { for (auto &a : m_areas) { if (a.area == area) { m_areas.remove(a); return true; } } return false; } vm_area * vm_space::get(uintptr_t addr, uintptr_t *base) { for (auto &a : m_areas) { uintptr_t end = a.base + a.area->size(); if (addr >= a.base && addr < end) { if (base) *base = a.base; return a.area; } } return nullptr; } void vm_space::copy_from(const vm_space &source, uintptr_t from, uintptr_t to, size_t count) { page_table::iterator sit {from, source.m_pml4}; page_table::iterator dit {to, m_pml4}; while (count--) { uint64_t &e = dit.entry(page_table::level::pt); if (e & page_table::flag::present) { // TODO: handle clobbering mapping } e = sit.entry(page_table::level::pt); } } void vm_space::page_in(uintptr_t virt, uintptr_t phys, size_t count) { page_table::iterator it {virt, m_pml4}; for (size_t i = 0; i < count; ++i) { uint64_t &e = it.entry(page_table::level::pt); bool allowed = (e & page_table::flag::allowed); e = (phys + i * memory::frame_size) | (allowed ? page_table::flag::allowed : page_table::flag::none); ++it; } } void vm_space::clear(uintptr_t addr, size_t count) { using memory::frame_size; uintptr_t free_start = 0; size_t free_count = 0; frame_allocator &fa = frame_allocator::get(); page_table::iterator it {addr, m_pml4}; while (count--) { uint64_t &e = it.entry(page_table::level::pt); bool allowed = (e & page_table::flag::allowed); uintptr_t phys = e & ~0xfffull; if (e & page_table::flag::present) { if (free_count && phys == free_start + (free_count * frame_size)) { ++free_count; } else { if (free_count) fa.free(free_start, free_count); free_start = phys; free_count = 1; } fa.free(e & ~0xfffull, 1); } e = 0 | (allowed ? page_table::flag::allowed : page_table::flag::none); ++it; } if (free_count) fa.free(free_start, free_count); } void vm_space::allow(uintptr_t start, size_t length, bool allow) { using level = page_table::level; kassert((start & 0xfff) == 0, "non-page-aligned address"); kassert((length & 0xfff) == 0, "non-page-aligned length"); const uintptr_t end = start + length; page_table::iterator it {start, m_pml4}; while (it.vaddress() < end) { level d = it.align(); while (it.end(d) > end) ++d; it.allow(d-1, allow); it.next(d); } } bool vm_space::active() const { uintptr_t pml4 = 0; __asm__ __volatile__ ( "mov %%cr3, %0" : "=r" (pml4) ); return memory::to_virtual(pml4 & ~0xfffull) == m_pml4; } void vm_space::activate() const { constexpr uint64_t phys_mask = ~memory::page_offset & ~0xfffull; uintptr_t p = reinterpret_cast(m_pml4) & phys_mask; __asm__ __volatile__ ( "mov %0, %%cr3" :: "r" (p) ); } void vm_space::initialize_tcb(TCB &tcb) { tcb.pml4 = reinterpret_cast(m_pml4) & ~memory::page_offset; } bool vm_space::handle_fault(uintptr_t addr, fault_type fault) { uintptr_t page = addr & ~0xfffull; page_table::iterator it {addr, m_pml4}; // TODO: Handle more fult types if (fault && fault_type::present) return false; uintptr_t base = 0; vm_area *area = get(addr, &base); if (!area && !it.allowed()) return false; uintptr_t phys = 0; size_t n = frame_allocator::get().allocate(1, &phys); kassert(n, "Failed to allocate a new page during page fault"); page_table::flag flags = page_table::flag::present | page_table::flag::write | (area ? page_table::flag::none : page_table::flag::allowed) | (is_kernel() ? page_table::flag::global : page_table::flag::user); it.entry(page_table::level::pt) = phys | flags; if (area) { uintptr_t offset = page - base; area->commit(phys, offset, 1); } return true; } size_t vm_space::copy(vm_space &source, vm_space &dest, void *from, void *to, size_t length) { uintptr_t ifrom = reinterpret_cast(from); uintptr_t ito = reinterpret_cast(to); page_table::iterator sit {ifrom, source.m_pml4}; page_table::iterator dit {ito, dest.m_pml4}; // TODO: iterate page mappings and continue copying. For now i'm blindly // assuming both buffers are fully contained within single pages kutil::memcpy( memory::to_virtual((*dit & ~0xfffull) | (ito & 0xffful)), memory::to_virtual((*sit & ~0xfffull) | (ifrom & 0xffful)), length); return length; }