Kernel: Don't expose a region's page directory to the outside world

Now that region manages its own mapping/unmapping, there's no need for
the outside world to be able to grab at its page directory.
This commit is contained in:
Andreas Kling 2019-11-04 00:26:00 +01:00
parent 6ed9cc4717
commit 0e8f1d7cb6
3 changed files with 8 additions and 11 deletions

View file

@ -365,7 +365,6 @@ bool MemoryManager::copy_on_write(Region& region, unsigned page_index_in_region)
bool MemoryManager::page_in_from_inode(Region& region, unsigned page_index_in_region)
{
ASSERT(region.page_directory());
ASSERT(region.vmobject().is_inode());
auto& vmobject = region.vmobject();

View file

@ -175,10 +175,10 @@ Bitmap& Region::ensure_cow_map() const
void Region::remap_page(size_t index)
{
ASSERT(page_directory());
ASSERT(m_page_directory);
InterruptDisabler disabler;
auto page_vaddr = vaddr().offset(index * PAGE_SIZE);
auto& pte = MM.ensure_pte(*page_directory(), page_vaddr);
auto& pte = MM.ensure_pte(*m_page_directory, page_vaddr);
auto& physical_page = vmobject().physical_pages()[first_page_index() + index];
ASSERT(physical_page);
pte.set_physical_page_base(physical_page->paddr().get());
@ -188,31 +188,31 @@ void Region::remap_page(size_t index)
else
pte.set_writable(is_writable());
pte.set_user_allowed(is_user_accessible());
page_directory()->flush(page_vaddr);
m_page_directory->flush(page_vaddr);
#ifdef MM_DEBUG
dbg() << "MM: >> region.remap_page (PD=" << page_directory()->cr3() << ", PTE=" << (void*)pte.raw() << "{" << &pte << "}) " << name() << " " << page_vaddr << " => " << physical_page->paddr() << " (@" << physical_page.ptr() << ")";
dbg() << "MM: >> region.remap_page (PD=" << m_page_directory->cr3() << ", PTE=" << (void*)pte.raw() << "{" << &pte << "}) " << name() << " " << page_vaddr << " => " << physical_page->paddr() << " (@" << physical_page.ptr() << ")";
#endif
}
void Region::unmap(ShouldDeallocateVirtualMemoryRange deallocate_range)
{
InterruptDisabler disabler;
ASSERT(page_directory());
ASSERT(m_page_directory);
for (size_t i = 0; i < page_count(); ++i) {
auto vaddr = this->vaddr().offset(i * PAGE_SIZE);
auto& pte = MM.ensure_pte(*page_directory(), vaddr);
auto& pte = MM.ensure_pte(*m_page_directory, vaddr);
pte.set_physical_page_base(0);
pte.set_present(false);
pte.set_writable(false);
pte.set_user_allowed(false);
page_directory()->flush(vaddr);
m_page_directory->flush(vaddr);
#ifdef MM_DEBUG
auto& physical_page = vmobject().physical_pages()[first_page_index() + i];
dbgprintf("MM: >> Unmapped V%p => P%p <<\n", vaddr.get(), physical_page ? physical_page->paddr().get() : 0);
#endif
}
if (deallocate_range == ShouldDeallocateVirtualMemoryRange::Yes)
page_directory()->range_allocator().deallocate(range());
m_page_directory->range_allocator().deallocate(range());
m_page_directory = nullptr;
}

View file

@ -89,8 +89,6 @@ public:
size_t amount_resident() const;
size_t amount_shared() const;
PageDirectory* page_directory() { return m_page_directory.ptr(); }
bool should_cow(size_t page_index) const;
void set_should_cow(size_t page_index, bool);