mirror of
https://github.com/starr-dusT/yuzu-mainline
synced 2024-03-05 21:12:25 -08:00
Merge pull request #7956 from bunnei/improve-mem-manager
Kernel Memory Updates (Part 4): Revamp KMemoryManager & other fixes
This commit is contained in:
commit
3ab82e7582
@ -152,6 +152,7 @@ add_library(core STATIC
|
|||||||
hle/api_version.h
|
hle/api_version.h
|
||||||
hle/ipc.h
|
hle/ipc.h
|
||||||
hle/ipc_helpers.h
|
hle/ipc_helpers.h
|
||||||
|
hle/kernel/board/nintendo/nx/k_memory_layout.h
|
||||||
hle/kernel/board/nintendo/nx/k_system_control.cpp
|
hle/kernel/board/nintendo/nx/k_system_control.cpp
|
||||||
hle/kernel/board/nintendo/nx/k_system_control.h
|
hle/kernel/board/nintendo/nx/k_system_control.h
|
||||||
hle/kernel/board/nintendo/nx/secure_monitor.h
|
hle/kernel/board/nintendo/nx/secure_monitor.h
|
||||||
@ -164,6 +165,7 @@ add_library(core STATIC
|
|||||||
hle/kernel/hle_ipc.h
|
hle/kernel/hle_ipc.h
|
||||||
hle/kernel/init/init_slab_setup.cpp
|
hle/kernel/init/init_slab_setup.cpp
|
||||||
hle/kernel/init/init_slab_setup.h
|
hle/kernel/init/init_slab_setup.h
|
||||||
|
hle/kernel/initial_process.h
|
||||||
hle/kernel/k_address_arbiter.cpp
|
hle/kernel/k_address_arbiter.cpp
|
||||||
hle/kernel/k_address_arbiter.h
|
hle/kernel/k_address_arbiter.h
|
||||||
hle/kernel/k_address_space_info.cpp
|
hle/kernel/k_address_space_info.cpp
|
||||||
|
13
src/core/hle/kernel/board/nintendo/nx/k_memory_layout.h
Normal file
13
src/core/hle/kernel/board/nintendo/nx/k_memory_layout.h
Normal file
@ -0,0 +1,13 @@
|
|||||||
|
// Copyright 2022 yuzu Emulator Project
|
||||||
|
// Licensed under GPLv2 or any later version
|
||||||
|
// Refer to the license.txt file included.
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include "common/common_types.h"
|
||||||
|
|
||||||
|
namespace Kernel {
|
||||||
|
|
||||||
|
constexpr inline PAddr MainMemoryAddress = 0x80000000;
|
||||||
|
|
||||||
|
} // namespace Kernel
|
@ -39,6 +39,10 @@ Smc::MemoryArrangement GetMemoryArrangeForInit() {
|
|||||||
}
|
}
|
||||||
} // namespace
|
} // namespace
|
||||||
|
|
||||||
|
size_t KSystemControl::Init::GetRealMemorySize() {
|
||||||
|
return GetIntendedMemorySize();
|
||||||
|
}
|
||||||
|
|
||||||
// Initialization.
|
// Initialization.
|
||||||
size_t KSystemControl::Init::GetIntendedMemorySize() {
|
size_t KSystemControl::Init::GetIntendedMemorySize() {
|
||||||
switch (GetMemorySizeForInit()) {
|
switch (GetMemorySizeForInit()) {
|
||||||
@ -53,7 +57,13 @@ size_t KSystemControl::Init::GetIntendedMemorySize() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
PAddr KSystemControl::Init::GetKernelPhysicalBaseAddress(u64 base_address) {
|
PAddr KSystemControl::Init::GetKernelPhysicalBaseAddress(u64 base_address) {
|
||||||
return base_address;
|
const size_t real_dram_size = KSystemControl::Init::GetRealMemorySize();
|
||||||
|
const size_t intended_dram_size = KSystemControl::Init::GetIntendedMemorySize();
|
||||||
|
if (intended_dram_size * 2 < real_dram_size) {
|
||||||
|
return base_address;
|
||||||
|
} else {
|
||||||
|
return base_address + ((real_dram_size - intended_dram_size) / 2);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
bool KSystemControl::Init::ShouldIncreaseThreadResourceLimit() {
|
bool KSystemControl::Init::ShouldIncreaseThreadResourceLimit() {
|
||||||
|
@ -13,6 +13,7 @@ public:
|
|||||||
class Init {
|
class Init {
|
||||||
public:
|
public:
|
||||||
// Initialization.
|
// Initialization.
|
||||||
|
static std::size_t GetRealMemorySize();
|
||||||
static std::size_t GetIntendedMemorySize();
|
static std::size_t GetIntendedMemorySize();
|
||||||
static PAddr GetKernelPhysicalBaseAddress(u64 base_address);
|
static PAddr GetKernelPhysicalBaseAddress(u64 base_address);
|
||||||
static bool ShouldIncreaseThreadResourceLimit();
|
static bool ShouldIncreaseThreadResourceLimit();
|
||||||
|
23
src/core/hle/kernel/initial_process.h
Normal file
23
src/core/hle/kernel/initial_process.h
Normal file
@ -0,0 +1,23 @@
|
|||||||
|
// Copyright 2022 yuzu Emulator Project
|
||||||
|
// Licensed under GPLv2 or any later version
|
||||||
|
// Refer to the license.txt file included.
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include "common/common_types.h"
|
||||||
|
#include "common/literals.h"
|
||||||
|
#include "core/hle/kernel/board/nintendo/nx/k_memory_layout.h"
|
||||||
|
#include "core/hle/kernel/board/nintendo/nx/k_system_control.h"
|
||||||
|
|
||||||
|
namespace Kernel {
|
||||||
|
|
||||||
|
using namespace Common::Literals;
|
||||||
|
|
||||||
|
constexpr std::size_t InitialProcessBinarySizeMax = 12_MiB;
|
||||||
|
|
||||||
|
static inline PAddr GetInitialProcessBinaryPhysicalAddress() {
|
||||||
|
return Kernel::Board::Nintendo::Nx::KSystemControl::Init::GetKernelPhysicalBaseAddress(
|
||||||
|
MainMemoryAddress);
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace Kernel
|
@ -173,6 +173,10 @@ public:
|
|||||||
return Dereference(FindVirtualLinear(address));
|
return Dereference(FindVirtualLinear(address));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const KMemoryRegion& GetPhysicalLinearRegion(PAddr address) const {
|
||||||
|
return Dereference(FindPhysicalLinear(address));
|
||||||
|
}
|
||||||
|
|
||||||
const KMemoryRegion* GetPhysicalKernelTraceBufferRegion() const {
|
const KMemoryRegion* GetPhysicalKernelTraceBufferRegion() const {
|
||||||
return GetPhysicalMemoryRegionTree().FindFirstDerived(KMemoryRegionType_KernelTraceBuffer);
|
return GetPhysicalMemoryRegionTree().FindFirstDerived(KMemoryRegionType_KernelTraceBuffer);
|
||||||
}
|
}
|
||||||
|
@ -10,189 +10,412 @@
|
|||||||
#include "common/scope_exit.h"
|
#include "common/scope_exit.h"
|
||||||
#include "core/core.h"
|
#include "core/core.h"
|
||||||
#include "core/device_memory.h"
|
#include "core/device_memory.h"
|
||||||
|
#include "core/hle/kernel/initial_process.h"
|
||||||
#include "core/hle/kernel/k_memory_manager.h"
|
#include "core/hle/kernel/k_memory_manager.h"
|
||||||
#include "core/hle/kernel/k_page_linked_list.h"
|
#include "core/hle/kernel/k_page_linked_list.h"
|
||||||
|
#include "core/hle/kernel/kernel.h"
|
||||||
#include "core/hle/kernel/svc_results.h"
|
#include "core/hle/kernel/svc_results.h"
|
||||||
|
#include "core/memory.h"
|
||||||
|
|
||||||
namespace Kernel {
|
namespace Kernel {
|
||||||
|
|
||||||
KMemoryManager::KMemoryManager(Core::System& system_) : system{system_} {}
|
namespace {
|
||||||
|
|
||||||
std::size_t KMemoryManager::Impl::Initialize(Pool new_pool, u64 start_address, u64 end_address) {
|
constexpr KMemoryManager::Pool GetPoolFromMemoryRegionType(u32 type) {
|
||||||
const auto size{end_address - start_address};
|
if ((type | KMemoryRegionType_DramApplicationPool) == type) {
|
||||||
|
return KMemoryManager::Pool::Application;
|
||||||
// Calculate metadata sizes
|
} else if ((type | KMemoryRegionType_DramAppletPool) == type) {
|
||||||
const auto ref_count_size{(size / PageSize) * sizeof(u16)};
|
return KMemoryManager::Pool::Applet;
|
||||||
const auto optimize_map_size{(Common::AlignUp((size / PageSize), 64) / 64) * sizeof(u64)};
|
} else if ((type | KMemoryRegionType_DramSystemPool) == type) {
|
||||||
const auto manager_size{Common::AlignUp(optimize_map_size + ref_count_size, PageSize)};
|
return KMemoryManager::Pool::System;
|
||||||
const auto page_heap_size{KPageHeap::CalculateManagementOverheadSize(size)};
|
} else if ((type | KMemoryRegionType_DramSystemNonSecurePool) == type) {
|
||||||
const auto total_metadata_size{manager_size + page_heap_size};
|
return KMemoryManager::Pool::SystemNonSecure;
|
||||||
ASSERT(manager_size <= total_metadata_size);
|
} else {
|
||||||
ASSERT(Common::IsAligned(total_metadata_size, PageSize));
|
UNREACHABLE_MSG("InvalidMemoryRegionType for conversion to Pool");
|
||||||
|
return {};
|
||||||
// Setup region
|
}
|
||||||
pool = new_pool;
|
|
||||||
|
|
||||||
// Initialize the manager's KPageHeap
|
|
||||||
heap.Initialize(start_address, size, page_heap_size);
|
|
||||||
|
|
||||||
// Free the memory to the heap
|
|
||||||
heap.Free(start_address, size / PageSize);
|
|
||||||
|
|
||||||
// Update the heap's used size
|
|
||||||
heap.UpdateUsedSize();
|
|
||||||
|
|
||||||
return total_metadata_size;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void KMemoryManager::InitializeManager(Pool pool, u64 start_address, u64 end_address) {
|
} // namespace
|
||||||
ASSERT(pool < Pool::Count);
|
|
||||||
managers[static_cast<std::size_t>(pool)].Initialize(pool, start_address, end_address);
|
KMemoryManager::KMemoryManager(Core::System& system_)
|
||||||
|
: system{system_}, pool_locks{
|
||||||
|
KLightLock{system_.Kernel()},
|
||||||
|
KLightLock{system_.Kernel()},
|
||||||
|
KLightLock{system_.Kernel()},
|
||||||
|
KLightLock{system_.Kernel()},
|
||||||
|
} {}
|
||||||
|
|
||||||
|
void KMemoryManager::Initialize(VAddr management_region, size_t management_region_size) {
|
||||||
|
|
||||||
|
// Clear the management region to zero.
|
||||||
|
const VAddr management_region_end = management_region + management_region_size;
|
||||||
|
|
||||||
|
// Reset our manager count.
|
||||||
|
num_managers = 0;
|
||||||
|
|
||||||
|
// Traverse the virtual memory layout tree, initializing each manager as appropriate.
|
||||||
|
while (num_managers != MaxManagerCount) {
|
||||||
|
// Locate the region that should initialize the current manager.
|
||||||
|
PAddr region_address = 0;
|
||||||
|
size_t region_size = 0;
|
||||||
|
Pool region_pool = Pool::Count;
|
||||||
|
for (const auto& it : system.Kernel().MemoryLayout().GetPhysicalMemoryRegionTree()) {
|
||||||
|
// We only care about regions that we need to create managers for.
|
||||||
|
if (!it.IsDerivedFrom(KMemoryRegionType_DramUserPool)) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// We want to initialize the managers in order.
|
||||||
|
if (it.GetAttributes() != num_managers) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
const PAddr cur_start = it.GetAddress();
|
||||||
|
const PAddr cur_end = it.GetEndAddress();
|
||||||
|
|
||||||
|
// Validate the region.
|
||||||
|
ASSERT(cur_end != 0);
|
||||||
|
ASSERT(cur_start != 0);
|
||||||
|
ASSERT(it.GetSize() > 0);
|
||||||
|
|
||||||
|
// Update the region's extents.
|
||||||
|
if (region_address == 0) {
|
||||||
|
region_address = cur_start;
|
||||||
|
region_size = it.GetSize();
|
||||||
|
region_pool = GetPoolFromMemoryRegionType(it.GetType());
|
||||||
|
} else {
|
||||||
|
ASSERT(cur_start == region_address + region_size);
|
||||||
|
|
||||||
|
// Update the size.
|
||||||
|
region_size = cur_end - region_address;
|
||||||
|
ASSERT(GetPoolFromMemoryRegionType(it.GetType()) == region_pool);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we didn't find a region, we're done.
|
||||||
|
if (region_size == 0) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Initialize a new manager for the region.
|
||||||
|
Impl* manager = std::addressof(managers[num_managers++]);
|
||||||
|
ASSERT(num_managers <= managers.size());
|
||||||
|
|
||||||
|
const size_t cur_size = manager->Initialize(region_address, region_size, management_region,
|
||||||
|
management_region_end, region_pool);
|
||||||
|
management_region += cur_size;
|
||||||
|
ASSERT(management_region <= management_region_end);
|
||||||
|
|
||||||
|
// Insert the manager into the pool list.
|
||||||
|
const auto region_pool_index = static_cast<u32>(region_pool);
|
||||||
|
if (pool_managers_tail[region_pool_index] == nullptr) {
|
||||||
|
pool_managers_head[region_pool_index] = manager;
|
||||||
|
} else {
|
||||||
|
pool_managers_tail[region_pool_index]->SetNext(manager);
|
||||||
|
manager->SetPrev(pool_managers_tail[region_pool_index]);
|
||||||
|
}
|
||||||
|
pool_managers_tail[region_pool_index] = manager;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Free each region to its corresponding heap.
|
||||||
|
size_t reserved_sizes[MaxManagerCount] = {};
|
||||||
|
const PAddr ini_start = GetInitialProcessBinaryPhysicalAddress();
|
||||||
|
const PAddr ini_end = ini_start + InitialProcessBinarySizeMax;
|
||||||
|
const PAddr ini_last = ini_end - 1;
|
||||||
|
for (const auto& it : system.Kernel().MemoryLayout().GetPhysicalMemoryRegionTree()) {
|
||||||
|
if (it.IsDerivedFrom(KMemoryRegionType_DramUserPool)) {
|
||||||
|
// Get the manager for the region.
|
||||||
|
auto index = it.GetAttributes();
|
||||||
|
auto& manager = managers[index];
|
||||||
|
|
||||||
|
const PAddr cur_start = it.GetAddress();
|
||||||
|
const PAddr cur_last = it.GetLastAddress();
|
||||||
|
const PAddr cur_end = it.GetEndAddress();
|
||||||
|
|
||||||
|
if (cur_start <= ini_start && ini_last <= cur_last) {
|
||||||
|
// Free memory before the ini to the heap.
|
||||||
|
if (cur_start != ini_start) {
|
||||||
|
manager.Free(cur_start, (ini_start - cur_start) / PageSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Open/reserve the ini memory.
|
||||||
|
manager.OpenFirst(ini_start, InitialProcessBinarySizeMax / PageSize);
|
||||||
|
reserved_sizes[it.GetAttributes()] += InitialProcessBinarySizeMax;
|
||||||
|
|
||||||
|
// Free memory after the ini to the heap.
|
||||||
|
if (ini_last != cur_last) {
|
||||||
|
ASSERT(cur_end != 0);
|
||||||
|
manager.Free(ini_end, cur_end - ini_end);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// Ensure there's no partial overlap with the ini image.
|
||||||
|
if (cur_start <= ini_last) {
|
||||||
|
ASSERT(cur_last < ini_start);
|
||||||
|
} else {
|
||||||
|
// Otherwise, check the region for general validity.
|
||||||
|
ASSERT(cur_end != 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Free the memory to the heap.
|
||||||
|
manager.Free(cur_start, it.GetSize() / PageSize);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update the used size for all managers.
|
||||||
|
for (size_t i = 0; i < num_managers; ++i) {
|
||||||
|
managers[i].SetInitialUsedHeapSize(reserved_sizes[i]);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
VAddr KMemoryManager::AllocateAndOpenContinuous(std::size_t num_pages, std::size_t align_pages,
|
PAddr KMemoryManager::AllocateAndOpenContinuous(size_t num_pages, size_t align_pages, u32 option) {
|
||||||
u32 option) {
|
// Early return if we're allocating no pages.
|
||||||
// Early return if we're allocating no pages
|
|
||||||
if (num_pages == 0) {
|
if (num_pages == 0) {
|
||||||
return {};
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Lock the pool that we're allocating from
|
// Lock the pool that we're allocating from.
|
||||||
const auto [pool, dir] = DecodeOption(option);
|
const auto [pool, dir] = DecodeOption(option);
|
||||||
const auto pool_index{static_cast<std::size_t>(pool)};
|
KScopedLightLock lk(pool_locks[static_cast<std::size_t>(pool)]);
|
||||||
std::lock_guard lock{pool_locks[pool_index]};
|
|
||||||
|
|
||||||
// Choose a heap based on our page size request
|
// Choose a heap based on our page size request.
|
||||||
const s32 heap_index{KPageHeap::GetAlignedBlockIndex(num_pages, align_pages)};
|
const s32 heap_index = KPageHeap::GetAlignedBlockIndex(num_pages, align_pages);
|
||||||
|
|
||||||
// Loop, trying to iterate from each block
|
// Loop, trying to iterate from each block.
|
||||||
// TODO (bunnei): Support multiple managers
|
Impl* chosen_manager = nullptr;
|
||||||
Impl& chosen_manager{managers[pool_index]};
|
PAddr allocated_block = 0;
|
||||||
VAddr allocated_block{chosen_manager.AllocateBlock(heap_index, false)};
|
for (chosen_manager = this->GetFirstManager(pool, dir); chosen_manager != nullptr;
|
||||||
|
chosen_manager = this->GetNextManager(chosen_manager, dir)) {
|
||||||
// If we failed to allocate, quit now
|
allocated_block = chosen_manager->AllocateBlock(heap_index, true);
|
||||||
if (!allocated_block) {
|
if (allocated_block != 0) {
|
||||||
return {};
|
break;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// If we allocated more than we need, free some
|
// If we failed to allocate, quit now.
|
||||||
const auto allocated_pages{KPageHeap::GetBlockNumPages(heap_index)};
|
if (allocated_block == 0) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we allocated more than we need, free some.
|
||||||
|
const size_t allocated_pages = KPageHeap::GetBlockNumPages(heap_index);
|
||||||
if (allocated_pages > num_pages) {
|
if (allocated_pages > num_pages) {
|
||||||
chosen_manager.Free(allocated_block + num_pages * PageSize, allocated_pages - num_pages);
|
chosen_manager->Free(allocated_block + num_pages * PageSize, allocated_pages - num_pages);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Open the first reference to the pages.
|
||||||
|
chosen_manager->OpenFirst(allocated_block, num_pages);
|
||||||
|
|
||||||
return allocated_block;
|
return allocated_block;
|
||||||
}
|
}
|
||||||
|
|
||||||
ResultCode KMemoryManager::Allocate(KPageLinkedList& page_list, std::size_t num_pages, Pool pool,
|
ResultCode KMemoryManager::AllocatePageGroupImpl(KPageLinkedList* out, size_t num_pages, Pool pool,
|
||||||
Direction dir, u32 heap_fill_value) {
|
Direction dir, bool random) {
|
||||||
ASSERT(page_list.GetNumPages() == 0);
|
// Choose a heap based on our page size request.
|
||||||
|
const s32 heap_index = KPageHeap::GetBlockIndex(num_pages);
|
||||||
|
R_UNLESS(0 <= heap_index, ResultOutOfMemory);
|
||||||
|
|
||||||
// Early return if we're allocating no pages
|
// Ensure that we don't leave anything un-freed.
|
||||||
if (num_pages == 0) {
|
auto group_guard = SCOPE_GUARD({
|
||||||
return ResultSuccess;
|
for (const auto& it : out->Nodes()) {
|
||||||
}
|
auto& manager = this->GetManager(system.Kernel().MemoryLayout(), it.GetAddress());
|
||||||
|
const size_t num_pages_to_free =
|
||||||
// Lock the pool that we're allocating from
|
std::min(it.GetNumPages(), (manager.GetEndAddress() - it.GetAddress()) / PageSize);
|
||||||
const auto pool_index{static_cast<std::size_t>(pool)};
|
manager.Free(it.GetAddress(), num_pages_to_free);
|
||||||
std::lock_guard lock{pool_locks[pool_index]};
|
|
||||||
|
|
||||||
// Choose a heap based on our page size request
|
|
||||||
const s32 heap_index{KPageHeap::GetBlockIndex(num_pages)};
|
|
||||||
if (heap_index < 0) {
|
|
||||||
return ResultOutOfMemory;
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO (bunnei): Support multiple managers
|
|
||||||
Impl& chosen_manager{managers[pool_index]};
|
|
||||||
|
|
||||||
// Ensure that we don't leave anything un-freed
|
|
||||||
auto group_guard = detail::ScopeExit([&] {
|
|
||||||
for (const auto& it : page_list.Nodes()) {
|
|
||||||
const auto min_num_pages{std::min<size_t>(
|
|
||||||
it.GetNumPages(), (chosen_manager.GetEndAddress() - it.GetAddress()) / PageSize)};
|
|
||||||
chosen_manager.Free(it.GetAddress(), min_num_pages);
|
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
// Keep allocating until we've allocated all our pages
|
// Keep allocating until we've allocated all our pages.
|
||||||
for (s32 index{heap_index}; index >= 0 && num_pages > 0; index--) {
|
for (s32 index = heap_index; index >= 0 && num_pages > 0; index--) {
|
||||||
const auto pages_per_alloc{KPageHeap::GetBlockNumPages(index)};
|
const size_t pages_per_alloc = KPageHeap::GetBlockNumPages(index);
|
||||||
|
for (Impl* cur_manager = this->GetFirstManager(pool, dir); cur_manager != nullptr;
|
||||||
while (num_pages >= pages_per_alloc) {
|
cur_manager = this->GetNextManager(cur_manager, dir)) {
|
||||||
// Allocate a block
|
while (num_pages >= pages_per_alloc) {
|
||||||
VAddr allocated_block{chosen_manager.AllocateBlock(index, false)};
|
// Allocate a block.
|
||||||
if (!allocated_block) {
|
PAddr allocated_block = cur_manager->AllocateBlock(index, random);
|
||||||
break;
|
if (allocated_block == 0) {
|
||||||
}
|
break;
|
||||||
|
|
||||||
// Safely add it to our group
|
|
||||||
{
|
|
||||||
auto block_guard = detail::ScopeExit(
|
|
||||||
[&] { chosen_manager.Free(allocated_block, pages_per_alloc); });
|
|
||||||
|
|
||||||
if (const ResultCode result{page_list.AddBlock(allocated_block, pages_per_alloc)};
|
|
||||||
result.IsError()) {
|
|
||||||
return result;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
block_guard.Cancel();
|
// Safely add it to our group.
|
||||||
}
|
{
|
||||||
|
auto block_guard =
|
||||||
|
SCOPE_GUARD({ cur_manager->Free(allocated_block, pages_per_alloc); });
|
||||||
|
R_TRY(out->AddBlock(allocated_block, pages_per_alloc));
|
||||||
|
block_guard.Cancel();
|
||||||
|
}
|
||||||
|
|
||||||
num_pages -= pages_per_alloc;
|
num_pages -= pages_per_alloc;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Clear allocated memory.
|
// Only succeed if we allocated as many pages as we wanted.
|
||||||
for (const auto& it : page_list.Nodes()) {
|
R_UNLESS(num_pages == 0, ResultOutOfMemory);
|
||||||
std::memset(system.DeviceMemory().GetPointer(it.GetAddress()), heap_fill_value,
|
|
||||||
it.GetSize());
|
|
||||||
}
|
|
||||||
|
|
||||||
// Only succeed if we allocated as many pages as we wanted
|
|
||||||
if (num_pages) {
|
|
||||||
return ResultOutOfMemory;
|
|
||||||
}
|
|
||||||
|
|
||||||
// We succeeded!
|
// We succeeded!
|
||||||
group_guard.Cancel();
|
group_guard.Cancel();
|
||||||
|
|
||||||
return ResultSuccess;
|
return ResultSuccess;
|
||||||
}
|
}
|
||||||
|
|
||||||
ResultCode KMemoryManager::Free(KPageLinkedList& page_list, std::size_t num_pages, Pool pool,
|
ResultCode KMemoryManager::AllocateAndOpen(KPageLinkedList* out, size_t num_pages, u32 option) {
|
||||||
Direction dir, u32 heap_fill_value) {
|
ASSERT(out != nullptr);
|
||||||
// Early return if we're freeing no pages
|
ASSERT(out->GetNumPages() == 0);
|
||||||
if (!num_pages) {
|
|
||||||
return ResultSuccess;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Lock the pool that we're freeing from
|
// Early return if we're allocating no pages.
|
||||||
const auto pool_index{static_cast<std::size_t>(pool)};
|
R_SUCCEED_IF(num_pages == 0);
|
||||||
std::lock_guard lock{pool_locks[pool_index]};
|
|
||||||
|
|
||||||
// TODO (bunnei): Support multiple managers
|
// Lock the pool that we're allocating from.
|
||||||
Impl& chosen_manager{managers[pool_index]};
|
const auto [pool, dir] = DecodeOption(option);
|
||||||
|
KScopedLightLock lk(pool_locks[static_cast<size_t>(pool)]);
|
||||||
|
|
||||||
// Free all of the pages
|
// Allocate the page group.
|
||||||
for (const auto& it : page_list.Nodes()) {
|
R_TRY(this->AllocatePageGroupImpl(out, num_pages, pool, dir, false));
|
||||||
const auto min_num_pages{std::min<size_t>(
|
|
||||||
it.GetNumPages(), (chosen_manager.GetEndAddress() - it.GetAddress()) / PageSize)};
|
// Open the first reference to the pages.
|
||||||
chosen_manager.Free(it.GetAddress(), min_num_pages);
|
for (const auto& block : out->Nodes()) {
|
||||||
|
PAddr cur_address = block.GetAddress();
|
||||||
|
size_t remaining_pages = block.GetNumPages();
|
||||||
|
while (remaining_pages > 0) {
|
||||||
|
// Get the manager for the current address.
|
||||||
|
auto& manager = this->GetManager(system.Kernel().MemoryLayout(), cur_address);
|
||||||
|
|
||||||
|
// Process part or all of the block.
|
||||||
|
const size_t cur_pages =
|
||||||
|
std::min(remaining_pages, manager.GetPageOffsetToEnd(cur_address));
|
||||||
|
manager.OpenFirst(cur_address, cur_pages);
|
||||||
|
|
||||||
|
// Advance.
|
||||||
|
cur_address += cur_pages * PageSize;
|
||||||
|
remaining_pages -= cur_pages;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return ResultSuccess;
|
return ResultSuccess;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::size_t KMemoryManager::Impl::CalculateManagementOverheadSize(std::size_t region_size) {
|
ResultCode KMemoryManager::AllocateAndOpenForProcess(KPageLinkedList* out, size_t num_pages,
|
||||||
const std::size_t ref_count_size = (region_size / PageSize) * sizeof(u16);
|
u32 option, u64 process_id, u8 fill_pattern) {
|
||||||
const std::size_t optimize_map_size =
|
ASSERT(out != nullptr);
|
||||||
|
ASSERT(out->GetNumPages() == 0);
|
||||||
|
|
||||||
|
// Decode the option.
|
||||||
|
const auto [pool, dir] = DecodeOption(option);
|
||||||
|
|
||||||
|
// Allocate the memory.
|
||||||
|
{
|
||||||
|
// Lock the pool that we're allocating from.
|
||||||
|
KScopedLightLock lk(pool_locks[static_cast<size_t>(pool)]);
|
||||||
|
|
||||||
|
// Allocate the page group.
|
||||||
|
R_TRY(this->AllocatePageGroupImpl(out, num_pages, pool, dir, false));
|
||||||
|
|
||||||
|
// Open the first reference to the pages.
|
||||||
|
for (const auto& block : out->Nodes()) {
|
||||||
|
PAddr cur_address = block.GetAddress();
|
||||||
|
size_t remaining_pages = block.GetNumPages();
|
||||||
|
while (remaining_pages > 0) {
|
||||||
|
// Get the manager for the current address.
|
||||||
|
auto& manager = this->GetManager(system.Kernel().MemoryLayout(), cur_address);
|
||||||
|
|
||||||
|
// Process part or all of the block.
|
||||||
|
const size_t cur_pages =
|
||||||
|
std::min(remaining_pages, manager.GetPageOffsetToEnd(cur_address));
|
||||||
|
manager.OpenFirst(cur_address, cur_pages);
|
||||||
|
|
||||||
|
// Advance.
|
||||||
|
cur_address += cur_pages * PageSize;
|
||||||
|
remaining_pages -= cur_pages;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set all the allocated memory.
|
||||||
|
for (const auto& block : out->Nodes()) {
|
||||||
|
std::memset(system.DeviceMemory().GetPointer(block.GetAddress()), fill_pattern,
|
||||||
|
block.GetSize());
|
||||||
|
}
|
||||||
|
|
||||||
|
return ResultSuccess;
|
||||||
|
}
|
||||||
|
|
||||||
|
void KMemoryManager::Open(PAddr address, size_t num_pages) {
|
||||||
|
// Repeatedly open references until we've done so for all pages.
|
||||||
|
while (num_pages) {
|
||||||
|
auto& manager = this->GetManager(system.Kernel().MemoryLayout(), address);
|
||||||
|
const size_t cur_pages = std::min(num_pages, manager.GetPageOffsetToEnd(address));
|
||||||
|
|
||||||
|
{
|
||||||
|
KScopedLightLock lk(pool_locks[static_cast<size_t>(manager.GetPool())]);
|
||||||
|
manager.Open(address, cur_pages);
|
||||||
|
}
|
||||||
|
|
||||||
|
num_pages -= cur_pages;
|
||||||
|
address += cur_pages * PageSize;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void KMemoryManager::Close(PAddr address, size_t num_pages) {
|
||||||
|
// Repeatedly close references until we've done so for all pages.
|
||||||
|
while (num_pages) {
|
||||||
|
auto& manager = this->GetManager(system.Kernel().MemoryLayout(), address);
|
||||||
|
const size_t cur_pages = std::min(num_pages, manager.GetPageOffsetToEnd(address));
|
||||||
|
|
||||||
|
{
|
||||||
|
KScopedLightLock lk(pool_locks[static_cast<size_t>(manager.GetPool())]);
|
||||||
|
manager.Close(address, cur_pages);
|
||||||
|
}
|
||||||
|
|
||||||
|
num_pages -= cur_pages;
|
||||||
|
address += cur_pages * PageSize;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void KMemoryManager::Close(const KPageLinkedList& pg) {
|
||||||
|
for (const auto& node : pg.Nodes()) {
|
||||||
|
Close(node.GetAddress(), node.GetNumPages());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
void KMemoryManager::Open(const KPageLinkedList& pg) {
|
||||||
|
for (const auto& node : pg.Nodes()) {
|
||||||
|
Open(node.GetAddress(), node.GetNumPages());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t KMemoryManager::Impl::Initialize(PAddr address, size_t size, VAddr management,
|
||||||
|
VAddr management_end, Pool p) {
|
||||||
|
// Calculate management sizes.
|
||||||
|
const size_t ref_count_size = (size / PageSize) * sizeof(u16);
|
||||||
|
const size_t optimize_map_size = CalculateOptimizedProcessOverheadSize(size);
|
||||||
|
const size_t manager_size = Common::AlignUp(optimize_map_size + ref_count_size, PageSize);
|
||||||
|
const size_t page_heap_size = KPageHeap::CalculateManagementOverheadSize(size);
|
||||||
|
const size_t total_management_size = manager_size + page_heap_size;
|
||||||
|
ASSERT(manager_size <= total_management_size);
|
||||||
|
ASSERT(management + total_management_size <= management_end);
|
||||||
|
ASSERT(Common::IsAligned(total_management_size, PageSize));
|
||||||
|
|
||||||
|
// Setup region.
|
||||||
|
pool = p;
|
||||||
|
management_region = management;
|
||||||
|
page_reference_counts.resize(
|
||||||
|
Kernel::Board::Nintendo::Nx::KSystemControl::Init::GetIntendedMemorySize() / PageSize);
|
||||||
|
ASSERT(Common::IsAligned(management_region, PageSize));
|
||||||
|
|
||||||
|
// Initialize the manager's KPageHeap.
|
||||||
|
heap.Initialize(address, size, management + manager_size, page_heap_size);
|
||||||
|
|
||||||
|
return total_management_size;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t KMemoryManager::Impl::CalculateManagementOverheadSize(size_t region_size) {
|
||||||
|
const size_t ref_count_size = (region_size / PageSize) * sizeof(u16);
|
||||||
|
const size_t optimize_map_size =
|
||||||
(Common::AlignUp((region_size / PageSize), Common::BitSize<u64>()) /
|
(Common::AlignUp((region_size / PageSize), Common::BitSize<u64>()) /
|
||||||
Common::BitSize<u64>()) *
|
Common::BitSize<u64>()) *
|
||||||
sizeof(u64);
|
sizeof(u64);
|
||||||
const std::size_t manager_meta_size =
|
const size_t manager_meta_size = Common::AlignUp(optimize_map_size + ref_count_size, PageSize);
|
||||||
Common::AlignUp(optimize_map_size + ref_count_size, PageSize);
|
const size_t page_heap_size = KPageHeap::CalculateManagementOverheadSize(region_size);
|
||||||
const std::size_t page_heap_size = KPageHeap::CalculateManagementOverheadSize(region_size);
|
|
||||||
return manager_meta_size + page_heap_size;
|
return manager_meta_size + page_heap_size;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -5,11 +5,12 @@
|
|||||||
#pragma once
|
#pragma once
|
||||||
|
|
||||||
#include <array>
|
#include <array>
|
||||||
#include <mutex>
|
|
||||||
#include <tuple>
|
#include <tuple>
|
||||||
|
|
||||||
#include "common/common_funcs.h"
|
#include "common/common_funcs.h"
|
||||||
#include "common/common_types.h"
|
#include "common/common_types.h"
|
||||||
|
#include "core/hle/kernel/k_light_lock.h"
|
||||||
|
#include "core/hle/kernel/k_memory_layout.h"
|
||||||
#include "core/hle/kernel/k_page_heap.h"
|
#include "core/hle/kernel/k_page_heap.h"
|
||||||
#include "core/hle/result.h"
|
#include "core/hle/result.h"
|
||||||
|
|
||||||
@ -52,22 +53,33 @@ public:
|
|||||||
|
|
||||||
explicit KMemoryManager(Core::System& system_);
|
explicit KMemoryManager(Core::System& system_);
|
||||||
|
|
||||||
constexpr std::size_t GetSize(Pool pool) const {
|
void Initialize(VAddr management_region, size_t management_region_size);
|
||||||
return managers[static_cast<std::size_t>(pool)].GetSize();
|
|
||||||
|
constexpr size_t GetSize(Pool pool) const {
|
||||||
|
constexpr Direction GetSizeDirection = Direction::FromFront;
|
||||||
|
size_t total = 0;
|
||||||
|
for (auto* manager = this->GetFirstManager(pool, GetSizeDirection); manager != nullptr;
|
||||||
|
manager = this->GetNextManager(manager, GetSizeDirection)) {
|
||||||
|
total += manager->GetSize();
|
||||||
|
}
|
||||||
|
return total;
|
||||||
}
|
}
|
||||||
|
|
||||||
void InitializeManager(Pool pool, u64 start_address, u64 end_address);
|
PAddr AllocateAndOpenContinuous(size_t num_pages, size_t align_pages, u32 option);
|
||||||
|
ResultCode AllocateAndOpen(KPageLinkedList* out, size_t num_pages, u32 option);
|
||||||
|
ResultCode AllocateAndOpenForProcess(KPageLinkedList* out, size_t num_pages, u32 option,
|
||||||
|
u64 process_id, u8 fill_pattern);
|
||||||
|
|
||||||
VAddr AllocateAndOpenContinuous(size_t num_pages, size_t align_pages, u32 option);
|
static constexpr size_t MaxManagerCount = 10;
|
||||||
ResultCode Allocate(KPageLinkedList& page_list, std::size_t num_pages, Pool pool, Direction dir,
|
|
||||||
u32 heap_fill_value = 0);
|
|
||||||
ResultCode Free(KPageLinkedList& page_list, std::size_t num_pages, Pool pool, Direction dir,
|
|
||||||
u32 heap_fill_value = 0);
|
|
||||||
|
|
||||||
static constexpr std::size_t MaxManagerCount = 10;
|
void Close(PAddr address, size_t num_pages);
|
||||||
|
void Close(const KPageLinkedList& pg);
|
||||||
|
|
||||||
|
void Open(PAddr address, size_t num_pages);
|
||||||
|
void Open(const KPageLinkedList& pg);
|
||||||
|
|
||||||
public:
|
public:
|
||||||
static std::size_t CalculateManagementOverheadSize(std::size_t region_size) {
|
static size_t CalculateManagementOverheadSize(size_t region_size) {
|
||||||
return Impl::CalculateManagementOverheadSize(region_size);
|
return Impl::CalculateManagementOverheadSize(region_size);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -100,17 +112,26 @@ private:
|
|||||||
Impl() = default;
|
Impl() = default;
|
||||||
~Impl() = default;
|
~Impl() = default;
|
||||||
|
|
||||||
std::size_t Initialize(Pool new_pool, u64 start_address, u64 end_address);
|
size_t Initialize(PAddr address, size_t size, VAddr management, VAddr management_end,
|
||||||
|
Pool p);
|
||||||
|
|
||||||
VAddr AllocateBlock(s32 index, bool random) {
|
VAddr AllocateBlock(s32 index, bool random) {
|
||||||
return heap.AllocateBlock(index, random);
|
return heap.AllocateBlock(index, random);
|
||||||
}
|
}
|
||||||
|
|
||||||
void Free(VAddr addr, std::size_t num_pages) {
|
void Free(VAddr addr, size_t num_pages) {
|
||||||
heap.Free(addr, num_pages);
|
heap.Free(addr, num_pages);
|
||||||
}
|
}
|
||||||
|
|
||||||
constexpr std::size_t GetSize() const {
|
void SetInitialUsedHeapSize(size_t reserved_size) {
|
||||||
|
heap.SetInitialUsedSize(reserved_size);
|
||||||
|
}
|
||||||
|
|
||||||
|
constexpr Pool GetPool() const {
|
||||||
|
return pool;
|
||||||
|
}
|
||||||
|
|
||||||
|
constexpr size_t GetSize() const {
|
||||||
return heap.GetSize();
|
return heap.GetSize();
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -122,10 +143,88 @@ private:
|
|||||||
return heap.GetEndAddress();
|
return heap.GetEndAddress();
|
||||||
}
|
}
|
||||||
|
|
||||||
static std::size_t CalculateManagementOverheadSize(std::size_t region_size);
|
constexpr size_t GetPageOffset(PAddr address) const {
|
||||||
|
return heap.GetPageOffset(address);
|
||||||
|
}
|
||||||
|
|
||||||
static constexpr std::size_t CalculateOptimizedProcessOverheadSize(
|
constexpr size_t GetPageOffsetToEnd(PAddr address) const {
|
||||||
std::size_t region_size) {
|
return heap.GetPageOffsetToEnd(address);
|
||||||
|
}
|
||||||
|
|
||||||
|
constexpr void SetNext(Impl* n) {
|
||||||
|
next = n;
|
||||||
|
}
|
||||||
|
|
||||||
|
constexpr void SetPrev(Impl* n) {
|
||||||
|
prev = n;
|
||||||
|
}
|
||||||
|
|
||||||
|
constexpr Impl* GetNext() const {
|
||||||
|
return next;
|
||||||
|
}
|
||||||
|
|
||||||
|
constexpr Impl* GetPrev() const {
|
||||||
|
return prev;
|
||||||
|
}
|
||||||
|
|
||||||
|
void OpenFirst(PAddr address, size_t num_pages) {
|
||||||
|
size_t index = this->GetPageOffset(address);
|
||||||
|
const size_t end = index + num_pages;
|
||||||
|
while (index < end) {
|
||||||
|
const RefCount ref_count = (++page_reference_counts[index]);
|
||||||
|
ASSERT(ref_count == 1);
|
||||||
|
|
||||||
|
index++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void Open(PAddr address, size_t num_pages) {
|
||||||
|
size_t index = this->GetPageOffset(address);
|
||||||
|
const size_t end = index + num_pages;
|
||||||
|
while (index < end) {
|
||||||
|
const RefCount ref_count = (++page_reference_counts[index]);
|
||||||
|
ASSERT(ref_count > 1);
|
||||||
|
|
||||||
|
index++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void Close(PAddr address, size_t num_pages) {
|
||||||
|
size_t index = this->GetPageOffset(address);
|
||||||
|
const size_t end = index + num_pages;
|
||||||
|
|
||||||
|
size_t free_start = 0;
|
||||||
|
size_t free_count = 0;
|
||||||
|
while (index < end) {
|
||||||
|
ASSERT(page_reference_counts[index] > 0);
|
||||||
|
const RefCount ref_count = (--page_reference_counts[index]);
|
||||||
|
|
||||||
|
// Keep track of how many zero refcounts we see in a row, to minimize calls to free.
|
||||||
|
if (ref_count == 0) {
|
||||||
|
if (free_count > 0) {
|
||||||
|
free_count++;
|
||||||
|
} else {
|
||||||
|
free_start = index;
|
||||||
|
free_count = 1;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if (free_count > 0) {
|
||||||
|
this->Free(heap.GetAddress() + free_start * PageSize, free_count);
|
||||||
|
free_count = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
index++;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (free_count > 0) {
|
||||||
|
this->Free(heap.GetAddress() + free_start * PageSize, free_count);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t CalculateManagementOverheadSize(size_t region_size);
|
||||||
|
|
||||||
|
static constexpr size_t CalculateOptimizedProcessOverheadSize(size_t region_size) {
|
||||||
return (Common::AlignUp((region_size / PageSize), Common::BitSize<u64>()) /
|
return (Common::AlignUp((region_size / PageSize), Common::BitSize<u64>()) /
|
||||||
Common::BitSize<u64>()) *
|
Common::BitSize<u64>()) *
|
||||||
sizeof(u64);
|
sizeof(u64);
|
||||||
@ -135,13 +234,45 @@ private:
|
|||||||
using RefCount = u16;
|
using RefCount = u16;
|
||||||
|
|
||||||
KPageHeap heap;
|
KPageHeap heap;
|
||||||
|
std::vector<RefCount> page_reference_counts;
|
||||||
|
VAddr management_region{};
|
||||||
Pool pool{};
|
Pool pool{};
|
||||||
|
Impl* next{};
|
||||||
|
Impl* prev{};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
private:
|
||||||
|
Impl& GetManager(const KMemoryLayout& memory_layout, PAddr address) {
|
||||||
|
return managers[memory_layout.GetPhysicalLinearRegion(address).GetAttributes()];
|
||||||
|
}
|
||||||
|
|
||||||
|
const Impl& GetManager(const KMemoryLayout& memory_layout, PAddr address) const {
|
||||||
|
return managers[memory_layout.GetPhysicalLinearRegion(address).GetAttributes()];
|
||||||
|
}
|
||||||
|
|
||||||
|
constexpr Impl* GetFirstManager(Pool pool, Direction dir) const {
|
||||||
|
return dir == Direction::FromBack ? pool_managers_tail[static_cast<size_t>(pool)]
|
||||||
|
: pool_managers_head[static_cast<size_t>(pool)];
|
||||||
|
}
|
||||||
|
|
||||||
|
constexpr Impl* GetNextManager(Impl* cur, Direction dir) const {
|
||||||
|
if (dir == Direction::FromBack) {
|
||||||
|
return cur->GetPrev();
|
||||||
|
} else {
|
||||||
|
return cur->GetNext();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
ResultCode AllocatePageGroupImpl(KPageLinkedList* out, size_t num_pages, Pool pool,
|
||||||
|
Direction dir, bool random);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
Core::System& system;
|
Core::System& system;
|
||||||
std::array<std::mutex, static_cast<std::size_t>(Pool::Count)> pool_locks;
|
std::array<KLightLock, static_cast<size_t>(Pool::Count)> pool_locks;
|
||||||
|
std::array<Impl*, MaxManagerCount> pool_managers_head{};
|
||||||
|
std::array<Impl*, MaxManagerCount> pool_managers_tail{};
|
||||||
std::array<Impl, MaxManagerCount> managers;
|
std::array<Impl, MaxManagerCount> managers;
|
||||||
|
size_t num_managers{};
|
||||||
};
|
};
|
||||||
|
|
||||||
} // namespace Kernel
|
} // namespace Kernel
|
||||||
|
@ -14,7 +14,8 @@
|
|||||||
namespace Kernel {
|
namespace Kernel {
|
||||||
|
|
||||||
enum KMemoryRegionType : u32 {
|
enum KMemoryRegionType : u32 {
|
||||||
KMemoryRegionAttr_CarveoutProtected = 0x04000000,
|
KMemoryRegionAttr_CarveoutProtected = 0x02000000,
|
||||||
|
KMemoryRegionAttr_Uncached = 0x04000000,
|
||||||
KMemoryRegionAttr_DidKernelMap = 0x08000000,
|
KMemoryRegionAttr_DidKernelMap = 0x08000000,
|
||||||
KMemoryRegionAttr_ShouldKernelMap = 0x10000000,
|
KMemoryRegionAttr_ShouldKernelMap = 0x10000000,
|
||||||
KMemoryRegionAttr_UserReadOnly = 0x20000000,
|
KMemoryRegionAttr_UserReadOnly = 0x20000000,
|
||||||
@ -239,6 +240,11 @@ static_assert(KMemoryRegionType_VirtualDramHeapBase.GetValue() == 0x1A);
|
|||||||
static_assert(KMemoryRegionType_VirtualDramKernelPtHeap.GetValue() == 0x2A);
|
static_assert(KMemoryRegionType_VirtualDramKernelPtHeap.GetValue() == 0x2A);
|
||||||
static_assert(KMemoryRegionType_VirtualDramKernelTraceBuffer.GetValue() == 0x4A);
|
static_assert(KMemoryRegionType_VirtualDramKernelTraceBuffer.GetValue() == 0x4A);
|
||||||
|
|
||||||
|
// UNUSED: .DeriveSparse(2, 2, 0);
|
||||||
|
constexpr auto KMemoryRegionType_VirtualDramUnknownDebug =
|
||||||
|
KMemoryRegionType_Dram.DeriveSparse(2, 2, 1);
|
||||||
|
static_assert(KMemoryRegionType_VirtualDramUnknownDebug.GetValue() == (0x52));
|
||||||
|
|
||||||
constexpr auto KMemoryRegionType_VirtualDramKernelInitPt =
|
constexpr auto KMemoryRegionType_VirtualDramKernelInitPt =
|
||||||
KMemoryRegionType_VirtualDramHeapBase.Derive(3, 0);
|
KMemoryRegionType_VirtualDramHeapBase.Derive(3, 0);
|
||||||
constexpr auto KMemoryRegionType_VirtualDramPoolManagement =
|
constexpr auto KMemoryRegionType_VirtualDramPoolManagement =
|
||||||
@ -330,6 +336,8 @@ constexpr KMemoryRegionType GetTypeForVirtualLinearMapping(u32 type_id) {
|
|||||||
return KMemoryRegionType_VirtualDramKernelTraceBuffer;
|
return KMemoryRegionType_VirtualDramKernelTraceBuffer;
|
||||||
} else if (KMemoryRegionType_DramKernelPtHeap.IsAncestorOf(type_id)) {
|
} else if (KMemoryRegionType_DramKernelPtHeap.IsAncestorOf(type_id)) {
|
||||||
return KMemoryRegionType_VirtualDramKernelPtHeap;
|
return KMemoryRegionType_VirtualDramKernelPtHeap;
|
||||||
|
} else if ((type_id | KMemoryRegionAttr_ShouldKernelMap) == type_id) {
|
||||||
|
return KMemoryRegionType_VirtualDramUnknownDebug;
|
||||||
} else {
|
} else {
|
||||||
return KMemoryRegionType_Dram;
|
return KMemoryRegionType_Dram;
|
||||||
}
|
}
|
||||||
|
@ -7,35 +7,51 @@
|
|||||||
|
|
||||||
namespace Kernel {
|
namespace Kernel {
|
||||||
|
|
||||||
void KPageHeap::Initialize(VAddr address, std::size_t size, std::size_t metadata_size) {
|
void KPageHeap::Initialize(PAddr address, size_t size, VAddr management_address,
|
||||||
// Check our assumptions
|
size_t management_size, const size_t* block_shifts,
|
||||||
ASSERT(Common::IsAligned((address), PageSize));
|
size_t num_block_shifts) {
|
||||||
|
// Check our assumptions.
|
||||||
|
ASSERT(Common::IsAligned(address, PageSize));
|
||||||
ASSERT(Common::IsAligned(size, PageSize));
|
ASSERT(Common::IsAligned(size, PageSize));
|
||||||
|
ASSERT(0 < num_block_shifts && num_block_shifts <= NumMemoryBlockPageShifts);
|
||||||
|
const VAddr management_end = management_address + management_size;
|
||||||
|
|
||||||
// Set our members
|
// Set our members.
|
||||||
heap_address = address;
|
m_heap_address = address;
|
||||||
heap_size = size;
|
m_heap_size = size;
|
||||||
|
m_num_blocks = num_block_shifts;
|
||||||
|
|
||||||
// Setup bitmaps
|
// Setup bitmaps.
|
||||||
metadata.resize(metadata_size / sizeof(u64));
|
m_management_data.resize(management_size / sizeof(u64));
|
||||||
u64* cur_bitmap_storage{metadata.data()};
|
u64* cur_bitmap_storage{m_management_data.data()};
|
||||||
for (std::size_t i = 0; i < MemoryBlockPageShifts.size(); i++) {
|
for (size_t i = 0; i < num_block_shifts; i++) {
|
||||||
const std::size_t cur_block_shift{MemoryBlockPageShifts[i]};
|
const size_t cur_block_shift = block_shifts[i];
|
||||||
const std::size_t next_block_shift{
|
const size_t next_block_shift = (i != num_block_shifts - 1) ? block_shifts[i + 1] : 0;
|
||||||
(i != MemoryBlockPageShifts.size() - 1) ? MemoryBlockPageShifts[i + 1] : 0};
|
cur_bitmap_storage = m_blocks[i].Initialize(m_heap_address, m_heap_size, cur_block_shift,
|
||||||
cur_bitmap_storage = blocks[i].Initialize(heap_address, heap_size, cur_block_shift,
|
next_block_shift, cur_bitmap_storage);
|
||||||
next_block_shift, cur_bitmap_storage);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Ensure we didn't overextend our bounds.
|
||||||
|
ASSERT(VAddr(cur_bitmap_storage) <= management_end);
|
||||||
}
|
}
|
||||||
|
|
||||||
VAddr KPageHeap::AllocateBlock(s32 index, bool random) {
|
size_t KPageHeap::GetNumFreePages() const {
|
||||||
const std::size_t needed_size{blocks[index].GetSize()};
|
size_t num_free = 0;
|
||||||
|
|
||||||
for (s32 i{index}; i < static_cast<s32>(MemoryBlockPageShifts.size()); i++) {
|
for (size_t i = 0; i < m_num_blocks; i++) {
|
||||||
if (const VAddr addr{blocks[i].PopBlock(random)}; addr) {
|
num_free += m_blocks[i].GetNumFreePages();
|
||||||
if (const std::size_t allocated_size{blocks[i].GetSize()};
|
}
|
||||||
allocated_size > needed_size) {
|
|
||||||
Free(addr + needed_size, (allocated_size - needed_size) / PageSize);
|
return num_free;
|
||||||
|
}
|
||||||
|
|
||||||
|
PAddr KPageHeap::AllocateBlock(s32 index, bool random) {
|
||||||
|
const size_t needed_size = m_blocks[index].GetSize();
|
||||||
|
|
||||||
|
for (s32 i = index; i < static_cast<s32>(m_num_blocks); i++) {
|
||||||
|
if (const PAddr addr = m_blocks[i].PopBlock(random); addr != 0) {
|
||||||
|
if (const size_t allocated_size = m_blocks[i].GetSize(); allocated_size > needed_size) {
|
||||||
|
this->Free(addr + needed_size, (allocated_size - needed_size) / PageSize);
|
||||||
}
|
}
|
||||||
return addr;
|
return addr;
|
||||||
}
|
}
|
||||||
@ -44,34 +60,34 @@ VAddr KPageHeap::AllocateBlock(s32 index, bool random) {
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
void KPageHeap::FreeBlock(VAddr block, s32 index) {
|
void KPageHeap::FreeBlock(PAddr block, s32 index) {
|
||||||
do {
|
do {
|
||||||
block = blocks[index++].PushBlock(block);
|
block = m_blocks[index++].PushBlock(block);
|
||||||
} while (block != 0);
|
} while (block != 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
void KPageHeap::Free(VAddr addr, std::size_t num_pages) {
|
void KPageHeap::Free(PAddr addr, size_t num_pages) {
|
||||||
// Freeing no pages is a no-op
|
// Freeing no pages is a no-op.
|
||||||
if (num_pages == 0) {
|
if (num_pages == 0) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Find the largest block size that we can free, and free as many as possible
|
// Find the largest block size that we can free, and free as many as possible.
|
||||||
s32 big_index{static_cast<s32>(MemoryBlockPageShifts.size()) - 1};
|
s32 big_index = static_cast<s32>(m_num_blocks) - 1;
|
||||||
const VAddr start{addr};
|
const PAddr start = addr;
|
||||||
const VAddr end{(num_pages * PageSize) + addr};
|
const PAddr end = addr + num_pages * PageSize;
|
||||||
VAddr before_start{start};
|
PAddr before_start = start;
|
||||||
VAddr before_end{start};
|
PAddr before_end = start;
|
||||||
VAddr after_start{end};
|
PAddr after_start = end;
|
||||||
VAddr after_end{end};
|
PAddr after_end = end;
|
||||||
while (big_index >= 0) {
|
while (big_index >= 0) {
|
||||||
const std::size_t block_size{blocks[big_index].GetSize()};
|
const size_t block_size = m_blocks[big_index].GetSize();
|
||||||
const VAddr big_start{Common::AlignUp((start), block_size)};
|
const PAddr big_start = Common::AlignUp(start, block_size);
|
||||||
const VAddr big_end{Common::AlignDown((end), block_size)};
|
const PAddr big_end = Common::AlignDown(end, block_size);
|
||||||
if (big_start < big_end) {
|
if (big_start < big_end) {
|
||||||
// Free as many big blocks as we can
|
// Free as many big blocks as we can.
|
||||||
for (auto block{big_start}; block < big_end; block += block_size) {
|
for (auto block = big_start; block < big_end; block += block_size) {
|
||||||
FreeBlock(block, big_index);
|
this->FreeBlock(block, big_index);
|
||||||
}
|
}
|
||||||
before_end = big_start;
|
before_end = big_start;
|
||||||
after_start = big_end;
|
after_start = big_end;
|
||||||
@ -81,31 +97,31 @@ void KPageHeap::Free(VAddr addr, std::size_t num_pages) {
|
|||||||
}
|
}
|
||||||
ASSERT(big_index >= 0);
|
ASSERT(big_index >= 0);
|
||||||
|
|
||||||
// Free space before the big blocks
|
// Free space before the big blocks.
|
||||||
for (s32 i{big_index - 1}; i >= 0; i--) {
|
for (s32 i = big_index - 1; i >= 0; i--) {
|
||||||
const std::size_t block_size{blocks[i].GetSize()};
|
const size_t block_size = m_blocks[i].GetSize();
|
||||||
while (before_start + block_size <= before_end) {
|
while (before_start + block_size <= before_end) {
|
||||||
before_end -= block_size;
|
before_end -= block_size;
|
||||||
FreeBlock(before_end, i);
|
this->FreeBlock(before_end, i);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Free space after the big blocks
|
// Free space after the big blocks.
|
||||||
for (s32 i{big_index - 1}; i >= 0; i--) {
|
for (s32 i = big_index - 1; i >= 0; i--) {
|
||||||
const std::size_t block_size{blocks[i].GetSize()};
|
const size_t block_size = m_blocks[i].GetSize();
|
||||||
while (after_start + block_size <= after_end) {
|
while (after_start + block_size <= after_end) {
|
||||||
FreeBlock(after_start, i);
|
this->FreeBlock(after_start, i);
|
||||||
after_start += block_size;
|
after_start += block_size;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::size_t KPageHeap::CalculateManagementOverheadSize(std::size_t region_size) {
|
size_t KPageHeap::CalculateManagementOverheadSize(size_t region_size, const size_t* block_shifts,
|
||||||
std::size_t overhead_size = 0;
|
size_t num_block_shifts) {
|
||||||
for (std::size_t i = 0; i < MemoryBlockPageShifts.size(); i++) {
|
size_t overhead_size = 0;
|
||||||
const std::size_t cur_block_shift{MemoryBlockPageShifts[i]};
|
for (size_t i = 0; i < num_block_shifts; i++) {
|
||||||
const std::size_t next_block_shift{
|
const size_t cur_block_shift = block_shifts[i];
|
||||||
(i != MemoryBlockPageShifts.size() - 1) ? MemoryBlockPageShifts[i + 1] : 0};
|
const size_t next_block_shift = (i != num_block_shifts - 1) ? block_shifts[i + 1] : 0;
|
||||||
overhead_size += KPageHeap::Block::CalculateManagementOverheadSize(
|
overhead_size += KPageHeap::Block::CalculateManagementOverheadSize(
|
||||||
region_size, cur_block_shift, next_block_shift);
|
region_size, cur_block_shift, next_block_shift);
|
||||||
}
|
}
|
||||||
|
@ -23,54 +23,73 @@ public:
|
|||||||
KPageHeap() = default;
|
KPageHeap() = default;
|
||||||
~KPageHeap() = default;
|
~KPageHeap() = default;
|
||||||
|
|
||||||
constexpr VAddr GetAddress() const {
|
constexpr PAddr GetAddress() const {
|
||||||
return heap_address;
|
return m_heap_address;
|
||||||
}
|
}
|
||||||
constexpr std::size_t GetSize() const {
|
constexpr size_t GetSize() const {
|
||||||
return heap_size;
|
return m_heap_size;
|
||||||
}
|
}
|
||||||
constexpr VAddr GetEndAddress() const {
|
constexpr PAddr GetEndAddress() const {
|
||||||
return GetAddress() + GetSize();
|
return this->GetAddress() + this->GetSize();
|
||||||
}
|
}
|
||||||
constexpr std::size_t GetPageOffset(VAddr block) const {
|
constexpr size_t GetPageOffset(PAddr block) const {
|
||||||
return (block - GetAddress()) / PageSize;
|
return (block - this->GetAddress()) / PageSize;
|
||||||
|
}
|
||||||
|
constexpr size_t GetPageOffsetToEnd(PAddr block) const {
|
||||||
|
return (this->GetEndAddress() - block) / PageSize;
|
||||||
}
|
}
|
||||||
|
|
||||||
void Initialize(VAddr heap_address, std::size_t heap_size, std::size_t metadata_size);
|
void Initialize(PAddr heap_address, size_t heap_size, VAddr management_address,
|
||||||
VAddr AllocateBlock(s32 index, bool random);
|
size_t management_size) {
|
||||||
void Free(VAddr addr, std::size_t num_pages);
|
return this->Initialize(heap_address, heap_size, management_address, management_size,
|
||||||
|
MemoryBlockPageShifts.data(), NumMemoryBlockPageShifts);
|
||||||
void UpdateUsedSize() {
|
|
||||||
used_size = heap_size - (GetNumFreePages() * PageSize);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static std::size_t CalculateManagementOverheadSize(std::size_t region_size);
|
size_t GetFreeSize() const {
|
||||||
|
return this->GetNumFreePages() * PageSize;
|
||||||
|
}
|
||||||
|
|
||||||
static constexpr s32 GetAlignedBlockIndex(std::size_t num_pages, std::size_t align_pages) {
|
void SetInitialUsedSize(size_t reserved_size) {
|
||||||
const auto target_pages{std::max(num_pages, align_pages)};
|
// Check that the reserved size is valid.
|
||||||
for (std::size_t i = 0; i < NumMemoryBlockPageShifts; i++) {
|
const size_t free_size = this->GetNumFreePages() * PageSize;
|
||||||
if (target_pages <=
|
ASSERT(m_heap_size >= free_size + reserved_size);
|
||||||
(static_cast<std::size_t>(1) << MemoryBlockPageShifts[i]) / PageSize) {
|
|
||||||
|
// Set the initial used size.
|
||||||
|
m_initial_used_size = m_heap_size - free_size - reserved_size;
|
||||||
|
}
|
||||||
|
|
||||||
|
PAddr AllocateBlock(s32 index, bool random);
|
||||||
|
void Free(PAddr addr, size_t num_pages);
|
||||||
|
|
||||||
|
static size_t CalculateManagementOverheadSize(size_t region_size) {
|
||||||
|
return CalculateManagementOverheadSize(region_size, MemoryBlockPageShifts.data(),
|
||||||
|
NumMemoryBlockPageShifts);
|
||||||
|
}
|
||||||
|
|
||||||
|
static constexpr s32 GetAlignedBlockIndex(size_t num_pages, size_t align_pages) {
|
||||||
|
const size_t target_pages = std::max(num_pages, align_pages);
|
||||||
|
for (size_t i = 0; i < NumMemoryBlockPageShifts; i++) {
|
||||||
|
if (target_pages <= (size_t(1) << MemoryBlockPageShifts[i]) / PageSize) {
|
||||||
return static_cast<s32>(i);
|
return static_cast<s32>(i);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static constexpr s32 GetBlockIndex(std::size_t num_pages) {
|
static constexpr s32 GetBlockIndex(size_t num_pages) {
|
||||||
for (s32 i{static_cast<s32>(NumMemoryBlockPageShifts) - 1}; i >= 0; i--) {
|
for (s32 i = static_cast<s32>(NumMemoryBlockPageShifts) - 1; i >= 0; i--) {
|
||||||
if (num_pages >= (static_cast<std::size_t>(1) << MemoryBlockPageShifts[i]) / PageSize) {
|
if (num_pages >= (size_t(1) << MemoryBlockPageShifts[i]) / PageSize) {
|
||||||
return i;
|
return i;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static constexpr std::size_t GetBlockSize(std::size_t index) {
|
static constexpr size_t GetBlockSize(size_t index) {
|
||||||
return static_cast<std::size_t>(1) << MemoryBlockPageShifts[index];
|
return size_t(1) << MemoryBlockPageShifts[index];
|
||||||
}
|
}
|
||||||
|
|
||||||
static constexpr std::size_t GetBlockNumPages(std::size_t index) {
|
static constexpr size_t GetBlockNumPages(size_t index) {
|
||||||
return GetBlockSize(index) / PageSize;
|
return GetBlockSize(index) / PageSize;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -83,114 +102,116 @@ private:
|
|||||||
Block() = default;
|
Block() = default;
|
||||||
~Block() = default;
|
~Block() = default;
|
||||||
|
|
||||||
constexpr std::size_t GetShift() const {
|
constexpr size_t GetShift() const {
|
||||||
return block_shift;
|
return m_block_shift;
|
||||||
}
|
}
|
||||||
constexpr std::size_t GetNextShift() const {
|
constexpr size_t GetNextShift() const {
|
||||||
return next_block_shift;
|
return m_next_block_shift;
|
||||||
}
|
}
|
||||||
constexpr std::size_t GetSize() const {
|
constexpr size_t GetSize() const {
|
||||||
return static_cast<std::size_t>(1) << GetShift();
|
return u64(1) << this->GetShift();
|
||||||
}
|
}
|
||||||
constexpr std::size_t GetNumPages() const {
|
constexpr size_t GetNumPages() const {
|
||||||
return GetSize() / PageSize;
|
return this->GetSize() / PageSize;
|
||||||
}
|
}
|
||||||
constexpr std::size_t GetNumFreeBlocks() const {
|
constexpr size_t GetNumFreeBlocks() const {
|
||||||
return bitmap.GetNumBits();
|
return m_bitmap.GetNumBits();
|
||||||
}
|
}
|
||||||
constexpr std::size_t GetNumFreePages() const {
|
constexpr size_t GetNumFreePages() const {
|
||||||
return GetNumFreeBlocks() * GetNumPages();
|
return this->GetNumFreeBlocks() * this->GetNumPages();
|
||||||
}
|
}
|
||||||
|
|
||||||
u64* Initialize(VAddr addr, std::size_t size, std::size_t bs, std::size_t nbs,
|
u64* Initialize(PAddr addr, size_t size, size_t bs, size_t nbs, u64* bit_storage) {
|
||||||
u64* bit_storage) {
|
// Set shifts.
|
||||||
// Set shifts
|
m_block_shift = bs;
|
||||||
block_shift = bs;
|
m_next_block_shift = nbs;
|
||||||
next_block_shift = nbs;
|
|
||||||
|
|
||||||
// Align up the address
|
// Align up the address.
|
||||||
VAddr end{addr + size};
|
PAddr end = addr + size;
|
||||||
const auto align{(next_block_shift != 0) ? (1ULL << next_block_shift)
|
const size_t align = (m_next_block_shift != 0) ? (u64(1) << m_next_block_shift)
|
||||||
: (1ULL << block_shift)};
|
: (u64(1) << m_block_shift);
|
||||||
addr = Common::AlignDown((addr), align);
|
addr = Common::AlignDown(addr, align);
|
||||||
end = Common::AlignUp((end), align);
|
end = Common::AlignUp(end, align);
|
||||||
|
|
||||||
heap_address = addr;
|
m_heap_address = addr;
|
||||||
end_offset = (end - addr) / (1ULL << block_shift);
|
m_end_offset = (end - addr) / (u64(1) << m_block_shift);
|
||||||
return bitmap.Initialize(bit_storage, end_offset);
|
return m_bitmap.Initialize(bit_storage, m_end_offset);
|
||||||
}
|
}
|
||||||
|
|
||||||
VAddr PushBlock(VAddr address) {
|
PAddr PushBlock(PAddr address) {
|
||||||
// Set the bit for the free block
|
// Set the bit for the free block.
|
||||||
std::size_t offset{(address - heap_address) >> GetShift()};
|
size_t offset = (address - m_heap_address) >> this->GetShift();
|
||||||
bitmap.SetBit(offset);
|
m_bitmap.SetBit(offset);
|
||||||
|
|
||||||
// If we have a next shift, try to clear the blocks below and return the address
|
// If we have a next shift, try to clear the blocks below this one and return the new
|
||||||
if (GetNextShift()) {
|
// address.
|
||||||
const auto diff{1ULL << (GetNextShift() - GetShift())};
|
if (this->GetNextShift()) {
|
||||||
|
const size_t diff = u64(1) << (this->GetNextShift() - this->GetShift());
|
||||||
offset = Common::AlignDown(offset, diff);
|
offset = Common::AlignDown(offset, diff);
|
||||||
if (bitmap.ClearRange(offset, diff)) {
|
if (m_bitmap.ClearRange(offset, diff)) {
|
||||||
return heap_address + (offset << GetShift());
|
return m_heap_address + (offset << this->GetShift());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// We couldn't coalesce, or we're already as big as possible
|
// We couldn't coalesce, or we're already as big as possible.
|
||||||
return 0;
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
VAddr PopBlock(bool random) {
|
PAddr PopBlock(bool random) {
|
||||||
// Find a free block
|
// Find a free block.
|
||||||
const s64 soffset{bitmap.FindFreeBlock(random)};
|
s64 soffset = m_bitmap.FindFreeBlock(random);
|
||||||
if (soffset < 0) {
|
if (soffset < 0) {
|
||||||
return 0;
|
return {};
|
||||||
}
|
}
|
||||||
const auto offset{static_cast<std::size_t>(soffset)};
|
const size_t offset = static_cast<size_t>(soffset);
|
||||||
|
|
||||||
// Update our tracking and return it
|
// Update our tracking and return it.
|
||||||
bitmap.ClearBit(offset);
|
m_bitmap.ClearBit(offset);
|
||||||
return heap_address + (offset << GetShift());
|
return m_heap_address + (offset << this->GetShift());
|
||||||
}
|
}
|
||||||
|
|
||||||
static constexpr std::size_t CalculateManagementOverheadSize(std::size_t region_size,
|
public:
|
||||||
std::size_t cur_block_shift,
|
static constexpr size_t CalculateManagementOverheadSize(size_t region_size,
|
||||||
std::size_t next_block_shift) {
|
size_t cur_block_shift,
|
||||||
const auto cur_block_size{(1ULL << cur_block_shift)};
|
size_t next_block_shift) {
|
||||||
const auto next_block_size{(1ULL << next_block_shift)};
|
const size_t cur_block_size = (u64(1) << cur_block_shift);
|
||||||
const auto align{(next_block_shift != 0) ? next_block_size : cur_block_size};
|
const size_t next_block_size = (u64(1) << next_block_shift);
|
||||||
|
const size_t align = (next_block_shift != 0) ? next_block_size : cur_block_size;
|
||||||
return KPageBitmap::CalculateManagementOverheadSize(
|
return KPageBitmap::CalculateManagementOverheadSize(
|
||||||
(align * 2 + Common::AlignUp(region_size, align)) / cur_block_size);
|
(align * 2 + Common::AlignUp(region_size, align)) / cur_block_size);
|
||||||
}
|
}
|
||||||
|
|
||||||
private:
|
private:
|
||||||
KPageBitmap bitmap;
|
KPageBitmap m_bitmap;
|
||||||
VAddr heap_address{};
|
PAddr m_heap_address{};
|
||||||
uintptr_t end_offset{};
|
uintptr_t m_end_offset{};
|
||||||
std::size_t block_shift{};
|
size_t m_block_shift{};
|
||||||
std::size_t next_block_shift{};
|
size_t m_next_block_shift{};
|
||||||
};
|
};
|
||||||
|
|
||||||
constexpr std::size_t GetNumFreePages() const {
|
private:
|
||||||
std::size_t num_free{};
|
void Initialize(PAddr heap_address, size_t heap_size, VAddr management_address,
|
||||||
|
size_t management_size, const size_t* block_shifts, size_t num_block_shifts);
|
||||||
|
size_t GetNumFreePages() const;
|
||||||
|
|
||||||
for (const auto& block : blocks) {
|
void FreeBlock(PAddr block, s32 index);
|
||||||
num_free += block.GetNumFreePages();
|
|
||||||
}
|
|
||||||
|
|
||||||
return num_free;
|
static constexpr size_t NumMemoryBlockPageShifts{7};
|
||||||
}
|
static constexpr std::array<size_t, NumMemoryBlockPageShifts> MemoryBlockPageShifts{
|
||||||
|
|
||||||
void FreeBlock(VAddr block, s32 index);
|
|
||||||
|
|
||||||
static constexpr std::size_t NumMemoryBlockPageShifts{7};
|
|
||||||
static constexpr std::array<std::size_t, NumMemoryBlockPageShifts> MemoryBlockPageShifts{
|
|
||||||
0xC, 0x10, 0x15, 0x16, 0x19, 0x1D, 0x1E,
|
0xC, 0x10, 0x15, 0x16, 0x19, 0x1D, 0x1E,
|
||||||
};
|
};
|
||||||
|
|
||||||
VAddr heap_address{};
|
private:
|
||||||
std::size_t heap_size{};
|
static size_t CalculateManagementOverheadSize(size_t region_size, const size_t* block_shifts,
|
||||||
std::size_t used_size{};
|
size_t num_block_shifts);
|
||||||
std::array<Block, NumMemoryBlockPageShifts> blocks{};
|
|
||||||
std::vector<u64> metadata;
|
private:
|
||||||
|
PAddr m_heap_address{};
|
||||||
|
size_t m_heap_size{};
|
||||||
|
size_t m_initial_used_size{};
|
||||||
|
size_t m_num_blocks{};
|
||||||
|
std::array<Block, NumMemoryBlockPageShifts> m_blocks{};
|
||||||
|
std::vector<u64> m_management_data;
|
||||||
};
|
};
|
||||||
|
|
||||||
} // namespace Kernel
|
} // namespace Kernel
|
||||||
|
@ -273,11 +273,12 @@ ResultCode KPageTable::MapProcessCode(VAddr addr, std::size_t num_pages, KMemory
|
|||||||
R_TRY(this->CheckMemoryState(addr, size, KMemoryState::All, KMemoryState::Free,
|
R_TRY(this->CheckMemoryState(addr, size, KMemoryState::All, KMemoryState::Free,
|
||||||
KMemoryPermission::None, KMemoryPermission::None,
|
KMemoryPermission::None, KMemoryPermission::None,
|
||||||
KMemoryAttribute::None, KMemoryAttribute::None));
|
KMemoryAttribute::None, KMemoryAttribute::None));
|
||||||
|
KPageLinkedList pg;
|
||||||
|
R_TRY(system.Kernel().MemoryManager().AllocateAndOpen(
|
||||||
|
&pg, num_pages,
|
||||||
|
KMemoryManager::EncodeOption(KMemoryManager::Pool::Application, allocation_option)));
|
||||||
|
|
||||||
KPageLinkedList page_linked_list;
|
R_TRY(Operate(addr, num_pages, pg, OperationType::MapGroup));
|
||||||
R_TRY(system.Kernel().MemoryManager().Allocate(page_linked_list, num_pages, memory_pool,
|
|
||||||
allocation_option));
|
|
||||||
R_TRY(Operate(addr, num_pages, page_linked_list, OperationType::MapGroup));
|
|
||||||
|
|
||||||
block_manager->Update(addr, num_pages, state, perm);
|
block_manager->Update(addr, num_pages, state, perm);
|
||||||
|
|
||||||
@ -443,9 +444,10 @@ ResultCode KPageTable::MapPhysicalMemory(VAddr address, std::size_t size) {
|
|||||||
R_UNLESS(memory_reservation.Succeeded(), ResultLimitReached);
|
R_UNLESS(memory_reservation.Succeeded(), ResultLimitReached);
|
||||||
|
|
||||||
// Allocate pages for the new memory.
|
// Allocate pages for the new memory.
|
||||||
KPageLinkedList page_linked_list;
|
KPageLinkedList pg;
|
||||||
R_TRY(system.Kernel().MemoryManager().Allocate(
|
R_TRY(system.Kernel().MemoryManager().AllocateAndOpenForProcess(
|
||||||
page_linked_list, (size - mapped_size) / PageSize, memory_pool, allocation_option));
|
&pg, (size - mapped_size) / PageSize,
|
||||||
|
KMemoryManager::EncodeOption(memory_pool, allocation_option), 0, 0));
|
||||||
|
|
||||||
// Map the memory.
|
// Map the memory.
|
||||||
{
|
{
|
||||||
@ -547,7 +549,7 @@ ResultCode KPageTable::MapPhysicalMemory(VAddr address, std::size_t size) {
|
|||||||
});
|
});
|
||||||
|
|
||||||
// Iterate over the memory.
|
// Iterate over the memory.
|
||||||
auto pg_it = page_linked_list.Nodes().begin();
|
auto pg_it = pg.Nodes().begin();
|
||||||
PAddr pg_phys_addr = pg_it->GetAddress();
|
PAddr pg_phys_addr = pg_it->GetAddress();
|
||||||
size_t pg_pages = pg_it->GetNumPages();
|
size_t pg_pages = pg_it->GetNumPages();
|
||||||
|
|
||||||
@ -571,7 +573,7 @@ ResultCode KPageTable::MapPhysicalMemory(VAddr address, std::size_t size) {
|
|||||||
// Check if we're at the end of the physical block.
|
// Check if we're at the end of the physical block.
|
||||||
if (pg_pages == 0) {
|
if (pg_pages == 0) {
|
||||||
// Ensure there are more pages to map.
|
// Ensure there are more pages to map.
|
||||||
ASSERT(pg_it != page_linked_list.Nodes().end());
|
ASSERT(pg_it != pg.Nodes().end());
|
||||||
|
|
||||||
// Advance our physical block.
|
// Advance our physical block.
|
||||||
++pg_it;
|
++pg_it;
|
||||||
@ -841,10 +843,14 @@ ResultCode KPageTable::UnmapPhysicalMemory(VAddr address, std::size_t size) {
|
|||||||
process->GetResourceLimit()->Release(LimitableResource::PhysicalMemory, mapped_size);
|
process->GetResourceLimit()->Release(LimitableResource::PhysicalMemory, mapped_size);
|
||||||
|
|
||||||
// Update memory blocks.
|
// Update memory blocks.
|
||||||
system.Kernel().MemoryManager().Free(pg, size / PageSize, memory_pool, allocation_option);
|
|
||||||
block_manager->Update(address, size / PageSize, KMemoryState::Free, KMemoryPermission::None,
|
block_manager->Update(address, size / PageSize, KMemoryState::Free, KMemoryPermission::None,
|
||||||
KMemoryAttribute::None);
|
KMemoryAttribute::None);
|
||||||
|
|
||||||
|
// TODO(bunnei): This is a workaround until the next set of changes, where we add reference
|
||||||
|
// counting for mapped pages. Until then, we must manually close the reference to the page
|
||||||
|
// group.
|
||||||
|
system.Kernel().MemoryManager().Close(pg);
|
||||||
|
|
||||||
// We succeeded.
|
// We succeeded.
|
||||||
remap_guard.Cancel();
|
remap_guard.Cancel();
|
||||||
|
|
||||||
@ -1270,9 +1276,16 @@ ResultCode KPageTable::SetHeapSize(VAddr* out, std::size_t size) {
|
|||||||
R_UNLESS(memory_reservation.Succeeded(), ResultLimitReached);
|
R_UNLESS(memory_reservation.Succeeded(), ResultLimitReached);
|
||||||
|
|
||||||
// Allocate pages for the heap extension.
|
// Allocate pages for the heap extension.
|
||||||
KPageLinkedList page_linked_list;
|
KPageLinkedList pg;
|
||||||
R_TRY(system.Kernel().MemoryManager().Allocate(page_linked_list, allocation_size / PageSize,
|
R_TRY(system.Kernel().MemoryManager().AllocateAndOpen(
|
||||||
memory_pool, allocation_option));
|
&pg, allocation_size / PageSize,
|
||||||
|
KMemoryManager::EncodeOption(memory_pool, allocation_option)));
|
||||||
|
|
||||||
|
// Clear all the newly allocated pages.
|
||||||
|
for (const auto& it : pg.Nodes()) {
|
||||||
|
std::memset(system.DeviceMemory().GetPointer(it.GetAddress()), heap_fill_value,
|
||||||
|
it.GetSize());
|
||||||
|
}
|
||||||
|
|
||||||
// Map the pages.
|
// Map the pages.
|
||||||
{
|
{
|
||||||
@ -1291,7 +1304,7 @@ ResultCode KPageTable::SetHeapSize(VAddr* out, std::size_t size) {
|
|||||||
|
|
||||||
// Map the pages.
|
// Map the pages.
|
||||||
const auto num_pages = allocation_size / PageSize;
|
const auto num_pages = allocation_size / PageSize;
|
||||||
R_TRY(Operate(current_heap_end, num_pages, page_linked_list, OperationType::MapGroup));
|
R_TRY(Operate(current_heap_end, num_pages, pg, OperationType::MapGroup));
|
||||||
|
|
||||||
// Clear all the newly allocated pages.
|
// Clear all the newly allocated pages.
|
||||||
for (std::size_t cur_page = 0; cur_page < num_pages; ++cur_page) {
|
for (std::size_t cur_page = 0; cur_page < num_pages; ++cur_page) {
|
||||||
@ -1339,8 +1352,9 @@ ResultVal<VAddr> KPageTable::AllocateAndMapMemory(std::size_t needed_num_pages,
|
|||||||
R_TRY(Operate(addr, needed_num_pages, perm, OperationType::Map, map_addr));
|
R_TRY(Operate(addr, needed_num_pages, perm, OperationType::Map, map_addr));
|
||||||
} else {
|
} else {
|
||||||
KPageLinkedList page_group;
|
KPageLinkedList page_group;
|
||||||
R_TRY(system.Kernel().MemoryManager().Allocate(page_group, needed_num_pages, memory_pool,
|
R_TRY(system.Kernel().MemoryManager().AllocateAndOpenForProcess(
|
||||||
allocation_option));
|
&page_group, needed_num_pages,
|
||||||
|
KMemoryManager::EncodeOption(memory_pool, allocation_option), 0, 0));
|
||||||
R_TRY(Operate(addr, needed_num_pages, page_group, OperationType::MapGroup));
|
R_TRY(Operate(addr, needed_num_pages, page_group, OperationType::MapGroup));
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1547,7 +1561,7 @@ ResultCode KPageTable::Operate(VAddr addr, std::size_t num_pages, KMemoryPermiss
|
|||||||
return ResultSuccess;
|
return ResultSuccess;
|
||||||
}
|
}
|
||||||
|
|
||||||
constexpr VAddr KPageTable::GetRegionAddress(KMemoryState state) const {
|
VAddr KPageTable::GetRegionAddress(KMemoryState state) const {
|
||||||
switch (state) {
|
switch (state) {
|
||||||
case KMemoryState::Free:
|
case KMemoryState::Free:
|
||||||
case KMemoryState::Kernel:
|
case KMemoryState::Kernel:
|
||||||
@ -1583,7 +1597,7 @@ constexpr VAddr KPageTable::GetRegionAddress(KMemoryState state) const {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
constexpr std::size_t KPageTable::GetRegionSize(KMemoryState state) const {
|
std::size_t KPageTable::GetRegionSize(KMemoryState state) const {
|
||||||
switch (state) {
|
switch (state) {
|
||||||
case KMemoryState::Free:
|
case KMemoryState::Free:
|
||||||
case KMemoryState::Kernel:
|
case KMemoryState::Kernel:
|
||||||
|
@ -102,8 +102,8 @@ private:
|
|||||||
OperationType operation);
|
OperationType operation);
|
||||||
ResultCode Operate(VAddr addr, std::size_t num_pages, KMemoryPermission perm,
|
ResultCode Operate(VAddr addr, std::size_t num_pages, KMemoryPermission perm,
|
||||||
OperationType operation, PAddr map_addr = 0);
|
OperationType operation, PAddr map_addr = 0);
|
||||||
constexpr VAddr GetRegionAddress(KMemoryState state) const;
|
VAddr GetRegionAddress(KMemoryState state) const;
|
||||||
constexpr std::size_t GetRegionSize(KMemoryState state) const;
|
std::size_t GetRegionSize(KMemoryState state) const;
|
||||||
|
|
||||||
ResultCode CheckMemoryStateContiguous(std::size_t* out_blocks_needed, VAddr addr,
|
ResultCode CheckMemoryStateContiguous(std::size_t* out_blocks_needed, VAddr addr,
|
||||||
std::size_t size, KMemoryState state_mask,
|
std::size_t size, KMemoryState state_mask,
|
||||||
@ -254,8 +254,7 @@ public:
|
|||||||
return !IsOutsideASLRRegion(address, size);
|
return !IsOutsideASLRRegion(address, size);
|
||||||
}
|
}
|
||||||
|
|
||||||
PAddr GetPhysicalAddr(VAddr addr) {
|
PAddr GetPhysicalAddr(VAddr addr) const {
|
||||||
ASSERT(IsLockedByCurrentThread());
|
|
||||||
const auto backing_addr = page_table_impl.backing_addr[addr >> PageBits];
|
const auto backing_addr = page_table_impl.backing_addr[addr >> PageBits];
|
||||||
ASSERT(backing_addr);
|
ASSERT(backing_addr);
|
||||||
return backing_addr + addr;
|
return backing_addr + addr;
|
||||||
@ -311,6 +310,8 @@ private:
|
|||||||
bool is_kernel{};
|
bool is_kernel{};
|
||||||
bool is_aslr_enabled{};
|
bool is_aslr_enabled{};
|
||||||
|
|
||||||
|
u32 heap_fill_value{};
|
||||||
|
|
||||||
KMemoryManager::Pool memory_pool{KMemoryManager::Pool::Application};
|
KMemoryManager::Pool memory_pool{KMemoryManager::Pool::Application};
|
||||||
KMemoryManager::Direction allocation_option{KMemoryManager::Direction::FromFront};
|
KMemoryManager::Direction allocation_option{KMemoryManager::Direction::FromFront};
|
||||||
|
|
||||||
|
@ -70,13 +70,12 @@ struct KernelCore::Impl {
|
|||||||
|
|
||||||
// Derive the initial memory layout from the emulated board
|
// Derive the initial memory layout from the emulated board
|
||||||
Init::InitializeSlabResourceCounts(kernel);
|
Init::InitializeSlabResourceCounts(kernel);
|
||||||
KMemoryLayout memory_layout;
|
DeriveInitialMemoryLayout();
|
||||||
DeriveInitialMemoryLayout(memory_layout);
|
Init::InitializeSlabHeaps(system, *memory_layout);
|
||||||
Init::InitializeSlabHeaps(system, memory_layout);
|
|
||||||
|
|
||||||
// Initialize kernel memory and resources.
|
// Initialize kernel memory and resources.
|
||||||
InitializeSystemResourceLimit(kernel, system.CoreTiming(), memory_layout);
|
InitializeSystemResourceLimit(kernel, system.CoreTiming());
|
||||||
InitializeMemoryLayout(memory_layout);
|
InitializeMemoryLayout();
|
||||||
InitializePageSlab();
|
InitializePageSlab();
|
||||||
InitializeSchedulers();
|
InitializeSchedulers();
|
||||||
InitializeSuspendThreads();
|
InitializeSuspendThreads();
|
||||||
@ -219,12 +218,11 @@ struct KernelCore::Impl {
|
|||||||
|
|
||||||
// Creates the default system resource limit
|
// Creates the default system resource limit
|
||||||
void InitializeSystemResourceLimit(KernelCore& kernel,
|
void InitializeSystemResourceLimit(KernelCore& kernel,
|
||||||
const Core::Timing::CoreTiming& core_timing,
|
const Core::Timing::CoreTiming& core_timing) {
|
||||||
const KMemoryLayout& memory_layout) {
|
|
||||||
system_resource_limit = KResourceLimit::Create(system.Kernel());
|
system_resource_limit = KResourceLimit::Create(system.Kernel());
|
||||||
system_resource_limit->Initialize(&core_timing);
|
system_resource_limit->Initialize(&core_timing);
|
||||||
|
|
||||||
const auto [total_size, kernel_size] = memory_layout.GetTotalAndKernelMemorySizes();
|
const auto [total_size, kernel_size] = memory_layout->GetTotalAndKernelMemorySizes();
|
||||||
|
|
||||||
// If setting the default system values fails, then something seriously wrong has occurred.
|
// If setting the default system values fails, then something seriously wrong has occurred.
|
||||||
ASSERT(system_resource_limit->SetLimitValue(LimitableResource::PhysicalMemory, total_size)
|
ASSERT(system_resource_limit->SetLimitValue(LimitableResource::PhysicalMemory, total_size)
|
||||||
@ -353,16 +351,18 @@ struct KernelCore::Impl {
|
|||||||
return schedulers[thread_id]->GetCurrentThread();
|
return schedulers[thread_id]->GetCurrentThread();
|
||||||
}
|
}
|
||||||
|
|
||||||
void DeriveInitialMemoryLayout(KMemoryLayout& memory_layout) {
|
void DeriveInitialMemoryLayout() {
|
||||||
|
memory_layout = std::make_unique<KMemoryLayout>();
|
||||||
|
|
||||||
// Insert the root region for the virtual memory tree, from which all other regions will
|
// Insert the root region for the virtual memory tree, from which all other regions will
|
||||||
// derive.
|
// derive.
|
||||||
memory_layout.GetVirtualMemoryRegionTree().InsertDirectly(
|
memory_layout->GetVirtualMemoryRegionTree().InsertDirectly(
|
||||||
KernelVirtualAddressSpaceBase,
|
KernelVirtualAddressSpaceBase,
|
||||||
KernelVirtualAddressSpaceBase + KernelVirtualAddressSpaceSize - 1);
|
KernelVirtualAddressSpaceBase + KernelVirtualAddressSpaceSize - 1);
|
||||||
|
|
||||||
// Insert the root region for the physical memory tree, from which all other regions will
|
// Insert the root region for the physical memory tree, from which all other regions will
|
||||||
// derive.
|
// derive.
|
||||||
memory_layout.GetPhysicalMemoryRegionTree().InsertDirectly(
|
memory_layout->GetPhysicalMemoryRegionTree().InsertDirectly(
|
||||||
KernelPhysicalAddressSpaceBase,
|
KernelPhysicalAddressSpaceBase,
|
||||||
KernelPhysicalAddressSpaceBase + KernelPhysicalAddressSpaceSize - 1);
|
KernelPhysicalAddressSpaceBase + KernelPhysicalAddressSpaceSize - 1);
|
||||||
|
|
||||||
@ -379,7 +379,7 @@ struct KernelCore::Impl {
|
|||||||
if (!(kernel_region_start + KernelRegionSize - 1 <= KernelVirtualAddressSpaceLast)) {
|
if (!(kernel_region_start + KernelRegionSize - 1 <= KernelVirtualAddressSpaceLast)) {
|
||||||
kernel_region_size = KernelVirtualAddressSpaceEnd - kernel_region_start;
|
kernel_region_size = KernelVirtualAddressSpaceEnd - kernel_region_start;
|
||||||
}
|
}
|
||||||
ASSERT(memory_layout.GetVirtualMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetVirtualMemoryRegionTree().Insert(
|
||||||
kernel_region_start, kernel_region_size, KMemoryRegionType_Kernel));
|
kernel_region_start, kernel_region_size, KMemoryRegionType_Kernel));
|
||||||
|
|
||||||
// Setup the code region.
|
// Setup the code region.
|
||||||
@ -388,11 +388,11 @@ struct KernelCore::Impl {
|
|||||||
Common::AlignDown(code_start_virt_addr, CodeRegionAlign);
|
Common::AlignDown(code_start_virt_addr, CodeRegionAlign);
|
||||||
constexpr VAddr code_region_end = Common::AlignUp(code_end_virt_addr, CodeRegionAlign);
|
constexpr VAddr code_region_end = Common::AlignUp(code_end_virt_addr, CodeRegionAlign);
|
||||||
constexpr size_t code_region_size = code_region_end - code_region_start;
|
constexpr size_t code_region_size = code_region_end - code_region_start;
|
||||||
ASSERT(memory_layout.GetVirtualMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetVirtualMemoryRegionTree().Insert(
|
||||||
code_region_start, code_region_size, KMemoryRegionType_KernelCode));
|
code_region_start, code_region_size, KMemoryRegionType_KernelCode));
|
||||||
|
|
||||||
// Setup board-specific device physical regions.
|
// Setup board-specific device physical regions.
|
||||||
Init::SetupDevicePhysicalMemoryRegions(memory_layout);
|
Init::SetupDevicePhysicalMemoryRegions(*memory_layout);
|
||||||
|
|
||||||
// Determine the amount of space needed for the misc region.
|
// Determine the amount of space needed for the misc region.
|
||||||
size_t misc_region_needed_size;
|
size_t misc_region_needed_size;
|
||||||
@ -401,7 +401,7 @@ struct KernelCore::Impl {
|
|||||||
misc_region_needed_size = Core::Hardware::NUM_CPU_CORES * (3 * (PageSize + PageSize));
|
misc_region_needed_size = Core::Hardware::NUM_CPU_CORES * (3 * (PageSize + PageSize));
|
||||||
|
|
||||||
// Account for each auto-map device.
|
// Account for each auto-map device.
|
||||||
for (const auto& region : memory_layout.GetPhysicalMemoryRegionTree()) {
|
for (const auto& region : memory_layout->GetPhysicalMemoryRegionTree()) {
|
||||||
if (region.HasTypeAttribute(KMemoryRegionAttr_ShouldKernelMap)) {
|
if (region.HasTypeAttribute(KMemoryRegionAttr_ShouldKernelMap)) {
|
||||||
// Check that the region is valid.
|
// Check that the region is valid.
|
||||||
ASSERT(region.GetEndAddress() != 0);
|
ASSERT(region.GetEndAddress() != 0);
|
||||||
@ -426,22 +426,22 @@ struct KernelCore::Impl {
|
|||||||
|
|
||||||
// Setup the misc region.
|
// Setup the misc region.
|
||||||
const VAddr misc_region_start =
|
const VAddr misc_region_start =
|
||||||
memory_layout.GetVirtualMemoryRegionTree().GetRandomAlignedRegion(
|
memory_layout->GetVirtualMemoryRegionTree().GetRandomAlignedRegion(
|
||||||
misc_region_size, MiscRegionAlign, KMemoryRegionType_Kernel);
|
misc_region_size, MiscRegionAlign, KMemoryRegionType_Kernel);
|
||||||
ASSERT(memory_layout.GetVirtualMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetVirtualMemoryRegionTree().Insert(
|
||||||
misc_region_start, misc_region_size, KMemoryRegionType_KernelMisc));
|
misc_region_start, misc_region_size, KMemoryRegionType_KernelMisc));
|
||||||
|
|
||||||
// Setup the stack region.
|
// Setup the stack region.
|
||||||
constexpr size_t StackRegionSize = 14_MiB;
|
constexpr size_t StackRegionSize = 14_MiB;
|
||||||
constexpr size_t StackRegionAlign = KernelAslrAlignment;
|
constexpr size_t StackRegionAlign = KernelAslrAlignment;
|
||||||
const VAddr stack_region_start =
|
const VAddr stack_region_start =
|
||||||
memory_layout.GetVirtualMemoryRegionTree().GetRandomAlignedRegion(
|
memory_layout->GetVirtualMemoryRegionTree().GetRandomAlignedRegion(
|
||||||
StackRegionSize, StackRegionAlign, KMemoryRegionType_Kernel);
|
StackRegionSize, StackRegionAlign, KMemoryRegionType_Kernel);
|
||||||
ASSERT(memory_layout.GetVirtualMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetVirtualMemoryRegionTree().Insert(
|
||||||
stack_region_start, StackRegionSize, KMemoryRegionType_KernelStack));
|
stack_region_start, StackRegionSize, KMemoryRegionType_KernelStack));
|
||||||
|
|
||||||
// Determine the size of the resource region.
|
// Determine the size of the resource region.
|
||||||
const size_t resource_region_size = memory_layout.GetResourceRegionSizeForInit();
|
const size_t resource_region_size = memory_layout->GetResourceRegionSizeForInit();
|
||||||
|
|
||||||
// Determine the size of the slab region.
|
// Determine the size of the slab region.
|
||||||
const size_t slab_region_size =
|
const size_t slab_region_size =
|
||||||
@ -458,23 +458,23 @@ struct KernelCore::Impl {
|
|||||||
Common::AlignUp(code_end_phys_addr + slab_region_size, SlabRegionAlign) -
|
Common::AlignUp(code_end_phys_addr + slab_region_size, SlabRegionAlign) -
|
||||||
Common::AlignDown(code_end_phys_addr, SlabRegionAlign);
|
Common::AlignDown(code_end_phys_addr, SlabRegionAlign);
|
||||||
const VAddr slab_region_start =
|
const VAddr slab_region_start =
|
||||||
memory_layout.GetVirtualMemoryRegionTree().GetRandomAlignedRegion(
|
memory_layout->GetVirtualMemoryRegionTree().GetRandomAlignedRegion(
|
||||||
slab_region_needed_size, SlabRegionAlign, KMemoryRegionType_Kernel) +
|
slab_region_needed_size, SlabRegionAlign, KMemoryRegionType_Kernel) +
|
||||||
(code_end_phys_addr % SlabRegionAlign);
|
(code_end_phys_addr % SlabRegionAlign);
|
||||||
ASSERT(memory_layout.GetVirtualMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetVirtualMemoryRegionTree().Insert(
|
||||||
slab_region_start, slab_region_size, KMemoryRegionType_KernelSlab));
|
slab_region_start, slab_region_size, KMemoryRegionType_KernelSlab));
|
||||||
|
|
||||||
// Setup the temp region.
|
// Setup the temp region.
|
||||||
constexpr size_t TempRegionSize = 128_MiB;
|
constexpr size_t TempRegionSize = 128_MiB;
|
||||||
constexpr size_t TempRegionAlign = KernelAslrAlignment;
|
constexpr size_t TempRegionAlign = KernelAslrAlignment;
|
||||||
const VAddr temp_region_start =
|
const VAddr temp_region_start =
|
||||||
memory_layout.GetVirtualMemoryRegionTree().GetRandomAlignedRegion(
|
memory_layout->GetVirtualMemoryRegionTree().GetRandomAlignedRegion(
|
||||||
TempRegionSize, TempRegionAlign, KMemoryRegionType_Kernel);
|
TempRegionSize, TempRegionAlign, KMemoryRegionType_Kernel);
|
||||||
ASSERT(memory_layout.GetVirtualMemoryRegionTree().Insert(temp_region_start, TempRegionSize,
|
ASSERT(memory_layout->GetVirtualMemoryRegionTree().Insert(temp_region_start, TempRegionSize,
|
||||||
KMemoryRegionType_KernelTemp));
|
KMemoryRegionType_KernelTemp));
|
||||||
|
|
||||||
// Automatically map in devices that have auto-map attributes.
|
// Automatically map in devices that have auto-map attributes.
|
||||||
for (auto& region : memory_layout.GetPhysicalMemoryRegionTree()) {
|
for (auto& region : memory_layout->GetPhysicalMemoryRegionTree()) {
|
||||||
// We only care about kernel regions.
|
// We only care about kernel regions.
|
||||||
if (!region.IsDerivedFrom(KMemoryRegionType_Kernel)) {
|
if (!region.IsDerivedFrom(KMemoryRegionType_Kernel)) {
|
||||||
continue;
|
continue;
|
||||||
@ -501,21 +501,21 @@ struct KernelCore::Impl {
|
|||||||
const size_t map_size =
|
const size_t map_size =
|
||||||
Common::AlignUp(region.GetEndAddress(), PageSize) - map_phys_addr;
|
Common::AlignUp(region.GetEndAddress(), PageSize) - map_phys_addr;
|
||||||
const VAddr map_virt_addr =
|
const VAddr map_virt_addr =
|
||||||
memory_layout.GetVirtualMemoryRegionTree().GetRandomAlignedRegionWithGuard(
|
memory_layout->GetVirtualMemoryRegionTree().GetRandomAlignedRegionWithGuard(
|
||||||
map_size, PageSize, KMemoryRegionType_KernelMisc, PageSize);
|
map_size, PageSize, KMemoryRegionType_KernelMisc, PageSize);
|
||||||
ASSERT(memory_layout.GetVirtualMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetVirtualMemoryRegionTree().Insert(
|
||||||
map_virt_addr, map_size, KMemoryRegionType_KernelMiscMappedDevice));
|
map_virt_addr, map_size, KMemoryRegionType_KernelMiscMappedDevice));
|
||||||
region.SetPairAddress(map_virt_addr + region.GetAddress() - map_phys_addr);
|
region.SetPairAddress(map_virt_addr + region.GetAddress() - map_phys_addr);
|
||||||
}
|
}
|
||||||
|
|
||||||
Init::SetupDramPhysicalMemoryRegions(memory_layout);
|
Init::SetupDramPhysicalMemoryRegions(*memory_layout);
|
||||||
|
|
||||||
// Insert a physical region for the kernel code region.
|
// Insert a physical region for the kernel code region.
|
||||||
ASSERT(memory_layout.GetPhysicalMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetPhysicalMemoryRegionTree().Insert(
|
||||||
code_start_phys_addr, code_region_size, KMemoryRegionType_DramKernelCode));
|
code_start_phys_addr, code_region_size, KMemoryRegionType_DramKernelCode));
|
||||||
|
|
||||||
// Insert a physical region for the kernel slab region.
|
// Insert a physical region for the kernel slab region.
|
||||||
ASSERT(memory_layout.GetPhysicalMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetPhysicalMemoryRegionTree().Insert(
|
||||||
slab_start_phys_addr, slab_region_size, KMemoryRegionType_DramKernelSlab));
|
slab_start_phys_addr, slab_region_size, KMemoryRegionType_DramKernelSlab));
|
||||||
|
|
||||||
// Determine size available for kernel page table heaps, requiring > 8 MB.
|
// Determine size available for kernel page table heaps, requiring > 8 MB.
|
||||||
@ -524,12 +524,12 @@ struct KernelCore::Impl {
|
|||||||
ASSERT(page_table_heap_size / 4_MiB > 2);
|
ASSERT(page_table_heap_size / 4_MiB > 2);
|
||||||
|
|
||||||
// Insert a physical region for the kernel page table heap region
|
// Insert a physical region for the kernel page table heap region
|
||||||
ASSERT(memory_layout.GetPhysicalMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetPhysicalMemoryRegionTree().Insert(
|
||||||
slab_end_phys_addr, page_table_heap_size, KMemoryRegionType_DramKernelPtHeap));
|
slab_end_phys_addr, page_table_heap_size, KMemoryRegionType_DramKernelPtHeap));
|
||||||
|
|
||||||
// All DRAM regions that we haven't tagged by this point will be mapped under the linear
|
// All DRAM regions that we haven't tagged by this point will be mapped under the linear
|
||||||
// mapping. Tag them.
|
// mapping. Tag them.
|
||||||
for (auto& region : memory_layout.GetPhysicalMemoryRegionTree()) {
|
for (auto& region : memory_layout->GetPhysicalMemoryRegionTree()) {
|
||||||
if (region.GetType() == KMemoryRegionType_Dram) {
|
if (region.GetType() == KMemoryRegionType_Dram) {
|
||||||
// Check that the region is valid.
|
// Check that the region is valid.
|
||||||
ASSERT(region.GetEndAddress() != 0);
|
ASSERT(region.GetEndAddress() != 0);
|
||||||
@ -541,7 +541,7 @@ struct KernelCore::Impl {
|
|||||||
|
|
||||||
// Get the linear region extents.
|
// Get the linear region extents.
|
||||||
const auto linear_extents =
|
const auto linear_extents =
|
||||||
memory_layout.GetPhysicalMemoryRegionTree().GetDerivedRegionExtents(
|
memory_layout->GetPhysicalMemoryRegionTree().GetDerivedRegionExtents(
|
||||||
KMemoryRegionAttr_LinearMapped);
|
KMemoryRegionAttr_LinearMapped);
|
||||||
ASSERT(linear_extents.GetEndAddress() != 0);
|
ASSERT(linear_extents.GetEndAddress() != 0);
|
||||||
|
|
||||||
@ -553,7 +553,7 @@ struct KernelCore::Impl {
|
|||||||
Common::AlignUp(linear_extents.GetEndAddress(), LinearRegionAlign) -
|
Common::AlignUp(linear_extents.GetEndAddress(), LinearRegionAlign) -
|
||||||
aligned_linear_phys_start;
|
aligned_linear_phys_start;
|
||||||
const VAddr linear_region_start =
|
const VAddr linear_region_start =
|
||||||
memory_layout.GetVirtualMemoryRegionTree().GetRandomAlignedRegionWithGuard(
|
memory_layout->GetVirtualMemoryRegionTree().GetRandomAlignedRegionWithGuard(
|
||||||
linear_region_size, LinearRegionAlign, KMemoryRegionType_None, LinearRegionAlign);
|
linear_region_size, LinearRegionAlign, KMemoryRegionType_None, LinearRegionAlign);
|
||||||
|
|
||||||
const u64 linear_region_phys_to_virt_diff = linear_region_start - aligned_linear_phys_start;
|
const u64 linear_region_phys_to_virt_diff = linear_region_start - aligned_linear_phys_start;
|
||||||
@ -562,7 +562,7 @@ struct KernelCore::Impl {
|
|||||||
{
|
{
|
||||||
PAddr cur_phys_addr = 0;
|
PAddr cur_phys_addr = 0;
|
||||||
u64 cur_size = 0;
|
u64 cur_size = 0;
|
||||||
for (auto& region : memory_layout.GetPhysicalMemoryRegionTree()) {
|
for (auto& region : memory_layout->GetPhysicalMemoryRegionTree()) {
|
||||||
if (!region.HasTypeAttribute(KMemoryRegionAttr_LinearMapped)) {
|
if (!region.HasTypeAttribute(KMemoryRegionAttr_LinearMapped)) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
@ -581,55 +581,49 @@ struct KernelCore::Impl {
|
|||||||
|
|
||||||
const VAddr region_virt_addr =
|
const VAddr region_virt_addr =
|
||||||
region.GetAddress() + linear_region_phys_to_virt_diff;
|
region.GetAddress() + linear_region_phys_to_virt_diff;
|
||||||
ASSERT(memory_layout.GetVirtualMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetVirtualMemoryRegionTree().Insert(
|
||||||
region_virt_addr, region.GetSize(),
|
region_virt_addr, region.GetSize(),
|
||||||
GetTypeForVirtualLinearMapping(region.GetType())));
|
GetTypeForVirtualLinearMapping(region.GetType())));
|
||||||
region.SetPairAddress(region_virt_addr);
|
region.SetPairAddress(region_virt_addr);
|
||||||
|
|
||||||
KMemoryRegion* virt_region =
|
KMemoryRegion* virt_region =
|
||||||
memory_layout.GetVirtualMemoryRegionTree().FindModifiable(region_virt_addr);
|
memory_layout->GetVirtualMemoryRegionTree().FindModifiable(region_virt_addr);
|
||||||
ASSERT(virt_region != nullptr);
|
ASSERT(virt_region != nullptr);
|
||||||
virt_region->SetPairAddress(region.GetAddress());
|
virt_region->SetPairAddress(region.GetAddress());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Insert regions for the initial page table region.
|
// Insert regions for the initial page table region.
|
||||||
ASSERT(memory_layout.GetPhysicalMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetPhysicalMemoryRegionTree().Insert(
|
||||||
resource_end_phys_addr, KernelPageTableHeapSize, KMemoryRegionType_DramKernelInitPt));
|
resource_end_phys_addr, KernelPageTableHeapSize, KMemoryRegionType_DramKernelInitPt));
|
||||||
ASSERT(memory_layout.GetVirtualMemoryRegionTree().Insert(
|
ASSERT(memory_layout->GetVirtualMemoryRegionTree().Insert(
|
||||||
resource_end_phys_addr + linear_region_phys_to_virt_diff, KernelPageTableHeapSize,
|
resource_end_phys_addr + linear_region_phys_to_virt_diff, KernelPageTableHeapSize,
|
||||||
KMemoryRegionType_VirtualDramKernelInitPt));
|
KMemoryRegionType_VirtualDramKernelInitPt));
|
||||||
|
|
||||||
// All linear-mapped DRAM regions that we haven't tagged by this point will be allocated to
|
// All linear-mapped DRAM regions that we haven't tagged by this point will be allocated to
|
||||||
// some pool partition. Tag them.
|
// some pool partition. Tag them.
|
||||||
for (auto& region : memory_layout.GetPhysicalMemoryRegionTree()) {
|
for (auto& region : memory_layout->GetPhysicalMemoryRegionTree()) {
|
||||||
if (region.GetType() == (KMemoryRegionType_Dram | KMemoryRegionAttr_LinearMapped)) {
|
if (region.GetType() == (KMemoryRegionType_Dram | KMemoryRegionAttr_LinearMapped)) {
|
||||||
region.SetType(KMemoryRegionType_DramPoolPartition);
|
region.SetType(KMemoryRegionType_DramPoolPartition);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Setup all other memory regions needed to arrange the pool partitions.
|
// Setup all other memory regions needed to arrange the pool partitions.
|
||||||
Init::SetupPoolPartitionMemoryRegions(memory_layout);
|
Init::SetupPoolPartitionMemoryRegions(*memory_layout);
|
||||||
|
|
||||||
// Cache all linear regions in their own trees for faster access, later.
|
// Cache all linear regions in their own trees for faster access, later.
|
||||||
memory_layout.InitializeLinearMemoryRegionTrees(aligned_linear_phys_start,
|
memory_layout->InitializeLinearMemoryRegionTrees(aligned_linear_phys_start,
|
||||||
linear_region_start);
|
linear_region_start);
|
||||||
}
|
}
|
||||||
|
|
||||||
void InitializeMemoryLayout(const KMemoryLayout& memory_layout) {
|
void InitializeMemoryLayout() {
|
||||||
const auto system_pool = memory_layout.GetKernelSystemPoolRegionPhysicalExtents();
|
const auto system_pool = memory_layout->GetKernelSystemPoolRegionPhysicalExtents();
|
||||||
const auto applet_pool = memory_layout.GetKernelAppletPoolRegionPhysicalExtents();
|
|
||||||
const auto application_pool = memory_layout.GetKernelApplicationPoolRegionPhysicalExtents();
|
|
||||||
|
|
||||||
// Initialize memory managers
|
// Initialize the memory manager.
|
||||||
memory_manager = std::make_unique<KMemoryManager>(system);
|
memory_manager = std::make_unique<KMemoryManager>(system);
|
||||||
memory_manager->InitializeManager(KMemoryManager::Pool::Application,
|
const auto& management_region = memory_layout->GetPoolManagementRegion();
|
||||||
application_pool.GetAddress(),
|
ASSERT(management_region.GetEndAddress() != 0);
|
||||||
application_pool.GetEndAddress());
|
memory_manager->Initialize(management_region.GetAddress(), management_region.GetSize());
|
||||||
memory_manager->InitializeManager(KMemoryManager::Pool::Applet, applet_pool.GetAddress(),
|
|
||||||
applet_pool.GetEndAddress());
|
|
||||||
memory_manager->InitializeManager(KMemoryManager::Pool::System, system_pool.GetAddress(),
|
|
||||||
system_pool.GetEndAddress());
|
|
||||||
|
|
||||||
// Setup memory regions for emulated processes
|
// Setup memory regions for emulated processes
|
||||||
// TODO(bunnei): These should not be hardcoded regions initialized within the kernel
|
// TODO(bunnei): These should not be hardcoded regions initialized within the kernel
|
||||||
@ -770,6 +764,9 @@ struct KernelCore::Impl {
|
|||||||
Kernel::KSharedMemory* irs_shared_mem{};
|
Kernel::KSharedMemory* irs_shared_mem{};
|
||||||
Kernel::KSharedMemory* time_shared_mem{};
|
Kernel::KSharedMemory* time_shared_mem{};
|
||||||
|
|
||||||
|
// Memory layout
|
||||||
|
std::unique_ptr<KMemoryLayout> memory_layout;
|
||||||
|
|
||||||
// Threads used for services
|
// Threads used for services
|
||||||
std::unordered_set<std::shared_ptr<Kernel::ServiceThread>> service_threads;
|
std::unordered_set<std::shared_ptr<Kernel::ServiceThread>> service_threads;
|
||||||
Common::ThreadWorker service_threads_manager;
|
Common::ThreadWorker service_threads_manager;
|
||||||
@ -1135,6 +1132,10 @@ const KWorkerTaskManager& KernelCore::WorkerTaskManager() const {
|
|||||||
return impl->worker_task_manager;
|
return impl->worker_task_manager;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const KMemoryLayout& KernelCore::MemoryLayout() const {
|
||||||
|
return *impl->memory_layout;
|
||||||
|
}
|
||||||
|
|
||||||
bool KernelCore::IsPhantomModeForSingleCore() const {
|
bool KernelCore::IsPhantomModeForSingleCore() const {
|
||||||
return impl->IsPhantomModeForSingleCore();
|
return impl->IsPhantomModeForSingleCore();
|
||||||
}
|
}
|
||||||
|
@ -41,6 +41,7 @@ class KClientSession;
|
|||||||
class KEvent;
|
class KEvent;
|
||||||
class KHandleTable;
|
class KHandleTable;
|
||||||
class KLinkedListNode;
|
class KLinkedListNode;
|
||||||
|
class KMemoryLayout;
|
||||||
class KMemoryManager;
|
class KMemoryManager;
|
||||||
class KPort;
|
class KPort;
|
||||||
class KProcess;
|
class KProcess;
|
||||||
@ -350,6 +351,9 @@ public:
|
|||||||
/// Gets the current worker task manager, used for dispatching KThread/KProcess tasks.
|
/// Gets the current worker task manager, used for dispatching KThread/KProcess tasks.
|
||||||
const KWorkerTaskManager& WorkerTaskManager() const;
|
const KWorkerTaskManager& WorkerTaskManager() const;
|
||||||
|
|
||||||
|
/// Gets the memory layout.
|
||||||
|
const KMemoryLayout& MemoryLayout() const;
|
||||||
|
|
||||||
private:
|
private:
|
||||||
friend class KProcess;
|
friend class KProcess;
|
||||||
friend class KThread;
|
friend class KThread;
|
||||||
|
Loading…
Reference in New Issue
Block a user