hle: kernel: memory: PageHeap: Migrate to KPageBitmap class.
This commit is contained in:
parent
a02566136c
commit
6a19086001
|
@ -21,7 +21,7 @@ std::size_t MemoryManager::Impl::Initialize(Pool new_pool, u64 start_address, u6
|
||||||
const auto ref_count_size{(size / PageSize) * sizeof(u16)};
|
const auto ref_count_size{(size / PageSize) * sizeof(u16)};
|
||||||
const auto optimize_map_size{(Common::AlignUp((size / PageSize), 64) / 64) * sizeof(u64)};
|
const auto optimize_map_size{(Common::AlignUp((size / PageSize), 64) / 64) * sizeof(u64)};
|
||||||
const auto manager_size{Common::AlignUp(optimize_map_size + ref_count_size, PageSize)};
|
const auto manager_size{Common::AlignUp(optimize_map_size + ref_count_size, PageSize)};
|
||||||
const auto page_heap_size{PageHeap::CalculateMetadataOverheadSize(size)};
|
const auto page_heap_size{PageHeap::CalculateManagementOverheadSize(size)};
|
||||||
const auto total_metadata_size{manager_size + page_heap_size};
|
const auto total_metadata_size{manager_size + page_heap_size};
|
||||||
ASSERT(manager_size <= total_metadata_size);
|
ASSERT(manager_size <= total_metadata_size);
|
||||||
ASSERT(Common::IsAligned(total_metadata_size, PageSize));
|
ASSERT(Common::IsAligned(total_metadata_size, PageSize));
|
||||||
|
@ -63,7 +63,7 @@ VAddr MemoryManager::AllocateContinuous(std::size_t num_pages, std::size_t align
|
||||||
// Loop, trying to iterate from each block
|
// Loop, trying to iterate from each block
|
||||||
// TODO (bunnei): Support multiple managers
|
// TODO (bunnei): Support multiple managers
|
||||||
Impl& chosen_manager{managers[pool_index]};
|
Impl& chosen_manager{managers[pool_index]};
|
||||||
VAddr allocated_block{chosen_manager.AllocateBlock(heap_index)};
|
VAddr allocated_block{chosen_manager.AllocateBlock(heap_index, false)};
|
||||||
|
|
||||||
// If we failed to allocate, quit now
|
// If we failed to allocate, quit now
|
||||||
if (!allocated_block) {
|
if (!allocated_block) {
|
||||||
|
@ -116,7 +116,7 @@ ResultCode MemoryManager::Allocate(PageLinkedList& page_list, std::size_t num_pa
|
||||||
|
|
||||||
while (num_pages >= pages_per_alloc) {
|
while (num_pages >= pages_per_alloc) {
|
||||||
// Allocate a block
|
// Allocate a block
|
||||||
VAddr allocated_block{chosen_manager.AllocateBlock(index)};
|
VAddr allocated_block{chosen_manager.AllocateBlock(index, false)};
|
||||||
if (!allocated_block) {
|
if (!allocated_block) {
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
|
@ -67,8 +67,8 @@ private:
|
||||||
|
|
||||||
std::size_t Initialize(Pool new_pool, u64 start_address, u64 end_address);
|
std::size_t Initialize(Pool new_pool, u64 start_address, u64 end_address);
|
||||||
|
|
||||||
VAddr AllocateBlock(s32 index) {
|
VAddr AllocateBlock(s32 index, bool random) {
|
||||||
return heap.AllocateBlock(index);
|
return heap.AllocateBlock(index, random);
|
||||||
}
|
}
|
||||||
|
|
||||||
void Free(VAddr addr, std::size_t num_pages) {
|
void Free(VAddr addr, std::size_t num_pages) {
|
||||||
|
|
|
@ -32,11 +32,11 @@ void PageHeap::Initialize(VAddr address, std::size_t size, std::size_t metadata_
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
VAddr PageHeap::AllocateBlock(s32 index) {
|
VAddr PageHeap::AllocateBlock(s32 index, bool random) {
|
||||||
const std::size_t needed_size{blocks[index].GetSize()};
|
const std::size_t needed_size{blocks[index].GetSize()};
|
||||||
|
|
||||||
for (s32 i{index}; i < static_cast<s32>(MemoryBlockPageShifts.size()); i++) {
|
for (s32 i{index}; i < static_cast<s32>(MemoryBlockPageShifts.size()); i++) {
|
||||||
if (const VAddr addr{blocks[i].PopBlock()}; addr) {
|
if (const VAddr addr{blocks[i].PopBlock(random)}; addr) {
|
||||||
if (const std::size_t allocated_size{blocks[i].GetSize()};
|
if (const std::size_t allocated_size{blocks[i].GetSize()};
|
||||||
allocated_size > needed_size) {
|
allocated_size > needed_size) {
|
||||||
Free(addr + needed_size, (allocated_size - needed_size) / PageSize);
|
Free(addr + needed_size, (allocated_size - needed_size) / PageSize);
|
||||||
|
@ -104,13 +104,13 @@ void PageHeap::Free(VAddr addr, std::size_t num_pages) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::size_t PageHeap::CalculateMetadataOverheadSize(std::size_t region_size) {
|
std::size_t PageHeap::CalculateManagementOverheadSize(std::size_t region_size) {
|
||||||
std::size_t overhead_size = 0;
|
std::size_t overhead_size = 0;
|
||||||
for (std::size_t i = 0; i < MemoryBlockPageShifts.size(); i++) {
|
for (std::size_t i = 0; i < MemoryBlockPageShifts.size(); i++) {
|
||||||
const std::size_t cur_block_shift{MemoryBlockPageShifts[i]};
|
const std::size_t cur_block_shift{MemoryBlockPageShifts[i]};
|
||||||
const std::size_t next_block_shift{
|
const std::size_t next_block_shift{
|
||||||
(i != MemoryBlockPageShifts.size() - 1) ? MemoryBlockPageShifts[i + 1] : 0};
|
(i != MemoryBlockPageShifts.size() - 1) ? MemoryBlockPageShifts[i + 1] : 0};
|
||||||
overhead_size += PageHeap::Block::CalculateMetadataOverheadSize(
|
overhead_size += PageHeap::Block::CalculateManagementOverheadSize(
|
||||||
region_size, cur_block_shift, next_block_shift);
|
region_size, cur_block_shift, next_block_shift);
|
||||||
}
|
}
|
||||||
return Common::AlignUp(overhead_size, PageSize);
|
return Common::AlignUp(overhead_size, PageSize);
|
||||||
|
|
|
@ -15,6 +15,7 @@
|
||||||
#include "common/assert.h"
|
#include "common/assert.h"
|
||||||
#include "common/common_funcs.h"
|
#include "common/common_funcs.h"
|
||||||
#include "common/common_types.h"
|
#include "common/common_types.h"
|
||||||
|
#include "core/hle/kernel/k_page_bitmap.h"
|
||||||
#include "core/hle/kernel/memory/memory_types.h"
|
#include "core/hle/kernel/memory/memory_types.h"
|
||||||
|
|
||||||
namespace Kernel::Memory {
|
namespace Kernel::Memory {
|
||||||
|
@ -57,189 +58,14 @@ private:
|
||||||
|
|
||||||
class Block final : NonCopyable {
|
class Block final : NonCopyable {
|
||||||
private:
|
private:
|
||||||
class Bitmap final : NonCopyable {
|
KPageBitmap bitmap;
|
||||||
public:
|
|
||||||
static constexpr std::size_t MaxDepth{4};
|
|
||||||
|
|
||||||
private:
|
|
||||||
std::array<u64*, MaxDepth> bit_storages{};
|
|
||||||
std::size_t num_bits{};
|
|
||||||
std::size_t used_depths{};
|
|
||||||
|
|
||||||
public:
|
|
||||||
constexpr Bitmap() = default;
|
|
||||||
|
|
||||||
constexpr std::size_t GetNumBits() const {
|
|
||||||
return num_bits;
|
|
||||||
}
|
|
||||||
constexpr s32 GetHighestDepthIndex() const {
|
|
||||||
return static_cast<s32>(used_depths) - 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
constexpr u64* Initialize(u64* storage, std::size_t size) {
|
|
||||||
//* Initially, everything is un-set
|
|
||||||
num_bits = 0;
|
|
||||||
|
|
||||||
// Calculate the needed bitmap depth
|
|
||||||
used_depths = static_cast<std::size_t>(GetRequiredDepth(size));
|
|
||||||
ASSERT(used_depths <= MaxDepth);
|
|
||||||
|
|
||||||
// Set the bitmap pointers
|
|
||||||
for (s32 depth{GetHighestDepthIndex()}; depth >= 0; depth--) {
|
|
||||||
bit_storages[depth] = storage;
|
|
||||||
size = Common::AlignUp(size, 64) / 64;
|
|
||||||
storage += size;
|
|
||||||
}
|
|
||||||
|
|
||||||
return storage;
|
|
||||||
}
|
|
||||||
|
|
||||||
s64 FindFreeBlock() const {
|
|
||||||
uintptr_t offset{};
|
|
||||||
s32 depth{};
|
|
||||||
|
|
||||||
do {
|
|
||||||
const u64 v{bit_storages[depth][offset]};
|
|
||||||
if (v == 0) {
|
|
||||||
// Non-zero depth indicates that a previous level had a free block
|
|
||||||
ASSERT(depth == 0);
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
offset = offset * 64 + static_cast<u32>(std::countr_zero(v));
|
|
||||||
++depth;
|
|
||||||
} while (depth < static_cast<s32>(used_depths));
|
|
||||||
|
|
||||||
return static_cast<s64>(offset);
|
|
||||||
}
|
|
||||||
|
|
||||||
constexpr void SetBit(std::size_t offset) {
|
|
||||||
SetBit(GetHighestDepthIndex(), offset);
|
|
||||||
num_bits++;
|
|
||||||
}
|
|
||||||
|
|
||||||
constexpr void ClearBit(std::size_t offset) {
|
|
||||||
ClearBit(GetHighestDepthIndex(), offset);
|
|
||||||
num_bits--;
|
|
||||||
}
|
|
||||||
|
|
||||||
constexpr bool ClearRange(std::size_t offset, std::size_t count) {
|
|
||||||
const s32 depth{GetHighestDepthIndex()};
|
|
||||||
const auto bit_ind{offset / 64};
|
|
||||||
u64* bits{bit_storages[depth]};
|
|
||||||
if (count < 64) {
|
|
||||||
const auto shift{offset % 64};
|
|
||||||
ASSERT(shift + count <= 64);
|
|
||||||
// Check that all the bits are set
|
|
||||||
const u64 mask{((1ULL << count) - 1) << shift};
|
|
||||||
u64 v{bits[bit_ind]};
|
|
||||||
if ((v & mask) != mask) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Clear the bits
|
|
||||||
v &= ~mask;
|
|
||||||
bits[bit_ind] = v;
|
|
||||||
if (v == 0) {
|
|
||||||
ClearBit(depth - 1, bit_ind);
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
ASSERT(offset % 64 == 0);
|
|
||||||
ASSERT(count % 64 == 0);
|
|
||||||
// Check that all the bits are set
|
|
||||||
std::size_t remaining{count};
|
|
||||||
std::size_t i = 0;
|
|
||||||
do {
|
|
||||||
if (bits[bit_ind + i++] != ~u64(0)) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
remaining -= 64;
|
|
||||||
} while (remaining > 0);
|
|
||||||
|
|
||||||
// Clear the bits
|
|
||||||
remaining = count;
|
|
||||||
i = 0;
|
|
||||||
do {
|
|
||||||
bits[bit_ind + i] = 0;
|
|
||||||
ClearBit(depth - 1, bit_ind + i);
|
|
||||||
i++;
|
|
||||||
remaining -= 64;
|
|
||||||
} while (remaining > 0);
|
|
||||||
}
|
|
||||||
|
|
||||||
num_bits -= count;
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
private:
|
|
||||||
constexpr void SetBit(s32 depth, std::size_t offset) {
|
|
||||||
while (depth >= 0) {
|
|
||||||
const auto ind{offset / 64};
|
|
||||||
const auto which{offset % 64};
|
|
||||||
const u64 mask{1ULL << which};
|
|
||||||
|
|
||||||
u64* bit{std::addressof(bit_storages[depth][ind])};
|
|
||||||
const u64 v{*bit};
|
|
||||||
ASSERT((v & mask) == 0);
|
|
||||||
*bit = v | mask;
|
|
||||||
if (v) {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
offset = ind;
|
|
||||||
depth--;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
constexpr void ClearBit(s32 depth, std::size_t offset) {
|
|
||||||
while (depth >= 0) {
|
|
||||||
const auto ind{offset / 64};
|
|
||||||
const auto which{offset % 64};
|
|
||||||
const u64 mask{1ULL << which};
|
|
||||||
|
|
||||||
u64* bit{std::addressof(bit_storages[depth][ind])};
|
|
||||||
u64 v{*bit};
|
|
||||||
ASSERT((v & mask) != 0);
|
|
||||||
v &= ~mask;
|
|
||||||
*bit = v;
|
|
||||||
if (v) {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
offset = ind;
|
|
||||||
depth--;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private:
|
|
||||||
static constexpr s32 GetRequiredDepth(std::size_t region_size) {
|
|
||||||
s32 depth = 0;
|
|
||||||
while (true) {
|
|
||||||
region_size /= 64;
|
|
||||||
depth++;
|
|
||||||
if (region_size == 0) {
|
|
||||||
return depth;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public:
|
|
||||||
static constexpr std::size_t CalculateMetadataOverheadSize(std::size_t region_size) {
|
|
||||||
std::size_t overhead_bits = 0;
|
|
||||||
for (s32 depth{GetRequiredDepth(region_size) - 1}; depth >= 0; depth--) {
|
|
||||||
region_size = Common::AlignUp(region_size, 64) / 64;
|
|
||||||
overhead_bits += region_size;
|
|
||||||
}
|
|
||||||
return overhead_bits * sizeof(u64);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
private:
|
|
||||||
Bitmap bitmap;
|
|
||||||
VAddr heap_address{};
|
VAddr heap_address{};
|
||||||
uintptr_t end_offset{};
|
uintptr_t end_offset{};
|
||||||
std::size_t block_shift{};
|
std::size_t block_shift{};
|
||||||
std::size_t next_block_shift{};
|
std::size_t next_block_shift{};
|
||||||
|
|
||||||
public:
|
public:
|
||||||
constexpr Block() = default;
|
Block() = default;
|
||||||
|
|
||||||
constexpr std::size_t GetShift() const {
|
constexpr std::size_t GetShift() const {
|
||||||
return block_shift;
|
return block_shift;
|
||||||
|
@ -260,8 +86,8 @@ private:
|
||||||
return GetNumFreeBlocks() * GetNumPages();
|
return GetNumFreeBlocks() * GetNumPages();
|
||||||
}
|
}
|
||||||
|
|
||||||
constexpr u64* Initialize(VAddr addr, std::size_t size, std::size_t bs, std::size_t nbs,
|
u64* Initialize(VAddr addr, std::size_t size, std::size_t bs, std::size_t nbs,
|
||||||
u64* bit_storage) {
|
u64* bit_storage) {
|
||||||
// Set shifts
|
// Set shifts
|
||||||
block_shift = bs;
|
block_shift = bs;
|
||||||
next_block_shift = nbs;
|
next_block_shift = nbs;
|
||||||
|
@ -278,7 +104,7 @@ private:
|
||||||
return bitmap.Initialize(bit_storage, end_offset);
|
return bitmap.Initialize(bit_storage, end_offset);
|
||||||
}
|
}
|
||||||
|
|
||||||
constexpr VAddr PushBlock(VAddr address) {
|
VAddr PushBlock(VAddr address) {
|
||||||
// Set the bit for the free block
|
// Set the bit for the free block
|
||||||
std::size_t offset{(address - heap_address) >> GetShift()};
|
std::size_t offset{(address - heap_address) >> GetShift()};
|
||||||
bitmap.SetBit(offset);
|
bitmap.SetBit(offset);
|
||||||
|
@ -296,9 +122,9 @@ private:
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
VAddr PopBlock() {
|
VAddr PopBlock(bool random) {
|
||||||
// Find a free block
|
// Find a free block
|
||||||
const s64 soffset{bitmap.FindFreeBlock()};
|
const s64 soffset{bitmap.FindFreeBlock(random)};
|
||||||
if (soffset < 0) {
|
if (soffset < 0) {
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@ -310,13 +136,13 @@ private:
|
||||||
}
|
}
|
||||||
|
|
||||||
public:
|
public:
|
||||||
static constexpr std::size_t CalculateMetadataOverheadSize(std::size_t region_size,
|
static constexpr std::size_t CalculateManagementOverheadSize(std::size_t region_size,
|
||||||
std::size_t cur_block_shift,
|
std::size_t cur_block_shift,
|
||||||
std::size_t next_block_shift) {
|
std::size_t next_block_shift) {
|
||||||
const auto cur_block_size{(1ULL << cur_block_shift)};
|
const auto cur_block_size{(1ULL << cur_block_shift)};
|
||||||
const auto next_block_size{(1ULL << next_block_shift)};
|
const auto next_block_size{(1ULL << next_block_shift)};
|
||||||
const auto align{(next_block_shift != 0) ? next_block_size : cur_block_size};
|
const auto align{(next_block_shift != 0) ? next_block_size : cur_block_size};
|
||||||
return Bitmap::CalculateMetadataOverheadSize(
|
return KPageBitmap::CalculateManagementOverheadSize(
|
||||||
(align * 2 + Common::AlignUp(region_size, align)) / cur_block_size);
|
(align * 2 + Common::AlignUp(region_size, align)) / cur_block_size);
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
@ -338,14 +164,14 @@ public:
|
||||||
}
|
}
|
||||||
|
|
||||||
void Initialize(VAddr heap_address, std::size_t heap_size, std::size_t metadata_size);
|
void Initialize(VAddr heap_address, std::size_t heap_size, std::size_t metadata_size);
|
||||||
VAddr AllocateBlock(s32 index);
|
VAddr AllocateBlock(s32 index, bool random);
|
||||||
void Free(VAddr addr, std::size_t num_pages);
|
void Free(VAddr addr, std::size_t num_pages);
|
||||||
|
|
||||||
void UpdateUsedSize() {
|
void UpdateUsedSize() {
|
||||||
used_size = heap_size - (GetNumFreePages() * PageSize);
|
used_size = heap_size - (GetNumFreePages() * PageSize);
|
||||||
}
|
}
|
||||||
|
|
||||||
static std::size_t CalculateMetadataOverheadSize(std::size_t region_size);
|
static std::size_t CalculateManagementOverheadSize(std::size_t region_size);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
constexpr std::size_t GetNumFreePages() const {
|
constexpr std::size_t GetNumFreePages() const {
|
||||||
|
|
Loading…
Reference in New Issue