Merge pull request #905 from lioncash/vma

kernel/vm_manager: Minor changes
This commit is contained in:
bunnei 2018-08-02 23:18:03 -04:00 committed by GitHub
commit 61ed68f3d0
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
3 changed files with 23 additions and 23 deletions

View file

@ -267,7 +267,7 @@ static ResultCode GetInfo(u64* result, u64 info_id, u64 handle, u64 info_sub_id)
LOG_TRACE(Kernel_SVC, "called info_id=0x{:X}, info_sub_id=0x{:X}, handle=0x{:08X}", info_id, LOG_TRACE(Kernel_SVC, "called info_id=0x{:X}, info_sub_id=0x{:X}, handle=0x{:08X}", info_id,
info_sub_id, handle); info_sub_id, handle);
auto& vm_manager = Core::CurrentProcess()->vm_manager; const auto& vm_manager = Core::CurrentProcess()->vm_manager;
switch (static_cast<GetInfoType>(info_id)) { switch (static_cast<GetInfoType>(info_id)) {
case GetInfoType::AllowedCpuIdBitmask: case GetInfoType::AllowedCpuIdBitmask:

View file

@ -2,6 +2,7 @@
// Licensed under GPLv2 or any later version // Licensed under GPLv2 or any later version
// Refer to the license.txt file included. // Refer to the license.txt file included.
#include <algorithm>
#include <iterator> #include <iterator>
#include <utility> #include <utility>
#include "common/assert.h" #include "common/assert.h"
@ -175,9 +176,9 @@ VMManager::VMAIter VMManager::Unmap(VMAIter vma_handle) {
ResultCode VMManager::UnmapRange(VAddr target, u64 size) { ResultCode VMManager::UnmapRange(VAddr target, u64 size) {
CASCADE_RESULT(VMAIter vma, CarveVMARange(target, size)); CASCADE_RESULT(VMAIter vma, CarveVMARange(target, size));
VAddr target_end = target + size; const VAddr target_end = target + size;
VMAIter end = vma_map.end(); const VMAIter end = vma_map.end();
// The comparison against the end of the range must be done using addresses since VMAs can be // The comparison against the end of the range must be done using addresses since VMAs can be
// merged during this process, causing invalidation of the iterators. // merged during this process, causing invalidation of the iterators.
while (vma != end && vma->second.base < target_end) { while (vma != end && vma->second.base < target_end) {
@ -207,9 +208,9 @@ VMManager::VMAHandle VMManager::Reprotect(VMAHandle vma_handle, VMAPermission ne
ResultCode VMManager::ReprotectRange(VAddr target, u64 size, VMAPermission new_perms) { ResultCode VMManager::ReprotectRange(VAddr target, u64 size, VMAPermission new_perms) {
CASCADE_RESULT(VMAIter vma, CarveVMARange(target, size)); CASCADE_RESULT(VMAIter vma, CarveVMARange(target, size));
VAddr target_end = target + size; const VAddr target_end = target + size;
VMAIter end = vma_map.end(); const VMAIter end = vma_map.end();
// The comparison against the end of the range must be done using addresses since VMAs can be // The comparison against the end of the range must be done using addresses since VMAs can be
// merged during this process, causing invalidation of the iterators. // merged during this process, causing invalidation of the iterators.
while (vma != end && vma->second.base < target_end) { while (vma != end && vma->second.base < target_end) {
@ -258,14 +259,14 @@ ResultVal<VMManager::VMAIter> VMManager::CarveVMA(VAddr base, u64 size) {
return ERR_INVALID_ADDRESS; return ERR_INVALID_ADDRESS;
} }
VirtualMemoryArea& vma = vma_handle->second; const VirtualMemoryArea& vma = vma_handle->second;
if (vma.type != VMAType::Free) { if (vma.type != VMAType::Free) {
// Region is already allocated // Region is already allocated
return ERR_INVALID_ADDRESS_STATE; return ERR_INVALID_ADDRESS_STATE;
} }
u64 start_in_vma = base - vma.base; const VAddr start_in_vma = base - vma.base;
u64 end_in_vma = start_in_vma + size; const VAddr end_in_vma = start_in_vma + size;
if (end_in_vma > vma.size) { if (end_in_vma > vma.size) {
// Requested allocation doesn't fit inside VMA // Requested allocation doesn't fit inside VMA
@ -288,18 +289,17 @@ ResultVal<VMManager::VMAIter> VMManager::CarveVMARange(VAddr target, u64 size) {
ASSERT_MSG((size & Memory::PAGE_MASK) == 0, "non-page aligned size: 0x{:016X}", size); ASSERT_MSG((size & Memory::PAGE_MASK) == 0, "non-page aligned size: 0x{:016X}", size);
ASSERT_MSG((target & Memory::PAGE_MASK) == 0, "non-page aligned base: 0x{:016X}", target); ASSERT_MSG((target & Memory::PAGE_MASK) == 0, "non-page aligned base: 0x{:016X}", target);
VAddr target_end = target + size; const VAddr target_end = target + size;
ASSERT(target_end >= target); ASSERT(target_end >= target);
ASSERT(target_end <= MAX_ADDRESS); ASSERT(target_end <= MAX_ADDRESS);
ASSERT(size > 0); ASSERT(size > 0);
VMAIter begin_vma = StripIterConstness(FindVMA(target)); VMAIter begin_vma = StripIterConstness(FindVMA(target));
VMAIter i_end = vma_map.lower_bound(target_end); const VMAIter i_end = vma_map.lower_bound(target_end);
for (auto i = begin_vma; i != i_end; ++i) { if (std::any_of(begin_vma, i_end,
if (i->second.type == VMAType::Free) { [](const auto& entry) { return entry.second.type == VMAType::Free; })) {
return ERR_INVALID_ADDRESS_STATE; return ERR_INVALID_ADDRESS_STATE;
} }
}
if (target != begin_vma->second.base) { if (target != begin_vma->second.base) {
begin_vma = SplitVMA(begin_vma, target - begin_vma->second.base); begin_vma = SplitVMA(begin_vma, target - begin_vma->second.base);
@ -346,7 +346,7 @@ VMManager::VMAIter VMManager::SplitVMA(VMAIter vma_handle, u64 offset_in_vma) {
} }
VMManager::VMAIter VMManager::MergeAdjacent(VMAIter iter) { VMManager::VMAIter VMManager::MergeAdjacent(VMAIter iter) {
VMAIter next_vma = std::next(iter); const VMAIter next_vma = std::next(iter);
if (next_vma != vma_map.end() && iter->second.CanBeMergedWith(next_vma->second)) { if (next_vma != vma_map.end() && iter->second.CanBeMergedWith(next_vma->second)) {
iter->second.size += next_vma->second.size; iter->second.size += next_vma->second.size;
vma_map.erase(next_vma); vma_map.erase(next_vma);
@ -382,22 +382,22 @@ void VMManager::UpdatePageTableForVMA(const VirtualMemoryArea& vma) {
} }
} }
u64 VMManager::GetTotalMemoryUsage() { u64 VMManager::GetTotalMemoryUsage() const {
LOG_WARNING(Kernel, "(STUBBED) called"); LOG_WARNING(Kernel, "(STUBBED) called");
return 0xF8000000; return 0xF8000000;
} }
u64 VMManager::GetTotalHeapUsage() { u64 VMManager::GetTotalHeapUsage() const {
LOG_WARNING(Kernel, "(STUBBED) called"); LOG_WARNING(Kernel, "(STUBBED) called");
return 0x0; return 0x0;
} }
VAddr VMManager::GetAddressSpaceBaseAddr() { VAddr VMManager::GetAddressSpaceBaseAddr() const {
LOG_WARNING(Kernel, "(STUBBED) called"); LOG_WARNING(Kernel, "(STUBBED) called");
return 0x8000000; return 0x8000000;
} }
u64 VMManager::GetAddressSpaceSize() { u64 VMManager::GetAddressSpaceSize() const {
LOG_WARNING(Kernel, "(STUBBED) called"); LOG_WARNING(Kernel, "(STUBBED) called");
return MAX_ADDRESS; return MAX_ADDRESS;
} }

View file

@ -190,16 +190,16 @@ public:
void LogLayout() const; void LogLayout() const;
/// Gets the total memory usage, used by svcGetInfo /// Gets the total memory usage, used by svcGetInfo
u64 GetTotalMemoryUsage(); u64 GetTotalMemoryUsage() const;
/// Gets the total heap usage, used by svcGetInfo /// Gets the total heap usage, used by svcGetInfo
u64 GetTotalHeapUsage(); u64 GetTotalHeapUsage() const;
/// Gets the total address space base address, used by svcGetInfo /// Gets the total address space base address, used by svcGetInfo
VAddr GetAddressSpaceBaseAddr(); VAddr GetAddressSpaceBaseAddr() const;
/// Gets the total address space address size, used by svcGetInfo /// Gets the total address space address size, used by svcGetInfo
u64 GetAddressSpaceSize(); u64 GetAddressSpaceSize() const;
/// Each VMManager has its own page table, which is set as the main one when the owning process /// Each VMManager has its own page table, which is set as the main one when the owning process
/// is scheduled. /// is scheduled.