forked from suyu/suyu
kernel: conver KScopedLock, KScopedResourceReservation, KSessionRequest, KSharedMemory, KSpinLock
This commit is contained in:
parent
7322c99e5f
commit
d1b53c8d82
9 changed files with 99 additions and 97 deletions
|
@ -18,15 +18,15 @@ std::is_reference_v<T>&& requires(T& t) {
|
|||
|
||||
template <typename T>
|
||||
requires KLockable<T>
|
||||
class [[nodiscard]] KScopedLock {
|
||||
class KScopedLock {
|
||||
public:
|
||||
explicit KScopedLock(T* l) : lock_ptr(l) {
|
||||
this->lock_ptr->Lock();
|
||||
explicit KScopedLock(T* l) : m_lock(*l) {}
|
||||
explicit KScopedLock(T& l) : m_lock(l) {
|
||||
m_lock.Lock();
|
||||
}
|
||||
explicit KScopedLock(T& l) : KScopedLock(std::addressof(l)) {}
|
||||
|
||||
~KScopedLock() {
|
||||
this->lock_ptr->Unlock();
|
||||
m_lock.Unlock();
|
||||
}
|
||||
|
||||
KScopedLock(const KScopedLock&) = delete;
|
||||
|
@ -36,7 +36,7 @@ public:
|
|||
KScopedLock& operator=(KScopedLock&&) = delete;
|
||||
|
||||
private:
|
||||
T* lock_ptr;
|
||||
T& m_lock;
|
||||
};
|
||||
|
||||
} // namespace Kernel
|
||||
|
|
|
@ -12,20 +12,20 @@ namespace Kernel {
|
|||
class KScopedResourceReservation {
|
||||
public:
|
||||
explicit KScopedResourceReservation(KResourceLimit* l, LimitableResource r, s64 v, s64 timeout)
|
||||
: resource_limit(std::move(l)), value(v), resource(r) {
|
||||
if (resource_limit && value) {
|
||||
success = resource_limit->Reserve(resource, value, timeout);
|
||||
: m_limit(l), m_value(v), m_resource(r) {
|
||||
if (m_limit && m_value) {
|
||||
m_succeeded = m_limit->Reserve(m_resource, m_value, timeout);
|
||||
} else {
|
||||
success = true;
|
||||
m_succeeded = true;
|
||||
}
|
||||
}
|
||||
|
||||
explicit KScopedResourceReservation(KResourceLimit* l, LimitableResource r, s64 v = 1)
|
||||
: resource_limit(std::move(l)), value(v), resource(r) {
|
||||
if (resource_limit && value) {
|
||||
success = resource_limit->Reserve(resource, value);
|
||||
: m_limit(l), m_value(v), m_resource(r) {
|
||||
if (m_limit && m_value) {
|
||||
m_succeeded = m_limit->Reserve(m_resource, m_value);
|
||||
} else {
|
||||
success = true;
|
||||
m_succeeded = true;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -36,26 +36,26 @@ public:
|
|||
: KScopedResourceReservation(p->GetResourceLimit(), r, v) {}
|
||||
|
||||
~KScopedResourceReservation() noexcept {
|
||||
if (resource_limit && value && success) {
|
||||
// resource was not committed, release the reservation.
|
||||
resource_limit->Release(resource, value);
|
||||
if (m_limit && m_value && m_succeeded) {
|
||||
// Resource was not committed, release the reservation.
|
||||
m_limit->Release(m_resource, m_value);
|
||||
}
|
||||
}
|
||||
|
||||
/// Commit the resource reservation, destruction of this object does not release the resource
|
||||
void Commit() {
|
||||
resource_limit = nullptr;
|
||||
m_limit = nullptr;
|
||||
}
|
||||
|
||||
[[nodiscard]] bool Succeeded() const {
|
||||
return success;
|
||||
bool Succeeded() const {
|
||||
return m_succeeded;
|
||||
}
|
||||
|
||||
private:
|
||||
KResourceLimit* resource_limit{};
|
||||
s64 value;
|
||||
LimitableResource resource;
|
||||
bool success;
|
||||
KResourceLimit* m_limit{};
|
||||
s64 m_value{};
|
||||
LimitableResource m_resource{};
|
||||
bool m_succeeded{};
|
||||
};
|
||||
|
||||
} // namespace Kernel
|
||||
|
|
|
@ -14,46 +14,46 @@ Result KSessionRequest::SessionMappings::PushMap(VAddr client, VAddr server, siz
|
|||
// Get the mapping.
|
||||
Mapping* mapping;
|
||||
if (index < NumStaticMappings) {
|
||||
mapping = &m_static_mappings[index];
|
||||
mapping = std::addressof(m_static_mappings[index]);
|
||||
} else {
|
||||
// Allocate a page for the extra mappings.
|
||||
if (m_mappings == nullptr) {
|
||||
KPageBuffer* page_buffer = KPageBuffer::Allocate(kernel);
|
||||
KPageBuffer* page_buffer = KPageBuffer::Allocate(m_kernel);
|
||||
R_UNLESS(page_buffer != nullptr, ResultOutOfMemory);
|
||||
|
||||
m_mappings = reinterpret_cast<Mapping*>(page_buffer);
|
||||
}
|
||||
|
||||
mapping = &m_mappings[index - NumStaticMappings];
|
||||
mapping = std::addressof(m_mappings[index - NumStaticMappings]);
|
||||
}
|
||||
|
||||
// Set the mapping.
|
||||
mapping->Set(client, server, size, state);
|
||||
|
||||
return ResultSuccess;
|
||||
R_SUCCEED();
|
||||
}
|
||||
|
||||
Result KSessionRequest::SessionMappings::PushSend(VAddr client, VAddr server, size_t size,
|
||||
KMemoryState state) {
|
||||
ASSERT(m_num_recv == 0);
|
||||
ASSERT(m_num_exch == 0);
|
||||
return this->PushMap(client, server, size, state, m_num_send++);
|
||||
R_RETURN(this->PushMap(client, server, size, state, m_num_send++));
|
||||
}
|
||||
|
||||
Result KSessionRequest::SessionMappings::PushReceive(VAddr client, VAddr server, size_t size,
|
||||
KMemoryState state) {
|
||||
ASSERT(m_num_exch == 0);
|
||||
return this->PushMap(client, server, size, state, m_num_send + m_num_recv++);
|
||||
R_RETURN(this->PushMap(client, server, size, state, m_num_send + m_num_recv++));
|
||||
}
|
||||
|
||||
Result KSessionRequest::SessionMappings::PushExchange(VAddr client, VAddr server, size_t size,
|
||||
KMemoryState state) {
|
||||
return this->PushMap(client, server, size, state, m_num_send + m_num_recv + m_num_exch++);
|
||||
R_RETURN(this->PushMap(client, server, size, state, m_num_send + m_num_recv + m_num_exch++));
|
||||
}
|
||||
|
||||
void KSessionRequest::SessionMappings::Finalize() {
|
||||
if (m_mappings) {
|
||||
KPageBuffer::Free(kernel, reinterpret_cast<KPageBuffer*>(m_mappings));
|
||||
KPageBuffer::Free(m_kernel, reinterpret_cast<KPageBuffer*>(m_mappings));
|
||||
m_mappings = nullptr;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -47,14 +47,14 @@ public:
|
|||
}
|
||||
|
||||
private:
|
||||
VAddr m_client_address;
|
||||
VAddr m_server_address;
|
||||
size_t m_size;
|
||||
KMemoryState m_state;
|
||||
VAddr m_client_address{};
|
||||
VAddr m_server_address{};
|
||||
size_t m_size{};
|
||||
KMemoryState m_state{};
|
||||
};
|
||||
|
||||
public:
|
||||
explicit SessionMappings(KernelCore& kernel_) : kernel(kernel_) {}
|
||||
explicit SessionMappings(KernelCore& kernel) : m_kernel(kernel) {}
|
||||
|
||||
void Initialize() {}
|
||||
void Finalize();
|
||||
|
@ -149,8 +149,8 @@ public:
|
|||
}
|
||||
|
||||
private:
|
||||
KernelCore& kernel;
|
||||
std::array<Mapping, NumStaticMappings> m_static_mappings;
|
||||
KernelCore& m_kernel;
|
||||
std::array<Mapping, NumStaticMappings> m_static_mappings{};
|
||||
Mapping* m_mappings{};
|
||||
u8 m_num_send{};
|
||||
u8 m_num_recv{};
|
||||
|
|
|
@ -15,15 +15,15 @@ namespace Kernel {
|
|||
KSharedMemory::KSharedMemory(KernelCore& kernel_) : KAutoObjectWithSlabHeapAndContainer{kernel_} {}
|
||||
KSharedMemory::~KSharedMemory() = default;
|
||||
|
||||
Result KSharedMemory::Initialize(Core::DeviceMemory& device_memory_, KProcess* owner_process_,
|
||||
Svc::MemoryPermission owner_permission_,
|
||||
Svc::MemoryPermission user_permission_, std::size_t size_) {
|
||||
Result KSharedMemory::Initialize(Core::DeviceMemory& device_memory, KProcess* owner_process,
|
||||
Svc::MemoryPermission owner_permission,
|
||||
Svc::MemoryPermission user_permission, std::size_t size) {
|
||||
// Set members.
|
||||
owner_process = owner_process_;
|
||||
device_memory = &device_memory_;
|
||||
owner_permission = owner_permission_;
|
||||
user_permission = user_permission_;
|
||||
size = Common::AlignUp(size_, PageSize);
|
||||
m_owner_process = owner_process;
|
||||
m_device_memory = std::addressof(device_memory);
|
||||
m_owner_permission = owner_permission;
|
||||
m_user_permission = user_permission;
|
||||
m_size = Common::AlignUp(size, PageSize);
|
||||
|
||||
const size_t num_pages = Common::DivideUp(size, PageSize);
|
||||
|
||||
|
@ -32,7 +32,7 @@ Result KSharedMemory::Initialize(Core::DeviceMemory& device_memory_, KProcess* o
|
|||
|
||||
// Reserve memory for ourselves.
|
||||
KScopedResourceReservation memory_reservation(reslimit, LimitableResource::PhysicalMemoryMax,
|
||||
size_);
|
||||
size);
|
||||
R_UNLESS(memory_reservation.Succeeded(), ResultLimitReached);
|
||||
|
||||
// Allocate the memory.
|
||||
|
@ -40,26 +40,26 @@ Result KSharedMemory::Initialize(Core::DeviceMemory& device_memory_, KProcess* o
|
|||
//! HACK: Open continuous mapping from sysmodule pool.
|
||||
auto option = KMemoryManager::EncodeOption(KMemoryManager::Pool::Secure,
|
||||
KMemoryManager::Direction::FromBack);
|
||||
physical_address = kernel.MemoryManager().AllocateAndOpenContinuous(num_pages, 1, option);
|
||||
R_UNLESS(physical_address != 0, ResultOutOfMemory);
|
||||
m_physical_address = kernel.MemoryManager().AllocateAndOpenContinuous(num_pages, 1, option);
|
||||
R_UNLESS(m_physical_address != 0, ResultOutOfMemory);
|
||||
|
||||
//! Insert the result into our page group.
|
||||
page_group.emplace(kernel, &kernel.GetSystemSystemResource().GetBlockInfoManager());
|
||||
page_group->AddBlock(physical_address, num_pages);
|
||||
m_page_group.emplace(kernel, &kernel.GetSystemSystemResource().GetBlockInfoManager());
|
||||
m_page_group->AddBlock(m_physical_address, num_pages);
|
||||
|
||||
// Commit our reservation.
|
||||
memory_reservation.Commit();
|
||||
|
||||
// Set our resource limit.
|
||||
resource_limit = reslimit;
|
||||
resource_limit->Open();
|
||||
m_resource_limit = reslimit;
|
||||
m_resource_limit->Open();
|
||||
|
||||
// Mark initialized.
|
||||
is_initialized = true;
|
||||
m_is_initialized = true;
|
||||
|
||||
// Clear all pages in the memory.
|
||||
for (const auto& block : *page_group) {
|
||||
std::memset(device_memory_.GetPointer<void>(block.GetAddress()), 0, block.GetSize());
|
||||
for (const auto& block : *m_page_group) {
|
||||
std::memset(m_device_memory->GetPointer<void>(block.GetAddress()), 0, block.GetSize());
|
||||
}
|
||||
|
||||
R_SUCCEED();
|
||||
|
@ -67,12 +67,12 @@ Result KSharedMemory::Initialize(Core::DeviceMemory& device_memory_, KProcess* o
|
|||
|
||||
void KSharedMemory::Finalize() {
|
||||
// Close and finalize the page group.
|
||||
page_group->Close();
|
||||
page_group->Finalize();
|
||||
m_page_group->Close();
|
||||
m_page_group->Finalize();
|
||||
|
||||
// Release the memory reservation.
|
||||
resource_limit->Release(LimitableResource::PhysicalMemoryMax, size);
|
||||
resource_limit->Close();
|
||||
m_resource_limit->Release(LimitableResource::PhysicalMemoryMax, m_size);
|
||||
m_resource_limit->Close();
|
||||
|
||||
// Perform inherited finalization.
|
||||
KAutoObjectWithSlabHeapAndContainer<KSharedMemory, KAutoObjectWithList>::Finalize();
|
||||
|
@ -81,26 +81,27 @@ void KSharedMemory::Finalize() {
|
|||
Result KSharedMemory::Map(KProcess& target_process, VAddr address, std::size_t map_size,
|
||||
Svc::MemoryPermission map_perm) {
|
||||
// Validate the size.
|
||||
R_UNLESS(size == map_size, ResultInvalidSize);
|
||||
R_UNLESS(m_size == map_size, ResultInvalidSize);
|
||||
|
||||
// Validate the permission.
|
||||
const Svc::MemoryPermission test_perm =
|
||||
&target_process == owner_process ? owner_permission : user_permission;
|
||||
std::addressof(target_process) == m_owner_process ? m_owner_permission : m_user_permission;
|
||||
if (test_perm == Svc::MemoryPermission::DontCare) {
|
||||
ASSERT(map_perm == Svc::MemoryPermission::Read || map_perm == Svc::MemoryPermission::Write);
|
||||
} else {
|
||||
R_UNLESS(map_perm == test_perm, ResultInvalidNewMemoryPermission);
|
||||
}
|
||||
|
||||
R_RETURN(target_process.PageTable().MapPageGroup(address, *page_group, KMemoryState::Shared,
|
||||
R_RETURN(target_process.PageTable().MapPageGroup(address, *m_page_group, KMemoryState::Shared,
|
||||
ConvertToKMemoryPermission(map_perm)));
|
||||
}
|
||||
|
||||
Result KSharedMemory::Unmap(KProcess& target_process, VAddr address, std::size_t unmap_size) {
|
||||
// Validate the size.
|
||||
R_UNLESS(size == unmap_size, ResultInvalidSize);
|
||||
R_UNLESS(m_size == unmap_size, ResultInvalidSize);
|
||||
|
||||
R_RETURN(target_process.PageTable().UnmapPageGroup(address, *page_group, KMemoryState::Shared));
|
||||
R_RETURN(
|
||||
target_process.PageTable().UnmapPageGroup(address, *m_page_group, KMemoryState::Shared));
|
||||
}
|
||||
|
||||
} // namespace Kernel
|
||||
|
|
|
@ -54,7 +54,7 @@ public:
|
|||
* @return A pointer to the shared memory block from the specified offset
|
||||
*/
|
||||
u8* GetPointer(std::size_t offset = 0) {
|
||||
return device_memory->GetPointer<u8>(physical_address + offset);
|
||||
return m_device_memory->GetPointer<u8>(m_physical_address + offset);
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -63,26 +63,26 @@ public:
|
|||
* @return A pointer to the shared memory block from the specified offset
|
||||
*/
|
||||
const u8* GetPointer(std::size_t offset = 0) const {
|
||||
return device_memory->GetPointer<u8>(physical_address + offset);
|
||||
return m_device_memory->GetPointer<u8>(m_physical_address + offset);
|
||||
}
|
||||
|
||||
void Finalize() override;
|
||||
|
||||
bool IsInitialized() const override {
|
||||
return is_initialized;
|
||||
return m_is_initialized;
|
||||
}
|
||||
static void PostDestroy([[maybe_unused]] uintptr_t arg) {}
|
||||
static void PostDestroy(uintptr_t arg) {}
|
||||
|
||||
private:
|
||||
Core::DeviceMemory* device_memory{};
|
||||
KProcess* owner_process{};
|
||||
std::optional<KPageGroup> page_group{};
|
||||
Svc::MemoryPermission owner_permission{};
|
||||
Svc::MemoryPermission user_permission{};
|
||||
PAddr physical_address{};
|
||||
std::size_t size{};
|
||||
KResourceLimit* resource_limit{};
|
||||
bool is_initialized{};
|
||||
Core::DeviceMemory* m_device_memory{};
|
||||
KProcess* m_owner_process{};
|
||||
std::optional<KPageGroup> m_page_group{};
|
||||
Svc::MemoryPermission m_owner_permission{};
|
||||
Svc::MemoryPermission m_user_permission{};
|
||||
PAddr m_physical_address{};
|
||||
std::size_t m_size{};
|
||||
KResourceLimit* m_resource_limit{};
|
||||
bool m_is_initialized{};
|
||||
};
|
||||
|
||||
} // namespace Kernel
|
||||
|
|
|
@ -18,25 +18,28 @@ public:
|
|||
explicit KSharedMemoryInfo(KernelCore&) {}
|
||||
KSharedMemoryInfo() = default;
|
||||
|
||||
constexpr void Initialize(KSharedMemory* shmem) {
|
||||
shared_memory = shmem;
|
||||
constexpr void Initialize(KSharedMemory* m) {
|
||||
m_shared_memory = m;
|
||||
m_reference_count = 0;
|
||||
}
|
||||
|
||||
constexpr KSharedMemory* GetSharedMemory() const {
|
||||
return shared_memory;
|
||||
return m_shared_memory;
|
||||
}
|
||||
|
||||
constexpr void Open() {
|
||||
++reference_count;
|
||||
++m_reference_count;
|
||||
ASSERT(m_reference_count > 0);
|
||||
}
|
||||
|
||||
constexpr bool Close() {
|
||||
return (--reference_count) == 0;
|
||||
ASSERT(m_reference_count > 0);
|
||||
return (--m_reference_count) == 0;
|
||||
}
|
||||
|
||||
private:
|
||||
KSharedMemory* shared_memory{};
|
||||
size_t reference_count{};
|
||||
KSharedMemory* m_shared_memory{};
|
||||
size_t m_reference_count{};
|
||||
};
|
||||
|
||||
} // namespace Kernel
|
||||
|
|
|
@ -6,15 +6,15 @@
|
|||
namespace Kernel {
|
||||
|
||||
void KSpinLock::Lock() {
|
||||
lck.lock();
|
||||
m_lock.lock();
|
||||
}
|
||||
|
||||
void KSpinLock::Unlock() {
|
||||
lck.unlock();
|
||||
m_lock.unlock();
|
||||
}
|
||||
|
||||
bool KSpinLock::TryLock() {
|
||||
return lck.try_lock();
|
||||
return m_lock.try_lock();
|
||||
}
|
||||
|
||||
} // namespace Kernel
|
||||
|
|
|
@ -5,26 +5,24 @@
|
|||
|
||||
#include <mutex>
|
||||
|
||||
#include "common/common_funcs.h"
|
||||
#include "core/hle/kernel/k_scoped_lock.h"
|
||||
|
||||
namespace Kernel {
|
||||
|
||||
class KSpinLock {
|
||||
public:
|
||||
KSpinLock() = default;
|
||||
explicit KSpinLock() = default;
|
||||
|
||||
KSpinLock(const KSpinLock&) = delete;
|
||||
KSpinLock& operator=(const KSpinLock&) = delete;
|
||||
|
||||
KSpinLock(KSpinLock&&) = delete;
|
||||
KSpinLock& operator=(KSpinLock&&) = delete;
|
||||
YUZU_NON_COPYABLE(KSpinLock);
|
||||
YUZU_NON_MOVEABLE(KSpinLock);
|
||||
|
||||
void Lock();
|
||||
void Unlock();
|
||||
[[nodiscard]] bool TryLock();
|
||||
bool TryLock();
|
||||
|
||||
private:
|
||||
std::mutex lck;
|
||||
std::mutex m_lock;
|
||||
};
|
||||
|
||||
// TODO(bunnei): Alias for now, in case we want to implement these accurately in the future.
|
||||
|
|
Loading…
Reference in a new issue