forked from suyu/suyu
Merge pull request #8475 from liamwhite/x18
kernel: make current thread pointer thread local
This commit is contained in:
commit
abfd690601
13 changed files with 69 additions and 52 deletions
|
@ -95,7 +95,7 @@ void ARM_Interface::Run() {
|
||||||
using Kernel::SuspendType;
|
using Kernel::SuspendType;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
Kernel::KThread* current_thread{system.Kernel().CurrentScheduler()->GetCurrentThread()};
|
Kernel::KThread* current_thread{Kernel::GetCurrentThreadPointer(system.Kernel())};
|
||||||
Dynarmic::HaltReason hr{};
|
Dynarmic::HaltReason hr{};
|
||||||
|
|
||||||
// Notify the debugger and go to sleep if a step was performed
|
// Notify the debugger and go to sleep if a step was performed
|
||||||
|
|
|
@ -95,7 +95,7 @@ void* CpuManager::GetStartFuncParameter() {
|
||||||
void CpuManager::MultiCoreRunGuestThread() {
|
void CpuManager::MultiCoreRunGuestThread() {
|
||||||
auto& kernel = system.Kernel();
|
auto& kernel = system.Kernel();
|
||||||
kernel.CurrentScheduler()->OnThreadStart();
|
kernel.CurrentScheduler()->OnThreadStart();
|
||||||
auto* thread = kernel.CurrentScheduler()->GetCurrentThread();
|
auto* thread = kernel.CurrentScheduler()->GetSchedulerCurrentThread();
|
||||||
auto& host_context = thread->GetHostContext();
|
auto& host_context = thread->GetHostContext();
|
||||||
host_context->SetRewindPoint(GuestRewindFunction, this);
|
host_context->SetRewindPoint(GuestRewindFunction, this);
|
||||||
MultiCoreRunGuestLoop();
|
MultiCoreRunGuestLoop();
|
||||||
|
@ -132,7 +132,7 @@ void CpuManager::MultiCoreRunIdleThread() {
|
||||||
void CpuManager::SingleCoreRunGuestThread() {
|
void CpuManager::SingleCoreRunGuestThread() {
|
||||||
auto& kernel = system.Kernel();
|
auto& kernel = system.Kernel();
|
||||||
kernel.CurrentScheduler()->OnThreadStart();
|
kernel.CurrentScheduler()->OnThreadStart();
|
||||||
auto* thread = kernel.CurrentScheduler()->GetCurrentThread();
|
auto* thread = kernel.CurrentScheduler()->GetSchedulerCurrentThread();
|
||||||
auto& host_context = thread->GetHostContext();
|
auto& host_context = thread->GetHostContext();
|
||||||
host_context->SetRewindPoint(GuestRewindFunction, this);
|
host_context->SetRewindPoint(GuestRewindFunction, this);
|
||||||
SingleCoreRunGuestLoop();
|
SingleCoreRunGuestLoop();
|
||||||
|
@ -172,7 +172,7 @@ void CpuManager::PreemptSingleCore(bool from_running_enviroment) {
|
||||||
{
|
{
|
||||||
auto& kernel = system.Kernel();
|
auto& kernel = system.Kernel();
|
||||||
auto& scheduler = kernel.Scheduler(current_core);
|
auto& scheduler = kernel.Scheduler(current_core);
|
||||||
Kernel::KThread* current_thread = scheduler.GetCurrentThread();
|
Kernel::KThread* current_thread = scheduler.GetSchedulerCurrentThread();
|
||||||
if (idle_count >= 4 || from_running_enviroment) {
|
if (idle_count >= 4 || from_running_enviroment) {
|
||||||
if (!from_running_enviroment) {
|
if (!from_running_enviroment) {
|
||||||
system.CoreTiming().Idle();
|
system.CoreTiming().Idle();
|
||||||
|
@ -184,7 +184,7 @@ void CpuManager::PreemptSingleCore(bool from_running_enviroment) {
|
||||||
}
|
}
|
||||||
current_core.store((current_core + 1) % Core::Hardware::NUM_CPU_CORES);
|
current_core.store((current_core + 1) % Core::Hardware::NUM_CPU_CORES);
|
||||||
system.CoreTiming().ResetTicks();
|
system.CoreTiming().ResetTicks();
|
||||||
scheduler.Unload(scheduler.GetCurrentThread());
|
scheduler.Unload(scheduler.GetSchedulerCurrentThread());
|
||||||
|
|
||||||
auto& next_scheduler = kernel.Scheduler(current_core);
|
auto& next_scheduler = kernel.Scheduler(current_core);
|
||||||
Common::Fiber::YieldTo(current_thread->GetHostContext(), *next_scheduler.ControlContext());
|
Common::Fiber::YieldTo(current_thread->GetHostContext(), *next_scheduler.ControlContext());
|
||||||
|
@ -193,11 +193,9 @@ void CpuManager::PreemptSingleCore(bool from_running_enviroment) {
|
||||||
// May have changed scheduler
|
// May have changed scheduler
|
||||||
{
|
{
|
||||||
auto& scheduler = system.Kernel().Scheduler(current_core);
|
auto& scheduler = system.Kernel().Scheduler(current_core);
|
||||||
scheduler.Reload(scheduler.GetCurrentThread());
|
scheduler.Reload(scheduler.GetSchedulerCurrentThread());
|
||||||
if (!scheduler.IsIdle()) {
|
|
||||||
idle_count = 0;
|
idle_count = 0;
|
||||||
}
|
}
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void CpuManager::ShutdownThread() {
|
void CpuManager::ShutdownThread() {
|
||||||
|
@ -237,7 +235,8 @@ void CpuManager::RunThread(std::size_t core) {
|
||||||
system.GPU().ObtainContext();
|
system.GPU().ObtainContext();
|
||||||
}
|
}
|
||||||
|
|
||||||
auto current_thread = system.Kernel().CurrentScheduler()->GetCurrentThread();
|
auto* current_thread = system.Kernel().CurrentScheduler()->GetIdleThread();
|
||||||
|
Kernel::SetCurrentThread(system.Kernel(), current_thread);
|
||||||
Common::Fiber::YieldTo(data.host_context, *current_thread->GetHostContext());
|
Common::Fiber::YieldTo(data.host_context, *current_thread->GetHostContext());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -234,7 +234,7 @@ ResultCode KAddressArbiter::SignalAndModifyByWaitingCountIfEqual(VAddr addr, s32
|
||||||
|
|
||||||
ResultCode KAddressArbiter::WaitIfLessThan(VAddr addr, s32 value, bool decrement, s64 timeout) {
|
ResultCode KAddressArbiter::WaitIfLessThan(VAddr addr, s32 value, bool decrement, s64 timeout) {
|
||||||
// Prepare to wait.
|
// Prepare to wait.
|
||||||
KThread* cur_thread = kernel.CurrentScheduler()->GetCurrentThread();
|
KThread* cur_thread = GetCurrentThreadPointer(kernel);
|
||||||
ThreadQueueImplForKAddressArbiter wait_queue(kernel, std::addressof(thread_tree));
|
ThreadQueueImplForKAddressArbiter wait_queue(kernel, std::addressof(thread_tree));
|
||||||
|
|
||||||
{
|
{
|
||||||
|
@ -287,7 +287,7 @@ ResultCode KAddressArbiter::WaitIfLessThan(VAddr addr, s32 value, bool decrement
|
||||||
|
|
||||||
ResultCode KAddressArbiter::WaitIfEqual(VAddr addr, s32 value, s64 timeout) {
|
ResultCode KAddressArbiter::WaitIfEqual(VAddr addr, s32 value, s64 timeout) {
|
||||||
// Prepare to wait.
|
// Prepare to wait.
|
||||||
KThread* cur_thread = kernel.CurrentScheduler()->GetCurrentThread();
|
KThread* cur_thread = GetCurrentThreadPointer(kernel);
|
||||||
ThreadQueueImplForKAddressArbiter wait_queue(kernel, std::addressof(thread_tree));
|
ThreadQueueImplForKAddressArbiter wait_queue(kernel, std::addressof(thread_tree));
|
||||||
|
|
||||||
{
|
{
|
||||||
|
|
|
@ -106,7 +106,7 @@ KConditionVariable::KConditionVariable(Core::System& system_)
|
||||||
KConditionVariable::~KConditionVariable() = default;
|
KConditionVariable::~KConditionVariable() = default;
|
||||||
|
|
||||||
ResultCode KConditionVariable::SignalToAddress(VAddr addr) {
|
ResultCode KConditionVariable::SignalToAddress(VAddr addr) {
|
||||||
KThread* owner_thread = kernel.CurrentScheduler()->GetCurrentThread();
|
KThread* owner_thread = GetCurrentThreadPointer(kernel);
|
||||||
|
|
||||||
// Signal the address.
|
// Signal the address.
|
||||||
{
|
{
|
||||||
|
@ -147,7 +147,7 @@ ResultCode KConditionVariable::SignalToAddress(VAddr addr) {
|
||||||
}
|
}
|
||||||
|
|
||||||
ResultCode KConditionVariable::WaitForAddress(Handle handle, VAddr addr, u32 value) {
|
ResultCode KConditionVariable::WaitForAddress(Handle handle, VAddr addr, u32 value) {
|
||||||
KThread* cur_thread = kernel.CurrentScheduler()->GetCurrentThread();
|
KThread* cur_thread = GetCurrentThreadPointer(kernel);
|
||||||
ThreadQueueImplForKConditionVariableWaitForAddress wait_queue(kernel);
|
ThreadQueueImplForKConditionVariableWaitForAddress wait_queue(kernel);
|
||||||
|
|
||||||
// Wait for the address.
|
// Wait for the address.
|
||||||
|
|
|
@ -15,8 +15,7 @@ void HandleInterrupt(KernelCore& kernel, s32 core_id) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
auto& scheduler = kernel.Scheduler(core_id);
|
auto& current_thread = GetCurrentThread(kernel);
|
||||||
auto& current_thread = *scheduler.GetCurrentThread();
|
|
||||||
|
|
||||||
// If the user disable count is set, we may need to pin the current thread.
|
// If the user disable count is set, we may need to pin the current thread.
|
||||||
if (current_thread.GetUserDisableCount() && !process->GetPinnedThread(core_id)) {
|
if (current_thread.GetUserDisableCount() && !process->GetPinnedThread(core_id)) {
|
||||||
|
@ -26,7 +25,7 @@ void HandleInterrupt(KernelCore& kernel, s32 core_id) {
|
||||||
process->PinCurrentThread(core_id);
|
process->PinCurrentThread(core_id);
|
||||||
|
|
||||||
// Set the interrupt flag for the thread.
|
// Set the interrupt flag for the thread.
|
||||||
scheduler.GetCurrentThread()->SetInterruptFlag();
|
GetCurrentThread(kernel).SetInterruptFlag();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -176,7 +176,8 @@ void KProcess::PinCurrentThread(s32 core_id) {
|
||||||
ASSERT(kernel.GlobalSchedulerContext().IsLocked());
|
ASSERT(kernel.GlobalSchedulerContext().IsLocked());
|
||||||
|
|
||||||
// Get the current thread.
|
// Get the current thread.
|
||||||
KThread* cur_thread = kernel.Scheduler(static_cast<std::size_t>(core_id)).GetCurrentThread();
|
KThread* cur_thread =
|
||||||
|
kernel.Scheduler(static_cast<std::size_t>(core_id)).GetSchedulerCurrentThread();
|
||||||
|
|
||||||
// If the thread isn't terminated, pin it.
|
// If the thread isn't terminated, pin it.
|
||||||
if (!cur_thread->IsTerminationRequested()) {
|
if (!cur_thread->IsTerminationRequested()) {
|
||||||
|
@ -193,7 +194,8 @@ void KProcess::UnpinCurrentThread(s32 core_id) {
|
||||||
ASSERT(kernel.GlobalSchedulerContext().IsLocked());
|
ASSERT(kernel.GlobalSchedulerContext().IsLocked());
|
||||||
|
|
||||||
// Get the current thread.
|
// Get the current thread.
|
||||||
KThread* cur_thread = kernel.Scheduler(static_cast<std::size_t>(core_id)).GetCurrentThread();
|
KThread* cur_thread =
|
||||||
|
kernel.Scheduler(static_cast<std::size_t>(core_id)).GetSchedulerCurrentThread();
|
||||||
|
|
||||||
// Unpin it.
|
// Unpin it.
|
||||||
cur_thread->Unpin();
|
cur_thread->Unpin();
|
||||||
|
@ -420,11 +422,11 @@ void KProcess::PrepareForTermination() {
|
||||||
ChangeStatus(ProcessStatus::Exiting);
|
ChangeStatus(ProcessStatus::Exiting);
|
||||||
|
|
||||||
const auto stop_threads = [this](const std::vector<KThread*>& in_thread_list) {
|
const auto stop_threads = [this](const std::vector<KThread*>& in_thread_list) {
|
||||||
for (auto& thread : in_thread_list) {
|
for (auto* thread : in_thread_list) {
|
||||||
if (thread->GetOwnerProcess() != this)
|
if (thread->GetOwnerProcess() != this)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
if (thread == kernel.CurrentScheduler()->GetCurrentThread())
|
if (thread == GetCurrentThreadPointer(kernel))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
// TODO(Subv): When are the other running/ready threads terminated?
|
// TODO(Subv): When are the other running/ready threads terminated?
|
||||||
|
|
|
@ -317,7 +317,7 @@ void KScheduler::RotateScheduledQueue(s32 cpu_core_id, s32 priority) {
|
||||||
|
|
||||||
{
|
{
|
||||||
KThread* best_thread = priority_queue.GetScheduledFront(cpu_core_id);
|
KThread* best_thread = priority_queue.GetScheduledFront(cpu_core_id);
|
||||||
if (best_thread == GetCurrentThread()) {
|
if (best_thread == GetCurrentThreadPointer(kernel)) {
|
||||||
best_thread = priority_queue.GetScheduledNext(cpu_core_id, best_thread);
|
best_thread = priority_queue.GetScheduledNext(cpu_core_id, best_thread);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -424,7 +424,7 @@ void KScheduler::YieldWithoutCoreMigration(KernelCore& kernel) {
|
||||||
ASSERT(kernel.CurrentProcess() != nullptr);
|
ASSERT(kernel.CurrentProcess() != nullptr);
|
||||||
|
|
||||||
// Get the current thread and process.
|
// Get the current thread and process.
|
||||||
KThread& cur_thread = Kernel::GetCurrentThread(kernel);
|
KThread& cur_thread = GetCurrentThread(kernel);
|
||||||
KProcess& cur_process = *kernel.CurrentProcess();
|
KProcess& cur_process = *kernel.CurrentProcess();
|
||||||
|
|
||||||
// If the thread's yield count matches, there's nothing for us to do.
|
// If the thread's yield count matches, there's nothing for us to do.
|
||||||
|
@ -463,7 +463,7 @@ void KScheduler::YieldWithCoreMigration(KernelCore& kernel) {
|
||||||
ASSERT(kernel.CurrentProcess() != nullptr);
|
ASSERT(kernel.CurrentProcess() != nullptr);
|
||||||
|
|
||||||
// Get the current thread and process.
|
// Get the current thread and process.
|
||||||
KThread& cur_thread = Kernel::GetCurrentThread(kernel);
|
KThread& cur_thread = GetCurrentThread(kernel);
|
||||||
KProcess& cur_process = *kernel.CurrentProcess();
|
KProcess& cur_process = *kernel.CurrentProcess();
|
||||||
|
|
||||||
// If the thread's yield count matches, there's nothing for us to do.
|
// If the thread's yield count matches, there's nothing for us to do.
|
||||||
|
@ -551,7 +551,7 @@ void KScheduler::YieldToAnyThread(KernelCore& kernel) {
|
||||||
ASSERT(kernel.CurrentProcess() != nullptr);
|
ASSERT(kernel.CurrentProcess() != nullptr);
|
||||||
|
|
||||||
// Get the current thread and process.
|
// Get the current thread and process.
|
||||||
KThread& cur_thread = Kernel::GetCurrentThread(kernel);
|
KThread& cur_thread = GetCurrentThread(kernel);
|
||||||
KProcess& cur_process = *kernel.CurrentProcess();
|
KProcess& cur_process = *kernel.CurrentProcess();
|
||||||
|
|
||||||
// If the thread's yield count matches, there's nothing for us to do.
|
// If the thread's yield count matches, there's nothing for us to do.
|
||||||
|
@ -642,7 +642,7 @@ KScheduler::~KScheduler() {
|
||||||
ASSERT(!idle_thread);
|
ASSERT(!idle_thread);
|
||||||
}
|
}
|
||||||
|
|
||||||
KThread* KScheduler::GetCurrentThread() const {
|
KThread* KScheduler::GetSchedulerCurrentThread() const {
|
||||||
if (auto result = current_thread.load(); result) {
|
if (auto result = current_thread.load(); result) {
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
@ -654,7 +654,7 @@ u64 KScheduler::GetLastContextSwitchTicks() const {
|
||||||
}
|
}
|
||||||
|
|
||||||
void KScheduler::RescheduleCurrentCore() {
|
void KScheduler::RescheduleCurrentCore() {
|
||||||
ASSERT(GetCurrentThread()->GetDisableDispatchCount() == 1);
|
ASSERT(GetCurrentThread(system.Kernel()).GetDisableDispatchCount() == 1);
|
||||||
|
|
||||||
auto& phys_core = system.Kernel().PhysicalCore(core_id);
|
auto& phys_core = system.Kernel().PhysicalCore(core_id);
|
||||||
if (phys_core.IsInterrupted()) {
|
if (phys_core.IsInterrupted()) {
|
||||||
|
@ -665,7 +665,7 @@ void KScheduler::RescheduleCurrentCore() {
|
||||||
if (state.needs_scheduling.load()) {
|
if (state.needs_scheduling.load()) {
|
||||||
Schedule();
|
Schedule();
|
||||||
} else {
|
} else {
|
||||||
GetCurrentThread()->EnableDispatch();
|
GetCurrentThread(system.Kernel()).EnableDispatch();
|
||||||
guard.Unlock();
|
guard.Unlock();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -718,13 +718,18 @@ void KScheduler::Reload(KThread* thread) {
|
||||||
|
|
||||||
void KScheduler::SwitchContextStep2() {
|
void KScheduler::SwitchContextStep2() {
|
||||||
// Load context of new thread
|
// Load context of new thread
|
||||||
Reload(GetCurrentThread());
|
Reload(GetCurrentThreadPointer(system.Kernel()));
|
||||||
|
|
||||||
RescheduleCurrentCore();
|
RescheduleCurrentCore();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void KScheduler::Schedule() {
|
||||||
|
ASSERT(GetCurrentThread(system.Kernel()).GetDisableDispatchCount() == 1);
|
||||||
|
this->ScheduleImpl();
|
||||||
|
}
|
||||||
|
|
||||||
void KScheduler::ScheduleImpl() {
|
void KScheduler::ScheduleImpl() {
|
||||||
KThread* previous_thread = GetCurrentThread();
|
KThread* previous_thread = GetCurrentThreadPointer(system.Kernel());
|
||||||
KThread* next_thread = state.highest_priority_thread;
|
KThread* next_thread = state.highest_priority_thread;
|
||||||
|
|
||||||
state.needs_scheduling.store(false);
|
state.needs_scheduling.store(false);
|
||||||
|
@ -762,6 +767,7 @@ void KScheduler::ScheduleImpl() {
|
||||||
old_context = &previous_thread->GetHostContext();
|
old_context = &previous_thread->GetHostContext();
|
||||||
|
|
||||||
// Set the new thread.
|
// Set the new thread.
|
||||||
|
SetCurrentThread(system.Kernel(), next_thread);
|
||||||
current_thread.store(next_thread);
|
current_thread.store(next_thread);
|
||||||
|
|
||||||
guard.Unlock();
|
guard.Unlock();
|
||||||
|
@ -805,6 +811,7 @@ void KScheduler::SwitchToCurrent() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
auto thread = next_thread ? next_thread : idle_thread;
|
auto thread = next_thread ? next_thread : idle_thread;
|
||||||
|
SetCurrentThread(system.Kernel(), thread);
|
||||||
Common::Fiber::YieldTo(switch_fiber, *thread->GetHostContext());
|
Common::Fiber::YieldTo(switch_fiber, *thread->GetHostContext());
|
||||||
} while (!is_switch_pending());
|
} while (!is_switch_pending());
|
||||||
}
|
}
|
||||||
|
|
|
@ -48,18 +48,13 @@ public:
|
||||||
void Reload(KThread* thread);
|
void Reload(KThread* thread);
|
||||||
|
|
||||||
/// Gets the current running thread
|
/// Gets the current running thread
|
||||||
[[nodiscard]] KThread* GetCurrentThread() const;
|
[[nodiscard]] KThread* GetSchedulerCurrentThread() const;
|
||||||
|
|
||||||
/// Gets the idle thread
|
/// Gets the idle thread
|
||||||
[[nodiscard]] KThread* GetIdleThread() const {
|
[[nodiscard]] KThread* GetIdleThread() const {
|
||||||
return idle_thread;
|
return idle_thread;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns true if the scheduler is idle
|
|
||||||
[[nodiscard]] bool IsIdle() const {
|
|
||||||
return GetCurrentThread() == idle_thread;
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Gets the timestamp for the last context switch in ticks.
|
/// Gets the timestamp for the last context switch in ticks.
|
||||||
[[nodiscard]] u64 GetLastContextSwitchTicks() const;
|
[[nodiscard]] u64 GetLastContextSwitchTicks() const;
|
||||||
|
|
||||||
|
@ -149,10 +144,7 @@ private:
|
||||||
|
|
||||||
void RotateScheduledQueue(s32 cpu_core_id, s32 priority);
|
void RotateScheduledQueue(s32 cpu_core_id, s32 priority);
|
||||||
|
|
||||||
void Schedule() {
|
void Schedule();
|
||||||
ASSERT(GetCurrentThread()->GetDisableDispatchCount() == 1);
|
|
||||||
this->ScheduleImpl();
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Switches the CPU's active thread context to that of the specified thread
|
/// Switches the CPU's active thread context to that of the specified thread
|
||||||
void ScheduleImpl();
|
void ScheduleImpl();
|
||||||
|
|
|
@ -382,7 +382,7 @@ void KThread::FinishTermination() {
|
||||||
for (std::size_t i = 0; i < static_cast<std::size_t>(Core::Hardware::NUM_CPU_CORES); ++i) {
|
for (std::size_t i = 0; i < static_cast<std::size_t>(Core::Hardware::NUM_CPU_CORES); ++i) {
|
||||||
KThread* core_thread{};
|
KThread* core_thread{};
|
||||||
do {
|
do {
|
||||||
core_thread = kernel.Scheduler(i).GetCurrentThread();
|
core_thread = kernel.Scheduler(i).GetSchedulerCurrentThread();
|
||||||
} while (core_thread == this);
|
} while (core_thread == this);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -631,7 +631,7 @@ ResultCode KThread::SetCoreMask(s32 core_id_, u64 v_affinity_mask) {
|
||||||
s32 thread_core;
|
s32 thread_core;
|
||||||
for (thread_core = 0; thread_core < static_cast<s32>(Core::Hardware::NUM_CPU_CORES);
|
for (thread_core = 0; thread_core < static_cast<s32>(Core::Hardware::NUM_CPU_CORES);
|
||||||
++thread_core) {
|
++thread_core) {
|
||||||
if (kernel.Scheduler(thread_core).GetCurrentThread() == this) {
|
if (kernel.Scheduler(thread_core).GetSchedulerCurrentThread() == this) {
|
||||||
thread_is_current = true;
|
thread_is_current = true;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
@ -756,7 +756,7 @@ void KThread::WaitUntilSuspended() {
|
||||||
for (std::size_t i = 0; i < static_cast<std::size_t>(Core::Hardware::NUM_CPU_CORES); ++i) {
|
for (std::size_t i = 0; i < static_cast<std::size_t>(Core::Hardware::NUM_CPU_CORES); ++i) {
|
||||||
KThread* core_thread{};
|
KThread* core_thread{};
|
||||||
do {
|
do {
|
||||||
core_thread = kernel.Scheduler(i).GetCurrentThread();
|
core_thread = kernel.Scheduler(i).GetSchedulerCurrentThread();
|
||||||
} while (core_thread == this);
|
} while (core_thread == this);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -822,7 +822,7 @@ ResultCode KThread::SetActivity(Svc::ThreadActivity activity) {
|
||||||
// Check if the thread is currently running.
|
// Check if the thread is currently running.
|
||||||
// If it is, we'll need to retry.
|
// If it is, we'll need to retry.
|
||||||
for (auto i = 0; i < static_cast<s32>(Core::Hardware::NUM_CPU_CORES); ++i) {
|
for (auto i = 0; i < static_cast<s32>(Core::Hardware::NUM_CPU_CORES); ++i) {
|
||||||
if (kernel.Scheduler(i).GetCurrentThread() == this) {
|
if (kernel.Scheduler(i).GetSchedulerCurrentThread() == this) {
|
||||||
thread_is_current = true;
|
thread_is_current = true;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
@ -1175,6 +1175,10 @@ std::shared_ptr<Common::Fiber>& KThread::GetHostContext() {
|
||||||
return host_context;
|
return host_context;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void SetCurrentThread(KernelCore& kernel, KThread* thread) {
|
||||||
|
kernel.SetCurrentEmuThread(thread);
|
||||||
|
}
|
||||||
|
|
||||||
KThread* GetCurrentThreadPointer(KernelCore& kernel) {
|
KThread* GetCurrentThreadPointer(KernelCore& kernel) {
|
||||||
return kernel.GetCurrentEmuThread();
|
return kernel.GetCurrentEmuThread();
|
||||||
}
|
}
|
||||||
|
|
|
@ -106,6 +106,7 @@ enum class StepState : u32 {
|
||||||
StepPerformed, ///< Thread has stepped, waiting to be scheduled again
|
StepPerformed, ///< Thread has stepped, waiting to be scheduled again
|
||||||
};
|
};
|
||||||
|
|
||||||
|
void SetCurrentThread(KernelCore& kernel, KThread* thread);
|
||||||
[[nodiscard]] KThread* GetCurrentThreadPointer(KernelCore& kernel);
|
[[nodiscard]] KThread* GetCurrentThreadPointer(KernelCore& kernel);
|
||||||
[[nodiscard]] KThread& GetCurrentThread(KernelCore& kernel);
|
[[nodiscard]] KThread& GetCurrentThread(KernelCore& kernel);
|
||||||
[[nodiscard]] s32 GetCurrentCoreId(KernelCore& kernel);
|
[[nodiscard]] s32 GetCurrentCoreId(KernelCore& kernel);
|
||||||
|
|
|
@ -331,6 +331,8 @@ struct KernelCore::Impl {
|
||||||
return is_shutting_down.load(std::memory_order_relaxed);
|
return is_shutting_down.load(std::memory_order_relaxed);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline thread_local KThread* current_thread{nullptr};
|
||||||
|
|
||||||
KThread* GetCurrentEmuThread() {
|
KThread* GetCurrentEmuThread() {
|
||||||
// If we are shutting down the kernel, none of this is relevant anymore.
|
// If we are shutting down the kernel, none of this is relevant anymore.
|
||||||
if (IsShuttingDown()) {
|
if (IsShuttingDown()) {
|
||||||
|
@ -341,7 +343,12 @@ struct KernelCore::Impl {
|
||||||
if (thread_id >= Core::Hardware::NUM_CPU_CORES) {
|
if (thread_id >= Core::Hardware::NUM_CPU_CORES) {
|
||||||
return GetHostDummyThread();
|
return GetHostDummyThread();
|
||||||
}
|
}
|
||||||
return schedulers[thread_id]->GetCurrentThread();
|
|
||||||
|
return current_thread;
|
||||||
|
}
|
||||||
|
|
||||||
|
void SetCurrentEmuThread(KThread* thread) {
|
||||||
|
current_thread = thread;
|
||||||
}
|
}
|
||||||
|
|
||||||
void DeriveInitialMemoryLayout() {
|
void DeriveInitialMemoryLayout() {
|
||||||
|
@ -1024,6 +1031,10 @@ KThread* KernelCore::GetCurrentEmuThread() const {
|
||||||
return impl->GetCurrentEmuThread();
|
return impl->GetCurrentEmuThread();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void KernelCore::SetCurrentEmuThread(KThread* thread) {
|
||||||
|
impl->SetCurrentEmuThread(thread);
|
||||||
|
}
|
||||||
|
|
||||||
KMemoryManager& KernelCore::MemoryManager() {
|
KMemoryManager& KernelCore::MemoryManager() {
|
||||||
return *impl->memory_manager;
|
return *impl->memory_manager;
|
||||||
}
|
}
|
||||||
|
|
|
@ -226,6 +226,9 @@ public:
|
||||||
/// Gets the current host_thread/guest_thread pointer.
|
/// Gets the current host_thread/guest_thread pointer.
|
||||||
KThread* GetCurrentEmuThread() const;
|
KThread* GetCurrentEmuThread() const;
|
||||||
|
|
||||||
|
/// Sets the current guest_thread pointer.
|
||||||
|
void SetCurrentEmuThread(KThread* thread);
|
||||||
|
|
||||||
/// Gets the current host_thread handle.
|
/// Gets the current host_thread handle.
|
||||||
u32 GetCurrentHostThreadID() const;
|
u32 GetCurrentHostThreadID() const;
|
||||||
|
|
||||||
|
|
|
@ -327,7 +327,6 @@ static ResultCode SendSyncRequest(Core::System& system, Handle handle) {
|
||||||
|
|
||||||
LOG_TRACE(Kernel_SVC, "called handle=0x{:08X}({})", handle, session->GetName());
|
LOG_TRACE(Kernel_SVC, "called handle=0x{:08X}({})", handle, session->GetName());
|
||||||
|
|
||||||
auto thread = kernel.CurrentScheduler()->GetCurrentThread();
|
|
||||||
{
|
{
|
||||||
KScopedSchedulerLock lock(kernel);
|
KScopedSchedulerLock lock(kernel);
|
||||||
|
|
||||||
|
@ -337,7 +336,7 @@ static ResultCode SendSyncRequest(Core::System& system, Handle handle) {
|
||||||
session->SendSyncRequest(&GetCurrentThread(kernel), system.Memory(), system.CoreTiming());
|
session->SendSyncRequest(&GetCurrentThread(kernel), system.Memory(), system.CoreTiming());
|
||||||
}
|
}
|
||||||
|
|
||||||
return thread->GetWaitResult();
|
return GetCurrentThread(kernel).GetWaitResult();
|
||||||
}
|
}
|
||||||
|
|
||||||
static ResultCode SendSyncRequest32(Core::System& system, Handle handle) {
|
static ResultCode SendSyncRequest32(Core::System& system, Handle handle) {
|
||||||
|
@ -624,7 +623,7 @@ static void Break(Core::System& system, u32 reason, u64 info1, u64 info2) {
|
||||||
|
|
||||||
handle_debug_buffer(info1, info2);
|
handle_debug_buffer(info1, info2);
|
||||||
|
|
||||||
auto* const current_thread = system.Kernel().CurrentScheduler()->GetCurrentThread();
|
auto* const current_thread = GetCurrentThreadPointer(system.Kernel());
|
||||||
const auto thread_processor_id = current_thread->GetActiveCore();
|
const auto thread_processor_id = current_thread->GetActiveCore();
|
||||||
system.ArmInterface(static_cast<std::size_t>(thread_processor_id)).LogBacktrace();
|
system.ArmInterface(static_cast<std::size_t>(thread_processor_id)).LogBacktrace();
|
||||||
}
|
}
|
||||||
|
@ -884,7 +883,7 @@ static ResultCode GetInfo(Core::System& system, u64* result, u64 info_id, Handle
|
||||||
|
|
||||||
const auto& core_timing = system.CoreTiming();
|
const auto& core_timing = system.CoreTiming();
|
||||||
const auto& scheduler = *system.Kernel().CurrentScheduler();
|
const auto& scheduler = *system.Kernel().CurrentScheduler();
|
||||||
const auto* const current_thread = scheduler.GetCurrentThread();
|
const auto* const current_thread = GetCurrentThreadPointer(system.Kernel());
|
||||||
const bool same_thread = current_thread == thread.GetPointerUnsafe();
|
const bool same_thread = current_thread == thread.GetPointerUnsafe();
|
||||||
|
|
||||||
const u64 prev_ctx_ticks = scheduler.GetLastContextSwitchTicks();
|
const u64 prev_ctx_ticks = scheduler.GetLastContextSwitchTicks();
|
||||||
|
@ -1103,7 +1102,7 @@ static ResultCode GetThreadContext(Core::System& system, VAddr out_context, Hand
|
||||||
if (thread->GetRawState() != ThreadState::Runnable) {
|
if (thread->GetRawState() != ThreadState::Runnable) {
|
||||||
bool current = false;
|
bool current = false;
|
||||||
for (auto i = 0; i < static_cast<s32>(Core::Hardware::NUM_CPU_CORES); ++i) {
|
for (auto i = 0; i < static_cast<s32>(Core::Hardware::NUM_CPU_CORES); ++i) {
|
||||||
if (thread.GetPointerUnsafe() == kernel.Scheduler(i).GetCurrentThread()) {
|
if (thread.GetPointerUnsafe() == kernel.Scheduler(i).GetSchedulerCurrentThread()) {
|
||||||
current = true;
|
current = true;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
@ -1851,7 +1850,7 @@ static ResultCode StartThread32(Core::System& system, Handle thread_handle) {
|
||||||
static void ExitThread(Core::System& system) {
|
static void ExitThread(Core::System& system) {
|
||||||
LOG_DEBUG(Kernel_SVC, "called, pc=0x{:08X}", system.CurrentArmInterface().GetPC());
|
LOG_DEBUG(Kernel_SVC, "called, pc=0x{:08X}", system.CurrentArmInterface().GetPC());
|
||||||
|
|
||||||
auto* const current_thread = system.Kernel().CurrentScheduler()->GetCurrentThread();
|
auto* const current_thread = GetCurrentThreadPointer(system.Kernel());
|
||||||
system.GlobalSchedulerContext().RemoveThread(current_thread);
|
system.GlobalSchedulerContext().RemoveThread(current_thread);
|
||||||
current_thread->Exit();
|
current_thread->Exit();
|
||||||
system.Kernel().UnregisterInUseObject(current_thread);
|
system.Kernel().UnregisterInUseObject(current_thread);
|
||||||
|
@ -2993,7 +2992,7 @@ void Call(Core::System& system, u32 immediate) {
|
||||||
auto& kernel = system.Kernel();
|
auto& kernel = system.Kernel();
|
||||||
kernel.EnterSVCProfile();
|
kernel.EnterSVCProfile();
|
||||||
|
|
||||||
auto* thread = kernel.CurrentScheduler()->GetCurrentThread();
|
auto* thread = GetCurrentThreadPointer(kernel);
|
||||||
thread->SetIsCallingSvc();
|
thread->SetIsCallingSvc();
|
||||||
|
|
||||||
const FunctionDef* info = system.CurrentProcess()->Is64BitProcess() ? GetSVCInfo64(immediate)
|
const FunctionDef* info = system.CurrentProcess()->Is64BitProcess() ? GetSVCInfo64(immediate)
|
||||||
|
|
Loading…
Reference in a new issue