1
0
Fork 0
forked from suyu/suyu

kernel: fix issues with single core mode

This commit is contained in:
Liam 2022-07-05 23:27:25 -04:00
parent 0624c880bd
commit 21945ae127
9 changed files with 229 additions and 193 deletions

View file

@ -42,14 +42,6 @@ void CpuManager::Shutdown() {
} }
} }
void CpuManager::GuestActivateFunction() {
if (is_multicore) {
MultiCoreGuestActivate();
} else {
SingleCoreGuestActivate();
}
}
void CpuManager::GuestThreadFunction() { void CpuManager::GuestThreadFunction() {
if (is_multicore) { if (is_multicore) {
MultiCoreRunGuestThread(); MultiCoreRunGuestThread();
@ -58,21 +50,16 @@ void CpuManager::GuestThreadFunction() {
} }
} }
void CpuManager::ShutdownThreadFunction() { void CpuManager::IdleThreadFunction() {
ShutdownThread(); if (is_multicore) {
MultiCoreRunIdleThread();
} else {
SingleCoreRunIdleThread();
}
} }
void CpuManager::WaitForAndHandleInterrupt() { void CpuManager::ShutdownThreadFunction() {
auto& kernel = system.Kernel(); ShutdownThread();
auto& physical_core = kernel.CurrentPhysicalCore();
ASSERT(Kernel::GetCurrentThread(kernel).GetDisableDispatchCount() == 1);
if (!physical_core.IsInterrupted()) {
physical_core.Idle();
}
HandleInterrupt();
} }
void CpuManager::HandleInterrupt() { void CpuManager::HandleInterrupt() {
@ -86,26 +73,10 @@ void CpuManager::HandleInterrupt() {
/// MultiCore /// /// MultiCore ///
/////////////////////////////////////////////////////////////////////////////// ///////////////////////////////////////////////////////////////////////////////
void CpuManager::MultiCoreGuestActivate() {
// Similar to the HorizonKernelMain callback in HOS
auto& kernel = system.Kernel();
auto* scheduler = kernel.CurrentScheduler();
scheduler->Activate();
UNREACHABLE();
}
void CpuManager::MultiCoreRunGuestThread() { void CpuManager::MultiCoreRunGuestThread() {
// Similar to UserModeThreadStarter in HOS // Similar to UserModeThreadStarter in HOS
auto& kernel = system.Kernel(); auto& kernel = system.Kernel();
auto* thread = kernel.GetCurrentEmuThread(); kernel.CurrentScheduler()->OnThreadStart();
thread->EnableDispatch();
MultiCoreRunGuestLoop();
}
void CpuManager::MultiCoreRunGuestLoop() {
auto& kernel = system.Kernel();
while (true) { while (true) {
auto* physical_core = &kernel.CurrentPhysicalCore(); auto* physical_core = &kernel.CurrentPhysicalCore();
@ -118,17 +89,105 @@ void CpuManager::MultiCoreRunGuestLoop() {
} }
} }
void CpuManager::MultiCoreRunIdleThread() {
// Not accurate to HOS. Remove this entire method when singlecore is removed.
// See notes in KScheduler::ScheduleImpl for more information about why this
// is inaccurate.
auto& kernel = system.Kernel();
kernel.CurrentScheduler()->OnThreadStart();
while (true) {
auto& physical_core = kernel.CurrentPhysicalCore();
if (!physical_core.IsInterrupted()) {
physical_core.Idle();
}
HandleInterrupt();
}
}
/////////////////////////////////////////////////////////////////////////////// ///////////////////////////////////////////////////////////////////////////////
/// SingleCore /// /// SingleCore ///
/////////////////////////////////////////////////////////////////////////////// ///////////////////////////////////////////////////////////////////////////////
void CpuManager::SingleCoreGuestActivate() {} void CpuManager::SingleCoreRunGuestThread() {
auto& kernel = system.Kernel();
kernel.CurrentScheduler()->OnThreadStart();
void CpuManager::SingleCoreRunGuestThread() {} while (true) {
auto* physical_core = &kernel.CurrentPhysicalCore();
if (!physical_core->IsInterrupted()) {
physical_core->Run();
physical_core = &kernel.CurrentPhysicalCore();
}
void CpuManager::SingleCoreRunGuestLoop() {} kernel.SetIsPhantomModeForSingleCore(true);
system.CoreTiming().Advance();
kernel.SetIsPhantomModeForSingleCore(false);
void CpuManager::PreemptSingleCore(bool from_running_enviroment) {} PreemptSingleCore();
HandleInterrupt();
}
}
void CpuManager::SingleCoreRunIdleThread() {
auto& kernel = system.Kernel();
kernel.CurrentScheduler()->OnThreadStart();
while (true) {
PreemptSingleCore(false);
system.CoreTiming().AddTicks(1000U);
idle_count++;
HandleInterrupt();
}
}
void CpuManager::PreemptSingleCore(bool from_running_environment) {
{
auto& kernel = system.Kernel();
auto& scheduler = kernel.Scheduler(current_core);
Kernel::KThread* current_thread = scheduler.GetSchedulerCurrentThread();
if (idle_count >= 4 || from_running_environment) {
if (!from_running_environment) {
system.CoreTiming().Idle();
idle_count = 0;
}
kernel.SetIsPhantomModeForSingleCore(true);
system.CoreTiming().Advance();
kernel.SetIsPhantomModeForSingleCore(false);
}
current_core.store((current_core + 1) % Core::Hardware::NUM_CPU_CORES);
system.CoreTiming().ResetTicks();
scheduler.Unload(scheduler.GetSchedulerCurrentThread());
auto& next_scheduler = kernel.Scheduler(current_core);
// Disable dispatch. We're about to preempt this thread.
Kernel::KScopedDisableDispatch dd{kernel};
Common::Fiber::YieldTo(current_thread->GetHostContext(), *next_scheduler.GetSwitchFiber());
}
// We've now been scheduled again, and we may have exchanged schedulers.
// Reload the scheduler in case it's different.
{
auto& scheduler = system.Kernel().Scheduler(current_core);
scheduler.Reload(scheduler.GetSchedulerCurrentThread());
if (!scheduler.IsIdle()) {
idle_count = 0;
}
}
}
void CpuManager::GuestActivate() {
// Similar to the HorizonKernelMain callback in HOS
auto& kernel = system.Kernel();
auto* scheduler = kernel.CurrentScheduler();
scheduler->Activate();
UNREACHABLE();
}
void CpuManager::ShutdownThread() { void CpuManager::ShutdownThread() {
auto& kernel = system.Kernel(); auto& kernel = system.Kernel();
@ -168,20 +227,11 @@ void CpuManager::RunThread(std::size_t core) {
} }
auto& kernel = system.Kernel(); auto& kernel = system.Kernel();
auto& scheduler = *kernel.CurrentScheduler();
auto* thread = scheduler.GetSchedulerCurrentThread();
Kernel::SetCurrentThread(kernel, thread);
auto* main_thread = Kernel::KThread::Create(kernel); Common::Fiber::YieldTo(data.host_context, *thread->GetHostContext());
main_thread->SetName(fmt::format("MainThread:{}", core));
ASSERT(Kernel::KThread::InitializeMainThread(system, main_thread, static_cast<s32>(core))
.IsSuccess());
auto* idle_thread = Kernel::KThread::Create(kernel);
ASSERT(Kernel::KThread::InitializeIdleThread(system, idle_thread, static_cast<s32>(core))
.IsSuccess());
kernel.SetCurrentEmuThread(main_thread);
kernel.CurrentScheduler()->Initialize(idle_thread);
Common::Fiber::YieldTo(data.host_context, *main_thread->GetHostContext());
} }
} // namespace Core } // namespace Core

View file

@ -48,12 +48,11 @@ public:
gpu_barrier->Sync(); gpu_barrier->Sync();
} }
void WaitForAndHandleInterrupt();
void Initialize(); void Initialize();
void Shutdown(); void Shutdown();
std::function<void()> GetGuestActivateFunc() { std::function<void()> GetGuestActivateFunc() {
return [this] { GuestActivateFunction(); }; return [this] { GuestActivate(); };
} }
std::function<void()> GetGuestThreadFunc() { std::function<void()> GetGuestThreadFunc() {
return [this] { GuestThreadFunction(); }; return [this] { GuestThreadFunction(); };
@ -72,21 +71,19 @@ public:
} }
private: private:
void GuestActivateFunction();
void GuestThreadFunction(); void GuestThreadFunction();
void IdleThreadFunction(); void IdleThreadFunction();
void ShutdownThreadFunction(); void ShutdownThreadFunction();
void MultiCoreGuestActivate();
void MultiCoreRunGuestThread(); void MultiCoreRunGuestThread();
void MultiCoreRunGuestLoop(); void MultiCoreRunIdleThread();
void SingleCoreGuestActivate();
void SingleCoreRunGuestThread(); void SingleCoreRunGuestThread();
void SingleCoreRunGuestLoop(); void SingleCoreRunIdleThread();
static void ThreadStart(std::stop_token stop_token, CpuManager& cpu_manager, std::size_t core); static void ThreadStart(std::stop_token stop_token, CpuManager& cpu_manager, std::size_t core);
void GuestActivate();
void HandleInterrupt(); void HandleInterrupt();
void ShutdownThread(); void ShutdownThread();
void RunThread(std::size_t core); void RunThread(std::size_t core);

View file

@ -42,11 +42,6 @@ void GlobalSchedulerContext::PreemptThreads() {
for (u32 core_id = 0; core_id < Core::Hardware::NUM_CPU_CORES; core_id++) { for (u32 core_id = 0; core_id < Core::Hardware::NUM_CPU_CORES; core_id++) {
const u32 priority = preemption_priorities[core_id]; const u32 priority = preemption_priorities[core_id];
KScheduler::RotateScheduledQueue(kernel, core_id, priority); KScheduler::RotateScheduledQueue(kernel, core_id, priority);
// Signal an interrupt occurred. For core 3, this is a certainty, as preemption will result
// in the rotator thread being scheduled. For cores 0-2, this is to simulate or system
// interrupts that may have occurred.
kernel.PhysicalCore(core_id).Interrupt();
} }
} }

View file

@ -28,9 +28,9 @@ static void IncrementScheduledCount(Kernel::KThread* thread) {
} }
KScheduler::KScheduler(KernelCore& kernel_) : kernel{kernel_} { KScheduler::KScheduler(KernelCore& kernel_) : kernel{kernel_} {
m_idle_stack = std::make_shared<Common::Fiber>([this] { m_switch_fiber = std::make_shared<Common::Fiber>([this] {
while (true) { while (true) {
ScheduleImplOffStack(); ScheduleImplFiber();
} }
}); });
@ -60,9 +60,9 @@ void KScheduler::DisableScheduling(KernelCore& kernel) {
void KScheduler::EnableScheduling(KernelCore& kernel, u64 cores_needing_scheduling) { void KScheduler::EnableScheduling(KernelCore& kernel, u64 cores_needing_scheduling) {
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() >= 1); ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() >= 1);
auto* scheduler = kernel.CurrentScheduler(); auto* scheduler{kernel.CurrentScheduler()};
if (!scheduler) { if (!scheduler || kernel.IsPhantomModeForSingleCore()) {
// HACK: we cannot schedule from this thread, it is not a core thread // HACK: we cannot schedule from this thread, it is not a core thread
RescheduleCores(kernel, cores_needing_scheduling); RescheduleCores(kernel, cores_needing_scheduling);
if (GetCurrentThread(kernel).GetDisableDispatchCount() == 1) { if (GetCurrentThread(kernel).GetDisableDispatchCount() == 1) {
@ -125,9 +125,9 @@ void KScheduler::RescheduleCurrentCoreImpl() {
} }
} }
void KScheduler::Initialize(KThread* idle_thread) { void KScheduler::Initialize(KThread* main_thread, KThread* idle_thread, s32 core_id) {
// Set core ID/idle thread/interrupt task manager. // Set core ID/idle thread/interrupt task manager.
m_core_id = GetCurrentCoreId(kernel); m_core_id = core_id;
m_idle_thread = idle_thread; m_idle_thread = idle_thread;
// m_state.idle_thread_stack = m_idle_thread->GetStackTop(); // m_state.idle_thread_stack = m_idle_thread->GetStackTop();
// m_state.interrupt_task_manager = &kernel.GetInterruptTaskManager(); // m_state.interrupt_task_manager = &kernel.GetInterruptTaskManager();
@ -142,10 +142,10 @@ void KScheduler::Initialize(KThread* idle_thread) {
// Bind interrupt handler. // Bind interrupt handler.
// kernel.GetInterruptManager().BindHandler( // kernel.GetInterruptManager().BindHandler(
// GetSchedulerInterruptHandler(kernel), KInterruptName::Scheduler, m_core_id, // GetSchedulerInterruptHandler(kernel), KInterruptName::Scheduler, m_core_id,
// KInterruptController::PriorityLevel_Scheduler, false, false); // KInterruptController::PriorityLevel::Scheduler, false, false);
// Set the current thread. // Set the current thread.
m_current_thread = GetCurrentThreadPointer(kernel); m_current_thread = main_thread;
} }
void KScheduler::Activate() { void KScheduler::Activate() {
@ -156,6 +156,10 @@ void KScheduler::Activate() {
RescheduleCurrentCore(); RescheduleCurrentCore();
} }
void KScheduler::OnThreadStart() {
GetCurrentThread(kernel).EnableDispatch();
}
u64 KScheduler::UpdateHighestPriorityThread(KThread* highest_thread) { u64 KScheduler::UpdateHighestPriorityThread(KThread* highest_thread) {
if (KThread* prev_highest_thread = m_state.highest_priority_thread; if (KThread* prev_highest_thread = m_state.highest_priority_thread;
prev_highest_thread != highest_thread) [[likely]] { prev_highest_thread != highest_thread) [[likely]] {
@ -372,37 +376,30 @@ void KScheduler::ScheduleImpl() {
} }
// The highest priority thread is not the same as the current thread. // The highest priority thread is not the same as the current thread.
// Switch to the idle thread stack and continue executing from there. // Jump to the switcher and continue executing from there.
m_idle_cur_thread = cur_thread; m_switch_cur_thread = cur_thread;
m_idle_highest_priority_thread = highest_priority_thread; m_switch_highest_priority_thread = highest_priority_thread;
Common::Fiber::YieldTo(cur_thread->host_context, *m_idle_stack); m_switch_from_schedule = true;
Common::Fiber::YieldTo(cur_thread->host_context, *m_switch_fiber);
// Returning from ScheduleImpl occurs after this thread has been scheduled again. // Returning from ScheduleImpl occurs after this thread has been scheduled again.
} }
void KScheduler::ScheduleImplOffStack() { void KScheduler::ScheduleImplFiber() {
KThread* const cur_thread{m_idle_cur_thread}; KThread* const cur_thread{m_switch_cur_thread};
KThread* highest_priority_thread{m_idle_highest_priority_thread}; KThread* highest_priority_thread{m_switch_highest_priority_thread};
// Get a reference to the current thread's stack parameters. // If we're not coming from scheduling (i.e., we came from SC preemption),
auto& sp{cur_thread->GetStackParameters()}; // we should restart the scheduling loop directly. Not accurate to HOS.
if (!m_switch_from_schedule) {
goto retry;
}
// Mark that we are not coming from scheduling anymore.
m_switch_from_schedule = false;
// Save the original thread context. // Save the original thread context.
{ Unload(cur_thread);
auto& physical_core = kernel.System().CurrentPhysicalCore();
auto& cpu_core = physical_core.ArmInterface();
cpu_core.SaveContext(cur_thread->GetContext32());
cpu_core.SaveContext(cur_thread->GetContext64());
// Save the TPIDR_EL0 system register in case it was modified.
cur_thread->SetTPIDR_EL0(cpu_core.GetTPIDR_EL0());
cpu_core.ClearExclusiveState();
}
// Check if the thread is terminated by checking the DPC flags.
if ((sp.dpc_flags & static_cast<u32>(DpcFlag::Terminated)) == 0) {
// The thread isn't terminated, so we want to unlock it.
sp.m_lock.store(false, std::memory_order_seq_cst);
}
// The current thread's context has been entirely taken care of. // The current thread's context has been entirely taken care of.
// Now we want to loop until we successfully switch the thread context. // Now we want to loop until we successfully switch the thread context.
@ -411,45 +408,24 @@ void KScheduler::ScheduleImplOffStack() {
// Check if the highest priority thread is null. // Check if the highest priority thread is null.
if (!highest_priority_thread) { if (!highest_priority_thread) {
// The next thread is nullptr! // The next thread is nullptr!
// Switch to nullptr. This will actually switch to the idle thread.
SwitchThread(nullptr);
// We've switched to the idle thread, so we want to process interrupt tasks until we // Switch to the idle thread. Note: HOS treats idling as a special case for
// schedule a non-idle thread. // performance. This is not *required* for yuzu's purposes, and for singlecore
while (!m_state.interrupt_task_runnable) { // compatibility, we can just move the logic that would go here into the execution
// Check if we need scheduling. // of the idle thread. If we ever remove singlecore, we should implement this
if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) { // accurately to HOS.
goto retry; highest_priority_thread = m_idle_thread;
} }
// Clear the previous thread.
m_state.prev_thread = nullptr;
// Wait for an interrupt before checking again.
kernel.System().GetCpuManager().WaitForAndHandleInterrupt();
}
// Execute any pending interrupt tasks.
// m_state.interrupt_task_manager->DoTasks();
// Clear the interrupt task thread as runnable.
m_state.interrupt_task_runnable = false;
// Retry the scheduling loop.
goto retry;
} else {
// We want to try to lock the highest priority thread's context. // We want to try to lock the highest priority thread's context.
// Try to take it. // Try to take it.
bool expected{false}; while (!highest_priority_thread->context_guard.try_lock()) {
while (!highest_priority_thread->stack_parameters.m_lock.compare_exchange_strong(
expected, true, std::memory_order_seq_cst)) {
// The highest priority thread's context is already locked. // The highest priority thread's context is already locked.
// Check if we need scheduling. If we don't, we can retry directly. // Check if we need scheduling. If we don't, we can retry directly.
if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) { if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) {
// If we do, another core is interfering, and we must start again. // If we do, another core is interfering, and we must start again.
goto retry; goto retry;
} }
expected = false;
} }
// It's time to switch the thread. // It's time to switch the thread.
@ -461,13 +437,11 @@ void KScheduler::ScheduleImplOffStack() {
if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) { if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) {
// Our switch failed. // Our switch failed.
// We should unlock the thread context, and then retry. // We should unlock the thread context, and then retry.
highest_priority_thread->stack_parameters.m_lock.store(false, highest_priority_thread->context_guard.unlock();
std::memory_order_seq_cst);
goto retry; goto retry;
} else { } else {
break; break;
} }
}
retry: retry:
@ -480,18 +454,35 @@ void KScheduler::ScheduleImplOffStack() {
} }
// Reload the guest thread context. // Reload the guest thread context.
{ Reload(highest_priority_thread);
auto& cpu_core = kernel.System().CurrentArmInterface();
cpu_core.LoadContext(highest_priority_thread->GetContext32());
cpu_core.LoadContext(highest_priority_thread->GetContext64());
cpu_core.SetTlsAddress(highest_priority_thread->GetTLSAddress());
cpu_core.SetTPIDR_EL0(highest_priority_thread->GetTPIDR_EL0());
cpu_core.LoadWatchpointArray(highest_priority_thread->GetOwnerProcess()->GetWatchpoints());
cpu_core.ClearExclusiveState();
}
// Reload the host thread. // Reload the host thread.
Common::Fiber::YieldTo(m_idle_stack, *highest_priority_thread->host_context); Common::Fiber::YieldTo(m_switch_fiber, *highest_priority_thread->host_context);
}
void KScheduler::Unload(KThread* thread) {
auto& cpu_core = kernel.System().ArmInterface(m_core_id);
cpu_core.SaveContext(thread->GetContext32());
cpu_core.SaveContext(thread->GetContext64());
// Save the TPIDR_EL0 system register in case it was modified.
thread->SetTPIDR_EL0(cpu_core.GetTPIDR_EL0());
cpu_core.ClearExclusiveState();
// Check if the thread is terminated by checking the DPC flags.
if ((thread->GetStackParameters().dpc_flags & static_cast<u32>(DpcFlag::Terminated)) == 0) {
// The thread isn't terminated, so we want to unlock it.
thread->context_guard.unlock();
}
}
void KScheduler::Reload(KThread* thread) {
auto& cpu_core = kernel.System().ArmInterface(m_core_id);
cpu_core.LoadContext(thread->GetContext32());
cpu_core.LoadContext(thread->GetContext64());
cpu_core.SetTlsAddress(thread->GetTLSAddress());
cpu_core.SetTPIDR_EL0(thread->GetTPIDR_EL0());
cpu_core.LoadWatchpointArray(thread->GetOwnerProcess()->GetWatchpoints());
cpu_core.ClearExclusiveState();
} }
void KScheduler::ClearPreviousThread(KernelCore& kernel, KThread* thread) { void KScheduler::ClearPreviousThread(KernelCore& kernel, KThread* thread) {

View file

@ -41,8 +41,11 @@ public:
explicit KScheduler(KernelCore& kernel); explicit KScheduler(KernelCore& kernel);
~KScheduler(); ~KScheduler();
void Initialize(KThread* idle_thread); void Initialize(KThread* main_thread, KThread* idle_thread, s32 core_id);
void Activate(); void Activate();
void OnThreadStart();
void Unload(KThread* thread);
void Reload(KThread* thread);
void SetInterruptTaskRunnable(); void SetInterruptTaskRunnable();
void RequestScheduleOnInterrupt(); void RequestScheduleOnInterrupt();
@ -55,6 +58,14 @@ public:
return m_idle_thread; return m_idle_thread;
} }
bool IsIdle() const {
return m_current_thread.load() == m_idle_thread;
}
std::shared_ptr<Common::Fiber> GetSwitchFiber() {
return m_switch_fiber;
}
KThread* GetPreviousThread() const { KThread* GetPreviousThread() const {
return m_state.prev_thread; return m_state.prev_thread;
} }
@ -69,7 +80,7 @@ public:
// Static public API. // Static public API.
static bool CanSchedule(KernelCore& kernel) { static bool CanSchedule(KernelCore& kernel) {
return kernel.GetCurrentEmuThread()->GetDisableDispatchCount() == 0; return GetCurrentThread(kernel).GetDisableDispatchCount() == 0;
} }
static bool IsSchedulerLockedByCurrentThread(KernelCore& kernel) { static bool IsSchedulerLockedByCurrentThread(KernelCore& kernel) {
return kernel.GlobalSchedulerContext().scheduler_lock.IsLockedByCurrentThread(); return kernel.GlobalSchedulerContext().scheduler_lock.IsLockedByCurrentThread();
@ -113,7 +124,7 @@ private:
// Instanced private API. // Instanced private API.
void ScheduleImpl(); void ScheduleImpl();
void ScheduleImplOffStack(); void ScheduleImplFiber();
void SwitchThread(KThread* next_thread); void SwitchThread(KThread* next_thread);
void Schedule(); void Schedule();
@ -147,9 +158,10 @@ private:
KThread* m_idle_thread{nullptr}; KThread* m_idle_thread{nullptr};
std::atomic<KThread*> m_current_thread{nullptr}; std::atomic<KThread*> m_current_thread{nullptr};
std::shared_ptr<Common::Fiber> m_idle_stack{}; std::shared_ptr<Common::Fiber> m_switch_fiber{};
KThread* m_idle_cur_thread{}; KThread* m_switch_cur_thread{};
KThread* m_idle_highest_priority_thread{}; KThread* m_switch_highest_priority_thread{};
bool m_switch_from_schedule{};
}; };
class KScopedSchedulerLock : public KScopedLock<KScheduler::LockType> { class KScopedSchedulerLock : public KScopedLock<KScheduler::LockType> {

View file

@ -268,7 +268,7 @@ Result KThread::InitializeMainThread(Core::System& system, KThread* thread, s32
Result KThread::InitializeIdleThread(Core::System& system, KThread* thread, s32 virt_core) { Result KThread::InitializeIdleThread(Core::System& system, KThread* thread, s32 virt_core) {
return InitializeThread(thread, {}, {}, {}, IdleThreadPriority, virt_core, {}, ThreadType::Main, return InitializeThread(thread, {}, {}, {}, IdleThreadPriority, virt_core, {}, ThreadType::Main,
abort); system.GetCpuManager().GetIdleThreadStartFunc());
} }
Result KThread::InitializeHighPriorityThread(Core::System& system, KThread* thread, Result KThread::InitializeHighPriorityThread(Core::System& system, KThread* thread,
@ -1204,8 +1204,9 @@ KScopedDisableDispatch::~KScopedDisableDispatch() {
return; return;
} }
// Skip the reschedule if single-core, as dispatch tracking is disabled here. // Skip the reschedule if single-core.
if (!Settings::values.use_multi_core.GetValue()) { if (!Settings::values.use_multi_core.GetValue()) {
GetCurrentThread(kernel).EnableDispatch();
return; return;
} }

View file

@ -439,7 +439,6 @@ public:
bool is_pinned; bool is_pinned;
s32 disable_count; s32 disable_count;
KThread* cur_thread; KThread* cur_thread;
std::atomic<bool> m_lock;
}; };
[[nodiscard]] StackParameters& GetStackParameters() { [[nodiscard]] StackParameters& GetStackParameters() {
@ -485,39 +484,16 @@ public:
return per_core_priority_queue_entry[core]; return per_core_priority_queue_entry[core];
} }
[[nodiscard]] bool IsKernelThread() const {
return GetActiveCore() == 3;
}
[[nodiscard]] bool IsDispatchTrackingDisabled() const {
return is_single_core || IsKernelThread();
}
[[nodiscard]] s32 GetDisableDispatchCount() const { [[nodiscard]] s32 GetDisableDispatchCount() const {
if (IsDispatchTrackingDisabled()) {
// TODO(bunnei): Until kernel threads are emulated, we cannot enable/disable dispatch.
return 1;
}
return this->GetStackParameters().disable_count; return this->GetStackParameters().disable_count;
} }
void DisableDispatch() { void DisableDispatch() {
if (IsDispatchTrackingDisabled()) {
// TODO(bunnei): Until kernel threads are emulated, we cannot enable/disable dispatch.
return;
}
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() >= 0); ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() >= 0);
this->GetStackParameters().disable_count++; this->GetStackParameters().disable_count++;
} }
void EnableDispatch() { void EnableDispatch() {
if (IsDispatchTrackingDisabled()) {
// TODO(bunnei): Until kernel threads are emulated, we cannot enable/disable dispatch.
return;
}
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() > 0); ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() > 0);
this->GetStackParameters().disable_count--; this->GetStackParameters().disable_count--;
} }

View file

@ -64,8 +64,6 @@ struct KernelCore::Impl {
is_phantom_mode_for_singlecore = false; is_phantom_mode_for_singlecore = false;
InitializePhysicalCores();
// Derive the initial memory layout from the emulated board // Derive the initial memory layout from the emulated board
Init::InitializeSlabResourceCounts(kernel); Init::InitializeSlabResourceCounts(kernel);
DeriveInitialMemoryLayout(); DeriveInitialMemoryLayout();
@ -77,6 +75,7 @@ struct KernelCore::Impl {
Init::InitializeKPageBufferSlabHeap(system); Init::InitializeKPageBufferSlabHeap(system);
InitializeShutdownThreads(); InitializeShutdownThreads();
InitializePreemption(kernel); InitializePreemption(kernel);
InitializePhysicalCores();
RegisterHostThread(); RegisterHostThread();
} }
@ -193,8 +192,21 @@ struct KernelCore::Impl {
exclusive_monitor = exclusive_monitor =
Core::MakeExclusiveMonitor(system.Memory(), Core::Hardware::NUM_CPU_CORES); Core::MakeExclusiveMonitor(system.Memory(), Core::Hardware::NUM_CPU_CORES);
for (u32 i = 0; i < Core::Hardware::NUM_CPU_CORES; i++) { for (u32 i = 0; i < Core::Hardware::NUM_CPU_CORES; i++) {
const s32 core{static_cast<s32>(i)};
schedulers[i] = std::make_unique<Kernel::KScheduler>(system.Kernel()); schedulers[i] = std::make_unique<Kernel::KScheduler>(system.Kernel());
cores.emplace_back(i, system, *schedulers[i], interrupts); cores.emplace_back(i, system, *schedulers[i], interrupts);
auto* main_thread{Kernel::KThread::Create(system.Kernel())};
main_thread->SetName(fmt::format("MainThread:{}", core));
main_thread->SetCurrentCore(core);
ASSERT(Kernel::KThread::InitializeMainThread(system, main_thread, core).IsSuccess());
auto* idle_thread{Kernel::KThread::Create(system.Kernel())};
idle_thread->SetCurrentCore(core);
ASSERT(Kernel::KThread::InitializeIdleThread(system, idle_thread, core).IsSuccess());
schedulers[i]->Initialize(main_thread, idle_thread, core);
} }
} }
@ -1093,10 +1105,11 @@ void KernelCore::Suspend(bool suspended) {
} }
void KernelCore::ShutdownCores() { void KernelCore::ShutdownCores() {
KScopedSchedulerLock lk{*this};
for (auto* thread : impl->shutdown_threads) { for (auto* thread : impl->shutdown_threads) {
void(thread->Run()); void(thread->Run());
} }
InterruptAllPhysicalCores();
} }
bool KernelCore::IsMulticore() const { bool KernelCore::IsMulticore() const {

View file

@ -43,6 +43,7 @@ void PhysicalCore::Initialize([[maybe_unused]] bool is_64_bit) {
void PhysicalCore::Run() { void PhysicalCore::Run() {
arm_interface->Run(); arm_interface->Run();
arm_interface->ClearExclusiveState();
} }
void PhysicalCore::Idle() { void PhysicalCore::Idle() {