1
0
Fork 0
forked from suyu/suyu

Kernel: Corrections to Scheduling.

This commit is contained in:
Fernando Sahmkow 2020-03-03 15:50:38 -04:00
parent b4dc01f16a
commit 07993ac8c8
5 changed files with 23 additions and 19 deletions

View file

@ -154,7 +154,7 @@ void CoreTiming::RemoveEvent(const std::shared_ptr<EventType>& event_type) {
basic_lock.unlock(); basic_lock.unlock();
} }
std::optional<u64> CoreTiming::Advance() { std::optional<s64> CoreTiming::Advance() {
advance_lock.lock(); advance_lock.lock();
basic_lock.lock(); basic_lock.lock();
global_timer = GetGlobalTimeNs().count(); global_timer = GetGlobalTimeNs().count();
@ -170,10 +170,11 @@ std::optional<u64> CoreTiming::Advance() {
} }
basic_lock.lock(); basic_lock.lock();
global_timer = GetGlobalTimeNs().count();
} }
if (!event_queue.empty()) { if (!event_queue.empty()) {
const u64 next_time = event_queue.front().time - global_timer; const s64 next_time = event_queue.front().time - global_timer;
basic_lock.unlock(); basic_lock.unlock();
advance_lock.unlock(); advance_lock.unlock();
return next_time; return next_time;
@ -191,8 +192,10 @@ void CoreTiming::ThreadLoop() {
paused_set = false; paused_set = false;
const auto next_time = Advance(); const auto next_time = Advance();
if (next_time) { if (next_time) {
if (*next_time > 0) {
std::chrono::nanoseconds next_time_ns = std::chrono::nanoseconds(*next_time); std::chrono::nanoseconds next_time_ns = std::chrono::nanoseconds(*next_time);
event.WaitFor(next_time_ns); event.WaitFor(next_time_ns);
}
} else { } else {
wait_set = true; wait_set = true;
event.Wait(); event.Wait();

View file

@ -110,7 +110,7 @@ public:
std::chrono::nanoseconds GetGlobalTimeNs() const; std::chrono::nanoseconds GetGlobalTimeNs() const;
/// Checks for events manually and returns time in nanoseconds for next event, threadsafe. /// Checks for events manually and returns time in nanoseconds for next event, threadsafe.
std::optional<u64> Advance(); std::optional<s64> Advance();
private: private:
struct Event; struct Event;

View file

@ -47,13 +47,13 @@ u32 GlobalScheduler::SelectThreads() {
ASSERT(is_locked); ASSERT(is_locked);
const auto update_thread = [](Thread* thread, Scheduler& sched) { const auto update_thread = [](Thread* thread, Scheduler& sched) {
sched.guard.lock(); sched.guard.lock();
if (thread != sched.selected_thread.get()) { if (thread != sched.selected_thread_set.get()) {
if (thread == nullptr) { if (thread == nullptr) {
++sched.idle_selection_count; ++sched.idle_selection_count;
} }
sched.selected_thread = SharedFrom(thread); sched.selected_thread_set = SharedFrom(thread);
} }
const bool reschedule_pending = sched.selected_thread != sched.current_thread; const bool reschedule_pending = sched.selected_thread_set != sched.current_thread;
sched.is_context_switch_pending = reschedule_pending; sched.is_context_switch_pending = reschedule_pending;
std::atomic_thread_fence(std::memory_order_seq_cst); std::atomic_thread_fence(std::memory_order_seq_cst);
sched.guard.unlock(); sched.guard.unlock();
@ -118,6 +118,8 @@ u32 GlobalScheduler::SelectThreads() {
suggested); suggested);
top_threads[candidate_core] = next; top_threads[candidate_core] = next;
break; break;
} else {
suggested = nullptr;
} }
} }
} }
@ -590,7 +592,7 @@ void Scheduler::OnThreadStart() {
} }
void Scheduler::SwitchContextStep2() { void Scheduler::SwitchContextStep2() {
Thread* previous_thread = current_thread.get(); Thread* previous_thread = current_thread_prev.get();
Thread* new_thread = selected_thread.get(); Thread* new_thread = selected_thread.get();
// Load context of new thread // Load context of new thread
@ -606,8 +608,6 @@ void Scheduler::SwitchContextStep2() {
"Thread must be ready to become running."); "Thread must be ready to become running.");
// Cancel any outstanding wakeup events for this thread // Cancel any outstanding wakeup events for this thread
current_thread = SharedFrom(new_thread);
new_thread->SetStatus(ThreadStatus::Running);
new_thread->SetIsRunning(true); new_thread->SetIsRunning(true);
auto* const thread_owner_process = current_thread->GetOwnerProcess(); auto* const thread_owner_process = current_thread->GetOwnerProcess();
@ -622,21 +622,21 @@ void Scheduler::SwitchContextStep2() {
cpu_core.SetTPIDR_EL0(new_thread->GetTPIDR_EL0()); cpu_core.SetTPIDR_EL0(new_thread->GetTPIDR_EL0());
cpu_core.ClearExclusiveState(); cpu_core.ClearExclusiveState();
} }
} else {
current_thread = nullptr;
// Note: We do not reset the current process and current page table when idling because
// technically we haven't changed processes, our threads are just paused.
} }
guard.unlock();
TryDoContextSwitch();
} }
void Scheduler::SwitchContext() { void Scheduler::SwitchContext() {
Thread* previous_thread = current_thread.get(); current_thread_prev = current_thread;
selected_thread = selected_thread_set;
Thread* previous_thread = current_thread_prev.get();
Thread* new_thread = selected_thread.get(); Thread* new_thread = selected_thread.get();
current_thread = selected_thread;
is_context_switch_pending = false; is_context_switch_pending = false;
if (new_thread == previous_thread) {
guard.unlock(); guard.unlock();
if (new_thread == previous_thread) {
return; return;
} }

View file

@ -249,6 +249,8 @@ private:
std::shared_ptr<Thread> current_thread = nullptr; std::shared_ptr<Thread> current_thread = nullptr;
std::shared_ptr<Thread> selected_thread = nullptr; std::shared_ptr<Thread> selected_thread = nullptr;
std::shared_ptr<Thread> current_thread_prev = nullptr;
std::shared_ptr<Thread> selected_thread_set = nullptr;
std::shared_ptr<Thread> idle_thread = nullptr; std::shared_ptr<Thread> idle_thread = nullptr;
Core::System& system; Core::System& system;

View file

@ -316,7 +316,6 @@ static ResultCode ConnectToNamedPort32(Core::System& system, Handle* out_handle,
/// Makes a blocking IPC call to an OS service. /// Makes a blocking IPC call to an OS service.
static ResultCode SendSyncRequest(Core::System& system, Handle handle) { static ResultCode SendSyncRequest(Core::System& system, Handle handle) {
std::lock_guard lock{HLE::g_hle_lock};
const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable(); const auto& handle_table = system.Kernel().CurrentProcess()->GetHandleTable();
std::shared_ptr<ClientSession> session = handle_table.Get<ClientSession>(handle); std::shared_ptr<ClientSession> session = handle_table.Get<ClientSession>(handle);
if (!session) { if (!session) {