forked from suyu/suyu
kernel/scheduler: Use deduction guides on mutex locks
Since C++17, we no longer need to explicitly specify the type of the mutex within the lock_guard. The type system can now deduce these with deduction guides.
This commit is contained in:
parent
f770c17d01
commit
cb805f45ae
1 changed files with 8 additions and 8 deletions
|
@ -29,7 +29,7 @@ Scheduler::~Scheduler() {
|
||||||
}
|
}
|
||||||
|
|
||||||
bool Scheduler::HaveReadyThreads() const {
|
bool Scheduler::HaveReadyThreads() const {
|
||||||
std::lock_guard<std::mutex> lock(scheduler_mutex);
|
std::lock_guard lock{scheduler_mutex};
|
||||||
return !ready_queue.empty();
|
return !ready_queue.empty();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -132,7 +132,7 @@ void Scheduler::UpdateLastContextSwitchTime(Thread* thread, Process* process) {
|
||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::Reschedule() {
|
void Scheduler::Reschedule() {
|
||||||
std::lock_guard<std::mutex> lock(scheduler_mutex);
|
std::lock_guard lock{scheduler_mutex};
|
||||||
|
|
||||||
Thread* cur = GetCurrentThread();
|
Thread* cur = GetCurrentThread();
|
||||||
Thread* next = PopNextReadyThread();
|
Thread* next = PopNextReadyThread();
|
||||||
|
@ -149,34 +149,34 @@ void Scheduler::Reschedule() {
|
||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::AddThread(SharedPtr<Thread> thread, u32 priority) {
|
void Scheduler::AddThread(SharedPtr<Thread> thread, u32 priority) {
|
||||||
std::lock_guard<std::mutex> lock(scheduler_mutex);
|
std::lock_guard lock{scheduler_mutex};
|
||||||
|
|
||||||
thread_list.push_back(std::move(thread));
|
thread_list.push_back(std::move(thread));
|
||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::RemoveThread(Thread* thread) {
|
void Scheduler::RemoveThread(Thread* thread) {
|
||||||
std::lock_guard<std::mutex> lock(scheduler_mutex);
|
std::lock_guard lock{scheduler_mutex};
|
||||||
|
|
||||||
thread_list.erase(std::remove(thread_list.begin(), thread_list.end(), thread),
|
thread_list.erase(std::remove(thread_list.begin(), thread_list.end(), thread),
|
||||||
thread_list.end());
|
thread_list.end());
|
||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::ScheduleThread(Thread* thread, u32 priority) {
|
void Scheduler::ScheduleThread(Thread* thread, u32 priority) {
|
||||||
std::lock_guard<std::mutex> lock(scheduler_mutex);
|
std::lock_guard lock{scheduler_mutex};
|
||||||
|
|
||||||
ASSERT(thread->GetStatus() == ThreadStatus::Ready);
|
ASSERT(thread->GetStatus() == ThreadStatus::Ready);
|
||||||
ready_queue.add(thread, priority);
|
ready_queue.add(thread, priority);
|
||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::UnscheduleThread(Thread* thread, u32 priority) {
|
void Scheduler::UnscheduleThread(Thread* thread, u32 priority) {
|
||||||
std::lock_guard<std::mutex> lock(scheduler_mutex);
|
std::lock_guard lock{scheduler_mutex};
|
||||||
|
|
||||||
ASSERT(thread->GetStatus() == ThreadStatus::Ready);
|
ASSERT(thread->GetStatus() == ThreadStatus::Ready);
|
||||||
ready_queue.remove(thread, priority);
|
ready_queue.remove(thread, priority);
|
||||||
}
|
}
|
||||||
|
|
||||||
void Scheduler::SetThreadPriority(Thread* thread, u32 priority) {
|
void Scheduler::SetThreadPriority(Thread* thread, u32 priority) {
|
||||||
std::lock_guard<std::mutex> lock(scheduler_mutex);
|
std::lock_guard lock{scheduler_mutex};
|
||||||
if (thread->GetPriority() == priority) {
|
if (thread->GetPriority() == priority) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
@ -187,7 +187,7 @@ void Scheduler::SetThreadPriority(Thread* thread, u32 priority) {
|
||||||
}
|
}
|
||||||
|
|
||||||
Thread* Scheduler::GetNextSuggestedThread(u32 core, u32 maximum_priority) const {
|
Thread* Scheduler::GetNextSuggestedThread(u32 core, u32 maximum_priority) const {
|
||||||
std::lock_guard<std::mutex> lock(scheduler_mutex);
|
std::lock_guard lock{scheduler_mutex};
|
||||||
|
|
||||||
const u32 mask = 1U << core;
|
const u32 mask = 1U << core;
|
||||||
for (auto* thread : ready_queue) {
|
for (auto* thread : ready_queue) {
|
||||||
|
|
Loading…
Reference in a new issue