mirror of
https://github.com/yuzu-emu/yuzu.git
synced 2024-07-04 23:31:19 +01:00
core_timing,scheduler: Use std::scoped_lock when possible
Simplifies the cognitive load of procedures using locks and makes locks safe against exceptions.
This commit is contained in:
parent
b05795d704
commit
8562b516c0
2 changed files with 12 additions and 15 deletions
|
@ -172,7 +172,7 @@ void CoreTiming::ClearPendingEvents() {
|
||||||
}
|
}
|
||||||
|
|
||||||
void CoreTiming::RemoveEvent(const std::shared_ptr<EventType>& event_type) {
|
void CoreTiming::RemoveEvent(const std::shared_ptr<EventType>& event_type) {
|
||||||
basic_lock.lock();
|
std::scoped_lock lock{basic_lock};
|
||||||
|
|
||||||
const auto itr = std::remove_if(event_queue.begin(), event_queue.end(), [&](const Event& e) {
|
const auto itr = std::remove_if(event_queue.begin(), event_queue.end(), [&](const Event& e) {
|
||||||
return e.type.lock().get() == event_type.get();
|
return e.type.lock().get() == event_type.get();
|
||||||
|
@ -183,12 +183,10 @@ void CoreTiming::RemoveEvent(const std::shared_ptr<EventType>& event_type) {
|
||||||
event_queue.erase(itr, event_queue.end());
|
event_queue.erase(itr, event_queue.end());
|
||||||
std::make_heap(event_queue.begin(), event_queue.end(), std::greater<>());
|
std::make_heap(event_queue.begin(), event_queue.end(), std::greater<>());
|
||||||
}
|
}
|
||||||
basic_lock.unlock();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
std::optional<s64> CoreTiming::Advance() {
|
std::optional<s64> CoreTiming::Advance() {
|
||||||
std::scoped_lock advance_scope{advance_lock};
|
std::scoped_lock lock{advance_lock, basic_lock};
|
||||||
std::scoped_lock basic_scope{basic_lock};
|
|
||||||
global_timer = GetGlobalTimeNs().count();
|
global_timer = GetGlobalTimeNs().count();
|
||||||
|
|
||||||
while (!event_queue.empty() && event_queue.front().time <= global_timer) {
|
while (!event_queue.empty() && event_queue.front().time <= global_timer) {
|
||||||
|
|
|
@ -6,6 +6,7 @@
|
||||||
// licensed under GPLv2 or later under exception provided by the author.
|
// licensed under GPLv2 or later under exception provided by the author.
|
||||||
|
|
||||||
#include <algorithm>
|
#include <algorithm>
|
||||||
|
#include <mutex>
|
||||||
#include <set>
|
#include <set>
|
||||||
#include <unordered_set>
|
#include <unordered_set>
|
||||||
#include <utility>
|
#include <utility>
|
||||||
|
@ -31,22 +32,20 @@ GlobalScheduler::GlobalScheduler(KernelCore& kernel) : kernel{kernel} {}
|
||||||
GlobalScheduler::~GlobalScheduler() = default;
|
GlobalScheduler::~GlobalScheduler() = default;
|
||||||
|
|
||||||
void GlobalScheduler::AddThread(std::shared_ptr<Thread> thread) {
|
void GlobalScheduler::AddThread(std::shared_ptr<Thread> thread) {
|
||||||
global_list_guard.lock();
|
std::scoped_lock lock{global_list_guard};
|
||||||
thread_list.push_back(std::move(thread));
|
thread_list.push_back(std::move(thread));
|
||||||
global_list_guard.unlock();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void GlobalScheduler::RemoveThread(std::shared_ptr<Thread> thread) {
|
void GlobalScheduler::RemoveThread(std::shared_ptr<Thread> thread) {
|
||||||
global_list_guard.lock();
|
std::scoped_lock lock{global_list_guard};
|
||||||
thread_list.erase(std::remove(thread_list.begin(), thread_list.end(), thread),
|
thread_list.erase(std::remove(thread_list.begin(), thread_list.end(), thread),
|
||||||
thread_list.end());
|
thread_list.end());
|
||||||
global_list_guard.unlock();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
u32 GlobalScheduler::SelectThreads() {
|
u32 GlobalScheduler::SelectThreads() {
|
||||||
ASSERT(is_locked);
|
ASSERT(is_locked);
|
||||||
const auto update_thread = [](Thread* thread, Scheduler& sched) {
|
const auto update_thread = [](Thread* thread, Scheduler& sched) {
|
||||||
sched.guard.lock();
|
std::scoped_lock lock{sched.guard};
|
||||||
if (thread != sched.selected_thread_set.get()) {
|
if (thread != sched.selected_thread_set.get()) {
|
||||||
if (thread == nullptr) {
|
if (thread == nullptr) {
|
||||||
++sched.idle_selection_count;
|
++sched.idle_selection_count;
|
||||||
|
@ -57,7 +56,6 @@ u32 GlobalScheduler::SelectThreads() {
|
||||||
sched.is_context_switch_pending || (sched.selected_thread_set != sched.current_thread);
|
sched.is_context_switch_pending || (sched.selected_thread_set != sched.current_thread);
|
||||||
sched.is_context_switch_pending = reschedule_pending;
|
sched.is_context_switch_pending = reschedule_pending;
|
||||||
std::atomic_thread_fence(std::memory_order_seq_cst);
|
std::atomic_thread_fence(std::memory_order_seq_cst);
|
||||||
sched.guard.unlock();
|
|
||||||
return reschedule_pending;
|
return reschedule_pending;
|
||||||
};
|
};
|
||||||
if (!is_reselection_pending.load()) {
|
if (!is_reselection_pending.load()) {
|
||||||
|
@ -757,11 +755,12 @@ void Scheduler::OnSwitch(void* this_scheduler) {
|
||||||
|
|
||||||
void Scheduler::SwitchToCurrent() {
|
void Scheduler::SwitchToCurrent() {
|
||||||
while (true) {
|
while (true) {
|
||||||
guard.lock();
|
{
|
||||||
|
std::scoped_lock lock{guard};
|
||||||
selected_thread = selected_thread_set;
|
selected_thread = selected_thread_set;
|
||||||
current_thread = selected_thread;
|
current_thread = selected_thread;
|
||||||
is_context_switch_pending = false;
|
is_context_switch_pending = false;
|
||||||
guard.unlock();
|
}
|
||||||
while (!is_context_switch_pending) {
|
while (!is_context_switch_pending) {
|
||||||
if (current_thread != nullptr && !current_thread->IsHLEThread()) {
|
if (current_thread != nullptr && !current_thread->IsHLEThread()) {
|
||||||
current_thread->context_guard.lock();
|
current_thread->context_guard.lock();
|
||||||
|
|
Loading…
Reference in a new issue