mirror of
https://github.com/yuzu-emu/yuzu.git
synced 2024-12-25 17:00:07 +00:00
Kernel/SVC: Signal the highest priority threads first in svcSignalProcessWideKey.
This commit is contained in:
parent
c74f2555b6
commit
2a35a36251
@ -608,61 +608,78 @@ static ResultCode SignalProcessWideKey(VAddr condition_variable_addr, s32 target
|
|||||||
NGLOG_TRACE(Kernel_SVC, "called, condition_variable_addr=0x{:X}, target=0x{:08X}",
|
NGLOG_TRACE(Kernel_SVC, "called, condition_variable_addr=0x{:X}, target=0x{:08X}",
|
||||||
condition_variable_addr, target);
|
condition_variable_addr, target);
|
||||||
|
|
||||||
u32 processed = 0;
|
auto RetrieveWaitingThreads =
|
||||||
|
[](size_t core_index, std::vector<SharedPtr<Thread>>& waiting_threads, VAddr condvar_addr) {
|
||||||
|
const auto& scheduler = Core::System::GetInstance().Scheduler(core_index);
|
||||||
|
auto& thread_list = scheduler->GetThreadList();
|
||||||
|
|
||||||
auto signal_process_wide_key = [&](size_t core_index) {
|
for (auto& thread : thread_list) {
|
||||||
const auto& scheduler = Core::System::GetInstance().Scheduler(core_index);
|
if (thread->condvar_wait_address == condvar_addr)
|
||||||
for (auto& thread : scheduler->GetThreadList()) {
|
waiting_threads.push_back(thread);
|
||||||
if (thread->condvar_wait_address != condition_variable_addr)
|
|
||||||
continue;
|
|
||||||
|
|
||||||
// Only process up to 'target' threads, unless 'target' is -1, in which case process
|
|
||||||
// them all.
|
|
||||||
if (target != -1 && processed >= target)
|
|
||||||
break;
|
|
||||||
|
|
||||||
// If the mutex is not yet acquired, acquire it.
|
|
||||||
u32 mutex_val = Memory::Read32(thread->mutex_wait_address);
|
|
||||||
|
|
||||||
if (mutex_val == 0) {
|
|
||||||
// We were able to acquire the mutex, resume this thread.
|
|
||||||
Memory::Write32(thread->mutex_wait_address, thread->wait_handle);
|
|
||||||
ASSERT(thread->status == THREADSTATUS_WAIT_MUTEX);
|
|
||||||
thread->ResumeFromWait();
|
|
||||||
|
|
||||||
auto lock_owner = thread->lock_owner;
|
|
||||||
if (lock_owner)
|
|
||||||
lock_owner->RemoveMutexWaiter(thread);
|
|
||||||
|
|
||||||
thread->lock_owner = nullptr;
|
|
||||||
thread->mutex_wait_address = 0;
|
|
||||||
thread->condvar_wait_address = 0;
|
|
||||||
thread->wait_handle = 0;
|
|
||||||
} else {
|
|
||||||
// Couldn't acquire the mutex, block the thread.
|
|
||||||
Handle owner_handle = static_cast<Handle>(mutex_val & Mutex::MutexOwnerMask);
|
|
||||||
auto owner = g_handle_table.Get<Thread>(owner_handle);
|
|
||||||
ASSERT(owner);
|
|
||||||
ASSERT(thread->status != THREADSTATUS_RUNNING);
|
|
||||||
thread->status = THREADSTATUS_WAIT_MUTEX;
|
|
||||||
thread->wakeup_callback = nullptr;
|
|
||||||
|
|
||||||
// Signal that the mutex now has a waiting thread.
|
|
||||||
Memory::Write32(thread->mutex_wait_address, mutex_val | Mutex::MutexHasWaitersFlag);
|
|
||||||
|
|
||||||
owner->AddMutexWaiter(thread);
|
|
||||||
|
|
||||||
Core::System::GetInstance().CpuCore(thread->processor_id).PrepareReschedule();
|
|
||||||
}
|
}
|
||||||
|
};
|
||||||
|
|
||||||
++processed;
|
// Retrieve a list of all threads that are waiting for this condition variable.
|
||||||
|
std::vector<SharedPtr<Thread>> waiting_threads;
|
||||||
|
RetrieveWaitingThreads(0, waiting_threads, condition_variable_addr);
|
||||||
|
RetrieveWaitingThreads(1, waiting_threads, condition_variable_addr);
|
||||||
|
RetrieveWaitingThreads(2, waiting_threads, condition_variable_addr);
|
||||||
|
RetrieveWaitingThreads(3, waiting_threads, condition_variable_addr);
|
||||||
|
// Sort them by priority, such that the highest priority ones come first.
|
||||||
|
std::sort(waiting_threads.begin(), waiting_threads.end(),
|
||||||
|
[](const SharedPtr<Thread>& lhs, const SharedPtr<Thread>& rhs) {
|
||||||
|
return lhs->current_priority < rhs->current_priority;
|
||||||
|
});
|
||||||
|
|
||||||
|
// Only process up to 'target' threads, unless 'target' is -1, in which case process
|
||||||
|
// them all.
|
||||||
|
size_t last = waiting_threads.size();
|
||||||
|
if (target != -1)
|
||||||
|
last = target;
|
||||||
|
|
||||||
|
// If there are no threads waiting on this condition variable, just exit
|
||||||
|
if (last > waiting_threads.size())
|
||||||
|
return RESULT_SUCCESS;
|
||||||
|
|
||||||
|
for (size_t index = 0; index < last; ++index) {
|
||||||
|
auto& thread = waiting_threads[index];
|
||||||
|
|
||||||
|
ASSERT(thread->condvar_wait_address == condition_variable_addr);
|
||||||
|
|
||||||
|
// If the mutex is not yet acquired, acquire it.
|
||||||
|
u32 mutex_val = Memory::Read32(thread->mutex_wait_address);
|
||||||
|
|
||||||
|
if (mutex_val == 0) {
|
||||||
|
// We were able to acquire the mutex, resume this thread.
|
||||||
|
Memory::Write32(thread->mutex_wait_address, thread->wait_handle);
|
||||||
|
ASSERT(thread->status == THREADSTATUS_WAIT_MUTEX);
|
||||||
|
thread->ResumeFromWait();
|
||||||
|
|
||||||
|
auto lock_owner = thread->lock_owner;
|
||||||
|
if (lock_owner)
|
||||||
|
lock_owner->RemoveMutexWaiter(thread);
|
||||||
|
|
||||||
|
thread->lock_owner = nullptr;
|
||||||
|
thread->mutex_wait_address = 0;
|
||||||
|
thread->condvar_wait_address = 0;
|
||||||
|
thread->wait_handle = 0;
|
||||||
|
} else {
|
||||||
|
// Couldn't acquire the mutex, block the thread.
|
||||||
|
Handle owner_handle = static_cast<Handle>(mutex_val & Mutex::MutexOwnerMask);
|
||||||
|
auto owner = g_handle_table.Get<Thread>(owner_handle);
|
||||||
|
ASSERT(owner);
|
||||||
|
ASSERT(thread->status != THREADSTATUS_RUNNING);
|
||||||
|
thread->status = THREADSTATUS_WAIT_MUTEX;
|
||||||
|
thread->wakeup_callback = nullptr;
|
||||||
|
|
||||||
|
// Signal that the mutex now has a waiting thread.
|
||||||
|
Memory::Write32(thread->mutex_wait_address, mutex_val | Mutex::MutexHasWaitersFlag);
|
||||||
|
|
||||||
|
owner->AddMutexWaiter(thread);
|
||||||
|
|
||||||
|
Core::System::GetInstance().CpuCore(thread->processor_id).PrepareReschedule();
|
||||||
}
|
}
|
||||||
};
|
}
|
||||||
|
|
||||||
signal_process_wide_key(0);
|
|
||||||
signal_process_wide_key(1);
|
|
||||||
signal_process_wide_key(2);
|
|
||||||
signal_process_wide_key(3);
|
|
||||||
|
|
||||||
return RESULT_SUCCESS;
|
return RESULT_SUCCESS;
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user