#include #include #include #include #include #include namespace mesosphere { bool KThread::IsAlive() const { return true; } void KThread::OnAlarm() { CancelKernelSync(); } void KThread::AdjustScheduling(ushort oldMaskFull) { if (currentSchedMaskFull == oldMaskFull) { return; } else if (CompareSchedulingStatusFull(oldMaskFull, SchedulingStatus::Running)) { KScheduler::Global::SetThreadPaused(*this); } else if (CompareSchedulingStatusFull(SchedulingStatus::Running)) { KScheduler::Global::SetThreadRunning(*this); } } void KThread::Reschedule(KThread::SchedulingStatus newStatus) { std::lock_guard criticalSection{KScheduler::GetCriticalSection()}; AdjustScheduling(SetSchedulingStatusField(newStatus)); } void KThread::RescheduleIfStatusEquals(SchedulingStatus expectedStatus, SchedulingStatus newStatus) { if(GetSchedulingStatus() == expectedStatus) { Reschedule(newStatus); } } void KThread::AddForcePauseReason(KThread::ForcePauseReason reason) { std::lock_guard criticalSection{KScheduler::GetCriticalSection()}; if (!IsDying()) { AddForcePauseReasonToField(reason); if (numKernelMutexWaiters == 0) { AdjustScheduling(CommitForcePauseToField()); } } } void KThread::RemoveForcePauseReason(KThread::ForcePauseReason reason) { std::lock_guard criticalSection{KScheduler::GetCriticalSection()}; if (!IsDying()) { RemoveForcePauseReasonToField(reason); if (!IsForcePaused() && numKernelMutexWaiters == 0) { AdjustScheduling(CommitForcePauseToField()); } } } bool KThread::WaitForKernelSync(KThread::WaitList &waitList) { // Has to be called from critical section currentWaitList = &waitList; Reschedule(SchedulingStatus::Paused); waitList.push_back(*this); if (IsDying()) { // Whoops ResumeFromKernelSync(); return false; } return true; } void KThread::ResumeFromKernelSync() { // Has to be called from critical section currentWaitList->erase(currentWaitList->iterator_to(*this)); currentWaitList = nullptr; Reschedule(SchedulingStatus::Running); } void KThread::ResumeAllFromKernelSync(KThread::WaitList &waitList) { // Has to be called from critical section waitList.clear_and_dispose( [](KThread *t) { t->currentWaitList = nullptr; t->Reschedule(SchedulingStatus::Running); } ); } void KThread::CancelKernelSync() { std::lock_guard criticalSection{KScheduler::GetCriticalSection()}; if (GetSchedulingStatus() == SchedulingStatus::Paused) { // Note: transparent to force-pause if (currentWaitList != nullptr) { ResumeFromKernelSync(); } else { Reschedule(SchedulingStatus::Running); } } } void KThread::CancelKernelSync(Result res) { syncResult = res; CancelKernelSync(); } void KThread::HandleSyncObjectSignaled(KSynchronizationObject *syncObj) { if (GetSchedulingStatus() == SchedulingStatus::Paused) { signaledSyncObject = syncObj; syncResult = ResultSuccess{}; Reschedule(SchedulingStatus::Running); } } void KThread::AddToMutexWaitList(KThread &thread) { // TODO: check&increment numKernelMutexWaiters // Ordered list insertion auto it = std::find_if( mutexWaitList.begin(), mutexWaitList.end(), [&thread](const KThread &t) { return t.GetPriority() > thread.GetPriority(); } ); if (it != mutexWaitList.end()) { mutexWaitList.insert(it, thread); } else { mutexWaitList.push_back(thread); } } KThread::MutexWaitList::iterator KThread::RemoveFromMutexWaitList(KThread::MutexWaitList::const_iterator it) { // TODO: check&decrement numKernelMutexWaiters return mutexWaitList.erase(it); } void KThread::RemoveFromMutexWaitList(const KThread &t) { RemoveFromMutexWaitList(mutexWaitList.iterator_to(t)); } void KThread::InheritDynamicPriority() { /* Do priority inheritance Since we're maybe changing the priority of the thread, we must go through the entire mutex owner chain. The invariant must be preserved: A thread holding a mutex must have a higher-or-same priority than all threads waiting for it to release the mutex. */ for (KThread *t = this; t != nullptr; t = t->wantedMutexOwner) { uint newPrio, oldPrio = priority; if (!mutexWaitList.empty() && mutexWaitList.front().priority < basePriority) { newPrio = mutexWaitList.front().priority; } else { newPrio = basePriority; } if (newPrio == oldPrio) { break; } else { // Update everything that depends on dynamic priority: // TODO update condvar // TODO update ctr arbiter priority = newPrio; // TODO update condvar // TODO update ctr arbiter if (CompareSchedulingStatusFull(SchedulingStatus::Running)) { KScheduler::Global::AdjustThreadPriorityChanged(*this, oldPrio, this == KCoreContext::GetCurrentInstance().GetCurrentThread()); } if (wantedMutexOwner != nullptr) { wantedMutexOwner->RemoveFromMutexWaitList(*this); wantedMutexOwner->AddToMutexWaitList(*this); } } } } void KThread::AddMutexWaiter(KThread &waiter) { AddToMutexWaitList(waiter); InheritDynamicPriority(); } void KThread::RemoveMutexWaiter(KThread &waiter) { RemoveFromMutexWaitList(waiter); InheritDynamicPriority(); } KThread *KThread::RelinquishMutex(size_t *count, uiptr mutexAddr) { KThread *newOwner = nullptr; *count = 0; // First in list wanting mutexAddr becomes owner, the rest is transferred for (auto it = mutexWaitList.begin(); it != mutexWaitList.end(); ) { if (it->wantedMutex != mutexAddr) { ++it; continue; } else { KThread &t = *it; ++(*count); it = RemoveFromMutexWaitList(it); if (newOwner == nullptr) { newOwner = &t; } else { newOwner->AddToMutexWaitList(t); } } } // Mutex waiters list have changed InheritDynamicPriority(); if (newOwner != nullptr) { newOwner->InheritDynamicPriority(); } return newOwner; } }