mirror of
https://git.suyu.dev/suyu/suyu.git
synced 2024-11-23 07:12:46 +01:00
kernel: fix issues with single core mode
This commit is contained in:
parent
0624c880bd
commit
21945ae127
9 changed files with 229 additions and 193 deletions
|
@ -42,14 +42,6 @@ void CpuManager::Shutdown() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void CpuManager::GuestActivateFunction() {
|
|
||||||
if (is_multicore) {
|
|
||||||
MultiCoreGuestActivate();
|
|
||||||
} else {
|
|
||||||
SingleCoreGuestActivate();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void CpuManager::GuestThreadFunction() {
|
void CpuManager::GuestThreadFunction() {
|
||||||
if (is_multicore) {
|
if (is_multicore) {
|
||||||
MultiCoreRunGuestThread();
|
MultiCoreRunGuestThread();
|
||||||
|
@ -58,21 +50,16 @@ void CpuManager::GuestThreadFunction() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void CpuManager::ShutdownThreadFunction() {
|
void CpuManager::IdleThreadFunction() {
|
||||||
ShutdownThread();
|
if (is_multicore) {
|
||||||
|
MultiCoreRunIdleThread();
|
||||||
|
} else {
|
||||||
|
SingleCoreRunIdleThread();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void CpuManager::WaitForAndHandleInterrupt() {
|
void CpuManager::ShutdownThreadFunction() {
|
||||||
auto& kernel = system.Kernel();
|
ShutdownThread();
|
||||||
auto& physical_core = kernel.CurrentPhysicalCore();
|
|
||||||
|
|
||||||
ASSERT(Kernel::GetCurrentThread(kernel).GetDisableDispatchCount() == 1);
|
|
||||||
|
|
||||||
if (!physical_core.IsInterrupted()) {
|
|
||||||
physical_core.Idle();
|
|
||||||
}
|
|
||||||
|
|
||||||
HandleInterrupt();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void CpuManager::HandleInterrupt() {
|
void CpuManager::HandleInterrupt() {
|
||||||
|
@ -86,26 +73,10 @@ void CpuManager::HandleInterrupt() {
|
||||||
/// MultiCore ///
|
/// MultiCore ///
|
||||||
///////////////////////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////////////////////
|
||||||
|
|
||||||
void CpuManager::MultiCoreGuestActivate() {
|
|
||||||
// Similar to the HorizonKernelMain callback in HOS
|
|
||||||
auto& kernel = system.Kernel();
|
|
||||||
auto* scheduler = kernel.CurrentScheduler();
|
|
||||||
|
|
||||||
scheduler->Activate();
|
|
||||||
UNREACHABLE();
|
|
||||||
}
|
|
||||||
|
|
||||||
void CpuManager::MultiCoreRunGuestThread() {
|
void CpuManager::MultiCoreRunGuestThread() {
|
||||||
// Similar to UserModeThreadStarter in HOS
|
// Similar to UserModeThreadStarter in HOS
|
||||||
auto& kernel = system.Kernel();
|
auto& kernel = system.Kernel();
|
||||||
auto* thread = kernel.GetCurrentEmuThread();
|
kernel.CurrentScheduler()->OnThreadStart();
|
||||||
thread->EnableDispatch();
|
|
||||||
|
|
||||||
MultiCoreRunGuestLoop();
|
|
||||||
}
|
|
||||||
|
|
||||||
void CpuManager::MultiCoreRunGuestLoop() {
|
|
||||||
auto& kernel = system.Kernel();
|
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
auto* physical_core = &kernel.CurrentPhysicalCore();
|
auto* physical_core = &kernel.CurrentPhysicalCore();
|
||||||
|
@ -118,17 +89,105 @@ void CpuManager::MultiCoreRunGuestLoop() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void CpuManager::MultiCoreRunIdleThread() {
|
||||||
|
// Not accurate to HOS. Remove this entire method when singlecore is removed.
|
||||||
|
// See notes in KScheduler::ScheduleImpl for more information about why this
|
||||||
|
// is inaccurate.
|
||||||
|
|
||||||
|
auto& kernel = system.Kernel();
|
||||||
|
kernel.CurrentScheduler()->OnThreadStart();
|
||||||
|
|
||||||
|
while (true) {
|
||||||
|
auto& physical_core = kernel.CurrentPhysicalCore();
|
||||||
|
if (!physical_core.IsInterrupted()) {
|
||||||
|
physical_core.Idle();
|
||||||
|
}
|
||||||
|
|
||||||
|
HandleInterrupt();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
///////////////////////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////////////////////
|
||||||
/// SingleCore ///
|
/// SingleCore ///
|
||||||
///////////////////////////////////////////////////////////////////////////////
|
///////////////////////////////////////////////////////////////////////////////
|
||||||
|
|
||||||
void CpuManager::SingleCoreGuestActivate() {}
|
void CpuManager::SingleCoreRunGuestThread() {
|
||||||
|
auto& kernel = system.Kernel();
|
||||||
|
kernel.CurrentScheduler()->OnThreadStart();
|
||||||
|
|
||||||
void CpuManager::SingleCoreRunGuestThread() {}
|
while (true) {
|
||||||
|
auto* physical_core = &kernel.CurrentPhysicalCore();
|
||||||
|
if (!physical_core->IsInterrupted()) {
|
||||||
|
physical_core->Run();
|
||||||
|
physical_core = &kernel.CurrentPhysicalCore();
|
||||||
|
}
|
||||||
|
|
||||||
void CpuManager::SingleCoreRunGuestLoop() {}
|
kernel.SetIsPhantomModeForSingleCore(true);
|
||||||
|
system.CoreTiming().Advance();
|
||||||
|
kernel.SetIsPhantomModeForSingleCore(false);
|
||||||
|
|
||||||
void CpuManager::PreemptSingleCore(bool from_running_enviroment) {}
|
PreemptSingleCore();
|
||||||
|
HandleInterrupt();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void CpuManager::SingleCoreRunIdleThread() {
|
||||||
|
auto& kernel = system.Kernel();
|
||||||
|
kernel.CurrentScheduler()->OnThreadStart();
|
||||||
|
|
||||||
|
while (true) {
|
||||||
|
PreemptSingleCore(false);
|
||||||
|
system.CoreTiming().AddTicks(1000U);
|
||||||
|
idle_count++;
|
||||||
|
HandleInterrupt();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void CpuManager::PreemptSingleCore(bool from_running_environment) {
|
||||||
|
{
|
||||||
|
auto& kernel = system.Kernel();
|
||||||
|
auto& scheduler = kernel.Scheduler(current_core);
|
||||||
|
|
||||||
|
Kernel::KThread* current_thread = scheduler.GetSchedulerCurrentThread();
|
||||||
|
if (idle_count >= 4 || from_running_environment) {
|
||||||
|
if (!from_running_environment) {
|
||||||
|
system.CoreTiming().Idle();
|
||||||
|
idle_count = 0;
|
||||||
|
}
|
||||||
|
kernel.SetIsPhantomModeForSingleCore(true);
|
||||||
|
system.CoreTiming().Advance();
|
||||||
|
kernel.SetIsPhantomModeForSingleCore(false);
|
||||||
|
}
|
||||||
|
current_core.store((current_core + 1) % Core::Hardware::NUM_CPU_CORES);
|
||||||
|
system.CoreTiming().ResetTicks();
|
||||||
|
scheduler.Unload(scheduler.GetSchedulerCurrentThread());
|
||||||
|
|
||||||
|
auto& next_scheduler = kernel.Scheduler(current_core);
|
||||||
|
|
||||||
|
// Disable dispatch. We're about to preempt this thread.
|
||||||
|
Kernel::KScopedDisableDispatch dd{kernel};
|
||||||
|
Common::Fiber::YieldTo(current_thread->GetHostContext(), *next_scheduler.GetSwitchFiber());
|
||||||
|
}
|
||||||
|
|
||||||
|
// We've now been scheduled again, and we may have exchanged schedulers.
|
||||||
|
// Reload the scheduler in case it's different.
|
||||||
|
{
|
||||||
|
auto& scheduler = system.Kernel().Scheduler(current_core);
|
||||||
|
scheduler.Reload(scheduler.GetSchedulerCurrentThread());
|
||||||
|
if (!scheduler.IsIdle()) {
|
||||||
|
idle_count = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void CpuManager::GuestActivate() {
|
||||||
|
// Similar to the HorizonKernelMain callback in HOS
|
||||||
|
auto& kernel = system.Kernel();
|
||||||
|
auto* scheduler = kernel.CurrentScheduler();
|
||||||
|
|
||||||
|
scheduler->Activate();
|
||||||
|
UNREACHABLE();
|
||||||
|
}
|
||||||
|
|
||||||
void CpuManager::ShutdownThread() {
|
void CpuManager::ShutdownThread() {
|
||||||
auto& kernel = system.Kernel();
|
auto& kernel = system.Kernel();
|
||||||
|
@ -168,20 +227,11 @@ void CpuManager::RunThread(std::size_t core) {
|
||||||
}
|
}
|
||||||
|
|
||||||
auto& kernel = system.Kernel();
|
auto& kernel = system.Kernel();
|
||||||
|
auto& scheduler = *kernel.CurrentScheduler();
|
||||||
|
auto* thread = scheduler.GetSchedulerCurrentThread();
|
||||||
|
Kernel::SetCurrentThread(kernel, thread);
|
||||||
|
|
||||||
auto* main_thread = Kernel::KThread::Create(kernel);
|
Common::Fiber::YieldTo(data.host_context, *thread->GetHostContext());
|
||||||
main_thread->SetName(fmt::format("MainThread:{}", core));
|
|
||||||
ASSERT(Kernel::KThread::InitializeMainThread(system, main_thread, static_cast<s32>(core))
|
|
||||||
.IsSuccess());
|
|
||||||
|
|
||||||
auto* idle_thread = Kernel::KThread::Create(kernel);
|
|
||||||
ASSERT(Kernel::KThread::InitializeIdleThread(system, idle_thread, static_cast<s32>(core))
|
|
||||||
.IsSuccess());
|
|
||||||
|
|
||||||
kernel.SetCurrentEmuThread(main_thread);
|
|
||||||
kernel.CurrentScheduler()->Initialize(idle_thread);
|
|
||||||
|
|
||||||
Common::Fiber::YieldTo(data.host_context, *main_thread->GetHostContext());
|
|
||||||
}
|
}
|
||||||
|
|
||||||
} // namespace Core
|
} // namespace Core
|
||||||
|
|
|
@ -48,12 +48,11 @@ public:
|
||||||
gpu_barrier->Sync();
|
gpu_barrier->Sync();
|
||||||
}
|
}
|
||||||
|
|
||||||
void WaitForAndHandleInterrupt();
|
|
||||||
void Initialize();
|
void Initialize();
|
||||||
void Shutdown();
|
void Shutdown();
|
||||||
|
|
||||||
std::function<void()> GetGuestActivateFunc() {
|
std::function<void()> GetGuestActivateFunc() {
|
||||||
return [this] { GuestActivateFunction(); };
|
return [this] { GuestActivate(); };
|
||||||
}
|
}
|
||||||
std::function<void()> GetGuestThreadFunc() {
|
std::function<void()> GetGuestThreadFunc() {
|
||||||
return [this] { GuestThreadFunction(); };
|
return [this] { GuestThreadFunction(); };
|
||||||
|
@ -72,21 +71,19 @@ public:
|
||||||
}
|
}
|
||||||
|
|
||||||
private:
|
private:
|
||||||
void GuestActivateFunction();
|
|
||||||
void GuestThreadFunction();
|
void GuestThreadFunction();
|
||||||
void IdleThreadFunction();
|
void IdleThreadFunction();
|
||||||
void ShutdownThreadFunction();
|
void ShutdownThreadFunction();
|
||||||
|
|
||||||
void MultiCoreGuestActivate();
|
|
||||||
void MultiCoreRunGuestThread();
|
void MultiCoreRunGuestThread();
|
||||||
void MultiCoreRunGuestLoop();
|
void MultiCoreRunIdleThread();
|
||||||
|
|
||||||
void SingleCoreGuestActivate();
|
|
||||||
void SingleCoreRunGuestThread();
|
void SingleCoreRunGuestThread();
|
||||||
void SingleCoreRunGuestLoop();
|
void SingleCoreRunIdleThread();
|
||||||
|
|
||||||
static void ThreadStart(std::stop_token stop_token, CpuManager& cpu_manager, std::size_t core);
|
static void ThreadStart(std::stop_token stop_token, CpuManager& cpu_manager, std::size_t core);
|
||||||
|
|
||||||
|
void GuestActivate();
|
||||||
void HandleInterrupt();
|
void HandleInterrupt();
|
||||||
void ShutdownThread();
|
void ShutdownThread();
|
||||||
void RunThread(std::size_t core);
|
void RunThread(std::size_t core);
|
||||||
|
|
|
@ -42,11 +42,6 @@ void GlobalSchedulerContext::PreemptThreads() {
|
||||||
for (u32 core_id = 0; core_id < Core::Hardware::NUM_CPU_CORES; core_id++) {
|
for (u32 core_id = 0; core_id < Core::Hardware::NUM_CPU_CORES; core_id++) {
|
||||||
const u32 priority = preemption_priorities[core_id];
|
const u32 priority = preemption_priorities[core_id];
|
||||||
KScheduler::RotateScheduledQueue(kernel, core_id, priority);
|
KScheduler::RotateScheduledQueue(kernel, core_id, priority);
|
||||||
|
|
||||||
// Signal an interrupt occurred. For core 3, this is a certainty, as preemption will result
|
|
||||||
// in the rotator thread being scheduled. For cores 0-2, this is to simulate or system
|
|
||||||
// interrupts that may have occurred.
|
|
||||||
kernel.PhysicalCore(core_id).Interrupt();
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -28,9 +28,9 @@ static void IncrementScheduledCount(Kernel::KThread* thread) {
|
||||||
}
|
}
|
||||||
|
|
||||||
KScheduler::KScheduler(KernelCore& kernel_) : kernel{kernel_} {
|
KScheduler::KScheduler(KernelCore& kernel_) : kernel{kernel_} {
|
||||||
m_idle_stack = std::make_shared<Common::Fiber>([this] {
|
m_switch_fiber = std::make_shared<Common::Fiber>([this] {
|
||||||
while (true) {
|
while (true) {
|
||||||
ScheduleImplOffStack();
|
ScheduleImplFiber();
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
|
@ -60,9 +60,9 @@ void KScheduler::DisableScheduling(KernelCore& kernel) {
|
||||||
void KScheduler::EnableScheduling(KernelCore& kernel, u64 cores_needing_scheduling) {
|
void KScheduler::EnableScheduling(KernelCore& kernel, u64 cores_needing_scheduling) {
|
||||||
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() >= 1);
|
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() >= 1);
|
||||||
|
|
||||||
auto* scheduler = kernel.CurrentScheduler();
|
auto* scheduler{kernel.CurrentScheduler()};
|
||||||
|
|
||||||
if (!scheduler) {
|
if (!scheduler || kernel.IsPhantomModeForSingleCore()) {
|
||||||
// HACK: we cannot schedule from this thread, it is not a core thread
|
// HACK: we cannot schedule from this thread, it is not a core thread
|
||||||
RescheduleCores(kernel, cores_needing_scheduling);
|
RescheduleCores(kernel, cores_needing_scheduling);
|
||||||
if (GetCurrentThread(kernel).GetDisableDispatchCount() == 1) {
|
if (GetCurrentThread(kernel).GetDisableDispatchCount() == 1) {
|
||||||
|
@ -125,9 +125,9 @@ void KScheduler::RescheduleCurrentCoreImpl() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void KScheduler::Initialize(KThread* idle_thread) {
|
void KScheduler::Initialize(KThread* main_thread, KThread* idle_thread, s32 core_id) {
|
||||||
// Set core ID/idle thread/interrupt task manager.
|
// Set core ID/idle thread/interrupt task manager.
|
||||||
m_core_id = GetCurrentCoreId(kernel);
|
m_core_id = core_id;
|
||||||
m_idle_thread = idle_thread;
|
m_idle_thread = idle_thread;
|
||||||
// m_state.idle_thread_stack = m_idle_thread->GetStackTop();
|
// m_state.idle_thread_stack = m_idle_thread->GetStackTop();
|
||||||
// m_state.interrupt_task_manager = &kernel.GetInterruptTaskManager();
|
// m_state.interrupt_task_manager = &kernel.GetInterruptTaskManager();
|
||||||
|
@ -142,10 +142,10 @@ void KScheduler::Initialize(KThread* idle_thread) {
|
||||||
// Bind interrupt handler.
|
// Bind interrupt handler.
|
||||||
// kernel.GetInterruptManager().BindHandler(
|
// kernel.GetInterruptManager().BindHandler(
|
||||||
// GetSchedulerInterruptHandler(kernel), KInterruptName::Scheduler, m_core_id,
|
// GetSchedulerInterruptHandler(kernel), KInterruptName::Scheduler, m_core_id,
|
||||||
// KInterruptController::PriorityLevel_Scheduler, false, false);
|
// KInterruptController::PriorityLevel::Scheduler, false, false);
|
||||||
|
|
||||||
// Set the current thread.
|
// Set the current thread.
|
||||||
m_current_thread = GetCurrentThreadPointer(kernel);
|
m_current_thread = main_thread;
|
||||||
}
|
}
|
||||||
|
|
||||||
void KScheduler::Activate() {
|
void KScheduler::Activate() {
|
||||||
|
@ -156,6 +156,10 @@ void KScheduler::Activate() {
|
||||||
RescheduleCurrentCore();
|
RescheduleCurrentCore();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void KScheduler::OnThreadStart() {
|
||||||
|
GetCurrentThread(kernel).EnableDispatch();
|
||||||
|
}
|
||||||
|
|
||||||
u64 KScheduler::UpdateHighestPriorityThread(KThread* highest_thread) {
|
u64 KScheduler::UpdateHighestPriorityThread(KThread* highest_thread) {
|
||||||
if (KThread* prev_highest_thread = m_state.highest_priority_thread;
|
if (KThread* prev_highest_thread = m_state.highest_priority_thread;
|
||||||
prev_highest_thread != highest_thread) [[likely]] {
|
prev_highest_thread != highest_thread) [[likely]] {
|
||||||
|
@ -372,37 +376,30 @@ void KScheduler::ScheduleImpl() {
|
||||||
}
|
}
|
||||||
|
|
||||||
// The highest priority thread is not the same as the current thread.
|
// The highest priority thread is not the same as the current thread.
|
||||||
// Switch to the idle thread stack and continue executing from there.
|
// Jump to the switcher and continue executing from there.
|
||||||
m_idle_cur_thread = cur_thread;
|
m_switch_cur_thread = cur_thread;
|
||||||
m_idle_highest_priority_thread = highest_priority_thread;
|
m_switch_highest_priority_thread = highest_priority_thread;
|
||||||
Common::Fiber::YieldTo(cur_thread->host_context, *m_idle_stack);
|
m_switch_from_schedule = true;
|
||||||
|
Common::Fiber::YieldTo(cur_thread->host_context, *m_switch_fiber);
|
||||||
|
|
||||||
// Returning from ScheduleImpl occurs after this thread has been scheduled again.
|
// Returning from ScheduleImpl occurs after this thread has been scheduled again.
|
||||||
}
|
}
|
||||||
|
|
||||||
void KScheduler::ScheduleImplOffStack() {
|
void KScheduler::ScheduleImplFiber() {
|
||||||
KThread* const cur_thread{m_idle_cur_thread};
|
KThread* const cur_thread{m_switch_cur_thread};
|
||||||
KThread* highest_priority_thread{m_idle_highest_priority_thread};
|
KThread* highest_priority_thread{m_switch_highest_priority_thread};
|
||||||
|
|
||||||
// Get a reference to the current thread's stack parameters.
|
// If we're not coming from scheduling (i.e., we came from SC preemption),
|
||||||
auto& sp{cur_thread->GetStackParameters()};
|
// we should restart the scheduling loop directly. Not accurate to HOS.
|
||||||
|
if (!m_switch_from_schedule) {
|
||||||
|
goto retry;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Mark that we are not coming from scheduling anymore.
|
||||||
|
m_switch_from_schedule = false;
|
||||||
|
|
||||||
// Save the original thread context.
|
// Save the original thread context.
|
||||||
{
|
Unload(cur_thread);
|
||||||
auto& physical_core = kernel.System().CurrentPhysicalCore();
|
|
||||||
auto& cpu_core = physical_core.ArmInterface();
|
|
||||||
cpu_core.SaveContext(cur_thread->GetContext32());
|
|
||||||
cpu_core.SaveContext(cur_thread->GetContext64());
|
|
||||||
// Save the TPIDR_EL0 system register in case it was modified.
|
|
||||||
cur_thread->SetTPIDR_EL0(cpu_core.GetTPIDR_EL0());
|
|
||||||
cpu_core.ClearExclusiveState();
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check if the thread is terminated by checking the DPC flags.
|
|
||||||
if ((sp.dpc_flags & static_cast<u32>(DpcFlag::Terminated)) == 0) {
|
|
||||||
// The thread isn't terminated, so we want to unlock it.
|
|
||||||
sp.m_lock.store(false, std::memory_order_seq_cst);
|
|
||||||
}
|
|
||||||
|
|
||||||
// The current thread's context has been entirely taken care of.
|
// The current thread's context has been entirely taken care of.
|
||||||
// Now we want to loop until we successfully switch the thread context.
|
// Now we want to loop until we successfully switch the thread context.
|
||||||
|
@ -411,62 +408,39 @@ void KScheduler::ScheduleImplOffStack() {
|
||||||
// Check if the highest priority thread is null.
|
// Check if the highest priority thread is null.
|
||||||
if (!highest_priority_thread) {
|
if (!highest_priority_thread) {
|
||||||
// The next thread is nullptr!
|
// The next thread is nullptr!
|
||||||
// Switch to nullptr. This will actually switch to the idle thread.
|
|
||||||
SwitchThread(nullptr);
|
|
||||||
|
|
||||||
// We've switched to the idle thread, so we want to process interrupt tasks until we
|
// Switch to the idle thread. Note: HOS treats idling as a special case for
|
||||||
// schedule a non-idle thread.
|
// performance. This is not *required* for yuzu's purposes, and for singlecore
|
||||||
while (!m_state.interrupt_task_runnable) {
|
// compatibility, we can just move the logic that would go here into the execution
|
||||||
// Check if we need scheduling.
|
// of the idle thread. If we ever remove singlecore, we should implement this
|
||||||
if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) {
|
// accurately to HOS.
|
||||||
goto retry;
|
highest_priority_thread = m_idle_thread;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Clear the previous thread.
|
// We want to try to lock the highest priority thread's context.
|
||||||
m_state.prev_thread = nullptr;
|
// Try to take it.
|
||||||
|
while (!highest_priority_thread->context_guard.try_lock()) {
|
||||||
// Wait for an interrupt before checking again.
|
// The highest priority thread's context is already locked.
|
||||||
kernel.System().GetCpuManager().WaitForAndHandleInterrupt();
|
// Check if we need scheduling. If we don't, we can retry directly.
|
||||||
|
if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) {
|
||||||
|
// If we do, another core is interfering, and we must start again.
|
||||||
|
goto retry;
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Execute any pending interrupt tasks.
|
// It's time to switch the thread.
|
||||||
// m_state.interrupt_task_manager->DoTasks();
|
// Switch to the highest priority thread.
|
||||||
|
SwitchThread(highest_priority_thread);
|
||||||
|
|
||||||
// Clear the interrupt task thread as runnable.
|
// Check if we need scheduling. If we do, then we can't complete the switch and should
|
||||||
m_state.interrupt_task_runnable = false;
|
// retry.
|
||||||
|
if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) {
|
||||||
// Retry the scheduling loop.
|
// Our switch failed.
|
||||||
|
// We should unlock the thread context, and then retry.
|
||||||
|
highest_priority_thread->context_guard.unlock();
|
||||||
goto retry;
|
goto retry;
|
||||||
} else {
|
} else {
|
||||||
// We want to try to lock the highest priority thread's context.
|
break;
|
||||||
// Try to take it.
|
|
||||||
bool expected{false};
|
|
||||||
while (!highest_priority_thread->stack_parameters.m_lock.compare_exchange_strong(
|
|
||||||
expected, true, std::memory_order_seq_cst)) {
|
|
||||||
// The highest priority thread's context is already locked.
|
|
||||||
// Check if we need scheduling. If we don't, we can retry directly.
|
|
||||||
if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) {
|
|
||||||
// If we do, another core is interfering, and we must start again.
|
|
||||||
goto retry;
|
|
||||||
}
|
|
||||||
expected = false;
|
|
||||||
}
|
|
||||||
|
|
||||||
// It's time to switch the thread.
|
|
||||||
// Switch to the highest priority thread.
|
|
||||||
SwitchThread(highest_priority_thread);
|
|
||||||
|
|
||||||
// Check if we need scheduling. If we do, then we can't complete the switch and should
|
|
||||||
// retry.
|
|
||||||
if (m_state.needs_scheduling.load(std::memory_order_seq_cst)) {
|
|
||||||
// Our switch failed.
|
|
||||||
// We should unlock the thread context, and then retry.
|
|
||||||
highest_priority_thread->stack_parameters.m_lock.store(false,
|
|
||||||
std::memory_order_seq_cst);
|
|
||||||
goto retry;
|
|
||||||
} else {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
retry:
|
retry:
|
||||||
|
@ -480,18 +454,35 @@ void KScheduler::ScheduleImplOffStack() {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Reload the guest thread context.
|
// Reload the guest thread context.
|
||||||
{
|
Reload(highest_priority_thread);
|
||||||
auto& cpu_core = kernel.System().CurrentArmInterface();
|
|
||||||
cpu_core.LoadContext(highest_priority_thread->GetContext32());
|
|
||||||
cpu_core.LoadContext(highest_priority_thread->GetContext64());
|
|
||||||
cpu_core.SetTlsAddress(highest_priority_thread->GetTLSAddress());
|
|
||||||
cpu_core.SetTPIDR_EL0(highest_priority_thread->GetTPIDR_EL0());
|
|
||||||
cpu_core.LoadWatchpointArray(highest_priority_thread->GetOwnerProcess()->GetWatchpoints());
|
|
||||||
cpu_core.ClearExclusiveState();
|
|
||||||
}
|
|
||||||
|
|
||||||
// Reload the host thread.
|
// Reload the host thread.
|
||||||
Common::Fiber::YieldTo(m_idle_stack, *highest_priority_thread->host_context);
|
Common::Fiber::YieldTo(m_switch_fiber, *highest_priority_thread->host_context);
|
||||||
|
}
|
||||||
|
|
||||||
|
void KScheduler::Unload(KThread* thread) {
|
||||||
|
auto& cpu_core = kernel.System().ArmInterface(m_core_id);
|
||||||
|
cpu_core.SaveContext(thread->GetContext32());
|
||||||
|
cpu_core.SaveContext(thread->GetContext64());
|
||||||
|
// Save the TPIDR_EL0 system register in case it was modified.
|
||||||
|
thread->SetTPIDR_EL0(cpu_core.GetTPIDR_EL0());
|
||||||
|
cpu_core.ClearExclusiveState();
|
||||||
|
|
||||||
|
// Check if the thread is terminated by checking the DPC flags.
|
||||||
|
if ((thread->GetStackParameters().dpc_flags & static_cast<u32>(DpcFlag::Terminated)) == 0) {
|
||||||
|
// The thread isn't terminated, so we want to unlock it.
|
||||||
|
thread->context_guard.unlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void KScheduler::Reload(KThread* thread) {
|
||||||
|
auto& cpu_core = kernel.System().ArmInterface(m_core_id);
|
||||||
|
cpu_core.LoadContext(thread->GetContext32());
|
||||||
|
cpu_core.LoadContext(thread->GetContext64());
|
||||||
|
cpu_core.SetTlsAddress(thread->GetTLSAddress());
|
||||||
|
cpu_core.SetTPIDR_EL0(thread->GetTPIDR_EL0());
|
||||||
|
cpu_core.LoadWatchpointArray(thread->GetOwnerProcess()->GetWatchpoints());
|
||||||
|
cpu_core.ClearExclusiveState();
|
||||||
}
|
}
|
||||||
|
|
||||||
void KScheduler::ClearPreviousThread(KernelCore& kernel, KThread* thread) {
|
void KScheduler::ClearPreviousThread(KernelCore& kernel, KThread* thread) {
|
||||||
|
|
|
@ -41,8 +41,11 @@ public:
|
||||||
explicit KScheduler(KernelCore& kernel);
|
explicit KScheduler(KernelCore& kernel);
|
||||||
~KScheduler();
|
~KScheduler();
|
||||||
|
|
||||||
void Initialize(KThread* idle_thread);
|
void Initialize(KThread* main_thread, KThread* idle_thread, s32 core_id);
|
||||||
void Activate();
|
void Activate();
|
||||||
|
void OnThreadStart();
|
||||||
|
void Unload(KThread* thread);
|
||||||
|
void Reload(KThread* thread);
|
||||||
|
|
||||||
void SetInterruptTaskRunnable();
|
void SetInterruptTaskRunnable();
|
||||||
void RequestScheduleOnInterrupt();
|
void RequestScheduleOnInterrupt();
|
||||||
|
@ -55,6 +58,14 @@ public:
|
||||||
return m_idle_thread;
|
return m_idle_thread;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
bool IsIdle() const {
|
||||||
|
return m_current_thread.load() == m_idle_thread;
|
||||||
|
}
|
||||||
|
|
||||||
|
std::shared_ptr<Common::Fiber> GetSwitchFiber() {
|
||||||
|
return m_switch_fiber;
|
||||||
|
}
|
||||||
|
|
||||||
KThread* GetPreviousThread() const {
|
KThread* GetPreviousThread() const {
|
||||||
return m_state.prev_thread;
|
return m_state.prev_thread;
|
||||||
}
|
}
|
||||||
|
@ -69,7 +80,7 @@ public:
|
||||||
|
|
||||||
// Static public API.
|
// Static public API.
|
||||||
static bool CanSchedule(KernelCore& kernel) {
|
static bool CanSchedule(KernelCore& kernel) {
|
||||||
return kernel.GetCurrentEmuThread()->GetDisableDispatchCount() == 0;
|
return GetCurrentThread(kernel).GetDisableDispatchCount() == 0;
|
||||||
}
|
}
|
||||||
static bool IsSchedulerLockedByCurrentThread(KernelCore& kernel) {
|
static bool IsSchedulerLockedByCurrentThread(KernelCore& kernel) {
|
||||||
return kernel.GlobalSchedulerContext().scheduler_lock.IsLockedByCurrentThread();
|
return kernel.GlobalSchedulerContext().scheduler_lock.IsLockedByCurrentThread();
|
||||||
|
@ -113,7 +124,7 @@ private:
|
||||||
|
|
||||||
// Instanced private API.
|
// Instanced private API.
|
||||||
void ScheduleImpl();
|
void ScheduleImpl();
|
||||||
void ScheduleImplOffStack();
|
void ScheduleImplFiber();
|
||||||
void SwitchThread(KThread* next_thread);
|
void SwitchThread(KThread* next_thread);
|
||||||
|
|
||||||
void Schedule();
|
void Schedule();
|
||||||
|
@ -147,9 +158,10 @@ private:
|
||||||
KThread* m_idle_thread{nullptr};
|
KThread* m_idle_thread{nullptr};
|
||||||
std::atomic<KThread*> m_current_thread{nullptr};
|
std::atomic<KThread*> m_current_thread{nullptr};
|
||||||
|
|
||||||
std::shared_ptr<Common::Fiber> m_idle_stack{};
|
std::shared_ptr<Common::Fiber> m_switch_fiber{};
|
||||||
KThread* m_idle_cur_thread{};
|
KThread* m_switch_cur_thread{};
|
||||||
KThread* m_idle_highest_priority_thread{};
|
KThread* m_switch_highest_priority_thread{};
|
||||||
|
bool m_switch_from_schedule{};
|
||||||
};
|
};
|
||||||
|
|
||||||
class KScopedSchedulerLock : public KScopedLock<KScheduler::LockType> {
|
class KScopedSchedulerLock : public KScopedLock<KScheduler::LockType> {
|
||||||
|
|
|
@ -268,7 +268,7 @@ Result KThread::InitializeMainThread(Core::System& system, KThread* thread, s32
|
||||||
|
|
||||||
Result KThread::InitializeIdleThread(Core::System& system, KThread* thread, s32 virt_core) {
|
Result KThread::InitializeIdleThread(Core::System& system, KThread* thread, s32 virt_core) {
|
||||||
return InitializeThread(thread, {}, {}, {}, IdleThreadPriority, virt_core, {}, ThreadType::Main,
|
return InitializeThread(thread, {}, {}, {}, IdleThreadPriority, virt_core, {}, ThreadType::Main,
|
||||||
abort);
|
system.GetCpuManager().GetIdleThreadStartFunc());
|
||||||
}
|
}
|
||||||
|
|
||||||
Result KThread::InitializeHighPriorityThread(Core::System& system, KThread* thread,
|
Result KThread::InitializeHighPriorityThread(Core::System& system, KThread* thread,
|
||||||
|
@ -1204,8 +1204,9 @@ KScopedDisableDispatch::~KScopedDisableDispatch() {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Skip the reschedule if single-core, as dispatch tracking is disabled here.
|
// Skip the reschedule if single-core.
|
||||||
if (!Settings::values.use_multi_core.GetValue()) {
|
if (!Settings::values.use_multi_core.GetValue()) {
|
||||||
|
GetCurrentThread(kernel).EnableDispatch();
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -439,7 +439,6 @@ public:
|
||||||
bool is_pinned;
|
bool is_pinned;
|
||||||
s32 disable_count;
|
s32 disable_count;
|
||||||
KThread* cur_thread;
|
KThread* cur_thread;
|
||||||
std::atomic<bool> m_lock;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
[[nodiscard]] StackParameters& GetStackParameters() {
|
[[nodiscard]] StackParameters& GetStackParameters() {
|
||||||
|
@ -485,39 +484,16 @@ public:
|
||||||
return per_core_priority_queue_entry[core];
|
return per_core_priority_queue_entry[core];
|
||||||
}
|
}
|
||||||
|
|
||||||
[[nodiscard]] bool IsKernelThread() const {
|
|
||||||
return GetActiveCore() == 3;
|
|
||||||
}
|
|
||||||
|
|
||||||
[[nodiscard]] bool IsDispatchTrackingDisabled() const {
|
|
||||||
return is_single_core || IsKernelThread();
|
|
||||||
}
|
|
||||||
|
|
||||||
[[nodiscard]] s32 GetDisableDispatchCount() const {
|
[[nodiscard]] s32 GetDisableDispatchCount() const {
|
||||||
if (IsDispatchTrackingDisabled()) {
|
|
||||||
// TODO(bunnei): Until kernel threads are emulated, we cannot enable/disable dispatch.
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
return this->GetStackParameters().disable_count;
|
return this->GetStackParameters().disable_count;
|
||||||
}
|
}
|
||||||
|
|
||||||
void DisableDispatch() {
|
void DisableDispatch() {
|
||||||
if (IsDispatchTrackingDisabled()) {
|
|
||||||
// TODO(bunnei): Until kernel threads are emulated, we cannot enable/disable dispatch.
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() >= 0);
|
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() >= 0);
|
||||||
this->GetStackParameters().disable_count++;
|
this->GetStackParameters().disable_count++;
|
||||||
}
|
}
|
||||||
|
|
||||||
void EnableDispatch() {
|
void EnableDispatch() {
|
||||||
if (IsDispatchTrackingDisabled()) {
|
|
||||||
// TODO(bunnei): Until kernel threads are emulated, we cannot enable/disable dispatch.
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() > 0);
|
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() > 0);
|
||||||
this->GetStackParameters().disable_count--;
|
this->GetStackParameters().disable_count--;
|
||||||
}
|
}
|
||||||
|
|
|
@ -64,8 +64,6 @@ struct KernelCore::Impl {
|
||||||
|
|
||||||
is_phantom_mode_for_singlecore = false;
|
is_phantom_mode_for_singlecore = false;
|
||||||
|
|
||||||
InitializePhysicalCores();
|
|
||||||
|
|
||||||
// Derive the initial memory layout from the emulated board
|
// Derive the initial memory layout from the emulated board
|
||||||
Init::InitializeSlabResourceCounts(kernel);
|
Init::InitializeSlabResourceCounts(kernel);
|
||||||
DeriveInitialMemoryLayout();
|
DeriveInitialMemoryLayout();
|
||||||
|
@ -77,6 +75,7 @@ struct KernelCore::Impl {
|
||||||
Init::InitializeKPageBufferSlabHeap(system);
|
Init::InitializeKPageBufferSlabHeap(system);
|
||||||
InitializeShutdownThreads();
|
InitializeShutdownThreads();
|
||||||
InitializePreemption(kernel);
|
InitializePreemption(kernel);
|
||||||
|
InitializePhysicalCores();
|
||||||
|
|
||||||
RegisterHostThread();
|
RegisterHostThread();
|
||||||
}
|
}
|
||||||
|
@ -193,8 +192,21 @@ struct KernelCore::Impl {
|
||||||
exclusive_monitor =
|
exclusive_monitor =
|
||||||
Core::MakeExclusiveMonitor(system.Memory(), Core::Hardware::NUM_CPU_CORES);
|
Core::MakeExclusiveMonitor(system.Memory(), Core::Hardware::NUM_CPU_CORES);
|
||||||
for (u32 i = 0; i < Core::Hardware::NUM_CPU_CORES; i++) {
|
for (u32 i = 0; i < Core::Hardware::NUM_CPU_CORES; i++) {
|
||||||
|
const s32 core{static_cast<s32>(i)};
|
||||||
|
|
||||||
schedulers[i] = std::make_unique<Kernel::KScheduler>(system.Kernel());
|
schedulers[i] = std::make_unique<Kernel::KScheduler>(system.Kernel());
|
||||||
cores.emplace_back(i, system, *schedulers[i], interrupts);
|
cores.emplace_back(i, system, *schedulers[i], interrupts);
|
||||||
|
|
||||||
|
auto* main_thread{Kernel::KThread::Create(system.Kernel())};
|
||||||
|
main_thread->SetName(fmt::format("MainThread:{}", core));
|
||||||
|
main_thread->SetCurrentCore(core);
|
||||||
|
ASSERT(Kernel::KThread::InitializeMainThread(system, main_thread, core).IsSuccess());
|
||||||
|
|
||||||
|
auto* idle_thread{Kernel::KThread::Create(system.Kernel())};
|
||||||
|
idle_thread->SetCurrentCore(core);
|
||||||
|
ASSERT(Kernel::KThread::InitializeIdleThread(system, idle_thread, core).IsSuccess());
|
||||||
|
|
||||||
|
schedulers[i]->Initialize(main_thread, idle_thread, core);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1093,10 +1105,11 @@ void KernelCore::Suspend(bool suspended) {
|
||||||
}
|
}
|
||||||
|
|
||||||
void KernelCore::ShutdownCores() {
|
void KernelCore::ShutdownCores() {
|
||||||
|
KScopedSchedulerLock lk{*this};
|
||||||
|
|
||||||
for (auto* thread : impl->shutdown_threads) {
|
for (auto* thread : impl->shutdown_threads) {
|
||||||
void(thread->Run());
|
void(thread->Run());
|
||||||
}
|
}
|
||||||
InterruptAllPhysicalCores();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
bool KernelCore::IsMulticore() const {
|
bool KernelCore::IsMulticore() const {
|
||||||
|
|
|
@ -43,6 +43,7 @@ void PhysicalCore::Initialize([[maybe_unused]] bool is_64_bit) {
|
||||||
|
|
||||||
void PhysicalCore::Run() {
|
void PhysicalCore::Run() {
|
||||||
arm_interface->Run();
|
arm_interface->Run();
|
||||||
|
arm_interface->ClearExclusiveState();
|
||||||
}
|
}
|
||||||
|
|
||||||
void PhysicalCore::Idle() {
|
void PhysicalCore::Idle() {
|
||||||
|
|
Loading…
Reference in a new issue