2018-11-22 06:27:23 +00:00
|
|
|
// Copyright 2018 yuzu emulator team
|
|
|
|
// Licensed under GPLv2 or any later version
|
|
|
|
// Refer to the license.txt file included.
|
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
#include "common/fiber.h"
|
|
|
|
#include "common/thread.h"
|
2018-11-22 06:27:23 +00:00
|
|
|
#include "core/arm/exclusive_monitor.h"
|
|
|
|
#include "core/core.h"
|
2019-09-10 01:37:29 +00:00
|
|
|
#include "core/core_timing.h"
|
2020-01-26 18:07:22 +00:00
|
|
|
#include "core/cpu_manager.h"
|
2018-11-22 06:27:23 +00:00
|
|
|
#include "core/gdbstub/gdbstub.h"
|
2020-02-25 02:04:12 +00:00
|
|
|
#include "core/hle/kernel/kernel.h"
|
|
|
|
#include "core/hle/kernel/physical_core.h"
|
|
|
|
#include "core/hle/kernel/scheduler.h"
|
|
|
|
#include "core/hle/kernel/thread.h"
|
2018-11-22 06:27:23 +00:00
|
|
|
|
|
|
|
namespace Core {
|
|
|
|
|
2020-01-26 18:07:22 +00:00
|
|
|
CpuManager::CpuManager(System& system) : system{system} {}
|
|
|
|
CpuManager::~CpuManager() = default;
|
2018-11-22 06:27:23 +00:00
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
void CpuManager::ThreadStart(CpuManager& cpu_manager, std::size_t core) {
|
|
|
|
cpu_manager.RunThread(core);
|
|
|
|
}
|
|
|
|
|
2020-01-26 18:07:22 +00:00
|
|
|
void CpuManager::Initialize() {
|
2020-02-25 02:04:12 +00:00
|
|
|
running_mode = true;
|
|
|
|
for (std::size_t core = 0; core < Core::Hardware::NUM_CPU_CORES; core++) {
|
|
|
|
core_data[core].host_thread =
|
|
|
|
std::make_unique<std::thread>(ThreadStart, std::ref(*this), core);
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-01-26 18:07:22 +00:00
|
|
|
void CpuManager::Shutdown() {
|
2020-02-25 02:04:12 +00:00
|
|
|
running_mode = false;
|
|
|
|
Pause(false);
|
|
|
|
for (std::size_t core = 0; core < Core::Hardware::NUM_CPU_CORES; core++) {
|
|
|
|
core_data[core].host_thread->join();
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
void CpuManager::GuestThreadFunction(void* cpu_manager_) {
|
|
|
|
CpuManager* cpu_manager = static_cast<CpuManager*>(cpu_manager_);
|
|
|
|
cpu_manager->RunGuestThread();
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
void CpuManager::IdleThreadFunction(void* cpu_manager_) {
|
|
|
|
CpuManager* cpu_manager = static_cast<CpuManager*>(cpu_manager_);
|
|
|
|
cpu_manager->RunIdleThread();
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
void CpuManager::SuspendThreadFunction(void* cpu_manager_) {
|
|
|
|
CpuManager* cpu_manager = static_cast<CpuManager*>(cpu_manager_);
|
|
|
|
cpu_manager->RunSuspendThread();
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
std::function<void(void*)> CpuManager::GetGuestThreadStartFunc() {
|
|
|
|
return std::function<void(void*)>(GuestThreadFunction);
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
std::function<void(void*)> CpuManager::GetIdleThreadStartFunc() {
|
|
|
|
return std::function<void(void*)>(IdleThreadFunction);
|
|
|
|
}
|
|
|
|
|
|
|
|
std::function<void(void*)> CpuManager::GetSuspendThreadStartFunc() {
|
|
|
|
return std::function<void(void*)>(SuspendThreadFunction);
|
|
|
|
}
|
|
|
|
|
|
|
|
void* CpuManager::GetStartFuncParamater() {
|
|
|
|
return static_cast<void*>(this);
|
|
|
|
}
|
|
|
|
|
|
|
|
void CpuManager::RunGuestThread() {
|
|
|
|
auto& kernel = system.Kernel();
|
|
|
|
{
|
|
|
|
auto& sched = kernel.CurrentScheduler();
|
|
|
|
sched.OnThreadStart();
|
|
|
|
}
|
|
|
|
while (true) {
|
|
|
|
auto& physical_core = kernel.CurrentPhysicalCore();
|
|
|
|
LOG_CRITICAL(Core_ARM, "Running Guest Thread");
|
|
|
|
physical_core.Idle();
|
|
|
|
LOG_CRITICAL(Core_ARM, "Leaving Guest Thread");
|
|
|
|
// physical_core.Run();
|
|
|
|
auto& scheduler = physical_core.Scheduler();
|
|
|
|
scheduler.TryDoContextSwitch();
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
2020-02-25 02:04:12 +00:00
|
|
|
}
|
2018-11-22 06:27:23 +00:00
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
void CpuManager::RunIdleThread() {
|
|
|
|
auto& kernel = system.Kernel();
|
|
|
|
while (true) {
|
|
|
|
auto& physical_core = kernel.CurrentPhysicalCore();
|
|
|
|
LOG_CRITICAL(Core_ARM, "Running Idle Thread");
|
|
|
|
physical_core.Idle();
|
|
|
|
auto& scheduler = physical_core.Scheduler();
|
|
|
|
scheduler.TryDoContextSwitch();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void CpuManager::RunSuspendThread() {
|
|
|
|
LOG_CRITICAL(Core_ARM, "Suspending Thread Entered");
|
|
|
|
auto& kernel = system.Kernel();
|
|
|
|
{
|
|
|
|
auto& sched = kernel.CurrentScheduler();
|
|
|
|
sched.OnThreadStart();
|
|
|
|
}
|
|
|
|
while (true) {
|
|
|
|
auto core = kernel.GetCurrentHostThreadID();
|
|
|
|
auto& scheduler = kernel.CurrentScheduler();
|
|
|
|
Kernel::Thread* current_thread = scheduler.GetCurrentThread();
|
|
|
|
LOG_CRITICAL(Core_ARM, "Suspending Core {}", core);
|
|
|
|
Common::Fiber::YieldTo(current_thread->GetHostContext(), core_data[core].host_context);
|
|
|
|
LOG_CRITICAL(Core_ARM, "Unsuspending Core {}", core);
|
|
|
|
ASSERT(scheduler.ContextSwitchPending());
|
|
|
|
ASSERT(core == kernel.GetCurrentHostThreadID());
|
|
|
|
scheduler.TryDoContextSwitch();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void CpuManager::Pause(bool paused) {
|
|
|
|
if (!paused) {
|
|
|
|
bool all_not_barrier = false;
|
|
|
|
while (!all_not_barrier) {
|
|
|
|
all_not_barrier = true;
|
|
|
|
for (std::size_t core = 0; core < Core::Hardware::NUM_CPU_CORES; core++) {
|
|
|
|
all_not_barrier &=
|
|
|
|
!core_data[core].is_running.load() && core_data[core].initialized.load();
|
2019-09-10 01:37:29 +00:00
|
|
|
}
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
2020-02-25 02:04:12 +00:00
|
|
|
for (std::size_t core = 0; core < Core::Hardware::NUM_CPU_CORES; core++) {
|
|
|
|
core_data[core].enter_barrier->Set();
|
|
|
|
}
|
|
|
|
if (paused_state.load()) {
|
|
|
|
bool all_barrier = false;
|
|
|
|
while (!all_barrier) {
|
|
|
|
all_barrier = true;
|
|
|
|
for (std::size_t core = 0; core < Core::Hardware::NUM_CPU_CORES; core++) {
|
|
|
|
all_barrier &=
|
|
|
|
core_data[core].is_paused.load() && core_data[core].initialized.load();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for (std::size_t core = 0; core < Core::Hardware::NUM_CPU_CORES; core++) {
|
|
|
|
core_data[core].exit_barrier->Set();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
/// Wait until all cores are paused.
|
|
|
|
bool all_barrier = false;
|
|
|
|
while (!all_barrier) {
|
|
|
|
all_barrier = true;
|
|
|
|
for (std::size_t core = 0; core < Core::Hardware::NUM_CPU_CORES; core++) {
|
|
|
|
all_barrier &=
|
|
|
|
core_data[core].is_paused.load() && core_data[core].initialized.load();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/// Don't release the barrier
|
|
|
|
}
|
|
|
|
paused_state = paused;
|
|
|
|
}
|
2018-11-22 06:27:23 +00:00
|
|
|
|
2020-02-25 02:04:12 +00:00
|
|
|
void CpuManager::RunThread(std::size_t core) {
|
|
|
|
/// Initialization
|
|
|
|
system.RegisterCoreThread(core);
|
|
|
|
std::string name = "yuzu:CoreHostThread_" + std::to_string(core);
|
|
|
|
Common::SetCurrentThreadName(name.c_str());
|
|
|
|
auto& data = core_data[core];
|
|
|
|
data.enter_barrier = std::make_unique<Common::Event>();
|
|
|
|
data.exit_barrier = std::make_unique<Common::Event>();
|
|
|
|
data.host_context = Common::Fiber::ThreadToFiber();
|
|
|
|
data.is_running = false;
|
|
|
|
data.initialized = true;
|
|
|
|
/// Running
|
|
|
|
while (running_mode) {
|
|
|
|
data.is_running = false;
|
|
|
|
data.enter_barrier->Wait();
|
|
|
|
auto& scheduler = system.Kernel().CurrentScheduler();
|
|
|
|
Kernel::Thread* current_thread = scheduler.GetCurrentThread();
|
|
|
|
data.is_running = true;
|
|
|
|
Common::Fiber::YieldTo(data.host_context, current_thread->GetHostContext());
|
|
|
|
data.is_running = false;
|
|
|
|
data.is_paused = true;
|
|
|
|
data.exit_barrier->Wait();
|
|
|
|
data.is_paused = false;
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
2020-02-25 02:04:12 +00:00
|
|
|
/// Time to cleanup
|
|
|
|
data.host_context->Exit();
|
|
|
|
data.enter_barrier.reset();
|
|
|
|
data.exit_barrier.reset();
|
|
|
|
data.initialized = false;
|
2018-11-22 06:27:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
} // namespace Core
|