hle: kernel: Refactor several threads/events/sharedmemory to use slab heaps.

This commit is contained in:
bunnei 2021-04-09 22:42:23 -07:00
parent b6156e735c
commit 89edbe8aa2
11 changed files with 53 additions and 59 deletions

View file

@ -618,14 +618,17 @@ KScheduler::KScheduler(Core::System& system, s32 core_id) : system(system), core
}
KScheduler::~KScheduler() {
idle_thread->Close();
if (idle_thread) {
idle_thread->Close();
idle_thread = nullptr;
}
}
KThread* KScheduler::GetCurrentThread() const {
if (auto result = current_thread.load(); result) {
return result;
}
return idle_thread.get();
return idle_thread;
}
u64 KScheduler::GetLastContextSwitchTicks() const {
@ -710,7 +713,7 @@ void KScheduler::ScheduleImpl() {
// We never want to schedule a null thread, so use the idle thread if we don't have a next.
if (next_thread == nullptr) {
next_thread = idle_thread.get();
next_thread = idle_thread;
}
// If we're not actually switching thread, there's nothing to do.
@ -771,7 +774,7 @@ void KScheduler::SwitchToCurrent() {
break;
}
}
auto thread = next_thread ? next_thread : idle_thread.get();
auto thread = next_thread ? next_thread : idle_thread;
Common::Fiber::YieldTo(switch_fiber, *thread->GetHostContext());
} while (!is_switch_pending());
}
@ -794,9 +797,8 @@ void KScheduler::UpdateLastContextSwitchTime(KThread* thread, Process* process)
}
void KScheduler::Initialize() {
idle_thread = std::make_unique<KThread>(system.Kernel());
KAutoObject::Create(idle_thread.get());
ASSERT(KThread::InitializeIdleThread(system, idle_thread.get(), core_id).IsSuccess());
idle_thread = KThread::Create(system.Kernel());
ASSERT(KThread::InitializeIdleThread(system, idle_thread, core_id).IsSuccess());
idle_thread->SetName(fmt::format("IdleThread:{}", core_id));
}

View file

@ -51,7 +51,7 @@ public:
/// Returns true if the scheduler is idle
[[nodiscard]] bool IsIdle() const {
return GetCurrentThread() == idle_thread.get();
return GetCurrentThread() == idle_thread;
}
/// Gets the timestamp for the last context switch in ticks.
@ -173,7 +173,7 @@ private:
KThread* prev_thread{};
std::atomic<KThread*> current_thread{};
std::unique_ptr<KThread> idle_thread;
KThread* idle_thread{};
std::shared_ptr<Common::Fiber> switch_fiber{};

View file

@ -102,15 +102,21 @@ struct KernelCore::Impl {
next_user_process_id = Process::ProcessIDMin;
next_thread_id = 1;
for (std::size_t i = 0; i < Core::Hardware::NUM_CPU_CORES; i++) {
if (suspend_threads[i]) {
suspend_threads[i]->Close();
for (s32 core_id = 0; core_id < Core::Hardware::NUM_CPU_CORES; core_id++) {
if (suspend_threads[core_id]) {
suspend_threads[core_id]->Close();
suspend_threads[core_id] = nullptr;
}
schedulers[core_id].reset();
}
cores.clear();
current_process = nullptr;
if (current_process) {
current_process->Close();
current_process = nullptr;
}
global_handle_table.Clear();
@ -195,10 +201,9 @@ struct KernelCore::Impl {
void InitializeSuspendThreads() {
for (s32 core_id = 0; core_id < Core::Hardware::NUM_CPU_CORES; core_id++) {
suspend_threads[core_id] = std::make_unique<KThread>(system.Kernel());
KAutoObject::Create(suspend_threads[core_id].get());
ASSERT(KThread::InitializeHighPriorityThread(system, suspend_threads[core_id].get(), {},
{}, core_id)
suspend_threads[core_id] = KThread::Create(system.Kernel());
ASSERT(KThread::InitializeHighPriorityThread(system, suspend_threads[core_id], {}, {},
core_id)
.IsSuccess());
suspend_threads[core_id]->SetName(fmt::format("SuspendThread:{}", core_id));
}
@ -577,15 +582,10 @@ struct KernelCore::Impl {
const PAddr irs_phys_addr{system_pool.GetAddress() + hid_size + font_size};
const PAddr time_phys_addr{system_pool.GetAddress() + hid_size + font_size + irs_size};
hid_shared_mem = std::make_unique<KSharedMemory>(system.Kernel());
font_shared_mem = std::make_unique<KSharedMemory>(system.Kernel());
irs_shared_mem = std::make_unique<KSharedMemory>(system.Kernel());
time_shared_mem = std::make_unique<KSharedMemory>(system.Kernel());
KAutoObject::Create(hid_shared_mem.get());
KAutoObject::Create(font_shared_mem.get());
KAutoObject::Create(irs_shared_mem.get());
KAutoObject::Create(time_shared_mem.get());
hid_shared_mem = KSharedMemory::Create(system.Kernel());
font_shared_mem = KSharedMemory::Create(system.Kernel());
irs_shared_mem = KSharedMemory::Create(system.Kernel());
time_shared_mem = KSharedMemory::Create(system.Kernel());
hid_shared_mem->Initialize(system.Kernel(), system.DeviceMemory(), nullptr,
{hid_phys_addr, hid_size / PageSize}, KMemoryPermission::None,
@ -656,10 +656,10 @@ struct KernelCore::Impl {
std::unique_ptr<KSlabHeap<Page>> user_slab_heap_pages;
// Shared memory for services
std::unique_ptr<Kernel::KSharedMemory> hid_shared_mem;
std::unique_ptr<Kernel::KSharedMemory> font_shared_mem;
std::unique_ptr<Kernel::KSharedMemory> irs_shared_mem;
std::unique_ptr<Kernel::KSharedMemory> time_shared_mem;
Kernel::KSharedMemory* hid_shared_mem{};
Kernel::KSharedMemory* font_shared_mem{};
Kernel::KSharedMemory* irs_shared_mem{};
Kernel::KSharedMemory* time_shared_mem{};
// Threads used for services
std::unordered_set<std::shared_ptr<Kernel::ServiceThread>> service_threads;
@ -668,7 +668,7 @@ struct KernelCore::Impl {
// the release of itself
std::unique_ptr<Common::ThreadWorker> service_thread_manager;
std::array<std::unique_ptr<KThread>, Core::Hardware::NUM_CPU_CORES> suspend_threads;
std::array<KThread*, Core::Hardware::NUM_CPU_CORES> suspend_threads;
std::array<Core::CPUInterruptHandler, Core::Hardware::NUM_CPU_CORES> interrupts{};
std::array<std::unique_ptr<Kernel::KScheduler>, Core::Hardware::NUM_CPU_CORES> schedulers{};
@ -938,9 +938,9 @@ void KernelCore::Suspend(bool in_suspention) {
{
KScopedSchedulerLock lock(*this);
const auto state = should_suspend ? ThreadState::Runnable : ThreadState::Waiting;
for (std::size_t i = 0; i < Core::Hardware::NUM_CPU_CORES; i++) {
impl->suspend_threads[i]->SetState(state);
impl->suspend_threads[i]->SetWaitReasonForDebugging(
for (s32 core_id = 0; core_id < Core::Hardware::NUM_CPU_CORES; core_id++) {
impl->suspend_threads[core_id]->SetState(state);
impl->suspend_threads[core_id]->SetWaitReasonForDebugging(
ThreadWaitReasonForDebugging::Suspended);
}
}

View file

@ -41,7 +41,7 @@ void SetupMainThread(Core::System& system, Process& owner_process, u32 priority,
const VAddr entry_point = owner_process.PageTable().GetCodeRegionStart();
ASSERT(owner_process.GetResourceLimit()->Reserve(LimitableResource::Threads, 1));
KThread* thread = KThread::CreateWithKernel(system.Kernel());
KThread* thread = KThread::Create(system.Kernel());
ASSERT(KThread::InitializeUserThread(system, thread, entry_point, 0, stack_top, priority,
owner_process.GetIdealCoreId(), &owner_process)
.IsSuccess());

View file

@ -67,10 +67,6 @@ class KAutoObjectWithSlabHeapAndContainer : public Base {
private:
static Derived* Allocate(KernelCore& kernel) {
return kernel.SlabHeap<Derived>().Allocate();
}
static Derived* AllocateWithKernel(KernelCore& kernel) {
return kernel.SlabHeap<Derived>().AllocateWithKernel(kernel);
}
@ -120,16 +116,8 @@ public:
kernel.ObjectListContainer().Initialize();
}
static Derived* Create() {
Derived* obj = Allocate();
if (obj != nullptr) {
KAutoObject::Create(obj);
}
return obj;
}
static Derived* CreateWithKernel(KernelCore& kernel) {
Derived* obj = AllocateWithKernel(kernel);
static Derived* Create(KernelCore& kernel) {
Derived* obj = Allocate(kernel);
if (obj != nullptr) {
KAutoObject::Create(obj);
}

View file

@ -1431,7 +1431,7 @@ static ResultCode CreateThread(Core::System& system, Handle* out_handle, VAddr e
}
// Create the thread.
KThread* thread = KThread::CreateWithKernel(kernel);
KThread* thread = KThread::Create(kernel);
if (!thread) {
LOG_ERROR(Kernel_SVC, "Unable to create new threads. Thread creation limit reached.");
return ResultOutOfResource;
@ -1953,7 +1953,7 @@ static ResultCode CreateEvent(Core::System& system, Handle* out_write, Handle* o
HandleTable& handle_table = kernel.CurrentProcess()->GetHandleTable();
// Create a new event.
KEvent* event = KEvent::CreateWithKernel(kernel);
KEvent* event = KEvent::Create(kernel);
R_UNLESS(event != nullptr, ResultOutOfResource);
// Initialize the event.