mirror of
https://github.com/RPCSX/rpcsx.git
synced 2025-12-06 07:12:14 +01:00
Implement helper functions balanced_wait_until and balanced_awaken They include new path for Windows 8.1+ (WaitOnAddress) shared_mutex, cond_variable, cond_one, cond_x16 modified to use it Added helper function utils::popcnt16 Replace most semaphore<> with shared_mutex
231 lines
4.6 KiB
C++
231 lines
4.6 KiB
C++
#pragma once
|
|
|
|
#include "Utilities/mutex.h"
|
|
#include "Utilities/sema.h"
|
|
#include "Utilities/cond.h"
|
|
|
|
#include "Emu/Memory/vm.h"
|
|
#include "Emu/CPU/CPUThread.h"
|
|
#include "Emu/Cell/ErrorCodes.h"
|
|
#include "Emu/IdManager.h"
|
|
#include "Emu/IPC.h"
|
|
|
|
#include <deque>
|
|
|
|
// attr_protocol (waiting scheduling policy)
|
|
enum
|
|
{
|
|
SYS_SYNC_FIFO = 0x1, // First In, First Out Order
|
|
SYS_SYNC_PRIORITY = 0x2, // Priority Order
|
|
SYS_SYNC_PRIORITY_INHERIT = 0x3, // Basic Priority Inheritance Protocol
|
|
SYS_SYNC_RETRY = 0x4, // Not selected while unlocking
|
|
SYS_SYNC_ATTR_PROTOCOL_MASK = 0xf,
|
|
};
|
|
|
|
// attr_recursive (recursive locks policy)
|
|
enum
|
|
{
|
|
SYS_SYNC_RECURSIVE = 0x10,
|
|
SYS_SYNC_NOT_RECURSIVE = 0x20,
|
|
SYS_SYNC_ATTR_RECURSIVE_MASK = 0xf0,
|
|
};
|
|
|
|
// attr_pshared (sharing among processes policy)
|
|
enum
|
|
{
|
|
SYS_SYNC_PROCESS_SHARED = 0x100,
|
|
SYS_SYNC_NOT_PROCESS_SHARED = 0x200,
|
|
SYS_SYNC_ATTR_PSHARED_MASK = 0xf00,
|
|
};
|
|
|
|
// attr_flags (creation policy)
|
|
enum
|
|
{
|
|
SYS_SYNC_NEWLY_CREATED = 0x1, // Create new object, fails if specified IPC key exists
|
|
SYS_SYNC_NOT_CREATE = 0x2, // Reference existing object, fails if IPC key not found
|
|
SYS_SYNC_NOT_CARE = 0x3, // Reference existing object, create new one if IPC key not found
|
|
SYS_SYNC_ATTR_FLAGS_MASK = 0xf,
|
|
};
|
|
|
|
// attr_adaptive
|
|
enum
|
|
{
|
|
SYS_SYNC_ADAPTIVE = 0x1000,
|
|
SYS_SYNC_NOT_ADAPTIVE = 0x2000,
|
|
SYS_SYNC_ATTR_ADAPTIVE_MASK = 0xf000,
|
|
};
|
|
|
|
// Base class for some kernel objects (shared set of 8192 objects).
|
|
struct lv2_obj
|
|
{
|
|
using id_type = lv2_obj;
|
|
|
|
static const u32 id_step = 0x100;
|
|
static const u32 id_count = 8192;
|
|
|
|
// Find and remove the object from the container (deque or vector)
|
|
template <typename T, typename E>
|
|
static bool unqueue(std::deque<T*>& queue, const E& object)
|
|
{
|
|
for (auto found = queue.cbegin(), end = queue.cend(); found != end; found++)
|
|
{
|
|
if (*found == object)
|
|
{
|
|
queue.erase(found);
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
template <typename E, typename T>
|
|
static T* schedule(std::deque<T*>& queue, u32 protocol)
|
|
{
|
|
if (queue.empty())
|
|
{
|
|
return nullptr;
|
|
}
|
|
|
|
if (protocol == SYS_SYNC_FIFO)
|
|
{
|
|
const auto res = queue.front();
|
|
queue.pop_front();
|
|
return res;
|
|
}
|
|
|
|
u32 prio = -1;
|
|
auto it = queue.cbegin();
|
|
|
|
for (auto found = it, end = queue.cend(); found != end; found++)
|
|
{
|
|
const u32 _prio = static_cast<E*>(*found)->prio;
|
|
|
|
if (_prio < prio)
|
|
{
|
|
it = found;
|
|
prio = _prio;
|
|
}
|
|
}
|
|
|
|
const auto res = *it;
|
|
queue.erase(it);
|
|
return res;
|
|
}
|
|
|
|
// Remove the current thread from the scheduling queue, register timeout
|
|
static void sleep_timeout(cpu_thread&, u64 timeout);
|
|
|
|
static void sleep(cpu_thread& thread, u64 timeout = 0)
|
|
{
|
|
vm::temporary_unlock(thread);
|
|
sleep_timeout(thread, timeout);
|
|
}
|
|
|
|
static void yield(cpu_thread& thread)
|
|
{
|
|
vm::temporary_unlock(thread);
|
|
awake(thread, -4);
|
|
}
|
|
|
|
// Schedule the thread
|
|
static void awake(cpu_thread&, u32 prio);
|
|
|
|
static void awake(cpu_thread& thread)
|
|
{
|
|
awake(thread, -1);
|
|
}
|
|
|
|
static void cleanup();
|
|
|
|
template <typename T, typename F>
|
|
static error_code create(u32 pshared, u64 ipc_key, s32 flags, F&& make)
|
|
{
|
|
switch (pshared)
|
|
{
|
|
case SYS_SYNC_PROCESS_SHARED:
|
|
{
|
|
switch (flags)
|
|
{
|
|
case SYS_SYNC_NEWLY_CREATED:
|
|
case SYS_SYNC_NOT_CARE:
|
|
{
|
|
std::shared_ptr<T> result = make();
|
|
|
|
if (!ipc_manager<T, u64>::add(ipc_key, [&] { if (!idm::import_existing<lv2_obj, T>(result)) result.reset(); return result; }, &result))
|
|
{
|
|
if (flags == SYS_SYNC_NEWLY_CREATED)
|
|
{
|
|
return CELL_EEXIST;
|
|
}
|
|
|
|
if (!idm::import_existing<lv2_obj, T>(result))
|
|
{
|
|
return CELL_EAGAIN;
|
|
}
|
|
|
|
return CELL_OK;
|
|
}
|
|
else if (!result)
|
|
{
|
|
return CELL_EAGAIN;
|
|
}
|
|
else
|
|
{
|
|
return CELL_OK;
|
|
}
|
|
}
|
|
case SYS_SYNC_NOT_CREATE:
|
|
{
|
|
auto result = ipc_manager<T, u64>::get(ipc_key);
|
|
|
|
if (!result)
|
|
{
|
|
return CELL_ESRCH;
|
|
}
|
|
|
|
if (!idm::import_existing<lv2_obj, T>(result))
|
|
{
|
|
return CELL_EAGAIN;
|
|
}
|
|
|
|
return CELL_OK;
|
|
}
|
|
default:
|
|
{
|
|
return CELL_EINVAL;
|
|
}
|
|
}
|
|
}
|
|
case SYS_SYNC_NOT_PROCESS_SHARED:
|
|
{
|
|
if (!idm::import<lv2_obj, T>(std::forward<F>(make)))
|
|
{
|
|
return CELL_EAGAIN;
|
|
}
|
|
|
|
return CELL_OK;
|
|
}
|
|
default:
|
|
{
|
|
return CELL_EINVAL;
|
|
}
|
|
}
|
|
}
|
|
|
|
private:
|
|
// Scheduler mutex
|
|
static shared_mutex g_mutex;
|
|
|
|
// Scheduler queue for active PPU threads
|
|
static std::deque<class ppu_thread*> g_ppu;
|
|
|
|
// Waiting for the response from
|
|
static std::deque<class cpu_thread*> g_pending;
|
|
|
|
// Scheduler queue for timeouts (wait until -> thread)
|
|
static std::deque<std::pair<u64, class cpu_thread*>> g_waiting;
|
|
|
|
static void schedule_all();
|
|
};
|