#include "stdafx.h" #include "sys_event.h" #include "Emu/IdManager.h" #include "Emu/IPC.h" #include "Emu/System.h" #include "Emu/Cell/ErrorCodes.h" #include "Emu/Cell/PPUThread.h" #include "Emu/Cell/SPUThread.h" #include "sys_process.h" #include "util/asm.hpp" LOG_CHANNEL(sys_event); lv2_event_queue::lv2_event_queue(u32 protocol, s32 type, s32 size, u64 name, u64 ipc_key) noexcept : id(idm::last_id()) , protocol{static_cast(protocol)} , type(static_cast(type)) , size(static_cast(size)) , name(name) , key(ipc_key) { } lv2_event_queue::lv2_event_queue(utils::serial& ar) noexcept : id(idm::last_id()) , protocol(ar) , type(ar) , size(ar) , name(ar) , key(ar) { ar(events); } std::shared_ptr lv2_event_queue::load(utils::serial& ar) { auto queue = std::make_shared(ar); return lv2_obj::load(queue->key, queue); } void lv2_event_queue::save(utils::serial& ar) { ar(protocol, type, size, name, key, events); } void lv2_event_queue::save_ptr(utils::serial& ar, lv2_event_queue* q) { if (!lv2_obj::check(q)) { ar(u32{0}); return; } ar(q->id); } std::shared_ptr lv2_event_queue::load_ptr(utils::serial& ar, std::shared_ptr& queue, std::string_view msg) { const u32 id = ar.pop(); if (!id) { return nullptr; } if (auto q = idm::get_unlocked(id)) { // Already initialized return q; } if (id >> 24 != id_base >> 24) { fmt::throw_exception("Failed in event queue pointer deserialization (invalid ID): location: %s, id=0x%x", msg, id); } Emu.PostponeInitCode([id, &queue, msg_str = std::string{msg}]() { // Defer resolving queue = idm::get_unlocked(id); if (!queue) { fmt::throw_exception("Failed in event queue pointer deserialization (not found): location: %s, id=0x%x", msg_str, id); } }); // Null until resolved return nullptr; } lv2_event_port::lv2_event_port(utils::serial& ar) : type(ar) , name(ar) , queue(lv2_event_queue::load_ptr(ar, queue, "eventport")) { } void lv2_event_port::save(utils::serial& ar) { ar(type, name); lv2_event_queue::save_ptr(ar, queue.get()); } std::shared_ptr lv2_event_queue::find(u64 ipc_key) { if (ipc_key == SYS_EVENT_QUEUE_LOCAL) { // Invalid IPC key return {}; } return g_fxo->get>().get(ipc_key); } extern void resume_spu_thread_group_from_waiting(spu_thread& spu); CellError lv2_event_queue::send(lv2_event event, bool* notified_thread, lv2_event_port* port) { if (notified_thread) { *notified_thread = false; } std::lock_guard lock(mutex); if (!exists) { return CELL_ENOTCONN; } if (!pq && !sq) { if (events.size() < this->size + 0u) { // Save event events.emplace_back(event); return {}; } return CELL_EBUSY; } if (type == SYS_PPU_QUEUE) { // Store event in registers auto& ppu = static_cast(*schedule(pq, protocol)); if (ppu.state & cpu_flag::again) { if (auto cpu = get_current_cpu_thread()) { cpu->state += cpu_flag::again; cpu->state += cpu_flag::exit; } sys_event.warning("Ignored event!"); // Fake error for abort return CELL_EAGAIN; } std::tie(ppu.gpr[4], ppu.gpr[5], ppu.gpr[6], ppu.gpr[7]) = event; awake(&ppu); if (port && ppu.prio.load().prio < ensure(cpu_thread::get_current())->prio.load().prio) { // Block event port disconnection for the time being of sending events // PPU -> lower prio PPU is the only case that can cause thread blocking port->is_busy++; ensure(notified_thread); *notified_thread = true; } } else { // Store event in In_MBox auto& spu = static_cast(*schedule(sq, protocol)); if (spu.state & cpu_flag::again) { if (auto cpu = get_current_cpu_thread()) { cpu->state += cpu_flag::exit + cpu_flag::again; } sys_event.warning("Ignored event!"); // Fake error for abort return CELL_EAGAIN; } const u32 data1 = static_cast(std::get<1>(event)); const u32 data2 = static_cast(std::get<2>(event)); const u32 data3 = static_cast(std::get<3>(event)); spu.ch_in_mbox.set_values(4, CELL_OK, data1, data2, data3); resume_spu_thread_group_from_waiting(spu); } return {}; } error_code sys_event_queue_create(cpu_thread& cpu, vm::ptr equeue_id, vm::ptr attr, u64 ipc_key, s32 size) { cpu.state += cpu_flag::wait; sys_event.warning("sys_event_queue_create(equeue_id=*0x%x, attr=*0x%x, ipc_key=0x%llx, size=%d)", equeue_id, attr, ipc_key, size); if (size <= 0 || size > 127) { return CELL_EINVAL; } const u32 protocol = attr->protocol; if (protocol != SYS_SYNC_FIFO && protocol != SYS_SYNC_PRIORITY) { sys_event.error("sys_event_queue_create(): unknown protocol (0x%x)", protocol); return CELL_EINVAL; } const u32 type = attr->type; if (type != SYS_PPU_QUEUE && type != SYS_SPU_QUEUE) { sys_event.error("sys_event_queue_create(): unknown type (0x%x)", type); return CELL_EINVAL; } const u32 pshared = ipc_key == SYS_EVENT_QUEUE_LOCAL ? SYS_SYNC_NOT_PROCESS_SHARED : SYS_SYNC_PROCESS_SHARED; constexpr u32 flags = SYS_SYNC_NEWLY_CREATED; const u64 name = attr->name_u64; if (const auto error = lv2_obj::create(pshared, ipc_key, flags, [&]() { return std::make_shared(protocol, type, size, name, ipc_key); })) { return error; } cpu.check_state(); *equeue_id = idm::last_id(); return CELL_OK; } error_code sys_event_queue_destroy(ppu_thread& ppu, u32 equeue_id, s32 mode) { ppu.state += cpu_flag::wait; sys_event.warning("sys_event_queue_destroy(equeue_id=0x%x, mode=%d)", equeue_id, mode); if (mode && mode != SYS_EVENT_QUEUE_DESTROY_FORCE) { return CELL_EINVAL; } std::vector events; std::unique_lock qlock; cpu_thread* head{}; const auto queue = idm::withdraw(equeue_id, [&](lv2_event_queue& queue) -> CellError { qlock = std::unique_lock{queue.mutex}; head = queue.type == SYS_PPU_QUEUE ? static_cast(+queue.pq) : +queue.sq; if (!mode && head) { return CELL_EBUSY; } if (!queue.events.empty()) { // Copy events for logging, does not empty events.insert(events.begin(), queue.events.begin(), queue.events.end()); } lv2_obj::on_id_destroy(queue, queue.key); if (!head) { qlock.unlock(); } else { for (auto cpu = head; cpu; cpu = cpu->get_next_cpu()) { if (cpu->state & cpu_flag::again) { ppu.state += cpu_flag::again; return CELL_EAGAIN; } } } return {}; }); if (!queue) { return CELL_ESRCH; } if (ppu.state & cpu_flag::again) { return {}; } if (queue.ret) { return queue.ret; } std::string lost_data; if (qlock.owns_lock()) { if (sys_event.warning) { u32 size = 0; for (auto cpu = head; cpu; cpu = cpu->get_next_cpu()) { size++; } fmt::append(lost_data, "Forcefully awaken waiters (%u):\n", size); for (auto cpu = head; cpu; cpu = cpu->get_next_cpu()) { lost_data += cpu->get_name(); lost_data += '\n'; } } if (queue->type == SYS_PPU_QUEUE) { for (auto cpu = +queue->pq; cpu; cpu = cpu->next_cpu) { cpu->gpr[3] = CELL_ECANCELED; queue->append(cpu); } atomic_storage::release(queue->pq, nullptr); lv2_obj::awake_all(); } else { for (auto cpu = +queue->sq; cpu; cpu = cpu->next_cpu) { cpu->ch_in_mbox.set_values(1, CELL_ECANCELED); resume_spu_thread_group_from_waiting(*cpu); } atomic_storage::release(queue->sq, nullptr); } qlock.unlock(); } if (sys_event.warning) { if (!events.empty()) { fmt::append(lost_data, "Unread queue events (%u):\n", events.size()); } for (const lv2_event& evt : events) { fmt::append(lost_data, "data0=0x%x, data1=0x%x, data2=0x%x, data3=0x%x\n" , std::get<0>(evt), std::get<1>(evt), std::get<2>(evt), std::get<3>(evt)); } if (!lost_data.empty()) { sys_event.warning("sys_event_queue_destroy(): %s", lost_data); } } return CELL_OK; } error_code sys_event_queue_tryreceive(ppu_thread& ppu, u32 equeue_id, vm::ptr event_array, s32 size, vm::ptr number) { ppu.state += cpu_flag::wait; sys_event.trace("sys_event_queue_tryreceive(equeue_id=0x%x, event_array=*0x%x, size=%d, number=*0x%x)", equeue_id, event_array, size, number); const auto queue = idm::get(equeue_id); if (!queue) { return CELL_ESRCH; } if (queue->type != SYS_PPU_QUEUE) { return CELL_EINVAL; } std::array events; std::unique_lock lock(queue->mutex); if (!queue->exists) { return CELL_ESRCH; } s32 count = 0; while (count < size && !queue->events.empty()) { auto& dest = events[count++]; const auto event = queue->events.front(); queue->events.pop_front(); std::tie(dest.source, dest.data1, dest.data2, dest.data3) = event; } lock.unlock(); ppu.check_state(); std::copy_n(events.begin(), count, event_array.get_ptr()); *number = count; return CELL_OK; } error_code sys_event_queue_receive(ppu_thread& ppu, u32 equeue_id, vm::ptr dummy_event, u64 timeout) { ppu.state += cpu_flag::wait; sys_event.trace("sys_event_queue_receive(equeue_id=0x%x, *0x%x, timeout=0x%llx)", equeue_id, dummy_event, timeout); ppu.gpr[3] = CELL_OK; const auto queue = idm::get(equeue_id, [&, notify = lv2_obj::notify_all_t()](lv2_event_queue& queue) -> CellError { if (queue.type != SYS_PPU_QUEUE) { return CELL_EINVAL; } lv2_obj::prepare_for_sleep(ppu); std::lock_guard lock(queue.mutex); // "/dev_flash/vsh/module/msmw2.sprx" seems to rely on some cryptic shared memory behaviour that we don't emulate correctly // This is a hack to avoid waiting for 1m40s every time we boot vsh if (queue.key == 0x8005911000000012 && Emu.IsVsh()) { sys_event.todo("sys_event_queue_receive(equeue_id=0x%x, *0x%x, timeout=0x%llx) Bypassing timeout for msmw2.sprx", equeue_id, dummy_event, timeout); timeout = 1; } if (queue.events.empty()) { queue.sleep(ppu, timeout); lv2_obj::emplace(queue.pq, &ppu); return CELL_EBUSY; } std::tie(ppu.gpr[4], ppu.gpr[5], ppu.gpr[6], ppu.gpr[7]) = queue.events.front(); queue.events.pop_front(); return {}; }); if (!queue) { return CELL_ESRCH; } if (queue.ret) { if (queue.ret != CELL_EBUSY) { return queue.ret; } } else { return CELL_OK; } // If cancelled, gpr[3] will be non-zero. Other registers must contain event data. while (auto state = +ppu.state) { if (state & cpu_flag::signal && ppu.state.test_and_reset(cpu_flag::signal)) { break; } if (is_stopped(state)) { std::lock_guard lock_rsx(queue->mutex); for (auto cpu = +queue->pq; cpu; cpu = cpu->next_cpu) { if (cpu == &ppu) { ppu.state += cpu_flag::again; return {}; } } break; } for (usz i = 0; cpu_flag::signal - ppu.state && i < 50; i++) { busy_wait(500); } if (ppu.state & cpu_flag::signal) { continue; } if (timeout) { if (lv2_obj::wait_timeout(timeout, &ppu)) { // Wait for rescheduling if (ppu.check_state()) { continue; } ppu.state += cpu_flag::wait; if (!atomic_storage::load(queue->pq)) { // Waiters queue is empty, so the thread must have been signaled queue->mutex.lock_unlock(); break; } std::lock_guard lock(queue->mutex); if (!queue->unqueue(queue->pq, &ppu)) { break; } ppu.gpr[3] = CELL_ETIMEDOUT; break; } } else { ppu.state.wait(state); } } return not_an_error(ppu.gpr[3]); } error_code sys_event_queue_drain(ppu_thread& ppu, u32 equeue_id) { ppu.state += cpu_flag::wait; sys_event.trace("sys_event_queue_drain(equeue_id=0x%x)", equeue_id); const auto queue = idm::check(equeue_id, [&](lv2_event_queue& queue) { std::lock_guard lock(queue.mutex); queue.events.clear(); }); if (!queue) { return CELL_ESRCH; } return CELL_OK; } error_code sys_event_port_create(cpu_thread& cpu, vm::ptr eport_id, s32 port_type, u64 name) { cpu.state += cpu_flag::wait; sys_event.warning("sys_event_port_create(eport_id=*0x%x, port_type=%d, name=0x%llx)", eport_id, port_type, name); if (port_type != SYS_EVENT_PORT_LOCAL && port_type != 3) { sys_event.error("sys_event_port_create(): unknown port type (%d)", port_type); return CELL_EINVAL; } if (const u32 id = idm::make(port_type, name)) { cpu.check_state(); *eport_id = id; return CELL_OK; } return CELL_EAGAIN; } error_code sys_event_port_destroy(ppu_thread& ppu, u32 eport_id) { ppu.state += cpu_flag::wait; sys_event.warning("sys_event_port_destroy(eport_id=0x%x)", eport_id); const auto port = idm::withdraw(eport_id, [](lv2_event_port& port) -> CellError { if (lv2_obj::check(port.queue)) { return CELL_EISCONN; } return {}; }); if (!port) { return CELL_ESRCH; } if (port.ret) { return port.ret; } return CELL_OK; } error_code sys_event_port_connect_local(cpu_thread& cpu, u32 eport_id, u32 equeue_id) { cpu.state += cpu_flag::wait; sys_event.warning("sys_event_port_connect_local(eport_id=0x%x, equeue_id=0x%x)", eport_id, equeue_id); std::lock_guard lock(id_manager::g_mutex); const auto port = idm::check_unlocked(eport_id); if (!port || !idm::check_unlocked(equeue_id)) { return CELL_ESRCH; } if (port->type != SYS_EVENT_PORT_LOCAL) { return CELL_EINVAL; } if (lv2_obj::check(port->queue)) { return CELL_EISCONN; } port->queue = idm::get_unlocked(equeue_id); return CELL_OK; } error_code sys_event_port_connect_ipc(ppu_thread& ppu, u32 eport_id, u64 ipc_key) { ppu.state += cpu_flag::wait; sys_event.warning("sys_event_port_connect_ipc(eport_id=0x%x, ipc_key=0x%x)", eport_id, ipc_key); if (ipc_key == 0) { return CELL_EINVAL; } auto queue = lv2_event_queue::find(ipc_key); std::lock_guard lock(id_manager::g_mutex); const auto port = idm::check_unlocked(eport_id); if (!port || !queue) { return CELL_ESRCH; } if (port->type != SYS_EVENT_PORT_IPC) { return CELL_EINVAL; } if (lv2_obj::check(port->queue)) { return CELL_EISCONN; } port->queue = std::move(queue); return CELL_OK; } error_code sys_event_port_disconnect(ppu_thread& ppu, u32 eport_id) { ppu.state += cpu_flag::wait; sys_event.warning("sys_event_port_disconnect(eport_id=0x%x)", eport_id); std::lock_guard lock(id_manager::g_mutex); const auto port = idm::check_unlocked(eport_id); if (!port) { return CELL_ESRCH; } if (!lv2_obj::check(port->queue)) { return CELL_ENOTCONN; } if (port->is_busy) { return CELL_EBUSY; } port->queue.reset(); return CELL_OK; } error_code sys_event_port_send(u32 eport_id, u64 data1, u64 data2, u64 data3) { const auto cpu = cpu_thread::get_current(); const auto ppu = cpu ? cpu->try_get() : nullptr; if (cpu) { cpu->state += cpu_flag::wait; } sys_event.trace("sys_event_port_send(eport_id=0x%x, data1=0x%llx, data2=0x%llx, data3=0x%llx)", eport_id, data1, data2, data3); bool notified_thread = false; const auto port = idm::check(eport_id, [&, notify = lv2_obj::notify_all_t()](lv2_event_port& port) -> CellError { if (ppu && ppu->loaded_from_savestate) { port.is_busy++; notified_thread = true; return {}; } if (lv2_obj::check(port.queue)) { const u64 source = port.name ? port.name : (u64{process_getpid() + 0u} << 32) | u64{eport_id}; return port.queue->send(source, data1, data2, data3, ¬ified_thread, ppu && port.queue->type == SYS_PPU_QUEUE ? &port : nullptr); } return CELL_ENOTCONN; }); if (!port) { return CELL_ESRCH; } if (ppu && port->queue->type == SYS_PPU_QUEUE && notified_thread) { // Wait to be requeued if (ppu->test_stopped()) { // Wait again on savestate load ppu->state += cpu_flag::again; } port->is_busy--; return CELL_OK; } if (port.ret) { if (port.ret == CELL_EAGAIN) { // Not really an error code exposed to games (thread has raised cpu_flag::again) return not_an_error(CELL_EAGAIN); } if (port.ret == CELL_EBUSY) { return not_an_error(CELL_EBUSY); } return port.ret; } return CELL_OK; }