#include "stdafx.h" #include "Emu/Memory/Memory.h" #include "Emu/System.h" #include "Emu/IdManager.h" #include "Crypto/unself.h" #include "Crypto/sha1.h" #include "Loader/ELF.h" #include "Utilities/bin_patch.h" #include "Emu/Cell/ErrorCodes.h" #include "Emu/Cell/PPUThread.h" #include "Emu/Cell/RawSPUThread.h" #include "sys_interrupt.h" #include "sys_event.h" #include "sys_spu.h" namespace vm { using namespace ps3; } logs::channel sys_spu("sys_spu"); void sys_spu_image::load(const fs::file& stream) { const spu_exec_object obj{stream, 0, elf_opt::no_sections + elf_opt::no_data}; if (obj != elf_error::ok) { fmt::throw_exception("Failed to load SPU image: %s" HERE, obj.get_error()); } for (const auto& shdr : obj.shdrs) { LOG_NOTICE(SPU, "** Section: sh_type=0x%x, addr=0x%llx, size=0x%llx, flags=0x%x", shdr.sh_type, shdr.sh_addr, shdr.sh_size, shdr.sh_flags); } for (const auto& prog : obj.progs) { LOG_NOTICE(SPU, "** Segment: p_type=0x%x, p_vaddr=0x%llx, p_filesz=0x%llx, p_memsz=0x%llx, flags=0x%x", prog.p_type, prog.p_vaddr, prog.p_filesz, prog.p_memsz, prog.p_flags); if (prog.p_type != SYS_SPU_SEGMENT_TYPE_COPY && prog.p_type != SYS_SPU_SEGMENT_TYPE_INFO) { LOG_ERROR(SPU, "Unknown program type (0x%x)", prog.p_type); } } type = SYS_SPU_IMAGE_TYPE_KERNEL; entry_point = obj.header.e_entry; nsegs = sys_spu_image::get_nsegs(obj.progs); segs = vm::cast(vm::alloc(nsegs * sizeof(sys_spu_segment) + ::size32(stream), vm::main)); const u32 src = segs.addr() + nsegs * sizeof(sys_spu_segment); stream.seek(0); stream.read(vm::base(src), stream.size()); if (nsegs < 0 || sys_spu_image::fill(segs, nsegs, obj.progs, src) != nsegs) { fmt::throw_exception("Failed to load SPU segments (%d)" HERE, nsegs); } } void sys_spu_image::free() { if (type == SYS_SPU_IMAGE_TYPE_KERNEL) { vm::dealloc_verbose_nothrow(segs.addr(), vm::main); } } void sys_spu_image::deploy(u32 loc, sys_spu_segment* segs, u32 nsegs) { // Segment info dump std::string dump; // Executable hash sha1_context sha; sha1_starts(&sha); u8 sha1_hash[20]; for (u32 i = 0; i < nsegs; i++) { auto& seg = segs[i]; fmt::append(dump, "\n\t[%d] t=0x%x, ls=0x%x, size=0x%x, addr=0x%x", i, seg.type, seg.ls, seg.size, seg.addr); sha1_update(&sha, (uchar*)&seg.type, sizeof(seg.type)); // Hash big-endian values if (seg.type == SYS_SPU_SEGMENT_TYPE_COPY) { std::memcpy(vm::base(loc + seg.ls), vm::base(seg.addr), seg.size); sha1_update(&sha, (uchar*)&seg.size, sizeof(seg.size)); sha1_update(&sha, (uchar*)&seg.ls, sizeof(seg.ls)); sha1_update(&sha, vm::g_base_addr + seg.addr, seg.size); } else if (seg.type == SYS_SPU_SEGMENT_TYPE_FILL) { if ((seg.ls | seg.size) % 4) { LOG_ERROR(SPU, "Unaligned SPU FILL type segment (ls=0x%x, size=0x%x)", seg.ls, seg.size); } std::fill_n(vm::_ptr(loc + seg.ls), seg.size / 4, seg.addr); sha1_update(&sha, (uchar*)&seg.size, sizeof(seg.size)); sha1_update(&sha, (uchar*)&seg.ls, sizeof(seg.ls)); sha1_update(&sha, (uchar*)&seg.addr, sizeof(seg.addr)); } else if (seg.type == SYS_SPU_SEGMENT_TYPE_INFO) { const be_t size = seg.size + 0x14; // Workaround sha1_update(&sha, (uchar*)&size, sizeof(size)); } } sha1_finish(&sha, sha1_hash); // Format patch name std::string hash("SPU-0000000000000000000000000000000000000000"); for (u32 i = 0; i < sizeof(sha1_hash); i++) { constexpr auto pal = "0123456789abcdef"; hash[4 + i * 2] = pal[sha1_hash[i] >> 4]; hash[5 + i * 2] = pal[sha1_hash[i] & 15]; } // Apply the patch auto applied = fxm::check_unlocked()->apply(hash, vm::g_base_addr + loc); if (!Emu.GetTitleID().empty()) { // Alternative patch applied += fxm::check_unlocked()->apply(Emu.GetTitleID() + '-' + hash, vm::g_base_addr + loc); } LOG_NOTICE(LOADER, "Loaded SPU image: %s (<- %u)%s", hash, applied, dump); } error_code sys_spu_initialize(u32 max_usable_spu, u32 max_raw_spu) { sys_spu.warning("sys_spu_initialize(max_usable_spu=%d, max_raw_spu=%d)", max_usable_spu, max_raw_spu); if (max_raw_spu > 5) { return CELL_EINVAL; } return CELL_OK; } error_code _sys_spu_image_get_information(vm::ptr img, u32 ptr1, u32 ptr2) { sys_spu.todo("_sys_spu_image_get_information(img=*0x%x, 1=*0x%x, 2=*0x%x)", img, ptr1, ptr2); fmt::throw_exception("Unimplemented syscall: _sys_spu_image_get_information"); } error_code sys_spu_image_open(vm::ptr img, vm::cptr path) { sys_spu.warning("sys_spu_image_open(img=*0x%x, path=%s)", img, path); const fs::file elf_file = decrypt_self(fs::file(vfs::get(path.get_ptr()))); if (!elf_file) { sys_spu.error("sys_spu_image_open() error: %s not found!", path); return CELL_ENOENT; } img->load(elf_file); return CELL_OK; } error_code _sys_spu_image_import(vm::ptr img, u32 src, u32 size, u32 arg4) { sys_spu.warning("_sys_spu_image_import(img=*0x%x, src=*0x%x, size=0x%x, arg4=0x%x)", img, src, size, arg4); img->load(fs::file{vm::base(src), size}); return CELL_OK; } error_code _sys_spu_image_close(vm::ptr img) { sys_spu.warning("_sys_spu_image_close(img=*0x%x)", img); vm::dealloc(img->segs.addr(), vm::main); return CELL_OK; } error_code _sys_raw_spu_image_load(vm::ptr img, u32 ptr, u32 arg3) { sys_spu.todo("_sys_raw_spu_image_load(img=*0x%x, ptr=*0x%x, arg3=0x%x)", img, ptr, arg3); fmt::throw_exception("Unimlemented syscall: _sys_raw_spu_image_load"); } error_code sys_spu_thread_initialize(vm::ptr thread, u32 group_id, u32 spu_num, vm::ptr img, vm::ptr attr, vm::ptr arg) { sys_spu.warning("sys_spu_thread_initialize(thread=*0x%x, group=0x%x, spu_num=%d, img=*0x%x, attr=*0x%x, arg=*0x%x)", thread, group_id, spu_num, img, attr, arg); // Read thread name const std::string thread_name(attr->name.get_ptr(), attr->name ? attr->name_len - 1 : 0); const auto group = idm::get(group_id); if (!group) { return CELL_ESRCH; } semaphore_lock lock(group->mutex); if (spu_num >= group->threads.size()) { return CELL_EINVAL; } if (group->threads[spu_num] || group->run_state != SPU_THREAD_GROUP_STATUS_NOT_INITIALIZED) { return CELL_EBUSY; } if (u32 option = attr->option) { sys_spu.todo("Unimplemented SPU Thread options (0x%x)", option); } auto spu = idm::make_ptr(thread_name, spu_num, group.get()); fxm::get_always()->add_spu(spu); *thread = spu->id; group->threads[spu_num] = std::move(spu); group->args[spu_num] = {arg->arg1, arg->arg2, arg->arg3, arg->arg4}; group->imgs[spu_num] = std::make_pair(*img, std::vector()); group->imgs[spu_num].second.assign(img->segs.get_ptr(), img->segs.get_ptr() + img->nsegs); if (++group->init == group->num) { group->run_state = SPU_THREAD_GROUP_STATUS_INITIALIZED; } return CELL_OK; } error_code sys_spu_thread_set_argument(u32 id, vm::ptr arg) { sys_spu.warning("sys_spu_thread_set_argument(id=0x%x, arg=*0x%x)", id, arg); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } const auto group = thread->group; semaphore_lock lock(group->mutex); group->args[thread->index] = {arg->arg1, arg->arg2, arg->arg3, arg->arg4}; return CELL_OK; } error_code sys_spu_thread_get_exit_status(u32 id, vm::ptr status) { sys_spu.warning("sys_spu_thread_get_exit_status(id=0x%x, status=*0x%x)", id, status); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } // TODO: check CELL_ESTAT condition *status = thread->ch_out_mbox.pop(*thread); return CELL_OK; } error_code sys_spu_thread_group_create(vm::ptr id, u32 num, s32 prio, vm::ptr attr) { sys_spu.warning("sys_spu_thread_group_create(id=*0x%x, num=%d, prio=%d, attr=*0x%x)", id, num, prio, attr); // TODO: max num value should be affected by sys_spu_initialize() settings if (!num || num > 6 || prio < 16 || prio > 255) { return CELL_EINVAL; } if (attr->type) { sys_spu.todo("Unsupported SPU Thread Group type (0x%x)", attr->type); } *id = idm::make(std::string(attr->name.get_ptr(), attr->nsize - 1), num, prio, attr->type, attr->ct); return CELL_OK; } error_code sys_spu_thread_group_destroy(u32 id) { sys_spu.warning("sys_spu_thread_group_destroy(id=0x%x)", id); const auto group = idm::withdraw(id, [](lv2_spu_group& group) -> CellError { const auto _old = group.run_state.compare_and_swap(SPU_THREAD_GROUP_STATUS_INITIALIZED, SPU_THREAD_GROUP_STATUS_NOT_INITIALIZED); if (_old > SPU_THREAD_GROUP_STATUS_INITIALIZED) { return CELL_EBUSY; } return {}; }); if (!group) { return CELL_ESRCH; } if (group.ret) { return group.ret; } // Cleanup for (auto& ptr : group->threads) { if (auto thread = std::move(ptr)) { idm::remove(thread->id); } } fxm::check_unlocked()->add_spu(nullptr); return CELL_OK; } error_code sys_spu_thread_group_start(ppu_thread& ppu, u32 id) { vm::temporary_unlock(ppu); sys_spu.warning("sys_spu_thread_group_start(id=0x%x)", id); const auto group = idm::get(id, [](lv2_spu_group& group) { // SPU_THREAD_GROUP_STATUS_READY state is not used return group.run_state.compare_and_swap_test(SPU_THREAD_GROUP_STATUS_INITIALIZED, SPU_THREAD_GROUP_STATUS_RUNNING); }); if (!group) { return CELL_ESRCH; } if (!group.ret) { return CELL_ESTAT; } semaphore_lock lock(group->mutex); group->join_state = 0; for (auto& thread : group->threads) { if (thread) { auto& args = group->args[thread->index]; auto& img = group->imgs[thread->index]; sys_spu_image::deploy(thread->offset, img.second.data(), img.first.nsegs); thread->pc = img.first.entry_point; thread->cpu_init(); thread->gpr[3] = v128::from64(0, args[0]); thread->gpr[4] = v128::from64(0, args[1]); thread->gpr[5] = v128::from64(0, args[2]); thread->gpr[6] = v128::from64(0, args[3]); thread->status.exchange(SPU_STATUS_RUNNING); } } // Because SPU_THREAD_GROUP_STATUS_READY is not possible, run event is delivered immediately // TODO: check data2 and data3 group->send_run_event(id, 0, 0); for (auto& thread : group->threads) { if (thread) { thread->run(); } } return CELL_OK; } error_code sys_spu_thread_group_suspend(u32 id) { sys_spu.trace("sys_spu_thread_group_suspend(id=0x%x)", id); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (group->type & SYS_SPU_THREAD_GROUP_TYPE_EXCLUSIVE_NON_CONTEXT) // this check may be inaccurate { return CELL_EINVAL; } semaphore_lock lock(group->mutex); if (group->run_state <= SPU_THREAD_GROUP_STATUS_INITIALIZED || group->run_state == SPU_THREAD_GROUP_STATUS_STOPPED) { return CELL_ESTAT; } // SPU_THREAD_GROUP_STATUS_READY state is not used if (group->run_state == SPU_THREAD_GROUP_STATUS_RUNNING) { group->run_state = SPU_THREAD_GROUP_STATUS_SUSPENDED; } else if (group->run_state == SPU_THREAD_GROUP_STATUS_WAITING) { group->run_state = SPU_THREAD_GROUP_STATUS_WAITING_AND_SUSPENDED; } else if (group->run_state == SPU_THREAD_GROUP_STATUS_SUSPENDED || group->run_state == SPU_THREAD_GROUP_STATUS_WAITING_AND_SUSPENDED) { return CELL_OK; } else { return CELL_ESTAT; } for (auto& thread : group->threads) { if (thread) { thread->state += cpu_flag::suspend; } } return CELL_OK; } error_code sys_spu_thread_group_resume(u32 id) { sys_spu.trace("sys_spu_thread_group_resume(id=0x%x)", id); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (group->type & SYS_SPU_THREAD_GROUP_TYPE_EXCLUSIVE_NON_CONTEXT) // this check may be inaccurate { return CELL_EINVAL; } semaphore_lock lock(group->mutex); // SPU_THREAD_GROUP_STATUS_READY state is not used if (group->run_state == SPU_THREAD_GROUP_STATUS_SUSPENDED) { group->run_state = SPU_THREAD_GROUP_STATUS_RUNNING; } else if (group->run_state == SPU_THREAD_GROUP_STATUS_WAITING_AND_SUSPENDED) { group->run_state = SPU_THREAD_GROUP_STATUS_WAITING; } else { return CELL_ESTAT; } for (auto& thread : group->threads) { if (thread) { thread->state -= cpu_flag::suspend; thread->notify(); } } return CELL_OK; } error_code sys_spu_thread_group_yield(u32 id) { sys_spu.trace("sys_spu_thread_group_yield(id=0x%x)", id); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (group->type & SYS_SPU_THREAD_GROUP_TYPE_EXCLUSIVE_NON_CONTEXT) // this check may be inaccurate { return CELL_OK; } if (group->run_state != SPU_THREAD_GROUP_STATUS_RUNNING) { return CELL_ESTAT; } // SPU_THREAD_GROUP_STATUS_READY state is not used, so this function does nothing return CELL_OK; } error_code sys_spu_thread_group_terminate(u32 id, s32 value) { sys_spu.warning("sys_spu_thread_group_terminate(id=0x%x, value=0x%x)", id, value); // The id can be either SPU Thread Group or SPU Thread const auto thread = idm::get(id); const auto _group = idm::get(id); const auto group = thread ? thread->group : _group.get(); if (!group && !thread) { return CELL_ESRCH; } if (thread) { for (auto& t : group->threads) { // find primary (?) thread and compare it with the one specified if (t) { if (t == thread) { break; } else { return CELL_EPERM; } } } } semaphore_lock lock(group->mutex); if (group->run_state <= SPU_THREAD_GROUP_STATUS_INITIALIZED || group->run_state == SPU_THREAD_GROUP_STATUS_WAITING || group->run_state == SPU_THREAD_GROUP_STATUS_WAITING_AND_SUSPENDED) { return CELL_ESTAT; } for (auto& thread : group->threads) { if (thread) { thread->state += cpu_flag::stop; thread->notify(); } } group->run_state = SPU_THREAD_GROUP_STATUS_INITIALIZED; group->exit_status = value; group->join_state |= SPU_TGJSF_TERMINATED; group->cv.notify_one(); return CELL_OK; } error_code sys_spu_thread_group_join(ppu_thread& ppu, u32 id, vm::ptr cause, vm::ptr status) { vm::temporary_unlock(ppu); sys_spu.warning("sys_spu_thread_group_join(id=0x%x, cause=*0x%x, status=*0x%x)", id, cause, status); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } u32 join_state = 0; s32 exit_value = 0; { semaphore_lock lock(group->mutex); if (group->run_state < SPU_THREAD_GROUP_STATUS_INITIALIZED) { return CELL_ESTAT; } if (group->join_state.fetch_or(SPU_TGJSF_IS_JOINING) & SPU_TGJSF_IS_JOINING) { // another PPU thread is joining this thread group return CELL_EBUSY; } lv2_obj::sleep(ppu); while ((group->join_state & ~SPU_TGJSF_IS_JOINING) == 0) { bool stopped = true; for (auto& t : group->threads) { if (t) { if ((t->status & SPU_STATUS_STOPPED_BY_STOP) == 0) { stopped = false; break; } } } if (stopped) { break; } // TODO group->cv.wait(lock, 1000); thread_ctrl::test(); } join_state = group->join_state; exit_value = group->exit_status; group->join_state &= ~SPU_TGJSF_IS_JOINING; group->run_state = SPU_THREAD_GROUP_STATUS_INITIALIZED; // hack } ppu.test_state(); switch (join_state & ~SPU_TGJSF_IS_JOINING) { case 0: { if (cause) *cause = SYS_SPU_THREAD_GROUP_JOIN_ALL_THREADS_EXIT; break; } case SPU_TGJSF_GROUP_EXIT: { if (cause) *cause = SYS_SPU_THREAD_GROUP_JOIN_GROUP_EXIT; break; } case SPU_TGJSF_TERMINATED: { if (cause) *cause = SYS_SPU_THREAD_GROUP_JOIN_TERMINATED; break; } default: { fmt::throw_exception("Unexpected join_state" HERE); } } if (status) { *status = group->exit_status; } return CELL_OK; } error_code sys_spu_thread_group_set_priority(u32 id, s32 priority) { sys_spu.trace("sys_spu_thread_group_set_priority(id=0x%x, priority=%d)", id, priority); if (priority < 16 || priority > 255) { return CELL_EINVAL; } const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (group->type == SYS_SPU_THREAD_GROUP_TYPE_EXCLUSIVE_NON_CONTEXT) { return CELL_EINVAL; } group->prio = priority; return CELL_OK; } error_code sys_spu_thread_group_get_priority(u32 id, vm::ptr priority) { sys_spu.trace("sys_spu_thread_group_get_priority(id=0x%x, priority=*0x%x)", id, priority); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (group->type == SYS_SPU_THREAD_GROUP_TYPE_EXCLUSIVE_NON_CONTEXT) { *priority = 0; } else { *priority = group->prio; } return CELL_OK; } error_code sys_spu_thread_write_ls(u32 id, u32 lsa, u64 value, u32 type) { sys_spu.trace("sys_spu_thread_write_ls(id=0x%x, lsa=0x%05x, value=0x%llx, type=%d)", id, lsa, value, type); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } if (lsa >= 0x40000 || lsa + type > 0x40000 || lsa % type) // check range and alignment { return CELL_EINVAL; } const auto group = thread->group; semaphore_lock lock(group->mutex); if (group->run_state < SPU_THREAD_GROUP_STATUS_WAITING || group->run_state > SPU_THREAD_GROUP_STATUS_RUNNING) { return CELL_ESTAT; } switch (type) { case 1: thread->_ref(lsa) = (u8)value; break; case 2: thread->_ref(lsa) = (u16)value; break; case 4: thread->_ref(lsa) = (u32)value; break; case 8: thread->_ref(lsa) = value; break; default: return CELL_EINVAL; } return CELL_OK; } error_code sys_spu_thread_read_ls(u32 id, u32 lsa, vm::ptr value, u32 type) { sys_spu.trace("sys_spu_thread_read_ls(id=0x%x, lsa=0x%05x, value=*0x%x, type=%d)", id, lsa, value, type); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } if (lsa >= 0x40000 || lsa + type > 0x40000 || lsa % type) // check range and alignment { return CELL_EINVAL; } const auto group = thread->group; semaphore_lock lock(group->mutex); if (group->run_state < SPU_THREAD_GROUP_STATUS_WAITING || group->run_state > SPU_THREAD_GROUP_STATUS_RUNNING) { return CELL_ESTAT; } switch (type) { case 1: *value = thread->_ref(lsa); break; case 2: *value = thread->_ref(lsa); break; case 4: *value = thread->_ref(lsa); break; case 8: *value = thread->_ref(lsa); break; default: return CELL_EINVAL; } return CELL_OK; } error_code sys_spu_thread_write_spu_mb(u32 id, u32 value) { sys_spu.warning("sys_spu_thread_write_spu_mb(id=0x%x, value=0x%x)", id, value); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } const auto group = thread->group; semaphore_lock lock(group->mutex); if (group->run_state < SPU_THREAD_GROUP_STATUS_WAITING || group->run_state > SPU_THREAD_GROUP_STATUS_RUNNING) { return CELL_ESTAT; } thread->ch_in_mbox.push(*thread, value); return CELL_OK; } error_code sys_spu_thread_set_spu_cfg(u32 id, u64 value) { sys_spu.warning("sys_spu_thread_set_spu_cfg(id=0x%x, value=0x%x)", id, value); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } if (value > 3) { return CELL_EINVAL; } thread->snr_config = value; return CELL_OK; } error_code sys_spu_thread_get_spu_cfg(u32 id, vm::ptr value) { sys_spu.warning("sys_spu_thread_get_spu_cfg(id=0x%x, value=*0x%x)", id, value); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } *value = thread->snr_config; return CELL_OK; } error_code sys_spu_thread_write_snr(u32 id, u32 number, u32 value) { sys_spu.trace("sys_spu_thread_write_snr(id=0x%x, number=%d, value=0x%x)", id, number, value); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } if (number > 1) { return CELL_EINVAL; } //if (group->state < SPU_THREAD_GROUP_STATUS_WAITING || group->state > SPU_THREAD_GROUP_STATUS_RUNNING) // ??? //{ // return CELL_ESTAT; //} thread->push_snr(number, value); return CELL_OK; } error_code sys_spu_thread_group_connect_event(u32 id, u32 eq, u32 et) { sys_spu.warning("sys_spu_thread_group_connect_event(id=0x%x, eq=0x%x, et=%d)", id, eq, et); const auto group = idm::get(id); const auto queue = idm::get(eq); if (!group || !queue) { return CELL_ESRCH; } semaphore_lock lock(group->mutex); switch (et) { case SYS_SPU_THREAD_GROUP_EVENT_RUN: { if (!group->ep_run.expired()) { return CELL_EBUSY; } group->ep_run = queue; break; } case SYS_SPU_THREAD_GROUP_EVENT_EXCEPTION: { if (!group->ep_exception.expired()) { return CELL_EBUSY; } group->ep_exception = queue; break; } case SYS_SPU_THREAD_GROUP_EVENT_SYSTEM_MODULE: { if (!group->ep_sysmodule.expired()) { return CELL_EBUSY; } group->ep_sysmodule = queue; break; } default: { sys_spu.error("sys_spu_thread_group_connect_event(): unknown event type (%d)", et); return CELL_EINVAL; } } return CELL_OK; } error_code sys_spu_thread_group_disconnect_event(u32 id, u32 et) { sys_spu.warning("sys_spu_thread_group_disconnect_event(id=0x%x, et=%d)", id, et); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } semaphore_lock lock(group->mutex); switch (et) { case SYS_SPU_THREAD_GROUP_EVENT_RUN: { if (group->ep_run.expired()) { return CELL_ENOTCONN; } group->ep_run.reset(); break; } case SYS_SPU_THREAD_GROUP_EVENT_EXCEPTION: { if (group->ep_exception.expired()) { return CELL_ENOTCONN; } group->ep_exception.reset(); break; } case SYS_SPU_THREAD_GROUP_EVENT_SYSTEM_MODULE: { if (group->ep_sysmodule.expired()) { return CELL_ENOTCONN; } group->ep_sysmodule.reset(); break; } default: { sys_spu.error("sys_spu_thread_group_disconnect_event(): unknown event type (%d)", et); return CELL_EINVAL; } } return CELL_OK; } error_code sys_spu_thread_connect_event(u32 id, u32 eq, u32 et, u8 spup) { sys_spu.warning("sys_spu_thread_connect_event(id=0x%x, eq=0x%x, et=%d, spup=%d)", id, eq, et, spup); const auto thread = idm::get(id); const auto queue = idm::get(eq); if (!thread || !queue) { return CELL_ESRCH; } if (et != SYS_SPU_THREAD_EVENT_USER || spup > 63 || queue->type != SYS_PPU_QUEUE) { sys_spu.error("sys_spu_thread_connect_event(): invalid arguments (et=%d, spup=%d, queue->type=%d)", et, spup, queue->type); return CELL_EINVAL; } semaphore_lock lock(thread->group->mutex); auto& port = thread->spup[spup]; if (!port.expired()) { return CELL_EISCONN; } port = queue; return CELL_OK; } error_code sys_spu_thread_disconnect_event(u32 id, u32 et, u8 spup) { sys_spu.warning("sys_spu_thread_disconnect_event(id=0x%x, et=%d, spup=%d)", id, et, spup); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } if (et != SYS_SPU_THREAD_EVENT_USER || spup > 63) { sys_spu.error("sys_spu_thread_disconnect_event(): invalid arguments (et=%d, spup=%d)", et, spup); return CELL_EINVAL; } semaphore_lock lock(thread->group->mutex); auto& port = thread->spup[spup]; if (port.expired()) { return CELL_ENOTCONN; } port.reset(); return CELL_OK; } error_code sys_spu_thread_bind_queue(u32 id, u32 spuq, u32 spuq_num) { sys_spu.warning("sys_spu_thread_bind_queue(id=0x%x, spuq=0x%x, spuq_num=0x%x)", id, spuq, spuq_num); const auto thread = idm::get(id); const auto queue = idm::get(spuq); if (!thread || !queue) { return CELL_ESRCH; } if (queue->type != SYS_SPU_QUEUE) { return CELL_EINVAL; } semaphore_lock lock(thread->group->mutex); for (auto& v : thread->spuq) { if (auto q = v.second.lock()) { if (v.first == spuq_num || q == queue) { return CELL_EBUSY; } } } for (auto& v : thread->spuq) { if (v.second.expired()) { v.first = spuq_num; v.second = queue; return CELL_OK; } } return CELL_EAGAIN; } error_code sys_spu_thread_unbind_queue(u32 id, u32 spuq_num) { sys_spu.warning("sys_spu_thread_unbind_queue(id=0x%x, spuq_num=0x%x)", id, spuq_num); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } semaphore_lock lock(thread->group->mutex); for (auto& v : thread->spuq) { if (v.first == spuq_num && !v.second.expired()) { v.second.reset(); return CELL_OK; } } return CELL_ESRCH; } error_code sys_spu_thread_group_connect_event_all_threads(u32 id, u32 eq, u64 req, vm::ptr spup) { sys_spu.warning("sys_spu_thread_group_connect_event_all_threads(id=0x%x, eq=0x%x, req=0x%llx, spup=*0x%x)", id, eq, req, spup); const auto group = idm::get(id); const auto queue = idm::get(eq); if (!group || !queue) { return CELL_ESRCH; } if (!req) { return CELL_EINVAL; } semaphore_lock lock(group->mutex); if (group->run_state < SPU_THREAD_GROUP_STATUS_INITIALIZED) { return CELL_ESTAT; } u8 port = 0; // SPU Port number for (; port < 64; port++) { if (!(req & (1ull << port))) { continue; } bool found = true; for (auto& t : group->threads) { if (t) { if (!t->spup[port].expired()) { found = false; break; } } } if (found) { break; } } if (port == 64) { return CELL_EISCONN; } for (auto& t : group->threads) { if (t) { t->spup[port] = queue; } } *spup = port; return CELL_OK; } error_code sys_spu_thread_group_disconnect_event_all_threads(u32 id, u8 spup) { sys_spu.warning("sys_spu_thread_group_disconnect_event_all_threads(id=0x%x, spup=%d)", id, spup); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (spup > 63) { return CELL_EINVAL; } semaphore_lock lock(group->mutex); for (auto& t : group->threads) { if (t) { t->spup[spup].reset(); } } return CELL_OK; } error_code sys_raw_spu_create(vm::ptr id, vm::ptr attr) { sys_spu.warning("sys_raw_spu_create(id=*0x%x, attr=*0x%x)", id, attr); // TODO: check number set by sys_spu_initialize() const auto thread = idm::make_ptr(""); if (!thread) { return CELL_EAGAIN; } thread->cpu_init(); const u32 _id = thread->index; fxm::get_always()->add_spu(std::move(thread)); *id = _id; return CELL_OK; } error_code sys_raw_spu_destroy(ppu_thread& ppu, u32 id) { sys_spu.warning("sys_raw_spu_destroy(id=%d)", id); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } // TODO: CELL_EBUSY is not returned // Stop thread thread->state += cpu_flag::stop; // Kernel objects which must be removed std::unordered_map> to_remove; // Clear interrupt handlers for (auto& intr : thread->int_ctrl) { if (intr.tag) { if (auto handler = intr.tag->handler.lock()) { // SLEEP handler->join(); to_remove.emplace(handler.get(), 0); } to_remove.emplace(intr.tag.get(), 0); } } // Scan all kernel objects to determine IDs idm::select([&](u32 id, lv2_obj& obj) { const auto found = to_remove.find(&obj); if (found != to_remove.end()) { found->second = id; } }); // Remove IDs for (auto&& pair : to_remove) { if (pair.second >> 24 == 0xa) idm::remove(pair.second); if (pair.second >> 24 == 0xb) idm::remove(pair.second); } idm::remove(thread->id); fxm::check_unlocked()->add_spu(nullptr); return CELL_OK; } error_code sys_raw_spu_create_interrupt_tag(u32 id, u32 class_id, u32 hwthread, vm::ptr intrtag) { sys_spu.warning("sys_raw_spu_create_interrupt_tag(id=%d, class_id=%d, hwthread=0x%x, intrtag=*0x%x)", id, class_id, hwthread, intrtag); if (class_id != 0 && class_id != 2) { return CELL_EINVAL; } CellError error = {}; const auto tag = idm::import([&]() { std::shared_ptr result; auto thread = idm::check_unlocked(id); if (!thread) { error = CELL_ESRCH; return result; } auto& int_ctrl = thread->int_ctrl[class_id]; if (int_ctrl.tag) { error = CELL_EAGAIN; return result; } result = std::make_shared(); int_ctrl.tag = result; return result; }); if (tag) { *intrtag = tag; return CELL_OK; } return error; } error_code sys_raw_spu_set_int_mask(u32 id, u32 class_id, u64 mask) { sys_spu.trace("sys_raw_spu_set_int_mask(id=%d, class_id=%d, mask=0x%llx)", id, class_id, mask); if (class_id != 0 && class_id != 2) { return CELL_EINVAL; } const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } thread->int_ctrl[class_id].mask.exchange(mask); return CELL_OK; } error_code sys_raw_spu_get_int_mask(u32 id, u32 class_id, vm::ptr mask) { sys_spu.trace("sys_raw_spu_get_int_mask(id=%d, class_id=%d, mask=*0x%x)", id, class_id, mask); if (class_id != 0 && class_id != 2) { return CELL_EINVAL; } const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } *mask = thread->int_ctrl[class_id].mask; return CELL_OK; } error_code sys_raw_spu_set_int_stat(u32 id, u32 class_id, u64 stat) { sys_spu.trace("sys_raw_spu_set_int_stat(id=%d, class_id=%d, stat=0x%llx)", id, class_id, stat); if (class_id != 0 && class_id != 2) { return CELL_EINVAL; } const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } thread->int_ctrl[class_id].clear(stat); return CELL_OK; } error_code sys_raw_spu_get_int_stat(u32 id, u32 class_id, vm::ptr stat) { sys_spu.trace("sys_raw_spu_get_int_stat(id=%d, class_id=%d, stat=*0x%x)", id, class_id, stat); if (class_id != 0 && class_id != 2) { return CELL_EINVAL; } const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } *stat = thread->int_ctrl[class_id].stat; return CELL_OK; } error_code sys_raw_spu_read_puint_mb(u32 id, vm::ptr value) { sys_spu.trace("sys_raw_spu_read_puint_mb(id=%d, value=*0x%x)", id, value); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } *value = thread->ch_out_intr_mbox.pop(*thread); return CELL_OK; } error_code sys_raw_spu_set_spu_cfg(u32 id, u32 value) { sys_spu.trace("sys_raw_spu_set_spu_cfg(id=%d, value=0x%x)", id, value); if (value > 3) { fmt::throw_exception("Unexpected value (0x%x)" HERE, value); } const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } thread->snr_config = value; return CELL_OK; } error_code sys_raw_spu_get_spu_cfg(u32 id, vm::ptr value) { sys_spu.trace("sys_raw_spu_get_spu_afg(id=%d, value=*0x%x)", id, value); const auto thread = idm::get(id); if (!thread) { return CELL_ESRCH; } *value = (u32)thread->snr_config; return CELL_OK; }