#include "stdafx.h" #include "sys_spu.h" #include "Emu/System.h" #include "Emu/VFS.h" #include "Emu/IdManager.h" #include "Crypto/unself.h" #include "Crypto/unedat.h" #include "Crypto/sha1.h" #include "Loader/ELF.h" #include "Utilities/bin_patch.h" #include "Emu/Cell/ErrorCodes.h" #include "Emu/Cell/PPUThread.h" #include "Emu/Cell/PPUModule.h" #include "Emu/Cell/RawSPUThread.h" #include "sys_interrupt.h" #include "sys_process.h" #include "sys_memory.h" #include "sys_mmapper.h" #include "sys_event.h" #include "sys_fs.h" #include "util/asm.hpp" LOG_CHANNEL(sys_spu); extern u64 get_timebased_time(); template <> void fmt_class_string::format(std::string& out, u64 arg) { format_enum(out, arg, [](spu_group_status value) { switch (value) { case SPU_THREAD_GROUP_STATUS_NOT_INITIALIZED: return "uninitialized"; case SPU_THREAD_GROUP_STATUS_INITIALIZED: return "initialized"; case SPU_THREAD_GROUP_STATUS_READY: return "ready"; case SPU_THREAD_GROUP_STATUS_WAITING: return "waiting"; case SPU_THREAD_GROUP_STATUS_SUSPENDED: return "suspended"; case SPU_THREAD_GROUP_STATUS_WAITING_AND_SUSPENDED: return "waiting and suspended"; case SPU_THREAD_GROUP_STATUS_RUNNING: return "running"; case SPU_THREAD_GROUP_STATUS_STOPPED: return "stopped"; case SPU_THREAD_GROUP_STATUS_DESTROYED: return "destroyed"; case SPU_THREAD_GROUP_STATUS_UNKNOWN: break; } return unknown; }); } void sys_spu_image::load(const fs::file& stream) { const spu_exec_object obj{stream, 0, elf_opt::no_sections + elf_opt::no_data}; if (obj != elf_error::ok) { fmt::throw_exception("Failed to load SPU image: %s", obj.get_error()); } for (const auto& shdr : obj.shdrs) { spu_log.notice("** Section: sh_type=0x%x, addr=0x%llx, size=0x%llx, flags=0x%x", shdr.sh_type, shdr.sh_addr, shdr.sh_size, shdr.sh_flags); } for (const auto& prog : obj.progs) { spu_log.notice("** Segment: p_type=0x%x, p_vaddr=0x%llx, p_filesz=0x%llx, p_memsz=0x%llx, flags=0x%x", prog.p_type, prog.p_vaddr, prog.p_filesz, prog.p_memsz, prog.p_flags); if (prog.p_type != u32{SYS_SPU_SEGMENT_TYPE_COPY} && prog.p_type != u32{SYS_SPU_SEGMENT_TYPE_INFO}) { spu_log.error("Unknown program type (0x%x)", prog.p_type); } } this->type = SYS_SPU_IMAGE_TYPE_KERNEL; const s32 nsegs = sys_spu_image::get_nsegs(obj.progs); const u32 mem_size = nsegs * sizeof(sys_spu_segment) + ::size32(stream); const vm::ptr segs = vm::cast(vm::alloc(mem_size, vm::main)); const u32 entry = obj.header.e_entry; const u32 src = (segs + nsegs).addr(); stream.seek(0); stream.read(vm::base(src), stream.size()); if (nsegs <= 0 || nsegs > 0x20 || sys_spu_image::fill(segs, nsegs, obj.progs, src) != nsegs) { fmt::throw_exception("Failed to load SPU segments (%d)", nsegs); } // Write ID and save entry this->entry_point = idm::make(+obj.header.e_entry, segs, nsegs); // Unused and set to 0 this->nsegs = 0; this->segs = vm::null; vm::page_protect(segs.addr(), ::align(mem_size, 4096), 0, 0, vm::page_writable); } void sys_spu_image::free() { if (type == SYS_SPU_IMAGE_TYPE_KERNEL) { vm::dealloc_verbose_nothrow(segs.addr(), vm::main); } } void sys_spu_image::deploy(u8* loc, sys_spu_segment* segs, u32 nsegs) { // Segment info dump std::string dump; // Executable hash sha1_context sha; sha1_starts(&sha); u8 sha1_hash[20]; for (u32 i = 0; i < nsegs; i++) { auto& seg = segs[i]; fmt::append(dump, "\n\t[%d] t=0x%x, ls=0x%x, size=0x%x, addr=0x%x", i, seg.type, seg.ls, seg.size, seg.addr); sha1_update(&sha, reinterpret_cast(&seg.type), sizeof(seg.type)); // Hash big-endian values if (seg.type == SYS_SPU_SEGMENT_TYPE_COPY) { std::memcpy(loc + seg.ls, vm::base(seg.addr), seg.size); sha1_update(&sha, reinterpret_cast(&seg.size), sizeof(seg.size)); sha1_update(&sha, reinterpret_cast(&seg.ls), sizeof(seg.ls)); sha1_update(&sha, vm::_ptr(seg.addr), seg.size); } else if (seg.type == SYS_SPU_SEGMENT_TYPE_FILL) { if ((seg.ls | seg.size) % 4) { spu_log.error("Unaligned SPU FILL type segment (ls=0x%x, size=0x%x)", seg.ls, seg.size); } std::fill_n(reinterpret_cast*>(loc + seg.ls), seg.size / 4, seg.addr); sha1_update(&sha, reinterpret_cast(&seg.size), sizeof(seg.size)); sha1_update(&sha, reinterpret_cast(&seg.ls), sizeof(seg.ls)); sha1_update(&sha, reinterpret_cast(&seg.addr), sizeof(seg.addr)); } else if (seg.type == SYS_SPU_SEGMENT_TYPE_INFO) { const be_t size = seg.size + 0x14; // Workaround sha1_update(&sha, reinterpret_cast(&size), sizeof(size)); } } sha1_finish(&sha, sha1_hash); // Format patch name std::string hash("SPU-0000000000000000000000000000000000000000"); for (u32 i = 0; i < sizeof(sha1_hash); i++) { constexpr auto pal = "0123456789abcdef"; hash[4 + i * 2] = pal[sha1_hash[i] >> 4]; hash[5 + i * 2] = pal[sha1_hash[i] & 15]; } // Apply the patch auto applied = g_fxo->get()->apply(hash, loc); if (!Emu.GetTitleID().empty()) { // Alternative patch applied += g_fxo->get()->apply(Emu.GetTitleID() + '-' + hash, loc); } spu_log.notice("Loaded SPU image: %s (<- %u)%s", hash, applied, dump); } // Get spu thread ptr, returns group ptr as well for refcounting std::pair*, std::shared_ptr> lv2_spu_group::get_thread(u32 id) { if (id >= 0x06000000) { // thread index is out of range (5 max) return {}; } // Bits 0-23 contain group id (without id base) decltype(get_thread(0)) res{nullptr, idm::get((id & 0xFFFFFF) | (lv2_spu_group::id_base & ~0xFFFFFF))}; // Bits 24-31 contain thread index within the group const u32 index = id >> 24; if (auto group = res.second.get(); group && group->init > index) { res.first = group->threads[index].get(); } return res; } error_code sys_spu_initialize(ppu_thread& ppu, u32 max_usable_spu, u32 max_raw_spu) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_initialize(max_usable_spu=%d, max_raw_spu=%d)", max_usable_spu, max_raw_spu); if (max_raw_spu > 5) { return CELL_EINVAL; } return CELL_OK; } error_code _sys_spu_image_get_information(ppu_thread& ppu, vm::ptr img, vm::ptr entry_point, vm::ptr nsegs) { ppu.state += cpu_flag::wait; sys_spu.warning("_sys_spu_image_get_information(img=*0x%x, entry_point=*0x%x, nsegs=*0x%x)", img, entry_point, nsegs); if (img->type != SYS_SPU_IMAGE_TYPE_KERNEL) { return CELL_EINVAL; } const auto image = idm::get(img->entry_point); if (!image) { return CELL_ESRCH; } *entry_point = image->e_entry; *nsegs = image->nsegs; return CELL_OK; } error_code sys_spu_image_open(ppu_thread& ppu, vm::ptr img, vm::cptr path) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_image_open(img=*0x%x, path=%s)", img, path); auto [fs_error, ppath, path0, file, type] = lv2_file::open(path.get_ptr(), 0, 0); if (fs_error) { return {fs_error, path}; } const fs::file elf_file = decrypt_self(std::move(file), g_fxo->get()->devKlic.load()._bytes); if (!elf_file) { sys_spu.error("sys_spu_image_open(): file %s is illegal for SPU image!", path); return {CELL_ENOEXEC, path}; } img->load(elf_file); return CELL_OK; } error_code _sys_spu_image_import(ppu_thread& ppu, vm::ptr img, u32 src, u32 size, u32 arg4) { ppu.state += cpu_flag::wait; sys_spu.warning("_sys_spu_image_import(img=*0x%x, src=*0x%x, size=0x%x, arg4=0x%x)", img, src, size, arg4); img->load(fs::file{vm::base(src), size}); return CELL_OK; } error_code _sys_spu_image_close(ppu_thread& ppu, vm::ptr img) { ppu.state += cpu_flag::wait; sys_spu.warning("_sys_spu_image_close(img=*0x%x)", img); if (img->type != SYS_SPU_IMAGE_TYPE_KERNEL) { return CELL_EINVAL; } const auto handle = idm::withdraw(img->entry_point); if (!handle) { return CELL_ESRCH; } ensure(vm::dealloc(handle->segs.addr(), vm::main)); return CELL_OK; } error_code _sys_spu_image_get_segments(ppu_thread& ppu, vm::ptr img, vm::ptr segments, s32 nseg) { ppu.state += cpu_flag::wait; sys_spu.error("_sys_spu_image_get_segments(img=*0x%x, segments=*0x%x, nseg=%d)", img, segments, nseg); if (nseg <= 0 || nseg > 0x20 || img->type != SYS_SPU_IMAGE_TYPE_KERNEL) { return CELL_EINVAL; } const auto handle = idm::get(img->entry_point); if (!handle) { return CELL_ESRCH; } // TODO: apply SPU patches std::memcpy(segments.get_ptr(), handle->segs.get_ptr(), sizeof(sys_spu_segment) * std::min(nseg, handle->nsegs)); return CELL_OK; } error_code sys_spu_thread_initialize(ppu_thread& ppu, vm::ptr thread, u32 group_id, u32 spu_num, vm::ptr img, vm::ptr attr, vm::ptr arg) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_initialize(thread=*0x%x, group=0x%x, spu_num=%d, img=*0x%x, attr=*0x%x, arg=*0x%x)", thread, group_id, spu_num, img, attr, arg); const u32 option = attr->option; if (attr->name_len > 0x80 || option & ~(SYS_SPU_THREAD_OPTION_DEC_SYNC_TB_ENABLE | SYS_SPU_THREAD_OPTION_ASYNC_INTR_ENABLE)) { return CELL_EINVAL; } sys_spu_image image; switch (img->type) { case SYS_SPU_IMAGE_TYPE_KERNEL: { const auto handle = idm::get(img->entry_point); if (!handle) { return CELL_ESRCH; } // Image information is stored in IDM image.entry_point = handle->e_entry; image.nsegs = handle->nsegs; image.segs = handle->segs; image.type = SYS_SPU_IMAGE_TYPE_KERNEL; break; } case SYS_SPU_IMAGE_TYPE_USER: { if (img->entry_point > 0x3fffc || img->nsegs <= 0 || img->nsegs > 0x20) { return CELL_EINVAL; } image = *img; break; } default: return CELL_EINVAL; } // Read thread name const std::string thread_name(attr->name.get_ptr(), std::max(attr->name_len, 1) - 1); const auto group = idm::get(group_id); if (!group) { return CELL_ESRCH; } if (spu_num >= group->threads_map.size()) { return CELL_EINVAL; } std::lock_guard lock(group->mutex); if (auto state = +group->run_state; state != SPU_THREAD_GROUP_STATUS_NOT_INITIALIZED) { if (state == SPU_THREAD_GROUP_STATUS_DESTROYED) { return CELL_ESRCH; } return CELL_EBUSY; } if (group->threads_map[spu_num] != -1) { return CELL_EBUSY; } if (option & SYS_SPU_THREAD_OPTION_ASYNC_INTR_ENABLE) { sys_spu.warning("Unimplemented SPU Thread options (0x%x)", option); } const u32 inited = group->init; const u32 tid = (inited << 24) | (group_id & 0xffffff); ensure(idm::import>([&]() { std::string full_name = fmt::format("SPU[0x%07x] ", tid); if (!thread_name.empty()) { fmt::append(full_name, "%s ", thread_name); } const auto spu = std::make_shared>(full_name, group.get(), spu_num, thread_name, tid, false, option); group->threads[inited] = spu; group->threads_map[spu_num] = static_cast(inited); return spu; })); *thread = tid; group->args[inited] = {arg->arg1, arg->arg2, arg->arg3, arg->arg4}; group->imgs[inited].first = image; group->imgs[inited].second.assign(image.segs.get_ptr(), image.segs.get_ptr() + image.nsegs); if (++group->init == group->max_num) { if (g_cfg.core.max_spurs_threads < 6 && group->max_num > 0u + g_cfg.core.max_spurs_threads) { if (group->name.ends_with("CellSpursKernelGroup")) { // Hack: don't run more SPURS threads than specified. group->max_run = g_cfg.core.max_spurs_threads; spu_log.success("HACK: '%s' (0x%x) limited to %u threads.", group->name, group_id, +g_cfg.core.max_spurs_threads); } } const auto old = group->run_state.compare_and_swap(SPU_THREAD_GROUP_STATUS_NOT_INITIALIZED, SPU_THREAD_GROUP_STATUS_INITIALIZED); if (old == SPU_THREAD_GROUP_STATUS_DESTROYED) { return CELL_ESRCH; } } sys_spu.warning(u8"sys_spu_thread_initialize(): Thread ā€œ%sā€ created (id=0x%x)", thread_name, tid); return CELL_OK; } error_code sys_spu_thread_set_argument(ppu_thread& ppu, u32 id, vm::ptr arg) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_set_argument(id=0x%x, arg=*0x%x)", id, arg); const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } std::lock_guard lock(group->mutex); group->args[id >> 24] = {arg->arg1, arg->arg2, arg->arg3, arg->arg4}; return CELL_OK; } error_code sys_spu_thread_get_exit_status(ppu_thread& ppu, u32 id, vm::ptr status) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_get_exit_status(id=0x%x, status=*0x%x)", id, status); const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } u32 data; if (thread->exit_status.try_read(data)) { *status = static_cast(data); return CELL_OK; } return CELL_ESTAT; } error_code sys_spu_thread_group_create(ppu_thread& ppu, vm::ptr id, u32 num, s32 prio, vm::ptr attr) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_group_create(id=*0x%x, num=%d, prio=%d, attr=*0x%x)", id, num, prio, attr); const s32 min_prio = g_ps3_process_info.has_root_perm() ? 0 : 16; if (attr->nsize > 0x80 || !num) { return CELL_EINVAL; } const s32 type = attr->type; bool use_scheduler = true; bool use_memct = !!(type & SYS_SPU_THREAD_GROUP_TYPE_MEMORY_FROM_CONTAINER); bool needs_root = false; u32 max_threads = 6; // TODO: max num value should be affected by sys_spu_initialize() settings u32 min_threads = 1; u32 mem_size = 0; lv2_memory_container* ct{}; if (type) { sys_spu.warning("sys_spu_thread_group_create(): SPU Thread Group type (0x%x)", type); } switch (type) { case 0x0: case 0x4: case 0x18: { break; } case 0x20: case 0x22: case 0x24: case 0x26: { if (type == 0x22 || type == 0x26) { needs_root = true; } min_threads = 2; // That's what appears from reversing break; } case 0x2: case 0x6: case 0xA: case 0x102: case 0x106: case 0x10A: case 0x202: case 0x206: case 0x20A: case 0x902: case 0x906: case 0xA02: case 0xA06: case 0xC02: case 0xC06: { if (type & 0x700) { max_threads = 1; } needs_root = true; break; } default: return CELL_EINVAL; } if (type & SYS_SPU_THREAD_GROUP_TYPE_COOPERATE_WITH_SYSTEM) { // Constant size, unknown what it means but it's definitely not for each spu thread alone mem_size = SPU_LS_SIZE; use_scheduler = false; } else if (type & SYS_SPU_THREAD_GROUP_TYPE_NON_CONTEXT) { // No memory consumed mem_size = 0; use_scheduler = false; } else { // 256kb for each spu thread, probably for saving and restoring SPU LS (used by scheduler?) mem_size = SPU_LS_SIZE * num; } if (num < min_threads || num > max_threads || (needs_root && min_prio == 0x10) || (use_scheduler && (prio > 255 || prio < min_prio))) { return CELL_EINVAL; } if (use_memct && mem_size) { const auto sct = idm::get(attr->ct); if (!sct) { return CELL_ESRCH; } if (sct->take(mem_size) != mem_size) { return CELL_ENOMEM; } ct = sct.get(); } else { ct = g_fxo->get(); if (ct->take(mem_size) != mem_size) { return CELL_ENOMEM; } } const auto group = idm::make_ptr(std::string(attr->name.get_ptr(), std::max(attr->nsize, 1) - 1), num, prio, type, ct, use_scheduler, mem_size); if (!group) { ct->used -= mem_size; return CELL_EAGAIN; } *id = idm::last_id(); sys_spu.warning(u8"sys_spu_thread_group_create(): Thread group ā€œ%sā€ created (id=0x%x)", group->name, idm::last_id()); return CELL_OK; } error_code sys_spu_thread_group_destroy(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_group_destroy(id=0x%x)", id); const auto group = idm::withdraw(id, [](lv2_spu_group& group) -> CellError { if (!group.run_state.fetch_op([](spu_group_status& state) { if (state <= SPU_THREAD_GROUP_STATUS_INITIALIZED) { state = SPU_THREAD_GROUP_STATUS_DESTROYED; return true; } return false; }).second) { return CELL_EBUSY; } group.ct->used -= group.mem_size; return {}; }); if (!group) { return CELL_ESRCH; } if (group.ret) { return group.ret; } group->mutex.lock_unlock(); for (const auto& t : group->threads) { if (auto thread = t.get()) { // Deallocate LS ensure(vm::get(vm::spu)->dealloc(SPU_FAKE_BASE_ADDR + SPU_LS_SIZE * (thread->id & 0xffffff), &thread->shm)); // Remove ID from IDM (destruction will occur in group destructor) idm::remove>(thread->id); } } return CELL_OK; } error_code sys_spu_thread_group_start(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_group_start(id=0x%x)", id); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } std::lock_guard lock(group->mutex); // SPU_THREAD_GROUP_STATUS_READY state is not used switch (group->run_state.compare_and_swap(SPU_THREAD_GROUP_STATUS_INITIALIZED, SPU_THREAD_GROUP_STATUS_RUNNING)) { case SPU_THREAD_GROUP_STATUS_INITIALIZED: break; case SPU_THREAD_GROUP_STATUS_DESTROYED: return CELL_ESRCH; default: return CELL_ESTAT; } const u32 max_threads = group->max_run; group->join_state = 0; group->exit_status = 0; group->running = max_threads; group->set_terminate = false; for (auto& thread : group->threads) { if (thread) { auto& args = group->args[thread->lv2_id >> 24]; auto& img = group->imgs[thread->lv2_id >> 24]; sys_spu_image::deploy(thread->ls, img.second.data(), img.first.nsegs); thread->cpu_init(); thread->gpr[3] = v128::from64(0, args[0]); thread->gpr[4] = v128::from64(0, args[1]); thread->gpr[5] = v128::from64(0, args[2]); thread->gpr[6] = v128::from64(0, args[3]); thread->status_npc = {SPU_STATUS_RUNNING, img.first.entry_point}; } } // Because SPU_THREAD_GROUP_STATUS_READY is not possible, run event is delivered immediately // TODO: check data2 and data3 group->send_run_event(id, 0, 0); u32 ran_threads = max_threads; for (auto& thread : group->threads) { if (!ran_threads) { break; } if (thread && ran_threads--) { thread->state -= cpu_flag::stop; thread_ctrl::notify(*thread); } } return CELL_OK; } error_code sys_spu_thread_group_suspend(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_group_suspend(id=0x%x)", id); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (!group->has_scheduler_context || group->type & 0xf00) { return CELL_EINVAL; } std::lock_guard lock(group->mutex); CellError error; group->run_state.fetch_op([&error](spu_group_status& state) { if (state == SPU_THREAD_GROUP_STATUS_DESTROYED) { error = CELL_ESRCH; return false; } if (state <= SPU_THREAD_GROUP_STATUS_INITIALIZED || state == SPU_THREAD_GROUP_STATUS_STOPPED) { error = CELL_ESTAT; return false; } // SPU_THREAD_GROUP_STATUS_READY state is not used if (state == SPU_THREAD_GROUP_STATUS_RUNNING) { state = SPU_THREAD_GROUP_STATUS_SUSPENDED; } else if (state == SPU_THREAD_GROUP_STATUS_WAITING) { state = SPU_THREAD_GROUP_STATUS_WAITING_AND_SUSPENDED; } else if (state == SPU_THREAD_GROUP_STATUS_SUSPENDED || state == SPU_THREAD_GROUP_STATUS_WAITING_AND_SUSPENDED) { error = {}; return false; } else { error = CELL_ESTAT; return false; } error = CellError{CELL_CANCEL + 0u}; return true; }); if (error != CELL_CANCEL + 0u) { if (!error) { return CELL_OK; } return error; } for (auto& thread : group->threads) { if (thread) { thread->state += cpu_flag::suspend; } } return CELL_OK; } error_code sys_spu_thread_group_resume(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_group_resume(id=0x%x)", id); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (!group->has_scheduler_context || group->type & 0xf00) { return CELL_EINVAL; } std::lock_guard lock(group->mutex); CellError error; group->run_state.fetch_op([&error](spu_group_status& state) { if (state == SPU_THREAD_GROUP_STATUS_DESTROYED) { error = CELL_ESRCH; return false; } // SPU_THREAD_GROUP_STATUS_READY state is not used if (state == SPU_THREAD_GROUP_STATUS_SUSPENDED) { state = SPU_THREAD_GROUP_STATUS_RUNNING; } else if (state == SPU_THREAD_GROUP_STATUS_WAITING_AND_SUSPENDED) { state = SPU_THREAD_GROUP_STATUS_WAITING; } else { error = CELL_ESTAT; return false; } error = CellError{CELL_CANCEL + 0u}; return true; }); if (error != CELL_CANCEL + 0u) { return error; } for (auto& thread : group->threads) { if (thread) { thread->state -= cpu_flag::suspend; thread_ctrl::notify(*thread); } } return CELL_OK; } error_code sys_spu_thread_group_yield(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_group_yield(id=0x%x)", id); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } // No effect on these group types if (!group->has_scheduler_context || group->type & 0xf00) { return CELL_OK; } if (auto state = +group->run_state; state != SPU_THREAD_GROUP_STATUS_RUNNING) { if (state == SPU_THREAD_GROUP_STATUS_DESTROYED) { return CELL_ESRCH; } return CELL_ESTAT; } // SPU_THREAD_GROUP_STATUS_READY state is not used, so this function does nothing return CELL_OK; } error_code sys_spu_thread_group_terminate(ppu_thread& ppu, u32 id, s32 value) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_group_terminate(id=0x%x, value=0x%x)", id, value); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } std::unique_lock lock(group->mutex); // There should be a small period of sleep when the PPU waits for a signal of termination auto short_sleep = [](ppu_thread& ppu) { lv2_obj::sleep(ppu); busy_wait(3000); ppu.check_state(); }; if (auto state = +group->run_state; state <= SPU_THREAD_GROUP_STATUS_INITIALIZED || state == SPU_THREAD_GROUP_STATUS_WAITING || state == SPU_THREAD_GROUP_STATUS_WAITING_AND_SUSPENDED || state == SPU_THREAD_GROUP_STATUS_DESTROYED) { if (state == SPU_THREAD_GROUP_STATUS_DESTROYED) { return CELL_ESRCH; } return CELL_ESTAT; } if (group->set_terminate) { // Wait for termination, only then return error code const u64 last_stop = group->stop_count; lock.unlock(); short_sleep(ppu); while (group->stop_count == last_stop) { group->stop_count.wait(last_stop); } return CELL_ESTAT; } group->set_terminate = true; for (auto& thread : group->threads) { if (thread) { thread->state.fetch_op([](bs_t& flags) { if (flags & cpu_flag::stop) { // In case the thread raised the ret flag itself at some point do not raise it again return false; } flags += cpu_flag::stop + cpu_flag::ret; return true; }); } } for (auto& thread : group->threads) { while (thread && group->running && thread->state & cpu_flag::wait) { // TODO: replace with proper solution if (atomic_wait_engine::raw_notify(nullptr, thread_ctrl::get_native_id(*thread))) break; } } group->exit_status = value; group->join_state = SYS_SPU_THREAD_GROUP_JOIN_TERMINATED; // Wait until the threads are actually stopped const u64 last_stop = group->stop_count; lock.unlock(); short_sleep(ppu); while (group->stop_count == last_stop) { group->stop_count.wait(last_stop); } return CELL_OK; } error_code sys_spu_thread_group_join(ppu_thread& ppu, u32 id, vm::ptr cause, vm::ptr status) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_group_join(id=0x%x, cause=*0x%x, status=*0x%x)", id, cause, status); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } do { std::unique_lock lock(group->mutex); const auto state = +group->run_state; if (state == SPU_THREAD_GROUP_STATUS_DESTROYED) { return CELL_ESRCH; } if (state < SPU_THREAD_GROUP_STATUS_INITIALIZED) { return CELL_ESTAT; } if (group->waiter) { // another PPU thread is joining this thread group return CELL_EBUSY; } if (group->join_state && state == SPU_THREAD_GROUP_STATUS_INITIALIZED) { // Already signaled ppu.gpr[4] = group->join_state; ppu.gpr[5] = group->exit_status; group->join_state.release(0); break; } else { // Subscribe to receive status in r4-r5 group->waiter = &ppu; } lv2_obj::sleep(ppu); lock.unlock(); while (!ppu.state.test_and_reset(cpu_flag::signal)) { if (ppu.is_stopped()) { return 0; } thread_ctrl::wait(); } } while (0); if (!cause) { if (status) { // Report unwritten data return CELL_EFAULT; } return not_an_error(CELL_EFAULT); } *cause = static_cast(ppu.gpr[4]); if (!status) { return not_an_error(CELL_EFAULT); } *status = static_cast(ppu.gpr[5]); return CELL_OK; } error_code sys_spu_thread_group_set_priority(ppu_thread& ppu, u32 id, s32 priority) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_group_set_priority(id=0x%x, priority=%d)", id, priority); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (!group->has_scheduler_context || priority < (g_ps3_process_info.has_root_perm() ? 0 : 16) || priority > 255) { return CELL_EINVAL; } group->prio = priority; return CELL_OK; } error_code sys_spu_thread_group_get_priority(ppu_thread& ppu, u32 id, vm::ptr priority) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_group_get_priority(id=0x%x, priority=*0x%x)", id, priority); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (!group->has_scheduler_context) { *priority = 0; } else { *priority = group->prio; } return CELL_OK; } error_code sys_spu_thread_group_set_cooperative_victims(ppu_thread& ppu, u32 id, u32 threads_mask) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_group_set_cooperative_victims(id=0x%x, threads_mask=0x%x)", id, threads_mask); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (!(group->type & SYS_SPU_THREAD_GROUP_TYPE_COOPERATE_WITH_SYSTEM)) { return CELL_EINVAL; } if (threads_mask >= 1u << group->max_num) { return CELL_EINVAL; } // TODO return CELL_OK; } error_code sys_spu_thread_group_syscall_253(ppu_thread& ppu, u32 id, vm::ptr info) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_group_syscall_253(id=0x%x, info=*0x%x)", id, info); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (!(group->type & SYS_SPU_THREAD_GROUP_TYPE_COOPERATE_WITH_SYSTEM)) { return CELL_EINVAL; } // TODO info->deadlineMissCounter = 0; info->deadlineMeetCounter = 0; info->timestamp = get_timebased_time(); return CELL_OK; } error_code sys_spu_thread_write_ls(ppu_thread& ppu, u32 id, u32 lsa, u64 value, u32 type) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_write_ls(id=0x%x, lsa=0x%05x, value=0x%llx, type=%d)", id, lsa, value, type); if (lsa >= SPU_LS_SIZE || type > 8 || !type || (type | lsa) & (type - 1)) // check range and alignment { return CELL_EINVAL; } const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } std::lock_guard lock(group->mutex); if (auto state = +group->run_state; state < SPU_THREAD_GROUP_STATUS_WAITING || state > SPU_THREAD_GROUP_STATUS_RUNNING) { if (state == SPU_THREAD_GROUP_STATUS_DESTROYED) { return CELL_ESRCH; } return CELL_ESTAT; } switch (type) { case 1: thread->_ref(lsa) = static_cast(value); break; case 2: thread->_ref(lsa) = static_cast(value); break; case 4: thread->_ref(lsa) = static_cast(value); break; case 8: thread->_ref(lsa) = value; break; default: fmt::throw_exception("Unreachable"); } return CELL_OK; } error_code sys_spu_thread_read_ls(ppu_thread& ppu, u32 id, u32 lsa, vm::ptr value, u32 type) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_read_ls(id=0x%x, lsa=0x%05x, value=*0x%x, type=%d)", id, lsa, value, type); if (lsa >= SPU_LS_SIZE || type > 8 || !type || (type | lsa) & (type - 1)) // check range and alignment { return CELL_EINVAL; } const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } std::lock_guard lock(group->mutex); if (auto state = +group->run_state; state < SPU_THREAD_GROUP_STATUS_WAITING || state > SPU_THREAD_GROUP_STATUS_RUNNING) { if (state == SPU_THREAD_GROUP_STATUS_DESTROYED) { return CELL_ESRCH; } return CELL_ESTAT; } switch (type) { case 1: *value = thread->_ref(lsa); break; case 2: *value = thread->_ref(lsa); break; case 4: *value = thread->_ref(lsa); break; case 8: *value = thread->_ref(lsa); break; default: fmt::throw_exception("Unreachable"); } return CELL_OK; } error_code sys_spu_thread_write_spu_mb(ppu_thread& ppu, u32 id, u32 value) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_write_spu_mb(id=0x%x, value=0x%x)", id, value); const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } std::lock_guard lock(group->mutex); thread->ch_in_mbox.push(*thread, value); return CELL_OK; } error_code sys_spu_thread_set_spu_cfg(ppu_thread& ppu, u32 id, u64 value) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_set_spu_cfg(id=0x%x, value=0x%x)", id, value); if (value > 3) { return CELL_EINVAL; } const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } thread->snr_config = value; return CELL_OK; } error_code sys_spu_thread_get_spu_cfg(ppu_thread& ppu, u32 id, vm::ptr value) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_get_spu_cfg(id=0x%x, value=*0x%x)", id, value); const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } *value = thread->snr_config; return CELL_OK; } error_code sys_spu_thread_write_snr(ppu_thread& ppu, u32 id, u32 number, u32 value) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_spu_thread_write_snr(id=0x%x, number=%d, value=0x%x)", id, number, value); const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } if (number > 1) { return CELL_EINVAL; } thread->push_snr(number, value); return CELL_OK; } error_code sys_spu_thread_group_connect_event(ppu_thread& ppu, u32 id, u32 eq, u32 et) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_group_connect_event(id=0x%x, eq=0x%x, et=%d)", id, eq, et); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } const auto ep = et == SYS_SPU_THREAD_GROUP_EVENT_SYSTEM_MODULE ? &group->ep_sysmodule : et == SYS_SPU_THREAD_GROUP_EVENT_EXCEPTION ? &group->ep_exception : et == SYS_SPU_THREAD_GROUP_EVENT_RUN ? &group->ep_run : nullptr; if (!ep) { sys_spu.error("sys_spu_thread_group_connect_event(): unknown event type (%d)", et); return CELL_EINVAL; } if (et == SYS_SPU_THREAD_GROUP_EVENT_SYSTEM_MODULE && !(group->type & SYS_SPU_THREAD_GROUP_TYPE_COOPERATE_WITH_SYSTEM)) { return CELL_EINVAL; } const auto queue = idm::get(eq); std::lock_guard lock(group->mutex); if (lv2_event_queue::check(*ep)) { return CELL_EBUSY; } // ESRCH of event queue after EBUSY if (!queue) { return CELL_ESRCH; } *ep = queue; return CELL_OK; } error_code sys_spu_thread_group_disconnect_event(ppu_thread& ppu, u32 id, u32 et) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_group_disconnect_event(id=0x%x, et=%d)", id, et); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } const auto ep = et == SYS_SPU_THREAD_GROUP_EVENT_SYSTEM_MODULE ? &group->ep_sysmodule : et == SYS_SPU_THREAD_GROUP_EVENT_EXCEPTION ? &group->ep_exception : et == SYS_SPU_THREAD_GROUP_EVENT_RUN ? &group->ep_run : nullptr; if (!ep) { sys_spu.error("sys_spu_thread_group_disconnect_event(): unknown event type (%d)", et); return CELL_EINVAL; } std::lock_guard lock(group->mutex); if (!lv2_event_queue::check(*ep)) { return CELL_EINVAL; } ep->reset(); return CELL_OK; } error_code sys_spu_thread_connect_event(ppu_thread& ppu, u32 id, u32 eq, u32 et, u8 spup) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_connect_event(id=0x%x, eq=0x%x, et=%d, spup=%d)", id, eq, et, spup); const auto [thread, group] = lv2_spu_group::get_thread(id); const auto queue = idm::get(eq); if (!queue || !thread) [[unlikely]] { return CELL_ESRCH; } if (et != SYS_SPU_THREAD_EVENT_USER || spup > 63) { sys_spu.error("sys_spu_thread_connect_event(): invalid arguments (et=%d, spup=%d, queue->type=%d)", et, spup, queue->type); return CELL_EINVAL; } std::lock_guard lock(group->mutex); auto& port = thread->spup[spup]; if (lv2_event_queue::check(port)) { return CELL_EISCONN; } port = queue; return CELL_OK; } error_code sys_spu_thread_disconnect_event(ppu_thread& ppu, u32 id, u32 et, u8 spup) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_disconnect_event(id=0x%x, et=%d, spup=%d)", id, et, spup); const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } if (et != SYS_SPU_THREAD_EVENT_USER || spup > 63) { sys_spu.error("sys_spu_thread_disconnect_event(): invalid arguments (et=%d, spup=%d)", et, spup); return CELL_EINVAL; } std::lock_guard lock(group->mutex); auto& port = thread->spup[spup]; if (!lv2_event_queue::check(port)) { return CELL_ENOTCONN; } port.reset(); return CELL_OK; } error_code sys_spu_thread_bind_queue(ppu_thread& ppu, u32 id, u32 spuq, u32 spuq_num) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_bind_queue(id=0x%x, spuq=0x%x, spuq_num=0x%x)", id, spuq, spuq_num); const auto [thread, group] = lv2_spu_group::get_thread(id); const auto queue = idm::get(spuq); if (!queue || !thread) [[unlikely]] { return CELL_ESRCH; } if (queue->type != SYS_SPU_QUEUE) { return CELL_EINVAL; } std::lock_guard lock(group->mutex); decltype(std::data(thread->spuq)) q{}; for (auto& v : thread->spuq) { // Check if the entry is assigned at all if (const decltype(v.second) test{}; !v.second.owner_before(test) && !test.owner_before(v.second)) { if (!q) { q = &v; } continue; } if (v.first == spuq_num || (!v.second.owner_before(queue) && !queue.owner_before(v.second))) { return CELL_EBUSY; } } if (!q) { return CELL_EAGAIN; } q->first = spuq_num; q->second = queue; return CELL_OK; } error_code sys_spu_thread_unbind_queue(ppu_thread& ppu, u32 id, u32 spuq_num) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_unbind_queue(id=0x%x, spuq_num=0x%x)", id, spuq_num); const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } std::lock_guard lock(group->mutex); for (auto& v : thread->spuq) { if (v.first != spuq_num) { continue; } if (const decltype(v.second) test{}; !v.second.owner_before(test) && !test.owner_before(v.second)) { continue; } v.second.reset(); return CELL_OK; } return CELL_ESRCH; } error_code sys_spu_thread_group_connect_event_all_threads(ppu_thread& ppu, u32 id, u32 eq, u64 req, vm::ptr spup) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_group_connect_event_all_threads(id=0x%x, eq=0x%x, req=0x%llx, spup=*0x%x)", id, eq, req, spup); const auto group = idm::get(id); const auto queue = idm::get(eq); if (!group || !queue) { return CELL_ESRCH; } if (!req) { return CELL_EINVAL; } std::lock_guard lock(group->mutex); if (auto state = +group->run_state; state < SPU_THREAD_GROUP_STATUS_INITIALIZED || state == SPU_THREAD_GROUP_STATUS_DESTROYED) { if (state == SPU_THREAD_GROUP_STATUS_DESTROYED) { return CELL_ESRCH; } return CELL_ESTAT; } u8 port = 0; // SPU Port number for (; port < 64; port++) { if (!(req & (1ull << port))) { continue; } bool found = true; for (auto& t : group->threads) { if (t) { if (lv2_event_queue::check(t->spup[port])) { found = false; break; } } } if (found) { break; } } if (port == 64) { return CELL_EISCONN; } for (auto& t : group->threads) { if (t) { t->spup[port] = queue; } } *spup = port; return CELL_OK; } error_code sys_spu_thread_group_disconnect_event_all_threads(ppu_thread& ppu, u32 id, u8 spup) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_group_disconnect_event_all_threads(id=0x%x, spup=%d)", id, spup); const auto group = idm::get(id); if (!group) { return CELL_ESRCH; } if (spup > 63) { return CELL_EINVAL; } std::lock_guard lock(group->mutex); for (auto& t : group->threads) { if (t) { t->spup[spup].reset(); } } return CELL_OK; } error_code sys_spu_thread_group_log(ppu_thread& ppu, s32 command, vm::ptr stat) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_group_log(command=0x%x, stat=*0x%x)", command, stat); struct spu_group_log_state_t { atomic_t state = SYS_SPU_THREAD_GROUP_LOG_ON; }; const auto state = g_fxo->get(); switch (command) { case SYS_SPU_THREAD_GROUP_LOG_GET_STATUS: { if (!stat) { return CELL_EFAULT; } *stat = state->state; break; } case SYS_SPU_THREAD_GROUP_LOG_ON: case SYS_SPU_THREAD_GROUP_LOG_OFF: { state->state.release(command); break; } default: return CELL_EINVAL; } return CELL_OK; } error_code sys_spu_thread_recover_page_fault(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_spu_thread_recover_page_fault(id=0x%x)", id); const auto [thread, group] = lv2_spu_group::get_thread(id); if (!thread) [[unlikely]] { return CELL_ESRCH; } return mmapper_thread_recover_page_fault(thread); } error_code sys_raw_spu_recover_page_fault(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_raw_spu_recover_page_fault(id=0x%x)", id); const auto thread = idm::get>(spu_thread::find_raw_spu(id)); if (!thread) [[unlikely]] { return CELL_ESRCH; } return mmapper_thread_recover_page_fault(thread.get()); } error_code sys_raw_spu_create(ppu_thread& ppu, vm::ptr id, vm::ptr attr) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_raw_spu_create(id=*0x%x, attr=*0x%x)", id, attr); // TODO: check number set by sys_spu_initialize() if (!spu_thread::g_raw_spu_ctr.try_inc(5)) { return CELL_EAGAIN; } u32 index = 0; // Find free RawSPU ID while (!spu_thread::g_raw_spu_id[index].try_inc(1)) { if (++index == 5) index = 0; } const u32 tid = idm::make>(fmt::format("RawSPU[0x%x] ", index), nullptr, index, "", index); spu_thread::g_raw_spu_id[index] = (ensure(tid)); *id = index; return CELL_OK; } error_code sys_isolated_spu_create(ppu_thread& ppu, vm::ptr id, vm::ptr image, u64 arg1, u64 arg2, u64 arg3, u64 arg4) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_create(id=*0x%x, image=*0x%x, arg1=0x%llx, arg2=0x%llx, arg3=0x%llx, arg4=0x%llx)", id, image, arg1, arg2, arg3, arg4); // TODO: More accurate SPU image memory size calculation u32 max = image.addr() & -4096; while (max != 0u - 4096 && vm::check_addr(max)) { max += 4096; } const auto obj = decrypt_self(fs::file{image.get_ptr(), max - image.addr()}); if (!obj) { return CELL_EAUTHFAIL; } // TODO: check number set by sys_spu_initialize() if (!spu_thread::g_raw_spu_ctr.try_inc(5)) { return CELL_EAGAIN; } u32 index = 0; // Find free RawSPU ID while (!spu_thread::g_raw_spu_id[index].try_inc(1)) { if (++index == 5) index = 0; } const vm::addr_t ls_addr{RAW_SPU_BASE_ADDR + RAW_SPU_OFFSET * index}; const auto thread = idm::make_ptr>(fmt::format("IsoSPU[0x%x] ", index), nullptr, index, "", index, true); thread->gpr[3] = v128::from64(0, arg1); thread->gpr[4] = v128::from64(0, arg2); thread->gpr[5] = v128::from64(0, arg3); thread->gpr[6] = v128::from64(0, arg4); spu_thread::g_raw_spu_id[index] = (ensure(thread->id)); sys_spu_image img; img.load(obj); auto image_info = idm::get(img.entry_point); img.deploy(thread->ls, image_info->segs.get_ptr(), image_info->nsegs); thread->write_reg(ls_addr + RAW_SPU_PROB_OFFSET + SPU_NPC_offs, image_info->e_entry); ensure(idm::remove_verify(img.entry_point, std::move(image_info))); *id = index; return CELL_OK; } template error_code raw_spu_destroy(ppu_thread& ppu, u32 id) { const u32 idm_id = spu_thread::find_raw_spu(id); auto thread = idm::get>(idm_id, [](named_thread& thread) { if (thread.get_type() != (isolated ? spu_type::isolated : spu_type::raw)) { return false; } // Stop thread thread.state += cpu_flag::exit; thread = thread_state::aborting; return true; }); if (!thread || !thread.ret) [[unlikely]] { return CELL_ESRCH; } // TODO: CELL_EBUSY is not returned // Kernel objects which must be removed std::vector, u32>> to_remove; // Clear interrupt handlers for (auto& intr : thread->int_ctrl) { if (auto tag = intr.tag.lock()) { if (auto handler = tag->handler.lock()) { // SLEEP lv2_obj::sleep(ppu); handler->join(); to_remove.emplace_back(std::move(handler), 0); } to_remove.emplace_back(std::move(tag), 0); } } // Scan all kernel objects to determine IDs idm::select([&](u32 id, lv2_obj& obj) { for (auto& pair : to_remove) { if (pair.first.get() == std::addressof(obj)) { pair.second = id; } } }); // Remove IDs for (auto&& pair : to_remove) { if (pair.second >> 24 == 0xa) idm::remove_verify(pair.second, std::move(pair.first)); if (pair.second >> 24 == 0xb) idm::remove_verify(pair.second, std::move(pair.first)); } (*thread)(); if (!idm::remove_verify>(idm_id, std::move(thread.ptr))) { // Other thread destroyed beforehead return CELL_ESRCH; } return CELL_OK; } error_code sys_raw_spu_destroy(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_raw_spu_destroy(id=%d)", id); return raw_spu_destroy(ppu, id); } error_code sys_isolated_spu_destroy(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_destroy(id=%d)", id); return raw_spu_destroy(ppu, id); } template error_code raw_spu_create_interrupt_tag(u32 id, u32 class_id, u32 hwthread, vm::ptr intrtag) { if (class_id != 0 && class_id != 2) { return CELL_EINVAL; } CellError error = {}; const auto tag = idm::import([&]() { std::shared_ptr result; auto thread = idm::check_unlocked>(spu_thread::find_raw_spu(id)); if (!thread || *thread == thread_state::aborting || thread->get_type() != (isolated ? spu_type::isolated : spu_type::raw)) { error = CELL_ESRCH; return result; } auto& int_ctrl = thread->int_ctrl[class_id]; if (!int_ctrl.tag.expired()) { error = CELL_EAGAIN; return result; } result = std::make_shared(); int_ctrl.tag = result; return result; }); if (tag) { *intrtag = tag; return CELL_OK; } return error; } error_code sys_raw_spu_create_interrupt_tag(ppu_thread& ppu, u32 id, u32 class_id, u32 hwthread, vm::ptr intrtag) { ppu.state += cpu_flag::wait; sys_spu.warning("sys_raw_spu_create_interrupt_tag(id=%d, class_id=%d, hwthread=0x%x, intrtag=*0x%x)", id, class_id, hwthread, intrtag); return raw_spu_create_interrupt_tag(id, class_id, hwthread, intrtag); } error_code sys_isolated_spu_create_interrupt_tag(ppu_thread& ppu, u32 id, u32 class_id, u32 hwthread, vm::ptr intrtag) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_create_interrupt_tag(id=%d, class_id=%d, hwthread=0x%x, intrtag=*0x%x)", id, class_id, hwthread, intrtag); return raw_spu_create_interrupt_tag(id, class_id, hwthread, intrtag); } template error_code raw_spu_set_int_mask(u32 id, u32 class_id, u64 mask) { if (class_id != 0 && class_id != 2) { return CELL_EINVAL; } const auto thread = idm::get>(spu_thread::find_raw_spu(id)); if (!thread || thread->get_type() != (isolated ? spu_type::isolated : spu_type::raw)) [[unlikely]] { return CELL_ESRCH; } thread->int_ctrl[class_id].mask.exchange(mask); return CELL_OK; } error_code sys_raw_spu_set_int_mask(ppu_thread& ppu, u32 id, u32 class_id, u64 mask) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_raw_spu_set_int_mask(id=%d, class_id=%d, mask=0x%llx)", id, class_id, mask); return raw_spu_set_int_mask(id, class_id, mask); } error_code sys_isolated_spu_set_int_mask(ppu_thread& ppu, u32 id, u32 class_id, u64 mask) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_set_int_mask(id=%d, class_id=%d, mask=0x%llx)", id, class_id, mask); return raw_spu_set_int_mask(id, class_id, mask); } template error_code raw_spu_set_int_stat(u32 id, u32 class_id, u64 stat) { if (class_id != 0 && class_id != 2) { return CELL_EINVAL; } const auto thread = idm::get>(spu_thread::find_raw_spu(id)); if (!thread || thread->get_type() != (isolated ? spu_type::isolated : spu_type::raw)) [[unlikely]] { return CELL_ESRCH; } thread->int_ctrl[class_id].clear(stat); return CELL_OK; } error_code sys_raw_spu_set_int_stat(ppu_thread& ppu, u32 id, u32 class_id, u64 stat) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_raw_spu_set_int_stat(id=%d, class_id=%d, stat=0x%llx)", id, class_id, stat); return raw_spu_set_int_stat(id, class_id, stat); } error_code sys_isolated_spu_set_int_stat(ppu_thread& ppu, u32 id, u32 class_id, u64 stat) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_set_int_stat(id=%d, class_id=%d, stat=0x%llx)", id, class_id, stat); return raw_spu_set_int_stat(id, class_id, stat); } template error_code raw_spu_get_int_control(u32 id, u32 class_id, vm::ptr value, atomic_t spu_int_ctrl_t::* control) { if (class_id != 0 && class_id != 2) { return CELL_EINVAL; } const auto thread = idm::get>(spu_thread::find_raw_spu(id)); if (!thread || thread->get_type() != (isolated ? spu_type::isolated : spu_type::raw)) [[unlikely]] { return CELL_ESRCH; } *value = thread->int_ctrl[class_id].*control; return CELL_OK; } error_code sys_raw_spu_get_int_mask(ppu_thread& ppu, u32 id, u32 class_id, vm::ptr mask) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_raw_spu_get_int_mask(id=%d, class_id=%d, mask=*0x%x)", id, class_id, mask); return raw_spu_get_int_control(id, class_id, mask, &spu_int_ctrl_t::mask); } error_code sys_isolated_spu_get_int_mask(ppu_thread& ppu, u32 id, u32 class_id, vm::ptr mask) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_isolated_spu_get_int_mask(id=%d, class_id=%d, mask=*0x%x)", id, class_id, mask); return raw_spu_get_int_control(id, class_id, mask, &spu_int_ctrl_t::mask); } error_code sys_raw_spu_get_int_stat(ppu_thread& ppu, u32 id, u32 class_id, vm::ptr stat) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_raw_spu_get_int_stat(id=%d, class_id=%d, stat=*0x%x)", id, class_id, stat); return raw_spu_get_int_control(id, class_id, stat, &spu_int_ctrl_t::stat); } error_code sys_isolated_spu_get_int_stat(ppu_thread& ppu, u32 id, u32 class_id, vm::ptr stat) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_get_int_stat(id=%d, class_id=%d, stat=*0x%x)", id, class_id, stat); return raw_spu_get_int_control(id, class_id, stat, &spu_int_ctrl_t::stat); } template error_code raw_spu_read_puint_mb(u32 id, vm::ptr value) { const auto thread = idm::get>(spu_thread::find_raw_spu(id)); if (!thread || thread->get_type() != (isolated ? spu_type::isolated : spu_type::raw)) [[unlikely]] { return CELL_ESRCH; } *value = thread->ch_out_intr_mbox.pop(); return CELL_OK; } error_code sys_raw_spu_read_puint_mb(ppu_thread& ppu, u32 id, vm::ptr value) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_raw_spu_read_puint_mb(id=%d, value=*0x%x)", id, value); return raw_spu_read_puint_mb(id, value); } error_code sys_isolated_spu_read_puint_mb(ppu_thread& ppu, u32 id, vm::ptr value) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_read_puint_mb(id=%d, value=*0x%x)", id, value); return raw_spu_read_puint_mb(id, value); } template error_code raw_spu_set_spu_cfg(u32 id, u32 value) { if (value > 3) { fmt::throw_exception("Unexpected value (0x%x)", value); } const auto thread = idm::get>(spu_thread::find_raw_spu(id)); if (!thread || thread->get_type() != (isolated ? spu_type::isolated : spu_type::raw)) [[unlikely]] { return CELL_ESRCH; } thread->snr_config = value; return CELL_OK; } error_code sys_raw_spu_set_spu_cfg(ppu_thread& ppu, u32 id, u32 value) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_raw_spu_set_spu_cfg(id=%d, value=0x%x)", id, value); return raw_spu_set_spu_cfg(id, value); } error_code sys_isolated_spu_set_spu_cfg(ppu_thread& ppu, u32 id, u32 value) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_set_spu_cfg(id=%d, value=0x%x)", id, value); return raw_spu_set_spu_cfg(id, value); } template error_code raw_spu_get_spu_cfg(u32 id, vm::ptr value) { const auto thread = idm::get>(spu_thread::find_raw_spu(id)); if (!thread || thread->get_type() != (isolated ? spu_type::isolated : spu_type::raw)) [[unlikely]] { return CELL_ESRCH; } *value = static_cast(thread->snr_config); return CELL_OK; } error_code sys_raw_spu_get_spu_cfg(ppu_thread& ppu, u32 id, vm::ptr value) { ppu.state += cpu_flag::wait; sys_spu.trace("sys_raw_spu_get_spu_afg(id=%d, value=*0x%x)", id, value); return raw_spu_get_spu_cfg(id, value); } error_code sys_isolated_spu_get_spu_cfg(ppu_thread& ppu, u32 id, vm::ptr value) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_get_spu_afg(id=%d, value=*0x%x)", id, value); return raw_spu_get_spu_cfg(id, value); } error_code sys_isolated_spu_start(ppu_thread& ppu, u32 id) { ppu.state += cpu_flag::wait; sys_spu.todo("sys_isolated_spu_start(id=%d)", id); const auto thread = idm::get>(spu_thread::find_raw_spu(id)); if (!thread) [[unlikely]] { return CELL_ESRCH; } // TODO: Can return ESTAT if called twice thread->write_reg(RAW_SPU_BASE_ADDR + thread->lv2_id * RAW_SPU_OFFSET + RAW_SPU_PROB_OFFSET + SPU_RunCntl_offs, SPU_RUNCNTL_RUN_REQUEST); return CELL_OK; }