#pragma once #include "Utilities/File.h" #include "Utilities/lockless.h" #include "Utilities/Thread.h" #include "Common/bitfield.hpp" #include "Emu/System.h" #include "Emu/cache_utils.hpp" #include "Program/ProgramStateCache.h" #include "Common/texture_cache_checker.h" #include "Overlays/Shaders/shader_loading_dialog.h" #include #include #include "util/sysinfo.hpp" #include "util/fnv_hash.hpp" namespace rsx { template class shaders_cache { using unpacked_type = lf_fifo, 1000>; // TODO: Determine best size struct pipeline_data { u64 vertex_program_hash; u64 fragment_program_hash; u64 pipeline_storage_hash; u32 vp_ctrl; u32 vp_texture_dimensions; u64 vp_instruction_mask[9]; u32 vp_base_address; u32 vp_entry; u16 vp_jump_table[32]; u16 vp_multisampled_textures; u16 vp_reserved_0; u32 vp_reserved_1; u32 fp_ctrl; u32 fp_texture_dimensions; u32 fp_texcoord_control; u16 fp_height; u16 fp_pixel_layout; u16 fp_lighting_flags; u16 fp_shadow_textures; u16 fp_redirected_textures; u16 fp_multisampled_textures; u64 fp_reserved_0; pipeline_storage_type pipeline_properties; }; std::string version_prefix; std::string root_path; std::string pipeline_class_name; lf_fifo, 100> fragment_program_data; backend_storage& m_storage; static std::string get_message(u32 index, u32 processed, u32 entry_count) { return fmt::format("%s pipeline object %u of %u", index == 0 ? "Loading" : "Compiling", processed, entry_count); } void load_shaders(uint nb_workers, unpacked_type& unpacked, std::string& directory_path, std::vector& entries, u32 entry_count, shader_loading_dialog* dlg) { atomic_t processed(0); std::function shader_load_worker = [&](u32 stop_at) { u32 pos; // Processed is incremented before work starts in order to avoid two workers working on the same shader while (((pos = processed++) < stop_at) && !Emu.IsStopped()) { fs::dir_entry tmp = entries[pos]; const auto filename = directory_path + "/" + tmp.name; fs::file f(filename); if (!f) { // Unexpected error, but avoid crash continue; } if (f.size() != sizeof(pipeline_data)) { rsx_log.error("Removing cached pipeline object %s since it's not binary compatible with the current shader cache", tmp.name.c_str()); fs::remove_file(filename); continue; } pipeline_data pdata{}; f.read(&pdata, f.size()); auto entry = unpack(pdata); if (std::get<1>(entry).data.empty() || !std::get<2>(entry).ucode_length) { continue; } m_storage.preload_programs(std::get<1>(entry), std::get<2>(entry)); unpacked[unpacked.push_begin()] = std::move(entry); } // Do not account for an extra shader that was never processed processed--; }; await_workers(nb_workers, 0, shader_load_worker, processed, entry_count, dlg); } template void compile_shaders(uint nb_workers, unpacked_type& unpacked, u32 entry_count, shader_loading_dialog* dlg, Args&&... args) { atomic_t processed(0); std::function shader_comp_worker = [&](u32 stop_at) { u32 pos; // Processed is incremented before work starts in order to avoid two workers working on the same shader while (((pos = processed++) < stop_at) && !Emu.IsStopped()) { auto& entry = unpacked[pos]; m_storage.add_pipeline_entry(std::get<1>(entry), std::get<2>(entry), std::get<0>(entry), std::forward(args)...); } // Do not account for an extra shader that was never processed processed--; }; await_workers(nb_workers, 1, shader_comp_worker, processed, entry_count, dlg); } void await_workers(uint nb_workers, u8 step, std::function& worker, atomic_t& processed, u32 entry_count, shader_loading_dialog* dlg) { if (nb_workers == 1) { steady_clock::time_point last_update; // Call the worker function directly, stopping it prematurely to be able update the screen u32 stop_at = 0; do { stop_at = std::min(stop_at + 10, entry_count); worker(stop_at); // Only update the screen at about 60fps since updating it everytime slows down the process steady_clock::time_point now = steady_clock::now(); if ((std::chrono::duration_cast(now - last_update) > 16ms) || (stop_at == entry_count)) { dlg->update_msg(step, get_message(step, stop_at, entry_count)); dlg->set_value(step, stop_at); last_update = now; } } while (stop_at < entry_count && !Emu.IsStopped()); } else { named_thread_group workers("RSX Worker ", nb_workers, [&]() { worker(entry_count); }); u32 current_progress = 0; u32 last_update_progress = 0; while ((current_progress < entry_count) && !Emu.IsStopped()) { thread_ctrl::wait_for(16'000); // Around 60fps should be good enough if (Emu.IsStopped()) break; current_progress = std::min(processed.load(), entry_count); if (last_update_progress != current_progress) { last_update_progress = current_progress; dlg->update_msg(step, get_message(step, current_progress, entry_count)); dlg->set_value(step, current_progress); } } } if (!Emu.IsStopped()) { ensure(processed == entry_count); } } public: shaders_cache(backend_storage& storage, std::string pipeline_class, std::string version_prefix_str = "v1") : version_prefix(std::move(version_prefix_str)) , pipeline_class_name(std::move(pipeline_class)) , m_storage(storage) { if (!g_cfg.video.disable_on_disk_shader_cache) { if (std::string cache_path = rpcs3::cache::get_ppu_cache(); !cache_path.empty()) { root_path = std::move(cache_path) + "shaders_cache/"; } } } template void load(shader_loading_dialog* dlg, Args&& ...args) { if (root_path.empty()) { return; } std::string directory_path = root_path + "/pipelines/" + pipeline_class_name + "/" + version_prefix; fs::dir root = fs::dir(directory_path); if (!root) { fs::create_path(directory_path); fs::create_path(root_path + "/raw"); return; } std::vector entries; for (auto&& tmp : root) { if (tmp.is_directory) continue; entries.push_back(tmp); } u32 entry_count = ::size32(entries); if (!entry_count) return; root.rewind(); // Progress dialog std::unique_ptr fallback_dlg; if (!dlg) { fallback_dlg = std::make_unique(); dlg = fallback_dlg.get(); } dlg->create("Preloading cached shaders from disk.\nPlease wait...", "Shader Compilation"); dlg->set_limit(0, entry_count); dlg->set_limit(1, entry_count); dlg->update_msg(0, get_message(0, 0, entry_count)); dlg->update_msg(1, get_message(1, 0, entry_count)); // Preload everything needed to compile the shaders unpacked_type unpacked; uint nb_workers = g_cfg.video.renderer == video_renderer::vulkan ? utils::get_thread_count() : 1; load_shaders(nb_workers, unpacked, directory_path, entries, entry_count, dlg); // Account for any invalid entries entry_count = unpacked.size(); compile_shaders(nb_workers, unpacked, entry_count, dlg, std::forward(args)...); dlg->refresh(); dlg->close(); } void store(const pipeline_storage_type &pipeline, const RSXVertexProgram &vp, const RSXFragmentProgram &fp) { if (root_path.empty()) { return; } if (vp.jump_table.size() > 32) { rsx_log.error("shaders_cache: vertex program has more than 32 jump addresses. Entry not saved to cache"); return; } pipeline_data data = pack(pipeline, vp, fp); std::string fp_name = root_path + "/raw/" + fmt::format("%llX.fp", data.fragment_program_hash); std::string vp_name = root_path + "/raw/" + fmt::format("%llX.vp", data.vertex_program_hash); // Writeback to cache either if file does not exist or it is invalid (unexpected size) // Note: fs::write_file is not atomic, if the process is terminated in the middle an empty file is created if (fs::stat_t s{}; !fs::stat(fp_name, s) || s.size != fp.ucode_length) { fs::write_file(fp_name, fs::rewrite, fp.get_data(), fp.ucode_length); } if (fs::stat_t s{}; !fs::stat(vp_name, s) || s.size != vp.data.size() * sizeof(u32)) { fs::write_file(vp_name, fs::rewrite, vp.data); } u64 state_hash = 0; state_hash ^= rpcs3::hash_base(data.vp_ctrl); state_hash ^= rpcs3::hash_base(data.fp_ctrl); state_hash ^= rpcs3::hash_base(data.vp_texture_dimensions); state_hash ^= rpcs3::hash_base(data.fp_texture_dimensions); state_hash ^= rpcs3::hash_base(data.fp_texcoord_control); state_hash ^= rpcs3::hash_base(data.fp_height); state_hash ^= rpcs3::hash_base(data.fp_pixel_layout); state_hash ^= rpcs3::hash_base(data.fp_lighting_flags); state_hash ^= rpcs3::hash_base(data.fp_shadow_textures); state_hash ^= rpcs3::hash_base(data.fp_redirected_textures); state_hash ^= rpcs3::hash_base(data.vp_multisampled_textures); state_hash ^= rpcs3::hash_base(data.fp_multisampled_textures); const std::string pipeline_file_name = fmt::format("%llX+%llX+%llX+%llX.bin", data.vertex_program_hash, data.fragment_program_hash, data.pipeline_storage_hash, state_hash); const std::string pipeline_path = root_path + "/pipelines/" + pipeline_class_name + "/" + version_prefix + "/" + pipeline_file_name; fs::write_file(pipeline_path, fs::rewrite, &data, sizeof(data)); } RSXVertexProgram load_vp_raw(u64 program_hash) const { RSXVertexProgram vp = {}; fs::file f(fmt::format("%s/raw/%llX.vp", root_path, program_hash)); if (f) f.read(vp.data, f.size() / sizeof(u32)); return vp; } RSXFragmentProgram load_fp_raw(u64 program_hash) { fs::file f(fmt::format("%s/raw/%llX.fp", root_path, program_hash)); RSXFragmentProgram fp = {}; const u32 size = fp.ucode_length = f ? ::size32(f) : 0; if (!size) { return fp; } auto buf = std::make_unique(size); fp.data = buf.get(); f.read(buf.get(), size); fragment_program_data[fragment_program_data.push_begin()] = std::move(buf); return fp; } std::tuple unpack(pipeline_data &data) { std::tuple result; auto& [pipeline, vp, fp] = result; vp = load_vp_raw(data.vertex_program_hash); fp = load_fp_raw(data.fragment_program_hash); pipeline = data.pipeline_properties; vp.output_mask = data.vp_ctrl; vp.texture_state.texture_dimensions = data.vp_texture_dimensions; vp.texture_state.multisampled_textures = data.vp_multisampled_textures; vp.base_address = data.vp_base_address; vp.entry = data.vp_entry; pack_bitset(vp.instruction_mask, data.vp_instruction_mask); for (u8 index = 0; index < 32; ++index) { const auto address = data.vp_jump_table[index]; if (address == u16{umax}) { // End of list marker break; } vp.jump_table.emplace(address); } fp.ctrl = data.fp_ctrl; fp.texture_state.texture_dimensions = data.fp_texture_dimensions; fp.texture_state.shadow_textures = data.fp_shadow_textures; fp.texture_state.redirected_textures = data.fp_redirected_textures; fp.texture_state.multisampled_textures = data.fp_multisampled_textures; fp.texcoord_control_mask = data.fp_texcoord_control; fp.two_sided_lighting = !!(data.fp_lighting_flags & 0x1); return result; } pipeline_data pack(const pipeline_storage_type &pipeline, const RSXVertexProgram &vp, const RSXFragmentProgram &fp) { pipeline_data data_block = {}; data_block.pipeline_properties = pipeline; data_block.vertex_program_hash = m_storage.get_hash(vp); data_block.fragment_program_hash = m_storage.get_hash(fp); data_block.pipeline_storage_hash = m_storage.get_hash(pipeline); data_block.vp_ctrl = vp.output_mask; data_block.vp_texture_dimensions = vp.texture_state.texture_dimensions; data_block.vp_multisampled_textures = vp.texture_state.multisampled_textures; data_block.vp_base_address = vp.base_address; data_block.vp_entry = vp.entry; unpack_bitset(vp.instruction_mask, data_block.vp_instruction_mask); u8 index = 0; while (index < 32) { if (!index && !vp.jump_table.empty()) { for (auto &address : vp.jump_table) { data_block.vp_jump_table[index++] = static_cast(address); } } else { // End of list marker data_block.vp_jump_table[index] = u16{umax}; break; } } data_block.fp_ctrl = fp.ctrl; data_block.fp_texture_dimensions = fp.texture_state.texture_dimensions; data_block.fp_texcoord_control = fp.texcoord_control_mask; data_block.fp_lighting_flags = u16(fp.two_sided_lighting); data_block.fp_shadow_textures = fp.texture_state.shadow_textures; data_block.fp_redirected_textures = fp.texture_state.redirected_textures; data_block.fp_multisampled_textures = fp.texture_state.multisampled_textures; return data_block; } }; namespace vertex_cache { // A null vertex cache template class default_vertex_cache { public: virtual ~default_vertex_cache() = default; virtual storage_type* find_vertex_range(uptr /*local_addr*/, upload_format, u32 /*data_length*/) { return nullptr; } virtual void store_range(uptr /*local_addr*/, upload_format, u32 /*data_length*/, u32 /*offset_in_heap*/) {} virtual void purge() {} }; // A weak vertex cache with no data checks or memory range locks // Of limited use since contents are only guaranteed to be valid once per frame // TODO: Strict vertex cache with range locks template struct uploaded_range { uptr local_address; upload_format buffer_format; u32 offset_in_heap; u32 data_length; }; template class weak_vertex_cache : public default_vertex_cache, upload_format> { using storage_type = uploaded_range; private: std::unordered_map> vertex_ranges; public: storage_type* find_vertex_range(uptr local_addr, upload_format fmt, u32 data_length) override { //const auto data_end = local_addr + data_length; for (auto &v : vertex_ranges[local_addr]) { // NOTE: This has to match exactly. Using sized shortcuts such as >= comparison causes artifacting in some applications (UC1) if (v.buffer_format == fmt && v.data_length == data_length) return &v; } return nullptr; } void store_range(uptr local_addr, upload_format fmt, u32 data_length, u32 offset_in_heap) override { storage_type v = {}; v.buffer_format = fmt; v.data_length = data_length; v.local_address = local_addr; v.offset_in_heap = offset_in_heap; vertex_ranges[local_addr].push_back(v); } void purge() override { vertex_ranges.clear(); } }; } }