// SPDX-FileCopyrightText: Copyright 2024 shadPS4 Emulator Project // SPDX-License-Identifier: GPL-2.0-or-later #pragma once #include #include #include #include #include #include "common/assert.h" #include "common/types.h" #include "shader_recompiler/backend/bindings.h" #include "shader_recompiler/frontend/copy_shader.h" #include "shader_recompiler/ir/attribute.h" #include "shader_recompiler/ir/passes/srt.h" #include "shader_recompiler/ir/reg.h" #include "shader_recompiler/ir/type.h" #include "shader_recompiler/params.h" #include "shader_recompiler/runtime_info.h" #include "video_core/amdgpu/liverpool.h" #include "video_core/amdgpu/resource.h" namespace Shader { static constexpr size_t NumUserDataRegs = 16; static constexpr size_t MaxUboSize = 65536; static constexpr size_t MaxUboDwords = MaxUboSize >> 2; enum class TextureType : u32 { Color1D, ColorArray1D, Color2D, ColorArray2D, Color3D, ColorCube, Buffer, }; constexpr u32 NUM_TEXTURE_TYPES = 7; struct Info; struct BufferResource { u32 sharp_idx; IR::Type used_types; AmdGpu::Buffer inline_cbuf; bool is_gds_buffer{}; bool is_instance_data{}; u8 instance_attrib{}; bool is_written{}; bool IsStorage(AmdGpu::Buffer buffer) const noexcept { return buffer.GetSize() > MaxUboSize || is_written || is_gds_buffer; } constexpr AmdGpu::Buffer GetSharp(const Info& info) const noexcept; }; using BufferResourceList = boost::container::small_vector; struct TextureBufferResource { u32 sharp_idx; bool is_written{}; constexpr AmdGpu::Buffer GetSharp(const Info& info) const noexcept; }; using TextureBufferResourceList = boost::container::small_vector; struct ImageResource { u32 sharp_idx; bool is_storage{}; bool is_depth{}; bool is_atomic{}; bool is_array{}; constexpr AmdGpu::Image GetSharp(const Info& info) const noexcept; }; using ImageResourceList = boost::container::small_vector; struct SamplerResource { u32 sharp_idx; AmdGpu::Sampler inline_sampler{}; u32 associated_image : 4; u32 disable_aniso : 1; constexpr AmdGpu::Sampler GetSharp(const Info& info) const noexcept; }; using SamplerResourceList = boost::container::small_vector; struct FMaskResource { u32 sharp_idx; constexpr AmdGpu::Image GetSharp(const Info& info) const noexcept; }; using FMaskResourceList = boost::container::small_vector; struct PushData { static constexpr u32 BufOffsetIndex = 2; static constexpr u32 UdRegsIndex = 4; u32 step0; u32 step1; std::array buf_offsets; std::array ud_regs; void AddOffset(u32 binding, u32 offset) { ASSERT(offset < 256 && binding < buf_offsets.size()); buf_offsets[binding] = offset; } void AddTexelOffset(u32 binding, u32 multiplier, u32 texel_offset) { ASSERT(texel_offset < 64 && multiplier < 16); buf_offsets[binding] = texel_offset | ((std::bit_width(multiplier) - 1) << 6); } }; static_assert(sizeof(PushData) <= 128, "PushData size is greater than minimum size guaranteed by Vulkan spec"); /** * Contains general information generated by the shader recompiler for an input program. */ struct Info { struct AttributeFlags { bool Get(IR::Attribute attrib, u32 comp = 0) const { return flags[Index(attrib)] & (1 << comp); } bool GetAny(IR::Attribute attrib) const { return flags[Index(attrib)]; } void Set(IR::Attribute attrib, u32 comp = 0) { flags[Index(attrib)] |= (1 << comp); } u32 NumComponents(IR::Attribute attrib) const { return 4; } static size_t Index(IR::Attribute attrib) { return static_cast(attrib); } std::array flags; }; AttributeFlags loads{}; AttributeFlags stores{}; struct UserDataMask { void Set(IR::ScalarReg reg) noexcept { mask |= 1 << static_cast(reg); } u32 Index(IR::ScalarReg reg) const noexcept { const u32 reg_mask = (1 << static_cast(reg)) - 1; return std::popcount(mask & reg_mask); } u32 NumRegs() const noexcept { return std::popcount(mask); } u32 mask; }; UserDataMask ud_mask{}; CopyShaderData gs_copy_data; BufferResourceList buffers; TextureBufferResourceList texture_buffers; ImageResourceList images; SamplerResourceList samplers; FMaskResourceList fmasks; PersistentSrtInfo srt_info; std::vector flattened_ud_buf; std::span user_data; Stage stage; u64 pgm_hash{}; VAddr pgm_base; bool has_storage_images{}; bool has_image_buffers{}; bool has_texel_buffers{}; bool has_discard{}; bool has_image_gather{}; bool has_image_query{}; bool uses_lane_id{}; bool uses_group_quad{}; bool uses_group_ballot{}; bool uses_shared{}; bool uses_fp16{}; bool uses_fp64{}; bool translation_failed{}; // indicates that shader has unsupported instructions bool has_readconst{}; u8 mrt_mask{0u}; bool has_fetch_shader{false}; u32 fetch_shader_sgpr_base{0u}; explicit Info(Stage stage_, ShaderParams params) : stage{stage_}, pgm_hash{params.hash}, pgm_base{params.Base()}, user_data{params.user_data} {} template inline T ReadUdSharp(u32 sharp_idx) const noexcept { return *reinterpret_cast(&flattened_ud_buf[sharp_idx]); } template T ReadUdReg(u32 ptr_index, u32 dword_offset) const noexcept { T data; const u32* base = user_data.data(); if (ptr_index != IR::NumScalarRegs) { std::memcpy(&base, &user_data[ptr_index], sizeof(base)); base = reinterpret_cast(VAddr(base) & 0xFFFFFFFFFFFFULL); } std::memcpy(&data, base + dword_offset, sizeof(T)); return data; } void PushUd(Backend::Bindings& bnd, PushData& push) const { u32 mask = ud_mask.mask; while (mask) { const u32 index = std::countr_zero(mask); ASSERT(bnd.user_data < NumUserDataRegs && index < NumUserDataRegs); mask &= ~(1U << index); push.ud_regs[bnd.user_data++] = user_data[index]; } } void AddBindings(Backend::Bindings& bnd) const { const auto total_buffers = buffers.size() + texture_buffers.size() + (has_readconst ? 1 : 0); bnd.buffer += total_buffers; bnd.unified += total_buffers + images.size() + samplers.size(); bnd.user_data += ud_mask.NumRegs(); } void RefreshFlatBuf() { flattened_ud_buf.resize(srt_info.flattened_bufsize_dw); ASSERT(user_data.size() <= NumUserDataRegs); std::memcpy(flattened_ud_buf.data(), user_data.data(), user_data.size_bytes()); // Run the JIT program to walk the SRT and write the leaves to a flat buffer if (srt_info.walker_func) { srt_info.walker_func(user_data.data(), flattened_ud_buf.data()); } } }; constexpr AmdGpu::Buffer BufferResource::GetSharp(const Info& info) const noexcept { return inline_cbuf ? inline_cbuf : info.ReadUdSharp(sharp_idx); } constexpr AmdGpu::Buffer TextureBufferResource::GetSharp(const Info& info) const noexcept { return info.ReadUdSharp(sharp_idx); } constexpr AmdGpu::Image ImageResource::GetSharp(const Info& info) const noexcept { const auto image = info.ReadUdSharp(sharp_idx); if (!image.Valid()) { // Fall back to null image if unbound. return AmdGpu::Image::Null(); } return image; } constexpr AmdGpu::Sampler SamplerResource::GetSharp(const Info& info) const noexcept { return inline_sampler ? inline_sampler : info.ReadUdSharp(sharp_idx); } constexpr AmdGpu::Image FMaskResource::GetSharp(const Info& info) const noexcept { return info.ReadUdSharp(sharp_idx); } } // namespace Shader template <> struct fmt::formatter { constexpr auto parse(format_parse_context& ctx) { return ctx.begin(); } auto format(const Shader::Stage stage, format_context& ctx) const { constexpr static std::array names = {"fs", "vs", "gs", "es", "hs", "ls", "cs"}; return fmt::format_to(ctx.out(), "{}", names[static_cast(stage)]); } };