Merge pull request #9535 from bylaws/master
Port over several shader-compiler fixes from skyline
This commit is contained in:
		
							
								
								
									
										2
									
								
								externals/sirit
									
									
									
									
										vendored
									
									
								
							
							
								
								
								
								
								
							
						
						
									
										2
									
								
								externals/sirit
									
									
									
									
										vendored
									
									
								
							 Submodule externals/sirit updated: d7ad93a888...ab75463999
									
								
							| @@ -321,8 +321,12 @@ Id EmitGetAttribute(EmitContext& ctx, IR::Attribute attr, Id vertex) { | ||||
|     case IR::Attribute::PositionY: | ||||
|     case IR::Attribute::PositionZ: | ||||
|     case IR::Attribute::PositionW: | ||||
|         return ctx.OpLoad(ctx.F32[1], AttrPointer(ctx, ctx.input_f32, vertex, ctx.input_position, | ||||
|                                                   ctx.Const(element))); | ||||
|         return ctx.OpLoad( | ||||
|             ctx.F32[1], | ||||
|             ctx.need_input_position_indirect | ||||
|                 ? AttrPointer(ctx, ctx.input_f32, vertex, ctx.input_position, ctx.u32_zero_value, | ||||
|                               ctx.Const(element)) | ||||
|                 : AttrPointer(ctx, ctx.input_f32, vertex, ctx.input_position, ctx.Const(element))); | ||||
|     case IR::Attribute::InstanceId: | ||||
|         if (ctx.profile.support_vertex_instance_id) { | ||||
|             return ctx.OpBitcast(ctx.F32[1], ctx.OpLoad(ctx.U32[1], ctx.instance_id)); | ||||
|   | ||||
| @@ -58,11 +58,10 @@ Id SelectValue(EmitContext& ctx, Id in_range, Id value, Id src_thread_id) { | ||||
|         ctx.OpGroupNonUniformShuffle(ctx.U32[1], SubgroupScope(ctx), value, src_thread_id), value); | ||||
| } | ||||
|  | ||||
| Id GetUpperClamp(EmitContext& ctx, Id invocation_id, Id clamp) { | ||||
|     const Id thirty_two{ctx.Const(32u)}; | ||||
|     const Id is_upper_partition{ctx.OpSGreaterThanEqual(ctx.U1, invocation_id, thirty_two)}; | ||||
|     const Id upper_clamp{ctx.OpIAdd(ctx.U32[1], thirty_two, clamp)}; | ||||
|     return ctx.OpSelect(ctx.U32[1], is_upper_partition, upper_clamp, clamp); | ||||
| Id AddPartitionBase(EmitContext& ctx, Id thread_id) { | ||||
|     const Id partition_idx{ctx.OpShiftRightLogical(ctx.U32[1], GetThreadId(ctx), ctx.Const(5u))}; | ||||
|     const Id partition_base{ctx.OpShiftLeftLogical(ctx.U32[1], partition_idx, ctx.Const(5u))}; | ||||
|     return ctx.OpIAdd(ctx.U32[1], thread_id, partition_base); | ||||
| } | ||||
| } // Anonymous namespace | ||||
|  | ||||
| @@ -145,64 +144,63 @@ Id EmitSubgroupGeMask(EmitContext& ctx) { | ||||
| Id EmitShuffleIndex(EmitContext& ctx, IR::Inst* inst, Id value, Id index, Id clamp, | ||||
|                     Id segmentation_mask) { | ||||
|     const Id not_seg_mask{ctx.OpNot(ctx.U32[1], segmentation_mask)}; | ||||
|     const Id thread_id{GetThreadId(ctx)}; | ||||
|     if (ctx.profile.warp_size_potentially_larger_than_guest) { | ||||
|         const Id thirty_two{ctx.Const(32u)}; | ||||
|         const Id is_upper_partition{ctx.OpSGreaterThanEqual(ctx.U1, thread_id, thirty_two)}; | ||||
|         const Id upper_index{ctx.OpIAdd(ctx.U32[1], thirty_two, index)}; | ||||
|         const Id upper_clamp{ctx.OpIAdd(ctx.U32[1], thirty_two, clamp)}; | ||||
|         index = ctx.OpSelect(ctx.U32[1], is_upper_partition, upper_index, index); | ||||
|         clamp = ctx.OpSelect(ctx.U32[1], is_upper_partition, upper_clamp, clamp); | ||||
|     } | ||||
|     const Id thread_id{EmitLaneId(ctx)}; | ||||
|     const Id min_thread_id{ComputeMinThreadId(ctx, thread_id, segmentation_mask)}; | ||||
|     const Id max_thread_id{ComputeMaxThreadId(ctx, min_thread_id, clamp, not_seg_mask)}; | ||||
|  | ||||
|     const Id lhs{ctx.OpBitwiseAnd(ctx.U32[1], index, not_seg_mask)}; | ||||
|     const Id src_thread_id{ctx.OpBitwiseOr(ctx.U32[1], lhs, min_thread_id)}; | ||||
|     Id src_thread_id{ctx.OpBitwiseOr(ctx.U32[1], lhs, min_thread_id)}; | ||||
|     const Id in_range{ctx.OpSLessThanEqual(ctx.U1, src_thread_id, max_thread_id)}; | ||||
|  | ||||
|     if (ctx.profile.warp_size_potentially_larger_than_guest) { | ||||
|         src_thread_id = AddPartitionBase(ctx, src_thread_id); | ||||
|     } | ||||
|  | ||||
|     SetInBoundsFlag(inst, in_range); | ||||
|     return SelectValue(ctx, in_range, value, src_thread_id); | ||||
| } | ||||
|  | ||||
| Id EmitShuffleUp(EmitContext& ctx, IR::Inst* inst, Id value, Id index, Id clamp, | ||||
|                  Id segmentation_mask) { | ||||
|     const Id thread_id{GetThreadId(ctx)}; | ||||
|     if (ctx.profile.warp_size_potentially_larger_than_guest) { | ||||
|         clamp = GetUpperClamp(ctx, thread_id, clamp); | ||||
|     } | ||||
|     const Id thread_id{EmitLaneId(ctx)}; | ||||
|     const Id max_thread_id{GetMaxThreadId(ctx, thread_id, clamp, segmentation_mask)}; | ||||
|     const Id src_thread_id{ctx.OpISub(ctx.U32[1], thread_id, index)}; | ||||
|     Id src_thread_id{ctx.OpISub(ctx.U32[1], thread_id, index)}; | ||||
|     const Id in_range{ctx.OpSGreaterThanEqual(ctx.U1, src_thread_id, max_thread_id)}; | ||||
|  | ||||
|     if (ctx.profile.warp_size_potentially_larger_than_guest) { | ||||
|         src_thread_id = AddPartitionBase(ctx, src_thread_id); | ||||
|     } | ||||
|  | ||||
|     SetInBoundsFlag(inst, in_range); | ||||
|     return SelectValue(ctx, in_range, value, src_thread_id); | ||||
| } | ||||
|  | ||||
| Id EmitShuffleDown(EmitContext& ctx, IR::Inst* inst, Id value, Id index, Id clamp, | ||||
|                    Id segmentation_mask) { | ||||
|     const Id thread_id{GetThreadId(ctx)}; | ||||
|     if (ctx.profile.warp_size_potentially_larger_than_guest) { | ||||
|         clamp = GetUpperClamp(ctx, thread_id, clamp); | ||||
|     } | ||||
|     const Id thread_id{EmitLaneId(ctx)}; | ||||
|     const Id max_thread_id{GetMaxThreadId(ctx, thread_id, clamp, segmentation_mask)}; | ||||
|     const Id src_thread_id{ctx.OpIAdd(ctx.U32[1], thread_id, index)}; | ||||
|     Id src_thread_id{ctx.OpIAdd(ctx.U32[1], thread_id, index)}; | ||||
|     const Id in_range{ctx.OpSLessThanEqual(ctx.U1, src_thread_id, max_thread_id)}; | ||||
|  | ||||
|     if (ctx.profile.warp_size_potentially_larger_than_guest) { | ||||
|         src_thread_id = AddPartitionBase(ctx, src_thread_id); | ||||
|     } | ||||
|  | ||||
|     SetInBoundsFlag(inst, in_range); | ||||
|     return SelectValue(ctx, in_range, value, src_thread_id); | ||||
| } | ||||
|  | ||||
| Id EmitShuffleButterfly(EmitContext& ctx, IR::Inst* inst, Id value, Id index, Id clamp, | ||||
|                         Id segmentation_mask) { | ||||
|     const Id thread_id{GetThreadId(ctx)}; | ||||
|     if (ctx.profile.warp_size_potentially_larger_than_guest) { | ||||
|         clamp = GetUpperClamp(ctx, thread_id, clamp); | ||||
|     } | ||||
|     const Id thread_id{EmitLaneId(ctx)}; | ||||
|     const Id max_thread_id{GetMaxThreadId(ctx, thread_id, clamp, segmentation_mask)}; | ||||
|     const Id src_thread_id{ctx.OpBitwiseXor(ctx.U32[1], thread_id, index)}; | ||||
|     Id src_thread_id{ctx.OpBitwiseXor(ctx.U32[1], thread_id, index)}; | ||||
|     const Id in_range{ctx.OpSLessThanEqual(ctx.U1, src_thread_id, max_thread_id)}; | ||||
|  | ||||
|     if (ctx.profile.warp_size_potentially_larger_than_guest) { | ||||
|         src_thread_id = AddPartitionBase(ctx, src_thread_id); | ||||
|     } | ||||
|  | ||||
|     SetInBoundsFlag(inst, in_range); | ||||
|     return SelectValue(ctx, in_range, value, src_thread_id); | ||||
| } | ||||
|   | ||||
| @@ -544,7 +544,7 @@ void EmitContext::DefineCommonTypes(const Info& info) { | ||||
|         U16 = Name(TypeInt(16, false), "u16"); | ||||
|         S16 = Name(TypeInt(16, true), "s16"); | ||||
|     } | ||||
|     if (info.uses_int64) { | ||||
|     if (info.uses_int64 && profile.support_int64) { | ||||
|         AddCapability(spv::Capability::Int64); | ||||
|         U64 = Name(TypeInt(64, false), "u64"); | ||||
|     } | ||||
| @@ -721,9 +721,21 @@ void EmitContext::DefineAttributeMemAccess(const Info& info) { | ||||
|         size_t label_index{0}; | ||||
|         if (info.loads.AnyComponent(IR::Attribute::PositionX)) { | ||||
|             AddLabel(labels[label_index]); | ||||
|             const Id pointer{is_array | ||||
|                                  ? OpAccessChain(input_f32, input_position, vertex, masked_index) | ||||
|                                  : OpAccessChain(input_f32, input_position, masked_index)}; | ||||
|             const Id pointer{[&]() { | ||||
|                 if (need_input_position_indirect) { | ||||
|                     if (is_array) | ||||
|                         return OpAccessChain(input_f32, input_position, vertex, u32_zero_value, | ||||
|                                              masked_index); | ||||
|                     else | ||||
|                         return OpAccessChain(input_f32, input_position, u32_zero_value, | ||||
|                                              masked_index); | ||||
|                 } else { | ||||
|                     if (is_array) | ||||
|                         return OpAccessChain(input_f32, input_position, vertex, masked_index); | ||||
|                     else | ||||
|                         return OpAccessChain(input_f32, input_position, masked_index); | ||||
|                 } | ||||
|             }()}; | ||||
|             const Id result{OpLoad(F32[1], pointer)}; | ||||
|             OpReturnValue(result); | ||||
|             ++label_index; | ||||
| @@ -1367,12 +1379,25 @@ void EmitContext::DefineInputs(const IR::Program& program) { | ||||
|         Decorate(layer, spv::Decoration::Flat); | ||||
|     } | ||||
|     if (loads.AnyComponent(IR::Attribute::PositionX)) { | ||||
|         const bool is_fragment{stage != Stage::Fragment}; | ||||
|         const spv::BuiltIn built_in{is_fragment ? spv::BuiltIn::Position : spv::BuiltIn::FragCoord}; | ||||
|         input_position = DefineInput(*this, F32[4], true, built_in); | ||||
|         if (profile.support_geometry_shader_passthrough) { | ||||
|             if (info.passthrough.AnyComponent(IR::Attribute::PositionX)) { | ||||
|                 Decorate(input_position, spv::Decoration::PassthroughNV); | ||||
|         const bool is_fragment{stage == Stage::Fragment}; | ||||
|         if (!is_fragment && profile.has_broken_spirv_position_input) { | ||||
|             need_input_position_indirect = true; | ||||
|  | ||||
|             const Id input_position_struct = TypeStruct(F32[4]); | ||||
|             input_position = DefineInput(*this, input_position_struct, true); | ||||
|  | ||||
|             MemberDecorate(input_position_struct, 0, spv::Decoration::BuiltIn, | ||||
|                            static_cast<unsigned>(spv::BuiltIn::Position)); | ||||
|             Decorate(input_position_struct, spv::Decoration::Block); | ||||
|         } else { | ||||
|             const spv::BuiltIn built_in{is_fragment ? spv::BuiltIn::FragCoord | ||||
|                                                     : spv::BuiltIn::Position}; | ||||
|             input_position = DefineInput(*this, F32[4], true, built_in); | ||||
|  | ||||
|             if (profile.support_geometry_shader_passthrough) { | ||||
|                 if (info.passthrough.AnyComponent(IR::Attribute::PositionX)) { | ||||
|                     Decorate(input_position, spv::Decoration::PassthroughNV); | ||||
|                 } | ||||
|             } | ||||
|         } | ||||
|     } | ||||
|   | ||||
| @@ -280,6 +280,7 @@ public: | ||||
|     Id write_global_func_u32x2{}; | ||||
|     Id write_global_func_u32x4{}; | ||||
|  | ||||
|     bool need_input_position_indirect{}; | ||||
|     Id input_position{}; | ||||
|     std::array<Id, 32> input_generics{}; | ||||
|  | ||||
|   | ||||
| @@ -171,6 +171,70 @@ std::map<IR::Attribute, IR::Attribute> GenerateLegacyToGenericMappings( | ||||
|     } | ||||
|     return mapping; | ||||
| } | ||||
|  | ||||
| void EmitGeometryPassthrough(IR::IREmitter& ir, const IR::Program& program, | ||||
|                              const Shader::VaryingState& passthrough_mask, | ||||
|                              bool passthrough_position, | ||||
|                              std::optional<IR::Attribute> passthrough_layer_attr) { | ||||
|     for (u32 i = 0; i < program.output_vertices; i++) { | ||||
|         // Assign generics from input | ||||
|         for (u32 j = 0; j < 32; j++) { | ||||
|             if (!passthrough_mask.Generic(j)) { | ||||
|                 continue; | ||||
|             } | ||||
|  | ||||
|             const IR::Attribute attr = IR::Attribute::Generic0X + (j * 4); | ||||
|             ir.SetAttribute(attr + 0, ir.GetAttribute(attr + 0, ir.Imm32(i)), ir.Imm32(0)); | ||||
|             ir.SetAttribute(attr + 1, ir.GetAttribute(attr + 1, ir.Imm32(i)), ir.Imm32(0)); | ||||
|             ir.SetAttribute(attr + 2, ir.GetAttribute(attr + 2, ir.Imm32(i)), ir.Imm32(0)); | ||||
|             ir.SetAttribute(attr + 3, ir.GetAttribute(attr + 3, ir.Imm32(i)), ir.Imm32(0)); | ||||
|         } | ||||
|  | ||||
|         if (passthrough_position) { | ||||
|             // Assign position from input | ||||
|             const IR::Attribute attr = IR::Attribute::PositionX; | ||||
|             ir.SetAttribute(attr + 0, ir.GetAttribute(attr + 0, ir.Imm32(i)), ir.Imm32(0)); | ||||
|             ir.SetAttribute(attr + 1, ir.GetAttribute(attr + 1, ir.Imm32(i)), ir.Imm32(0)); | ||||
|             ir.SetAttribute(attr + 2, ir.GetAttribute(attr + 2, ir.Imm32(i)), ir.Imm32(0)); | ||||
|             ir.SetAttribute(attr + 3, ir.GetAttribute(attr + 3, ir.Imm32(i)), ir.Imm32(0)); | ||||
|         } | ||||
|  | ||||
|         if (passthrough_layer_attr) { | ||||
|             // Assign layer | ||||
|             ir.SetAttribute(IR::Attribute::Layer, ir.GetAttribute(*passthrough_layer_attr), | ||||
|                             ir.Imm32(0)); | ||||
|         } | ||||
|  | ||||
|         // Emit vertex | ||||
|         ir.EmitVertex(ir.Imm32(0)); | ||||
|     } | ||||
|     ir.EndPrimitive(ir.Imm32(0)); | ||||
| } | ||||
|  | ||||
| u32 GetOutputTopologyVertices(OutputTopology output_topology) { | ||||
|     switch (output_topology) { | ||||
|     case OutputTopology::PointList: | ||||
|         return 1; | ||||
|     case OutputTopology::LineStrip: | ||||
|         return 2; | ||||
|     default: | ||||
|         return 3; | ||||
|     } | ||||
| } | ||||
|  | ||||
| void LowerGeometryPassthrough(const IR::Program& program, const HostTranslateInfo& host_info) { | ||||
|     for (IR::Block* const block : program.blocks) { | ||||
|         for (IR::Inst& inst : block->Instructions()) { | ||||
|             if (inst.GetOpcode() == IR::Opcode::Epilogue) { | ||||
|                 IR::IREmitter ir{*block, IR::Block::InstructionList::s_iterator_to(inst)}; | ||||
|                 EmitGeometryPassthrough( | ||||
|                     ir, program, program.info.passthrough, | ||||
|                     program.info.passthrough.AnyComponent(IR::Attribute::PositionX), {}); | ||||
|             } | ||||
|         } | ||||
|     } | ||||
| } | ||||
|  | ||||
| } // Anonymous namespace | ||||
|  | ||||
| IR::Program TranslateProgram(ObjectPool<IR::Inst>& inst_pool, ObjectPool<IR::Block>& block_pool, | ||||
| @@ -198,6 +262,11 @@ IR::Program TranslateProgram(ObjectPool<IR::Inst>& inst_pool, ObjectPool<IR::Blo | ||||
|             for (size_t i = 0; i < program.info.passthrough.mask.size(); ++i) { | ||||
|                 program.info.passthrough.mask[i] = ((mask[i / 32] >> (i % 32)) & 1) == 0; | ||||
|             } | ||||
|  | ||||
|             if (!host_info.support_geometry_shader_passthrough) { | ||||
|                 program.output_vertices = GetOutputTopologyVertices(program.output_topology); | ||||
|                 LowerGeometryPassthrough(program, host_info); | ||||
|             } | ||||
|         } | ||||
|         break; | ||||
|     } | ||||
| @@ -223,7 +292,7 @@ IR::Program TranslateProgram(ObjectPool<IR::Inst>& inst_pool, ObjectPool<IR::Blo | ||||
|  | ||||
|     Optimization::PositionPass(env, program); | ||||
|  | ||||
|     Optimization::GlobalMemoryToStorageBufferPass(program); | ||||
|     Optimization::GlobalMemoryToStorageBufferPass(program, host_info); | ||||
|     Optimization::TexturePass(env, program, host_info); | ||||
|  | ||||
|     if (Settings::values.resolution_info.active) { | ||||
| @@ -342,17 +411,7 @@ IR::Program GenerateGeometryPassthrough(ObjectPool<IR::Inst>& inst_pool, | ||||
|     IR::Program program; | ||||
|     program.stage = Stage::Geometry; | ||||
|     program.output_topology = output_topology; | ||||
|     switch (output_topology) { | ||||
|     case OutputTopology::PointList: | ||||
|         program.output_vertices = 1; | ||||
|         break; | ||||
|     case OutputTopology::LineStrip: | ||||
|         program.output_vertices = 2; | ||||
|         break; | ||||
|     default: | ||||
|         program.output_vertices = 3; | ||||
|         break; | ||||
|     } | ||||
|     program.output_vertices = GetOutputTopologyVertices(output_topology); | ||||
|  | ||||
|     program.is_geometry_passthrough = false; | ||||
|     program.info.loads.mask = source_program.info.stores.mask; | ||||
| @@ -366,35 +425,8 @@ IR::Program GenerateGeometryPassthrough(ObjectPool<IR::Inst>& inst_pool, | ||||
|     node.data.block = current_block; | ||||
|  | ||||
|     IR::IREmitter ir{*current_block}; | ||||
|     for (u32 i = 0; i < program.output_vertices; i++) { | ||||
|         // Assign generics from input | ||||
|         for (u32 j = 0; j < 32; j++) { | ||||
|             if (!program.info.stores.Generic(j)) { | ||||
|                 continue; | ||||
|             } | ||||
|  | ||||
|             const IR::Attribute attr = IR::Attribute::Generic0X + (j * 4); | ||||
|             ir.SetAttribute(attr + 0, ir.GetAttribute(attr + 0, ir.Imm32(i)), ir.Imm32(0)); | ||||
|             ir.SetAttribute(attr + 1, ir.GetAttribute(attr + 1, ir.Imm32(i)), ir.Imm32(0)); | ||||
|             ir.SetAttribute(attr + 2, ir.GetAttribute(attr + 2, ir.Imm32(i)), ir.Imm32(0)); | ||||
|             ir.SetAttribute(attr + 3, ir.GetAttribute(attr + 3, ir.Imm32(i)), ir.Imm32(0)); | ||||
|         } | ||||
|  | ||||
|         // Assign position from input | ||||
|         const IR::Attribute attr = IR::Attribute::PositionX; | ||||
|         ir.SetAttribute(attr + 0, ir.GetAttribute(attr + 0, ir.Imm32(i)), ir.Imm32(0)); | ||||
|         ir.SetAttribute(attr + 1, ir.GetAttribute(attr + 1, ir.Imm32(i)), ir.Imm32(0)); | ||||
|         ir.SetAttribute(attr + 2, ir.GetAttribute(attr + 2, ir.Imm32(i)), ir.Imm32(0)); | ||||
|         ir.SetAttribute(attr + 3, ir.GetAttribute(attr + 3, ir.Imm32(i)), ir.Imm32(0)); | ||||
|  | ||||
|         // Assign layer | ||||
|         ir.SetAttribute(IR::Attribute::Layer, ir.GetAttribute(source_program.info.emulated_layer), | ||||
|                         ir.Imm32(0)); | ||||
|  | ||||
|         // Emit vertex | ||||
|         ir.EmitVertex(ir.Imm32(0)); | ||||
|     } | ||||
|     ir.EndPrimitive(ir.Imm32(0)); | ||||
|     EmitGeometryPassthrough(ir, program, program.info.stores, true, | ||||
|                             source_program.info.emulated_layer); | ||||
|  | ||||
|     IR::Block* return_block{block_pool.Create(inst_pool)}; | ||||
|     IR::IREmitter{*return_block}.Epilogue(); | ||||
|   | ||||
| @@ -15,6 +15,9 @@ struct HostTranslateInfo { | ||||
|     bool needs_demote_reorder{}; ///< True when the device needs DemoteToHelperInvocation reordered | ||||
|     bool support_snorm_render_buffer{};  ///< True when the device supports SNORM render buffers | ||||
|     bool support_viewport_index_layer{}; ///< True when the device supports gl_Layer in VS | ||||
|     u32 min_ssbo_alignment{};            ///< Minimum alignment supported by the device for SSBOs | ||||
|     bool support_geometry_shader_passthrough{}; ///< True when the device supports geometry | ||||
|                                                 ///< passthrough shaders | ||||
| }; | ||||
|  | ||||
| } // namespace Shader | ||||
|   | ||||
| @@ -11,6 +11,7 @@ | ||||
| #include "shader_recompiler/frontend/ir/breadth_first_search.h" | ||||
| #include "shader_recompiler/frontend/ir/ir_emitter.h" | ||||
| #include "shader_recompiler/frontend/ir/value.h" | ||||
| #include "shader_recompiler/host_translate_info.h" | ||||
| #include "shader_recompiler/ir_opt/passes.h" | ||||
|  | ||||
| namespace Shader::Optimization { | ||||
| @@ -402,7 +403,7 @@ void CollectStorageBuffers(IR::Block& block, IR::Inst& inst, StorageInfo& info) | ||||
| } | ||||
|  | ||||
| /// Returns the offset in indices (not bytes) for an equivalent storage instruction | ||||
| IR::U32 StorageOffset(IR::Block& block, IR::Inst& inst, StorageBufferAddr buffer) { | ||||
| IR::U32 StorageOffset(IR::Block& block, IR::Inst& inst, StorageBufferAddr buffer, u32 alignment) { | ||||
|     IR::IREmitter ir{block, IR::Block::InstructionList::s_iterator_to(inst)}; | ||||
|     IR::U32 offset; | ||||
|     if (const std::optional<LowAddrInfo> low_addr{TrackLowAddress(&inst)}) { | ||||
| @@ -415,7 +416,10 @@ IR::U32 StorageOffset(IR::Block& block, IR::Inst& inst, StorageBufferAddr buffer | ||||
|     } | ||||
|     // Subtract the least significant 32 bits from the guest offset. The result is the storage | ||||
|     // buffer offset in bytes. | ||||
|     const IR::U32 low_cbuf{ir.GetCbuf(ir.Imm32(buffer.index), ir.Imm32(buffer.offset))}; | ||||
|     IR::U32 low_cbuf{ir.GetCbuf(ir.Imm32(buffer.index), ir.Imm32(buffer.offset))}; | ||||
|  | ||||
|     // Align the offset base to match the host alignment requirements | ||||
|     low_cbuf = ir.BitwiseAnd(low_cbuf, ir.Imm32(~(alignment - 1U))); | ||||
|     return ir.ISub(offset, low_cbuf); | ||||
| } | ||||
|  | ||||
| @@ -510,7 +514,7 @@ void Replace(IR::Block& block, IR::Inst& inst, const IR::U32& storage_index, | ||||
| } | ||||
| } // Anonymous namespace | ||||
|  | ||||
| void GlobalMemoryToStorageBufferPass(IR::Program& program) { | ||||
| void GlobalMemoryToStorageBufferPass(IR::Program& program, const HostTranslateInfo& host_info) { | ||||
|     StorageInfo info; | ||||
|     for (IR::Block* const block : program.post_order_blocks) { | ||||
|         for (IR::Inst& inst : block->Instructions()) { | ||||
| @@ -534,7 +538,8 @@ void GlobalMemoryToStorageBufferPass(IR::Program& program) { | ||||
|         const IR::U32 index{IR::Value{static_cast<u32>(info.set.index_of(it))}}; | ||||
|         IR::Block* const block{storage_inst.block}; | ||||
|         IR::Inst* const inst{storage_inst.inst}; | ||||
|         const IR::U32 offset{StorageOffset(*block, *inst, storage_buffer)}; | ||||
|         const IR::U32 offset{ | ||||
|             StorageOffset(*block, *inst, storage_buffer, host_info.min_ssbo_alignment)}; | ||||
|         Replace(*block, *inst, index, offset); | ||||
|     } | ||||
| } | ||||
|   | ||||
| @@ -15,7 +15,7 @@ namespace Shader::Optimization { | ||||
| void CollectShaderInfoPass(Environment& env, IR::Program& program); | ||||
| void ConstantPropagationPass(Environment& env, IR::Program& program); | ||||
| void DeadCodeEliminationPass(IR::Program& program); | ||||
| void GlobalMemoryToStorageBufferPass(IR::Program& program); | ||||
| void GlobalMemoryToStorageBufferPass(IR::Program& program, const HostTranslateInfo& host_info); | ||||
| void IdentityRemovalPass(IR::Program& program); | ||||
| void LowerFp16ToFp32(IR::Program& program); | ||||
| void LowerInt64ToInt32(IR::Program& program); | ||||
|   | ||||
| @@ -55,6 +55,8 @@ struct Profile { | ||||
|  | ||||
|     /// OpFClamp is broken and OpFMax + OpFMin should be used instead | ||||
|     bool has_broken_spirv_clamp{}; | ||||
|     /// The Position builtin needs to be wrapped in a struct when used as an input | ||||
|     bool has_broken_spirv_position_input{}; | ||||
|     /// Offset image operands with an unsigned type do not work | ||||
|     bool has_broken_unsigned_image_offsets{}; | ||||
|     /// Signed instructions with unsigned data types are misinterpreted | ||||
|   | ||||
| @@ -65,6 +65,8 @@ enum class Interpolation { | ||||
| struct ConstantBufferDescriptor { | ||||
|     u32 index; | ||||
|     u32 count; | ||||
|  | ||||
|     auto operator<=>(const ConstantBufferDescriptor&) const = default; | ||||
| }; | ||||
|  | ||||
| struct StorageBufferDescriptor { | ||||
| @@ -72,6 +74,8 @@ struct StorageBufferDescriptor { | ||||
|     u32 cbuf_offset; | ||||
|     u32 count; | ||||
|     bool is_written; | ||||
|  | ||||
|     auto operator<=>(const StorageBufferDescriptor&) const = default; | ||||
| }; | ||||
|  | ||||
| struct TextureBufferDescriptor { | ||||
| @@ -84,6 +88,8 @@ struct TextureBufferDescriptor { | ||||
|     u32 secondary_shift_left; | ||||
|     u32 count; | ||||
|     u32 size_shift; | ||||
|  | ||||
|     auto operator<=>(const TextureBufferDescriptor&) const = default; | ||||
| }; | ||||
| using TextureBufferDescriptors = boost::container::small_vector<TextureBufferDescriptor, 6>; | ||||
|  | ||||
| @@ -95,6 +101,8 @@ struct ImageBufferDescriptor { | ||||
|     u32 cbuf_offset; | ||||
|     u32 count; | ||||
|     u32 size_shift; | ||||
|  | ||||
|     auto operator<=>(const ImageBufferDescriptor&) const = default; | ||||
| }; | ||||
| using ImageBufferDescriptors = boost::container::small_vector<ImageBufferDescriptor, 2>; | ||||
|  | ||||
| @@ -110,6 +118,8 @@ struct TextureDescriptor { | ||||
|     u32 secondary_shift_left; | ||||
|     u32 count; | ||||
|     u32 size_shift; | ||||
|  | ||||
|     auto operator<=>(const TextureDescriptor&) const = default; | ||||
| }; | ||||
| using TextureDescriptors = boost::container::small_vector<TextureDescriptor, 12>; | ||||
|  | ||||
| @@ -122,6 +132,8 @@ struct ImageDescriptor { | ||||
|     u32 cbuf_offset; | ||||
|     u32 count; | ||||
|     u32 size_shift; | ||||
|  | ||||
|     auto operator<=>(const ImageDescriptor&) const = default; | ||||
| }; | ||||
| using ImageDescriptors = boost::container::small_vector<ImageDescriptor, 4>; | ||||
|  | ||||
|   | ||||
| @@ -1938,14 +1938,21 @@ typename BufferCache<P>::Binding BufferCache<P>::StorageBufferBinding(GPUVAddr s | ||||
|                                                                       bool is_written) const { | ||||
|     const GPUVAddr gpu_addr = gpu_memory->Read<u64>(ssbo_addr); | ||||
|     const u32 size = gpu_memory->Read<u32>(ssbo_addr + 8); | ||||
|     const std::optional<VAddr> cpu_addr = gpu_memory->GpuToCpuAddress(gpu_addr); | ||||
|     const u32 alignment = runtime.GetStorageBufferAlignment(); | ||||
|  | ||||
|     const GPUVAddr aligned_gpu_addr = Common::AlignDown(gpu_addr, alignment); | ||||
|     const u32 aligned_size = | ||||
|         Common::AlignUp(static_cast<u32>(gpu_addr - aligned_gpu_addr) + size, alignment); | ||||
|  | ||||
|     const std::optional<VAddr> cpu_addr = gpu_memory->GpuToCpuAddress(aligned_gpu_addr); | ||||
|     if (!cpu_addr || size == 0) { | ||||
|         return NULL_BINDING; | ||||
|     } | ||||
|     const VAddr cpu_end = Common::AlignUp(*cpu_addr + size, Core::Memory::YUZU_PAGESIZE); | ||||
|  | ||||
|     const VAddr cpu_end = Common::AlignUp(*cpu_addr + aligned_size, Core::Memory::YUZU_PAGESIZE); | ||||
|     const Binding binding{ | ||||
|         .cpu_addr = *cpu_addr, | ||||
|         .size = is_written ? size : static_cast<u32>(cpu_end - *cpu_addr), | ||||
|         .size = is_written ? aligned_size : static_cast<u32>(cpu_end - *cpu_addr), | ||||
|         .buffer_id = BufferId{}, | ||||
|     }; | ||||
|     return binding; | ||||
|   | ||||
| @@ -160,6 +160,10 @@ public: | ||||
|         return device.CanReportMemoryUsage(); | ||||
|     } | ||||
|  | ||||
|     u32 GetStorageBufferAlignment() const { | ||||
|         return static_cast<u32>(device.GetShaderStorageBufferAlignment()); | ||||
|     } | ||||
|  | ||||
| private: | ||||
|     static constexpr std::array PABO_LUT{ | ||||
|         GL_VERTEX_PROGRAM_PARAMETER_BUFFER_NV,          GL_TESS_CONTROL_PROGRAM_PARAMETER_BUFFER_NV, | ||||
|   | ||||
| @@ -236,6 +236,8 @@ ShaderCache::ShaderCache(RasterizerOpenGL& rasterizer_, Core::Frontend::EmuWindo | ||||
|           .needs_demote_reorder = device.IsAmd(), | ||||
|           .support_snorm_render_buffer = false, | ||||
|           .support_viewport_index_layer = device.HasVertexViewportLayer(), | ||||
|           .min_ssbo_alignment = static_cast<u32>(device.GetShaderStorageBufferAlignment()), | ||||
|           .support_geometry_shader_passthrough = device.HasGeometryShaderPassthrough(), | ||||
|       } { | ||||
|     if (use_asynchronous_shaders) { | ||||
|         workers = CreateWorkers(); | ||||
|   | ||||
| @@ -330,6 +330,10 @@ bool BufferCacheRuntime::CanReportMemoryUsage() const { | ||||
|     return device.CanReportMemoryUsage(); | ||||
| } | ||||
|  | ||||
| u32 BufferCacheRuntime::GetStorageBufferAlignment() const { | ||||
|     return static_cast<u32>(device.GetStorageBufferAlignment()); | ||||
| } | ||||
|  | ||||
| void BufferCacheRuntime::Finish() { | ||||
|     scheduler.Finish(); | ||||
| } | ||||
|   | ||||
| @@ -73,6 +73,8 @@ public: | ||||
|  | ||||
|     bool CanReportMemoryUsage() const; | ||||
|  | ||||
|     u32 GetStorageBufferAlignment() const; | ||||
|  | ||||
|     [[nodiscard]] StagingBufferRef UploadStagingBuffer(size_t size); | ||||
|  | ||||
|     [[nodiscard]] StagingBufferRef DownloadStagingBuffer(size_t size); | ||||
|   | ||||
| @@ -331,6 +331,7 @@ PipelineCache::PipelineCache(RasterizerVulkan& rasterizer_, const Device& device | ||||
|         .need_declared_frag_colors = false, | ||||
|  | ||||
|         .has_broken_spirv_clamp = driver_id == VK_DRIVER_ID_INTEL_PROPRIETARY_WINDOWS, | ||||
|         .has_broken_spirv_position_input = driver_id == VK_DRIVER_ID_QUALCOMM_PROPRIETARY, | ||||
|         .has_broken_unsigned_image_offsets = false, | ||||
|         .has_broken_signed_operations = false, | ||||
|         .has_broken_fp16_float_controls = driver_id == VK_DRIVER_ID_NVIDIA_PROPRIETARY, | ||||
| @@ -343,6 +344,8 @@ PipelineCache::PipelineCache(RasterizerVulkan& rasterizer_, const Device& device | ||||
|             driver_id == VK_DRIVER_ID_AMD_PROPRIETARY || driver_id == VK_DRIVER_ID_AMD_OPEN_SOURCE, | ||||
|         .support_snorm_render_buffer = true, | ||||
|         .support_viewport_index_layer = device.IsExtShaderViewportIndexLayerSupported(), | ||||
|         .min_ssbo_alignment = static_cast<u32>(device.GetStorageBufferAlignment()), | ||||
|         .support_geometry_shader_passthrough = device.IsNvGeometryShaderPassthroughSupported(), | ||||
|     }; | ||||
|  | ||||
|     if (device.GetMaxVertexInputAttributes() < Maxwell::NumVertexAttributes) { | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Fernando S
					Fernando S