summaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
Diffstat (limited to 'src')
-rw-r--r--src/core/hle/kernel/process.cpp4
-rw-r--r--src/video_core/renderer_vulkan/maxwell_to_vk.cpp9
-rw-r--r--src/video_core/renderer_vulkan/maxwell_to_vk.h2
-rw-r--r--src/video_core/renderer_vulkan/vk_sampler_cache.cpp6
-rw-r--r--src/video_core/shader/decode/memory.cpp84
-rw-r--r--src/video_core/shader/decode/texture.cpp16
-rw-r--r--src/video_core/shader/shader_ir.h2
7 files changed, 71 insertions, 52 deletions
diff --git a/src/core/hle/kernel/process.cpp b/src/core/hle/kernel/process.cpp
index 12ea4ebe3..b9035a0be 100644
--- a/src/core/hle/kernel/process.cpp
+++ b/src/core/hle/kernel/process.cpp
@@ -317,6 +317,8 @@ void Process::FreeTLSRegion(VAddr tls_address) {
317} 317}
318 318
319void Process::LoadModule(CodeSet module_, VAddr base_addr) { 319void Process::LoadModule(CodeSet module_, VAddr base_addr) {
320 code_memory_size += module_.memory.size();
321
320 const auto memory = std::make_shared<PhysicalMemory>(std::move(module_.memory)); 322 const auto memory = std::make_shared<PhysicalMemory>(std::move(module_.memory));
321 323
322 const auto MapSegment = [&](const CodeSet::Segment& segment, VMAPermission permissions, 324 const auto MapSegment = [&](const CodeSet::Segment& segment, VMAPermission permissions,
@@ -332,8 +334,6 @@ void Process::LoadModule(CodeSet module_, VAddr base_addr) {
332 MapSegment(module_.CodeSegment(), VMAPermission::ReadExecute, MemoryState::Code); 334 MapSegment(module_.CodeSegment(), VMAPermission::ReadExecute, MemoryState::Code);
333 MapSegment(module_.RODataSegment(), VMAPermission::Read, MemoryState::CodeData); 335 MapSegment(module_.RODataSegment(), VMAPermission::Read, MemoryState::CodeData);
334 MapSegment(module_.DataSegment(), VMAPermission::ReadWrite, MemoryState::CodeData); 336 MapSegment(module_.DataSegment(), VMAPermission::ReadWrite, MemoryState::CodeData);
335
336 code_memory_size += module_.memory.size();
337} 337}
338 338
339Process::Process(Core::System& system) 339Process::Process(Core::System& system)
diff --git a/src/video_core/renderer_vulkan/maxwell_to_vk.cpp b/src/video_core/renderer_vulkan/maxwell_to_vk.cpp
index 000e3616d..331808113 100644
--- a/src/video_core/renderer_vulkan/maxwell_to_vk.cpp
+++ b/src/video_core/renderer_vulkan/maxwell_to_vk.cpp
@@ -44,7 +44,7 @@ vk::SamplerMipmapMode MipmapMode(Tegra::Texture::TextureMipmapFilter mipmap_filt
44 return {}; 44 return {};
45} 45}
46 46
47vk::SamplerAddressMode WrapMode(Tegra::Texture::WrapMode wrap_mode, 47vk::SamplerAddressMode WrapMode(const VKDevice& device, Tegra::Texture::WrapMode wrap_mode,
48 Tegra::Texture::TextureFilter filter) { 48 Tegra::Texture::TextureFilter filter) {
49 switch (wrap_mode) { 49 switch (wrap_mode) {
50 case Tegra::Texture::WrapMode::Wrap: 50 case Tegra::Texture::WrapMode::Wrap:
@@ -56,7 +56,12 @@ vk::SamplerAddressMode WrapMode(Tegra::Texture::WrapMode wrap_mode,
56 case Tegra::Texture::WrapMode::Border: 56 case Tegra::Texture::WrapMode::Border:
57 return vk::SamplerAddressMode::eClampToBorder; 57 return vk::SamplerAddressMode::eClampToBorder;
58 case Tegra::Texture::WrapMode::Clamp: 58 case Tegra::Texture::WrapMode::Clamp:
59 // TODO(Rodrigo): Emulate GL_CLAMP properly 59 if (device.GetDriverID() == vk::DriverIdKHR::eNvidiaProprietary) {
60 // Nvidia's Vulkan driver defaults to GL_CLAMP on invalid enumerations, we can hack this
61 // by sending an invalid enumeration.
62 return static_cast<vk::SamplerAddressMode>(0xcafe);
63 }
64 // TODO(Rodrigo): Emulate GL_CLAMP properly on other vendors
60 switch (filter) { 65 switch (filter) {
61 case Tegra::Texture::TextureFilter::Nearest: 66 case Tegra::Texture::TextureFilter::Nearest:
62 return vk::SamplerAddressMode::eClampToEdge; 67 return vk::SamplerAddressMode::eClampToEdge;
diff --git a/src/video_core/renderer_vulkan/maxwell_to_vk.h b/src/video_core/renderer_vulkan/maxwell_to_vk.h
index 1534b738b..7e9678b7b 100644
--- a/src/video_core/renderer_vulkan/maxwell_to_vk.h
+++ b/src/video_core/renderer_vulkan/maxwell_to_vk.h
@@ -22,7 +22,7 @@ vk::Filter Filter(Tegra::Texture::TextureFilter filter);
22 22
23vk::SamplerMipmapMode MipmapMode(Tegra::Texture::TextureMipmapFilter mipmap_filter); 23vk::SamplerMipmapMode MipmapMode(Tegra::Texture::TextureMipmapFilter mipmap_filter);
24 24
25vk::SamplerAddressMode WrapMode(Tegra::Texture::WrapMode wrap_mode, 25vk::SamplerAddressMode WrapMode(const VKDevice& device, Tegra::Texture::WrapMode wrap_mode,
26 Tegra::Texture::TextureFilter filter); 26 Tegra::Texture::TextureFilter filter);
27 27
28vk::CompareOp DepthCompareFunction(Tegra::Texture::DepthCompareFunc depth_compare_func); 28vk::CompareOp DepthCompareFunction(Tegra::Texture::DepthCompareFunc depth_compare_func);
diff --git a/src/video_core/renderer_vulkan/vk_sampler_cache.cpp b/src/video_core/renderer_vulkan/vk_sampler_cache.cpp
index 1ce583f75..0a8ec8398 100644
--- a/src/video_core/renderer_vulkan/vk_sampler_cache.cpp
+++ b/src/video_core/renderer_vulkan/vk_sampler_cache.cpp
@@ -46,9 +46,9 @@ UniqueSampler VKSamplerCache::CreateSampler(const Tegra::Texture::TSCEntry& tsc)
46 {}, MaxwellToVK::Sampler::Filter(tsc.mag_filter), 46 {}, MaxwellToVK::Sampler::Filter(tsc.mag_filter),
47 MaxwellToVK::Sampler::Filter(tsc.min_filter), 47 MaxwellToVK::Sampler::Filter(tsc.min_filter),
48 MaxwellToVK::Sampler::MipmapMode(tsc.mipmap_filter), 48 MaxwellToVK::Sampler::MipmapMode(tsc.mipmap_filter),
49 MaxwellToVK::Sampler::WrapMode(tsc.wrap_u, tsc.mag_filter), 49 MaxwellToVK::Sampler::WrapMode(device, tsc.wrap_u, tsc.mag_filter),
50 MaxwellToVK::Sampler::WrapMode(tsc.wrap_v, tsc.mag_filter), 50 MaxwellToVK::Sampler::WrapMode(device, tsc.wrap_v, tsc.mag_filter),
51 MaxwellToVK::Sampler::WrapMode(tsc.wrap_p, tsc.mag_filter), tsc.GetLodBias(), 51 MaxwellToVK::Sampler::WrapMode(device, tsc.wrap_p, tsc.mag_filter), tsc.GetLodBias(),
52 has_anisotropy, max_anisotropy, tsc.depth_compare_enabled, 52 has_anisotropy, max_anisotropy, tsc.depth_compare_enabled,
53 MaxwellToVK::Sampler::DepthCompareFunction(tsc.depth_compare_func), tsc.GetMinLod(), 53 MaxwellToVK::Sampler::DepthCompareFunction(tsc.depth_compare_func), tsc.GetMinLod(),
54 tsc.GetMaxLod(), vk_border_color.value_or(vk::BorderColor::eFloatTransparentBlack), 54 tsc.GetMaxLod(), vk_border_color.value_or(vk::BorderColor::eFloatTransparentBlack),
diff --git a/src/video_core/shader/decode/memory.cpp b/src/video_core/shader/decode/memory.cpp
index c934d0719..8cc84e935 100644
--- a/src/video_core/shader/decode/memory.cpp
+++ b/src/video_core/shader/decode/memory.cpp
@@ -6,6 +6,7 @@
6#include <vector> 6#include <vector>
7#include <fmt/format.h> 7#include <fmt/format.h>
8 8
9#include "common/alignment.h"
9#include "common/assert.h" 10#include "common/assert.h"
10#include "common/common_types.h" 11#include "common/common_types.h"
11#include "common/logging/log.h" 12#include "common/logging/log.h"
@@ -22,34 +23,39 @@ using Tegra::Shader::Register;
22 23
23namespace { 24namespace {
24 25
25u32 GetLdgMemorySize(Tegra::Shader::UniformType uniform_type) { 26bool IsUnaligned(Tegra::Shader::UniformType uniform_type) {
27 return uniform_type == Tegra::Shader::UniformType::UnsignedByte ||
28 uniform_type == Tegra::Shader::UniformType::UnsignedShort;
29}
30
31u32 GetUnalignedMask(Tegra::Shader::UniformType uniform_type) {
26 switch (uniform_type) { 32 switch (uniform_type) {
27 case Tegra::Shader::UniformType::UnsignedByte: 33 case Tegra::Shader::UniformType::UnsignedByte:
28 case Tegra::Shader::UniformType::Single: 34 return 0b11;
29 return 1; 35 case Tegra::Shader::UniformType::UnsignedShort:
30 case Tegra::Shader::UniformType::Double: 36 return 0b10;
31 return 2;
32 case Tegra::Shader::UniformType::Quad:
33 case Tegra::Shader::UniformType::UnsignedQuad:
34 return 4;
35 default: 37 default:
36 UNIMPLEMENTED_MSG("Unimplemented size={}!", static_cast<u32>(uniform_type)); 38 UNREACHABLE();
37 return 1; 39 return 0;
38 } 40 }
39} 41}
40 42
41u32 GetStgMemorySize(Tegra::Shader::UniformType uniform_type) { 43u32 GetMemorySize(Tegra::Shader::UniformType uniform_type) {
42 switch (uniform_type) { 44 switch (uniform_type) {
45 case Tegra::Shader::UniformType::UnsignedByte:
46 return 8;
47 case Tegra::Shader::UniformType::UnsignedShort:
48 return 16;
43 case Tegra::Shader::UniformType::Single: 49 case Tegra::Shader::UniformType::Single:
44 return 1; 50 return 32;
45 case Tegra::Shader::UniformType::Double: 51 case Tegra::Shader::UniformType::Double:
46 return 2; 52 return 64;
47 case Tegra::Shader::UniformType::Quad: 53 case Tegra::Shader::UniformType::Quad:
48 case Tegra::Shader::UniformType::UnsignedQuad: 54 case Tegra::Shader::UniformType::UnsignedQuad:
49 return 4; 55 return 128;
50 default: 56 default:
51 UNIMPLEMENTED_MSG("Unimplemented size={}!", static_cast<u32>(uniform_type)); 57 UNIMPLEMENTED_MSG("Unimplemented size={}!", static_cast<u32>(uniform_type));
52 return 1; 58 return 32;
53 } 59 }
54} 60}
55 61
@@ -184,9 +190,10 @@ u32 ShaderIR::DecodeMemory(NodeBlock& bb, u32 pc) {
184 }(); 190 }();
185 191
186 const auto [real_address_base, base_address, descriptor] = 192 const auto [real_address_base, base_address, descriptor] =
187 TrackGlobalMemory(bb, instr, false); 193 TrackGlobalMemory(bb, instr, true, false);
188 194
189 const u32 count = GetLdgMemorySize(type); 195 const u32 size = GetMemorySize(type);
196 const u32 count = Common::AlignUp(size, 32) / 32;
190 if (!real_address_base || !base_address) { 197 if (!real_address_base || !base_address) {
191 // Tracking failed, load zeroes. 198 // Tracking failed, load zeroes.
192 for (u32 i = 0; i < count; ++i) { 199 for (u32 i = 0; i < count; ++i) {
@@ -200,14 +207,15 @@ u32 ShaderIR::DecodeMemory(NodeBlock& bb, u32 pc) {
200 const Node real_address = Operation(OperationCode::UAdd, real_address_base, it_offset); 207 const Node real_address = Operation(OperationCode::UAdd, real_address_base, it_offset);
201 Node gmem = MakeNode<GmemNode>(real_address, base_address, descriptor); 208 Node gmem = MakeNode<GmemNode>(real_address, base_address, descriptor);
202 209
203 if (type == Tegra::Shader::UniformType::UnsignedByte) { 210 // To handle unaligned loads get the bytes used to dereference global memory and extract
204 // To handle unaligned loads get the byte used to dereferenced global memory 211 // those bytes from the loaded u32.
205 // and extract that byte from the loaded uint32. 212 if (IsUnaligned(type)) {
206 Node byte = Operation(OperationCode::UBitwiseAnd, real_address, Immediate(3)); 213 Node mask = Immediate(GetUnalignedMask(type));
207 byte = Operation(OperationCode::ULogicalShiftLeft, std::move(byte), Immediate(3)); 214 Node offset = Operation(OperationCode::UBitwiseAnd, real_address, std::move(mask));
215 offset = Operation(OperationCode::ULogicalShiftLeft, offset, Immediate(3));
208 216
209 gmem = Operation(OperationCode::UBitfieldExtract, std::move(gmem), std::move(byte), 217 gmem = Operation(OperationCode::UBitfieldExtract, std::move(gmem),
210 Immediate(8)); 218 std::move(offset), Immediate(size));
211 } 219 }
212 220
213 SetTemporary(bb, i, gmem); 221 SetTemporary(bb, i, gmem);
@@ -295,19 +303,32 @@ u32 ShaderIR::DecodeMemory(NodeBlock& bb, u32 pc) {
295 } 303 }
296 }(); 304 }();
297 305
306 // For unaligned reads we have to read memory too.
307 const bool is_read = IsUnaligned(type);
298 const auto [real_address_base, base_address, descriptor] = 308 const auto [real_address_base, base_address, descriptor] =
299 TrackGlobalMemory(bb, instr, true); 309 TrackGlobalMemory(bb, instr, is_read, true);
300 if (!real_address_base || !base_address) { 310 if (!real_address_base || !base_address) {
301 // Tracking failed, skip the store. 311 // Tracking failed, skip the store.
302 break; 312 break;
303 } 313 }
304 314
305 const u32 count = GetStgMemorySize(type); 315 const u32 size = GetMemorySize(type);
316 const u32 count = Common::AlignUp(size, 32) / 32;
306 for (u32 i = 0; i < count; ++i) { 317 for (u32 i = 0; i < count; ++i) {
307 const Node it_offset = Immediate(i * 4); 318 const Node it_offset = Immediate(i * 4);
308 const Node real_address = Operation(OperationCode::UAdd, real_address_base, it_offset); 319 const Node real_address = Operation(OperationCode::UAdd, real_address_base, it_offset);
309 const Node gmem = MakeNode<GmemNode>(real_address, base_address, descriptor); 320 const Node gmem = MakeNode<GmemNode>(real_address, base_address, descriptor);
310 const Node value = GetRegister(instr.gpr0.Value() + i); 321 Node value = GetRegister(instr.gpr0.Value() + i);
322
323 if (IsUnaligned(type)) {
324 Node mask = Immediate(GetUnalignedMask(type));
325 Node offset = Operation(OperationCode::UBitwiseAnd, real_address, std::move(mask));
326 offset = Operation(OperationCode::ULogicalShiftLeft, offset, Immediate(3));
327
328 value = Operation(OperationCode::UBitfieldInsert, gmem, std::move(value), offset,
329 Immediate(size));
330 }
331
311 bb.push_back(Operation(OperationCode::Assign, gmem, value)); 332 bb.push_back(Operation(OperationCode::Assign, gmem, value));
312 } 333 }
313 break; 334 break;
@@ -336,7 +357,7 @@ u32 ShaderIR::DecodeMemory(NodeBlock& bb, u32 pc) {
336 357
337std::tuple<Node, Node, GlobalMemoryBase> ShaderIR::TrackGlobalMemory(NodeBlock& bb, 358std::tuple<Node, Node, GlobalMemoryBase> ShaderIR::TrackGlobalMemory(NodeBlock& bb,
338 Instruction instr, 359 Instruction instr,
339 bool is_write) { 360 bool is_read, bool is_write) {
340 const auto addr_register{GetRegister(instr.gmem.gpr)}; 361 const auto addr_register{GetRegister(instr.gmem.gpr)};
341 const auto immediate_offset{static_cast<u32>(instr.gmem.offset)}; 362 const auto immediate_offset{static_cast<u32>(instr.gmem.offset)};
342 363
@@ -351,11 +372,8 @@ std::tuple<Node, Node, GlobalMemoryBase> ShaderIR::TrackGlobalMemory(NodeBlock&
351 const GlobalMemoryBase descriptor{index, offset}; 372 const GlobalMemoryBase descriptor{index, offset};
352 const auto& [entry, is_new] = used_global_memory.try_emplace(descriptor); 373 const auto& [entry, is_new] = used_global_memory.try_emplace(descriptor);
353 auto& usage = entry->second; 374 auto& usage = entry->second;
354 if (is_write) { 375 usage.is_written |= is_write;
355 usage.is_written = true; 376 usage.is_read |= is_read;
356 } else {
357 usage.is_read = true;
358 }
359 377
360 const auto real_address = 378 const auto real_address =
361 Operation(OperationCode::UAdd, NO_PRECISE, Immediate(immediate_offset), addr_register); 379 Operation(OperationCode::UAdd, NO_PRECISE, Immediate(immediate_offset), addr_register);
diff --git a/src/video_core/shader/decode/texture.cpp b/src/video_core/shader/decode/texture.cpp
index 4b14cdf58..cd984f763 100644
--- a/src/video_core/shader/decode/texture.cpp
+++ b/src/video_core/shader/decode/texture.cpp
@@ -794,14 +794,10 @@ std::tuple<std::size_t, std::size_t> ShaderIR::ValidateAndGetCoordinateElement(
794 794
795std::vector<Node> ShaderIR::GetAoffiCoordinates(Node aoffi_reg, std::size_t coord_count, 795std::vector<Node> ShaderIR::GetAoffiCoordinates(Node aoffi_reg, std::size_t coord_count,
796 bool is_tld4) { 796 bool is_tld4) {
797 const auto [coord_offsets, size, wrap_value, 797 const std::array coord_offsets = is_tld4 ? std::array{0U, 8U, 16U} : std::array{0U, 4U, 8U};
798 diff_value] = [is_tld4]() -> std::tuple<std::array<u32, 3>, u32, s32, s32> { 798 const u32 size = is_tld4 ? 6 : 4;
799 if (is_tld4) { 799 const s32 wrap_value = is_tld4 ? 32 : 8;
800 return {{0, 8, 16}, 6, 32, 64}; 800 const s32 diff_value = is_tld4 ? 64 : 16;
801 } else {
802 return {{0, 4, 8}, 4, 8, 16};
803 }
804 }();
805 const u32 mask = (1U << size) - 1; 801 const u32 mask = (1U << size) - 1;
806 802
807 std::vector<Node> aoffi; 803 std::vector<Node> aoffi;
@@ -814,7 +810,7 @@ std::vector<Node> ShaderIR::GetAoffiCoordinates(Node aoffi_reg, std::size_t coor
814 LOG_WARNING(HW_GPU, 810 LOG_WARNING(HW_GPU,
815 "AOFFI constant folding failed, some hardware might have graphical issues"); 811 "AOFFI constant folding failed, some hardware might have graphical issues");
816 for (std::size_t coord = 0; coord < coord_count; ++coord) { 812 for (std::size_t coord = 0; coord < coord_count; ++coord) {
817 const Node value = BitfieldExtract(aoffi_reg, coord_offsets.at(coord), size); 813 const Node value = BitfieldExtract(aoffi_reg, coord_offsets[coord], size);
818 const Node condition = 814 const Node condition =
819 Operation(OperationCode::LogicalIGreaterEqual, value, Immediate(wrap_value)); 815 Operation(OperationCode::LogicalIGreaterEqual, value, Immediate(wrap_value));
820 const Node negative = Operation(OperationCode::IAdd, value, Immediate(-diff_value)); 816 const Node negative = Operation(OperationCode::IAdd, value, Immediate(-diff_value));
@@ -824,7 +820,7 @@ std::vector<Node> ShaderIR::GetAoffiCoordinates(Node aoffi_reg, std::size_t coor
824 } 820 }
825 821
826 for (std::size_t coord = 0; coord < coord_count; ++coord) { 822 for (std::size_t coord = 0; coord < coord_count; ++coord) {
827 s32 value = (*aoffi_immediate >> coord_offsets.at(coord)) & mask; 823 s32 value = (*aoffi_immediate >> coord_offsets[coord]) & mask;
828 if (value >= wrap_value) { 824 if (value >= wrap_value) {
829 value -= diff_value; 825 value -= diff_value;
830 } 826 }
diff --git a/src/video_core/shader/shader_ir.h b/src/video_core/shader/shader_ir.h
index aacd0a0da..ba1db4c11 100644
--- a/src/video_core/shader/shader_ir.h
+++ b/src/video_core/shader/shader_ir.h
@@ -394,7 +394,7 @@ private:
394 394
395 std::tuple<Node, Node, GlobalMemoryBase> TrackGlobalMemory(NodeBlock& bb, 395 std::tuple<Node, Node, GlobalMemoryBase> TrackGlobalMemory(NodeBlock& bb,
396 Tegra::Shader::Instruction instr, 396 Tegra::Shader::Instruction instr,
397 bool is_write); 397 bool is_read, bool is_write);
398 398
399 /// Register new amending code and obtain the reference id. 399 /// Register new amending code and obtain the reference id.
400 std::size_t DeclareAmend(Node new_amend); 400 std::size_t DeclareAmend(Node new_amend);