summaryrefslogtreecommitdiff
path: root/src/video_core/renderer_vulkan
diff options
context:
space:
mode:
authorGravatar ameerj2023-03-06 21:16:17 -0500
committerGravatar ameerj2023-03-06 22:57:45 -0500
commit6b9cc0ed23b15a1b96b322b03feff2153e44a4a9 (patch)
tree6949cedbe90b6d34e2092c8eba57b243defb21dc /src/video_core/renderer_vulkan
parentMerge pull request #9890 from Kelebek1/reverb_fix (diff)
downloadyuzu-6b9cc0ed23b15a1b96b322b03feff2153e44a4a9.tar.gz
yuzu-6b9cc0ed23b15a1b96b322b03feff2153e44a4a9.tar.xz
yuzu-6b9cc0ed23b15a1b96b322b03feff2153e44a4a9.zip
Refactor AccelerateDMA code
Diffstat (limited to 'src/video_core/renderer_vulkan')
-rw-r--r--src/video_core/renderer_vulkan/vk_rasterizer.cpp248
-rw-r--r--src/video_core/renderer_vulkan/vk_rasterizer.h5
-rw-r--r--src/video_core/renderer_vulkan/vk_texture_cache.cpp22
-rw-r--r--src/video_core/renderer_vulkan/vk_texture_cache.h6
4 files changed, 57 insertions, 224 deletions
diff --git a/src/video_core/renderer_vulkan/vk_rasterizer.cpp b/src/video_core/renderer_vulkan/vk_rasterizer.cpp
index f085d53a1..a00cf1569 100644
--- a/src/video_core/renderer_vulkan/vk_rasterizer.cpp
+++ b/src/video_core/renderer_vulkan/vk_rasterizer.cpp
@@ -770,232 +770,44 @@ bool AccelerateDMA::BufferCopy(GPUVAddr src_address, GPUVAddr dest_address, u64
770 return buffer_cache.DMACopy(src_address, dest_address, amount); 770 return buffer_cache.DMACopy(src_address, dest_address, amount);
771} 771}
772 772
773bool AccelerateDMA::ImageToBuffer(const Tegra::DMA::ImageCopy& copy_info, 773template <bool IS_IMAGE_UPLOAD>
774 const Tegra::DMA::ImageOperand& src, 774bool AccelerateDMA::DmaBufferImageCopy(const Tegra::DMA::ImageCopy& copy_info,
775 const Tegra::DMA::BufferOperand& dst) { 775 const Tegra::DMA::BufferOperand& buffer_operand,
776 const Tegra::DMA::ImageOperand& image_operand) {
776 std::scoped_lock lock{buffer_cache.mutex, texture_cache.mutex}; 777 std::scoped_lock lock{buffer_cache.mutex, texture_cache.mutex};
777 auto query_image = texture_cache.ObtainImage(src, false); 778 const auto image_id = texture_cache.DmaImageId(image_operand);
778 if (!query_image) { 779 if (image_id == VideoCommon::NULL_IMAGE_ID) {
779 return false; 780 return false;
780 } 781 }
781 auto* image = query_image->first; 782 const u32 buffer_size = static_cast<u32>(buffer_operand.pitch * buffer_operand.height);
782 auto [level, base] = query_image->second; 783 static constexpr auto sync_info = VideoCommon::ObtainBufferSynchronize::FullSynchronize;
783 const u32 buffer_size = static_cast<u32>(dst.pitch * dst.height); 784 const auto post_op = IS_IMAGE_UPLOAD ? VideoCommon::ObtainBufferOperation::DoNothing
784 const auto [buffer, offset] = buffer_cache.ObtainBuffer( 785 : VideoCommon::ObtainBufferOperation::MarkAsWritten;
785 dst.address, buffer_size, VideoCommon::ObtainBufferSynchronize::FullSynchronize, 786 const auto [buffer, offset] =
786 VideoCommon::ObtainBufferOperation::MarkAsWritten); 787 buffer_cache.ObtainBuffer(buffer_operand.address, buffer_size, sync_info, post_op);
787 788
788 const bool is_rescaled = image->IsRescaled(); 789 const auto [image, copy] = texture_cache.DmaBufferImageCopy(
789 if (is_rescaled) { 790 copy_info, buffer_operand, image_operand, image_id, IS_IMAGE_UPLOAD);
790 image->ScaleDown(); 791 const std::span copy_span{&copy, 1};
791 } 792
792 VkImageSubresourceLayers subresources{ 793 if constexpr (IS_IMAGE_UPLOAD) {
793 .aspectMask = image->AspectMask(), 794 image->UploadMemory(buffer->Handle(), offset, copy_span);
794 .mipLevel = level, 795 } else {
795 .baseArrayLayer = base, 796 image->DownloadMemory(buffer->Handle(), offset, copy_span);
796 .layerCount = 1,
797 };
798 const u32 bpp = VideoCore::Surface::BytesPerBlock(image->info.format);
799 const auto convert = [old_bpp = src.bytes_per_pixel, bpp](u32 value) {
800 return (old_bpp * value) / bpp;
801 };
802 const u32 base_x = convert(src.params.origin.x.Value());
803 const u32 base_y = src.params.origin.y.Value();
804 const u32 length_x = convert(copy_info.length_x);
805 const u32 length_y = copy_info.length_y;
806 VkOffset3D image_offset{
807 .x = static_cast<s32>(base_x),
808 .y = static_cast<s32>(base_y),
809 .z = 0,
810 };
811 VkExtent3D image_extent{
812 .width = length_x,
813 .height = length_y,
814 .depth = 1,
815 };
816 auto buff_info(dst);
817 buff_info.pitch = convert(dst.pitch);
818 scheduler.RequestOutsideRenderPassOperationContext();
819 scheduler.Record([src_image = image->Handle(), dst_buffer = buffer->Handle(),
820 buffer_offset = offset, subresources, image_offset, image_extent,
821 buff_info](vk::CommandBuffer cmdbuf) {
822 const std::array buffer_copy_info{
823 VkBufferImageCopy{
824 .bufferOffset = buffer_offset,
825 .bufferRowLength = buff_info.pitch,
826 .bufferImageHeight = buff_info.height,
827 .imageSubresource = subresources,
828 .imageOffset = image_offset,
829 .imageExtent = image_extent,
830 },
831 };
832 const VkImageSubresourceRange range{
833 .aspectMask = subresources.aspectMask,
834 .baseMipLevel = subresources.mipLevel,
835 .levelCount = 1,
836 .baseArrayLayer = subresources.baseArrayLayer,
837 .layerCount = 1,
838 };
839 static constexpr VkMemoryBarrier WRITE_BARRIER{
840 .sType = VK_STRUCTURE_TYPE_MEMORY_BARRIER,
841 .pNext = nullptr,
842 .srcAccessMask = VK_ACCESS_TRANSFER_WRITE_BIT,
843 .dstAccessMask = VK_ACCESS_MEMORY_READ_BIT | VK_ACCESS_MEMORY_WRITE_BIT,
844 };
845 const std::array pre_barriers{
846 VkImageMemoryBarrier{
847 .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER,
848 .pNext = nullptr,
849 .srcAccessMask = VK_ACCESS_SHADER_WRITE_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT |
850 VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT |
851 VK_ACCESS_TRANSFER_WRITE_BIT,
852 .dstAccessMask = VK_ACCESS_TRANSFER_READ_BIT,
853 .oldLayout = VK_IMAGE_LAYOUT_GENERAL,
854 .newLayout = VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
855 .srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED,
856 .dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED,
857 .image = src_image,
858 .subresourceRange = range,
859 },
860 };
861 const std::array post_barriers{
862 VkImageMemoryBarrier{
863 .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER,
864 .pNext = nullptr,
865 .srcAccessMask = 0,
866 .dstAccessMask = 0,
867 .oldLayout = VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
868 .newLayout = VK_IMAGE_LAYOUT_GENERAL,
869 .srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED,
870 .dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED,
871 .image = src_image,
872 .subresourceRange = range,
873 },
874 };
875 cmdbuf.PipelineBarrier(VK_PIPELINE_STAGE_ALL_COMMANDS_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT,
876 0, {}, {}, pre_barriers);
877 cmdbuf.CopyImageToBuffer(src_image, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL, dst_buffer,
878 buffer_copy_info);
879 cmdbuf.PipelineBarrier(VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_ALL_COMMANDS_BIT,
880 0, WRITE_BARRIER, nullptr, post_barriers);
881 });
882 if (is_rescaled) {
883 image->ScaleUp(true);
884 } 797 }
885 return true; 798 return true;
886} 799}
887 800
801bool AccelerateDMA::ImageToBuffer(const Tegra::DMA::ImageCopy& copy_info,
802 const Tegra::DMA::ImageOperand& image_operand,
803 const Tegra::DMA::BufferOperand& buffer_operand) {
804 return DmaBufferImageCopy<false>(copy_info, buffer_operand, image_operand);
805}
806
888bool AccelerateDMA::BufferToImage(const Tegra::DMA::ImageCopy& copy_info, 807bool AccelerateDMA::BufferToImage(const Tegra::DMA::ImageCopy& copy_info,
889 const Tegra::DMA::BufferOperand& src, 808 const Tegra::DMA::BufferOperand& buffer_operand,
890 const Tegra::DMA::ImageOperand& dst) { 809 const Tegra::DMA::ImageOperand& image_operand) {
891 std::scoped_lock lock{buffer_cache.mutex, texture_cache.mutex}; 810 return DmaBufferImageCopy<true>(copy_info, buffer_operand, image_operand);
892 auto query_image = texture_cache.ObtainImage(dst, true);
893 if (!query_image) {
894 return false;
895 }
896 auto* image = query_image->first;
897 auto [level, base] = query_image->second;
898 const u32 buffer_size = static_cast<u32>(src.pitch * src.height);
899 const auto [buffer, offset] = buffer_cache.ObtainBuffer(
900 src.address, buffer_size, VideoCommon::ObtainBufferSynchronize::FullSynchronize,
901 VideoCommon::ObtainBufferOperation::DoNothing);
902 const bool is_rescaled = image->IsRescaled();
903 if (is_rescaled) {
904 image->ScaleDown(true);
905 }
906 VkImageSubresourceLayers subresources{
907 .aspectMask = image->AspectMask(),
908 .mipLevel = level,
909 .baseArrayLayer = base,
910 .layerCount = 1,
911 };
912 const u32 bpp = VideoCore::Surface::BytesPerBlock(image->info.format);
913 const auto convert = [old_bpp = dst.bytes_per_pixel, bpp](u32 value) {
914 return (old_bpp * value) / bpp;
915 };
916 const u32 base_x = convert(dst.params.origin.x.Value());
917 const u32 base_y = dst.params.origin.y.Value();
918 const u32 length_x = convert(copy_info.length_x);
919 const u32 length_y = copy_info.length_y;
920 VkOffset3D image_offset{
921 .x = static_cast<s32>(base_x),
922 .y = static_cast<s32>(base_y),
923 .z = 0,
924 };
925 VkExtent3D image_extent{
926 .width = length_x,
927 .height = length_y,
928 .depth = 1,
929 };
930 auto buff_info(src);
931 buff_info.pitch = convert(src.pitch);
932 scheduler.RequestOutsideRenderPassOperationContext();
933 scheduler.Record([dst_image = image->Handle(), src_buffer = buffer->Handle(),
934 buffer_offset = offset, subresources, image_offset, image_extent,
935 buff_info](vk::CommandBuffer cmdbuf) {
936 const std::array buffer_copy_info{
937 VkBufferImageCopy{
938 .bufferOffset = buffer_offset,
939 .bufferRowLength = buff_info.pitch,
940 .bufferImageHeight = buff_info.height,
941 .imageSubresource = subresources,
942 .imageOffset = image_offset,
943 .imageExtent = image_extent,
944 },
945 };
946 const VkImageSubresourceRange range{
947 .aspectMask = subresources.aspectMask,
948 .baseMipLevel = subresources.mipLevel,
949 .levelCount = 1,
950 .baseArrayLayer = subresources.baseArrayLayer,
951 .layerCount = 1,
952 };
953 static constexpr VkMemoryBarrier READ_BARRIER{
954 .sType = VK_STRUCTURE_TYPE_MEMORY_BARRIER,
955 .pNext = nullptr,
956 .srcAccessMask = VK_ACCESS_MEMORY_WRITE_BIT,
957 .dstAccessMask = VK_ACCESS_TRANSFER_READ_BIT | VK_ACCESS_TRANSFER_WRITE_BIT,
958 };
959 const std::array pre_barriers{
960 VkImageMemoryBarrier{
961 .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER,
962 .pNext = nullptr,
963 .srcAccessMask = VK_ACCESS_SHADER_WRITE_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT |
964 VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT |
965 VK_ACCESS_TRANSFER_WRITE_BIT,
966 .dstAccessMask = VK_ACCESS_TRANSFER_READ_BIT,
967 .oldLayout = VK_IMAGE_LAYOUT_GENERAL,
968 .newLayout = VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL,
969 .srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED,
970 .dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED,
971 .image = dst_image,
972 .subresourceRange = range,
973 },
974 };
975 const std::array post_barriers{
976 VkImageMemoryBarrier{
977 .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER,
978 .pNext = nullptr,
979 .srcAccessMask = 0,
980 .dstAccessMask = 0,
981 .oldLayout = VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL,
982 .newLayout = VK_IMAGE_LAYOUT_GENERAL,
983 .srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED,
984 .dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED,
985 .image = dst_image,
986 .subresourceRange = range,
987 },
988 };
989 cmdbuf.PipelineBarrier(VK_PIPELINE_STAGE_ALL_COMMANDS_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT,
990 0, READ_BARRIER, {}, pre_barriers);
991 cmdbuf.CopyBufferToImage(src_buffer, dst_image, VK_IMAGE_LAYOUT_GENERAL, buffer_copy_info);
992 cmdbuf.PipelineBarrier(VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_ALL_COMMANDS_BIT,
993 0, nullptr, nullptr, post_barriers);
994 });
995 if (is_rescaled) {
996 image->ScaleUp();
997 }
998 return true;
999} 811}
1000 812
1001void RasterizerVulkan::UpdateDynamicStates() { 813void RasterizerVulkan::UpdateDynamicStates() {
diff --git a/src/video_core/renderer_vulkan/vk_rasterizer.h b/src/video_core/renderer_vulkan/vk_rasterizer.h
index 7746c5434..1659fbc13 100644
--- a/src/video_core/renderer_vulkan/vk_rasterizer.h
+++ b/src/video_core/renderer_vulkan/vk_rasterizer.h
@@ -59,6 +59,11 @@ public:
59 const Tegra::DMA::ImageOperand& dst) override; 59 const Tegra::DMA::ImageOperand& dst) override;
60 60
61private: 61private:
62 template <bool IS_IMAGE_UPLOAD>
63 bool DmaBufferImageCopy(const Tegra::DMA::ImageCopy& copy_info,
64 const Tegra::DMA::BufferOperand& src,
65 const Tegra::DMA::ImageOperand& dst);
66
62 BufferCache& buffer_cache; 67 BufferCache& buffer_cache;
63 TextureCache& texture_cache; 68 TextureCache& texture_cache;
64 Scheduler& scheduler; 69 Scheduler& scheduler;
diff --git a/src/video_core/renderer_vulkan/vk_texture_cache.cpp b/src/video_core/renderer_vulkan/vk_texture_cache.cpp
index 8a204f93f..bf6389ff1 100644
--- a/src/video_core/renderer_vulkan/vk_texture_cache.cpp
+++ b/src/video_core/renderer_vulkan/vk_texture_cache.cpp
@@ -1312,15 +1312,16 @@ Image::Image(const VideoCommon::NullImageParams& params) : VideoCommon::ImageBas
1312 1312
1313Image::~Image() = default; 1313Image::~Image() = default;
1314 1314
1315void Image::UploadMemory(const StagingBufferRef& map, std::span<const BufferImageCopy> copies) { 1315void Image::UploadMemory(VkBuffer buffer, VkDeviceSize offset,
1316 std::span<const VideoCommon::BufferImageCopy> copies) {
1316 // TODO: Move this to another API 1317 // TODO: Move this to another API
1317 const bool is_rescaled = True(flags & ImageFlagBits::Rescaled); 1318 const bool is_rescaled = True(flags & ImageFlagBits::Rescaled);
1318 if (is_rescaled) { 1319 if (is_rescaled) {
1319 ScaleDown(true); 1320 ScaleDown(true);
1320 } 1321 }
1321 scheduler->RequestOutsideRenderPassOperationContext(); 1322 scheduler->RequestOutsideRenderPassOperationContext();
1322 std::vector vk_copies = TransformBufferImageCopies(copies, map.offset, aspect_mask); 1323 std::vector vk_copies = TransformBufferImageCopies(copies, offset, aspect_mask);
1323 const VkBuffer src_buffer = map.buffer; 1324 const VkBuffer src_buffer = buffer;
1324 const VkImage vk_image = *original_image; 1325 const VkImage vk_image = *original_image;
1325 const VkImageAspectFlags vk_aspect_mask = aspect_mask; 1326 const VkImageAspectFlags vk_aspect_mask = aspect_mask;
1326 const bool is_initialized = std::exchange(initialized, true); 1327 const bool is_initialized = std::exchange(initialized, true);
@@ -1333,14 +1334,19 @@ void Image::UploadMemory(const StagingBufferRef& map, std::span<const BufferImag
1333 } 1334 }
1334} 1335}
1335 1336
1336void Image::DownloadMemory(const StagingBufferRef& map, std::span<const BufferImageCopy> copies) { 1337void Image::UploadMemory(const StagingBufferRef& map, std::span<const BufferImageCopy> copies) {
1338 UploadMemory(map.buffer, map.offset, copies);
1339}
1340
1341void Image::DownloadMemory(VkBuffer buffer, VkDeviceSize offset,
1342 std::span<const VideoCommon::BufferImageCopy> copies) {
1337 const bool is_rescaled = True(flags & ImageFlagBits::Rescaled); 1343 const bool is_rescaled = True(flags & ImageFlagBits::Rescaled);
1338 if (is_rescaled) { 1344 if (is_rescaled) {
1339 ScaleDown(); 1345 ScaleDown();
1340 } 1346 }
1341 std::vector vk_copies = TransformBufferImageCopies(copies, map.offset, aspect_mask); 1347 std::vector vk_copies = TransformBufferImageCopies(copies, offset, aspect_mask);
1342 scheduler->RequestOutsideRenderPassOperationContext(); 1348 scheduler->RequestOutsideRenderPassOperationContext();
1343 scheduler->Record([buffer = map.buffer, image = *original_image, aspect_mask = aspect_mask, 1349 scheduler->Record([buffer, image = *original_image, aspect_mask = aspect_mask,
1344 vk_copies](vk::CommandBuffer cmdbuf) { 1350 vk_copies](vk::CommandBuffer cmdbuf) {
1345 const VkImageMemoryBarrier read_barrier{ 1351 const VkImageMemoryBarrier read_barrier{
1346 .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, 1352 .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER,
@@ -1395,6 +1401,10 @@ void Image::DownloadMemory(const StagingBufferRef& map, std::span<const BufferIm
1395 } 1401 }
1396} 1402}
1397 1403
1404void Image::DownloadMemory(const StagingBufferRef& map, std::span<const BufferImageCopy> copies) {
1405 DownloadMemory(map.buffer, map.offset, copies);
1406}
1407
1398bool Image::IsRescaled() const noexcept { 1408bool Image::IsRescaled() const noexcept {
1399 return True(flags & ImageFlagBits::Rescaled); 1409 return True(flags & ImageFlagBits::Rescaled);
1400} 1410}
diff --git a/src/video_core/renderer_vulkan/vk_texture_cache.h b/src/video_core/renderer_vulkan/vk_texture_cache.h
index 0ce39616f..d5ee23f8d 100644
--- a/src/video_core/renderer_vulkan/vk_texture_cache.h
+++ b/src/video_core/renderer_vulkan/vk_texture_cache.h
@@ -132,9 +132,15 @@ public:
132 Image(Image&&) = default; 132 Image(Image&&) = default;
133 Image& operator=(Image&&) = default; 133 Image& operator=(Image&&) = default;
134 134
135 void UploadMemory(VkBuffer buffer, VkDeviceSize offset,
136 std::span<const VideoCommon::BufferImageCopy> copies);
137
135 void UploadMemory(const StagingBufferRef& map, 138 void UploadMemory(const StagingBufferRef& map,
136 std::span<const VideoCommon::BufferImageCopy> copies); 139 std::span<const VideoCommon::BufferImageCopy> copies);
137 140
141 void DownloadMemory(VkBuffer buffer, VkDeviceSize offset,
142 std::span<const VideoCommon::BufferImageCopy> copies);
143
138 void DownloadMemory(const StagingBufferRef& map, 144 void DownloadMemory(const StagingBufferRef& map,
139 std::span<const VideoCommon::BufferImageCopy> copies); 145 std::span<const VideoCommon::BufferImageCopy> copies);
140 146