diff options
Diffstat (limited to 'src/shader_recompiler/frontend/ir/microinstruction.cpp')
| -rw-r--r-- | src/shader_recompiler/frontend/ir/microinstruction.cpp | 411 |
1 files changed, 411 insertions, 0 deletions
diff --git a/src/shader_recompiler/frontend/ir/microinstruction.cpp b/src/shader_recompiler/frontend/ir/microinstruction.cpp new file mode 100644 index 000000000..3dfa5a880 --- /dev/null +++ b/src/shader_recompiler/frontend/ir/microinstruction.cpp | |||
| @@ -0,0 +1,411 @@ | |||
| 1 | // Copyright 2021 yuzu Emulator Project | ||
| 2 | // Licensed under GPLv2 or any later version | ||
| 3 | // Refer to the license.txt file included. | ||
| 4 | |||
| 5 | #include <algorithm> | ||
| 6 | #include <memory> | ||
| 7 | |||
| 8 | #include "shader_recompiler/exception.h" | ||
| 9 | #include "shader_recompiler/frontend/ir/type.h" | ||
| 10 | #include "shader_recompiler/frontend/ir/value.h" | ||
| 11 | |||
| 12 | namespace Shader::IR { | ||
| 13 | namespace { | ||
| 14 | void CheckPseudoInstruction(IR::Inst* inst, IR::Opcode opcode) { | ||
| 15 | if (inst && inst->GetOpcode() != opcode) { | ||
| 16 | throw LogicError("Invalid pseudo-instruction"); | ||
| 17 | } | ||
| 18 | } | ||
| 19 | |||
| 20 | void SetPseudoInstruction(IR::Inst*& dest_inst, IR::Inst* pseudo_inst) { | ||
| 21 | if (dest_inst) { | ||
| 22 | throw LogicError("Only one of each type of pseudo-op allowed"); | ||
| 23 | } | ||
| 24 | dest_inst = pseudo_inst; | ||
| 25 | } | ||
| 26 | |||
| 27 | void RemovePseudoInstruction(IR::Inst*& inst, IR::Opcode expected_opcode) { | ||
| 28 | if (inst->GetOpcode() != expected_opcode) { | ||
| 29 | throw LogicError("Undoing use of invalid pseudo-op"); | ||
| 30 | } | ||
| 31 | inst = nullptr; | ||
| 32 | } | ||
| 33 | |||
| 34 | void AllocAssociatedInsts(std::unique_ptr<AssociatedInsts>& associated_insts) { | ||
| 35 | if (!associated_insts) { | ||
| 36 | associated_insts = std::make_unique<AssociatedInsts>(); | ||
| 37 | } | ||
| 38 | } | ||
| 39 | } // Anonymous namespace | ||
| 40 | |||
| 41 | Inst::Inst(IR::Opcode op_, u32 flags_) noexcept : op{op_}, flags{flags_} { | ||
| 42 | if (op == Opcode::Phi) { | ||
| 43 | std::construct_at(&phi_args); | ||
| 44 | } else { | ||
| 45 | std::construct_at(&args); | ||
| 46 | } | ||
| 47 | } | ||
| 48 | |||
| 49 | Inst::~Inst() { | ||
| 50 | if (op == Opcode::Phi) { | ||
| 51 | std::destroy_at(&phi_args); | ||
| 52 | } else { | ||
| 53 | std::destroy_at(&args); | ||
| 54 | } | ||
| 55 | } | ||
| 56 | |||
| 57 | bool Inst::MayHaveSideEffects() const noexcept { | ||
| 58 | switch (op) { | ||
| 59 | case Opcode::ConditionRef: | ||
| 60 | case Opcode::Reference: | ||
| 61 | case Opcode::PhiMove: | ||
| 62 | case Opcode::Prologue: | ||
| 63 | case Opcode::Epilogue: | ||
| 64 | case Opcode::Join: | ||
| 65 | case Opcode::DemoteToHelperInvocation: | ||
| 66 | case Opcode::Barrier: | ||
| 67 | case Opcode::WorkgroupMemoryBarrier: | ||
| 68 | case Opcode::DeviceMemoryBarrier: | ||
| 69 | case Opcode::EmitVertex: | ||
| 70 | case Opcode::EndPrimitive: | ||
| 71 | case Opcode::SetAttribute: | ||
| 72 | case Opcode::SetAttributeIndexed: | ||
| 73 | case Opcode::SetPatch: | ||
| 74 | case Opcode::SetFragColor: | ||
| 75 | case Opcode::SetSampleMask: | ||
| 76 | case Opcode::SetFragDepth: | ||
| 77 | case Opcode::WriteGlobalU8: | ||
| 78 | case Opcode::WriteGlobalS8: | ||
| 79 | case Opcode::WriteGlobalU16: | ||
| 80 | case Opcode::WriteGlobalS16: | ||
| 81 | case Opcode::WriteGlobal32: | ||
| 82 | case Opcode::WriteGlobal64: | ||
| 83 | case Opcode::WriteGlobal128: | ||
| 84 | case Opcode::WriteStorageU8: | ||
| 85 | case Opcode::WriteStorageS8: | ||
| 86 | case Opcode::WriteStorageU16: | ||
| 87 | case Opcode::WriteStorageS16: | ||
| 88 | case Opcode::WriteStorage32: | ||
| 89 | case Opcode::WriteStorage64: | ||
| 90 | case Opcode::WriteStorage128: | ||
| 91 | case Opcode::WriteLocal: | ||
| 92 | case Opcode::WriteSharedU8: | ||
| 93 | case Opcode::WriteSharedU16: | ||
| 94 | case Opcode::WriteSharedU32: | ||
| 95 | case Opcode::WriteSharedU64: | ||
| 96 | case Opcode::WriteSharedU128: | ||
| 97 | case Opcode::SharedAtomicIAdd32: | ||
| 98 | case Opcode::SharedAtomicSMin32: | ||
| 99 | case Opcode::SharedAtomicUMin32: | ||
| 100 | case Opcode::SharedAtomicSMax32: | ||
| 101 | case Opcode::SharedAtomicUMax32: | ||
| 102 | case Opcode::SharedAtomicInc32: | ||
| 103 | case Opcode::SharedAtomicDec32: | ||
| 104 | case Opcode::SharedAtomicAnd32: | ||
| 105 | case Opcode::SharedAtomicOr32: | ||
| 106 | case Opcode::SharedAtomicXor32: | ||
| 107 | case Opcode::SharedAtomicExchange32: | ||
| 108 | case Opcode::SharedAtomicExchange64: | ||
| 109 | case Opcode::GlobalAtomicIAdd32: | ||
| 110 | case Opcode::GlobalAtomicSMin32: | ||
| 111 | case Opcode::GlobalAtomicUMin32: | ||
| 112 | case Opcode::GlobalAtomicSMax32: | ||
| 113 | case Opcode::GlobalAtomicUMax32: | ||
| 114 | case Opcode::GlobalAtomicInc32: | ||
| 115 | case Opcode::GlobalAtomicDec32: | ||
| 116 | case Opcode::GlobalAtomicAnd32: | ||
| 117 | case Opcode::GlobalAtomicOr32: | ||
| 118 | case Opcode::GlobalAtomicXor32: | ||
| 119 | case Opcode::GlobalAtomicExchange32: | ||
| 120 | case Opcode::GlobalAtomicIAdd64: | ||
| 121 | case Opcode::GlobalAtomicSMin64: | ||
| 122 | case Opcode::GlobalAtomicUMin64: | ||
| 123 | case Opcode::GlobalAtomicSMax64: | ||
| 124 | case Opcode::GlobalAtomicUMax64: | ||
| 125 | case Opcode::GlobalAtomicAnd64: | ||
| 126 | case Opcode::GlobalAtomicOr64: | ||
| 127 | case Opcode::GlobalAtomicXor64: | ||
| 128 | case Opcode::GlobalAtomicExchange64: | ||
| 129 | case Opcode::GlobalAtomicAddF32: | ||
| 130 | case Opcode::GlobalAtomicAddF16x2: | ||
| 131 | case Opcode::GlobalAtomicAddF32x2: | ||
| 132 | case Opcode::GlobalAtomicMinF16x2: | ||
| 133 | case Opcode::GlobalAtomicMinF32x2: | ||
| 134 | case Opcode::GlobalAtomicMaxF16x2: | ||
| 135 | case Opcode::GlobalAtomicMaxF32x2: | ||
| 136 | case Opcode::StorageAtomicIAdd32: | ||
| 137 | case Opcode::StorageAtomicSMin32: | ||
| 138 | case Opcode::StorageAtomicUMin32: | ||
| 139 | case Opcode::StorageAtomicSMax32: | ||
| 140 | case Opcode::StorageAtomicUMax32: | ||
| 141 | case Opcode::StorageAtomicInc32: | ||
| 142 | case Opcode::StorageAtomicDec32: | ||
| 143 | case Opcode::StorageAtomicAnd32: | ||
| 144 | case Opcode::StorageAtomicOr32: | ||
| 145 | case Opcode::StorageAtomicXor32: | ||
| 146 | case Opcode::StorageAtomicExchange32: | ||
| 147 | case Opcode::StorageAtomicIAdd64: | ||
| 148 | case Opcode::StorageAtomicSMin64: | ||
| 149 | case Opcode::StorageAtomicUMin64: | ||
| 150 | case Opcode::StorageAtomicSMax64: | ||
| 151 | case Opcode::StorageAtomicUMax64: | ||
| 152 | case Opcode::StorageAtomicAnd64: | ||
| 153 | case Opcode::StorageAtomicOr64: | ||
| 154 | case Opcode::StorageAtomicXor64: | ||
| 155 | case Opcode::StorageAtomicExchange64: | ||
| 156 | case Opcode::StorageAtomicAddF32: | ||
| 157 | case Opcode::StorageAtomicAddF16x2: | ||
| 158 | case Opcode::StorageAtomicAddF32x2: | ||
| 159 | case Opcode::StorageAtomicMinF16x2: | ||
| 160 | case Opcode::StorageAtomicMinF32x2: | ||
| 161 | case Opcode::StorageAtomicMaxF16x2: | ||
| 162 | case Opcode::StorageAtomicMaxF32x2: | ||
| 163 | case Opcode::BindlessImageWrite: | ||
| 164 | case Opcode::BoundImageWrite: | ||
| 165 | case Opcode::ImageWrite: | ||
| 166 | case IR::Opcode::BindlessImageAtomicIAdd32: | ||
| 167 | case IR::Opcode::BindlessImageAtomicSMin32: | ||
| 168 | case IR::Opcode::BindlessImageAtomicUMin32: | ||
| 169 | case IR::Opcode::BindlessImageAtomicSMax32: | ||
| 170 | case IR::Opcode::BindlessImageAtomicUMax32: | ||
| 171 | case IR::Opcode::BindlessImageAtomicInc32: | ||
| 172 | case IR::Opcode::BindlessImageAtomicDec32: | ||
| 173 | case IR::Opcode::BindlessImageAtomicAnd32: | ||
| 174 | case IR::Opcode::BindlessImageAtomicOr32: | ||
| 175 | case IR::Opcode::BindlessImageAtomicXor32: | ||
| 176 | case IR::Opcode::BindlessImageAtomicExchange32: | ||
| 177 | case IR::Opcode::BoundImageAtomicIAdd32: | ||
| 178 | case IR::Opcode::BoundImageAtomicSMin32: | ||
| 179 | case IR::Opcode::BoundImageAtomicUMin32: | ||
| 180 | case IR::Opcode::BoundImageAtomicSMax32: | ||
| 181 | case IR::Opcode::BoundImageAtomicUMax32: | ||
| 182 | case IR::Opcode::BoundImageAtomicInc32: | ||
| 183 | case IR::Opcode::BoundImageAtomicDec32: | ||
| 184 | case IR::Opcode::BoundImageAtomicAnd32: | ||
| 185 | case IR::Opcode::BoundImageAtomicOr32: | ||
| 186 | case IR::Opcode::BoundImageAtomicXor32: | ||
| 187 | case IR::Opcode::BoundImageAtomicExchange32: | ||
| 188 | case IR::Opcode::ImageAtomicIAdd32: | ||
| 189 | case IR::Opcode::ImageAtomicSMin32: | ||
| 190 | case IR::Opcode::ImageAtomicUMin32: | ||
| 191 | case IR::Opcode::ImageAtomicSMax32: | ||
| 192 | case IR::Opcode::ImageAtomicUMax32: | ||
| 193 | case IR::Opcode::ImageAtomicInc32: | ||
| 194 | case IR::Opcode::ImageAtomicDec32: | ||
| 195 | case IR::Opcode::ImageAtomicAnd32: | ||
| 196 | case IR::Opcode::ImageAtomicOr32: | ||
| 197 | case IR::Opcode::ImageAtomicXor32: | ||
| 198 | case IR::Opcode::ImageAtomicExchange32: | ||
| 199 | return true; | ||
| 200 | default: | ||
| 201 | return false; | ||
| 202 | } | ||
| 203 | } | ||
| 204 | |||
| 205 | bool Inst::IsPseudoInstruction() const noexcept { | ||
| 206 | switch (op) { | ||
| 207 | case Opcode::GetZeroFromOp: | ||
| 208 | case Opcode::GetSignFromOp: | ||
| 209 | case Opcode::GetCarryFromOp: | ||
| 210 | case Opcode::GetOverflowFromOp: | ||
| 211 | case Opcode::GetSparseFromOp: | ||
| 212 | case Opcode::GetInBoundsFromOp: | ||
| 213 | return true; | ||
| 214 | default: | ||
| 215 | return false; | ||
| 216 | } | ||
| 217 | } | ||
| 218 | |||
| 219 | bool Inst::AreAllArgsImmediates() const { | ||
| 220 | if (op == Opcode::Phi) { | ||
| 221 | throw LogicError("Testing for all arguments are immediates on phi instruction"); | ||
| 222 | } | ||
| 223 | return std::all_of(args.begin(), args.begin() + NumArgs(), | ||
| 224 | [](const IR::Value& value) { return value.IsImmediate(); }); | ||
| 225 | } | ||
| 226 | |||
| 227 | Inst* Inst::GetAssociatedPseudoOperation(IR::Opcode opcode) { | ||
| 228 | if (!associated_insts) { | ||
| 229 | return nullptr; | ||
| 230 | } | ||
| 231 | switch (opcode) { | ||
| 232 | case Opcode::GetZeroFromOp: | ||
| 233 | CheckPseudoInstruction(associated_insts->zero_inst, Opcode::GetZeroFromOp); | ||
| 234 | return associated_insts->zero_inst; | ||
| 235 | case Opcode::GetSignFromOp: | ||
| 236 | CheckPseudoInstruction(associated_insts->sign_inst, Opcode::GetSignFromOp); | ||
| 237 | return associated_insts->sign_inst; | ||
| 238 | case Opcode::GetCarryFromOp: | ||
| 239 | CheckPseudoInstruction(associated_insts->carry_inst, Opcode::GetCarryFromOp); | ||
| 240 | return associated_insts->carry_inst; | ||
| 241 | case Opcode::GetOverflowFromOp: | ||
| 242 | CheckPseudoInstruction(associated_insts->overflow_inst, Opcode::GetOverflowFromOp); | ||
| 243 | return associated_insts->overflow_inst; | ||
| 244 | case Opcode::GetSparseFromOp: | ||
| 245 | CheckPseudoInstruction(associated_insts->sparse_inst, Opcode::GetSparseFromOp); | ||
| 246 | return associated_insts->sparse_inst; | ||
| 247 | case Opcode::GetInBoundsFromOp: | ||
| 248 | CheckPseudoInstruction(associated_insts->in_bounds_inst, Opcode::GetInBoundsFromOp); | ||
| 249 | return associated_insts->in_bounds_inst; | ||
| 250 | default: | ||
| 251 | throw InvalidArgument("{} is not a pseudo-instruction", opcode); | ||
| 252 | } | ||
| 253 | } | ||
| 254 | |||
| 255 | IR::Type Inst::Type() const { | ||
| 256 | return TypeOf(op); | ||
| 257 | } | ||
| 258 | |||
| 259 | void Inst::SetArg(size_t index, Value value) { | ||
| 260 | if (index >= NumArgs()) { | ||
| 261 | throw InvalidArgument("Out of bounds argument index {} in opcode {}", index, op); | ||
| 262 | } | ||
| 263 | const IR::Value arg{Arg(index)}; | ||
| 264 | if (!arg.IsImmediate()) { | ||
| 265 | UndoUse(arg); | ||
| 266 | } | ||
| 267 | if (!value.IsImmediate()) { | ||
| 268 | Use(value); | ||
| 269 | } | ||
| 270 | if (op == Opcode::Phi) { | ||
| 271 | phi_args[index].second = value; | ||
| 272 | } else { | ||
| 273 | args[index] = value; | ||
| 274 | } | ||
| 275 | } | ||
| 276 | |||
| 277 | Block* Inst::PhiBlock(size_t index) const { | ||
| 278 | if (op != Opcode::Phi) { | ||
| 279 | throw LogicError("{} is not a Phi instruction", op); | ||
| 280 | } | ||
| 281 | if (index >= phi_args.size()) { | ||
| 282 | throw InvalidArgument("Out of bounds argument index {} in phi instruction"); | ||
| 283 | } | ||
| 284 | return phi_args[index].first; | ||
| 285 | } | ||
| 286 | |||
| 287 | void Inst::AddPhiOperand(Block* predecessor, const Value& value) { | ||
| 288 | if (!value.IsImmediate()) { | ||
| 289 | Use(value); | ||
| 290 | } | ||
| 291 | phi_args.emplace_back(predecessor, value); | ||
| 292 | } | ||
| 293 | |||
| 294 | void Inst::Invalidate() { | ||
| 295 | ClearArgs(); | ||
| 296 | ReplaceOpcode(Opcode::Void); | ||
| 297 | } | ||
| 298 | |||
| 299 | void Inst::ClearArgs() { | ||
| 300 | if (op == Opcode::Phi) { | ||
| 301 | for (auto& pair : phi_args) { | ||
| 302 | IR::Value& value{pair.second}; | ||
| 303 | if (!value.IsImmediate()) { | ||
| 304 | UndoUse(value); | ||
| 305 | } | ||
| 306 | } | ||
| 307 | phi_args.clear(); | ||
| 308 | } else { | ||
| 309 | for (auto& value : args) { | ||
| 310 | if (!value.IsImmediate()) { | ||
| 311 | UndoUse(value); | ||
| 312 | } | ||
| 313 | } | ||
| 314 | // Reset arguments to null | ||
| 315 | // std::memset was measured to be faster on MSVC than std::ranges:fill | ||
| 316 | std::memset(reinterpret_cast<char*>(&args), 0, sizeof(args)); | ||
| 317 | } | ||
| 318 | } | ||
| 319 | |||
| 320 | void Inst::ReplaceUsesWith(Value replacement) { | ||
| 321 | Invalidate(); | ||
| 322 | ReplaceOpcode(Opcode::Identity); | ||
| 323 | if (!replacement.IsImmediate()) { | ||
| 324 | Use(replacement); | ||
| 325 | } | ||
| 326 | args[0] = replacement; | ||
| 327 | } | ||
| 328 | |||
| 329 | void Inst::ReplaceOpcode(IR::Opcode opcode) { | ||
| 330 | if (opcode == IR::Opcode::Phi) { | ||
| 331 | throw LogicError("Cannot transition into Phi"); | ||
| 332 | } | ||
| 333 | if (op == Opcode::Phi) { | ||
| 334 | // Transition out of phi arguments into non-phi | ||
| 335 | std::destroy_at(&phi_args); | ||
| 336 | std::construct_at(&args); | ||
| 337 | } | ||
| 338 | op = opcode; | ||
| 339 | } | ||
| 340 | |||
| 341 | void Inst::Use(const Value& value) { | ||
| 342 | Inst* const inst{value.Inst()}; | ||
| 343 | ++inst->use_count; | ||
| 344 | |||
| 345 | std::unique_ptr<AssociatedInsts>& assoc_inst{inst->associated_insts}; | ||
| 346 | switch (op) { | ||
| 347 | case Opcode::GetZeroFromOp: | ||
| 348 | AllocAssociatedInsts(assoc_inst); | ||
| 349 | SetPseudoInstruction(assoc_inst->zero_inst, this); | ||
| 350 | break; | ||
| 351 | case Opcode::GetSignFromOp: | ||
| 352 | AllocAssociatedInsts(assoc_inst); | ||
| 353 | SetPseudoInstruction(assoc_inst->sign_inst, this); | ||
| 354 | break; | ||
| 355 | case Opcode::GetCarryFromOp: | ||
| 356 | AllocAssociatedInsts(assoc_inst); | ||
| 357 | SetPseudoInstruction(assoc_inst->carry_inst, this); | ||
| 358 | break; | ||
| 359 | case Opcode::GetOverflowFromOp: | ||
| 360 | AllocAssociatedInsts(assoc_inst); | ||
| 361 | SetPseudoInstruction(assoc_inst->overflow_inst, this); | ||
| 362 | break; | ||
| 363 | case Opcode::GetSparseFromOp: | ||
| 364 | AllocAssociatedInsts(assoc_inst); | ||
| 365 | SetPseudoInstruction(assoc_inst->sparse_inst, this); | ||
| 366 | break; | ||
| 367 | case Opcode::GetInBoundsFromOp: | ||
| 368 | AllocAssociatedInsts(assoc_inst); | ||
| 369 | SetPseudoInstruction(assoc_inst->in_bounds_inst, this); | ||
| 370 | break; | ||
| 371 | default: | ||
| 372 | break; | ||
| 373 | } | ||
| 374 | } | ||
| 375 | |||
| 376 | void Inst::UndoUse(const Value& value) { | ||
| 377 | Inst* const inst{value.Inst()}; | ||
| 378 | --inst->use_count; | ||
| 379 | |||
| 380 | std::unique_ptr<AssociatedInsts>& assoc_inst{inst->associated_insts}; | ||
| 381 | switch (op) { | ||
| 382 | case Opcode::GetZeroFromOp: | ||
| 383 | AllocAssociatedInsts(assoc_inst); | ||
| 384 | RemovePseudoInstruction(assoc_inst->zero_inst, Opcode::GetZeroFromOp); | ||
| 385 | break; | ||
| 386 | case Opcode::GetSignFromOp: | ||
| 387 | AllocAssociatedInsts(assoc_inst); | ||
| 388 | RemovePseudoInstruction(assoc_inst->sign_inst, Opcode::GetSignFromOp); | ||
| 389 | break; | ||
| 390 | case Opcode::GetCarryFromOp: | ||
| 391 | AllocAssociatedInsts(assoc_inst); | ||
| 392 | RemovePseudoInstruction(assoc_inst->carry_inst, Opcode::GetCarryFromOp); | ||
| 393 | break; | ||
| 394 | case Opcode::GetOverflowFromOp: | ||
| 395 | AllocAssociatedInsts(assoc_inst); | ||
| 396 | RemovePseudoInstruction(assoc_inst->overflow_inst, Opcode::GetOverflowFromOp); | ||
| 397 | break; | ||
| 398 | case Opcode::GetSparseFromOp: | ||
| 399 | AllocAssociatedInsts(assoc_inst); | ||
| 400 | RemovePseudoInstruction(assoc_inst->sparse_inst, Opcode::GetSparseFromOp); | ||
| 401 | break; | ||
| 402 | case Opcode::GetInBoundsFromOp: | ||
| 403 | AllocAssociatedInsts(assoc_inst); | ||
| 404 | RemovePseudoInstruction(assoc_inst->in_bounds_inst, Opcode::GetInBoundsFromOp); | ||
| 405 | break; | ||
| 406 | default: | ||
| 407 | break; | ||
| 408 | } | ||
| 409 | } | ||
| 410 | |||
| 411 | } // namespace Shader::IR | ||