diff options
Diffstat (limited to 'src/shader_recompiler')
| -rw-r--r-- | src/shader_recompiler/frontend/maxwell/translate/impl/atomic_operations_global_memory.cpp | 80 |
1 files changed, 36 insertions, 44 deletions
diff --git a/src/shader_recompiler/frontend/maxwell/translate/impl/atomic_operations_global_memory.cpp b/src/shader_recompiler/frontend/maxwell/translate/impl/atomic_operations_global_memory.cpp index 7a32c5eb3..66f39e44e 100644 --- a/src/shader_recompiler/frontend/maxwell/translate/impl/atomic_operations_global_memory.cpp +++ b/src/shader_recompiler/frontend/maxwell/translate/impl/atomic_operations_global_memory.cpp | |||
| @@ -154,69 +154,61 @@ void StoreResult(TranslatorVisitor& v, IR::Reg dest_reg, const IR::Value& result | |||
| 154 | break; | 154 | break; |
| 155 | } | 155 | } |
| 156 | } | 156 | } |
| 157 | |||
| 158 | IR::Value ApplyAtomOp(TranslatorVisitor& v, IR::Reg operand_reg, const IR::U64& offset, | ||
| 159 | AtomSize size, AtomOp op) { | ||
| 160 | switch (size) { | ||
| 161 | case AtomSize::U32: | ||
| 162 | case AtomSize::S32: | ||
| 163 | return ApplyIntegerAtomOp(v.ir, offset, v.X(operand_reg), op, size == AtomSize::S32); | ||
| 164 | case AtomSize::U64: | ||
| 165 | case AtomSize::S64: | ||
| 166 | return ApplyIntegerAtomOp(v.ir, offset, v.L(operand_reg), op, size == AtomSize::S64); | ||
| 167 | case AtomSize::F32: | ||
| 168 | return ApplyFpAtomOp(v.ir, offset, v.F(operand_reg), op, size); | ||
| 169 | case AtomSize::F16x2: { | ||
| 170 | return ApplyFpAtomOp(v.ir, offset, v.ir.UnpackFloat2x16(v.X(operand_reg)), op, size); | ||
| 171 | } | ||
| 172 | default: | ||
| 173 | throw NotImplementedException("Atom Size {}", size); | ||
| 174 | } | ||
| 175 | } | ||
| 176 | |||
| 177 | void GlobalAtomic(TranslatorVisitor& v, IR::Reg dest_reg, IR::Reg operand_reg, | ||
| 178 | const IR::U64& offset, AtomSize size, AtomOp op, bool write_dest) { | ||
| 179 | IR::Value result; | ||
| 180 | if (AtomOpNotApplicable(size, op)) { | ||
| 181 | result = LoadGlobal(v.ir, offset, size); | ||
| 182 | } else { | ||
| 183 | result = ApplyAtomOp(v, operand_reg, offset, size, op); | ||
| 184 | } | ||
| 185 | if (write_dest) { | ||
| 186 | StoreResult(v, dest_reg, result, size); | ||
| 187 | } | ||
| 188 | } | ||
| 157 | } // Anonymous namespace | 189 | } // Anonymous namespace |
| 158 | 190 | ||
| 159 | void TranslatorVisitor::ATOM(u64 insn) { | 191 | void TranslatorVisitor::ATOM(u64 insn) { |
| 160 | union { | 192 | union { |
| 161 | u64 raw; | 193 | u64 raw; |
| 162 | BitField<0, 8, IR::Reg> dest_reg; | 194 | BitField<0, 8, IR::Reg> dest_reg; |
| 163 | BitField<8, 8, IR::Reg> addr_reg; | 195 | BitField<20, 8, IR::Reg> operand_reg; |
| 164 | BitField<20, 8, IR::Reg> src_reg_b; | ||
| 165 | BitField<49, 3, AtomSize> size; | 196 | BitField<49, 3, AtomSize> size; |
| 166 | BitField<52, 4, AtomOp> op; | 197 | BitField<52, 4, AtomOp> op; |
| 167 | } const atom{insn}; | 198 | } const atom{insn}; |
| 168 | |||
| 169 | const bool size_64{atom.size == AtomSize::U64 || atom.size == AtomSize::S64}; | ||
| 170 | const bool is_signed{atom.size == AtomSize::S32 || atom.size == AtomSize::S64}; | ||
| 171 | const bool is_integer{atom.size != AtomSize::F32 && atom.size != AtomSize::F16x2}; | ||
| 172 | const IR::U64 offset{AtomOffset(*this, insn)}; | 199 | const IR::U64 offset{AtomOffset(*this, insn)}; |
| 173 | IR::Value result; | 200 | GlobalAtomic(*this, atom.dest_reg, atom.operand_reg, offset, atom.size, atom.op, true); |
| 174 | |||
| 175 | if (AtomOpNotApplicable(atom.size, atom.op)) { | ||
| 176 | result = LoadGlobal(ir, offset, atom.size); | ||
| 177 | } else if (!is_integer) { | ||
| 178 | if (atom.size == AtomSize::F32) { | ||
| 179 | result = ApplyFpAtomOp(ir, offset, F(atom.src_reg_b), atom.op, atom.size); | ||
| 180 | } else { | ||
| 181 | const IR::Value src_b{ir.UnpackFloat2x16(X(atom.src_reg_b))}; | ||
| 182 | result = ApplyFpAtomOp(ir, offset, src_b, atom.op, atom.size); | ||
| 183 | } | ||
| 184 | } else if (size_64) { | ||
| 185 | result = ApplyIntegerAtomOp(ir, offset, L(atom.src_reg_b), atom.op, is_signed); | ||
| 186 | } else { | ||
| 187 | result = ApplyIntegerAtomOp(ir, offset, X(atom.src_reg_b), atom.op, is_signed); | ||
| 188 | } | ||
| 189 | StoreResult(*this, atom.dest_reg, result, atom.size); | ||
| 190 | } | 201 | } |
| 191 | 202 | ||
| 192 | void TranslatorVisitor::RED(u64 insn) { | 203 | void TranslatorVisitor::RED(u64 insn) { |
| 193 | union { | 204 | union { |
| 194 | u64 raw; | 205 | u64 raw; |
| 195 | BitField<0, 8, IR::Reg> src_reg_b; | 206 | BitField<0, 8, IR::Reg> operand_reg; |
| 196 | BitField<8, 8, IR::Reg> addr_reg; | ||
| 197 | BitField<20, 3, AtomSize> size; | 207 | BitField<20, 3, AtomSize> size; |
| 198 | BitField<23, 3, AtomOp> op; | 208 | BitField<23, 3, AtomOp> op; |
| 199 | } const red{insn}; | 209 | } const red{insn}; |
| 200 | |||
| 201 | if (AtomOpNotApplicable(red.size, red.op)) { | ||
| 202 | return; | ||
| 203 | } | ||
| 204 | const bool size_64{red.size == AtomSize::U64 || red.size == AtomSize::S64}; | ||
| 205 | const bool is_signed{red.size == AtomSize::S32 || red.size == AtomSize::S64}; | ||
| 206 | const bool is_integer{red.size != AtomSize::F32 && red.size != AtomSize::F16x2}; | ||
| 207 | const IR::U64 offset{AtomOffset(*this, insn)}; | 210 | const IR::U64 offset{AtomOffset(*this, insn)}; |
| 208 | if (!is_integer) { | 211 | GlobalAtomic(*this, IR::Reg::RZ, red.operand_reg, offset, red.size, red.op, true); |
| 209 | if (red.size == AtomSize::F32) { | ||
| 210 | ApplyFpAtomOp(ir, offset, F(red.src_reg_b), red.op, red.size); | ||
| 211 | } else { | ||
| 212 | const IR::Value src_b{ir.UnpackFloat2x16(X(red.src_reg_b))}; | ||
| 213 | ApplyFpAtomOp(ir, offset, src_b, red.op, red.size); | ||
| 214 | } | ||
| 215 | } else if (size_64) { | ||
| 216 | ApplyIntegerAtomOp(ir, offset, L(red.src_reg_b), red.op, is_signed); | ||
| 217 | } else { | ||
| 218 | ApplyIntegerAtomOp(ir, offset, X(red.src_reg_b), red.op, is_signed); | ||
| 219 | } | ||
| 220 | } | 212 | } |
| 221 | 213 | ||
| 222 | } // namespace Shader::Maxwell | 214 | } // namespace Shader::Maxwell |