summaryrefslogtreecommitdiffstats
path: root/src/shader_recompiler/backend/spirv/emit_spirv_atomic.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/shader_recompiler/backend/spirv/emit_spirv_atomic.cpp')
-rw-r--r--src/shader_recompiler/backend/spirv/emit_spirv_atomic.cpp333
1 files changed, 126 insertions, 207 deletions
diff --git a/src/shader_recompiler/backend/spirv/emit_spirv_atomic.cpp b/src/shader_recompiler/backend/spirv/emit_spirv_atomic.cpp
index 03d891419..aab32dc52 100644
--- a/src/shader_recompiler/backend/spirv/emit_spirv_atomic.cpp
+++ b/src/shader_recompiler/backend/spirv/emit_spirv_atomic.cpp
@@ -6,11 +6,12 @@
namespace Shader::Backend::SPIRV {
namespace {
-
-Id GetSharedPointer(EmitContext& ctx, Id offset, u32 index_offset = 0) {
+Id SharedPointer(EmitContext& ctx, Id offset, u32 index_offset = 0) {
const Id shift_id{ctx.Constant(ctx.U32[1], 2U)};
- const Id shifted_value{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)};
- const Id index{ctx.OpIAdd(ctx.U32[1], shifted_value, ctx.Constant(ctx.U32[1], index_offset))};
+ Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)};
+ if (index_offset > 0) {
+ index = ctx.OpIAdd(ctx.U32[1], index, ctx.Constant(ctx.U32[1], index_offset));
+ }
return ctx.profile.support_explicit_workgroup_layout
? ctx.OpAccessChain(ctx.shared_u32, ctx.shared_memory_u32, ctx.u32_zero_value, index)
: ctx.OpAccessChain(ctx.shared_u32, ctx.shared_memory_u32, index);
@@ -30,340 +31,258 @@ Id StorageIndex(EmitContext& ctx, const IR::Value& offset, size_t element_size)
return ctx.OpShiftRightLogical(ctx.U32[1], index, shift_id);
}
-Id GetStoragePointer(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
- u32 index_offset = 0) {
- // TODO: Support reinterpreting bindings, guaranteed to be aligned
+Id StoragePointer(EmitContext& ctx, const StorageTypeDefinition& type_def,
+ Id StorageDefinitions::*member_ptr, const IR::Value& binding,
+ const IR::Value& offset, size_t element_size) {
if (!binding.IsImmediate()) {
throw NotImplementedException("Dynamic storage buffer indexing");
}
- const Id ssbo{ctx.ssbos[binding.U32()]};
- const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
- const Id index{ctx.OpIAdd(ctx.U32[1], base_index, ctx.Constant(ctx.U32[1], index_offset))};
- return ctx.OpAccessChain(ctx.storage_u32, ssbo, ctx.u32_zero_value, index);
+ const Id ssbo{ctx.ssbos[binding.U32()].*member_ptr};
+ const Id index{StorageIndex(ctx, offset, element_size)};
+ return ctx.OpAccessChain(type_def.element, ssbo, ctx.u32_zero_value, index);
}
-std::pair<Id, Id> GetAtomicArgs(EmitContext& ctx) {
+std::pair<Id, Id> AtomicArgs(EmitContext& ctx) {
const Id scope{ctx.Constant(ctx.U32[1], static_cast<u32>(spv::Scope::Device))};
const Id semantics{ctx.u32_zero_value};
return {scope, semantics};
}
-Id LoadU64(EmitContext& ctx, Id pointer_1, Id pointer_2) {
- const Id value_1{ctx.OpLoad(ctx.U32[1], pointer_1)};
- const Id value_2{ctx.OpLoad(ctx.U32[1], pointer_2)};
- const Id original_composite{ctx.OpCompositeConstruct(ctx.U32[2], value_1, value_2)};
- return ctx.OpBitcast(ctx.U64, original_composite);
+Id SharedAtomicU32(EmitContext& ctx, Id offset, Id value,
+ Id (Sirit::Module::*atomic_func)(Id, Id, Id, Id, Id)) {
+ const Id pointer{SharedPointer(ctx, offset)};
+ const auto [scope, semantics]{AtomicArgs(ctx)};
+ return (ctx.*atomic_func)(ctx.U32[1], pointer, scope, semantics, value);
}
-void StoreResult(EmitContext& ctx, Id pointer_1, Id pointer_2, Id result) {
- const Id composite{ctx.OpBitcast(ctx.U32[2], result)};
- ctx.OpStore(pointer_1, ctx.OpCompositeExtract(ctx.U32[1], composite, 0));
- ctx.OpStore(pointer_2, ctx.OpCompositeExtract(ctx.U32[1], composite, 1));
+Id StorageAtomicU32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, Id value,
+ Id (Sirit::Module::*atomic_func)(Id, Id, Id, Id, Id)) {
+ const Id pointer{StoragePointer(ctx, ctx.storage_types.U32, &StorageDefinitions::U32, binding,
+ offset, sizeof(u32))};
+ const auto [scope, semantics]{AtomicArgs(ctx)};
+ return (ctx.*atomic_func)(ctx.U32[1], pointer, scope, semantics, value);
+}
+
+Id StorageAtomicU64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset, Id value,
+ Id (Sirit::Module::*atomic_func)(Id, Id, Id, Id, Id),
+ Id (Sirit::Module::*non_atomic_func)(Id, Id, Id)) {
+ if (ctx.profile.support_int64_atomics) {
+ const Id pointer{StoragePointer(ctx, ctx.storage_types.U64, &StorageDefinitions::U64,
+ binding, offset, sizeof(u64))};
+ const auto [scope, semantics]{AtomicArgs(ctx)};
+ return (ctx.*atomic_func)(ctx.U64, pointer, scope, semantics, value);
+ }
+ // LOG_WARNING(..., "Int64 Atomics not supported, fallback to non-atomic");
+ const Id pointer{StoragePointer(ctx, ctx.storage_types.U32x2, &StorageDefinitions::U32x2,
+ binding, offset, sizeof(u32[2]))};
+ const Id original_value{ctx.OpBitcast(ctx.U64, ctx.OpLoad(ctx.U32[2], pointer))};
+ const Id result{(ctx.*non_atomic_func)(ctx.U64, value, original_value)};
+ ctx.OpStore(pointer, result);
+ return original_value;
}
} // Anonymous namespace
-Id EmitSharedAtomicIAdd32(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer{GetSharedPointer(ctx, pointer_offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicIAdd(ctx.U32[1], pointer, scope, semantics, value);
+Id EmitSharedAtomicIAdd32(EmitContext& ctx, Id offset, Id value) {
+ return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicIAdd);
}
-Id EmitSharedAtomicSMin32(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer{GetSharedPointer(ctx, pointer_offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicSMin(ctx.U32[1], pointer, scope, semantics, value);
+Id EmitSharedAtomicSMin32(EmitContext& ctx, Id offset, Id value) {
+ return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicSMin);
}
-Id EmitSharedAtomicUMin32(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer{GetSharedPointer(ctx, pointer_offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicUMin(ctx.U32[1], pointer, scope, semantics, value);
+Id EmitSharedAtomicUMin32(EmitContext& ctx, Id offset, Id value) {
+ return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicUMin);
}
-Id EmitSharedAtomicSMax32(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer{GetSharedPointer(ctx, pointer_offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicSMax(ctx.U32[1], pointer, scope, semantics, value);
+Id EmitSharedAtomicSMax32(EmitContext& ctx, Id offset, Id value) {
+ return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicSMax);
}
-Id EmitSharedAtomicUMax32(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer{GetSharedPointer(ctx, pointer_offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicUMax(ctx.U32[1], pointer, scope, semantics, value);
+Id EmitSharedAtomicUMax32(EmitContext& ctx, Id offset, Id value) {
+ return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicUMax);
}
-Id EmitSharedAtomicInc32(EmitContext& ctx, Id pointer_offset, Id value) {
+Id EmitSharedAtomicInc32(EmitContext& ctx, Id offset, Id value) {
const Id shift_id{ctx.Constant(ctx.U32[1], 2U)};
- const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], pointer_offset, shift_id)};
- return ctx.OpFunctionCall(ctx.U32[1], ctx.increment_cas_shared, index, value,
- ctx.shared_memory_u32);
+ const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)};
+ return ctx.OpFunctionCall(ctx.U32[1], ctx.increment_cas_shared, index, value);
}
-Id EmitSharedAtomicDec32(EmitContext& ctx, Id pointer_offset, Id value) {
+Id EmitSharedAtomicDec32(EmitContext& ctx, Id offset, Id value) {
const Id shift_id{ctx.Constant(ctx.U32[1], 2U)};
- const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], pointer_offset, shift_id)};
- return ctx.OpFunctionCall(ctx.U32[1], ctx.decrement_cas_shared, index, value,
- ctx.shared_memory_u32);
+ const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)};
+ return ctx.OpFunctionCall(ctx.U32[1], ctx.decrement_cas_shared, index, value);
}
-Id EmitSharedAtomicAnd32(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer{GetSharedPointer(ctx, pointer_offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicAnd(ctx.U32[1], pointer, scope, semantics, value);
+Id EmitSharedAtomicAnd32(EmitContext& ctx, Id offset, Id value) {
+ return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicAnd);
}
-Id EmitSharedAtomicOr32(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer{GetSharedPointer(ctx, pointer_offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicOr(ctx.U32[1], pointer, scope, semantics, value);
+Id EmitSharedAtomicOr32(EmitContext& ctx, Id offset, Id value) {
+ return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicOr);
}
-Id EmitSharedAtomicXor32(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer{GetSharedPointer(ctx, pointer_offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicXor(ctx.U32[1], pointer, scope, semantics, value);
+Id EmitSharedAtomicXor32(EmitContext& ctx, Id offset, Id value) {
+ return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicXor);
}
-Id EmitSharedAtomicExchange32(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer{GetSharedPointer(ctx, pointer_offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicExchange(ctx.U32[1], pointer, scope, semantics, value);
+Id EmitSharedAtomicExchange32(EmitContext& ctx, Id offset, Id value) {
+ return SharedAtomicU32(ctx, offset, value, &Sirit::Module::OpAtomicExchange);
}
-Id EmitSharedAtomicExchange64(EmitContext& ctx, Id pointer_offset, Id value) {
- const Id pointer_1{GetSharedPointer(ctx, pointer_offset)};
- if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicExchange(ctx.U64, pointer_1, scope, semantics, value);
+Id EmitSharedAtomicExchange64(EmitContext& ctx, Id offset, Id value) {
+ if (ctx.profile.support_int64_atomics && ctx.profile.support_explicit_workgroup_layout) {
+ const Id shift_id{ctx.Constant(ctx.U32[1], 3U)};
+ const Id index{ctx.OpShiftRightArithmetic(ctx.U32[1], offset, shift_id)};
+ const Id pointer{
+ ctx.OpAccessChain(ctx.shared_u64, ctx.shared_memory_u64, ctx.u32_zero_value, index)};
+ const auto [scope, semantics]{AtomicArgs(ctx)};
+ return ctx.OpAtomicExchange(ctx.U64, pointer, scope, semantics, value);
}
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetSharedPointer(ctx, pointer_offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- StoreResult(ctx, pointer_1, pointer_2, value);
- return original_value;
+ // LOG_WARNING("Int64 Atomics not supported, fallback to non-atomic");
+ const Id pointer_1{SharedPointer(ctx, offset, 0)};
+ const Id pointer_2{SharedPointer(ctx, offset, 1)};
+ const Id value_1{ctx.OpLoad(ctx.U32[1], pointer_1)};
+ const Id value_2{ctx.OpLoad(ctx.U32[1], pointer_2)};
+ const Id new_vector{ctx.OpBitcast(ctx.U32[2], value)};
+ ctx.OpStore(pointer_1, ctx.OpCompositeExtract(ctx.U32[1], new_vector, 0U));
+ ctx.OpStore(pointer_2, ctx.OpCompositeExtract(ctx.U32[1], new_vector, 1U));
+ return ctx.OpBitcast(ctx.U64, ctx.OpCompositeConstruct(ctx.U32[2], value_1, value_2));
}
Id EmitStorageAtomicIAdd32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer{GetStoragePointer(ctx, binding, offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicIAdd(ctx.U32[1], pointer, scope, semantics, value);
+ return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicIAdd);
}
Id EmitStorageAtomicSMin32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer{GetStoragePointer(ctx, binding, offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicSMin(ctx.U32[1], pointer, scope, semantics, value);
+ return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicSMin);
}
Id EmitStorageAtomicUMin32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer{GetStoragePointer(ctx, binding, offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicUMin(ctx.U32[1], pointer, scope, semantics, value);
+ return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicUMin);
}
Id EmitStorageAtomicSMax32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer{GetStoragePointer(ctx, binding, offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicSMax(ctx.U32[1], pointer, scope, semantics, value);
+ return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicSMax);
}
Id EmitStorageAtomicUMax32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer{GetStoragePointer(ctx, binding, offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicUMax(ctx.U32[1], pointer, scope, semantics, value);
+ return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicUMax);
}
Id EmitStorageAtomicInc32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id ssbo{ctx.ssbos[binding.U32()]};
+ const Id ssbo{ctx.ssbos[binding.U32()].U32};
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
return ctx.OpFunctionCall(ctx.U32[1], ctx.increment_cas_ssbo, base_index, value, ssbo);
}
Id EmitStorageAtomicDec32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id ssbo{ctx.ssbos[binding.U32()]};
+ const Id ssbo{ctx.ssbos[binding.U32()].U32};
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
return ctx.OpFunctionCall(ctx.U32[1], ctx.decrement_cas_ssbo, base_index, value, ssbo);
}
Id EmitStorageAtomicAnd32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer{GetStoragePointer(ctx, binding, offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicAnd(ctx.U32[1], pointer, scope, semantics, value);
+ return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicAnd);
}
Id EmitStorageAtomicOr32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer{GetStoragePointer(ctx, binding, offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicOr(ctx.U32[1], pointer, scope, semantics, value);
+ return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicOr);
}
Id EmitStorageAtomicXor32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer{GetStoragePointer(ctx, binding, offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicXor(ctx.U32[1], pointer, scope, semantics, value);
+ return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicXor);
}
Id EmitStorageAtomicExchange32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer{GetStoragePointer(ctx, binding, offset)};
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicExchange(ctx.U32[1], pointer, scope, semantics, value);
+ return StorageAtomicU32(ctx, binding, offset, value, &Sirit::Module::OpAtomicExchange);
}
Id EmitStorageAtomicIAdd64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer_1{GetStoragePointer(ctx, binding, offset)};
- if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicIAdd(ctx.U64, pointer_1, scope, semantics, value);
- }
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- const Id result{ctx.OpIAdd(ctx.U64, value, original_value)};
- StoreResult(ctx, pointer_1, pointer_2, result);
- return original_value;
+ return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicIAdd,
+ &Sirit::Module::OpIAdd);
}
Id EmitStorageAtomicSMin64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer_1{GetStoragePointer(ctx, binding, offset)};
- if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicSMin(ctx.U64, pointer_1, scope, semantics, value);
- }
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- const Id result{ctx.OpSMin(ctx.U64, value, original_value)};
- StoreResult(ctx, pointer_1, pointer_2, result);
- return original_value;
+ return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicSMin,
+ &Sirit::Module::OpSMin);
}
Id EmitStorageAtomicUMin64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer_1{GetStoragePointer(ctx, binding, offset)};
- if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicUMin(ctx.U64, pointer_1, scope, semantics, value);
- }
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- const Id result{ctx.OpUMin(ctx.U64, value, original_value)};
- StoreResult(ctx, pointer_1, pointer_2, result);
- return original_value;
+ return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicUMin,
+ &Sirit::Module::OpUMin);
}
Id EmitStorageAtomicSMax64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer_1{GetStoragePointer(ctx, binding, offset)};
- if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicSMax(ctx.U64, pointer_1, scope, semantics, value);
- }
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- const Id result{ctx.OpSMax(ctx.U64, value, original_value)};
- StoreResult(ctx, pointer_1, pointer_2, result);
- return original_value;
+ return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicSMax,
+ &Sirit::Module::OpSMax);
}
Id EmitStorageAtomicUMax64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer_1{GetStoragePointer(ctx, binding, offset)};
- if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicUMax(ctx.U64, pointer_1, scope, semantics, value);
- }
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- const Id result{ctx.OpUMax(ctx.U64, value, original_value)};
- StoreResult(ctx, pointer_1, pointer_2, result);
- return original_value;
+ return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicUMax,
+ &Sirit::Module::OpUMax);
}
Id EmitStorageAtomicAnd64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer_1{GetStoragePointer(ctx, binding, offset)};
- if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicAnd(ctx.U64, pointer_1, scope, semantics, value);
- }
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- const Id result{ctx.OpBitwiseAnd(ctx.U64, value, original_value)};
- StoreResult(ctx, pointer_1, pointer_2, result);
- return original_value;
+ return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicAnd,
+ &Sirit::Module::OpBitwiseAnd);
}
Id EmitStorageAtomicOr64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer_1{GetStoragePointer(ctx, binding, offset)};
- if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicOr(ctx.U64, pointer_1, scope, semantics, value);
- }
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- const Id result{ctx.OpBitwiseOr(ctx.U64, value, original_value)};
- StoreResult(ctx, pointer_1, pointer_2, result);
- return original_value;
+ return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicOr,
+ &Sirit::Module::OpBitwiseOr);
}
Id EmitStorageAtomicXor64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer_1{GetStoragePointer(ctx, binding, offset)};
- if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicXor(ctx.U64, pointer_1, scope, semantics, value);
- }
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- const Id result{ctx.OpBitwiseXor(ctx.U64, value, original_value)};
- StoreResult(ctx, pointer_1, pointer_2, result);
- return original_value;
+ return StorageAtomicU64(ctx, binding, offset, value, &Sirit::Module::OpAtomicXor,
+ &Sirit::Module::OpBitwiseXor);
}
Id EmitStorageAtomicExchange64(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id pointer_1{GetStoragePointer(ctx, binding, offset)};
if (ctx.profile.support_int64_atomics) {
- const auto [scope, semantics]{GetAtomicArgs(ctx)};
- return ctx.OpAtomicExchange(ctx.U64, pointer_1, scope, semantics, value);
+ const Id pointer{StoragePointer(ctx, ctx.storage_types.U64, &StorageDefinitions::U64,
+ binding, offset, sizeof(u64))};
+ const auto [scope, semantics]{AtomicArgs(ctx)};
+ return ctx.OpAtomicExchange(ctx.U64, pointer, scope, semantics, value);
}
- // LOG_WARNING(Render_Vulkan, "Int64 Atomics not supported, fallback to non-atomic");
- const Id pointer_2{GetStoragePointer(ctx, binding, offset, 1)};
- const Id original_value{LoadU64(ctx, pointer_1, pointer_2)};
- StoreResult(ctx, pointer_1, pointer_2, value);
- return original_value;
+ // LOG_WARNING(..., "Int64 Atomics not supported, fallback to non-atomic");
+ const Id pointer{StoragePointer(ctx, ctx.storage_types.U32x2, &StorageDefinitions::U32x2,
+ binding, offset, sizeof(u32[2]))};
+ const Id original{ctx.OpBitcast(ctx.U64, ctx.OpLoad(ctx.U32[2], pointer))};
+ ctx.OpStore(pointer, value);
+ return original;
}
Id EmitStorageAtomicAddF32(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id ssbo{ctx.ssbos[binding.U32()]};
+ const Id ssbo{ctx.ssbos[binding.U32()].U32};
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
return ctx.OpFunctionCall(ctx.F32[1], ctx.f32_add_cas, base_index, value, ssbo);
}
Id EmitStorageAtomicAddF16x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id ssbo{ctx.ssbos[binding.U32()]};
+ const Id ssbo{ctx.ssbos[binding.U32()].U32};
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
const Id result{ctx.OpFunctionCall(ctx.F16[2], ctx.f16x2_add_cas, base_index, value, ssbo)};
return ctx.OpBitcast(ctx.U32[1], result);
@@ -371,7 +290,7 @@ Id EmitStorageAtomicAddF16x2(EmitContext& ctx, const IR::Value& binding, const I
Id EmitStorageAtomicAddF32x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id ssbo{ctx.ssbos[binding.U32()]};
+ const Id ssbo{ctx.ssbos[binding.U32()].U32};
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
const Id result{ctx.OpFunctionCall(ctx.F32[2], ctx.f32x2_add_cas, base_index, value, ssbo)};
return ctx.OpPackHalf2x16(ctx.U32[1], result);
@@ -379,7 +298,7 @@ Id EmitStorageAtomicAddF32x2(EmitContext& ctx, const IR::Value& binding, const I
Id EmitStorageAtomicMinF16x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id ssbo{ctx.ssbos[binding.U32()]};
+ const Id ssbo{ctx.ssbos[binding.U32()].U32};
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
const Id result{ctx.OpFunctionCall(ctx.F16[2], ctx.f16x2_min_cas, base_index, value, ssbo)};
return ctx.OpBitcast(ctx.U32[1], result);
@@ -387,7 +306,7 @@ Id EmitStorageAtomicMinF16x2(EmitContext& ctx, const IR::Value& binding, const I
Id EmitStorageAtomicMinF32x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id ssbo{ctx.ssbos[binding.U32()]};
+ const Id ssbo{ctx.ssbos[binding.U32()].U32};
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
const Id result{ctx.OpFunctionCall(ctx.F32[2], ctx.f32x2_min_cas, base_index, value, ssbo)};
return ctx.OpPackHalf2x16(ctx.U32[1], result);
@@ -395,7 +314,7 @@ Id EmitStorageAtomicMinF32x2(EmitContext& ctx, const IR::Value& binding, const I
Id EmitStorageAtomicMaxF16x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id ssbo{ctx.ssbos[binding.U32()]};
+ const Id ssbo{ctx.ssbos[binding.U32()].U32};
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
const Id result{ctx.OpFunctionCall(ctx.F16[2], ctx.f16x2_max_cas, base_index, value, ssbo)};
return ctx.OpBitcast(ctx.U32[1], result);
@@ -403,7 +322,7 @@ Id EmitStorageAtomicMaxF16x2(EmitContext& ctx, const IR::Value& binding, const I
Id EmitStorageAtomicMaxF32x2(EmitContext& ctx, const IR::Value& binding, const IR::Value& offset,
Id value) {
- const Id ssbo{ctx.ssbos[binding.U32()]};
+ const Id ssbo{ctx.ssbos[binding.U32()].U32};
const Id base_index{StorageIndex(ctx, offset, sizeof(u32))};
const Id result{ctx.OpFunctionCall(ctx.F32[2], ctx.f32x2_max_cas, base_index, value, ssbo)};
return ctx.OpPackHalf2x16(ctx.U32[1], result);