| // Copyright (c) 2017 Google Inc. |
| // |
| // Licensed under the Apache License, Version 2.0 (the "License"); |
| // you may not use this file except in compliance with the License. |
| // You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, software |
| // distributed under the License is distributed on an "AS IS" BASIS, |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| // See the License for the specific language governing permissions and |
| // limitations under the License. |
| |
| // Validates correctness of atomic SPIR-V instructions. |
| |
| #include "source/val/validate.h" |
| |
| #include "source/diagnostic.h" |
| #include "source/opcode.h" |
| #include "source/spirv_target_env.h" |
| #include "source/util/bitutils.h" |
| #include "source/val/instruction.h" |
| #include "source/val/validate_memory_semantics.h" |
| #include "source/val/validate_scopes.h" |
| #include "source/val/validation_state.h" |
| |
| namespace { |
| |
| bool IsStorageClassAllowedByUniversalRules(uint32_t storage_class) { |
| switch (storage_class) { |
| case SpvStorageClassUniform: |
| case SpvStorageClassStorageBuffer: |
| case SpvStorageClassWorkgroup: |
| case SpvStorageClassCrossWorkgroup: |
| case SpvStorageClassGeneric: |
| case SpvStorageClassAtomicCounter: |
| case SpvStorageClassImage: |
| case SpvStorageClassFunction: |
| case SpvStorageClassPhysicalStorageBufferEXT: |
| return true; |
| break; |
| default: |
| return false; |
| } |
| } |
| |
| } // namespace |
| |
| namespace spvtools { |
| namespace val { |
| |
| // Validates correctness of atomic instructions. |
| spv_result_t AtomicsPass(ValidationState_t& _, const Instruction* inst) { |
| const SpvOp opcode = inst->opcode(); |
| const uint32_t result_type = inst->type_id(); |
| bool is_atomic_float_opcode = false; |
| if (opcode == SpvOpAtomicLoad || opcode == SpvOpAtomicStore || |
| opcode == SpvOpAtomicFAddEXT || opcode == SpvOpAtomicExchange) { |
| is_atomic_float_opcode = true; |
| } |
| switch (opcode) { |
| case SpvOpAtomicLoad: |
| case SpvOpAtomicStore: |
| case SpvOpAtomicExchange: |
| case SpvOpAtomicFAddEXT: |
| case SpvOpAtomicCompareExchange: |
| case SpvOpAtomicCompareExchangeWeak: |
| case SpvOpAtomicIIncrement: |
| case SpvOpAtomicIDecrement: |
| case SpvOpAtomicIAdd: |
| case SpvOpAtomicISub: |
| case SpvOpAtomicSMin: |
| case SpvOpAtomicUMin: |
| case SpvOpAtomicSMax: |
| case SpvOpAtomicUMax: |
| case SpvOpAtomicAnd: |
| case SpvOpAtomicOr: |
| case SpvOpAtomicXor: |
| case SpvOpAtomicFlagTestAndSet: |
| case SpvOpAtomicFlagClear: { |
| if (_.HasCapability(SpvCapabilityKernel) && |
| (opcode == SpvOpAtomicLoad || opcode == SpvOpAtomicExchange || |
| opcode == SpvOpAtomicCompareExchange)) { |
| if (!_.IsFloatScalarType(result_type) && |
| !_.IsIntScalarType(result_type)) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Result Type to be int or float scalar type"; |
| } |
| } else if (opcode == SpvOpAtomicFlagTestAndSet) { |
| if (!_.IsBoolScalarType(result_type)) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Result Type to be bool scalar type"; |
| } |
| } else if (opcode == SpvOpAtomicFlagClear || opcode == SpvOpAtomicStore) { |
| assert(result_type == 0); |
| } else { |
| if (_.IsFloatScalarType(result_type)) { |
| if (is_atomic_float_opcode) { |
| if (opcode == SpvOpAtomicFAddEXT) { |
| if ((_.GetBitWidth(result_type) == 32) && |
| (!_.HasCapability(SpvCapabilityAtomicFloat32AddEXT))) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": float add atomics require the AtomicFloat32AddEXT " |
| "capability"; |
| } |
| if ((_.GetBitWidth(result_type) == 64) && |
| (!_.HasCapability(SpvCapabilityAtomicFloat64AddEXT))) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": float add atomics require the AtomicFloat64AddEXT " |
| "capability"; |
| } |
| } |
| } else { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Result Type to be int scalar type"; |
| } |
| } else if (_.IsIntScalarType(result_type) && |
| opcode == SpvOpAtomicFAddEXT) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Result Type to be float scalar type"; |
| } else if (!_.IsFloatScalarType(result_type) && |
| !_.IsIntScalarType(result_type)) { |
| switch (opcode) { |
| case SpvOpAtomicFAddEXT: |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Result Type to be float scalar type"; |
| case SpvOpAtomicIIncrement: |
| case SpvOpAtomicIDecrement: |
| case SpvOpAtomicIAdd: |
| case SpvOpAtomicISub: |
| case SpvOpAtomicSMin: |
| case SpvOpAtomicSMax: |
| case SpvOpAtomicUMin: |
| case SpvOpAtomicUMax: |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Result Type to be integer scalar type"; |
| default: |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Result Type to be int or float scalar type"; |
| } |
| } |
| |
| if (spvIsVulkanEnv(_.context()->target_env) && |
| _.GetBitWidth(result_type) != 32) { |
| switch (opcode) { |
| case SpvOpAtomicSMin: |
| case SpvOpAtomicUMin: |
| case SpvOpAtomicSMax: |
| case SpvOpAtomicUMax: |
| case SpvOpAtomicAnd: |
| case SpvOpAtomicOr: |
| case SpvOpAtomicXor: |
| case SpvOpAtomicIAdd: |
| case SpvOpAtomicISub: |
| case SpvOpAtomicFAddEXT: |
| case SpvOpAtomicLoad: |
| case SpvOpAtomicStore: |
| case SpvOpAtomicExchange: |
| case SpvOpAtomicIIncrement: |
| case SpvOpAtomicIDecrement: |
| case SpvOpAtomicCompareExchangeWeak: |
| case SpvOpAtomicCompareExchange: { |
| if (_.GetBitWidth(result_type) == 64 && |
| _.IsIntScalarType(result_type) && |
| !_.HasCapability(SpvCapabilityInt64Atomics)) |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": 64-bit atomics require the Int64Atomics " |
| "capability"; |
| } break; |
| default: |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": according to the Vulkan spec atomic Result Type " |
| "needs " |
| "to be a 32-bit int scalar type"; |
| } |
| } |
| } |
| |
| uint32_t operand_index = |
| opcode == SpvOpAtomicFlagClear || opcode == SpvOpAtomicStore ? 0 : 2; |
| const uint32_t pointer_type = _.GetOperandTypeId(inst, operand_index++); |
| |
| uint32_t data_type = 0; |
| uint32_t storage_class = 0; |
| if (!_.GetPointerTypeInfo(pointer_type, &data_type, &storage_class)) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Pointer to be of type OpTypePointer"; |
| } |
| |
| // Validate storage class against universal rules |
| if (!IsStorageClassAllowedByUniversalRules(storage_class)) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": storage class forbidden by universal validation rules."; |
| } |
| |
| // Then Shader rules |
| if (_.HasCapability(SpvCapabilityShader)) { |
| if (storage_class == SpvStorageClassFunction) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": Function storage class forbidden when the Shader " |
| "capability is declared."; |
| } |
| } |
| |
| // And finally OpenCL environment rules |
| if (spvIsOpenCLEnv(_.context()->target_env)) { |
| if ((storage_class != SpvStorageClassFunction) && |
| (storage_class != SpvStorageClassWorkgroup) && |
| (storage_class != SpvStorageClassCrossWorkgroup) && |
| (storage_class != SpvStorageClassGeneric)) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": storage class must be Function, Workgroup, " |
| "CrossWorkGroup or Generic in the OpenCL environment."; |
| } |
| |
| if (_.context()->target_env == SPV_ENV_OPENCL_1_2) { |
| if (storage_class == SpvStorageClassGeneric) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << "Storage class cannot be Generic in OpenCL 1.2 " |
| "environment"; |
| } |
| } |
| } |
| |
| if (opcode == SpvOpAtomicFlagTestAndSet || |
| opcode == SpvOpAtomicFlagClear) { |
| if (!_.IsIntScalarType(data_type) || _.GetBitWidth(data_type) != 32) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Pointer to point to a value of 32-bit int type"; |
| } |
| } else if (opcode == SpvOpAtomicStore) { |
| if (!_.IsFloatScalarType(data_type) && !_.IsIntScalarType(data_type)) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Pointer to be a pointer to int or float " |
| << "scalar type"; |
| } |
| } else { |
| if (data_type != result_type) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Pointer to point to a value of type Result " |
| "Type"; |
| } |
| } |
| |
| auto memory_scope = inst->GetOperandAs<const uint32_t>(operand_index++); |
| if (auto error = ValidateMemoryScope(_, inst, memory_scope)) { |
| return error; |
| } |
| |
| const auto equal_semantics_index = operand_index++; |
| if (auto error = ValidateMemorySemantics(_, inst, equal_semantics_index)) |
| return error; |
| |
| if (opcode == SpvOpAtomicCompareExchange || |
| opcode == SpvOpAtomicCompareExchangeWeak) { |
| const auto unequal_semantics_index = operand_index++; |
| if (auto error = |
| ValidateMemorySemantics(_, inst, unequal_semantics_index)) |
| return error; |
| |
| // Volatile bits must match for equal and unequal semantics. Previous |
| // checks guarantee they are 32-bit constants, but we need to recheck |
| // whether they are evaluatable constants. |
| bool is_int32 = false; |
| bool is_equal_const = false; |
| bool is_unequal_const = false; |
| uint32_t equal_value = 0; |
| uint32_t unequal_value = 0; |
| std::tie(is_int32, is_equal_const, equal_value) = _.EvalInt32IfConst( |
| inst->GetOperandAs<uint32_t>(equal_semantics_index)); |
| std::tie(is_int32, is_unequal_const, unequal_value) = |
| _.EvalInt32IfConst( |
| inst->GetOperandAs<uint32_t>(unequal_semantics_index)); |
| if (is_equal_const && is_unequal_const && |
| ((equal_value & SpvMemorySemanticsVolatileMask) ^ |
| (unequal_value & SpvMemorySemanticsVolatileMask))) { |
| return _.diag(SPV_ERROR_INVALID_ID, inst) |
| << "Volatile mask setting must match for Equal and Unequal " |
| "memory semantics"; |
| } |
| } |
| |
| if (opcode == SpvOpAtomicStore) { |
| const uint32_t value_type = _.GetOperandTypeId(inst, 3); |
| if (value_type != data_type) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Value type and the type pointed to by " |
| "Pointer to be the same"; |
| } |
| } else if (opcode != SpvOpAtomicLoad && opcode != SpvOpAtomicIIncrement && |
| opcode != SpvOpAtomicIDecrement && |
| opcode != SpvOpAtomicFlagTestAndSet && |
| opcode != SpvOpAtomicFlagClear) { |
| const uint32_t value_type = _.GetOperandTypeId(inst, operand_index++); |
| if (value_type != result_type) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Value to be of type Result Type"; |
| } |
| } |
| |
| if (opcode == SpvOpAtomicCompareExchange || |
| opcode == SpvOpAtomicCompareExchangeWeak) { |
| const uint32_t comparator_type = |
| _.GetOperandTypeId(inst, operand_index++); |
| if (comparator_type != result_type) { |
| return _.diag(SPV_ERROR_INVALID_DATA, inst) |
| << spvOpcodeString(opcode) |
| << ": expected Comparator to be of type Result Type"; |
| } |
| } |
| |
| break; |
| } |
| |
| default: |
| break; |
| } |
| |
| return SPV_SUCCESS; |
| } |
| |
| } // namespace val |
| } // namespace spvtools |