398 lines
15 KiB
C++
398 lines
15 KiB
C++
// Copyright (c) 2017 Google Inc.
|
|
// Modifications Copyright (C) 2020 Advanced Micro Devices, Inc. All rights
|
|
// reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
// Validates correctness of atomic SPIR-V instructions.
|
|
|
|
#include "source/val/validate.h"
|
|
|
|
#include "source/diagnostic.h"
|
|
#include "source/opcode.h"
|
|
#include "source/spirv_target_env.h"
|
|
#include "source/util/bitutils.h"
|
|
#include "source/val/instruction.h"
|
|
#include "source/val/validate_memory_semantics.h"
|
|
#include "source/val/validate_scopes.h"
|
|
#include "source/val/validation_state.h"
|
|
|
|
namespace {
|
|
|
|
bool IsStorageClassAllowedByUniversalRules(uint32_t storage_class) {
|
|
switch (storage_class) {
|
|
case SpvStorageClassUniform:
|
|
case SpvStorageClassStorageBuffer:
|
|
case SpvStorageClassWorkgroup:
|
|
case SpvStorageClassCrossWorkgroup:
|
|
case SpvStorageClassGeneric:
|
|
case SpvStorageClassAtomicCounter:
|
|
case SpvStorageClassImage:
|
|
case SpvStorageClassFunction:
|
|
case SpvStorageClassPhysicalStorageBuffer:
|
|
return true;
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
bool HasReturnType(uint32_t opcode) {
|
|
switch (opcode) {
|
|
case SpvOpAtomicStore:
|
|
case SpvOpAtomicFlagClear:
|
|
return false;
|
|
break;
|
|
default:
|
|
return true;
|
|
}
|
|
}
|
|
|
|
bool HasOnlyFloatReturnType(uint32_t opcode) {
|
|
switch (opcode) {
|
|
case SpvOpAtomicFAddEXT:
|
|
case SpvOpAtomicFMinEXT:
|
|
case SpvOpAtomicFMaxEXT:
|
|
return true;
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
bool HasOnlyIntReturnType(uint32_t opcode) {
|
|
switch (opcode) {
|
|
case SpvOpAtomicCompareExchange:
|
|
case SpvOpAtomicCompareExchangeWeak:
|
|
case SpvOpAtomicIIncrement:
|
|
case SpvOpAtomicIDecrement:
|
|
case SpvOpAtomicIAdd:
|
|
case SpvOpAtomicISub:
|
|
case SpvOpAtomicSMin:
|
|
case SpvOpAtomicUMin:
|
|
case SpvOpAtomicSMax:
|
|
case SpvOpAtomicUMax:
|
|
case SpvOpAtomicAnd:
|
|
case SpvOpAtomicOr:
|
|
case SpvOpAtomicXor:
|
|
return true;
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
bool HasIntOrFloatReturnType(uint32_t opcode) {
|
|
switch (opcode) {
|
|
case SpvOpAtomicLoad:
|
|
case SpvOpAtomicExchange:
|
|
return true;
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
bool HasOnlyBoolReturnType(uint32_t opcode) {
|
|
switch (opcode) {
|
|
case SpvOpAtomicFlagTestAndSet:
|
|
return true;
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
} // namespace
|
|
|
|
namespace spvtools {
|
|
namespace val {
|
|
|
|
// Validates correctness of atomic instructions.
|
|
spv_result_t AtomicsPass(ValidationState_t& _, const Instruction* inst) {
|
|
const SpvOp opcode = inst->opcode();
|
|
switch (opcode) {
|
|
case SpvOpAtomicLoad:
|
|
case SpvOpAtomicStore:
|
|
case SpvOpAtomicExchange:
|
|
case SpvOpAtomicFAddEXT:
|
|
case SpvOpAtomicCompareExchange:
|
|
case SpvOpAtomicCompareExchangeWeak:
|
|
case SpvOpAtomicIIncrement:
|
|
case SpvOpAtomicIDecrement:
|
|
case SpvOpAtomicIAdd:
|
|
case SpvOpAtomicISub:
|
|
case SpvOpAtomicSMin:
|
|
case SpvOpAtomicUMin:
|
|
case SpvOpAtomicFMinEXT:
|
|
case SpvOpAtomicSMax:
|
|
case SpvOpAtomicUMax:
|
|
case SpvOpAtomicFMaxEXT:
|
|
case SpvOpAtomicAnd:
|
|
case SpvOpAtomicOr:
|
|
case SpvOpAtomicXor:
|
|
case SpvOpAtomicFlagTestAndSet:
|
|
case SpvOpAtomicFlagClear: {
|
|
const uint32_t result_type = inst->type_id();
|
|
|
|
// All current atomics only are scalar result
|
|
// Validate return type first so can just check if pointer type is same
|
|
// (if applicable)
|
|
if (HasReturnType(opcode)) {
|
|
if (HasOnlyFloatReturnType(opcode) &&
|
|
!_.IsFloatScalarType(result_type)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be float scalar type";
|
|
} else if (HasOnlyIntReturnType(opcode) &&
|
|
!_.IsIntScalarType(result_type)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be integer scalar type";
|
|
} else if (HasIntOrFloatReturnType(opcode) &&
|
|
!_.IsFloatScalarType(result_type) &&
|
|
!_.IsIntScalarType(result_type)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be integer or float scalar type";
|
|
} else if (HasOnlyBoolReturnType(opcode) &&
|
|
!_.IsBoolScalarType(result_type)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Result Type to be bool scalar type";
|
|
}
|
|
}
|
|
|
|
uint32_t operand_index = HasReturnType(opcode) ? 2 : 0;
|
|
const uint32_t pointer_type = _.GetOperandTypeId(inst, operand_index++);
|
|
uint32_t data_type = 0;
|
|
uint32_t storage_class = 0;
|
|
if (!_.GetPointerTypeInfo(pointer_type, &data_type, &storage_class)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Pointer to be of type OpTypePointer";
|
|
}
|
|
|
|
// Can't use result_type because OpAtomicStore doesn't have a result
|
|
if ( _.IsIntScalarType(data_type) &&_.GetBitWidth(data_type) == 64 &&
|
|
!_.HasCapability(SpvCapabilityInt64Atomics)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": 64-bit atomics require the Int64Atomics capability";
|
|
}
|
|
|
|
// Validate storage class against universal rules
|
|
if (!IsStorageClassAllowedByUniversalRules(storage_class)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": storage class forbidden by universal validation rules.";
|
|
}
|
|
|
|
// Then Shader rules
|
|
if (_.HasCapability(SpvCapabilityShader)) {
|
|
// Vulkan environment rule
|
|
if (spvIsVulkanEnv(_.context()->target_env)) {
|
|
if ((storage_class != SpvStorageClassUniform) &&
|
|
(storage_class != SpvStorageClassStorageBuffer) &&
|
|
(storage_class != SpvStorageClassWorkgroup) &&
|
|
(storage_class != SpvStorageClassImage) &&
|
|
(storage_class != SpvStorageClassPhysicalStorageBuffer)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< _.VkErrorID(4686) << spvOpcodeString(opcode)
|
|
<< ": Vulkan spec only allows storage classes for atomic to "
|
|
"be: Uniform, Workgroup, Image, StorageBuffer, or "
|
|
"PhysicalStorageBuffer.";
|
|
}
|
|
} else if (storage_class == SpvStorageClassFunction) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": Function storage class forbidden when the Shader "
|
|
"capability is declared.";
|
|
}
|
|
|
|
if (opcode == SpvOpAtomicFAddEXT) {
|
|
// result type being float checked already
|
|
if ((_.GetBitWidth(result_type) == 16) &&
|
|
(!_.HasCapability(SpvCapabilityAtomicFloat16AddEXT))) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": float add atomics require the AtomicFloat32AddEXT "
|
|
"capability";
|
|
}
|
|
if ((_.GetBitWidth(result_type) == 32) &&
|
|
(!_.HasCapability(SpvCapabilityAtomicFloat32AddEXT))) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": float add atomics require the AtomicFloat32AddEXT "
|
|
"capability";
|
|
}
|
|
if ((_.GetBitWidth(result_type) == 64) &&
|
|
(!_.HasCapability(SpvCapabilityAtomicFloat64AddEXT))) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": float add atomics require the AtomicFloat64AddEXT "
|
|
"capability";
|
|
}
|
|
} else if (opcode == SpvOpAtomicFMinEXT ||
|
|
opcode == SpvOpAtomicFMaxEXT) {
|
|
if ((_.GetBitWidth(result_type) == 16) &&
|
|
(!_.HasCapability(SpvCapabilityAtomicFloat16MinMaxEXT))) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": float min/max atomics require the "
|
|
"AtomicFloat16MinMaxEXT capability";
|
|
}
|
|
if ((_.GetBitWidth(result_type) == 32) &&
|
|
(!_.HasCapability(SpvCapabilityAtomicFloat32MinMaxEXT))) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": float min/max atomics require the "
|
|
"AtomicFloat32MinMaxEXT capability";
|
|
}
|
|
if ((_.GetBitWidth(result_type) == 64) &&
|
|
(!_.HasCapability(SpvCapabilityAtomicFloat64MinMaxEXT))) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": float min/max atomics require the "
|
|
"AtomicFloat64MinMaxEXT capability";
|
|
}
|
|
}
|
|
}
|
|
|
|
// And finally OpenCL environment rules
|
|
if (spvIsOpenCLEnv(_.context()->target_env)) {
|
|
if ((storage_class != SpvStorageClassFunction) &&
|
|
(storage_class != SpvStorageClassWorkgroup) &&
|
|
(storage_class != SpvStorageClassCrossWorkgroup) &&
|
|
(storage_class != SpvStorageClassGeneric)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": storage class must be Function, Workgroup, "
|
|
"CrossWorkGroup or Generic in the OpenCL environment.";
|
|
}
|
|
|
|
if (_.context()->target_env == SPV_ENV_OPENCL_1_2) {
|
|
if (storage_class == SpvStorageClassGeneric) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< "Storage class cannot be Generic in OpenCL 1.2 "
|
|
"environment";
|
|
}
|
|
}
|
|
}
|
|
|
|
// If result and pointer type are different, need to do special check here
|
|
if (opcode == SpvOpAtomicFlagTestAndSet ||
|
|
opcode == SpvOpAtomicFlagClear) {
|
|
if (!_.IsIntScalarType(data_type) || _.GetBitWidth(data_type) != 32) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Pointer to point to a value of 32-bit integer "
|
|
"type";
|
|
}
|
|
} else if (opcode == SpvOpAtomicStore) {
|
|
if (!_.IsFloatScalarType(data_type) && !_.IsIntScalarType(data_type)) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Pointer to be a pointer to integer or float "
|
|
<< "scalar type";
|
|
}
|
|
} else if (data_type != result_type) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Pointer to point to a value of type Result "
|
|
"Type";
|
|
}
|
|
|
|
auto memory_scope = inst->GetOperandAs<const uint32_t>(operand_index++);
|
|
if (auto error = ValidateMemoryScope(_, inst, memory_scope)) {
|
|
return error;
|
|
}
|
|
|
|
const auto equal_semantics_index = operand_index++;
|
|
if (auto error = ValidateMemorySemantics(_, inst, equal_semantics_index,
|
|
memory_scope))
|
|
return error;
|
|
|
|
if (opcode == SpvOpAtomicCompareExchange ||
|
|
opcode == SpvOpAtomicCompareExchangeWeak) {
|
|
const auto unequal_semantics_index = operand_index++;
|
|
if (auto error = ValidateMemorySemantics(
|
|
_, inst, unequal_semantics_index, memory_scope))
|
|
return error;
|
|
|
|
// Volatile bits must match for equal and unequal semantics. Previous
|
|
// checks guarantee they are 32-bit constants, but we need to recheck
|
|
// whether they are evaluatable constants.
|
|
bool is_int32 = false;
|
|
bool is_equal_const = false;
|
|
bool is_unequal_const = false;
|
|
uint32_t equal_value = 0;
|
|
uint32_t unequal_value = 0;
|
|
std::tie(is_int32, is_equal_const, equal_value) = _.EvalInt32IfConst(
|
|
inst->GetOperandAs<uint32_t>(equal_semantics_index));
|
|
std::tie(is_int32, is_unequal_const, unequal_value) =
|
|
_.EvalInt32IfConst(
|
|
inst->GetOperandAs<uint32_t>(unequal_semantics_index));
|
|
if (is_equal_const && is_unequal_const &&
|
|
((equal_value & SpvMemorySemanticsVolatileMask) ^
|
|
(unequal_value & SpvMemorySemanticsVolatileMask))) {
|
|
return _.diag(SPV_ERROR_INVALID_ID, inst)
|
|
<< "Volatile mask setting must match for Equal and Unequal "
|
|
"memory semantics";
|
|
}
|
|
}
|
|
|
|
if (opcode == SpvOpAtomicStore) {
|
|
const uint32_t value_type = _.GetOperandTypeId(inst, 3);
|
|
if (value_type != data_type) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Value type and the type pointed to by "
|
|
"Pointer to be the same";
|
|
}
|
|
} else if (opcode != SpvOpAtomicLoad && opcode != SpvOpAtomicIIncrement &&
|
|
opcode != SpvOpAtomicIDecrement &&
|
|
opcode != SpvOpAtomicFlagTestAndSet &&
|
|
opcode != SpvOpAtomicFlagClear) {
|
|
const uint32_t value_type = _.GetOperandTypeId(inst, operand_index++);
|
|
if (value_type != result_type) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Value to be of type Result Type";
|
|
}
|
|
}
|
|
|
|
if (opcode == SpvOpAtomicCompareExchange ||
|
|
opcode == SpvOpAtomicCompareExchangeWeak) {
|
|
const uint32_t comparator_type =
|
|
_.GetOperandTypeId(inst, operand_index++);
|
|
if (comparator_type != result_type) {
|
|
return _.diag(SPV_ERROR_INVALID_DATA, inst)
|
|
<< spvOpcodeString(opcode)
|
|
<< ": expected Comparator to be of type Result Type";
|
|
}
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return SPV_SUCCESS;
|
|
}
|
|
|
|
} // namespace val
|
|
} // namespace spvtools
|