bgfx/3rdparty/spirv-cross/spirv_msl.hpp

942 lines
40 KiB
C++
Raw Normal View History

2019-01-10 06:23:47 +03:00
/*
2020-02-05 08:38:12 +03:00
* Copyright 2016-2020 The Brenwill Workshop Ltd.
2019-01-10 06:23:47 +03:00
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef SPIRV_CROSS_MSL_HPP
#define SPIRV_CROSS_MSL_HPP
#include "spirv_glsl.hpp"
#include <map>
#include <set>
2019-06-22 17:30:47 +03:00
#include <stddef.h>
2019-01-10 06:23:47 +03:00
#include <unordered_map>
#include <unordered_set>
2019-03-31 07:03:28 +03:00
namespace SPIRV_CROSS_NAMESPACE
2019-01-10 06:23:47 +03:00
{
// Indicates the format of the vertex attribute. Currently limited to specifying
// if the attribute is an 8-bit unsigned integer, 16-bit unsigned integer, or
// some other format.
enum MSLVertexFormat
{
2019-03-09 21:20:48 +03:00
MSL_VERTEX_FORMAT_OTHER = 0,
MSL_VERTEX_FORMAT_UINT8 = 1,
MSL_VERTEX_FORMAT_UINT16 = 2,
MSL_VERTEX_FORMAT_INT_MAX = 0x7fffffff
2019-01-10 06:23:47 +03:00
};
// Defines MSL characteristics of a vertex attribute at a particular location.
2019-03-09 21:20:48 +03:00
// After compilation, it is possible to query whether or not this location was used.
2019-01-10 06:23:47 +03:00
struct MSLVertexAttr
{
uint32_t location = 0;
uint32_t msl_buffer = 0;
uint32_t msl_offset = 0;
uint32_t msl_stride = 0;
bool per_instance = false;
MSLVertexFormat format = MSL_VERTEX_FORMAT_OTHER;
2019-02-23 07:54:12 +03:00
spv::BuiltIn builtin = spv::BuiltInMax;
2019-01-10 06:23:47 +03:00
};
// Matches the binding index of a MSL resource for a binding within a descriptor set.
// Taken together, the stage, desc_set and binding combine to form a reference to a resource
2019-03-09 21:20:48 +03:00
// descriptor used in a particular shading stage.
2019-09-07 18:07:48 +03:00
// If using MSL 2.0 argument buffers, the descriptor set is not marked as a discrete descriptor set,
// and (for iOS only) the resource is not a storage image (sampled != 2), the binding reference we
// remap to will become an [[id(N)]] attribute within the "descriptor set" argument buffer structure.
2019-03-17 06:30:56 +03:00
// For resources which are bound in the "classic" MSL 1.0 way or discrete descriptors, the remap will become a
// [[buffer(N)]], [[texture(N)]] or [[sampler(N)]] depending on the resource types used.
2019-01-10 06:23:47 +03:00
struct MSLResourceBinding
{
2019-03-09 21:20:48 +03:00
spv::ExecutionModel stage = spv::ExecutionModelMax;
2019-01-10 06:23:47 +03:00
uint32_t desc_set = 0;
uint32_t binding = 0;
uint32_t msl_buffer = 0;
uint32_t msl_texture = 0;
uint32_t msl_sampler = 0;
};
enum MSLSamplerCoord
{
2019-03-09 21:20:48 +03:00
MSL_SAMPLER_COORD_NORMALIZED = 0,
MSL_SAMPLER_COORD_PIXEL = 1,
MSL_SAMPLER_INT_MAX = 0x7fffffff
2019-01-10 06:23:47 +03:00
};
enum MSLSamplerFilter
{
2019-03-09 21:20:48 +03:00
MSL_SAMPLER_FILTER_NEAREST = 0,
MSL_SAMPLER_FILTER_LINEAR = 1,
MSL_SAMPLER_FILTER_INT_MAX = 0x7fffffff
2019-01-10 06:23:47 +03:00
};
enum MSLSamplerMipFilter
{
2019-03-09 21:20:48 +03:00
MSL_SAMPLER_MIP_FILTER_NONE = 0,
MSL_SAMPLER_MIP_FILTER_NEAREST = 1,
MSL_SAMPLER_MIP_FILTER_LINEAR = 2,
MSL_SAMPLER_MIP_FILTER_INT_MAX = 0x7fffffff
2019-01-10 06:23:47 +03:00
};
enum MSLSamplerAddress
{
2019-03-09 21:20:48 +03:00
MSL_SAMPLER_ADDRESS_CLAMP_TO_ZERO = 0,
MSL_SAMPLER_ADDRESS_CLAMP_TO_EDGE = 1,
MSL_SAMPLER_ADDRESS_CLAMP_TO_BORDER = 2,
MSL_SAMPLER_ADDRESS_REPEAT = 3,
MSL_SAMPLER_ADDRESS_MIRRORED_REPEAT = 4,
MSL_SAMPLER_ADDRESS_INT_MAX = 0x7fffffff
2019-01-10 06:23:47 +03:00
};
enum MSLSamplerCompareFunc
{
2019-03-09 21:20:48 +03:00
MSL_SAMPLER_COMPARE_FUNC_NEVER = 0,
MSL_SAMPLER_COMPARE_FUNC_LESS = 1,
MSL_SAMPLER_COMPARE_FUNC_LESS_EQUAL = 2,
MSL_SAMPLER_COMPARE_FUNC_GREATER = 3,
MSL_SAMPLER_COMPARE_FUNC_GREATER_EQUAL = 4,
MSL_SAMPLER_COMPARE_FUNC_EQUAL = 5,
MSL_SAMPLER_COMPARE_FUNC_NOT_EQUAL = 6,
MSL_SAMPLER_COMPARE_FUNC_ALWAYS = 7,
MSL_SAMPLER_COMPARE_FUNC_INT_MAX = 0x7fffffff
2019-01-10 06:23:47 +03:00
};
enum MSLSamplerBorderColor
{
2019-03-09 21:20:48 +03:00
MSL_SAMPLER_BORDER_COLOR_TRANSPARENT_BLACK = 0,
MSL_SAMPLER_BORDER_COLOR_OPAQUE_BLACK = 1,
MSL_SAMPLER_BORDER_COLOR_OPAQUE_WHITE = 2,
MSL_SAMPLER_BORDER_COLOR_INT_MAX = 0x7fffffff
2019-01-10 06:23:47 +03:00
};
2019-09-07 18:07:48 +03:00
enum MSLFormatResolution
{
MSL_FORMAT_RESOLUTION_444 = 0,
MSL_FORMAT_RESOLUTION_422,
MSL_FORMAT_RESOLUTION_420,
MSL_FORMAT_RESOLUTION_INT_MAX = 0x7fffffff
};
enum MSLChromaLocation
{
MSL_CHROMA_LOCATION_COSITED_EVEN = 0,
MSL_CHROMA_LOCATION_MIDPOINT,
MSL_CHROMA_LOCATION_INT_MAX = 0x7fffffff
};
enum MSLComponentSwizzle
{
MSL_COMPONENT_SWIZZLE_IDENTITY = 0,
MSL_COMPONENT_SWIZZLE_ZERO,
MSL_COMPONENT_SWIZZLE_ONE,
MSL_COMPONENT_SWIZZLE_R,
MSL_COMPONENT_SWIZZLE_G,
MSL_COMPONENT_SWIZZLE_B,
MSL_COMPONENT_SWIZZLE_A,
MSL_COMPONENT_SWIZZLE_INT_MAX = 0x7fffffff
};
enum MSLSamplerYCbCrModelConversion
{
MSL_SAMPLER_YCBCR_MODEL_CONVERSION_RGB_IDENTITY = 0,
MSL_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_IDENTITY,
MSL_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_BT_709,
MSL_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_BT_601,
MSL_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_BT_2020,
MSL_SAMPLER_YCBCR_MODEL_CONVERSION_INT_MAX = 0x7fffffff
};
enum MSLSamplerYCbCrRange
{
MSL_SAMPLER_YCBCR_RANGE_ITU_FULL = 0,
MSL_SAMPLER_YCBCR_RANGE_ITU_NARROW,
MSL_SAMPLER_YCBCR_RANGE_INT_MAX = 0x7fffffff
};
2019-01-10 06:23:47 +03:00
struct MSLConstexprSampler
{
MSLSamplerCoord coord = MSL_SAMPLER_COORD_NORMALIZED;
MSLSamplerFilter min_filter = MSL_SAMPLER_FILTER_NEAREST;
MSLSamplerFilter mag_filter = MSL_SAMPLER_FILTER_NEAREST;
MSLSamplerMipFilter mip_filter = MSL_SAMPLER_MIP_FILTER_NONE;
MSLSamplerAddress s_address = MSL_SAMPLER_ADDRESS_CLAMP_TO_EDGE;
MSLSamplerAddress t_address = MSL_SAMPLER_ADDRESS_CLAMP_TO_EDGE;
MSLSamplerAddress r_address = MSL_SAMPLER_ADDRESS_CLAMP_TO_EDGE;
MSLSamplerCompareFunc compare_func = MSL_SAMPLER_COMPARE_FUNC_NEVER;
MSLSamplerBorderColor border_color = MSL_SAMPLER_BORDER_COLOR_TRANSPARENT_BLACK;
float lod_clamp_min = 0.0f;
float lod_clamp_max = 1000.0f;
int max_anisotropy = 1;
2019-09-07 18:07:48 +03:00
// Sampler Y'CbCr conversion parameters
uint32_t planes = 0;
MSLFormatResolution resolution = MSL_FORMAT_RESOLUTION_444;
MSLSamplerFilter chroma_filter = MSL_SAMPLER_FILTER_NEAREST;
MSLChromaLocation x_chroma_offset = MSL_CHROMA_LOCATION_COSITED_EVEN;
MSLChromaLocation y_chroma_offset = MSL_CHROMA_LOCATION_COSITED_EVEN;
MSLComponentSwizzle swizzle[4]; // IDENTITY, IDENTITY, IDENTITY, IDENTITY
MSLSamplerYCbCrModelConversion ycbcr_model = MSL_SAMPLER_YCBCR_MODEL_CONVERSION_RGB_IDENTITY;
MSLSamplerYCbCrRange ycbcr_range = MSL_SAMPLER_YCBCR_RANGE_ITU_FULL;
uint32_t bpc = 8;
2019-01-10 06:23:47 +03:00
bool compare_enable = false;
bool lod_clamp_enable = false;
bool anisotropy_enable = false;
2019-09-07 18:07:48 +03:00
bool ycbcr_conversion_enable = false;
MSLConstexprSampler()
{
for (uint32_t i = 0; i < 4; i++)
swizzle[i] = MSL_COMPONENT_SWIZZLE_IDENTITY;
}
bool swizzle_is_identity() const
{
return (swizzle[0] == MSL_COMPONENT_SWIZZLE_IDENTITY && swizzle[1] == MSL_COMPONENT_SWIZZLE_IDENTITY &&
swizzle[2] == MSL_COMPONENT_SWIZZLE_IDENTITY && swizzle[3] == MSL_COMPONENT_SWIZZLE_IDENTITY);
}
bool swizzle_has_one_or_zero() const
{
return (swizzle[0] == MSL_COMPONENT_SWIZZLE_ZERO || swizzle[0] == MSL_COMPONENT_SWIZZLE_ONE ||
swizzle[1] == MSL_COMPONENT_SWIZZLE_ZERO || swizzle[1] == MSL_COMPONENT_SWIZZLE_ONE ||
swizzle[2] == MSL_COMPONENT_SWIZZLE_ZERO || swizzle[2] == MSL_COMPONENT_SWIZZLE_ONE ||
swizzle[3] == MSL_COMPONENT_SWIZZLE_ZERO || swizzle[3] == MSL_COMPONENT_SWIZZLE_ONE);
}
2019-01-10 06:23:47 +03:00
};
// Special constant used in a MSLResourceBinding desc_set
// element to indicate the bindings for the push constants.
2020-02-05 08:38:12 +03:00
// Kinda deprecated. Just use ResourceBindingPushConstant{DescriptorSet,Binding} directly.
static const uint32_t kPushConstDescSet = ResourceBindingPushConstantDescriptorSet;
2019-01-10 06:23:47 +03:00
// Special constant used in a MSLResourceBinding binding
// element to indicate the bindings for the push constants.
2020-02-05 08:38:12 +03:00
// Kinda deprecated. Just use ResourceBindingPushConstant{DescriptorSet,Binding} directly.
static const uint32_t kPushConstBinding = ResourceBindingPushConstantBinding;
2019-01-10 06:23:47 +03:00
2019-05-18 22:28:43 +03:00
// Special constant used in a MSLResourceBinding binding
// element to indicate the buffer binding for swizzle buffers.
static const uint32_t kSwizzleBufferBinding = ~(1u);
2019-03-17 06:30:56 +03:00
2019-06-07 08:04:16 +03:00
// Special constant used in a MSLResourceBinding binding
// element to indicate the buffer binding for buffer size buffers to support OpArrayLength.
static const uint32_t kBufferSizeBufferBinding = ~(2u);
2019-07-03 18:10:12 +03:00
// Special constant used in a MSLResourceBinding binding
// element to indicate the buffer binding used for the argument buffer itself.
// This buffer binding should be kept as small as possible as all automatic bindings for buffers
// will start at max(kArgumentBufferBinding) + 1.
static const uint32_t kArgumentBufferBinding = ~(3u);
2019-05-18 22:28:43 +03:00
static const uint32_t kMaxArgumentBuffers = 8;
2019-02-09 23:44:48 +03:00
2019-01-10 06:23:47 +03:00
// Decompiles SPIR-V to Metal Shading Language
class CompilerMSL : public CompilerGLSL
{
public:
// Options for compiling to Metal Shading Language
struct Options
{
typedef enum
{
2019-03-09 21:20:48 +03:00
iOS = 0,
macOS = 1
2019-01-10 06:23:47 +03:00
} Platform;
Platform platform = macOS;
uint32_t msl_version = make_msl_version(1, 2);
uint32_t texel_buffer_texture_width = 4096; // Width of 2D Metal textures used as 1D texel buffers
2019-05-18 22:28:43 +03:00
uint32_t swizzle_buffer_index = 30;
2019-02-09 23:44:48 +03:00
uint32_t indirect_params_buffer_index = 29;
uint32_t shader_output_buffer_index = 28;
2019-02-16 06:52:34 +03:00
uint32_t shader_patch_output_buffer_index = 27;
uint32_t shader_tess_factor_buffer_index = 26;
2019-06-07 08:04:16 +03:00
uint32_t buffer_size_buffer_index = 25;
2019-07-03 18:10:12 +03:00
uint32_t view_mask_buffer_index = 24;
2019-09-07 18:07:48 +03:00
uint32_t dynamic_offsets_buffer_index = 23;
2019-02-16 06:52:34 +03:00
uint32_t shader_input_wg_index = 0;
2019-07-20 20:45:34 +03:00
uint32_t device_index = 0;
2019-01-10 06:23:47 +03:00
bool enable_point_size_builtin = true;
bool disable_rasterization = false;
2019-02-09 23:44:48 +03:00
bool capture_output_to_buffer = false;
2019-01-10 06:23:47 +03:00
bool swizzle_texture_samples = false;
2019-02-23 07:54:12 +03:00
bool tess_domain_origin_lower_left = false;
2019-07-03 18:10:12 +03:00
bool multiview = false;
2019-07-20 20:45:34 +03:00
bool view_index_from_device_index = false;
2019-08-10 06:33:38 +03:00
bool dispatch_base = false;
2019-10-26 18:03:21 +03:00
bool texture_1D_as_2D = false;
2019-01-10 06:23:47 +03:00
2019-03-17 06:30:56 +03:00
// Enable use of MSL 2.0 indirect argument buffers.
// MSL 2.0 must also be enabled.
bool argument_buffers = false;
2019-10-29 05:55:41 +03:00
// Ensures vertex and instance indices start at zero. This reflects the behavior of HLSL with SV_VertexID and SV_InstanceID.
bool enable_base_index_zero = false;
2019-01-17 05:50:01 +03:00
// Fragment output in MSL must have at least as many components as the render pass.
// Add support to explicit pad out components.
bool pad_fragment_output_components = false;
2019-10-29 05:55:41 +03:00
// Specifies whether the iOS target version supports the [[base_vertex]] and [[base_instance]] attributes.
bool ios_support_base_vertex_instance = false;
// Use Metal's native frame-buffer fetch API for subpass inputs.
bool ios_use_framebuffer_fetch_subpasses = false;
// Enables use of "fma" intrinsic for invariant float math
bool invariant_float_math = false;
// Emulate texturecube_array with texture2d_array for iOS where this type is not available
bool emulate_cube_array = false;
// Allow user to enable decoration binding
bool enable_decoration_binding = false;
2019-04-28 00:14:44 +03:00
// Requires MSL 2.1, use the native support for texel buffers.
bool texture_buffer_native = false;
2020-02-05 08:38:12 +03:00
// Forces all resources which are part of an argument buffer to be considered active.
// This ensures ABI compatibility between shaders where some resources might be unused,
// and would otherwise declare a different IAB.
bool force_active_argument_buffer_resources = false;
2019-01-10 06:23:47 +03:00
bool is_ios()
{
return platform == iOS;
}
bool is_macos()
{
return platform == macOS;
}
void set_msl_version(uint32_t major, uint32_t minor = 0, uint32_t patch = 0)
{
msl_version = make_msl_version(major, minor, patch);
}
2019-08-10 06:33:38 +03:00
bool supports_msl_version(uint32_t major, uint32_t minor = 0, uint32_t patch = 0) const
2019-01-10 06:23:47 +03:00
{
return msl_version >= make_msl_version(major, minor, patch);
}
static uint32_t make_msl_version(uint32_t major, uint32_t minor = 0, uint32_t patch = 0)
{
return (major * 10000) + (minor * 100) + patch;
}
};
const Options &get_msl_options() const
{
return msl_options;
}
void set_msl_options(const Options &opts)
{
msl_options = opts;
}
// Provide feedback to calling API to allow runtime to disable pipeline
// rasterization if vertex shader requires rasterization to be disabled.
bool get_is_rasterization_disabled() const
{
2019-02-16 06:52:34 +03:00
return is_rasterization_disabled && (get_entry_point().model == spv::ExecutionModelVertex ||
2019-02-23 07:54:12 +03:00
get_entry_point().model == spv::ExecutionModelTessellationControl ||
get_entry_point().model == spv::ExecutionModelTessellationEvaluation);
2019-01-10 06:23:47 +03:00
}
// Provide feedback to calling API to allow it to pass an auxiliary
2019-05-18 22:28:43 +03:00
// swizzle buffer if the shader needs it.
bool needs_swizzle_buffer() const
2019-01-10 06:23:47 +03:00
{
2019-05-18 22:28:43 +03:00
return used_swizzle_buffer;
2019-01-10 06:23:47 +03:00
}
2019-06-07 08:04:16 +03:00
// Provide feedback to calling API to allow it to pass a buffer
// containing STORAGE_BUFFER buffer sizes to support OpArrayLength.
bool needs_buffer_size_buffer() const
{
return !buffers_requiring_array_length.empty();
}
2019-07-03 18:10:12 +03:00
// Provide feedback to calling API to allow it to pass a buffer
// containing the view mask for the current multiview subpass.
bool needs_view_mask_buffer() const
{
2019-07-20 20:45:34 +03:00
return msl_options.multiview && !msl_options.view_index_from_device_index;
2019-07-03 18:10:12 +03:00
}
2019-08-10 06:33:38 +03:00
// Provide feedback to calling API to allow it to pass a buffer
// containing the dispatch base workgroup ID.
bool needs_dispatch_base_buffer() const
{
return msl_options.dispatch_base && !msl_options.supports_msl_version(1, 2);
}
2019-02-09 23:44:48 +03:00
// Provide feedback to calling API to allow it to pass an output
// buffer if the shader needs it.
bool needs_output_buffer() const
{
2019-09-07 18:07:48 +03:00
return capture_output_to_buffer && stage_out_var_id != ID(0);
2019-02-09 23:44:48 +03:00
}
2019-02-16 06:52:34 +03:00
// Provide feedback to calling API to allow it to pass a patch output
// buffer if the shader needs it.
bool needs_patch_output_buffer() const
{
2019-09-07 18:07:48 +03:00
return capture_output_to_buffer && patch_stage_out_var_id != ID(0);
2019-02-16 06:52:34 +03:00
}
// Provide feedback to calling API to allow it to pass an input threadgroup
// buffer if the shader needs it.
bool needs_input_threadgroup_mem() const
{
2019-09-07 18:07:48 +03:00
return capture_output_to_buffer && stage_in_var_id != ID(0);
2019-02-16 06:52:34 +03:00
}
2019-03-09 21:20:48 +03:00
explicit CompilerMSL(std::vector<uint32_t> spirv);
CompilerMSL(const uint32_t *ir, size_t word_count);
explicit CompilerMSL(const ParsedIR &ir);
explicit CompilerMSL(ParsedIR &&ir);
// attr is a vertex attribute binding used to match
// vertex content locations to MSL attributes. If vertex attributes are provided,
// is_msl_vertex_attribute_used() will return true after calling ::compile() if
// the location was used by the MSL code.
void add_msl_vertex_attribute(const MSLVertexAttr &attr);
// resource is a resource binding to indicate the MSL buffer,
// texture or sampler index to use for a particular SPIR-V description set
// and binding. If resource bindings are provided,
// is_msl_resource_binding_used() will return true after calling ::compile() if
// the set/binding combination was used by the MSL code.
void add_msl_resource_binding(const MSLResourceBinding &resource);
2019-09-07 18:07:48 +03:00
// desc_set and binding are the SPIR-V descriptor set and binding of a buffer resource
// in this shader. index is the index within the dynamic offset buffer to use. This
// function marks that resource as using a dynamic offset (VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC
// or VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC). This function only has any effect if argument buffers
// are enabled. If so, the buffer will have its address adjusted at the beginning of the shader with
// an offset taken from the dynamic offset buffer.
void add_dynamic_buffer(uint32_t desc_set, uint32_t binding, uint32_t index);
2020-02-05 08:38:12 +03:00
// desc_set and binding are the SPIR-V descriptor set and binding of a buffer resource
// in this shader. This function marks that resource as an inline uniform block
// (VK_DESCRIPTOR_TYPE_INLINE_UNIFORM_BLOCK_EXT). This function only has any effect if argument buffers
// are enabled. If so, the buffer block will be directly embedded into the argument
// buffer, instead of being referenced indirectly via pointer.
void add_inline_uniform_block(uint32_t desc_set, uint32_t binding);
2019-03-17 06:30:56 +03:00
// When using MSL argument buffers, we can force "classic" MSL 1.0 binding schemes for certain descriptor sets.
// This corresponds to VK_KHR_push_descriptor in Vulkan.
void add_discrete_descriptor_set(uint32_t desc_set);
2019-10-20 08:33:06 +03:00
// If an argument buffer is large enough, it may need to be in the device storage space rather than
// constant. Opt-in to this behavior here on a per set basis.
void set_argument_buffer_device_address_space(uint32_t desc_set, bool device_storage);
2019-03-09 21:20:48 +03:00
// Query after compilation is done. This allows you to check if a location or set/binding combination was used by the shader.
bool is_msl_vertex_attribute_used(uint32_t location);
2019-06-17 03:47:55 +03:00
// NOTE: Only resources which are remapped using add_msl_resource_binding will be reported here.
// Constexpr samplers are always assumed to be emitted.
// No specific MSLResourceBinding remapping is required for constexpr samplers as long as they are remapped
// by remap_constexpr_sampler(_by_binding).
2020-02-05 08:38:12 +03:00
bool is_msl_resource_binding_used(spv::ExecutionModel model, uint32_t set, uint32_t binding) const;
2019-03-09 21:20:48 +03:00
2019-06-22 17:30:47 +03:00
// This must only be called after a successful call to CompilerMSL::compile().
// For a variable resource ID obtained through reflection API, report the automatically assigned resource index.
// If the descriptor set was part of an argument buffer, report the [[id(N)]],
// or [[buffer/texture/sampler]] binding for other resources.
// If the resource was a combined image sampler, report the image binding here,
// use the _secondary version of this call to query the sampler half of the resource.
// If no binding exists, uint32_t(-1) is returned.
uint32_t get_automatic_msl_resource_binding(uint32_t id) const;
// Same as get_automatic_msl_resource_binding, but should only be used for combined image samplers, in which case the
// sampler's binding is returned instead. For any other resource type, -1 is returned.
uint32_t get_automatic_msl_resource_binding_secondary(uint32_t id) const;
2019-09-07 18:07:48 +03:00
// Same as get_automatic_msl_resource_binding, but should only be used for combined image samplers for multiplanar images,
// in which case the second plane's binding is returned instead. For any other resource type, -1 is returned.
uint32_t get_automatic_msl_resource_binding_tertiary(uint32_t id) const;
// Same as get_automatic_msl_resource_binding, but should only be used for combined image samplers for triplanar images,
// in which case the third plane's binding is returned instead. For any other resource type, -1 is returned.
uint32_t get_automatic_msl_resource_binding_quaternary(uint32_t id) const;
2019-03-09 21:20:48 +03:00
// Compiles the SPIR-V code into Metal Shading Language.
std::string compile() override;
// Remap a sampler with ID to a constexpr sampler.
// Older iOS targets must use constexpr samplers in certain cases (PCF),
// so a static sampler must be used.
// The sampler will not consume a binding, but be declared in the entry point as a constexpr sampler.
// This can be used on both combined image/samplers (sampler2D) or standalone samplers.
// The remapped sampler must not be an array of samplers.
2019-06-17 03:47:55 +03:00
// Prefer remap_constexpr_sampler_by_binding unless you're also doing reflection anyways.
2019-09-07 18:07:48 +03:00
void remap_constexpr_sampler(VariableID id, const MSLConstexprSampler &sampler);
2019-03-09 21:20:48 +03:00
2019-06-17 03:47:55 +03:00
// Same as remap_constexpr_sampler, except you provide set/binding, rather than variable ID.
// Remaps based on ID take priority over set/binding remaps.
void remap_constexpr_sampler_by_binding(uint32_t desc_set, uint32_t binding, const MSLConstexprSampler &sampler);
2019-03-09 21:20:48 +03:00
// If using CompilerMSL::Options::pad_fragment_output_components, override the number of components we expect
// to use for a particular location. The default is 4 if number of components is not overridden.
void set_fragment_output_components(uint32_t location, uint32_t components);
protected:
2019-01-10 06:23:47 +03:00
// An enum of SPIR-V functions that are implemented in additional
// source code that is added to the shader if necessary.
enum SPVFuncImpl
{
SPVFuncImplNone,
SPVFuncImplMod,
SPVFuncImplRadians,
SPVFuncImplDegrees,
SPVFuncImplFindILsb,
SPVFuncImplFindSMsb,
SPVFuncImplFindUMsb,
SPVFuncImplSSign,
SPVFuncImplArrayCopyMultidimBase,
// Unfortunately, we cannot use recursive templates in the MSL compiler properly,
// so stamp out variants up to some arbitrary maximum.
SPVFuncImplArrayCopy = SPVFuncImplArrayCopyMultidimBase + 1,
SPVFuncImplArrayOfArrayCopy2Dim = SPVFuncImplArrayCopyMultidimBase + 2,
SPVFuncImplArrayOfArrayCopy3Dim = SPVFuncImplArrayCopyMultidimBase + 3,
SPVFuncImplArrayOfArrayCopy4Dim = SPVFuncImplArrayCopyMultidimBase + 4,
SPVFuncImplArrayOfArrayCopy5Dim = SPVFuncImplArrayCopyMultidimBase + 5,
SPVFuncImplArrayOfArrayCopy6Dim = SPVFuncImplArrayCopyMultidimBase + 6,
SPVFuncImplTexelBufferCoords,
2019-10-29 05:55:41 +03:00
SPVFuncImplImage2DAtomicCoords, // Emulate texture2D atomic operations
SPVFuncImplFMul,
SPVFuncImplFAdd,
SPVFuncImplCubemapTo2DArrayFace,
SPVFuncImplUnsafeArray, // Allow Metal to use the array<T> template to make arrays a value type
2019-01-10 06:23:47 +03:00
SPVFuncImplInverse4x4,
SPVFuncImplInverse3x3,
SPVFuncImplInverse2x2,
2019-09-07 18:07:48 +03:00
// It is very important that this come before *Swizzle and ChromaReconstruct*, to ensure it's
// emitted before them.
SPVFuncImplForwardArgs,
// Likewise, this must come before *Swizzle.
SPVFuncImplGetSwizzle,
2019-01-10 06:23:47 +03:00
SPVFuncImplTextureSwizzle,
2019-09-07 18:07:48 +03:00
SPVFuncImplGatherSwizzle,
SPVFuncImplGatherCompareSwizzle,
2019-05-18 22:28:43 +03:00
SPVFuncImplSubgroupBallot,
SPVFuncImplSubgroupBallotBitExtract,
SPVFuncImplSubgroupBallotFindLSB,
SPVFuncImplSubgroupBallotFindMSB,
SPVFuncImplSubgroupBallotBitCount,
SPVFuncImplSubgroupAllEqual,
2019-07-03 18:10:12 +03:00
SPVFuncImplReflectScalar,
SPVFuncImplRefractScalar,
2019-07-20 20:45:34 +03:00
SPVFuncImplFaceForwardScalar,
2019-09-07 18:07:48 +03:00
SPVFuncImplChromaReconstructNearest2Plane,
SPVFuncImplChromaReconstructNearest3Plane,
SPVFuncImplChromaReconstructLinear422CositedEven2Plane,
SPVFuncImplChromaReconstructLinear422CositedEven3Plane,
SPVFuncImplChromaReconstructLinear422Midpoint2Plane,
SPVFuncImplChromaReconstructLinear422Midpoint3Plane,
SPVFuncImplChromaReconstructLinear420XCositedEvenYCositedEven2Plane,
SPVFuncImplChromaReconstructLinear420XCositedEvenYCositedEven3Plane,
SPVFuncImplChromaReconstructLinear420XMidpointYCositedEven2Plane,
SPVFuncImplChromaReconstructLinear420XMidpointYCositedEven3Plane,
SPVFuncImplChromaReconstructLinear420XCositedEvenYMidpoint2Plane,
SPVFuncImplChromaReconstructLinear420XCositedEvenYMidpoint3Plane,
SPVFuncImplChromaReconstructLinear420XMidpointYMidpoint2Plane,
SPVFuncImplChromaReconstructLinear420XMidpointYMidpoint3Plane,
SPVFuncImplExpandITUFullRange,
SPVFuncImplExpandITUNarrowRange,
SPVFuncImplConvertYCbCrBT709,
SPVFuncImplConvertYCbCrBT601,
SPVFuncImplConvertYCbCrBT2020,
SPVFuncImplDynamicImageSampler,
2019-01-10 06:23:47 +03:00
SPVFuncImplArrayCopyMultidimMax = 6
};
2019-10-29 05:55:41 +03:00
// If the underlying resource has been used for comparison then duplicate loads of that resource must be too
// Use Metal's native frame-buffer fetch API for subpass inputs.
void emit_texture_op(const Instruction &i) override;
2019-01-10 06:23:47 +03:00
void emit_binary_unord_op(uint32_t result_type, uint32_t result_id, uint32_t op0, uint32_t op1, const char *op);
void emit_instruction(const Instruction &instr) override;
void emit_glsl_op(uint32_t result_type, uint32_t result_id, uint32_t op, const uint32_t *args,
uint32_t count) override;
2019-07-20 20:45:34 +03:00
void emit_spv_amd_shader_trinary_minmax_op(uint32_t result_type, uint32_t result_id, uint32_t op,
const uint32_t *args, uint32_t count) override;
2019-01-10 06:23:47 +03:00
void emit_header() override;
void emit_function_prototype(SPIRFunction &func, const Bitset &return_flags) override;
void emit_sampled_image_op(uint32_t result_type, uint32_t result_id, uint32_t image_id, uint32_t samp_id) override;
2019-05-18 22:28:43 +03:00
void emit_subgroup_op(const Instruction &i) override;
2019-09-07 18:07:48 +03:00
std::string to_texture_op(const Instruction &i, bool *forward,
SmallVector<uint32_t> &inherited_expressions) override;
2019-01-10 06:23:47 +03:00
void emit_fixup() override;
2019-02-09 23:44:48 +03:00
std::string to_struct_member(const SPIRType &type, uint32_t member_type_id, uint32_t index,
const std::string &qualifier = "");
2019-01-10 06:23:47 +03:00
void emit_struct_member(const SPIRType &type, uint32_t member_type_id, uint32_t index,
const std::string &qualifier = "", uint32_t base_offset = 0) override;
2019-07-25 06:25:58 +03:00
void emit_struct_padding_target(const SPIRType &type) override;
2019-01-10 06:23:47 +03:00
std::string type_to_glsl(const SPIRType &type, uint32_t id = 0) override;
2019-10-29 05:55:41 +03:00
// Allow Metal to use the array<T> template to make arrays a value type
std::string type_to_array_glsl(const SPIRType &type) override;
// Threadgroup arrays can't have a wrapper type
std::string variable_decl(const SPIRVariable &variable) override;
// GCC workaround of lambdas calling protected functions (for older GCC versions)
std::string variable_decl(const SPIRType &type, const std::string &name, uint32_t id = 0) override;
2019-01-10 06:23:47 +03:00
std::string image_type_glsl(const SPIRType &type, uint32_t id = 0) override;
std::string sampler_type(const SPIRType &type);
std::string builtin_to_glsl(spv::BuiltIn builtin, spv::StorageClass storage) override;
2019-09-07 18:07:48 +03:00
std::string to_func_call_arg(const SPIRFunction::Parameter &arg, uint32_t id) override;
2019-01-10 06:23:47 +03:00
std::string to_name(uint32_t id, bool allow_alias = true) const override;
2019-09-07 18:07:48 +03:00
std::string to_function_name(VariableID img, const SPIRType &imgtype, bool is_fetch, bool is_gather, bool is_proj,
2019-06-22 17:30:47 +03:00
bool has_array_offsets, bool has_offset, bool has_grad, bool has_dref, uint32_t lod,
uint32_t minlod) override;
2019-09-07 18:07:48 +03:00
std::string to_function_args(VariableID img, const SPIRType &imgtype, bool is_fetch, bool is_gather, bool is_proj,
2019-01-10 06:23:47 +03:00
uint32_t coord, uint32_t coord_components, uint32_t dref, uint32_t grad_x,
uint32_t grad_y, uint32_t lod, uint32_t coffset, uint32_t offset, uint32_t bias,
2019-06-22 17:30:47 +03:00
uint32_t comp, uint32_t sample, uint32_t minlod, bool *p_forward) override;
2019-01-10 06:23:47 +03:00
std::string to_initializer_expression(const SPIRVariable &var) override;
2019-10-29 05:55:41 +03:00
2019-07-25 06:25:58 +03:00
std::string unpack_expression_type(std::string expr_str, const SPIRType &type, uint32_t physical_type_id,
bool is_packed, bool row_major) override;
2019-10-29 05:55:41 +03:00
// Returns true for BuiltInSampleMask because gl_SampleMask[] is an array in SPIR-V, but [[sample_mask]] is a scalar in Metal.
bool builtin_translates_to_nonarray(spv::BuiltIn builtin) const override;
2019-01-10 06:23:47 +03:00
std::string bitcast_glsl_op(const SPIRType &result_type, const SPIRType &argument_type) override;
bool skip_argument(uint32_t id) const override;
std::string to_member_reference(uint32_t base, const SPIRType &type, uint32_t index, bool ptr_chain) override;
std::string to_qualifiers_glsl(uint32_t id) override;
void replace_illegal_names() override;
void declare_undefined_values() override;
void declare_constant_arrays();
2019-10-29 05:55:41 +03:00
// Constant arrays of non-primitive types (i.e. matrices) won't link properly into Metal libraries
void declare_complex_constant_arrays();
2019-02-23 07:54:12 +03:00
bool is_patch_block(const SPIRType &type);
2019-01-10 06:23:47 +03:00
bool is_non_native_row_major_matrix(uint32_t id) override;
bool member_is_non_native_row_major_matrix(const SPIRType &type, uint32_t index) override;
2019-07-25 06:25:58 +03:00
std::string convert_row_major_matrix(std::string exp_str, const SPIRType &exp_type, uint32_t physical_type_id,
bool is_packed) override;
2019-01-10 06:23:47 +03:00
void preprocess_op_codes();
void localize_global_variables();
void extract_global_variables_from_functions();
void mark_packable_structs();
void mark_as_packable(SPIRType &type);
std::unordered_map<uint32_t, std::set<uint32_t>> function_global_vars;
void extract_global_variables_from_function(uint32_t func_id, std::set<uint32_t> &added_arg_ids,
std::unordered_set<uint32_t> &global_var_ids,
std::unordered_set<uint32_t> &processed_func_ids);
2019-02-16 06:52:34 +03:00
uint32_t add_interface_block(spv::StorageClass storage, bool patch = false);
uint32_t add_interface_block_pointer(uint32_t ib_var_id, spv::StorageClass storage);
2019-01-10 06:23:47 +03:00
2020-02-05 08:38:12 +03:00
struct InterfaceBlockMeta
{
struct LocationMeta
{
uint32_t num_components = 0;
uint32_t ib_index = ~0u;
};
std::unordered_map<uint32_t, LocationMeta> location_meta;
bool strip_array = false;
};
2019-01-10 06:23:47 +03:00
void add_variable_to_interface_block(spv::StorageClass storage, const std::string &ib_var_ref, SPIRType &ib_type,
2020-02-05 08:38:12 +03:00
SPIRVariable &var, InterfaceBlockMeta &meta);
2019-01-10 06:23:47 +03:00
void add_composite_variable_to_interface_block(spv::StorageClass storage, const std::string &ib_var_ref,
2020-02-05 08:38:12 +03:00
SPIRType &ib_type, SPIRVariable &var, InterfaceBlockMeta &meta);
2019-01-10 06:23:47 +03:00
void add_plain_variable_to_interface_block(spv::StorageClass storage, const std::string &ib_var_ref,
2020-02-05 08:38:12 +03:00
SPIRType &ib_type, SPIRVariable &var, InterfaceBlockMeta &meta);
2019-01-10 06:23:47 +03:00
void add_plain_member_variable_to_interface_block(spv::StorageClass storage, const std::string &ib_var_ref,
2019-02-16 06:52:34 +03:00
SPIRType &ib_type, SPIRVariable &var, uint32_t index,
2020-02-05 08:38:12 +03:00
InterfaceBlockMeta &meta);
2019-01-10 06:23:47 +03:00
void add_composite_member_variable_to_interface_block(spv::StorageClass storage, const std::string &ib_var_ref,
2019-02-16 06:52:34 +03:00
SPIRType &ib_type, SPIRVariable &var, uint32_t index,
2020-02-05 08:38:12 +03:00
InterfaceBlockMeta &meta);
2019-02-16 06:52:34 +03:00
uint32_t get_accumulated_member_location(const SPIRVariable &var, uint32_t mbr_idx, bool strip_array);
2019-03-03 06:16:00 +03:00
void add_tess_level_input_to_interface_block(const std::string &ib_var_ref, SPIRType &ib_type, SPIRVariable &var);
2019-02-16 06:52:34 +03:00
void fix_up_interface_member_indices(spv::StorageClass storage, uint32_t ib_type_id);
2019-01-10 06:23:47 +03:00
void mark_location_as_used_by_shader(uint32_t location, spv::StorageClass storage);
uint32_t ensure_correct_builtin_type(uint32_t type_id, spv::BuiltIn builtin);
2020-02-05 08:38:12 +03:00
uint32_t ensure_correct_attribute_type(uint32_t type_id, uint32_t location, uint32_t num_components = 0);
2019-01-10 06:23:47 +03:00
2019-10-29 05:55:41 +03:00
void emit_custom_templates();
2019-01-10 06:23:47 +03:00
void emit_custom_functions();
void emit_resources();
2019-01-17 05:50:01 +03:00
void emit_specialization_constants_and_structs();
2019-01-10 06:23:47 +03:00
void emit_interface_block(uint32_t ib_var_id);
bool maybe_emit_array_assignment(uint32_t id_lhs, uint32_t id_rhs);
2019-07-25 06:25:58 +03:00
2019-01-19 04:21:31 +03:00
void fix_up_shader_inputs_outputs();
2019-01-10 06:23:47 +03:00
std::string func_type_decl(SPIRType &type);
2019-03-17 06:30:56 +03:00
std::string entry_point_args_classic(bool append_comma);
std::string entry_point_args_argument_buffer(bool append_comma);
std::string entry_point_arg_stage_in();
void entry_point_args_builtin(std::string &args);
void entry_point_args_discrete_descriptors(std::string &args);
2019-01-10 06:23:47 +03:00
std::string to_qualified_member_name(const SPIRType &type, uint32_t index);
std::string ensure_valid_name(std::string name, std::string pfx);
std::string to_sampler_expression(uint32_t id);
2019-01-17 05:50:01 +03:00
std::string to_swizzle_expression(uint32_t id);
2019-06-07 08:04:16 +03:00
std::string to_buffer_size_expression(uint32_t id);
2019-01-10 06:23:47 +03:00
std::string builtin_qualifier(spv::BuiltIn builtin);
2019-06-22 17:30:47 +03:00
std::string builtin_type_decl(spv::BuiltIn builtin, uint32_t id = 0);
2019-01-10 06:23:47 +03:00
std::string built_in_func_arg(spv::BuiltIn builtin, bool prefix_comma);
std::string member_attribute_qualifier(const SPIRType &type, uint32_t index);
std::string argument_decl(const SPIRFunction::Parameter &arg);
std::string round_fp_tex_coords(std::string tex_coords, bool coord_is_fp);
2019-09-07 18:07:48 +03:00
uint32_t get_metal_resource_index(SPIRVariable &var, SPIRType::BaseType basetype, uint32_t plane = 0);
2019-01-10 06:23:47 +03:00
uint32_t get_ordered_member_location(uint32_t type_id, uint32_t index, uint32_t *comp = nullptr);
2019-07-25 06:25:58 +03:00
// MSL packing rules. These compute the effective packing rules as observed by the MSL compiler in the MSL output.
// These values can change depending on various extended decorations which control packing rules.
// We need to make these rules match up with SPIR-V declared rules.
uint32_t get_declared_type_size_msl(const SPIRType &type, bool packed, bool row_major) const;
uint32_t get_declared_type_array_stride_msl(const SPIRType &type, bool packed, bool row_major) const;
uint32_t get_declared_type_matrix_stride_msl(const SPIRType &type, bool packed, bool row_major) const;
uint32_t get_declared_type_alignment_msl(const SPIRType &type, bool packed, bool row_major) const;
uint32_t get_declared_struct_member_size_msl(const SPIRType &struct_type, uint32_t index) const;
uint32_t get_declared_struct_member_array_stride_msl(const SPIRType &struct_type, uint32_t index) const;
uint32_t get_declared_struct_member_matrix_stride_msl(const SPIRType &struct_type, uint32_t index) const;
uint32_t get_declared_struct_member_alignment_msl(const SPIRType &struct_type, uint32_t index) const;
const SPIRType &get_physical_member_type(const SPIRType &struct_type, uint32_t index) const;
uint32_t get_declared_struct_size_msl(const SPIRType &struct_type, bool ignore_alignment = false,
bool ignore_padding = false) const;
2019-01-10 06:23:47 +03:00
std::string to_component_argument(uint32_t id);
2019-07-25 06:25:58 +03:00
void align_struct(SPIRType &ib_type, std::unordered_set<uint32_t> &aligned_structs);
void mark_scalar_layout_structs(const SPIRType &ib_type);
void mark_struct_members_packed(const SPIRType &type);
void ensure_member_packing_rules_msl(SPIRType &ib_type, uint32_t index);
bool validate_member_packing_rules_msl(const SPIRType &type, uint32_t index) const;
2019-01-10 06:23:47 +03:00
std::string get_argument_address_space(const SPIRVariable &argument);
2019-08-10 06:33:38 +03:00
std::string get_type_address_space(const SPIRType &type, uint32_t id, bool argument = false);
2019-07-20 20:45:34 +03:00
const char *to_restrict(uint32_t id, bool space = true);
2019-02-16 06:52:34 +03:00
SPIRType &get_stage_in_struct_type();
2019-02-09 23:44:48 +03:00
SPIRType &get_stage_out_struct_type();
2019-02-23 07:54:12 +03:00
SPIRType &get_patch_stage_in_struct_type();
2019-02-16 06:52:34 +03:00
SPIRType &get_patch_stage_out_struct_type();
std::string get_tess_factor_struct_name();
2019-01-10 06:23:47 +03:00
void emit_atomic_func_op(uint32_t result_type, uint32_t result_id, const char *op, uint32_t mem_order_1,
uint32_t mem_order_2, bool has_mem_order_2, uint32_t op0, uint32_t op1 = 0,
bool op1_is_pointer = false, bool op1_is_literal = false, uint32_t op2 = 0);
const char *get_memory_order(uint32_t spv_mem_sem);
void add_pragma_line(const std::string &line);
void add_typedef_line(const std::string &line);
void emit_barrier(uint32_t id_exe_scope, uint32_t id_mem_scope, uint32_t id_mem_sem);
2019-09-02 08:06:03 +03:00
void emit_array_copy(const std::string &lhs, uint32_t rhs_id, spv::StorageClass lhs_storage,
spv::StorageClass rhs_storage) override;
2019-01-10 06:23:47 +03:00
void build_implicit_builtins();
2019-06-07 08:04:16 +03:00
uint32_t build_constant_uint_array_pointer();
2019-01-10 06:23:47 +03:00
void emit_entry_point_declarations() override;
uint32_t builtin_frag_coord_id = 0;
uint32_t builtin_sample_id_id = 0;
2019-02-09 23:44:48 +03:00
uint32_t builtin_vertex_idx_id = 0;
uint32_t builtin_base_vertex_id = 0;
uint32_t builtin_instance_idx_id = 0;
uint32_t builtin_base_instance_id = 0;
2019-07-03 18:10:12 +03:00
uint32_t builtin_view_idx_id = 0;
uint32_t builtin_layer_id = 0;
2019-02-16 06:52:34 +03:00
uint32_t builtin_invocation_id_id = 0;
uint32_t builtin_primitive_id_id = 0;
2019-05-18 22:28:43 +03:00
uint32_t builtin_subgroup_invocation_id_id = 0;
uint32_t builtin_subgroup_size_id = 0;
2019-08-10 06:33:38 +03:00
uint32_t builtin_dispatch_base_id = 0;
2019-05-18 22:28:43 +03:00
uint32_t swizzle_buffer_id = 0;
2019-06-07 08:04:16 +03:00
uint32_t buffer_size_buffer_id = 0;
2019-07-03 18:10:12 +03:00
uint32_t view_mask_buffer_id = 0;
2019-09-07 18:07:48 +03:00
uint32_t dynamic_offsets_buffer_id = 0;
2019-01-10 06:23:47 +03:00
void bitcast_to_builtin_store(uint32_t target_id, std::string &expr, const SPIRType &expr_type) override;
void bitcast_from_builtin_load(uint32_t source_id, std::string &expr, const SPIRType &expr_type) override;
2019-01-17 19:43:47 +03:00
void emit_store_statement(uint32_t lhs_expression, uint32_t rhs_expression) override;
2019-01-10 06:23:47 +03:00
void analyze_sampled_image_usage();
2019-02-23 07:54:12 +03:00
bool emit_tessellation_access_chain(const uint32_t *ops, uint32_t length);
2019-10-29 05:55:41 +03:00
bool emit_tessellation_io_load(uint32_t result_type, uint32_t id, uint32_t ptr);
2019-02-23 07:54:12 +03:00
bool is_out_of_bounds_tessellation_level(uint32_t id_lhs);
2019-02-16 06:52:34 +03:00
2019-10-29 05:55:41 +03:00
void ensure_builtin(spv::StorageClass storage, spv::BuiltIn builtin);
2019-06-07 08:04:16 +03:00
void mark_implicit_builtin(spv::StorageClass storage, spv::BuiltIn builtin, uint32_t id);
2019-07-03 18:10:12 +03:00
std::string convert_to_f32(const std::string &expr, uint32_t components);
2019-01-10 06:23:47 +03:00
Options msl_options;
std::set<SPVFuncImpl> spv_function_implementations;
2019-03-09 21:20:48 +03:00
std::unordered_map<uint32_t, MSLVertexAttr> vtx_attrs_by_location;
std::unordered_map<uint32_t, MSLVertexAttr> vtx_attrs_by_builtin;
std::unordered_set<uint32_t> vtx_attrs_in_use;
2019-01-17 05:50:01 +03:00
std::unordered_map<uint32_t, uint32_t> fragment_output_components;
2019-01-10 06:23:47 +03:00
std::set<std::string> pragma_lines;
std::set<std::string> typedef_lines;
2019-04-15 07:53:50 +03:00
SmallVector<uint32_t> vars_needing_early_declaration;
2019-03-09 21:20:48 +03:00
2019-06-17 03:47:55 +03:00
std::unordered_map<StageSetBinding, std::pair<MSLResourceBinding, bool>, InternalHasher> resource_bindings;
2019-07-12 07:08:58 +03:00
2019-03-09 21:20:48 +03:00
uint32_t next_metal_resource_index_buffer = 0;
uint32_t next_metal_resource_index_texture = 0;
uint32_t next_metal_resource_index_sampler = 0;
2019-07-12 07:08:58 +03:00
// Intentionally uninitialized, works around MSVC 2013 bug.
uint32_t next_metal_resource_ids[kMaxArgumentBuffers];
2019-03-09 21:20:48 +03:00
2019-09-07 18:07:48 +03:00
VariableID stage_in_var_id = 0;
VariableID stage_out_var_id = 0;
VariableID patch_stage_in_var_id = 0;
VariableID patch_stage_out_var_id = 0;
VariableID stage_in_ptr_var_id = 0;
VariableID stage_out_ptr_var_id = 0;
2019-10-29 05:55:41 +03:00
// Handle HLSL-style 0-based vertex/instance index.
enum class TriState
{
Neutral,
No,
Yes
};
TriState needs_base_vertex_arg = TriState::Neutral;
TriState needs_base_instance_arg = TriState::Neutral;
2019-01-10 06:23:47 +03:00
bool has_sampled_images = false;
2019-10-29 05:55:41 +03:00
bool builtin_declaration = false; // Handle HLSL-style 0-based vertex/instance index.
bool use_builtin_array = false; // Force the use of C style array declaration.
2019-01-10 06:23:47 +03:00
bool is_rasterization_disabled = false;
2019-02-09 23:44:48 +03:00
bool capture_output_to_buffer = false;
2019-05-18 22:28:43 +03:00
bool needs_swizzle_buffer_def = false;
bool used_swizzle_buffer = false;
2019-03-03 06:16:00 +03:00
bool added_builtin_tess_level = false;
2019-05-18 22:28:43 +03:00
bool needs_subgroup_invocation_id = false;
2019-01-10 06:23:47 +03:00
std::string qual_pos_var_name;
std::string stage_in_var_name = "in";
std::string stage_out_var_name = "out";
2019-02-23 07:54:12 +03:00
std::string patch_stage_in_var_name = "patchIn";
2019-02-16 06:52:34 +03:00
std::string patch_stage_out_var_name = "patchOut";
2019-01-10 06:23:47 +03:00
std::string sampler_name_suffix = "Smplr";
2019-01-17 05:50:01 +03:00
std::string swizzle_name_suffix = "Swzl";
2019-06-07 08:04:16 +03:00
std::string buffer_size_name_suffix = "BufferSize";
2019-09-07 18:07:48 +03:00
std::string plane_name_suffix = "Plane";
2019-02-16 06:52:34 +03:00
std::string input_wg_var_name = "gl_in";
2019-02-09 23:44:48 +03:00
std::string output_buffer_var_name = "spvOut";
2019-02-16 06:52:34 +03:00
std::string patch_output_buffer_var_name = "spvPatchOut";
std::string tess_factor_buffer_var_name = "spvTessLevel";
2019-01-10 06:23:47 +03:00
spv::Op previous_instruction_opcode = spv::OpNop;
2019-06-17 03:47:55 +03:00
// Must be ordered since declaration is in a specific order.
std::map<uint32_t, MSLConstexprSampler> constexpr_samplers_by_id;
std::unordered_map<SetBindingPair, MSLConstexprSampler, InternalHasher> constexpr_samplers_by_binding;
const MSLConstexprSampler *find_constexpr_sampler(uint32_t id) const;
2019-06-07 08:04:16 +03:00
std::unordered_set<uint32_t> buffers_requiring_array_length;
2019-04-15 07:53:50 +03:00
SmallVector<uint32_t> buffer_arrays;
2019-10-29 05:55:41 +03:00
std::unordered_set<uint32_t> atomic_image_vars; // Emulate texture2D atomic operations
2019-01-10 06:23:47 +03:00
2019-09-07 18:07:48 +03:00
// Must be ordered since array is in a specific order.
std::map<SetBindingPair, std::pair<uint32_t, uint32_t>> buffers_requiring_dynamic_offset;
2020-02-05 08:38:12 +03:00
std::unordered_set<SetBindingPair, InternalHasher> inline_uniform_blocks;
2019-03-17 06:30:56 +03:00
uint32_t argument_buffer_ids[kMaxArgumentBuffers];
uint32_t argument_buffer_discrete_mask = 0;
2019-10-20 08:33:06 +03:00
uint32_t argument_buffer_device_storage_mask = 0;
2019-03-17 06:30:56 +03:00
void analyze_argument_buffers();
bool descriptor_set_is_argument_buffer(uint32_t desc_set) const;
2019-01-17 05:50:01 +03:00
uint32_t get_target_components_for_fragment_location(uint32_t location) const;
uint32_t build_extended_vector_type(uint32_t type_id, uint32_t components);
2019-04-15 07:53:50 +03:00
bool suppress_missing_prototypes = false;
2019-09-07 18:07:48 +03:00
void add_spv_func_and_recompile(SPVFuncImpl spv_func);
2020-02-05 08:38:12 +03:00
void activate_argument_buffer_resources();
2019-01-10 06:23:47 +03:00
// OpcodeHandler that handles several MSL preprocessing operations.
struct OpCodePreprocessor : OpcodeHandler
{
OpCodePreprocessor(CompilerMSL &compiler_)
: compiler(compiler_)
{
}
bool handle(spv::Op opcode, const uint32_t *args, uint32_t length) override;
CompilerMSL::SPVFuncImpl get_spv_func_impl(spv::Op opcode, const uint32_t *args);
void check_resource_write(uint32_t var_id);
CompilerMSL &compiler;
std::unordered_map<uint32_t, uint32_t> result_types;
2019-10-29 05:55:41 +03:00
std::unordered_map<uint32_t, uint32_t> image_pointers; // Emulate texture2D atomic operations
2019-01-10 06:23:47 +03:00
bool suppress_missing_prototypes = false;
bool uses_atomics = false;
bool uses_resource_write = false;
2019-05-18 22:28:43 +03:00
bool needs_subgroup_invocation_id = false;
2019-01-10 06:23:47 +03:00
};
// OpcodeHandler that scans for uses of sampled images
struct SampledImageScanner : OpcodeHandler
{
SampledImageScanner(CompilerMSL &compiler_)
: compiler(compiler_)
{
}
bool handle(spv::Op opcode, const uint32_t *args, uint32_t) override;
CompilerMSL &compiler;
};
// Sorts the members of a SPIRType and associated Meta info based on a settable sorting
// aspect, which defines which aspect of the struct members will be used to sort them.
// Regardless of the sorting aspect, built-in members always appear at the end of the struct.
struct MemberSorter
{
enum SortAspect
{
Location,
LocationReverse,
Offset,
OffsetThenLocationReverse,
Alphabetical
};
void sort();
bool operator()(uint32_t mbr_idx1, uint32_t mbr_idx2);
MemberSorter(SPIRType &t, Meta &m, SortAspect sa);
SPIRType &type;
Meta &meta;
SortAspect sort_aspect;
};
};
2019-04-15 07:53:50 +03:00
} // namespace SPIRV_CROSS_NAMESPACE
2019-01-10 06:23:47 +03:00
#endif