mirror of
https://github.com/KhronosGroup/Vulkan-Hpp.git
synced 2024-10-14 16:32:17 +00:00
Update Vulkan-Headers to v1.3.298 (#1971)
Some checks failed
Set Version Tag / set-version-tag (push) Has been cancelled
Some checks failed
Set Version Tag / set-version-tag (push) Has been cancelled
Co-authored-by: GitHub <noreply@github.com>
This commit is contained in:
parent
56fdc0b070
commit
8592ed9743
@ -1 +1 @@
|
||||
Subproject commit 14345dab231912ee9601136e96ca67a6e1f632e7
|
||||
Subproject commit d91597a82f881d473887b560a03a7edf2720b72c
|
@ -63,7 +63,7 @@ extern "C" __declspec( dllimport ) FARPROC __stdcall GetProcAddress( HINSTANCE h
|
||||
# include <span>
|
||||
#endif
|
||||
|
||||
static_assert( VK_HEADER_VERSION == 297, "Wrong VK_HEADER_VERSION!" );
|
||||
static_assert( VK_HEADER_VERSION == 298, "Wrong VK_HEADER_VERSION!" );
|
||||
|
||||
// <tuple> includes <sys/sysmacros.h> through some other header
|
||||
// this results in major(x) being resolved to gnu_dev_major(x)
|
||||
@ -3632,28 +3632,36 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
return ::vkGetExecutionGraphPipelineNodeIndexAMDX( device, executionGraph, pNodeInfo, pNodeIndex );
|
||||
}
|
||||
|
||||
void vkCmdInitializeGraphScratchMemoryAMDX( VkCommandBuffer commandBuffer, VkDeviceAddress scratch ) const VULKAN_HPP_NOEXCEPT
|
||||
void vkCmdInitializeGraphScratchMemoryAMDX( VkCommandBuffer commandBuffer,
|
||||
VkPipeline executionGraph,
|
||||
VkDeviceAddress scratch,
|
||||
VkDeviceSize scratchSize ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkCmdInitializeGraphScratchMemoryAMDX( commandBuffer, scratch );
|
||||
return ::vkCmdInitializeGraphScratchMemoryAMDX( commandBuffer, executionGraph, scratch, scratchSize );
|
||||
}
|
||||
|
||||
void vkCmdDispatchGraphAMDX( VkCommandBuffer commandBuffer,
|
||||
VkDeviceAddress scratch,
|
||||
VkDeviceSize scratchSize,
|
||||
const VkDispatchGraphCountInfoAMDX * pCountInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkCmdDispatchGraphAMDX( commandBuffer, scratch, pCountInfo );
|
||||
return ::vkCmdDispatchGraphAMDX( commandBuffer, scratch, scratchSize, pCountInfo );
|
||||
}
|
||||
|
||||
void vkCmdDispatchGraphIndirectAMDX( VkCommandBuffer commandBuffer,
|
||||
VkDeviceAddress scratch,
|
||||
VkDeviceSize scratchSize,
|
||||
const VkDispatchGraphCountInfoAMDX * pCountInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkCmdDispatchGraphIndirectAMDX( commandBuffer, scratch, pCountInfo );
|
||||
return ::vkCmdDispatchGraphIndirectAMDX( commandBuffer, scratch, scratchSize, pCountInfo );
|
||||
}
|
||||
|
||||
void vkCmdDispatchGraphIndirectCountAMDX( VkCommandBuffer commandBuffer, VkDeviceAddress scratch, VkDeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
void vkCmdDispatchGraphIndirectCountAMDX( VkCommandBuffer commandBuffer,
|
||||
VkDeviceAddress scratch,
|
||||
VkDeviceSize scratchSize,
|
||||
VkDeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkCmdDispatchGraphIndirectCountAMDX( commandBuffer, scratch, countInfo );
|
||||
return ::vkCmdDispatchGraphIndirectCountAMDX( commandBuffer, scratch, scratchSize, countInfo );
|
||||
}
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
|
@ -7028,6 +7028,9 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
eDisableOptimization = VK_PIPELINE_CREATE_2_DISABLE_OPTIMIZATION_BIT_KHR,
|
||||
eAllowDerivatives = VK_PIPELINE_CREATE_2_ALLOW_DERIVATIVES_BIT_KHR,
|
||||
eDerivative = VK_PIPELINE_CREATE_2_DERIVATIVE_BIT_KHR,
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
eExecutionGraphAMDX = VK_PIPELINE_CREATE_2_EXECUTION_GRAPH_BIT_AMDX,
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
eEnableLegacyDitheringEXT = VK_PIPELINE_CREATE_2_ENABLE_LEGACY_DITHERING_BIT_EXT,
|
||||
eViewIndexFromDeviceIndex = VK_PIPELINE_CREATE_2_VIEW_INDEX_FROM_DEVICE_INDEX_BIT_KHR,
|
||||
eDispatchBase = VK_PIPELINE_CREATE_2_DISPATCH_BASE_BIT_KHR,
|
||||
@ -7068,8 +7071,11 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
{
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR bool isBitmask = true;
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR PipelineCreateFlags2KHR allFlags =
|
||||
PipelineCreateFlagBits2KHR::eDisableOptimization | PipelineCreateFlagBits2KHR::eAllowDerivatives | PipelineCreateFlagBits2KHR::eDerivative |
|
||||
PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT | PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex |
|
||||
PipelineCreateFlagBits2KHR::eDisableOptimization | PipelineCreateFlagBits2KHR::eAllowDerivatives | PipelineCreateFlagBits2KHR::eDerivative
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
| PipelineCreateFlagBits2KHR::eExecutionGraphAMDX
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
| PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT | PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex |
|
||||
PipelineCreateFlagBits2KHR::eDispatchBase | PipelineCreateFlagBits2KHR::eDeferCompileNV | PipelineCreateFlagBits2KHR::eCaptureStatistics |
|
||||
PipelineCreateFlagBits2KHR::eCaptureInternalRepresentations | PipelineCreateFlagBits2KHR::eFailOnPipelineCompileRequired |
|
||||
PipelineCreateFlagBits2KHR::eEarlyReturnOnFailure | PipelineCreateFlagBits2KHR::eLinkTimeOptimizationEXT |
|
||||
|
@ -993,17 +993,15 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
{ "VK_AMDX_shader_enqueue",
|
||||
{ { "VK_VERSION_1_0",
|
||||
{ {
|
||||
"VK_KHR_get_physical_device_properties2",
|
||||
} } },
|
||||
{ "VK_VERSION_1_1",
|
||||
{ {
|
||||
"VK_KHR_synchronization2",
|
||||
"VK_KHR_spirv_1_4",
|
||||
"VK_EXT_extended_dynamic_state",
|
||||
} } },
|
||||
{ "VK_VERSION_1_3",
|
||||
{ {
|
||||
"VK_KHR_maintenance5",
|
||||
"VK_KHR_pipeline_library",
|
||||
"VK_KHR_spirv_1_4",
|
||||
} } } } },
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
{ "VK_EXT_inline_uniform_block",
|
||||
|
@ -15575,27 +15575,35 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
d.vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ) );
|
||||
d.vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkPipeline>( executionGraph ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ) );
|
||||
}
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
d.vkCmdDispatchGraphAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( pCountInfo ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
@ -15604,24 +15612,30 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
VULKAN_HPP_ASSERT( d.vkCmdDispatchGraphAMDX && "Function <vkCmdDispatchGraphAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
# endif
|
||||
|
||||
d.vkCmdDispatchGraphAMDX( m_commandBuffer, static_cast<VkDeviceAddress>( scratch ), reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||
d.vkCmdDispatchGraphAMDX( m_commandBuffer,
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||
}
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
d.vkCmdDispatchGraphIndirectAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( pCountInfo ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
@ -15630,19 +15644,24 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
VULKAN_HPP_ASSERT( d.vkCmdDispatchGraphIndirectAMDX && "Function <vkCmdDispatchGraphIndirectAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
# endif
|
||||
|
||||
d.vkCmdDispatchGraphIndirectAMDX(
|
||||
m_commandBuffer, static_cast<VkDeviceAddress>( scratch ), reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||
d.vkCmdDispatchGraphIndirectAMDX( m_commandBuffer,
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||
}
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
d.vkCmdDispatchGraphIndirectCountAMDX(
|
||||
static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ), static_cast<VkDeviceAddress>( countInfo ) );
|
||||
d.vkCmdDispatchGraphIndirectCountAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
static_cast<VkDeviceAddress>( countInfo ) );
|
||||
}
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
|
@ -6078,33 +6078,40 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_AMDX_shader_enqueue ===
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX * pCountInfo,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
@ -4626,7 +4626,9 @@ namespace std
|
||||
std::size_t seed = 0;
|
||||
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.size );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.minSize );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.maxSize );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, executionGraphPipelineScratchSizeAMDX.sizeGranularity );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
@ -11865,6 +11867,7 @@ namespace std
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.shaderEnqueue );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueueFeaturesAMDX.shaderMeshEnqueue );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
@ -11885,6 +11888,11 @@ namespace std
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphShaderPayloadSize );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphShaderPayloadCount );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.executionGraphDispatchAddressAlignment );
|
||||
for ( size_t i = 0; i < 3; ++i )
|
||||
{
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphWorkgroupCount[i] );
|
||||
}
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceShaderEnqueuePropertiesAMDX.maxExecutionGraphWorkgroups );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
|
@ -5947,15 +5947,20 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_AMDX_shader_enqueue ===
|
||||
|
||||
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch ) const VULKAN_HPP_NOEXCEPT;
|
||||
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
@ -18376,43 +18381,55 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
return nodeIndex;
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch ) const VULKAN_HPP_NOEXCEPT
|
||||
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX &&
|
||||
"Function <vkCmdInitializeGraphScratchMemoryAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
|
||||
getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ) );
|
||||
getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkPipeline>( executionGraph ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphAMDX && "Function <vkCmdDispatchGraphAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
|
||||
getDispatcher()->vkCmdDispatchGraphAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void
|
||||
CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphIndirectAMDX && "Function <vkCmdDispatchGraphIndirectAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
|
||||
getDispatcher()->vkCmdDispatchGraphIndirectAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX &&
|
||||
"Function <vkCmdDispatchGraphIndirectCountAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
|
||||
getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX(
|
||||
static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ), static_cast<VkDeviceAddress>( countInfo ) );
|
||||
getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
static_cast<VkDeviceAddress>( countInfo ) );
|
||||
}
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
|
@ -34505,9 +34505,14 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR StructureType structureType = StructureType::eExecutionGraphPipelineScratchSizeAMDX;
|
||||
|
||||
# if !defined( VULKAN_HPP_NO_STRUCT_CONSTRUCTORS )
|
||||
VULKAN_HPP_CONSTEXPR ExecutionGraphPipelineScratchSizeAMDX( VULKAN_HPP_NAMESPACE::DeviceSize size_ = {}, void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||
VULKAN_HPP_CONSTEXPR ExecutionGraphPipelineScratchSizeAMDX( VULKAN_HPP_NAMESPACE::DeviceSize minSize_ = {},
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize maxSize_ = {},
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize sizeGranularity_ = {},
|
||||
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||
: pNext{ pNext_ }
|
||||
, size{ size_ }
|
||||
, minSize{ minSize_ }
|
||||
, maxSize{ maxSize_ }
|
||||
, sizeGranularity{ sizeGranularity_ }
|
||||
{
|
||||
}
|
||||
|
||||
@ -34534,9 +34539,21 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_CONSTEXPR_14 ExecutionGraphPipelineScratchSizeAMDX & setSize( VULKAN_HPP_NAMESPACE::DeviceSize size_ ) VULKAN_HPP_NOEXCEPT
|
||||
VULKAN_HPP_CONSTEXPR_14 ExecutionGraphPipelineScratchSizeAMDX & setMinSize( VULKAN_HPP_NAMESPACE::DeviceSize minSize_ ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
size = size_;
|
||||
minSize = minSize_;
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_CONSTEXPR_14 ExecutionGraphPipelineScratchSizeAMDX & setMaxSize( VULKAN_HPP_NAMESPACE::DeviceSize maxSize_ ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
maxSize = maxSize_;
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_CONSTEXPR_14 ExecutionGraphPipelineScratchSizeAMDX & setSizeGranularity( VULKAN_HPP_NAMESPACE::DeviceSize sizeGranularity_ ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
sizeGranularity = sizeGranularity_;
|
||||
return *this;
|
||||
}
|
||||
# endif /*VULKAN_HPP_NO_STRUCT_SETTERS*/
|
||||
@ -34555,11 +34572,15 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
# if 14 <= VULKAN_HPP_CPP_VERSION
|
||||
auto
|
||||
# else
|
||||
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &, void * const &, VULKAN_HPP_NAMESPACE::DeviceSize const &>
|
||||
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &,
|
||||
void * const &,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize const &,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize const &,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize const &>
|
||||
# endif
|
||||
reflect() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return std::tie( sType, pNext, size );
|
||||
return std::tie( sType, pNext, minSize, maxSize, sizeGranularity );
|
||||
}
|
||||
# endif
|
||||
|
||||
@ -34571,7 +34592,8 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
# if defined( VULKAN_HPP_USE_REFLECT )
|
||||
return this->reflect() == rhs.reflect();
|
||||
# else
|
||||
return ( sType == rhs.sType ) && ( pNext == rhs.pNext ) && ( size == rhs.size );
|
||||
return ( sType == rhs.sType ) && ( pNext == rhs.pNext ) && ( minSize == rhs.minSize ) && ( maxSize == rhs.maxSize ) &&
|
||||
( sizeGranularity == rhs.sizeGranularity );
|
||||
# endif
|
||||
}
|
||||
|
||||
@ -34584,7 +34606,9 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
public:
|
||||
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::eExecutionGraphPipelineScratchSizeAMDX;
|
||||
void * pNext = {};
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize size = {};
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize minSize = {};
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize maxSize = {};
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize sizeGranularity = {};
|
||||
};
|
||||
|
||||
template <>
|
||||
@ -39792,7 +39816,7 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT indirectCommandsLayout_ = {},
|
||||
uint32_t maxSequenceCount_ = {},
|
||||
uint32_t maxDrawCount_ = {},
|
||||
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||
const void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||
: pNext{ pNext_ }
|
||||
, indirectExecutionSet{ indirectExecutionSet_ }
|
||||
, indirectCommandsLayout{ indirectCommandsLayout_ }
|
||||
@ -39818,7 +39842,7 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
}
|
||||
|
||||
#if !defined( VULKAN_HPP_NO_STRUCT_SETTERS )
|
||||
VULKAN_HPP_CONSTEXPR_14 GeneratedCommandsMemoryRequirementsInfoEXT & setPNext( void * pNext_ ) VULKAN_HPP_NOEXCEPT
|
||||
VULKAN_HPP_CONSTEXPR_14 GeneratedCommandsMemoryRequirementsInfoEXT & setPNext( const void * pNext_ ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
pNext = pNext_;
|
||||
return *this;
|
||||
@ -39866,7 +39890,7 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
auto
|
||||
# else
|
||||
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &,
|
||||
void * const &,
|
||||
const void * const &,
|
||||
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT const &,
|
||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT const &,
|
||||
uint32_t const &,
|
||||
@ -39899,7 +39923,7 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
|
||||
public:
|
||||
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::eGeneratedCommandsMemoryRequirementsInfoEXT;
|
||||
void * pNext = {};
|
||||
const void * pNext = {};
|
||||
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT indirectExecutionSet = {};
|
||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT indirectCommandsLayout = {};
|
||||
uint32_t maxSequenceCount = {};
|
||||
@ -87117,9 +87141,11 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
|
||||
# if !defined( VULKAN_HPP_NO_STRUCT_CONSTRUCTORS )
|
||||
VULKAN_HPP_CONSTEXPR PhysicalDeviceShaderEnqueueFeaturesAMDX( VULKAN_HPP_NAMESPACE::Bool32 shaderEnqueue_ = {},
|
||||
VULKAN_HPP_NAMESPACE::Bool32 shaderMeshEnqueue_ = {},
|
||||
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||
: pNext{ pNext_ }
|
||||
, shaderEnqueue{ shaderEnqueue_ }
|
||||
, shaderMeshEnqueue{ shaderMeshEnqueue_ }
|
||||
{
|
||||
}
|
||||
|
||||
@ -87151,6 +87177,13 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
shaderEnqueue = shaderEnqueue_;
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueueFeaturesAMDX &
|
||||
setShaderMeshEnqueue( VULKAN_HPP_NAMESPACE::Bool32 shaderMeshEnqueue_ ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
shaderMeshEnqueue = shaderMeshEnqueue_;
|
||||
return *this;
|
||||
}
|
||||
# endif /*VULKAN_HPP_NO_STRUCT_SETTERS*/
|
||||
|
||||
operator VkPhysicalDeviceShaderEnqueueFeaturesAMDX const &() const VULKAN_HPP_NOEXCEPT
|
||||
@ -87167,11 +87200,11 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
# if 14 <= VULKAN_HPP_CPP_VERSION
|
||||
auto
|
||||
# else
|
||||
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &, void * const &, VULKAN_HPP_NAMESPACE::Bool32 const &>
|
||||
std::tuple<VULKAN_HPP_NAMESPACE::StructureType const &, void * const &, VULKAN_HPP_NAMESPACE::Bool32 const &, VULKAN_HPP_NAMESPACE::Bool32 const &>
|
||||
# endif
|
||||
reflect() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return std::tie( sType, pNext, shaderEnqueue );
|
||||
return std::tie( sType, pNext, shaderEnqueue, shaderMeshEnqueue );
|
||||
}
|
||||
# endif
|
||||
|
||||
@ -87183,7 +87216,7 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
# if defined( VULKAN_HPP_USE_REFLECT )
|
||||
return this->reflect() == rhs.reflect();
|
||||
# else
|
||||
return ( sType == rhs.sType ) && ( pNext == rhs.pNext ) && ( shaderEnqueue == rhs.shaderEnqueue );
|
||||
return ( sType == rhs.sType ) && ( pNext == rhs.pNext ) && ( shaderEnqueue == rhs.shaderEnqueue ) && ( shaderMeshEnqueue == rhs.shaderMeshEnqueue );
|
||||
# endif
|
||||
}
|
||||
|
||||
@ -87197,6 +87230,7 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
VULKAN_HPP_NAMESPACE::StructureType sType = StructureType::ePhysicalDeviceShaderEnqueueFeaturesAMDX;
|
||||
void * pNext = {};
|
||||
VULKAN_HPP_NAMESPACE::Bool32 shaderEnqueue = {};
|
||||
VULKAN_HPP_NAMESPACE::Bool32 shaderMeshEnqueue = {};
|
||||
};
|
||||
|
||||
template <>
|
||||
@ -87215,11 +87249,13 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR StructureType structureType = StructureType::ePhysicalDeviceShaderEnqueuePropertiesAMDX;
|
||||
|
||||
# if !defined( VULKAN_HPP_NO_STRUCT_CONSTRUCTORS )
|
||||
VULKAN_HPP_CONSTEXPR PhysicalDeviceShaderEnqueuePropertiesAMDX( uint32_t maxExecutionGraphDepth_ = {},
|
||||
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueuePropertiesAMDX( uint32_t maxExecutionGraphDepth_ = {},
|
||||
uint32_t maxExecutionGraphShaderOutputNodes_ = {},
|
||||
uint32_t maxExecutionGraphShaderPayloadSize_ = {},
|
||||
uint32_t maxExecutionGraphShaderPayloadCount_ = {},
|
||||
uint32_t executionGraphDispatchAddressAlignment_ = {},
|
||||
std::array<uint32_t, 3> const & maxExecutionGraphWorkgroupCount_ = {},
|
||||
uint32_t maxExecutionGraphWorkgroups_ = {},
|
||||
void * pNext_ = nullptr ) VULKAN_HPP_NOEXCEPT
|
||||
: pNext{ pNext_ }
|
||||
, maxExecutionGraphDepth{ maxExecutionGraphDepth_ }
|
||||
@ -87227,10 +87263,12 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
, maxExecutionGraphShaderPayloadSize{ maxExecutionGraphShaderPayloadSize_ }
|
||||
, maxExecutionGraphShaderPayloadCount{ maxExecutionGraphShaderPayloadCount_ }
|
||||
, executionGraphDispatchAddressAlignment{ executionGraphDispatchAddressAlignment_ }
|
||||
, maxExecutionGraphWorkgroupCount{ maxExecutionGraphWorkgroupCount_ }
|
||||
, maxExecutionGraphWorkgroups{ maxExecutionGraphWorkgroups_ }
|
||||
{
|
||||
}
|
||||
|
||||
VULKAN_HPP_CONSTEXPR PhysicalDeviceShaderEnqueuePropertiesAMDX( PhysicalDeviceShaderEnqueuePropertiesAMDX const & rhs ) VULKAN_HPP_NOEXCEPT = default;
|
||||
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueuePropertiesAMDX( PhysicalDeviceShaderEnqueuePropertiesAMDX const & rhs ) VULKAN_HPP_NOEXCEPT = default;
|
||||
|
||||
PhysicalDeviceShaderEnqueuePropertiesAMDX( VkPhysicalDeviceShaderEnqueuePropertiesAMDX const & rhs ) VULKAN_HPP_NOEXCEPT
|
||||
: PhysicalDeviceShaderEnqueuePropertiesAMDX( *reinterpret_cast<PhysicalDeviceShaderEnqueuePropertiesAMDX const *>( &rhs ) )
|
||||
@ -87286,6 +87324,20 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
executionGraphDispatchAddressAlignment = executionGraphDispatchAddressAlignment_;
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueuePropertiesAMDX &
|
||||
setMaxExecutionGraphWorkgroupCount( std::array<uint32_t, 3> maxExecutionGraphWorkgroupCount_ ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
maxExecutionGraphWorkgroupCount = maxExecutionGraphWorkgroupCount_;
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_CONSTEXPR_14 PhysicalDeviceShaderEnqueuePropertiesAMDX &
|
||||
setMaxExecutionGraphWorkgroups( uint32_t maxExecutionGraphWorkgroups_ ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
maxExecutionGraphWorkgroups = maxExecutionGraphWorkgroups_;
|
||||
return *this;
|
||||
}
|
||||
# endif /*VULKAN_HPP_NO_STRUCT_SETTERS*/
|
||||
|
||||
operator VkPhysicalDeviceShaderEnqueuePropertiesAMDX const &() const VULKAN_HPP_NOEXCEPT
|
||||
@ -87308,6 +87360,8 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
uint32_t const &,
|
||||
uint32_t const &,
|
||||
uint32_t const &,
|
||||
uint32_t const &,
|
||||
VULKAN_HPP_NAMESPACE::ArrayWrapper1D<uint32_t, 3> const &,
|
||||
uint32_t const &>
|
||||
# endif
|
||||
reflect() const VULKAN_HPP_NOEXCEPT
|
||||
@ -87318,7 +87372,9 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
maxExecutionGraphShaderOutputNodes,
|
||||
maxExecutionGraphShaderPayloadSize,
|
||||
maxExecutionGraphShaderPayloadCount,
|
||||
executionGraphDispatchAddressAlignment );
|
||||
executionGraphDispatchAddressAlignment,
|
||||
maxExecutionGraphWorkgroupCount,
|
||||
maxExecutionGraphWorkgroups );
|
||||
}
|
||||
# endif
|
||||
|
||||
@ -87334,7 +87390,8 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
( maxExecutionGraphShaderOutputNodes == rhs.maxExecutionGraphShaderOutputNodes ) &&
|
||||
( maxExecutionGraphShaderPayloadSize == rhs.maxExecutionGraphShaderPayloadSize ) &&
|
||||
( maxExecutionGraphShaderPayloadCount == rhs.maxExecutionGraphShaderPayloadCount ) &&
|
||||
( executionGraphDispatchAddressAlignment == rhs.executionGraphDispatchAddressAlignment );
|
||||
( executionGraphDispatchAddressAlignment == rhs.executionGraphDispatchAddressAlignment ) &&
|
||||
( maxExecutionGraphWorkgroupCount == rhs.maxExecutionGraphWorkgroupCount ) && ( maxExecutionGraphWorkgroups == rhs.maxExecutionGraphWorkgroups );
|
||||
# endif
|
||||
}
|
||||
|
||||
@ -87352,6 +87409,8 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
uint32_t maxExecutionGraphShaderPayloadSize = {};
|
||||
uint32_t maxExecutionGraphShaderPayloadCount = {};
|
||||
uint32_t executionGraphDispatchAddressAlignment = {};
|
||||
VULKAN_HPP_NAMESPACE::ArrayWrapper1D<uint32_t, 3> maxExecutionGraphWorkgroupCount = {};
|
||||
uint32_t maxExecutionGraphWorkgroups = {};
|
||||
};
|
||||
|
||||
template <>
|
||||
|
@ -3385,6 +3385,10 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
result += "AllowDerivatives | ";
|
||||
if ( value & PipelineCreateFlagBits2KHR::eDerivative )
|
||||
result += "Derivative | ";
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
if ( value & PipelineCreateFlagBits2KHR::eExecutionGraphAMDX )
|
||||
result += "ExecutionGraphAMDX | ";
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
if ( value & PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT )
|
||||
result += "EnableLegacyDitheringEXT | ";
|
||||
if ( value & PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex )
|
||||
@ -8919,6 +8923,9 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
case PipelineCreateFlagBits2KHR::eDisableOptimization: return "DisableOptimization";
|
||||
case PipelineCreateFlagBits2KHR::eAllowDerivatives: return "AllowDerivatives";
|
||||
case PipelineCreateFlagBits2KHR::eDerivative: return "Derivative";
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
case PipelineCreateFlagBits2KHR::eExecutionGraphAMDX: return "ExecutionGraphAMDX";
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
case PipelineCreateFlagBits2KHR::eEnableLegacyDitheringEXT: return "EnableLegacyDitheringEXT";
|
||||
case PipelineCreateFlagBits2KHR::eViewIndexFromDeviceIndex: return "ViewIndexFromDeviceIndex";
|
||||
case PipelineCreateFlagBits2KHR::eDispatchBase: return "DispatchBase";
|
||||
|
Loading…
Reference in New Issue
Block a user