mirror of
https://github.com/KhronosGroup/Vulkan-Hpp.git
synced 2024-10-14 16:32:17 +00:00
Update Vulkan-Headers to v1.3.269 (#1707)
Co-authored-by: GitHub <noreply@github.com>
This commit is contained in:
parent
3e0d722ca5
commit
9c15a1e1ef
@ -1 +1 @@
|
||||
Subproject commit 07ff4233bc69e573ae0a6b4b48ad16451ac4e37f
|
||||
Subproject commit 374f9fd97520f6dd1b80745de09208d878ab4a52
|
@ -784,6 +784,10 @@ export namespace VULKAN_HPP_NAMESPACE
|
||||
using VULKAN_HPP_NAMESPACE::DisplacementMicromapFormatNV;
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
//=== VK_ARM_scheduling_controls ===
|
||||
using VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsFlagBitsARM;
|
||||
using VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsFlagsARM;
|
||||
|
||||
//=== VK_NV_memory_decompression ===
|
||||
using VULKAN_HPP_NAMESPACE::MemoryDecompressionMethodFlagBitsNV;
|
||||
using VULKAN_HPP_NAMESPACE::MemoryDecompressionMethodFlagsNV;
|
||||
@ -2260,6 +2264,15 @@ export namespace VULKAN_HPP_NAMESPACE
|
||||
using VULKAN_HPP_NAMESPACE::DeviceDiagnosticsConfigCreateInfoNV;
|
||||
using VULKAN_HPP_NAMESPACE::PhysicalDeviceDiagnosticsConfigFeaturesNV;
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
using VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV;
|
||||
using VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV;
|
||||
using VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV;
|
||||
using VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchFeaturesNV;
|
||||
using VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchPropertiesNV;
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
//=== VK_NV_low_latency ===
|
||||
using VULKAN_HPP_NAMESPACE::QueryLowLatencySupportNV;
|
||||
|
||||
@ -2517,6 +2530,11 @@ export namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_ARM_shader_core_properties ===
|
||||
using VULKAN_HPP_NAMESPACE::PhysicalDeviceShaderCorePropertiesARM;
|
||||
|
||||
//=== VK_ARM_scheduling_controls ===
|
||||
using VULKAN_HPP_NAMESPACE::DeviceQueueShaderCoreControlCreateInfoARM;
|
||||
using VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsFeaturesARM;
|
||||
using VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsPropertiesARM;
|
||||
|
||||
//=== VK_EXT_image_sliced_view_of_3d ===
|
||||
using VULKAN_HPP_NAMESPACE::ImageViewSlicedCreateInfoEXT;
|
||||
using VULKAN_HPP_NAMESPACE::PhysicalDeviceImageSlicedViewOf3DFeaturesEXT;
|
||||
@ -2816,6 +2834,12 @@ export namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_NV_device_generated_commands ===
|
||||
using VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV;
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
using VULKAN_HPP_NAMESPACE::CudaFunctionNV;
|
||||
using VULKAN_HPP_NAMESPACE::CudaModuleNV;
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
using VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA;
|
||||
@ -2909,6 +2933,12 @@ export namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_NV_device_generated_commands ===
|
||||
using VULKAN_HPP_NAMESPACE::UniqueIndirectCommandsLayoutNV;
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
using VULKAN_HPP_NAMESPACE::UniqueCudaFunctionNV;
|
||||
using VULKAN_HPP_NAMESPACE::UniqueCudaModuleNV;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
using VULKAN_HPP_NAMESPACE::UniqueBufferCollectionFUCHSIA;
|
||||
@ -3007,6 +3037,12 @@ export namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_NV_device_generated_commands ===
|
||||
using VULKAN_HPP_NAMESPACE::SharedIndirectCommandsLayoutNV;
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
using VULKAN_HPP_NAMESPACE::SharedCudaFunctionNV;
|
||||
using VULKAN_HPP_NAMESPACE::SharedCudaModuleNV;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
using VULKAN_HPP_NAMESPACE::SharedBufferCollectionFUCHSIA;
|
||||
@ -3181,6 +3217,12 @@ export namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_NV_device_generated_commands ===
|
||||
using VULKAN_HPP_RAII_NAMESPACE::IndirectCommandsLayoutNV;
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
using VULKAN_HPP_RAII_NAMESPACE::CudaFunctionNV;
|
||||
using VULKAN_HPP_RAII_NAMESPACE::CudaModuleNV;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
using VULKAN_HPP_RAII_NAMESPACE::BufferCollectionFUCHSIA;
|
||||
|
@ -55,7 +55,7 @@ extern "C" __declspec( dllimport ) FARPROC __stdcall GetProcAddress( HINSTANCE h
|
||||
# include <span>
|
||||
#endif
|
||||
|
||||
static_assert( VK_HEADER_VERSION == 268, "Wrong VK_HEADER_VERSION!" );
|
||||
static_assert( VK_HEADER_VERSION == 269, "Wrong VK_HEADER_VERSION!" );
|
||||
|
||||
// <tuple> includes <sys/sysmacros.h> through some other header
|
||||
// this results in major(x) being resolved to gnu_dev_major(x)
|
||||
@ -4783,6 +4783,46 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
}
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
VkResult vkCreateCudaModuleNV( VkDevice device,
|
||||
const VkCudaModuleCreateInfoNV * pCreateInfo,
|
||||
const VkAllocationCallbacks * pAllocator,
|
||||
VkCudaModuleNV * pModule ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkCreateCudaModuleNV( device, pCreateInfo, pAllocator, pModule );
|
||||
}
|
||||
|
||||
VkResult vkGetCudaModuleCacheNV( VkDevice device, VkCudaModuleNV module, size_t * pCacheSize, void * pCacheData ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkGetCudaModuleCacheNV( device, module, pCacheSize, pCacheData );
|
||||
}
|
||||
|
||||
VkResult vkCreateCudaFunctionNV( VkDevice device,
|
||||
const VkCudaFunctionCreateInfoNV * pCreateInfo,
|
||||
const VkAllocationCallbacks * pAllocator,
|
||||
VkCudaFunctionNV * pFunction ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkCreateCudaFunctionNV( device, pCreateInfo, pAllocator, pFunction );
|
||||
}
|
||||
|
||||
void vkDestroyCudaModuleNV( VkDevice device, VkCudaModuleNV module, const VkAllocationCallbacks * pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkDestroyCudaModuleNV( device, module, pAllocator );
|
||||
}
|
||||
|
||||
void vkDestroyCudaFunctionNV( VkDevice device, VkCudaFunctionNV function, const VkAllocationCallbacks * pAllocator ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkDestroyCudaFunctionNV( device, function, pAllocator );
|
||||
}
|
||||
|
||||
void vkCmdCudaLaunchKernelNV( VkCommandBuffer commandBuffer, const VkCudaLaunchInfoNV * pLaunchInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return ::vkCmdCudaLaunchKernelNV( commandBuffer, pLaunchInfo );
|
||||
}
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
|
||||
@ -11431,6 +11471,36 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
};
|
||||
};
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
template <>
|
||||
struct StructExtends<PhysicalDeviceCudaKernelLaunchFeaturesNV, PhysicalDeviceFeatures2>
|
||||
{
|
||||
enum
|
||||
{
|
||||
value = true
|
||||
};
|
||||
};
|
||||
|
||||
template <>
|
||||
struct StructExtends<PhysicalDeviceCudaKernelLaunchFeaturesNV, DeviceCreateInfo>
|
||||
{
|
||||
enum
|
||||
{
|
||||
value = true
|
||||
};
|
||||
};
|
||||
|
||||
template <>
|
||||
struct StructExtends<PhysicalDeviceCudaKernelLaunchPropertiesNV, PhysicalDeviceProperties2>
|
||||
{
|
||||
enum
|
||||
{
|
||||
value = true
|
||||
};
|
||||
};
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
//=== VK_NV_low_latency ===
|
||||
template <>
|
||||
struct StructExtends<QueryLowLatencySupportNV, SemaphoreCreateInfo>
|
||||
@ -12845,6 +12915,52 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
};
|
||||
};
|
||||
|
||||
//=== VK_ARM_scheduling_controls ===
|
||||
template <>
|
||||
struct StructExtends<DeviceQueueShaderCoreControlCreateInfoARM, DeviceQueueCreateInfo>
|
||||
{
|
||||
enum
|
||||
{
|
||||
value = true
|
||||
};
|
||||
};
|
||||
|
||||
template <>
|
||||
struct StructExtends<DeviceQueueShaderCoreControlCreateInfoARM, DeviceCreateInfo>
|
||||
{
|
||||
enum
|
||||
{
|
||||
value = true
|
||||
};
|
||||
};
|
||||
|
||||
template <>
|
||||
struct StructExtends<PhysicalDeviceSchedulingControlsFeaturesARM, PhysicalDeviceFeatures2>
|
||||
{
|
||||
enum
|
||||
{
|
||||
value = true
|
||||
};
|
||||
};
|
||||
|
||||
template <>
|
||||
struct StructExtends<PhysicalDeviceSchedulingControlsFeaturesARM, DeviceCreateInfo>
|
||||
{
|
||||
enum
|
||||
{
|
||||
value = true
|
||||
};
|
||||
};
|
||||
|
||||
template <>
|
||||
struct StructExtends<PhysicalDeviceSchedulingControlsPropertiesARM, PhysicalDeviceProperties2>
|
||||
{
|
||||
enum
|
||||
{
|
||||
value = true
|
||||
};
|
||||
};
|
||||
|
||||
//=== VK_EXT_image_sliced_view_of_3d ===
|
||||
template <>
|
||||
struct StructExtends<PhysicalDeviceImageSlicedViewOf3DFeaturesEXT, PhysicalDeviceFeatures2>
|
||||
@ -15102,6 +15218,23 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
PFN_dummy vkCmdEncodeVideoKHR_placeholder = 0;
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
PFN_vkCreateCudaModuleNV vkCreateCudaModuleNV = 0;
|
||||
PFN_vkGetCudaModuleCacheNV vkGetCudaModuleCacheNV = 0;
|
||||
PFN_vkCreateCudaFunctionNV vkCreateCudaFunctionNV = 0;
|
||||
PFN_vkDestroyCudaModuleNV vkDestroyCudaModuleNV = 0;
|
||||
PFN_vkDestroyCudaFunctionNV vkDestroyCudaFunctionNV = 0;
|
||||
PFN_vkCmdCudaLaunchKernelNV vkCmdCudaLaunchKernelNV = 0;
|
||||
#else
|
||||
PFN_dummy vkCreateCudaModuleNV_placeholder = 0;
|
||||
PFN_dummy vkGetCudaModuleCacheNV_placeholder = 0;
|
||||
PFN_dummy vkCreateCudaFunctionNV_placeholder = 0;
|
||||
PFN_dummy vkDestroyCudaModuleNV_placeholder = 0;
|
||||
PFN_dummy vkDestroyCudaFunctionNV_placeholder = 0;
|
||||
PFN_dummy vkCmdCudaLaunchKernelNV_placeholder = 0;
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
PFN_vkExportMetalObjectsEXT vkExportMetalObjectsEXT = 0;
|
||||
@ -16426,6 +16559,16 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
vkCmdEncodeVideoKHR = PFN_vkCmdEncodeVideoKHR( vkGetInstanceProcAddr( instance, "vkCmdEncodeVideoKHR" ) );
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
vkCreateCudaModuleNV = PFN_vkCreateCudaModuleNV( vkGetInstanceProcAddr( instance, "vkCreateCudaModuleNV" ) );
|
||||
vkGetCudaModuleCacheNV = PFN_vkGetCudaModuleCacheNV( vkGetInstanceProcAddr( instance, "vkGetCudaModuleCacheNV" ) );
|
||||
vkCreateCudaFunctionNV = PFN_vkCreateCudaFunctionNV( vkGetInstanceProcAddr( instance, "vkCreateCudaFunctionNV" ) );
|
||||
vkDestroyCudaModuleNV = PFN_vkDestroyCudaModuleNV( vkGetInstanceProcAddr( instance, "vkDestroyCudaModuleNV" ) );
|
||||
vkDestroyCudaFunctionNV = PFN_vkDestroyCudaFunctionNV( vkGetInstanceProcAddr( instance, "vkDestroyCudaFunctionNV" ) );
|
||||
vkCmdCudaLaunchKernelNV = PFN_vkCmdCudaLaunchKernelNV( vkGetInstanceProcAddr( instance, "vkCmdCudaLaunchKernelNV" ) );
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
vkExportMetalObjectsEXT = PFN_vkExportMetalObjectsEXT( vkGetInstanceProcAddr( instance, "vkExportMetalObjectsEXT" ) );
|
||||
@ -17467,6 +17610,16 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
vkCmdEncodeVideoKHR = PFN_vkCmdEncodeVideoKHR( vkGetDeviceProcAddr( device, "vkCmdEncodeVideoKHR" ) );
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
vkCreateCudaModuleNV = PFN_vkCreateCudaModuleNV( vkGetDeviceProcAddr( device, "vkCreateCudaModuleNV" ) );
|
||||
vkGetCudaModuleCacheNV = PFN_vkGetCudaModuleCacheNV( vkGetDeviceProcAddr( device, "vkGetCudaModuleCacheNV" ) );
|
||||
vkCreateCudaFunctionNV = PFN_vkCreateCudaFunctionNV( vkGetDeviceProcAddr( device, "vkCreateCudaFunctionNV" ) );
|
||||
vkDestroyCudaModuleNV = PFN_vkDestroyCudaModuleNV( vkGetDeviceProcAddr( device, "vkDestroyCudaModuleNV" ) );
|
||||
vkDestroyCudaFunctionNV = PFN_vkDestroyCudaFunctionNV( vkGetDeviceProcAddr( device, "vkDestroyCudaFunctionNV" ) );
|
||||
vkCmdCudaLaunchKernelNV = PFN_vkCmdCudaLaunchKernelNV( vkGetDeviceProcAddr( device, "vkCmdCudaLaunchKernelNV" ) );
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
vkExportMetalObjectsEXT = PFN_vkExportMetalObjectsEXT( vkGetDeviceProcAddr( device, "vkExportMetalObjectsEXT" ) );
|
||||
|
@ -1089,6 +1089,13 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
ePhysicalDeviceDiagnosticsConfigFeaturesNV = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_DIAGNOSTICS_CONFIG_FEATURES_NV,
|
||||
eDeviceDiagnosticsConfigCreateInfoNV = VK_STRUCTURE_TYPE_DEVICE_DIAGNOSTICS_CONFIG_CREATE_INFO_NV,
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
eCudaModuleCreateInfoNV = VK_STRUCTURE_TYPE_CUDA_MODULE_CREATE_INFO_NV,
|
||||
eCudaFunctionCreateInfoNV = VK_STRUCTURE_TYPE_CUDA_FUNCTION_CREATE_INFO_NV,
|
||||
eCudaLaunchInfoNV = VK_STRUCTURE_TYPE_CUDA_LAUNCH_INFO_NV,
|
||||
ePhysicalDeviceCudaKernelLaunchFeaturesNV = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_CUDA_KERNEL_LAUNCH_FEATURES_NV,
|
||||
ePhysicalDeviceCudaKernelLaunchPropertiesNV = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_CUDA_KERNEL_LAUNCH_PROPERTIES_NV,
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
eQueryLowLatencySupportNV = VK_STRUCTURE_TYPE_QUERY_LOW_LATENCY_SUPPORT_NV,
|
||||
#if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
eExportMetalObjectCreateInfoEXT = VK_STRUCTURE_TYPE_EXPORT_METAL_OBJECT_CREATE_INFO_EXT,
|
||||
@ -1260,6 +1267,9 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
eDeviceBufferMemoryRequirementsKHR = VK_STRUCTURE_TYPE_DEVICE_BUFFER_MEMORY_REQUIREMENTS_KHR,
|
||||
eDeviceImageMemoryRequirementsKHR = VK_STRUCTURE_TYPE_DEVICE_IMAGE_MEMORY_REQUIREMENTS_KHR,
|
||||
ePhysicalDeviceShaderCorePropertiesARM = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_SHADER_CORE_PROPERTIES_ARM,
|
||||
eDeviceQueueShaderCoreControlCreateInfoARM = VK_STRUCTURE_TYPE_DEVICE_QUEUE_SHADER_CORE_CONTROL_CREATE_INFO_ARM,
|
||||
ePhysicalDeviceSchedulingControlsFeaturesARM = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_SCHEDULING_CONTROLS_FEATURES_ARM,
|
||||
ePhysicalDeviceSchedulingControlsPropertiesARM = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_SCHEDULING_CONTROLS_PROPERTIES_ARM,
|
||||
ePhysicalDeviceImageSlicedViewOf3DFeaturesEXT = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_IMAGE_SLICED_VIEW_OF_3D_FEATURES_EXT,
|
||||
eImageViewSlicedCreateInfoEXT = VK_STRUCTURE_TYPE_IMAGE_VIEW_SLICED_CREATE_INFO_EXT,
|
||||
ePhysicalDeviceDescriptorSetHostMappingFeaturesVALVE = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_DESCRIPTOR_SET_HOST_MAPPING_FEATURES_VALVE,
|
||||
@ -1430,6 +1440,10 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
eDeferredOperationKHR = VK_OBJECT_TYPE_DEFERRED_OPERATION_KHR,
|
||||
eIndirectCommandsLayoutNV = VK_OBJECT_TYPE_INDIRECT_COMMANDS_LAYOUT_NV,
|
||||
ePrivateDataSlotEXT = VK_OBJECT_TYPE_PRIVATE_DATA_SLOT_EXT,
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
eCudaModuleNV = VK_OBJECT_TYPE_CUDA_MODULE_NV,
|
||||
eCudaFunctionNV = VK_OBJECT_TYPE_CUDA_FUNCTION_NV,
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
#if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
eBufferCollectionFUCHSIA = VK_OBJECT_TYPE_BUFFER_COLLECTION_FUCHSIA,
|
||||
#endif /*VK_USE_PLATFORM_FUCHSIA*/
|
||||
@ -4660,6 +4674,10 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
eAccelerationStructureKHR = VK_DEBUG_REPORT_OBJECT_TYPE_ACCELERATION_STRUCTURE_KHR_EXT,
|
||||
eSamplerYcbcrConversionKHR = VK_DEBUG_REPORT_OBJECT_TYPE_SAMPLER_YCBCR_CONVERSION_KHR_EXT,
|
||||
eAccelerationStructureNV = VK_DEBUG_REPORT_OBJECT_TYPE_ACCELERATION_STRUCTURE_NV_EXT,
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
eCudaModuleNV = VK_DEBUG_REPORT_OBJECT_TYPE_CUDA_MODULE_NV,
|
||||
eCudaFunctionNV = VK_DEBUG_REPORT_OBJECT_TYPE_CUDA_FUNCTION_NV,
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
#if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
eBufferCollectionFUCHSIA = VK_DEBUG_REPORT_OBJECT_TYPE_BUFFER_COLLECTION_FUCHSIA_EXT
|
||||
#endif /*VK_USE_PLATFORM_FUCHSIA*/
|
||||
@ -6753,6 +6771,22 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
};
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
//=== VK_ARM_scheduling_controls ===
|
||||
|
||||
enum class PhysicalDeviceSchedulingControlsFlagBitsARM : VkPhysicalDeviceSchedulingControlsFlagsARM
|
||||
{
|
||||
eShaderCoreCount = VK_PHYSICAL_DEVICE_SCHEDULING_CONTROLS_SHADER_CORE_COUNT_ARM
|
||||
};
|
||||
|
||||
using PhysicalDeviceSchedulingControlsFlagsARM = Flags<PhysicalDeviceSchedulingControlsFlagBitsARM>;
|
||||
|
||||
template <>
|
||||
struct FlagTraits<PhysicalDeviceSchedulingControlsFlagBitsARM>
|
||||
{
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR bool isBitmask = true;
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR PhysicalDeviceSchedulingControlsFlagsARM allFlags = PhysicalDeviceSchedulingControlsFlagBitsARM::eShaderCoreCount;
|
||||
};
|
||||
|
||||
//=== VK_NV_memory_decompression ===
|
||||
|
||||
enum class MemoryDecompressionMethodFlagBitsNV : VkMemoryDecompressionMethodFlagsNV
|
||||
@ -7289,6 +7323,12 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_NV_device_generated_commands ===
|
||||
case VULKAN_HPP_NAMESPACE::ObjectType::eIndirectCommandsLayoutNV: return VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eUnknown;
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
case VULKAN_HPP_NAMESPACE::ObjectType::eCudaModuleNV: return VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eUnknown;
|
||||
case VULKAN_HPP_NAMESPACE::ObjectType::eCudaFunctionNV: return VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eUnknown;
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
case VULKAN_HPP_NAMESPACE::ObjectType::eBufferCollectionFUCHSIA: return VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eBufferCollectionFUCHSIA;
|
||||
|
@ -299,6 +299,9 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
"VK_NV_device_diagnostics_config",
|
||||
"VK_QCOM_render_pass_store_ops",
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
"VK_NV_cuda_kernel_launch",
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
"VK_NV_low_latency",
|
||||
#if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
"VK_EXT_metal_objects",
|
||||
@ -365,6 +368,7 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
"VK_EXT_pageable_device_local_memory",
|
||||
"VK_KHR_maintenance4",
|
||||
"VK_ARM_shader_core_properties",
|
||||
"VK_ARM_scheduling_controls",
|
||||
"VK_EXT_image_sliced_view_of_3d",
|
||||
"VK_VALVE_descriptor_set_host_mapping",
|
||||
"VK_EXT_depth_clamp_zero_one",
|
||||
@ -1774,6 +1778,11 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
} } } } },
|
||||
{ "VK_KHR_maintenance4", { { "VK_VERSION_1_1", { {} } } } },
|
||||
{ "VK_ARM_shader_core_properties", { { "VK_VERSION_1_1", { {} } } } },
|
||||
{ "VK_ARM_scheduling_controls",
|
||||
{ { "VK_VERSION_1_0",
|
||||
{ {
|
||||
"VK_ARM_shader_core_builtins",
|
||||
} } } } },
|
||||
{ "VK_EXT_image_sliced_view_of_3d",
|
||||
{ { "VK_VERSION_1_0",
|
||||
{ {
|
||||
@ -2674,7 +2683,11 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
|| ( extension == "VK_KHR_video_encode_queue" )
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|| ( extension == "VK_NV_device_diagnostics_config" ) || ( extension == "VK_QCOM_render_pass_store_ops" ) || ( extension == "VK_NV_low_latency" )
|
||||
|| ( extension == "VK_NV_device_diagnostics_config" ) || ( extension == "VK_QCOM_render_pass_store_ops" )
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
|| ( extension == "VK_NV_cuda_kernel_launch" )
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|| ( extension == "VK_NV_low_latency" )
|
||||
#if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
|| ( extension == "VK_EXT_metal_objects" )
|
||||
#endif /*VK_USE_PLATFORM_METAL_EXT*/
|
||||
@ -2710,16 +2723,17 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|| ( extension == "VK_EXT_load_store_op_none" ) || ( extension == "VK_HUAWEI_cluster_culling_shader" ) ||
|
||||
( extension == "VK_EXT_border_color_swizzle" ) || ( extension == "VK_EXT_pageable_device_local_memory" ) || ( extension == "VK_KHR_maintenance4" ) ||
|
||||
( extension == "VK_ARM_shader_core_properties" ) || ( extension == "VK_EXT_image_sliced_view_of_3d" ) ||
|
||||
( extension == "VK_VALVE_descriptor_set_host_mapping" ) || ( extension == "VK_EXT_depth_clamp_zero_one" ) ||
|
||||
( extension == "VK_EXT_non_seamless_cube_map" ) || ( extension == "VK_QCOM_fragment_density_map_offset" ) ||
|
||||
( extension == "VK_NV_copy_memory_indirect" ) || ( extension == "VK_NV_memory_decompression" ) ||
|
||||
( extension == "VK_NV_device_generated_commands_compute" ) || ( extension == "VK_NV_linear_color_attachment" ) ||
|
||||
( extension == "VK_EXT_image_compression_control_swapchain" ) || ( extension == "VK_QCOM_image_processing" ) ||
|
||||
( extension == "VK_EXT_nested_command_buffer" ) || ( extension == "VK_EXT_external_memory_acquire_unmodified" ) ||
|
||||
( extension == "VK_EXT_extended_dynamic_state3" ) || ( extension == "VK_EXT_subpass_merge_feedback" ) ||
|
||||
( extension == "VK_EXT_shader_module_identifier" ) || ( extension == "VK_EXT_rasterization_order_attachment_access" ) ||
|
||||
( extension == "VK_NV_optical_flow" ) || ( extension == "VK_EXT_legacy_dithering" ) || ( extension == "VK_EXT_pipeline_protected_access" )
|
||||
( extension == "VK_ARM_shader_core_properties" ) || ( extension == "VK_ARM_scheduling_controls" ) ||
|
||||
( extension == "VK_EXT_image_sliced_view_of_3d" ) || ( extension == "VK_VALVE_descriptor_set_host_mapping" ) ||
|
||||
( extension == "VK_EXT_depth_clamp_zero_one" ) || ( extension == "VK_EXT_non_seamless_cube_map" ) ||
|
||||
( extension == "VK_QCOM_fragment_density_map_offset" ) || ( extension == "VK_NV_copy_memory_indirect" ) ||
|
||||
( extension == "VK_NV_memory_decompression" ) || ( extension == "VK_NV_device_generated_commands_compute" ) ||
|
||||
( extension == "VK_NV_linear_color_attachment" ) || ( extension == "VK_EXT_image_compression_control_swapchain" ) ||
|
||||
( extension == "VK_QCOM_image_processing" ) || ( extension == "VK_EXT_nested_command_buffer" ) ||
|
||||
( extension == "VK_EXT_external_memory_acquire_unmodified" ) || ( extension == "VK_EXT_extended_dynamic_state3" ) ||
|
||||
( extension == "VK_EXT_subpass_merge_feedback" ) || ( extension == "VK_EXT_shader_module_identifier" ) ||
|
||||
( extension == "VK_EXT_rasterization_order_attachment_access" ) || ( extension == "VK_NV_optical_flow" ) ||
|
||||
( extension == "VK_EXT_legacy_dithering" ) || ( extension == "VK_EXT_pipeline_protected_access" )
|
||||
#if defined( VK_USE_PLATFORM_ANDROID_KHR )
|
||||
|| ( extension == "VK_ANDROID_external_format_resolve" )
|
||||
#endif /*VK_USE_PLATFORM_ANDROID_KHR*/
|
||||
|
@ -19211,6 +19211,304 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createCudaModuleNV( const VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV * pCreateInfo,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
VULKAN_HPP_NAMESPACE::CudaModuleNV * pModule,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
return static_cast<Result>( d.vkCreateCudaModuleNV( m_device,
|
||||
reinterpret_cast<const VkCudaModuleCreateInfoNV *>( pCreateInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( pAllocator ),
|
||||
reinterpret_cast<VkCudaModuleNV *>( pModule ) ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::CudaModuleNV>::type
|
||||
Device::createCudaModuleNV( const VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV & createInfo,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
|
||||
Dispatch const & d ) const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
VULKAN_HPP_NAMESPACE::CudaModuleNV module;
|
||||
VkResult result =
|
||||
d.vkCreateCudaModuleNV( m_device,
|
||||
reinterpret_cast<const VkCudaModuleCreateInfoNV *>( &createInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ),
|
||||
reinterpret_cast<VkCudaModuleNV *>( &module ) );
|
||||
resultCheck( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), VULKAN_HPP_NAMESPACE_STRING "::Device::createCudaModuleNV" );
|
||||
|
||||
return createResultValueType( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), module );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_NO_SMART_HANDLE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::CudaModuleNV, Dispatch>>::type
|
||||
Device::createCudaModuleNVUnique( const VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV & createInfo,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
|
||||
Dispatch const & d ) const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
VULKAN_HPP_NAMESPACE::CudaModuleNV module;
|
||||
VkResult result =
|
||||
d.vkCreateCudaModuleNV( m_device,
|
||||
reinterpret_cast<const VkCudaModuleCreateInfoNV *>( &createInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ),
|
||||
reinterpret_cast<VkCudaModuleNV *>( &module ) );
|
||||
resultCheck( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), VULKAN_HPP_NAMESPACE_STRING "::Device::createCudaModuleNVUnique" );
|
||||
|
||||
return createResultValueType(
|
||||
static_cast<VULKAN_HPP_NAMESPACE::Result>( result ),
|
||||
UniqueHandle<VULKAN_HPP_NAMESPACE::CudaModuleNV, Dispatch>( module, ObjectDestroy<Device, Dispatch>( *this, allocator, d ) ) );
|
||||
}
|
||||
# endif /* VULKAN_HPP_NO_SMART_HANDLE */
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getCudaModuleCacheNV( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
size_t * pCacheSize,
|
||||
void * pCacheData,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
return static_cast<Result>( d.vkGetCudaModuleCacheNV( m_device, static_cast<VkCudaModuleNV>( module ), pCacheSize, pCacheData ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Uint8_tAllocator, typename Dispatch>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type
|
||||
Device::getCudaModuleCacheNV( VULKAN_HPP_NAMESPACE::CudaModuleNV module, Dispatch const & d ) const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
std::vector<uint8_t, Uint8_tAllocator> cacheData;
|
||||
size_t cacheSize;
|
||||
VkResult result;
|
||||
do
|
||||
{
|
||||
result = d.vkGetCudaModuleCacheNV( m_device, static_cast<VkCudaModuleNV>( module ), &cacheSize, nullptr );
|
||||
if ( ( result == VK_SUCCESS ) && cacheSize )
|
||||
{
|
||||
cacheData.resize( cacheSize );
|
||||
result = d.vkGetCudaModuleCacheNV( m_device, static_cast<VkCudaModuleNV>( module ), &cacheSize, reinterpret_cast<void *>( cacheData.data() ) );
|
||||
}
|
||||
} while ( result == VK_INCOMPLETE );
|
||||
resultCheck( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), VULKAN_HPP_NAMESPACE_STRING "::Device::getCudaModuleCacheNV" );
|
||||
VULKAN_HPP_ASSERT( cacheSize <= cacheData.size() );
|
||||
if ( cacheSize < cacheData.size() )
|
||||
{
|
||||
cacheData.resize( cacheSize );
|
||||
}
|
||||
return createResultValueType( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), cacheData );
|
||||
}
|
||||
|
||||
template <typename Uint8_tAllocator,
|
||||
typename Dispatch,
|
||||
typename B1,
|
||||
typename std::enable_if<std::is_same<typename B1::value_type, uint8_t>::value, int>::type>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type
|
||||
Device::getCudaModuleCacheNV( VULKAN_HPP_NAMESPACE::CudaModuleNV module, Uint8_tAllocator & uint8_tAllocator, Dispatch const & d ) const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
std::vector<uint8_t, Uint8_tAllocator> cacheData( uint8_tAllocator );
|
||||
size_t cacheSize;
|
||||
VkResult result;
|
||||
do
|
||||
{
|
||||
result = d.vkGetCudaModuleCacheNV( m_device, static_cast<VkCudaModuleNV>( module ), &cacheSize, nullptr );
|
||||
if ( ( result == VK_SUCCESS ) && cacheSize )
|
||||
{
|
||||
cacheData.resize( cacheSize );
|
||||
result = d.vkGetCudaModuleCacheNV( m_device, static_cast<VkCudaModuleNV>( module ), &cacheSize, reinterpret_cast<void *>( cacheData.data() ) );
|
||||
}
|
||||
} while ( result == VK_INCOMPLETE );
|
||||
resultCheck( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), VULKAN_HPP_NAMESPACE_STRING "::Device::getCudaModuleCacheNV" );
|
||||
VULKAN_HPP_ASSERT( cacheSize <= cacheData.size() );
|
||||
if ( cacheSize < cacheData.size() )
|
||||
{
|
||||
cacheData.resize( cacheSize );
|
||||
}
|
||||
return createResultValueType( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), cacheData );
|
||||
}
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createCudaFunctionNV( const VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV * pCreateInfo,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
VULKAN_HPP_NAMESPACE::CudaFunctionNV * pFunction,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
return static_cast<Result>( d.vkCreateCudaFunctionNV( m_device,
|
||||
reinterpret_cast<const VkCudaFunctionCreateInfoNV *>( pCreateInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( pAllocator ),
|
||||
reinterpret_cast<VkCudaFunctionNV *>( pFunction ) ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::CudaFunctionNV>::type
|
||||
Device::createCudaFunctionNV( const VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV & createInfo,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
|
||||
Dispatch const & d ) const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
VULKAN_HPP_NAMESPACE::CudaFunctionNV function;
|
||||
VkResult result =
|
||||
d.vkCreateCudaFunctionNV( m_device,
|
||||
reinterpret_cast<const VkCudaFunctionCreateInfoNV *>( &createInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ),
|
||||
reinterpret_cast<VkCudaFunctionNV *>( &function ) );
|
||||
resultCheck( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), VULKAN_HPP_NAMESPACE_STRING "::Device::createCudaFunctionNV" );
|
||||
|
||||
return createResultValueType( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), function );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_NO_SMART_HANDLE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::CudaFunctionNV, Dispatch>>::type
|
||||
Device::createCudaFunctionNVUnique( const VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV & createInfo,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
|
||||
Dispatch const & d ) const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
VULKAN_HPP_NAMESPACE::CudaFunctionNV function;
|
||||
VkResult result =
|
||||
d.vkCreateCudaFunctionNV( m_device,
|
||||
reinterpret_cast<const VkCudaFunctionCreateInfoNV *>( &createInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ),
|
||||
reinterpret_cast<VkCudaFunctionNV *>( &function ) );
|
||||
resultCheck( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), VULKAN_HPP_NAMESPACE_STRING "::Device::createCudaFunctionNVUnique" );
|
||||
|
||||
return createResultValueType(
|
||||
static_cast<VULKAN_HPP_NAMESPACE::Result>( result ),
|
||||
UniqueHandle<VULKAN_HPP_NAMESPACE::CudaFunctionNV, Dispatch>( function, ObjectDestroy<Device, Dispatch>( *this, allocator, d ) ) );
|
||||
}
|
||||
# endif /* VULKAN_HPP_NO_SMART_HANDLE */
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void Device::destroyCudaModuleNV( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
d.vkDestroyCudaModuleNV( m_device, static_cast<VkCudaModuleNV>( module ), reinterpret_cast<const VkAllocationCallbacks *>( pAllocator ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void Device::destroyCudaModuleNV( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
d.vkDestroyCudaModuleNV( m_device,
|
||||
static_cast<VkCudaModuleNV>( module ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ) );
|
||||
}
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
d.vkDestroyCudaModuleNV( m_device, static_cast<VkCudaModuleNV>( module ), reinterpret_cast<const VkAllocationCallbacks *>( pAllocator ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
d.vkDestroyCudaModuleNV( m_device,
|
||||
static_cast<VkCudaModuleNV>( module ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ) );
|
||||
}
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void Device::destroyCudaFunctionNV( VULKAN_HPP_NAMESPACE::CudaFunctionNV function,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
d.vkDestroyCudaFunctionNV( m_device, static_cast<VkCudaFunctionNV>( function ), reinterpret_cast<const VkAllocationCallbacks *>( pAllocator ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void Device::destroyCudaFunctionNV( VULKAN_HPP_NAMESPACE::CudaFunctionNV function,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
d.vkDestroyCudaFunctionNV( m_device,
|
||||
static_cast<VkCudaFunctionNV>( function ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ) );
|
||||
}
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::CudaFunctionNV function,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
d.vkDestroyCudaFunctionNV( m_device, static_cast<VkCudaFunctionNV>( function ), reinterpret_cast<const VkAllocationCallbacks *>( pAllocator ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void Device::destroy( VULKAN_HPP_NAMESPACE::CudaFunctionNV function,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
d.vkDestroyCudaFunctionNV( m_device,
|
||||
static_cast<VkCudaFunctionNV>( function ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ) );
|
||||
}
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void CommandBuffer::cudaLaunchKernelNV( const VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV * pLaunchInfo,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
d.vkCmdCudaLaunchKernelNV( m_commandBuffer, reinterpret_cast<const VkCudaLaunchInfoNV *>( pLaunchInfo ) );
|
||||
}
|
||||
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch>
|
||||
VULKAN_HPP_INLINE void CommandBuffer::cudaLaunchKernelNV( const VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV & launchInfo,
|
||||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
|
||||
d.vkCmdCudaLaunchKernelNV( m_commandBuffer, reinterpret_cast<const VkCudaLaunchInfoNV *>( &launchInfo ) );
|
||||
}
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
|
||||
|
@ -1262,6 +1262,15 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
struct PhysicalDeviceDiagnosticsConfigFeaturesNV;
|
||||
struct DeviceDiagnosticsConfigCreateInfoNV;
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
struct CudaModuleCreateInfoNV;
|
||||
struct CudaFunctionCreateInfoNV;
|
||||
struct CudaLaunchInfoNV;
|
||||
struct PhysicalDeviceCudaKernelLaunchFeaturesNV;
|
||||
struct PhysicalDeviceCudaKernelLaunchPropertiesNV;
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
//=== VK_NV_low_latency ===
|
||||
struct QueryLowLatencySupportNV;
|
||||
|
||||
@ -1519,6 +1528,11 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_ARM_shader_core_properties ===
|
||||
struct PhysicalDeviceShaderCorePropertiesARM;
|
||||
|
||||
//=== VK_ARM_scheduling_controls ===
|
||||
struct DeviceQueueShaderCoreControlCreateInfoARM;
|
||||
struct PhysicalDeviceSchedulingControlsFeaturesARM;
|
||||
struct PhysicalDeviceSchedulingControlsPropertiesARM;
|
||||
|
||||
//=== VK_EXT_image_sliced_view_of_3d ===
|
||||
struct PhysicalDeviceImageSlicedViewOf3DFeaturesEXT;
|
||||
struct ImageViewSlicedCreateInfoEXT;
|
||||
@ -1816,6 +1830,12 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_NV_device_generated_commands ===
|
||||
class IndirectCommandsLayoutNV;
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
class CudaModuleNV;
|
||||
class CudaFunctionNV;
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
class BufferCollectionFUCHSIA;
|
||||
@ -2223,6 +2243,27 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
|
||||
using UniqueIndirectCommandsLayoutNV = UniqueHandle<IndirectCommandsLayoutNV, VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>;
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
template <typename Dispatch>
|
||||
class UniqueHandleTraits<CudaModuleNV, Dispatch>
|
||||
{
|
||||
public:
|
||||
using deleter = ObjectDestroy<Device, Dispatch>;
|
||||
};
|
||||
|
||||
using UniqueCudaModuleNV = UniqueHandle<CudaModuleNV, VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>;
|
||||
|
||||
template <typename Dispatch>
|
||||
class UniqueHandleTraits<CudaFunctionNV, Dispatch>
|
||||
{
|
||||
public:
|
||||
using deleter = ObjectDestroy<Device, Dispatch>;
|
||||
};
|
||||
|
||||
using UniqueCudaFunctionNV = UniqueHandle<CudaFunctionNV, VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
template <typename Dispatch>
|
||||
@ -5822,6 +5863,19 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
void encodeVideoKHR( const VULKAN_HPP_NAMESPACE::VideoEncodeInfoKHR & encodeInfo,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void cudaLaunchKernelNV( const VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV * pLaunchInfo,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void cudaLaunchKernelNV( const VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV & launchInfo,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
//=== VK_KHR_synchronization2 ===
|
||||
@ -7254,6 +7308,172 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR bool value = true;
|
||||
};
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
class CudaFunctionNV
|
||||
{
|
||||
public:
|
||||
using CType = VkCudaFunctionNV;
|
||||
using NativeType = VkCudaFunctionNV;
|
||||
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::ObjectType objectType = VULKAN_HPP_NAMESPACE::ObjectType::eCudaFunctionNV;
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT debugReportObjectType =
|
||||
VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eUnknown;
|
||||
|
||||
public:
|
||||
VULKAN_HPP_CONSTEXPR CudaFunctionNV() = default;
|
||||
|
||||
VULKAN_HPP_CONSTEXPR CudaFunctionNV( std::nullptr_t ) VULKAN_HPP_NOEXCEPT {}
|
||||
|
||||
VULKAN_HPP_TYPESAFE_EXPLICIT CudaFunctionNV( VkCudaFunctionNV cudaFunctionNV ) VULKAN_HPP_NOEXCEPT : m_cudaFunctionNV( cudaFunctionNV ) {}
|
||||
|
||||
# if defined( VULKAN_HPP_TYPESAFE_CONVERSION )
|
||||
CudaFunctionNV & operator=( VkCudaFunctionNV cudaFunctionNV ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
m_cudaFunctionNV = cudaFunctionNV;
|
||||
return *this;
|
||||
}
|
||||
# endif
|
||||
|
||||
CudaFunctionNV & operator=( std::nullptr_t ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
m_cudaFunctionNV = {};
|
||||
return *this;
|
||||
}
|
||||
|
||||
# if defined( VULKAN_HPP_HAS_SPACESHIP_OPERATOR )
|
||||
auto operator<=>( CudaFunctionNV const & ) const = default;
|
||||
# else
|
||||
bool operator==( CudaFunctionNV const & rhs ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaFunctionNV == rhs.m_cudaFunctionNV;
|
||||
}
|
||||
|
||||
bool operator!=( CudaFunctionNV const & rhs ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaFunctionNV != rhs.m_cudaFunctionNV;
|
||||
}
|
||||
|
||||
bool operator<( CudaFunctionNV const & rhs ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaFunctionNV < rhs.m_cudaFunctionNV;
|
||||
}
|
||||
# endif
|
||||
|
||||
VULKAN_HPP_TYPESAFE_EXPLICIT operator VkCudaFunctionNV() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaFunctionNV;
|
||||
}
|
||||
|
||||
explicit operator bool() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaFunctionNV != VK_NULL_HANDLE;
|
||||
}
|
||||
|
||||
bool operator!() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaFunctionNV == VK_NULL_HANDLE;
|
||||
}
|
||||
|
||||
private:
|
||||
VkCudaFunctionNV m_cudaFunctionNV = {};
|
||||
};
|
||||
|
||||
template <>
|
||||
struct CppType<VULKAN_HPP_NAMESPACE::ObjectType, VULKAN_HPP_NAMESPACE::ObjectType::eCudaFunctionNV>
|
||||
{
|
||||
using Type = VULKAN_HPP_NAMESPACE::CudaFunctionNV;
|
||||
};
|
||||
|
||||
template <>
|
||||
struct isVulkanHandleType<VULKAN_HPP_NAMESPACE::CudaFunctionNV>
|
||||
{
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR bool value = true;
|
||||
};
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
class CudaModuleNV
|
||||
{
|
||||
public:
|
||||
using CType = VkCudaModuleNV;
|
||||
using NativeType = VkCudaModuleNV;
|
||||
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::ObjectType objectType = VULKAN_HPP_NAMESPACE::ObjectType::eCudaModuleNV;
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT debugReportObjectType =
|
||||
VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eUnknown;
|
||||
|
||||
public:
|
||||
VULKAN_HPP_CONSTEXPR CudaModuleNV() = default;
|
||||
|
||||
VULKAN_HPP_CONSTEXPR CudaModuleNV( std::nullptr_t ) VULKAN_HPP_NOEXCEPT {}
|
||||
|
||||
VULKAN_HPP_TYPESAFE_EXPLICIT CudaModuleNV( VkCudaModuleNV cudaModuleNV ) VULKAN_HPP_NOEXCEPT : m_cudaModuleNV( cudaModuleNV ) {}
|
||||
|
||||
# if defined( VULKAN_HPP_TYPESAFE_CONVERSION )
|
||||
CudaModuleNV & operator=( VkCudaModuleNV cudaModuleNV ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
m_cudaModuleNV = cudaModuleNV;
|
||||
return *this;
|
||||
}
|
||||
# endif
|
||||
|
||||
CudaModuleNV & operator=( std::nullptr_t ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
m_cudaModuleNV = {};
|
||||
return *this;
|
||||
}
|
||||
|
||||
# if defined( VULKAN_HPP_HAS_SPACESHIP_OPERATOR )
|
||||
auto operator<=>( CudaModuleNV const & ) const = default;
|
||||
# else
|
||||
bool operator==( CudaModuleNV const & rhs ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaModuleNV == rhs.m_cudaModuleNV;
|
||||
}
|
||||
|
||||
bool operator!=( CudaModuleNV const & rhs ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaModuleNV != rhs.m_cudaModuleNV;
|
||||
}
|
||||
|
||||
bool operator<( CudaModuleNV const & rhs ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaModuleNV < rhs.m_cudaModuleNV;
|
||||
}
|
||||
# endif
|
||||
|
||||
VULKAN_HPP_TYPESAFE_EXPLICIT operator VkCudaModuleNV() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaModuleNV;
|
||||
}
|
||||
|
||||
explicit operator bool() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaModuleNV != VK_NULL_HANDLE;
|
||||
}
|
||||
|
||||
bool operator!() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_cudaModuleNV == VK_NULL_HANDLE;
|
||||
}
|
||||
|
||||
private:
|
||||
VkCudaModuleNV m_cudaModuleNV = {};
|
||||
};
|
||||
|
||||
template <>
|
||||
struct CppType<VULKAN_HPP_NAMESPACE::ObjectType, VULKAN_HPP_NAMESPACE::ObjectType::eCudaModuleNV>
|
||||
{
|
||||
using Type = VULKAN_HPP_NAMESPACE::CudaModuleNV;
|
||||
};
|
||||
|
||||
template <>
|
||||
struct isVulkanHandleType<VULKAN_HPP_NAMESPACE::CudaModuleNV>
|
||||
{
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR bool value = true;
|
||||
};
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
class DescriptorPool
|
||||
{
|
||||
public:
|
||||
@ -12563,6 +12783,111 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
VULKAN_HPP_NODISCARD Result createCudaModuleNV( const VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV * pCreateInfo,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
VULKAN_HPP_NAMESPACE::CudaModuleNV * pModule,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
VULKAN_HPP_NODISCARD typename ResultValueType<VULKAN_HPP_NAMESPACE::CudaModuleNV>::type
|
||||
createCudaModuleNV( const VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV & createInfo,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator VULKAN_HPP_DEFAULT_ARGUMENT_NULLPTR_ASSIGNMENT,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const;
|
||||
# ifndef VULKAN_HPP_NO_SMART_HANDLE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
VULKAN_HPP_NODISCARD typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::CudaModuleNV, Dispatch>>::type
|
||||
createCudaModuleNVUnique( const VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV & createInfo,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator VULKAN_HPP_DEFAULT_ARGUMENT_NULLPTR_ASSIGNMENT,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const;
|
||||
# endif /* VULKAN_HPP_NO_SMART_HANDLE */
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
VULKAN_HPP_NODISCARD Result getCudaModuleCacheNV( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
size_t * pCacheSize,
|
||||
void * pCacheData,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Uint8_tAllocator = std::allocator<uint8_t>, typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
VULKAN_HPP_NODISCARD typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type
|
||||
getCudaModuleCacheNV( VULKAN_HPP_NAMESPACE::CudaModuleNV module, Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const;
|
||||
template <typename Uint8_tAllocator = std::allocator<uint8_t>,
|
||||
typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE,
|
||||
typename B1 = Uint8_tAllocator,
|
||||
typename std::enable_if<std::is_same<typename B1::value_type, uint8_t>::value, int>::type = 0>
|
||||
VULKAN_HPP_NODISCARD typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type getCudaModuleCacheNV(
|
||||
VULKAN_HPP_NAMESPACE::CudaModuleNV module, Uint8_tAllocator & uint8_tAllocator, Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const;
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
VULKAN_HPP_NODISCARD Result createCudaFunctionNV( const VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV * pCreateInfo,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
VULKAN_HPP_NAMESPACE::CudaFunctionNV * pFunction,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
VULKAN_HPP_NODISCARD typename ResultValueType<VULKAN_HPP_NAMESPACE::CudaFunctionNV>::type
|
||||
createCudaFunctionNV( const VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV & createInfo,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator VULKAN_HPP_DEFAULT_ARGUMENT_NULLPTR_ASSIGNMENT,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const;
|
||||
# ifndef VULKAN_HPP_NO_SMART_HANDLE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
VULKAN_HPP_NODISCARD typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::CudaFunctionNV, Dispatch>>::type
|
||||
createCudaFunctionNVUnique( const VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV & createInfo,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator VULKAN_HPP_DEFAULT_ARGUMENT_NULLPTR_ASSIGNMENT,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const;
|
||||
# endif /* VULKAN_HPP_NO_SMART_HANDLE */
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void destroyCudaModuleNV( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void destroyCudaModuleNV( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator VULKAN_HPP_DEFAULT_ARGUMENT_NULLPTR_ASSIGNMENT,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void destroy( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void destroy( VULKAN_HPP_NAMESPACE::CudaModuleNV module,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator VULKAN_HPP_DEFAULT_ARGUMENT_NULLPTR_ASSIGNMENT,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void destroyCudaFunctionNV( VULKAN_HPP_NAMESPACE::CudaFunctionNV function,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void destroyCudaFunctionNV( VULKAN_HPP_NAMESPACE::CudaFunctionNV function,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator VULKAN_HPP_DEFAULT_ARGUMENT_NULLPTR_ASSIGNMENT,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void destroy( VULKAN_HPP_NAMESPACE::CudaFunctionNV function,
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * pAllocator,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE
|
||||
template <typename Dispatch = VULKAN_HPP_DEFAULT_DISPATCHER_TYPE>
|
||||
void destroy( VULKAN_HPP_NAMESPACE::CudaFunctionNV function,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator VULKAN_HPP_DEFAULT_ARGUMENT_NULLPTR_ASSIGNMENT,
|
||||
Dispatch const & d VULKAN_HPP_DEFAULT_DISPATCHER_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /* VULKAN_HPP_DISABLE_ENHANCED_MODE */
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
|
||||
|
@ -457,6 +457,28 @@ namespace std
|
||||
}
|
||||
};
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::CudaModuleNV>
|
||||
{
|
||||
std::size_t operator()( VULKAN_HPP_NAMESPACE::CudaModuleNV const & cudaModuleNV ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return std::hash<VkCudaModuleNV>{}( static_cast<VkCudaModuleNV>( cudaModuleNV ) );
|
||||
}
|
||||
};
|
||||
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::CudaFunctionNV>
|
||||
{
|
||||
std::size_t operator()( VULKAN_HPP_NAMESPACE::CudaFunctionNV const & cudaFunctionNV ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return std::hash<VkCudaFunctionNV>{}( static_cast<VkCudaFunctionNV>( cudaFunctionNV ) );
|
||||
}
|
||||
};
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
#if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
|
||||
@ -2732,6 +2754,67 @@ namespace std
|
||||
}
|
||||
};
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV>
|
||||
{
|
||||
std::size_t operator()( VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV const & cudaFunctionCreateInfoNV ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::size_t seed = 0;
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaFunctionCreateInfoNV.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaFunctionCreateInfoNV.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaFunctionCreateInfoNV.module );
|
||||
for ( const char * p = cudaFunctionCreateInfoNV.pName; *p != '\0'; ++p )
|
||||
{
|
||||
VULKAN_HPP_HASH_COMBINE( seed, *p );
|
||||
}
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV>
|
||||
{
|
||||
std::size_t operator()( VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV const & cudaLaunchInfoNV ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::size_t seed = 0;
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.function );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.gridDimX );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.gridDimY );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.gridDimZ );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.blockDimX );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.blockDimY );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.blockDimZ );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.sharedMemBytes );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.paramCount );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.pParams );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.extraCount );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaLaunchInfoNV.pExtras );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV>
|
||||
{
|
||||
std::size_t operator()( VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV const & cudaModuleCreateInfoNV ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::size_t seed = 0;
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaModuleCreateInfoNV.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaModuleCreateInfoNV.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaModuleCreateInfoNV.dataSize );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, cudaModuleCreateInfoNV.pData );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_WIN32_KHR )
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::D3D12FenceSubmitInfoKHR>
|
||||
@ -3901,6 +3984,20 @@ namespace std
|
||||
}
|
||||
};
|
||||
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::DeviceQueueShaderCoreControlCreateInfoARM>
|
||||
{
|
||||
std::size_t
|
||||
operator()( VULKAN_HPP_NAMESPACE::DeviceQueueShaderCoreControlCreateInfoARM const & deviceQueueShaderCoreControlCreateInfoARM ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::size_t seed = 0;
|
||||
VULKAN_HPP_HASH_COMBINE( seed, deviceQueueShaderCoreControlCreateInfoARM.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, deviceQueueShaderCoreControlCreateInfoARM.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, deviceQueueShaderCoreControlCreateInfoARM.shaderCoreCount );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::DirectDriverLoadingInfoLUNARG>
|
||||
{
|
||||
@ -7652,6 +7749,39 @@ namespace std
|
||||
}
|
||||
};
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchFeaturesNV>
|
||||
{
|
||||
std::size_t
|
||||
operator()( VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchFeaturesNV const & physicalDeviceCudaKernelLaunchFeaturesNV ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::size_t seed = 0;
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceCudaKernelLaunchFeaturesNV.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceCudaKernelLaunchFeaturesNV.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceCudaKernelLaunchFeaturesNV.cudaKernelLaunchFeatures );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchPropertiesNV>
|
||||
{
|
||||
std::size_t operator()( VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchPropertiesNV const & physicalDeviceCudaKernelLaunchPropertiesNV ) const
|
||||
VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::size_t seed = 0;
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceCudaKernelLaunchPropertiesNV.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceCudaKernelLaunchPropertiesNV.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceCudaKernelLaunchPropertiesNV.computeCapabilityMinor );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceCudaKernelLaunchPropertiesNV.computeCapabilityMajor );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::PhysicalDeviceCustomBorderColorFeaturesEXT>
|
||||
{
|
||||
@ -10575,6 +10705,34 @@ namespace std
|
||||
}
|
||||
};
|
||||
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsFeaturesARM>
|
||||
{
|
||||
std::size_t operator()( VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsFeaturesARM const & physicalDeviceSchedulingControlsFeaturesARM ) const
|
||||
VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::size_t seed = 0;
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceSchedulingControlsFeaturesARM.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceSchedulingControlsFeaturesARM.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceSchedulingControlsFeaturesARM.schedulingControls );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsPropertiesARM>
|
||||
{
|
||||
std::size_t operator()( VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsPropertiesARM const & physicalDeviceSchedulingControlsPropertiesARM ) const
|
||||
VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::size_t seed = 0;
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceSchedulingControlsPropertiesARM.sType );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceSchedulingControlsPropertiesARM.pNext );
|
||||
VULKAN_HPP_HASH_COMBINE( seed, physicalDeviceSchedulingControlsPropertiesARM.schedulingControlsFlags );
|
||||
return seed;
|
||||
}
|
||||
};
|
||||
|
||||
template <>
|
||||
struct hash<VULKAN_HPP_NAMESPACE::PhysicalDeviceSeparateDepthStencilLayoutsFeatures>
|
||||
{
|
||||
|
@ -1394,6 +1394,16 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
vkCmdEncodeVideoKHR = PFN_vkCmdEncodeVideoKHR( vkGetDeviceProcAddr( device, "vkCmdEncodeVideoKHR" ) );
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
vkCreateCudaModuleNV = PFN_vkCreateCudaModuleNV( vkGetDeviceProcAddr( device, "vkCreateCudaModuleNV" ) );
|
||||
vkGetCudaModuleCacheNV = PFN_vkGetCudaModuleCacheNV( vkGetDeviceProcAddr( device, "vkGetCudaModuleCacheNV" ) );
|
||||
vkCreateCudaFunctionNV = PFN_vkCreateCudaFunctionNV( vkGetDeviceProcAddr( device, "vkCreateCudaFunctionNV" ) );
|
||||
vkDestroyCudaModuleNV = PFN_vkDestroyCudaModuleNV( vkGetDeviceProcAddr( device, "vkDestroyCudaModuleNV" ) );
|
||||
vkDestroyCudaFunctionNV = PFN_vkDestroyCudaFunctionNV( vkGetDeviceProcAddr( device, "vkDestroyCudaFunctionNV" ) );
|
||||
vkCmdCudaLaunchKernelNV = PFN_vkCmdCudaLaunchKernelNV( vkGetDeviceProcAddr( device, "vkCmdCudaLaunchKernelNV" ) );
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
vkExportMetalObjectsEXT = PFN_vkExportMetalObjectsEXT( vkGetDeviceProcAddr( device, "vkExportMetalObjectsEXT" ) );
|
||||
@ -2295,6 +2305,23 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
PFN_dummy vkCmdEncodeVideoKHR_placeholder = 0;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
PFN_vkCreateCudaModuleNV vkCreateCudaModuleNV = 0;
|
||||
PFN_vkGetCudaModuleCacheNV vkGetCudaModuleCacheNV = 0;
|
||||
PFN_vkCreateCudaFunctionNV vkCreateCudaFunctionNV = 0;
|
||||
PFN_vkDestroyCudaModuleNV vkDestroyCudaModuleNV = 0;
|
||||
PFN_vkDestroyCudaFunctionNV vkDestroyCudaFunctionNV = 0;
|
||||
PFN_vkCmdCudaLaunchKernelNV vkCmdCudaLaunchKernelNV = 0;
|
||||
# else
|
||||
PFN_dummy vkCreateCudaModuleNV_placeholder = 0;
|
||||
PFN_dummy vkGetCudaModuleCacheNV_placeholder = 0;
|
||||
PFN_dummy vkCreateCudaFunctionNV_placeholder = 0;
|
||||
PFN_dummy vkDestroyCudaModuleNV_placeholder = 0;
|
||||
PFN_dummy vkDestroyCudaFunctionNV_placeholder = 0;
|
||||
PFN_dummy vkCmdCudaLaunchKernelNV_placeholder = 0;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
PFN_vkExportMetalObjectsEXT vkExportMetalObjectsEXT = 0;
|
||||
@ -2612,6 +2639,12 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
//=== VK_NV_device_generated_commands ===
|
||||
class IndirectCommandsLayoutNV;
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
class CudaModuleNV;
|
||||
class CudaFunctionNV;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
class BufferCollectionFUCHSIA;
|
||||
@ -4195,6 +4228,18 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
getEncodedVideoSessionParametersKHR( const VULKAN_HPP_NAMESPACE::VideoEncodeSessionParametersGetInfoKHR & videoSessionParametersInfo ) const;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_RAII_NAMESPACE::CudaModuleNV
|
||||
createCudaModuleNV( VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr ) const;
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_RAII_NAMESPACE::CudaFunctionNV
|
||||
createCudaFunctionNV( VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr ) const;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
|
||||
@ -5915,6 +5960,12 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
void encodeVideoKHR( const VULKAN_HPP_NAMESPACE::VideoEncodeInfoKHR & encodeInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
void cudaLaunchKernelNV( const VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV & launchInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
//=== VK_KHR_synchronization2 ===
|
||||
|
||||
void setEvent2KHR( VULKAN_HPP_NAMESPACE::Event event, const VULKAN_HPP_NAMESPACE::DependencyInfo & dependencyInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
@ -6460,6 +6511,260 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * m_dispatcher = nullptr;
|
||||
};
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
class CudaFunctionNV
|
||||
{
|
||||
public:
|
||||
using CType = VkCudaFunctionNV;
|
||||
using CppType = vk::CudaFunctionNV;
|
||||
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::ObjectType objectType = VULKAN_HPP_NAMESPACE::ObjectType::eCudaFunctionNV;
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT debugReportObjectType =
|
||||
VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eUnknown;
|
||||
|
||||
public:
|
||||
CudaFunctionNV( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::Device const & device,
|
||||
VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr )
|
||||
: m_device( *device )
|
||||
, m_allocator( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) )
|
||||
, m_dispatcher( device.getDispatcher() )
|
||||
{
|
||||
VULKAN_HPP_NAMESPACE::Result result = static_cast<VULKAN_HPP_NAMESPACE::Result>(
|
||||
device.getDispatcher()->vkCreateCudaFunctionNV( static_cast<VkDevice>( *device ),
|
||||
reinterpret_cast<const VkCudaFunctionCreateInfoNV *>( &createInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( m_allocator ),
|
||||
reinterpret_cast<VkCudaFunctionNV *>( &m_function ) ) );
|
||||
if ( result != VULKAN_HPP_NAMESPACE::Result::eSuccess )
|
||||
{
|
||||
detail::throwResultException( result, "vkCreateCudaFunctionNV" );
|
||||
}
|
||||
}
|
||||
|
||||
CudaFunctionNV( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::Device const & device,
|
||||
VkCudaFunctionNV function,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr )
|
||||
: m_device( *device )
|
||||
, m_function( function )
|
||||
, m_allocator( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) )
|
||||
, m_dispatcher( device.getDispatcher() )
|
||||
{
|
||||
}
|
||||
|
||||
CudaFunctionNV( std::nullptr_t ) {}
|
||||
|
||||
~CudaFunctionNV()
|
||||
{
|
||||
clear();
|
||||
}
|
||||
|
||||
CudaFunctionNV() = delete;
|
||||
CudaFunctionNV( CudaFunctionNV const & ) = delete;
|
||||
|
||||
CudaFunctionNV( CudaFunctionNV && rhs ) VULKAN_HPP_NOEXCEPT
|
||||
: m_device( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( rhs.m_device, {} ) )
|
||||
, m_function( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( rhs.m_function, {} ) )
|
||||
, m_allocator( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( rhs.m_allocator, {} ) )
|
||||
, m_dispatcher( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( rhs.m_dispatcher, nullptr ) )
|
||||
{
|
||||
}
|
||||
|
||||
CudaFunctionNV & operator=( CudaFunctionNV const & ) = delete;
|
||||
|
||||
CudaFunctionNV & operator=( CudaFunctionNV && rhs ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
if ( this != &rhs )
|
||||
{
|
||||
std::swap( m_device, rhs.m_device );
|
||||
std::swap( m_function, rhs.m_function );
|
||||
std::swap( m_allocator, rhs.m_allocator );
|
||||
std::swap( m_dispatcher, rhs.m_dispatcher );
|
||||
}
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::CudaFunctionNV const & operator*() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_function;
|
||||
}
|
||||
|
||||
void clear() VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
if ( m_function )
|
||||
{
|
||||
getDispatcher()->vkDestroyCudaFunctionNV(
|
||||
static_cast<VkDevice>( m_device ), static_cast<VkCudaFunctionNV>( m_function ), reinterpret_cast<const VkAllocationCallbacks *>( m_allocator ) );
|
||||
}
|
||||
m_device = nullptr;
|
||||
m_function = nullptr;
|
||||
m_allocator = nullptr;
|
||||
m_dispatcher = nullptr;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::CudaFunctionNV release()
|
||||
{
|
||||
m_device = nullptr;
|
||||
m_allocator = nullptr;
|
||||
m_dispatcher = nullptr;
|
||||
return VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( m_function, nullptr );
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::Device getDevice() const
|
||||
{
|
||||
return m_device;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * getDispatcher() const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( m_dispatcher->getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
return m_dispatcher;
|
||||
}
|
||||
|
||||
void swap( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::CudaFunctionNV & rhs ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::swap( m_device, rhs.m_device );
|
||||
std::swap( m_function, rhs.m_function );
|
||||
std::swap( m_allocator, rhs.m_allocator );
|
||||
std::swap( m_dispatcher, rhs.m_dispatcher );
|
||||
}
|
||||
|
||||
private:
|
||||
VULKAN_HPP_NAMESPACE::Device m_device = {};
|
||||
VULKAN_HPP_NAMESPACE::CudaFunctionNV m_function = {};
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * m_allocator = {};
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * m_dispatcher = nullptr;
|
||||
};
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
class CudaModuleNV
|
||||
{
|
||||
public:
|
||||
using CType = VkCudaModuleNV;
|
||||
using CppType = vk::CudaModuleNV;
|
||||
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::ObjectType objectType = VULKAN_HPP_NAMESPACE::ObjectType::eCudaModuleNV;
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT debugReportObjectType =
|
||||
VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eUnknown;
|
||||
|
||||
public:
|
||||
CudaModuleNV( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::Device const & device,
|
||||
VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr )
|
||||
: m_device( *device )
|
||||
, m_allocator( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) )
|
||||
, m_dispatcher( device.getDispatcher() )
|
||||
{
|
||||
VULKAN_HPP_NAMESPACE::Result result = static_cast<VULKAN_HPP_NAMESPACE::Result>(
|
||||
device.getDispatcher()->vkCreateCudaModuleNV( static_cast<VkDevice>( *device ),
|
||||
reinterpret_cast<const VkCudaModuleCreateInfoNV *>( &createInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( m_allocator ),
|
||||
reinterpret_cast<VkCudaModuleNV *>( &m_module ) ) );
|
||||
if ( result != VULKAN_HPP_NAMESPACE::Result::eSuccess )
|
||||
{
|
||||
detail::throwResultException( result, "vkCreateCudaModuleNV" );
|
||||
}
|
||||
}
|
||||
|
||||
CudaModuleNV( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::Device const & device,
|
||||
VkCudaModuleNV module,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr )
|
||||
: m_device( *device )
|
||||
, m_module( module )
|
||||
, m_allocator( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) )
|
||||
, m_dispatcher( device.getDispatcher() )
|
||||
{
|
||||
}
|
||||
|
||||
CudaModuleNV( std::nullptr_t ) {}
|
||||
|
||||
~CudaModuleNV()
|
||||
{
|
||||
clear();
|
||||
}
|
||||
|
||||
CudaModuleNV() = delete;
|
||||
CudaModuleNV( CudaModuleNV const & ) = delete;
|
||||
|
||||
CudaModuleNV( CudaModuleNV && rhs ) VULKAN_HPP_NOEXCEPT
|
||||
: m_device( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( rhs.m_device, {} ) )
|
||||
, m_module( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( rhs.m_module, {} ) )
|
||||
, m_allocator( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( rhs.m_allocator, {} ) )
|
||||
, m_dispatcher( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( rhs.m_dispatcher, nullptr ) )
|
||||
{
|
||||
}
|
||||
|
||||
CudaModuleNV & operator=( CudaModuleNV const & ) = delete;
|
||||
|
||||
CudaModuleNV & operator=( CudaModuleNV && rhs ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
if ( this != &rhs )
|
||||
{
|
||||
std::swap( m_device, rhs.m_device );
|
||||
std::swap( m_module, rhs.m_module );
|
||||
std::swap( m_allocator, rhs.m_allocator );
|
||||
std::swap( m_dispatcher, rhs.m_dispatcher );
|
||||
}
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::CudaModuleNV const & operator*() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_module;
|
||||
}
|
||||
|
||||
void clear() VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
if ( m_module )
|
||||
{
|
||||
getDispatcher()->vkDestroyCudaModuleNV(
|
||||
static_cast<VkDevice>( m_device ), static_cast<VkCudaModuleNV>( m_module ), reinterpret_cast<const VkAllocationCallbacks *>( m_allocator ) );
|
||||
}
|
||||
m_device = nullptr;
|
||||
m_module = nullptr;
|
||||
m_allocator = nullptr;
|
||||
m_dispatcher = nullptr;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::CudaModuleNV release()
|
||||
{
|
||||
m_device = nullptr;
|
||||
m_allocator = nullptr;
|
||||
m_dispatcher = nullptr;
|
||||
return VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::exchange( m_module, nullptr );
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::Device getDevice() const
|
||||
{
|
||||
return m_device;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * getDispatcher() const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( m_dispatcher->getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
return m_dispatcher;
|
||||
}
|
||||
|
||||
void swap( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::CudaModuleNV & rhs ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::swap( m_device, rhs.m_device );
|
||||
std::swap( m_module, rhs.m_module );
|
||||
std::swap( m_allocator, rhs.m_allocator );
|
||||
std::swap( m_dispatcher, rhs.m_dispatcher );
|
||||
}
|
||||
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
VULKAN_HPP_NODISCARD std::vector<uint8_t> getCache() const;
|
||||
|
||||
private:
|
||||
VULKAN_HPP_NAMESPACE::Device m_device = {};
|
||||
VULKAN_HPP_NAMESPACE::CudaModuleNV m_module = {};
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * m_allocator = {};
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * m_dispatcher = nullptr;
|
||||
};
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
class DebugReportCallbackEXT
|
||||
{
|
||||
public:
|
||||
@ -19049,6 +19354,57 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
}
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_RAII_NAMESPACE::CudaModuleNV
|
||||
Device::createCudaModuleNV( VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator ) const
|
||||
{
|
||||
return VULKAN_HPP_RAII_NAMESPACE::CudaModuleNV( *this, createInfo, allocator );
|
||||
}
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<uint8_t> CudaModuleNV::getCache() const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkGetCudaModuleCacheNV && "Function <vkGetCudaModuleCacheNV> requires <VK_NV_cuda_kernel_launch>" );
|
||||
|
||||
std::vector<uint8_t> cacheData;
|
||||
size_t cacheSize;
|
||||
VkResult result;
|
||||
do
|
||||
{
|
||||
result = getDispatcher()->vkGetCudaModuleCacheNV( static_cast<VkDevice>( m_device ), static_cast<VkCudaModuleNV>( m_module ), &cacheSize, nullptr );
|
||||
if ( ( result == VK_SUCCESS ) && cacheSize )
|
||||
{
|
||||
cacheData.resize( cacheSize );
|
||||
result = getDispatcher()->vkGetCudaModuleCacheNV(
|
||||
static_cast<VkDevice>( m_device ), static_cast<VkCudaModuleNV>( m_module ), &cacheSize, reinterpret_cast<void *>( cacheData.data() ) );
|
||||
}
|
||||
} while ( result == VK_INCOMPLETE );
|
||||
resultCheck( static_cast<VULKAN_HPP_NAMESPACE::Result>( result ), VULKAN_HPP_NAMESPACE_STRING "::CudaModuleNV::getCache" );
|
||||
VULKAN_HPP_ASSERT( cacheSize <= cacheData.size() );
|
||||
if ( cacheSize < cacheData.size() )
|
||||
{
|
||||
cacheData.resize( cacheSize );
|
||||
}
|
||||
return cacheData;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_RAII_NAMESPACE::CudaFunctionNV
|
||||
Device::createCudaFunctionNV( VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator ) const
|
||||
{
|
||||
return VULKAN_HPP_RAII_NAMESPACE::CudaFunctionNV( *this, createInfo, allocator );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::cudaLaunchKernelNV( const VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV & launchInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdCudaLaunchKernelNV && "Function <vkCmdCudaLaunchKernelNV> requires <VK_NV_cuda_kernel_launch>" );
|
||||
|
||||
getDispatcher()->vkCmdCudaLaunchKernelNV( static_cast<VkCommandBuffer>( m_commandBuffer ), reinterpret_cast<const VkCudaLaunchInfoNV *>( &launchInfo ) );
|
||||
}
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
//=== VK_EXT_metal_objects ===
|
||||
|
||||
|
@ -849,6 +849,29 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
|
||||
using SharedIndirectCommandsLayoutNV = SharedHandle<IndirectCommandsLayoutNV>;
|
||||
|
||||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
template <>
|
||||
class SharedHandleTraits<CudaModuleNV>
|
||||
{
|
||||
public:
|
||||
using DestructorType = Device;
|
||||
using deleter = ObjectDestroyShared<CudaModuleNV>;
|
||||
};
|
||||
|
||||
using SharedCudaModuleNV = SharedHandle<CudaModuleNV>;
|
||||
|
||||
template <>
|
||||
class SharedHandleTraits<CudaFunctionNV>
|
||||
{
|
||||
public:
|
||||
using DestructorType = Device;
|
||||
using deleter = ObjectDestroyShared<CudaFunctionNV>;
|
||||
};
|
||||
|
||||
using SharedCudaFunctionNV = SharedHandle<CudaFunctionNV>;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
# if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
//=== VK_FUCHSIA_buffer_collection ===
|
||||
template <>
|
||||
|
@ -5122,6 +5122,48 @@ VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::DeviceDi
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::DeviceDiagnosticsConfigCreateInfoNV>::value,
|
||||
"DeviceDiagnosticsConfigCreateInfoNV is not nothrow_move_constructible!" );
|
||||
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_NV_cuda_kernel_launch ===
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::CudaModuleNV ) == sizeof( VkCudaModuleNV ), "handle and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::CudaModuleNV>::value, "CudaModuleNV is not nothrow_move_constructible!" );
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::CudaFunctionNV ) == sizeof( VkCudaFunctionNV ), "handle and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::CudaFunctionNV>::value,
|
||||
"CudaFunctionNV is not nothrow_move_constructible!" );
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV ) == sizeof( VkCudaModuleCreateInfoNV ),
|
||||
"struct and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV>::value, "struct wrapper is not a standard layout!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::CudaModuleCreateInfoNV>::value,
|
||||
"CudaModuleCreateInfoNV is not nothrow_move_constructible!" );
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV ) == sizeof( VkCudaFunctionCreateInfoNV ),
|
||||
"struct and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV>::value, "struct wrapper is not a standard layout!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::CudaFunctionCreateInfoNV>::value,
|
||||
"CudaFunctionCreateInfoNV is not nothrow_move_constructible!" );
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV ) == sizeof( VkCudaLaunchInfoNV ), "struct and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV>::value, "struct wrapper is not a standard layout!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::CudaLaunchInfoNV>::value,
|
||||
"CudaLaunchInfoNV is not nothrow_move_constructible!" );
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchFeaturesNV ) == sizeof( VkPhysicalDeviceCudaKernelLaunchFeaturesNV ),
|
||||
"struct and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchFeaturesNV>::value,
|
||||
"struct wrapper is not a standard layout!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchFeaturesNV>::value,
|
||||
"PhysicalDeviceCudaKernelLaunchFeaturesNV is not nothrow_move_constructible!" );
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchPropertiesNV ) == sizeof( VkPhysicalDeviceCudaKernelLaunchPropertiesNV ),
|
||||
"struct and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchPropertiesNV>::value,
|
||||
"struct wrapper is not a standard layout!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::PhysicalDeviceCudaKernelLaunchPropertiesNV>::value,
|
||||
"PhysicalDeviceCudaKernelLaunchPropertiesNV is not nothrow_move_constructible!" );
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
//=== VK_NV_low_latency ===
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::QueryLowLatencySupportNV ) == sizeof( VkQueryLowLatencySupportNV ),
|
||||
@ -6207,6 +6249,31 @@ VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::Physical
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::PhysicalDeviceShaderCorePropertiesARM>::value,
|
||||
"PhysicalDeviceShaderCorePropertiesARM is not nothrow_move_constructible!" );
|
||||
|
||||
//=== VK_ARM_scheduling_controls ===
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::DeviceQueueShaderCoreControlCreateInfoARM ) == sizeof( VkDeviceQueueShaderCoreControlCreateInfoARM ),
|
||||
"struct and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::DeviceQueueShaderCoreControlCreateInfoARM>::value,
|
||||
"struct wrapper is not a standard layout!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::DeviceQueueShaderCoreControlCreateInfoARM>::value,
|
||||
"DeviceQueueShaderCoreControlCreateInfoARM is not nothrow_move_constructible!" );
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsFeaturesARM ) ==
|
||||
sizeof( VkPhysicalDeviceSchedulingControlsFeaturesARM ),
|
||||
"struct and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsFeaturesARM>::value,
|
||||
"struct wrapper is not a standard layout!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsFeaturesARM>::value,
|
||||
"PhysicalDeviceSchedulingControlsFeaturesARM is not nothrow_move_constructible!" );
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsPropertiesARM ) ==
|
||||
sizeof( VkPhysicalDeviceSchedulingControlsPropertiesARM ),
|
||||
"struct and wrapper have different size!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_standard_layout<VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsPropertiesARM>::value,
|
||||
"struct wrapper is not a standard layout!" );
|
||||
VULKAN_HPP_STATIC_ASSERT( std::is_nothrow_move_constructible<VULKAN_HPP_NAMESPACE::PhysicalDeviceSchedulingControlsPropertiesARM>::value,
|
||||
"PhysicalDeviceSchedulingControlsPropertiesARM is not nothrow_move_constructible!" );
|
||||
|
||||
//=== VK_EXT_image_sliced_view_of_3d ===
|
||||
|
||||
VULKAN_HPP_STATIC_ASSERT( sizeof( VULKAN_HPP_NAMESPACE::PhysicalDeviceImageSlicedViewOf3DFeaturesEXT ) ==
|
||||
|
File diff suppressed because it is too large
Load Diff
@ -3249,6 +3249,20 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
return "{ " + result.substr( 0, result.size() - 3 ) + " }";
|
||||
}
|
||||
|
||||
//=== VK_ARM_scheduling_controls ===
|
||||
|
||||
VULKAN_HPP_INLINE std::string to_string( PhysicalDeviceSchedulingControlsFlagsARM value )
|
||||
{
|
||||
if ( !value )
|
||||
return "{}";
|
||||
|
||||
std::string result;
|
||||
if ( value & PhysicalDeviceSchedulingControlsFlagBitsARM::eShaderCoreCount )
|
||||
result += "ShaderCoreCount | ";
|
||||
|
||||
return "{ " + result.substr( 0, result.size() - 3 ) + " }";
|
||||
}
|
||||
|
||||
//=== VK_NV_memory_decompression ===
|
||||
|
||||
VULKAN_HPP_INLINE std::string to_string( MemoryDecompressionMethodFlagsNV value )
|
||||
@ -4243,6 +4257,13 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
case StructureType::ePhysicalDeviceDiagnosticsConfigFeaturesNV: return "PhysicalDeviceDiagnosticsConfigFeaturesNV";
|
||||
case StructureType::eDeviceDiagnosticsConfigCreateInfoNV: return "DeviceDiagnosticsConfigCreateInfoNV";
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
case StructureType::eCudaModuleCreateInfoNV: return "CudaModuleCreateInfoNV";
|
||||
case StructureType::eCudaFunctionCreateInfoNV: return "CudaFunctionCreateInfoNV";
|
||||
case StructureType::eCudaLaunchInfoNV: return "CudaLaunchInfoNV";
|
||||
case StructureType::ePhysicalDeviceCudaKernelLaunchFeaturesNV: return "PhysicalDeviceCudaKernelLaunchFeaturesNV";
|
||||
case StructureType::ePhysicalDeviceCudaKernelLaunchPropertiesNV: return "PhysicalDeviceCudaKernelLaunchPropertiesNV";
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
case StructureType::eQueryLowLatencySupportNV: return "QueryLowLatencySupportNV";
|
||||
#if defined( VK_USE_PLATFORM_METAL_EXT )
|
||||
case StructureType::eExportMetalObjectCreateInfoEXT: return "ExportMetalObjectCreateInfoEXT";
|
||||
@ -4380,6 +4401,9 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
case StructureType::eSamplerBorderColorComponentMappingCreateInfoEXT: return "SamplerBorderColorComponentMappingCreateInfoEXT";
|
||||
case StructureType::ePhysicalDevicePageableDeviceLocalMemoryFeaturesEXT: return "PhysicalDevicePageableDeviceLocalMemoryFeaturesEXT";
|
||||
case StructureType::ePhysicalDeviceShaderCorePropertiesARM: return "PhysicalDeviceShaderCorePropertiesARM";
|
||||
case StructureType::eDeviceQueueShaderCoreControlCreateInfoARM: return "DeviceQueueShaderCoreControlCreateInfoARM";
|
||||
case StructureType::ePhysicalDeviceSchedulingControlsFeaturesARM: return "PhysicalDeviceSchedulingControlsFeaturesARM";
|
||||
case StructureType::ePhysicalDeviceSchedulingControlsPropertiesARM: return "PhysicalDeviceSchedulingControlsPropertiesARM";
|
||||
case StructureType::ePhysicalDeviceImageSlicedViewOf3DFeaturesEXT: return "PhysicalDeviceImageSlicedViewOf3DFeaturesEXT";
|
||||
case StructureType::eImageViewSlicedCreateInfoEXT: return "ImageViewSlicedCreateInfoEXT";
|
||||
case StructureType::ePhysicalDeviceDescriptorSetHostMappingFeaturesVALVE: return "PhysicalDeviceDescriptorSetHostMappingFeaturesVALVE";
|
||||
@ -4554,6 +4578,10 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
case ObjectType::ePerformanceConfigurationINTEL: return "PerformanceConfigurationINTEL";
|
||||
case ObjectType::eDeferredOperationKHR: return "DeferredOperationKHR";
|
||||
case ObjectType::eIndirectCommandsLayoutNV: return "IndirectCommandsLayoutNV";
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
case ObjectType::eCudaModuleNV: return "CudaModuleNV";
|
||||
case ObjectType::eCudaFunctionNV: return "CudaFunctionNV";
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
#if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
case ObjectType::eBufferCollectionFUCHSIA: return "BufferCollectionFUCHSIA";
|
||||
#endif /*VK_USE_PLATFORM_FUCHSIA*/
|
||||
@ -6948,6 +6976,10 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
case DebugReportObjectTypeEXT::eCuFunctionNVX: return "CuFunctionNVX";
|
||||
case DebugReportObjectTypeEXT::eAccelerationStructureKHR: return "AccelerationStructureKHR";
|
||||
case DebugReportObjectTypeEXT::eAccelerationStructureNV: return "AccelerationStructureNV";
|
||||
#if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
case DebugReportObjectTypeEXT::eCudaModuleNV: return "CudaModuleNV";
|
||||
case DebugReportObjectTypeEXT::eCudaFunctionNV: return "CudaFunctionNV";
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
#if defined( VK_USE_PLATFORM_FUCHSIA )
|
||||
case DebugReportObjectTypeEXT::eBufferCollectionFUCHSIA: return "BufferCollectionFUCHSIA";
|
||||
#endif /*VK_USE_PLATFORM_FUCHSIA*/
|
||||
@ -8639,6 +8671,17 @@ namespace VULKAN_HPP_NAMESPACE
|
||||
}
|
||||
#endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
//=== VK_ARM_scheduling_controls ===
|
||||
|
||||
VULKAN_HPP_INLINE std::string to_string( PhysicalDeviceSchedulingControlsFlagBitsARM value )
|
||||
{
|
||||
switch ( value )
|
||||
{
|
||||
case PhysicalDeviceSchedulingControlsFlagBitsARM::eShaderCoreCount: return "ShaderCoreCount";
|
||||
default: return "invalid ( " + VULKAN_HPP_NAMESPACE::toHexString( static_cast<uint32_t>( value ) ) + " )";
|
||||
}
|
||||
}
|
||||
|
||||
//=== VK_NV_memory_decompression ===
|
||||
|
||||
VULKAN_HPP_INLINE std::string to_string( MemoryDecompressionMethodFlagBitsNV value )
|
||||
|
Loading…
Reference in New Issue
Block a user