| // Copyright 2015-2022 The Khronos Group Inc. |
| // |
| // SPDX-License-Identifier: Apache-2.0 OR MIT |
| // |
| |
| // This header is generated from the Khronos Vulkan XML API Registry. |
| |
| #ifndef VULKAN_FUNCS_HPP |
| #define VULKAN_FUNCS_HPP |
| |
| namespace VULKAN_HPP_NAMESPACE { |
| //=========================== |
| //=== COMMAND Definitions === |
| //=========================== |
| |
| //=== VK_VERSION_1_0 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result createInstance(const VULKAN_HPP_NAMESPACE::InstanceCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Instance *pInstance, |
| Dispatch const &d) VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateInstance(reinterpret_cast<const VkInstanceCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkInstance *>(pInstance))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Instance>::type createInstance( |
| const VULKAN_HPP_NAMESPACE::InstanceCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Instance instance; |
| VkResult result = |
| d.vkCreateInstance(reinterpret_cast<const VkInstanceCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkInstance *>(&instance)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::createInstance"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), instance); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Instance, Dispatch>>::type createInstanceUnique( |
| const VULKAN_HPP_NAMESPACE::InstanceCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Instance instance; |
| Result result = static_cast<Result>( |
| d.vkCreateInstance(reinterpret_cast<const VkInstanceCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkInstance *>(&instance))); |
| ObjectDestroy<NoParent, Dispatch> deleter(allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Instance, Dispatch>(result, instance, VULKAN_HPP_NAMESPACE_STRING "::createInstanceUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroy(const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyInstance(m_instance, reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroy(Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyInstance(m_instance, |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::enumeratePhysicalDevices(uint32_t *pPhysicalDeviceCount, |
| VULKAN_HPP_NAMESPACE::PhysicalDevice *pPhysicalDevices, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkEnumeratePhysicalDevices(m_instance, pPhysicalDeviceCount, reinterpret_cast<VkPhysicalDevice *>(pPhysicalDevices))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PhysicalDeviceAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDevice, PhysicalDeviceAllocator>>::type |
| Instance::enumeratePhysicalDevices(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDevice, PhysicalDeviceAllocator> physicalDevices; |
| uint32_t physicalDeviceCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumeratePhysicalDevices(m_instance, &physicalDeviceCount, nullptr)); |
| if((result == Result::eSuccess) && physicalDeviceCount) |
| { |
| physicalDevices.resize(physicalDeviceCount); |
| result = static_cast<Result>( |
| d.vkEnumeratePhysicalDevices(m_instance, &physicalDeviceCount, reinterpret_cast<VkPhysicalDevice *>(physicalDevices.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(physicalDeviceCount <= physicalDevices.size()); |
| if(physicalDeviceCount < physicalDevices.size()) |
| { |
| physicalDevices.resize(physicalDeviceCount); |
| } |
| } |
| return createResultValue(result, physicalDevices, VULKAN_HPP_NAMESPACE_STRING "::Instance::enumeratePhysicalDevices"); |
| } |
| |
| template<typename PhysicalDeviceAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PhysicalDevice>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDevice, PhysicalDeviceAllocator>>::type |
| Instance::enumeratePhysicalDevices(PhysicalDeviceAllocator &physicalDeviceAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDevice, PhysicalDeviceAllocator> physicalDevices(physicalDeviceAllocator); |
| uint32_t physicalDeviceCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumeratePhysicalDevices(m_instance, &physicalDeviceCount, nullptr)); |
| if((result == Result::eSuccess) && physicalDeviceCount) |
| { |
| physicalDevices.resize(physicalDeviceCount); |
| result = static_cast<Result>( |
| d.vkEnumeratePhysicalDevices(m_instance, &physicalDeviceCount, reinterpret_cast<VkPhysicalDevice *>(physicalDevices.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(physicalDeviceCount <= physicalDevices.size()); |
| if(physicalDeviceCount < physicalDevices.size()) |
| { |
| physicalDevices.resize(physicalDeviceCount); |
| } |
| } |
| return createResultValue(result, physicalDevices, VULKAN_HPP_NAMESPACE_STRING "::Instance::enumeratePhysicalDevices"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getFeatures(VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures *pFeatures, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceFeatures(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceFeatures *>(pFeatures)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures |
| PhysicalDevice::getFeatures(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures features; |
| d.vkGetPhysicalDeviceFeatures(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceFeatures *>(&features)); |
| |
| return features; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getFormatProperties(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::FormatProperties *pFormatProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceFormatProperties(m_physicalDevice, static_cast<VkFormat>(format), reinterpret_cast<VkFormatProperties *>(pFormatProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::FormatProperties |
| PhysicalDevice::getFormatProperties(VULKAN_HPP_NAMESPACE::Format format, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::FormatProperties formatProperties; |
| d.vkGetPhysicalDeviceFormatProperties(m_physicalDevice, static_cast<VkFormat>(format), reinterpret_cast<VkFormatProperties *>(&formatProperties)); |
| |
| return formatProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getImageFormatProperties(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::ImageType type, |
| VULKAN_HPP_NAMESPACE::ImageTiling tiling, |
| VULKAN_HPP_NAMESPACE::ImageUsageFlags usage, |
| VULKAN_HPP_NAMESPACE::ImageCreateFlags flags, |
| VULKAN_HPP_NAMESPACE::ImageFormatProperties *pImageFormatProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceImageFormatProperties(m_physicalDevice, |
| static_cast<VkFormat>(format), |
| static_cast<VkImageType>(type), |
| static_cast<VkImageTiling>(tiling), |
| static_cast<VkImageUsageFlags>(usage), |
| static_cast<VkImageCreateFlags>(flags), |
| reinterpret_cast<VkImageFormatProperties *>(pImageFormatProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::ImageFormatProperties>::type |
| PhysicalDevice::getImageFormatProperties(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::ImageType type, |
| VULKAN_HPP_NAMESPACE::ImageTiling tiling, |
| VULKAN_HPP_NAMESPACE::ImageUsageFlags usage, |
| VULKAN_HPP_NAMESPACE::ImageCreateFlags flags, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ImageFormatProperties imageFormatProperties; |
| VkResult result = d.vkGetPhysicalDeviceImageFormatProperties(m_physicalDevice, |
| static_cast<VkFormat>(format), |
| static_cast<VkImageType>(type), |
| static_cast<VkImageTiling>(tiling), |
| static_cast<VkImageUsageFlags>(usage), |
| static_cast<VkImageCreateFlags>(flags), |
| reinterpret_cast<VkImageFormatProperties *>(&imageFormatProperties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getImageFormatProperties"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), imageFormatProperties); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getProperties(VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceProperties(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceProperties *>(pProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties |
| PhysicalDevice::getProperties(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties properties; |
| d.vkGetPhysicalDeviceProperties(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceProperties *>(&properties)); |
| |
| return properties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getQueueFamilyProperties(uint32_t *pQueueFamilyPropertyCount, |
| VULKAN_HPP_NAMESPACE::QueueFamilyProperties *pQueueFamilyProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceQueueFamilyProperties( |
| m_physicalDevice, pQueueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties *>(pQueueFamilyProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename QueueFamilyPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<QueueFamilyProperties, QueueFamilyPropertiesAllocator> |
| PhysicalDevice::getQueueFamilyProperties(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<QueueFamilyProperties, QueueFamilyPropertiesAllocator> queueFamilyProperties; |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| queueFamilyProperties.resize(queueFamilyPropertyCount); |
| d.vkGetPhysicalDeviceQueueFamilyProperties( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| return queueFamilyProperties; |
| } |
| |
| template<typename QueueFamilyPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, QueueFamilyProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<QueueFamilyProperties, QueueFamilyPropertiesAllocator> |
| PhysicalDevice::getQueueFamilyProperties(QueueFamilyPropertiesAllocator &queueFamilyPropertiesAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<QueueFamilyProperties, QueueFamilyPropertiesAllocator> queueFamilyProperties(queueFamilyPropertiesAllocator); |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| queueFamilyProperties.resize(queueFamilyPropertyCount); |
| d.vkGetPhysicalDeviceQueueFamilyProperties( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| return queueFamilyProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getMemoryProperties(VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties *pMemoryProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceMemoryProperties(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceMemoryProperties *>(pMemoryProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties |
| PhysicalDevice::getMemoryProperties(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties memoryProperties; |
| d.vkGetPhysicalDeviceMemoryProperties(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceMemoryProperties *>(&memoryProperties)); |
| |
| return memoryProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE PFN_vkVoidFunction Instance::getProcAddr(const char *pName, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return d.vkGetInstanceProcAddr(m_instance, pName); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE PFN_vkVoidFunction Instance::getProcAddr(const std::string &name, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| PFN_vkVoidFunction result = d.vkGetInstanceProcAddr(m_instance, name.c_str()); |
| |
| return result; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE PFN_vkVoidFunction Device::getProcAddr(const char *pName, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return d.vkGetDeviceProcAddr(m_device, pName); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE PFN_vkVoidFunction Device::getProcAddr(const std::string &name, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| PFN_vkVoidFunction result = d.vkGetDeviceProcAddr(m_device, name.c_str()); |
| |
| return result; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::createDevice(const VULKAN_HPP_NAMESPACE::DeviceCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Device *pDevice, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDevice(m_physicalDevice, |
| reinterpret_cast<const VkDeviceCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkDevice *>(pDevice))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Device>::type PhysicalDevice::createDevice( |
| const VULKAN_HPP_NAMESPACE::DeviceCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Device device; |
| VkResult result = |
| d.vkCreateDevice(m_physicalDevice, |
| reinterpret_cast<const VkDeviceCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDevice *>(&device)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::createDevice"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), device); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Device, Dispatch>>::type |
| PhysicalDevice::createDeviceUnique(const VULKAN_HPP_NAMESPACE::DeviceCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Device device; |
| Result result = static_cast<Result>( |
| d.vkCreateDevice(m_physicalDevice, |
| reinterpret_cast<const VkDeviceCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDevice *>(&device))); |
| ObjectDestroy<NoParent, Dispatch> deleter(allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Device, Dispatch>( |
| result, device, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::createDeviceUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDevice(m_device, reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDevice(m_device, |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result enumerateInstanceExtensionProperties(const char *pLayerName, |
| uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::ExtensionProperties *pProperties, |
| Dispatch const &d) VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkEnumerateInstanceExtensionProperties(pLayerName, pPropertyCount, reinterpret_cast<VkExtensionProperties *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename ExtensionPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<ExtensionProperties, ExtensionPropertiesAllocator>>::type |
| enumerateInstanceExtensionProperties(Optional<const std::string> layerName, Dispatch const &d) |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<ExtensionProperties, ExtensionPropertiesAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumerateInstanceExtensionProperties(layerName ? layerName->c_str() : nullptr, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkEnumerateInstanceExtensionProperties( |
| layerName ? layerName->c_str() : nullptr, &propertyCount, reinterpret_cast<VkExtensionProperties *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::enumerateInstanceExtensionProperties"); |
| } |
| |
| template<typename ExtensionPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, ExtensionProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<ExtensionProperties, ExtensionPropertiesAllocator>>::type |
| enumerateInstanceExtensionProperties(Optional<const std::string> layerName, |
| ExtensionPropertiesAllocator &extensionPropertiesAllocator, |
| Dispatch const &d) |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<ExtensionProperties, ExtensionPropertiesAllocator> properties(extensionPropertiesAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumerateInstanceExtensionProperties(layerName ? layerName->c_str() : nullptr, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkEnumerateInstanceExtensionProperties( |
| layerName ? layerName->c_str() : nullptr, &propertyCount, reinterpret_cast<VkExtensionProperties *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::enumerateInstanceExtensionProperties"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::enumerateDeviceExtensionProperties(const char *pLayerName, |
| uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::ExtensionProperties *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkEnumerateDeviceExtensionProperties(m_physicalDevice, pLayerName, pPropertyCount, reinterpret_cast<VkExtensionProperties *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename ExtensionPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<ExtensionProperties, ExtensionPropertiesAllocator>>::type |
| PhysicalDevice::enumerateDeviceExtensionProperties(Optional<const std::string> layerName, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<ExtensionProperties, ExtensionPropertiesAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = |
| static_cast<Result>(d.vkEnumerateDeviceExtensionProperties(m_physicalDevice, layerName ? layerName->c_str() : nullptr, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkEnumerateDeviceExtensionProperties( |
| m_physicalDevice, layerName ? layerName->c_str() : nullptr, &propertyCount, reinterpret_cast<VkExtensionProperties *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::enumerateDeviceExtensionProperties"); |
| } |
| |
| template<typename ExtensionPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, ExtensionProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<ExtensionProperties, ExtensionPropertiesAllocator>>::type |
| PhysicalDevice::enumerateDeviceExtensionProperties(Optional<const std::string> layerName, |
| ExtensionPropertiesAllocator &extensionPropertiesAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<ExtensionProperties, ExtensionPropertiesAllocator> properties(extensionPropertiesAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = |
| static_cast<Result>(d.vkEnumerateDeviceExtensionProperties(m_physicalDevice, layerName ? layerName->c_str() : nullptr, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkEnumerateDeviceExtensionProperties( |
| m_physicalDevice, layerName ? layerName->c_str() : nullptr, &propertyCount, reinterpret_cast<VkExtensionProperties *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::enumerateDeviceExtensionProperties"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result enumerateInstanceLayerProperties(uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::LayerProperties *pProperties, |
| Dispatch const &d) VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkEnumerateInstanceLayerProperties(pPropertyCount, reinterpret_cast<VkLayerProperties *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename LayerPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<LayerProperties, LayerPropertiesAllocator>>::type |
| enumerateInstanceLayerProperties(Dispatch const &d) |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<LayerProperties, LayerPropertiesAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumerateInstanceLayerProperties(&propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkEnumerateInstanceLayerProperties(&propertyCount, reinterpret_cast<VkLayerProperties *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::enumerateInstanceLayerProperties"); |
| } |
| |
| template<typename LayerPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, LayerProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<LayerProperties, LayerPropertiesAllocator>>::type |
| enumerateInstanceLayerProperties(LayerPropertiesAllocator &layerPropertiesAllocator, Dispatch const &d) |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<LayerProperties, LayerPropertiesAllocator> properties(layerPropertiesAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumerateInstanceLayerProperties(&propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkEnumerateInstanceLayerProperties(&propertyCount, reinterpret_cast<VkLayerProperties *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::enumerateInstanceLayerProperties"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::enumerateDeviceLayerProperties(uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::LayerProperties *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkEnumerateDeviceLayerProperties(m_physicalDevice, pPropertyCount, reinterpret_cast<VkLayerProperties *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename LayerPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<LayerProperties, LayerPropertiesAllocator>>::type |
| PhysicalDevice::enumerateDeviceLayerProperties(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<LayerProperties, LayerPropertiesAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumerateDeviceLayerProperties(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>( |
| d.vkEnumerateDeviceLayerProperties(m_physicalDevice, &propertyCount, reinterpret_cast<VkLayerProperties *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::enumerateDeviceLayerProperties"); |
| } |
| |
| template<typename LayerPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, LayerProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<LayerProperties, LayerPropertiesAllocator>>::type |
| PhysicalDevice::enumerateDeviceLayerProperties(LayerPropertiesAllocator &layerPropertiesAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<LayerProperties, LayerPropertiesAllocator> properties(layerPropertiesAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumerateDeviceLayerProperties(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>( |
| d.vkEnumerateDeviceLayerProperties(m_physicalDevice, &propertyCount, reinterpret_cast<VkLayerProperties *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::enumerateDeviceLayerProperties"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| Device::getQueue(uint32_t queueFamilyIndex, uint32_t queueIndex, VULKAN_HPP_NAMESPACE::Queue *pQueue, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceQueue(m_device, queueFamilyIndex, queueIndex, reinterpret_cast<VkQueue *>(pQueue)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Queue |
| Device::getQueue(uint32_t queueFamilyIndex, uint32_t queueIndex, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Queue queue; |
| d.vkGetDeviceQueue(m_device, queueFamilyIndex, queueIndex, reinterpret_cast<VkQueue *>(&queue)); |
| |
| return queue; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Queue::submit(uint32_t submitCount, |
| const VULKAN_HPP_NAMESPACE::SubmitInfo *pSubmits, |
| VULKAN_HPP_NAMESPACE::Fence fence, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkQueueSubmit(m_queue, submitCount, reinterpret_cast<const VkSubmitInfo *>(pSubmits), static_cast<VkFence>(fence))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Queue::submit(ArrayProxy<const VULKAN_HPP_NAMESPACE::SubmitInfo> const &submits, VULKAN_HPP_NAMESPACE::Fence fence, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkQueueSubmit(m_queue, submits.size(), reinterpret_cast<const VkSubmitInfo *>(submits.data()), static_cast<VkFence>(fence)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Queue::submit"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Queue::waitIdle(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkQueueWaitIdle(m_queue)); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type Queue::waitIdle(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkQueueWaitIdle(m_queue); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Queue::waitIdle"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::waitIdle(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkDeviceWaitIdle(m_device)); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::waitIdle(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkDeviceWaitIdle(m_device); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::waitIdle"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::allocateMemory(const VULKAN_HPP_NAMESPACE::MemoryAllocateInfo *pAllocateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::DeviceMemory *pMemory, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAllocateMemory(m_device, |
| reinterpret_cast<const VkMemoryAllocateInfo *>(pAllocateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkDeviceMemory *>(pMemory))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DeviceMemory>::type |
| Device::allocateMemory(const VULKAN_HPP_NAMESPACE::MemoryAllocateInfo &allocateInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DeviceMemory memory; |
| VkResult result = |
| d.vkAllocateMemory(m_device, |
| reinterpret_cast<const VkMemoryAllocateInfo *>(&allocateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDeviceMemory *>(&memory)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::allocateMemory"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), memory); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DeviceMemory, Dispatch>>::type |
| Device::allocateMemoryUnique(const VULKAN_HPP_NAMESPACE::MemoryAllocateInfo &allocateInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DeviceMemory memory; |
| Result result = static_cast<Result>( |
| d.vkAllocateMemory(m_device, |
| reinterpret_cast<const VkMemoryAllocateInfo *>(&allocateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDeviceMemory *>(&memory))); |
| ObjectFree<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DeviceMemory, Dispatch>( |
| result, memory, VULKAN_HPP_NAMESPACE_STRING "::Device::allocateMemoryUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::freeMemory(VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkFreeMemory(m_device, static_cast<VkDeviceMemory>(memory), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::freeMemory(VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkFreeMemory(m_device, |
| static_cast<VkDeviceMemory>(memory), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void(Device::free)(VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkFreeMemory(m_device, static_cast<VkDeviceMemory>(memory), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void(Device::free)(VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkFreeMemory(m_device, |
| static_cast<VkDeviceMemory>(memory), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::mapMemory(VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::DeviceSize size, |
| VULKAN_HPP_NAMESPACE::MemoryMapFlags flags, |
| void **ppData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkMapMemory(m_device, |
| static_cast<VkDeviceMemory>(memory), |
| static_cast<VkDeviceSize>(offset), |
| static_cast<VkDeviceSize>(size), |
| static_cast<VkMemoryMapFlags>(flags), |
| ppData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<void *>::type Device::mapMemory(VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::DeviceSize size, |
| VULKAN_HPP_NAMESPACE::MemoryMapFlags flags, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| void *pData; |
| VkResult result = d.vkMapMemory(m_device, |
| static_cast<VkDeviceMemory>(memory), |
| static_cast<VkDeviceSize>(offset), |
| static_cast<VkDeviceSize>(size), |
| static_cast<VkMemoryMapFlags>(flags), |
| &pData); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::mapMemory"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pData); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::unmapMemory(VULKAN_HPP_NAMESPACE::DeviceMemory memory, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkUnmapMemory(m_device, static_cast<VkDeviceMemory>(memory)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::flushMappedMemoryRanges(uint32_t memoryRangeCount, |
| const VULKAN_HPP_NAMESPACE::MappedMemoryRange *pMemoryRanges, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkFlushMappedMemoryRanges(m_device, memoryRangeCount, reinterpret_cast<const VkMappedMemoryRange *>(pMemoryRanges))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::flushMappedMemoryRanges(ArrayProxy<const VULKAN_HPP_NAMESPACE::MappedMemoryRange> const &memoryRanges, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkFlushMappedMemoryRanges(m_device, memoryRanges.size(), reinterpret_cast<const VkMappedMemoryRange *>(memoryRanges.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::flushMappedMemoryRanges"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::invalidateMappedMemoryRanges(uint32_t memoryRangeCount, |
| const VULKAN_HPP_NAMESPACE::MappedMemoryRange *pMemoryRanges, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkInvalidateMappedMemoryRanges(m_device, memoryRangeCount, reinterpret_cast<const VkMappedMemoryRange *>(pMemoryRanges))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::invalidateMappedMemoryRanges(ArrayProxy<const VULKAN_HPP_NAMESPACE::MappedMemoryRange> const &memoryRanges, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkInvalidateMappedMemoryRanges(m_device, memoryRanges.size(), reinterpret_cast<const VkMappedMemoryRange *>(memoryRanges.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::invalidateMappedMemoryRanges"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getMemoryCommitment(VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| VULKAN_HPP_NAMESPACE::DeviceSize *pCommittedMemoryInBytes, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceMemoryCommitment(m_device, static_cast<VkDeviceMemory>(memory), reinterpret_cast<VkDeviceSize *>(pCommittedMemoryInBytes)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::DeviceSize Device::getMemoryCommitment(VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DeviceSize committedMemoryInBytes; |
| d.vkGetDeviceMemoryCommitment(m_device, static_cast<VkDeviceMemory>(memory), reinterpret_cast<VkDeviceSize *>(&committedMemoryInBytes)); |
| |
| return committedMemoryInBytes; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::bindBufferMemory(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| VULKAN_HPP_NAMESPACE::DeviceSize memoryOffset, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkBindBufferMemory(m_device, static_cast<VkBuffer>(buffer), static_cast<VkDeviceMemory>(memory), static_cast<VkDeviceSize>(memoryOffset))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::bindBufferMemory( |
| VULKAN_HPP_NAMESPACE::Buffer buffer, VULKAN_HPP_NAMESPACE::DeviceMemory memory, VULKAN_HPP_NAMESPACE::DeviceSize memoryOffset, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = |
| d.vkBindBufferMemory(m_device, static_cast<VkBuffer>(buffer), static_cast<VkDeviceMemory>(memory), static_cast<VkDeviceSize>(memoryOffset)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::bindBufferMemory"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::bindImageMemory(VULKAN_HPP_NAMESPACE::Image image, |
| VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| VULKAN_HPP_NAMESPACE::DeviceSize memoryOffset, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkBindImageMemory(m_device, static_cast<VkImage>(image), static_cast<VkDeviceMemory>(memory), static_cast<VkDeviceSize>(memoryOffset))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::bindImageMemory( |
| VULKAN_HPP_NAMESPACE::Image image, VULKAN_HPP_NAMESPACE::DeviceMemory memory, VULKAN_HPP_NAMESPACE::DeviceSize memoryOffset, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = |
| d.vkBindImageMemory(m_device, static_cast<VkImage>(image), static_cast<VkDeviceMemory>(memory), static_cast<VkDeviceSize>(memoryOffset)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::bindImageMemory"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getBufferMemoryRequirements(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetBufferMemoryRequirements(m_device, static_cast<VkBuffer>(buffer), reinterpret_cast<VkMemoryRequirements *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements |
| Device::getBufferMemoryRequirements(VULKAN_HPP_NAMESPACE::Buffer buffer, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements memoryRequirements; |
| d.vkGetBufferMemoryRequirements(m_device, static_cast<VkBuffer>(buffer), reinterpret_cast<VkMemoryRequirements *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageMemoryRequirements(VULKAN_HPP_NAMESPACE::Image image, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetImageMemoryRequirements(m_device, static_cast<VkImage>(image), reinterpret_cast<VkMemoryRequirements *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements |
| Device::getImageMemoryRequirements(VULKAN_HPP_NAMESPACE::Image image, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements memoryRequirements; |
| d.vkGetImageMemoryRequirements(m_device, static_cast<VkImage>(image), reinterpret_cast<VkMemoryRequirements *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageSparseMemoryRequirements(VULKAN_HPP_NAMESPACE::Image image, |
| uint32_t *pSparseMemoryRequirementCount, |
| VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements *pSparseMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetImageSparseMemoryRequirements(m_device, |
| static_cast<VkImage>(image), |
| pSparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements *>(pSparseMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SparseImageMemoryRequirementsAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements, SparseImageMemoryRequirementsAllocator> |
| Device::getImageSparseMemoryRequirements(VULKAN_HPP_NAMESPACE::Image image, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements, SparseImageMemoryRequirementsAllocator> sparseMemoryRequirements; |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetImageSparseMemoryRequirements(m_device, static_cast<VkImage>(image), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetImageSparseMemoryRequirements(m_device, |
| static_cast<VkImage>(image), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| |
| template<typename SparseImageMemoryRequirementsAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SparseImageMemoryRequirements>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements, SparseImageMemoryRequirementsAllocator> |
| Device::getImageSparseMemoryRequirements(VULKAN_HPP_NAMESPACE::Image image, |
| SparseImageMemoryRequirementsAllocator &sparseImageMemoryRequirementsAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements, SparseImageMemoryRequirementsAllocator> sparseMemoryRequirements(sparseImageMemoryRequirementsAllocator); |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetImageSparseMemoryRequirements(m_device, static_cast<VkImage>(image), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetImageSparseMemoryRequirements(m_device, |
| static_cast<VkImage>(image), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getSparseImageFormatProperties(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::ImageType type, |
| VULKAN_HPP_NAMESPACE::SampleCountFlagBits samples, |
| VULKAN_HPP_NAMESPACE::ImageUsageFlags usage, |
| VULKAN_HPP_NAMESPACE::ImageTiling tiling, |
| uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::SparseImageFormatProperties *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceSparseImageFormatProperties(m_physicalDevice, |
| static_cast<VkFormat>(format), |
| static_cast<VkImageType>(type), |
| static_cast<VkSampleCountFlagBits>(samples), |
| static_cast<VkImageUsageFlags>(usage), |
| static_cast<VkImageTiling>(tiling), |
| pPropertyCount, |
| reinterpret_cast<VkSparseImageFormatProperties *>(pProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SparseImageFormatPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageFormatProperties, SparseImageFormatPropertiesAllocator> |
| PhysicalDevice::getSparseImageFormatProperties(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::ImageType type, |
| VULKAN_HPP_NAMESPACE::SampleCountFlagBits samples, |
| VULKAN_HPP_NAMESPACE::ImageUsageFlags usage, |
| VULKAN_HPP_NAMESPACE::ImageTiling tiling, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageFormatProperties, SparseImageFormatPropertiesAllocator> properties; |
| uint32_t propertyCount; |
| d.vkGetPhysicalDeviceSparseImageFormatProperties(m_physicalDevice, |
| static_cast<VkFormat>(format), |
| static_cast<VkImageType>(type), |
| static_cast<VkSampleCountFlagBits>(samples), |
| static_cast<VkImageUsageFlags>(usage), |
| static_cast<VkImageTiling>(tiling), |
| &propertyCount, |
| nullptr); |
| properties.resize(propertyCount); |
| d.vkGetPhysicalDeviceSparseImageFormatProperties(m_physicalDevice, |
| static_cast<VkFormat>(format), |
| static_cast<VkImageType>(type), |
| static_cast<VkSampleCountFlagBits>(samples), |
| static_cast<VkImageUsageFlags>(usage), |
| static_cast<VkImageTiling>(tiling), |
| &propertyCount, |
| reinterpret_cast<VkSparseImageFormatProperties *>(properties.data())); |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| return properties; |
| } |
| |
| template<typename SparseImageFormatPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SparseImageFormatProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageFormatProperties, SparseImageFormatPropertiesAllocator> |
| PhysicalDevice::getSparseImageFormatProperties(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::ImageType type, |
| VULKAN_HPP_NAMESPACE::SampleCountFlagBits samples, |
| VULKAN_HPP_NAMESPACE::ImageUsageFlags usage, |
| VULKAN_HPP_NAMESPACE::ImageTiling tiling, |
| SparseImageFormatPropertiesAllocator &sparseImageFormatPropertiesAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageFormatProperties, SparseImageFormatPropertiesAllocator> properties(sparseImageFormatPropertiesAllocator); |
| uint32_t propertyCount; |
| d.vkGetPhysicalDeviceSparseImageFormatProperties(m_physicalDevice, |
| static_cast<VkFormat>(format), |
| static_cast<VkImageType>(type), |
| static_cast<VkSampleCountFlagBits>(samples), |
| static_cast<VkImageUsageFlags>(usage), |
| static_cast<VkImageTiling>(tiling), |
| &propertyCount, |
| nullptr); |
| properties.resize(propertyCount); |
| d.vkGetPhysicalDeviceSparseImageFormatProperties(m_physicalDevice, |
| static_cast<VkFormat>(format), |
| static_cast<VkImageType>(type), |
| static_cast<VkSampleCountFlagBits>(samples), |
| static_cast<VkImageUsageFlags>(usage), |
| static_cast<VkImageTiling>(tiling), |
| &propertyCount, |
| reinterpret_cast<VkSparseImageFormatProperties *>(properties.data())); |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| return properties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Queue::bindSparse(uint32_t bindInfoCount, |
| const VULKAN_HPP_NAMESPACE::BindSparseInfo *pBindInfo, |
| VULKAN_HPP_NAMESPACE::Fence fence, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkQueueBindSparse(m_queue, bindInfoCount, reinterpret_cast<const VkBindSparseInfo *>(pBindInfo), static_cast<VkFence>(fence))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Queue::bindSparse(ArrayProxy<const VULKAN_HPP_NAMESPACE::BindSparseInfo> const &bindInfo, VULKAN_HPP_NAMESPACE::Fence fence, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = |
| d.vkQueueBindSparse(m_queue, bindInfo.size(), reinterpret_cast<const VkBindSparseInfo *>(bindInfo.data()), static_cast<VkFence>(fence)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Queue::bindSparse"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createFence(const VULKAN_HPP_NAMESPACE::FenceCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Fence *pFence, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateFence(m_device, |
| reinterpret_cast<const VkFenceCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkFence *>(pFence))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Fence>::type Device::createFence( |
| const VULKAN_HPP_NAMESPACE::FenceCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Fence fence; |
| VkResult result = |
| d.vkCreateFence(m_device, |
| reinterpret_cast<const VkFenceCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkFence *>(&fence)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createFence"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), fence); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Fence, Dispatch>>::type Device::createFenceUnique( |
| const VULKAN_HPP_NAMESPACE::FenceCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Fence fence; |
| Result result = static_cast<Result>( |
| d.vkCreateFence(m_device, |
| reinterpret_cast<const VkFenceCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkFence *>(&fence))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Fence, Dispatch>(result, fence, VULKAN_HPP_NAMESPACE_STRING "::Device::createFenceUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyFence(VULKAN_HPP_NAMESPACE::Fence fence, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyFence(m_device, static_cast<VkFence>(fence), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyFence(VULKAN_HPP_NAMESPACE::Fence fence, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyFence(m_device, |
| static_cast<VkFence>(fence), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Fence fence, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyFence(m_device, static_cast<VkFence>(fence), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Fence fence, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyFence(m_device, |
| static_cast<VkFence>(fence), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::resetFences(uint32_t fenceCount, |
| const VULKAN_HPP_NAMESPACE::Fence *pFences, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkResetFences(m_device, fenceCount, reinterpret_cast<const VkFence *>(pFences))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::resetFences(ArrayProxy<const VULKAN_HPP_NAMESPACE::Fence> const &fences, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkResetFences(m_device, fences.size(), reinterpret_cast<const VkFence *>(fences.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::resetFences"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getFenceStatus(VULKAN_HPP_NAMESPACE::Fence fence, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetFenceStatus(m_device, static_cast<VkFence>(fence))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result Device::getFenceStatus(VULKAN_HPP_NAMESPACE::Fence fence, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkGetFenceStatus(m_device, static_cast<VkFence>(fence)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::getFenceStatus", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eNotReady }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::waitForFences(uint32_t fenceCount, |
| const VULKAN_HPP_NAMESPACE::Fence *pFences, |
| VULKAN_HPP_NAMESPACE::Bool32 waitAll, |
| uint64_t timeout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkWaitForFences(m_device, fenceCount, reinterpret_cast<const VkFence *>(pFences), static_cast<VkBool32>(waitAll), timeout)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result Device::waitForFences(ArrayProxy<const VULKAN_HPP_NAMESPACE::Fence> const &fences, |
| VULKAN_HPP_NAMESPACE::Bool32 waitAll, |
| uint64_t timeout, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = |
| d.vkWaitForFences(m_device, fences.size(), reinterpret_cast<const VkFence *>(fences.data()), static_cast<VkBool32>(waitAll), timeout); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::waitForFences", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eTimeout }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createSemaphore(const VULKAN_HPP_NAMESPACE::SemaphoreCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Semaphore *pSemaphore, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateSemaphore(m_device, |
| reinterpret_cast<const VkSemaphoreCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSemaphore *>(pSemaphore))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Semaphore>::type |
| Device::createSemaphore(const VULKAN_HPP_NAMESPACE::SemaphoreCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Semaphore semaphore; |
| VkResult result = |
| d.vkCreateSemaphore(m_device, |
| reinterpret_cast<const VkSemaphoreCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSemaphore *>(&semaphore)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createSemaphore"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), semaphore); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Semaphore, Dispatch>>::type |
| Device::createSemaphoreUnique(const VULKAN_HPP_NAMESPACE::SemaphoreCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Semaphore semaphore; |
| Result result = static_cast<Result>( |
| d.vkCreateSemaphore(m_device, |
| reinterpret_cast<const VkSemaphoreCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSemaphore *>(&semaphore))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Semaphore, Dispatch>( |
| result, semaphore, VULKAN_HPP_NAMESPACE_STRING "::Device::createSemaphoreUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySemaphore(VULKAN_HPP_NAMESPACE::Semaphore semaphore, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySemaphore(m_device, static_cast<VkSemaphore>(semaphore), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySemaphore(VULKAN_HPP_NAMESPACE::Semaphore semaphore, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySemaphore(m_device, |
| static_cast<VkSemaphore>(semaphore), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Semaphore semaphore, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySemaphore(m_device, static_cast<VkSemaphore>(semaphore), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Semaphore semaphore, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySemaphore(m_device, |
| static_cast<VkSemaphore>(semaphore), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createEvent(const VULKAN_HPP_NAMESPACE::EventCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Event *pEvent, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateEvent(m_device, |
| reinterpret_cast<const VkEventCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkEvent *>(pEvent))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Event>::type Device::createEvent( |
| const VULKAN_HPP_NAMESPACE::EventCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Event event; |
| VkResult result = |
| d.vkCreateEvent(m_device, |
| reinterpret_cast<const VkEventCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkEvent *>(&event)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createEvent"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), event); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Event, Dispatch>>::type Device::createEventUnique( |
| const VULKAN_HPP_NAMESPACE::EventCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Event event; |
| Result result = static_cast<Result>( |
| d.vkCreateEvent(m_device, |
| reinterpret_cast<const VkEventCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkEvent *>(&event))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Event, Dispatch>(result, event, VULKAN_HPP_NAMESPACE_STRING "::Device::createEventUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyEvent(VULKAN_HPP_NAMESPACE::Event event, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyEvent(m_device, static_cast<VkEvent>(event), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyEvent(VULKAN_HPP_NAMESPACE::Event event, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyEvent(m_device, |
| static_cast<VkEvent>(event), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Event event, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyEvent(m_device, static_cast<VkEvent>(event), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Event event, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyEvent(m_device, |
| static_cast<VkEvent>(event), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getEventStatus(VULKAN_HPP_NAMESPACE::Event event, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetEventStatus(m_device, static_cast<VkEvent>(event))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result Device::getEventStatus(VULKAN_HPP_NAMESPACE::Event event, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkGetEventStatus(m_device, static_cast<VkEvent>(event)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::getEventStatus", |
| { VULKAN_HPP_NAMESPACE::Result::eEventSet, VULKAN_HPP_NAMESPACE::Result::eEventReset }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::setEvent(VULKAN_HPP_NAMESPACE::Event event, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkSetEvent(m_device, static_cast<VkEvent>(event))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::setEvent(VULKAN_HPP_NAMESPACE::Event event, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkSetEvent(m_device, static_cast<VkEvent>(event)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::setEvent"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::resetEvent(VULKAN_HPP_NAMESPACE::Event event, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkResetEvent(m_device, static_cast<VkEvent>(event))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::resetEvent(VULKAN_HPP_NAMESPACE::Event event, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkResetEvent(m_device, static_cast<VkEvent>(event)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::resetEvent"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createQueryPool(const VULKAN_HPP_NAMESPACE::QueryPoolCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::QueryPool *pQueryPool, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateQueryPool(m_device, |
| reinterpret_cast<const VkQueryPoolCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkQueryPool *>(pQueryPool))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::QueryPool>::type |
| Device::createQueryPool(const VULKAN_HPP_NAMESPACE::QueryPoolCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::QueryPool queryPool; |
| VkResult result = |
| d.vkCreateQueryPool(m_device, |
| reinterpret_cast<const VkQueryPoolCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkQueryPool *>(&queryPool)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createQueryPool"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), queryPool); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::QueryPool, Dispatch>>::type |
| Device::createQueryPoolUnique(const VULKAN_HPP_NAMESPACE::QueryPoolCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::QueryPool queryPool; |
| Result result = static_cast<Result>( |
| d.vkCreateQueryPool(m_device, |
| reinterpret_cast<const VkQueryPoolCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkQueryPool *>(&queryPool))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::QueryPool, Dispatch>( |
| result, queryPool, VULKAN_HPP_NAMESPACE_STRING "::Device::createQueryPoolUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyQueryPool(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyQueryPool(m_device, static_cast<VkQueryPool>(queryPool), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyQueryPool(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyQueryPool(m_device, |
| static_cast<VkQueryPool>(queryPool), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyQueryPool(m_device, static_cast<VkQueryPool>(queryPool), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyQueryPool(m_device, |
| static_cast<VkQueryPool>(queryPool), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getQueryPoolResults(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| uint32_t queryCount, |
| size_t dataSize, |
| void *pData, |
| VULKAN_HPP_NAMESPACE::DeviceSize stride, |
| VULKAN_HPP_NAMESPACE::QueryResultFlags flags, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetQueryPoolResults(m_device, |
| static_cast<VkQueryPool>(queryPool), |
| firstQuery, |
| queryCount, |
| dataSize, |
| pData, |
| static_cast<VkDeviceSize>(stride), |
| static_cast<VkQueryResultFlags>(flags))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename DataTypeAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<DataType, DataTypeAllocator>> |
| Device::getQueryPoolResults(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| uint32_t queryCount, |
| size_t dataSize, |
| VULKAN_HPP_NAMESPACE::DeviceSize stride, |
| VULKAN_HPP_NAMESPACE::QueryResultFlags flags, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_ASSERT(dataSize % sizeof(DataType) == 0); |
| std::vector<DataType, DataTypeAllocator> data(dataSize / sizeof(DataType)); |
| Result result = static_cast<Result>(d.vkGetQueryPoolResults(m_device, |
| static_cast<VkQueryPool>(queryPool), |
| firstQuery, |
| queryCount, |
| data.size() * sizeof(DataType), |
| reinterpret_cast<void *>(data.data()), |
| static_cast<VkDeviceSize>(stride), |
| static_cast<VkQueryResultFlags>(flags))); |
| return createResultValue(result, |
| data, |
| VULKAN_HPP_NAMESPACE_STRING "::Device::getQueryPoolResults", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eNotReady }); |
| } |
| |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<DataType> Device::getQueryPoolResult(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| uint32_t queryCount, |
| VULKAN_HPP_NAMESPACE::DeviceSize stride, |
| VULKAN_HPP_NAMESPACE::QueryResultFlags flags, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| DataType data; |
| VkResult result = d.vkGetQueryPoolResults(m_device, |
| static_cast<VkQueryPool>(queryPool), |
| firstQuery, |
| queryCount, |
| sizeof(DataType), |
| reinterpret_cast<void *>(&data), |
| static_cast<VkDeviceSize>(stride), |
| static_cast<VkQueryResultFlags>(flags)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::getQueryPoolResult", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eNotReady }); |
| return ResultValue<DataType>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), data); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createBuffer(const VULKAN_HPP_NAMESPACE::BufferCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Buffer *pBuffer, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateBuffer(m_device, |
| reinterpret_cast<const VkBufferCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkBuffer *>(pBuffer))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Buffer>::type Device::createBuffer( |
| const VULKAN_HPP_NAMESPACE::BufferCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Buffer buffer; |
| VkResult result = |
| d.vkCreateBuffer(m_device, |
| reinterpret_cast<const VkBufferCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkBuffer *>(&buffer)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createBuffer"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), buffer); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Buffer, Dispatch>>::type Device::createBufferUnique( |
| const VULKAN_HPP_NAMESPACE::BufferCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Buffer buffer; |
| Result result = static_cast<Result>( |
| d.vkCreateBuffer(m_device, |
| reinterpret_cast<const VkBufferCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkBuffer *>(&buffer))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Buffer, Dispatch>(result, buffer, VULKAN_HPP_NAMESPACE_STRING "::Device::createBufferUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyBuffer(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyBuffer(m_device, static_cast<VkBuffer>(buffer), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyBuffer(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyBuffer(m_device, |
| static_cast<VkBuffer>(buffer), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyBuffer(m_device, static_cast<VkBuffer>(buffer), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyBuffer(m_device, |
| static_cast<VkBuffer>(buffer), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createBufferView(const VULKAN_HPP_NAMESPACE::BufferViewCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::BufferView *pView, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateBufferView(m_device, |
| reinterpret_cast<const VkBufferViewCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkBufferView *>(pView))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::BufferView>::type |
| Device::createBufferView(const VULKAN_HPP_NAMESPACE::BufferViewCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::BufferView view; |
| VkResult result = |
| d.vkCreateBufferView(m_device, |
| reinterpret_cast<const VkBufferViewCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkBufferView *>(&view)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createBufferView"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), view); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::BufferView, Dispatch>>::type |
| Device::createBufferViewUnique(const VULKAN_HPP_NAMESPACE::BufferViewCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::BufferView view; |
| Result result = static_cast<Result>( |
| d.vkCreateBufferView(m_device, |
| reinterpret_cast<const VkBufferViewCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkBufferView *>(&view))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::BufferView, Dispatch>( |
| result, view, VULKAN_HPP_NAMESPACE_STRING "::Device::createBufferViewUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyBufferView(VULKAN_HPP_NAMESPACE::BufferView bufferView, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyBufferView(m_device, static_cast<VkBufferView>(bufferView), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyBufferView(VULKAN_HPP_NAMESPACE::BufferView bufferView, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyBufferView(m_device, |
| static_cast<VkBufferView>(bufferView), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::BufferView bufferView, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyBufferView(m_device, static_cast<VkBufferView>(bufferView), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::BufferView bufferView, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyBufferView(m_device, |
| static_cast<VkBufferView>(bufferView), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createImage(const VULKAN_HPP_NAMESPACE::ImageCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Image *pImage, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateImage(m_device, |
| reinterpret_cast<const VkImageCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkImage *>(pImage))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Image>::type Device::createImage( |
| const VULKAN_HPP_NAMESPACE::ImageCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Image image; |
| VkResult result = |
| d.vkCreateImage(m_device, |
| reinterpret_cast<const VkImageCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkImage *>(&image)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createImage"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), image); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Image, Dispatch>>::type Device::createImageUnique( |
| const VULKAN_HPP_NAMESPACE::ImageCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Image image; |
| Result result = static_cast<Result>( |
| d.vkCreateImage(m_device, |
| reinterpret_cast<const VkImageCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkImage *>(&image))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Image, Dispatch>(result, image, VULKAN_HPP_NAMESPACE_STRING "::Device::createImageUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyImage(VULKAN_HPP_NAMESPACE::Image image, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyImage(m_device, static_cast<VkImage>(image), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyImage(VULKAN_HPP_NAMESPACE::Image image, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyImage(m_device, |
| static_cast<VkImage>(image), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Image image, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyImage(m_device, static_cast<VkImage>(image), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Image image, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyImage(m_device, |
| static_cast<VkImage>(image), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageSubresourceLayout(VULKAN_HPP_NAMESPACE::Image image, |
| const VULKAN_HPP_NAMESPACE::ImageSubresource *pSubresource, |
| VULKAN_HPP_NAMESPACE::SubresourceLayout *pLayout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetImageSubresourceLayout(m_device, |
| static_cast<VkImage>(image), |
| reinterpret_cast<const VkImageSubresource *>(pSubresource), |
| reinterpret_cast<VkSubresourceLayout *>(pLayout)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::SubresourceLayout Device::getImageSubresourceLayout( |
| VULKAN_HPP_NAMESPACE::Image image, const VULKAN_HPP_NAMESPACE::ImageSubresource &subresource, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SubresourceLayout layout; |
| d.vkGetImageSubresourceLayout(m_device, |
| static_cast<VkImage>(image), |
| reinterpret_cast<const VkImageSubresource *>(&subresource), |
| reinterpret_cast<VkSubresourceLayout *>(&layout)); |
| |
| return layout; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createImageView(const VULKAN_HPP_NAMESPACE::ImageViewCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::ImageView *pView, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateImageView(m_device, |
| reinterpret_cast<const VkImageViewCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkImageView *>(pView))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::ImageView>::type |
| Device::createImageView(const VULKAN_HPP_NAMESPACE::ImageViewCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ImageView view; |
| VkResult result = |
| d.vkCreateImageView(m_device, |
| reinterpret_cast<const VkImageViewCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkImageView *>(&view)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createImageView"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), view); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::ImageView, Dispatch>>::type |
| Device::createImageViewUnique(const VULKAN_HPP_NAMESPACE::ImageViewCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::ImageView view; |
| Result result = static_cast<Result>( |
| d.vkCreateImageView(m_device, |
| reinterpret_cast<const VkImageViewCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkImageView *>(&view))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::ImageView, Dispatch>(result, view, VULKAN_HPP_NAMESPACE_STRING "::Device::createImageViewUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyImageView(VULKAN_HPP_NAMESPACE::ImageView imageView, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyImageView(m_device, static_cast<VkImageView>(imageView), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyImageView(VULKAN_HPP_NAMESPACE::ImageView imageView, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyImageView(m_device, |
| static_cast<VkImageView>(imageView), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::ImageView imageView, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyImageView(m_device, static_cast<VkImageView>(imageView), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::ImageView imageView, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyImageView(m_device, |
| static_cast<VkImageView>(imageView), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createShaderModule(const VULKAN_HPP_NAMESPACE::ShaderModuleCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::ShaderModule *pShaderModule, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateShaderModule(m_device, |
| reinterpret_cast<const VkShaderModuleCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkShaderModule *>(pShaderModule))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::ShaderModule>::type |
| Device::createShaderModule(const VULKAN_HPP_NAMESPACE::ShaderModuleCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ShaderModule shaderModule; |
| VkResult result = |
| d.vkCreateShaderModule(m_device, |
| reinterpret_cast<const VkShaderModuleCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkShaderModule *>(&shaderModule)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createShaderModule"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), shaderModule); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::ShaderModule, Dispatch>>::type |
| Device::createShaderModuleUnique(const VULKAN_HPP_NAMESPACE::ShaderModuleCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::ShaderModule shaderModule; |
| Result result = static_cast<Result>( |
| d.vkCreateShaderModule(m_device, |
| reinterpret_cast<const VkShaderModuleCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkShaderModule *>(&shaderModule))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::ShaderModule, Dispatch>( |
| result, shaderModule, VULKAN_HPP_NAMESPACE_STRING "::Device::createShaderModuleUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyShaderModule(VULKAN_HPP_NAMESPACE::ShaderModule shaderModule, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyShaderModule(m_device, static_cast<VkShaderModule>(shaderModule), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyShaderModule(VULKAN_HPP_NAMESPACE::ShaderModule shaderModule, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyShaderModule(m_device, |
| static_cast<VkShaderModule>(shaderModule), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::ShaderModule shaderModule, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyShaderModule(m_device, static_cast<VkShaderModule>(shaderModule), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::ShaderModule shaderModule, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyShaderModule(m_device, |
| static_cast<VkShaderModule>(shaderModule), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createPipelineCache(const VULKAN_HPP_NAMESPACE::PipelineCacheCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::PipelineCache *pPipelineCache, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreatePipelineCache(m_device, |
| reinterpret_cast<const VkPipelineCacheCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkPipelineCache *>(pPipelineCache))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::PipelineCache>::type |
| Device::createPipelineCache(const VULKAN_HPP_NAMESPACE::PipelineCacheCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache; |
| VkResult result = |
| d.vkCreatePipelineCache(m_device, |
| reinterpret_cast<const VkPipelineCacheCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipelineCache *>(&pipelineCache)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createPipelineCache"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelineCache); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::PipelineCache, Dispatch>>::type |
| Device::createPipelineCacheUnique(const VULKAN_HPP_NAMESPACE::PipelineCacheCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache; |
| Result result = static_cast<Result>( |
| d.vkCreatePipelineCache(m_device, |
| reinterpret_cast<const VkPipelineCacheCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipelineCache *>(&pipelineCache))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::PipelineCache, Dispatch>( |
| result, pipelineCache, VULKAN_HPP_NAMESPACE_STRING "::Device::createPipelineCacheUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPipelineCache(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyPipelineCache(m_device, static_cast<VkPipelineCache>(pipelineCache), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPipelineCache(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyPipelineCache(m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyPipelineCache(m_device, static_cast<VkPipelineCache>(pipelineCache), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyPipelineCache(m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getPipelineCacheData(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| size_t *pDataSize, |
| void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPipelineCacheData(m_device, static_cast<VkPipelineCache>(pipelineCache), pDataSize, pData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Uint8_tAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type |
| Device::getPipelineCacheData(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<uint8_t, Uint8_tAllocator> data; |
| size_t dataSize; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPipelineCacheData(m_device, static_cast<VkPipelineCache>(pipelineCache), &dataSize, nullptr)); |
| if((result == Result::eSuccess) && dataSize) |
| { |
| data.resize(dataSize); |
| result = static_cast<Result>( |
| d.vkGetPipelineCacheData(m_device, static_cast<VkPipelineCache>(pipelineCache), &dataSize, reinterpret_cast<void *>(data.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(dataSize <= data.size()); |
| if(dataSize < data.size()) |
| { |
| data.resize(dataSize); |
| } |
| } |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getPipelineCacheData"); |
| } |
| |
| template<typename Uint8_tAllocator, typename Dispatch, typename B, typename std::enable_if<std::is_same<typename B::value_type, uint8_t>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type |
| Device::getPipelineCacheData(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, Uint8_tAllocator &uint8_tAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<uint8_t, Uint8_tAllocator> data(uint8_tAllocator); |
| size_t dataSize; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPipelineCacheData(m_device, static_cast<VkPipelineCache>(pipelineCache), &dataSize, nullptr)); |
| if((result == Result::eSuccess) && dataSize) |
| { |
| data.resize(dataSize); |
| result = static_cast<Result>( |
| d.vkGetPipelineCacheData(m_device, static_cast<VkPipelineCache>(pipelineCache), &dataSize, reinterpret_cast<void *>(data.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(dataSize <= data.size()); |
| if(dataSize < data.size()) |
| { |
| data.resize(dataSize); |
| } |
| } |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getPipelineCacheData"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::mergePipelineCaches(VULKAN_HPP_NAMESPACE::PipelineCache dstCache, |
| uint32_t srcCacheCount, |
| const VULKAN_HPP_NAMESPACE::PipelineCache *pSrcCaches, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkMergePipelineCaches(m_device, static_cast<VkPipelineCache>(dstCache), srcCacheCount, reinterpret_cast<const VkPipelineCache *>(pSrcCaches))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::mergePipelineCaches( |
| VULKAN_HPP_NAMESPACE::PipelineCache dstCache, ArrayProxy<const VULKAN_HPP_NAMESPACE::PipelineCache> const &srcCaches, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkMergePipelineCaches( |
| m_device, static_cast<VkPipelineCache>(dstCache), srcCaches.size(), reinterpret_cast<const VkPipelineCache *>(srcCaches.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::mergePipelineCaches"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createGraphicsPipelines(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| uint32_t createInfoCount, |
| const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo *pCreateInfos, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Pipeline *pPipelines, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateGraphicsPipelines(m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfoCount, |
| reinterpret_cast<const VkGraphicsPipelineCreateInfo *>(pCreateInfos), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkPipeline *>(pPipelines))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PipelineAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>> |
| Device::createGraphicsPipelines(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines(createInfos.size()); |
| VkResult result = d.vkCreateGraphicsPipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkGraphicsPipelineCreateInfo *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createGraphicsPipelines", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelines); |
| } |
| |
| template<typename PipelineAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, Pipeline>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>> |
| Device::createGraphicsPipelines(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| PipelineAllocator &pipelineAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines(createInfos.size(), pipelineAllocator); |
| VkResult result = d.vkCreateGraphicsPipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkGraphicsPipelineCreateInfo *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createGraphicsPipelines", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelines); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<VULKAN_HPP_NAMESPACE::Pipeline> |
| Device::createGraphicsPipeline(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline; |
| VkResult result = d.vkCreateGraphicsPipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| 1, |
| reinterpret_cast<const VkGraphicsPipelineCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(&pipeline)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createGraphicsPipeline", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<VULKAN_HPP_NAMESPACE::Pipeline>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipeline); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch, typename PipelineAllocator> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator>> |
| Device::createGraphicsPipelinesUnique(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator> uniquePipelines; |
| std::vector<Pipeline> pipelines(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateGraphicsPipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkGraphicsPipelineCreateInfo *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data()))); |
| if((result == VULKAN_HPP_NAMESPACE::Result::eSuccess) || (result == VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT)) |
| { |
| uniquePipelines.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniquePipelines.push_back(UniqueHandle<Pipeline, Dispatch>(pipelines[i], deleter)); |
| } |
| } |
| return createResultValue(result, |
| std::move(uniquePipelines), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createGraphicsPipelinesUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| } |
| |
| template<typename Dispatch, |
| typename PipelineAllocator, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, UniqueHandle<Pipeline, Dispatch>>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator>> |
| Device::createGraphicsPipelinesUnique(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| PipelineAllocator &pipelineAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator> uniquePipelines(pipelineAllocator); |
| std::vector<Pipeline> pipelines(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateGraphicsPipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkGraphicsPipelineCreateInfo *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data()))); |
| if((result == VULKAN_HPP_NAMESPACE::Result::eSuccess) || (result == VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT)) |
| { |
| uniquePipelines.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniquePipelines.push_back(UniqueHandle<Pipeline, Dispatch>(pipelines[i], deleter)); |
| } |
| } |
| return createResultValue(result, |
| std::move(uniquePipelines), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createGraphicsPipelinesUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<UniqueHandle<Pipeline, Dispatch>> |
| Device::createGraphicsPipelineUnique(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::GraphicsPipelineCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| Pipeline pipeline; |
| Result result = static_cast<Result>(d.vkCreateGraphicsPipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| 1, |
| reinterpret_cast<const VkGraphicsPipelineCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(&pipeline))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<Pipeline, Dispatch>(result, |
| pipeline, |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createGraphicsPipelineUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }, |
| deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createComputePipelines(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| uint32_t createInfoCount, |
| const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo *pCreateInfos, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Pipeline *pPipelines, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateComputePipelines(m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfoCount, |
| reinterpret_cast<const VkComputePipelineCreateInfo *>(pCreateInfos), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkPipeline *>(pPipelines))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PipelineAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>> |
| Device::createComputePipelines(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines(createInfos.size()); |
| VkResult result = d.vkCreateComputePipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkComputePipelineCreateInfo *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createComputePipelines", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelines); |
| } |
| |
| template<typename PipelineAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, Pipeline>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>> |
| Device::createComputePipelines(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| PipelineAllocator &pipelineAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines(createInfos.size(), pipelineAllocator); |
| VkResult result = d.vkCreateComputePipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkComputePipelineCreateInfo *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createComputePipelines", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelines); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<VULKAN_HPP_NAMESPACE::Pipeline> |
| Device::createComputePipeline(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline; |
| VkResult result = d.vkCreateComputePipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| 1, |
| reinterpret_cast<const VkComputePipelineCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(&pipeline)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createComputePipeline", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<VULKAN_HPP_NAMESPACE::Pipeline>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipeline); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch, typename PipelineAllocator> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator>> |
| Device::createComputePipelinesUnique(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator> uniquePipelines; |
| std::vector<Pipeline> pipelines(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateComputePipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkComputePipelineCreateInfo *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data()))); |
| if((result == VULKAN_HPP_NAMESPACE::Result::eSuccess) || (result == VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT)) |
| { |
| uniquePipelines.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniquePipelines.push_back(UniqueHandle<Pipeline, Dispatch>(pipelines[i], deleter)); |
| } |
| } |
| return createResultValue(result, |
| std::move(uniquePipelines), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createComputePipelinesUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| } |
| |
| template<typename Dispatch, |
| typename PipelineAllocator, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, UniqueHandle<Pipeline, Dispatch>>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator>> |
| Device::createComputePipelinesUnique(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| PipelineAllocator &pipelineAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator> uniquePipelines(pipelineAllocator); |
| std::vector<Pipeline> pipelines(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateComputePipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkComputePipelineCreateInfo *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data()))); |
| if((result == VULKAN_HPP_NAMESPACE::Result::eSuccess) || (result == VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT)) |
| { |
| uniquePipelines.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniquePipelines.push_back(UniqueHandle<Pipeline, Dispatch>(pipelines[i], deleter)); |
| } |
| } |
| return createResultValue(result, |
| std::move(uniquePipelines), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createComputePipelinesUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<UniqueHandle<Pipeline, Dispatch>> |
| Device::createComputePipelineUnique(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::ComputePipelineCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| Pipeline pipeline; |
| Result result = static_cast<Result>(d.vkCreateComputePipelines( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| 1, |
| reinterpret_cast<const VkComputePipelineCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(&pipeline))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<Pipeline, Dispatch>(result, |
| pipeline, |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createComputePipelineUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }, |
| deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPipeline(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyPipeline(m_device, static_cast<VkPipeline>(pipeline), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPipeline(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyPipeline(m_device, |
| static_cast<VkPipeline>(pipeline), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyPipeline(m_device, static_cast<VkPipeline>(pipeline), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyPipeline(m_device, |
| static_cast<VkPipeline>(pipeline), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createPipelineLayout(const VULKAN_HPP_NAMESPACE::PipelineLayoutCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::PipelineLayout *pPipelineLayout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreatePipelineLayout(m_device, |
| reinterpret_cast<const VkPipelineLayoutCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkPipelineLayout *>(pPipelineLayout))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::PipelineLayout>::type |
| Device::createPipelineLayout(const VULKAN_HPP_NAMESPACE::PipelineLayoutCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout; |
| VkResult result = |
| d.vkCreatePipelineLayout(m_device, |
| reinterpret_cast<const VkPipelineLayoutCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipelineLayout *>(&pipelineLayout)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createPipelineLayout"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelineLayout); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::PipelineLayout, Dispatch>>::type |
| Device::createPipelineLayoutUnique(const VULKAN_HPP_NAMESPACE::PipelineLayoutCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout; |
| Result result = static_cast<Result>( |
| d.vkCreatePipelineLayout(m_device, |
| reinterpret_cast<const VkPipelineLayoutCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipelineLayout *>(&pipelineLayout))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::PipelineLayout, Dispatch>( |
| result, pipelineLayout, VULKAN_HPP_NAMESPACE_STRING "::Device::createPipelineLayoutUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPipelineLayout(VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyPipelineLayout(m_device, static_cast<VkPipelineLayout>(pipelineLayout), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPipelineLayout(VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyPipelineLayout(m_device, |
| static_cast<VkPipelineLayout>(pipelineLayout), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyPipelineLayout(m_device, static_cast<VkPipelineLayout>(pipelineLayout), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::PipelineLayout pipelineLayout, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyPipelineLayout(m_device, |
| static_cast<VkPipelineLayout>(pipelineLayout), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createSampler(const VULKAN_HPP_NAMESPACE::SamplerCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Sampler *pSampler, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateSampler(m_device, |
| reinterpret_cast<const VkSamplerCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSampler *>(pSampler))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Sampler>::type Device::createSampler( |
| const VULKAN_HPP_NAMESPACE::SamplerCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Sampler sampler; |
| VkResult result = |
| d.vkCreateSampler(m_device, |
| reinterpret_cast<const VkSamplerCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSampler *>(&sampler)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createSampler"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), sampler); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Sampler, Dispatch>>::type Device::createSamplerUnique( |
| const VULKAN_HPP_NAMESPACE::SamplerCreateInfo &createInfo, Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Sampler sampler; |
| Result result = static_cast<Result>( |
| d.vkCreateSampler(m_device, |
| reinterpret_cast<const VkSamplerCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSampler *>(&sampler))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Sampler, Dispatch>(result, sampler, VULKAN_HPP_NAMESPACE_STRING "::Device::createSamplerUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySampler(VULKAN_HPP_NAMESPACE::Sampler sampler, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySampler(m_device, static_cast<VkSampler>(sampler), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySampler(VULKAN_HPP_NAMESPACE::Sampler sampler, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySampler(m_device, |
| static_cast<VkSampler>(sampler), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Sampler sampler, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySampler(m_device, static_cast<VkSampler>(sampler), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Sampler sampler, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySampler(m_device, |
| static_cast<VkSampler>(sampler), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createDescriptorSetLayout(const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayout *pSetLayout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDescriptorSetLayout(m_device, |
| reinterpret_cast<const VkDescriptorSetLayoutCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkDescriptorSetLayout *>(pSetLayout))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DescriptorSetLayout>::type |
| Device::createDescriptorSetLayout(const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayout setLayout; |
| VkResult result = d.vkCreateDescriptorSetLayout( |
| m_device, |
| reinterpret_cast<const VkDescriptorSetLayoutCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDescriptorSetLayout *>(&setLayout)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createDescriptorSetLayout"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), setLayout); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DescriptorSetLayout, Dispatch>>::type |
| Device::createDescriptorSetLayoutUnique(const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayout setLayout; |
| Result result = static_cast<Result>(d.vkCreateDescriptorSetLayout( |
| m_device, |
| reinterpret_cast<const VkDescriptorSetLayoutCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDescriptorSetLayout *>(&setLayout))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DescriptorSetLayout, Dispatch>( |
| result, setLayout, VULKAN_HPP_NAMESPACE_STRING "::Device::createDescriptorSetLayoutUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDescriptorSetLayout(VULKAN_HPP_NAMESPACE::DescriptorSetLayout descriptorSetLayout, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDescriptorSetLayout( |
| m_device, static_cast<VkDescriptorSetLayout>(descriptorSetLayout), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDescriptorSetLayout(VULKAN_HPP_NAMESPACE::DescriptorSetLayout descriptorSetLayout, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDescriptorSetLayout( |
| m_device, |
| static_cast<VkDescriptorSetLayout>(descriptorSetLayout), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::DescriptorSetLayout descriptorSetLayout, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDescriptorSetLayout( |
| m_device, static_cast<VkDescriptorSetLayout>(descriptorSetLayout), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::DescriptorSetLayout descriptorSetLayout, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDescriptorSetLayout( |
| m_device, |
| static_cast<VkDescriptorSetLayout>(descriptorSetLayout), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createDescriptorPool(const VULKAN_HPP_NAMESPACE::DescriptorPoolCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::DescriptorPool *pDescriptorPool, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDescriptorPool(m_device, |
| reinterpret_cast<const VkDescriptorPoolCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkDescriptorPool *>(pDescriptorPool))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DescriptorPool>::type |
| Device::createDescriptorPool(const VULKAN_HPP_NAMESPACE::DescriptorPoolCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool; |
| VkResult result = |
| d.vkCreateDescriptorPool(m_device, |
| reinterpret_cast<const VkDescriptorPoolCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDescriptorPool *>(&descriptorPool)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createDescriptorPool"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), descriptorPool); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DescriptorPool, Dispatch>>::type |
| Device::createDescriptorPoolUnique(const VULKAN_HPP_NAMESPACE::DescriptorPoolCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool; |
| Result result = static_cast<Result>( |
| d.vkCreateDescriptorPool(m_device, |
| reinterpret_cast<const VkDescriptorPoolCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDescriptorPool *>(&descriptorPool))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DescriptorPool, Dispatch>( |
| result, descriptorPool, VULKAN_HPP_NAMESPACE_STRING "::Device::createDescriptorPoolUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDescriptorPool(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDescriptorPool(m_device, static_cast<VkDescriptorPool>(descriptorPool), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDescriptorPool(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDescriptorPool(m_device, |
| static_cast<VkDescriptorPool>(descriptorPool), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDescriptorPool(m_device, static_cast<VkDescriptorPool>(descriptorPool), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDescriptorPool(m_device, |
| static_cast<VkDescriptorPool>(descriptorPool), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Result Device::resetDescriptorPool(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| VULKAN_HPP_NAMESPACE::DescriptorPoolResetFlags flags, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkResetDescriptorPool(m_device, static_cast<VkDescriptorPool>(descriptorPool), static_cast<VkDescriptorPoolResetFlags>(flags))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::resetDescriptorPool(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| VULKAN_HPP_NAMESPACE::DescriptorPoolResetFlags flags, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkResetDescriptorPool(m_device, static_cast<VkDescriptorPool>(descriptorPool), static_cast<VkDescriptorPoolResetFlags>(flags)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::allocateDescriptorSets(const VULKAN_HPP_NAMESPACE::DescriptorSetAllocateInfo *pAllocateInfo, |
| VULKAN_HPP_NAMESPACE::DescriptorSet *pDescriptorSets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAllocateDescriptorSets( |
| m_device, reinterpret_cast<const VkDescriptorSetAllocateInfo *>(pAllocateInfo), reinterpret_cast<VkDescriptorSet *>(pDescriptorSets))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DescriptorSetAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VULKAN_HPP_NAMESPACE::DescriptorSet, DescriptorSetAllocator>>::type |
| Device::allocateDescriptorSets(const VULKAN_HPP_NAMESPACE::DescriptorSetAllocateInfo &allocateInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::DescriptorSet, DescriptorSetAllocator> descriptorSets(allocateInfo.descriptorSetCount); |
| VkResult result = d.vkAllocateDescriptorSets( |
| m_device, reinterpret_cast<const VkDescriptorSetAllocateInfo *>(&allocateInfo), reinterpret_cast<VkDescriptorSet *>(descriptorSets.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::allocateDescriptorSets"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), descriptorSets); |
| } |
| |
| template<typename DescriptorSetAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, DescriptorSet>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VULKAN_HPP_NAMESPACE::DescriptorSet, DescriptorSetAllocator>>::type |
| Device::allocateDescriptorSets(const VULKAN_HPP_NAMESPACE::DescriptorSetAllocateInfo &allocateInfo, |
| DescriptorSetAllocator &descriptorSetAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::DescriptorSet, DescriptorSetAllocator> descriptorSets(allocateInfo.descriptorSetCount, descriptorSetAllocator); |
| VkResult result = d.vkAllocateDescriptorSets( |
| m_device, reinterpret_cast<const VkDescriptorSetAllocateInfo *>(&allocateInfo), reinterpret_cast<VkDescriptorSet *>(descriptorSets.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::allocateDescriptorSets"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), descriptorSets); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch, typename DescriptorSetAllocator> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<UniqueHandle<DescriptorSet, Dispatch>, DescriptorSetAllocator>>::type |
| Device::allocateDescriptorSetsUnique(const VULKAN_HPP_NAMESPACE::DescriptorSetAllocateInfo &allocateInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<DescriptorSet, Dispatch>, DescriptorSetAllocator> uniqueDescriptorSets; |
| std::vector<DescriptorSet> descriptorSets(allocateInfo.descriptorSetCount); |
| Result result = static_cast<Result>(d.vkAllocateDescriptorSets( |
| m_device, reinterpret_cast<const VkDescriptorSetAllocateInfo *>(&allocateInfo), reinterpret_cast<VkDescriptorSet *>(descriptorSets.data()))); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| uniqueDescriptorSets.reserve(allocateInfo.descriptorSetCount); |
| PoolFree<Device, DescriptorPool, Dispatch> deleter(*this, allocateInfo.descriptorPool, d); |
| for(size_t i = 0; i < allocateInfo.descriptorSetCount; i++) |
| { |
| uniqueDescriptorSets.push_back(UniqueHandle<DescriptorSet, Dispatch>(descriptorSets[i], deleter)); |
| } |
| } |
| return createResultValue(result, std::move(uniqueDescriptorSets), VULKAN_HPP_NAMESPACE_STRING "::Device::allocateDescriptorSetsUnique"); |
| } |
| |
| template<typename Dispatch, |
| typename DescriptorSetAllocator, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, UniqueHandle<DescriptorSet, Dispatch>>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<UniqueHandle<DescriptorSet, Dispatch>, DescriptorSetAllocator>>::type |
| Device::allocateDescriptorSetsUnique(const VULKAN_HPP_NAMESPACE::DescriptorSetAllocateInfo &allocateInfo, |
| DescriptorSetAllocator &descriptorSetAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<DescriptorSet, Dispatch>, DescriptorSetAllocator> uniqueDescriptorSets(descriptorSetAllocator); |
| std::vector<DescriptorSet> descriptorSets(allocateInfo.descriptorSetCount); |
| Result result = static_cast<Result>(d.vkAllocateDescriptorSets( |
| m_device, reinterpret_cast<const VkDescriptorSetAllocateInfo *>(&allocateInfo), reinterpret_cast<VkDescriptorSet *>(descriptorSets.data()))); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| uniqueDescriptorSets.reserve(allocateInfo.descriptorSetCount); |
| PoolFree<Device, DescriptorPool, Dispatch> deleter(*this, allocateInfo.descriptorPool, d); |
| for(size_t i = 0; i < allocateInfo.descriptorSetCount; i++) |
| { |
| uniqueDescriptorSets.push_back(UniqueHandle<DescriptorSet, Dispatch>(descriptorSets[i], deleter)); |
| } |
| } |
| return createResultValue(result, std::move(uniqueDescriptorSets), VULKAN_HPP_NAMESPACE_STRING "::Device::allocateDescriptorSetsUnique"); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Result Device::freeDescriptorSets(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| uint32_t descriptorSetCount, |
| const VULKAN_HPP_NAMESPACE::DescriptorSet *pDescriptorSets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkFreeDescriptorSets( |
| m_device, static_cast<VkDescriptorPool>(descriptorPool), descriptorSetCount, reinterpret_cast<const VkDescriptorSet *>(pDescriptorSets))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::freeDescriptorSets(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DescriptorSet> const &descriptorSets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkFreeDescriptorSets( |
| m_device, static_cast<VkDescriptorPool>(descriptorPool), descriptorSets.size(), reinterpret_cast<const VkDescriptorSet *>(descriptorSets.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Result(Device::free)(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| uint32_t descriptorSetCount, |
| const VULKAN_HPP_NAMESPACE::DescriptorSet *pDescriptorSets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkFreeDescriptorSets( |
| m_device, static_cast<VkDescriptorPool>(descriptorPool), descriptorSetCount, reinterpret_cast<const VkDescriptorSet *>(pDescriptorSets))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void(Device::free)(VULKAN_HPP_NAMESPACE::DescriptorPool descriptorPool, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DescriptorSet> const &descriptorSets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkFreeDescriptorSets( |
| m_device, static_cast<VkDescriptorPool>(descriptorPool), descriptorSets.size(), reinterpret_cast<const VkDescriptorSet *>(descriptorSets.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::updateDescriptorSets(uint32_t descriptorWriteCount, |
| const VULKAN_HPP_NAMESPACE::WriteDescriptorSet *pDescriptorWrites, |
| uint32_t descriptorCopyCount, |
| const VULKAN_HPP_NAMESPACE::CopyDescriptorSet *pDescriptorCopies, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkUpdateDescriptorSets(m_device, |
| descriptorWriteCount, |
| reinterpret_cast<const VkWriteDescriptorSet *>(pDescriptorWrites), |
| descriptorCopyCount, |
| reinterpret_cast<const VkCopyDescriptorSet *>(pDescriptorCopies)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::updateDescriptorSets(ArrayProxy<const VULKAN_HPP_NAMESPACE::WriteDescriptorSet> const &descriptorWrites, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::CopyDescriptorSet> const &descriptorCopies, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkUpdateDescriptorSets(m_device, |
| descriptorWrites.size(), |
| reinterpret_cast<const VkWriteDescriptorSet *>(descriptorWrites.data()), |
| descriptorCopies.size(), |
| reinterpret_cast<const VkCopyDescriptorSet *>(descriptorCopies.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createFramebuffer(const VULKAN_HPP_NAMESPACE::FramebufferCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Framebuffer *pFramebuffer, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateFramebuffer(m_device, |
| reinterpret_cast<const VkFramebufferCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkFramebuffer *>(pFramebuffer))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Framebuffer>::type |
| Device::createFramebuffer(const VULKAN_HPP_NAMESPACE::FramebufferCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Framebuffer framebuffer; |
| VkResult result = |
| d.vkCreateFramebuffer(m_device, |
| reinterpret_cast<const VkFramebufferCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkFramebuffer *>(&framebuffer)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createFramebuffer"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), framebuffer); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Framebuffer, Dispatch>>::type |
| Device::createFramebufferUnique(const VULKAN_HPP_NAMESPACE::FramebufferCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Framebuffer framebuffer; |
| Result result = static_cast<Result>( |
| d.vkCreateFramebuffer(m_device, |
| reinterpret_cast<const VkFramebufferCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkFramebuffer *>(&framebuffer))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Framebuffer, Dispatch>( |
| result, framebuffer, VULKAN_HPP_NAMESPACE_STRING "::Device::createFramebufferUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyFramebuffer(VULKAN_HPP_NAMESPACE::Framebuffer framebuffer, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyFramebuffer(m_device, static_cast<VkFramebuffer>(framebuffer), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyFramebuffer(VULKAN_HPP_NAMESPACE::Framebuffer framebuffer, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyFramebuffer(m_device, |
| static_cast<VkFramebuffer>(framebuffer), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Framebuffer framebuffer, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyFramebuffer(m_device, static_cast<VkFramebuffer>(framebuffer), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::Framebuffer framebuffer, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyFramebuffer(m_device, |
| static_cast<VkFramebuffer>(framebuffer), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createRenderPass(const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::RenderPass *pRenderPass, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateRenderPass(m_device, |
| reinterpret_cast<const VkRenderPassCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkRenderPass *>(pRenderPass))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::RenderPass>::type |
| Device::createRenderPass(const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::RenderPass renderPass; |
| VkResult result = |
| d.vkCreateRenderPass(m_device, |
| reinterpret_cast<const VkRenderPassCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkRenderPass *>(&renderPass)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createRenderPass"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), renderPass); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::RenderPass, Dispatch>>::type |
| Device::createRenderPassUnique(const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::RenderPass renderPass; |
| Result result = static_cast<Result>( |
| d.vkCreateRenderPass(m_device, |
| reinterpret_cast<const VkRenderPassCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkRenderPass *>(&renderPass))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::RenderPass, Dispatch>( |
| result, renderPass, VULKAN_HPP_NAMESPACE_STRING "::Device::createRenderPassUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyRenderPass(VULKAN_HPP_NAMESPACE::RenderPass renderPass, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyRenderPass(m_device, static_cast<VkRenderPass>(renderPass), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyRenderPass(VULKAN_HPP_NAMESPACE::RenderPass renderPass, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyRenderPass(m_device, |
| static_cast<VkRenderPass>(renderPass), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::RenderPass renderPass, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyRenderPass(m_device, static_cast<VkRenderPass>(renderPass), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::RenderPass renderPass, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyRenderPass(m_device, |
| static_cast<VkRenderPass>(renderPass), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getRenderAreaGranularity(VULKAN_HPP_NAMESPACE::RenderPass renderPass, |
| VULKAN_HPP_NAMESPACE::Extent2D *pGranularity, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetRenderAreaGranularity(m_device, static_cast<VkRenderPass>(renderPass), reinterpret_cast<VkExtent2D *>(pGranularity)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Extent2D Device::getRenderAreaGranularity(VULKAN_HPP_NAMESPACE::RenderPass renderPass, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Extent2D granularity; |
| d.vkGetRenderAreaGranularity(m_device, static_cast<VkRenderPass>(renderPass), reinterpret_cast<VkExtent2D *>(&granularity)); |
| |
| return granularity; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createCommandPool(const VULKAN_HPP_NAMESPACE::CommandPoolCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::CommandPool *pCommandPool, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateCommandPool(m_device, |
| reinterpret_cast<const VkCommandPoolCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkCommandPool *>(pCommandPool))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::CommandPool>::type |
| Device::createCommandPool(const VULKAN_HPP_NAMESPACE::CommandPoolCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::CommandPool commandPool; |
| VkResult result = |
| d.vkCreateCommandPool(m_device, |
| reinterpret_cast<const VkCommandPoolCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkCommandPool *>(&commandPool)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createCommandPool"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), commandPool); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::CommandPool, Dispatch>>::type |
| Device::createCommandPoolUnique(const VULKAN_HPP_NAMESPACE::CommandPoolCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::CommandPool commandPool; |
| Result result = static_cast<Result>( |
| d.vkCreateCommandPool(m_device, |
| reinterpret_cast<const VkCommandPoolCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkCommandPool *>(&commandPool))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::CommandPool, Dispatch>( |
| result, commandPool, VULKAN_HPP_NAMESPACE_STRING "::Device::createCommandPoolUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyCommandPool(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyCommandPool(m_device, static_cast<VkCommandPool>(commandPool), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyCommandPool(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyCommandPool(m_device, |
| static_cast<VkCommandPool>(commandPool), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyCommandPool(m_device, static_cast<VkCommandPool>(commandPool), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyCommandPool(m_device, |
| static_cast<VkCommandPool>(commandPool), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::resetCommandPool(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| VULKAN_HPP_NAMESPACE::CommandPoolResetFlags flags, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkResetCommandPool(m_device, static_cast<VkCommandPool>(commandPool), static_cast<VkCommandPoolResetFlags>(flags))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::resetCommandPool(VULKAN_HPP_NAMESPACE::CommandPool commandPool, VULKAN_HPP_NAMESPACE::CommandPoolResetFlags flags, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkResetCommandPool(m_device, static_cast<VkCommandPool>(commandPool), static_cast<VkCommandPoolResetFlags>(flags)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::resetCommandPool"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::allocateCommandBuffers(const VULKAN_HPP_NAMESPACE::CommandBufferAllocateInfo *pAllocateInfo, |
| VULKAN_HPP_NAMESPACE::CommandBuffer *pCommandBuffers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAllocateCommandBuffers( |
| m_device, reinterpret_cast<const VkCommandBufferAllocateInfo *>(pAllocateInfo), reinterpret_cast<VkCommandBuffer *>(pCommandBuffers))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename CommandBufferAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VULKAN_HPP_NAMESPACE::CommandBuffer, CommandBufferAllocator>>::type |
| Device::allocateCommandBuffers(const VULKAN_HPP_NAMESPACE::CommandBufferAllocateInfo &allocateInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::CommandBuffer, CommandBufferAllocator> commandBuffers(allocateInfo.commandBufferCount); |
| VkResult result = d.vkAllocateCommandBuffers( |
| m_device, reinterpret_cast<const VkCommandBufferAllocateInfo *>(&allocateInfo), reinterpret_cast<VkCommandBuffer *>(commandBuffers.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::allocateCommandBuffers"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), commandBuffers); |
| } |
| |
| template<typename CommandBufferAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, CommandBuffer>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VULKAN_HPP_NAMESPACE::CommandBuffer, CommandBufferAllocator>>::type |
| Device::allocateCommandBuffers(const VULKAN_HPP_NAMESPACE::CommandBufferAllocateInfo &allocateInfo, |
| CommandBufferAllocator &commandBufferAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::CommandBuffer, CommandBufferAllocator> commandBuffers(allocateInfo.commandBufferCount, commandBufferAllocator); |
| VkResult result = d.vkAllocateCommandBuffers( |
| m_device, reinterpret_cast<const VkCommandBufferAllocateInfo *>(&allocateInfo), reinterpret_cast<VkCommandBuffer *>(commandBuffers.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::allocateCommandBuffers"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), commandBuffers); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch, typename CommandBufferAllocator> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<UniqueHandle<CommandBuffer, Dispatch>, CommandBufferAllocator>>::type |
| Device::allocateCommandBuffersUnique(const VULKAN_HPP_NAMESPACE::CommandBufferAllocateInfo &allocateInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<CommandBuffer, Dispatch>, CommandBufferAllocator> uniqueCommandBuffers; |
| std::vector<CommandBuffer> commandBuffers(allocateInfo.commandBufferCount); |
| Result result = static_cast<Result>(d.vkAllocateCommandBuffers( |
| m_device, reinterpret_cast<const VkCommandBufferAllocateInfo *>(&allocateInfo), reinterpret_cast<VkCommandBuffer *>(commandBuffers.data()))); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| uniqueCommandBuffers.reserve(allocateInfo.commandBufferCount); |
| PoolFree<Device, CommandPool, Dispatch> deleter(*this, allocateInfo.commandPool, d); |
| for(size_t i = 0; i < allocateInfo.commandBufferCount; i++) |
| { |
| uniqueCommandBuffers.push_back(UniqueHandle<CommandBuffer, Dispatch>(commandBuffers[i], deleter)); |
| } |
| } |
| return createResultValue(result, std::move(uniqueCommandBuffers), VULKAN_HPP_NAMESPACE_STRING "::Device::allocateCommandBuffersUnique"); |
| } |
| |
| template<typename Dispatch, |
| typename CommandBufferAllocator, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, UniqueHandle<CommandBuffer, Dispatch>>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<UniqueHandle<CommandBuffer, Dispatch>, CommandBufferAllocator>>::type |
| Device::allocateCommandBuffersUnique(const VULKAN_HPP_NAMESPACE::CommandBufferAllocateInfo &allocateInfo, |
| CommandBufferAllocator &commandBufferAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<CommandBuffer, Dispatch>, CommandBufferAllocator> uniqueCommandBuffers(commandBufferAllocator); |
| std::vector<CommandBuffer> commandBuffers(allocateInfo.commandBufferCount); |
| Result result = static_cast<Result>(d.vkAllocateCommandBuffers( |
| m_device, reinterpret_cast<const VkCommandBufferAllocateInfo *>(&allocateInfo), reinterpret_cast<VkCommandBuffer *>(commandBuffers.data()))); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| uniqueCommandBuffers.reserve(allocateInfo.commandBufferCount); |
| PoolFree<Device, CommandPool, Dispatch> deleter(*this, allocateInfo.commandPool, d); |
| for(size_t i = 0; i < allocateInfo.commandBufferCount; i++) |
| { |
| uniqueCommandBuffers.push_back(UniqueHandle<CommandBuffer, Dispatch>(commandBuffers[i], deleter)); |
| } |
| } |
| return createResultValue(result, std::move(uniqueCommandBuffers), VULKAN_HPP_NAMESPACE_STRING "::Device::allocateCommandBuffersUnique"); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::freeCommandBuffers(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| uint32_t commandBufferCount, |
| const VULKAN_HPP_NAMESPACE::CommandBuffer *pCommandBuffers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkFreeCommandBuffers( |
| m_device, static_cast<VkCommandPool>(commandPool), commandBufferCount, reinterpret_cast<const VkCommandBuffer *>(pCommandBuffers)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::freeCommandBuffers(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::CommandBuffer> const &commandBuffers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkFreeCommandBuffers( |
| m_device, static_cast<VkCommandPool>(commandPool), commandBuffers.size(), reinterpret_cast<const VkCommandBuffer *>(commandBuffers.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void(Device::free)(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| uint32_t commandBufferCount, |
| const VULKAN_HPP_NAMESPACE::CommandBuffer *pCommandBuffers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkFreeCommandBuffers( |
| m_device, static_cast<VkCommandPool>(commandPool), commandBufferCount, reinterpret_cast<const VkCommandBuffer *>(pCommandBuffers)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void(Device::free)(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::CommandBuffer> const &commandBuffers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkFreeCommandBuffers( |
| m_device, static_cast<VkCommandPool>(commandPool), commandBuffers.size(), reinterpret_cast<const VkCommandBuffer *>(commandBuffers.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result CommandBuffer::begin(const VULKAN_HPP_NAMESPACE::CommandBufferBeginInfo *pBeginInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkBeginCommandBuffer(m_commandBuffer, reinterpret_cast<const VkCommandBufferBeginInfo *>(pBeginInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| CommandBuffer::begin(const VULKAN_HPP_NAMESPACE::CommandBufferBeginInfo &beginInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkBeginCommandBuffer(m_commandBuffer, reinterpret_cast<const VkCommandBufferBeginInfo *>(&beginInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::begin"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result CommandBuffer::end(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkEndCommandBuffer(m_commandBuffer)); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type CommandBuffer::end(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkEndCommandBuffer(m_commandBuffer); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::end"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result CommandBuffer::reset(VULKAN_HPP_NAMESPACE::CommandBufferResetFlags flags, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkResetCommandBuffer(m_commandBuffer, static_cast<VkCommandBufferResetFlags>(flags))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE typename ResultValueType<void>::type CommandBuffer::reset(VULKAN_HPP_NAMESPACE::CommandBufferResetFlags flags, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkResetCommandBuffer(m_commandBuffer, static_cast<VkCommandBufferResetFlags>(flags)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::reset"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindPipeline(VULKAN_HPP_NAMESPACE::PipelineBindPoint pipelineBindPoint, |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindPipeline(m_commandBuffer, static_cast<VkPipelineBindPoint>(pipelineBindPoint), static_cast<VkPipeline>(pipeline)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setViewport(uint32_t firstViewport, |
| uint32_t viewportCount, |
| const VULKAN_HPP_NAMESPACE::Viewport *pViewports, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetViewport(m_commandBuffer, firstViewport, viewportCount, reinterpret_cast<const VkViewport *>(pViewports)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setViewport(uint32_t firstViewport, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Viewport> const &viewports, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetViewport(m_commandBuffer, firstViewport, viewports.size(), reinterpret_cast<const VkViewport *>(viewports.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setScissor(uint32_t firstScissor, |
| uint32_t scissorCount, |
| const VULKAN_HPP_NAMESPACE::Rect2D *pScissors, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetScissor(m_commandBuffer, firstScissor, scissorCount, reinterpret_cast<const VkRect2D *>(pScissors)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setScissor(uint32_t firstScissor, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D> const &scissors, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetScissor(m_commandBuffer, firstScissor, scissors.size(), reinterpret_cast<const VkRect2D *>(scissors.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setLineWidth(float lineWidth, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetLineWidth(m_commandBuffer, lineWidth); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::setDepthBias(float depthBiasConstantFactor, float depthBiasClamp, float depthBiasSlopeFactor, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthBias(m_commandBuffer, depthBiasConstantFactor, depthBiasClamp, depthBiasSlopeFactor); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setBlendConstants(const float blendConstants[4], Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetBlendConstants(m_commandBuffer, blendConstants); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthBounds(float minDepthBounds, float maxDepthBounds, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthBounds(m_commandBuffer, minDepthBounds, maxDepthBounds); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::setStencilCompareMask(VULKAN_HPP_NAMESPACE::StencilFaceFlags faceMask, uint32_t compareMask, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetStencilCompareMask(m_commandBuffer, static_cast<VkStencilFaceFlags>(faceMask), compareMask); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::setStencilWriteMask(VULKAN_HPP_NAMESPACE::StencilFaceFlags faceMask, uint32_t writeMask, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetStencilWriteMask(m_commandBuffer, static_cast<VkStencilFaceFlags>(faceMask), writeMask); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::setStencilReference(VULKAN_HPP_NAMESPACE::StencilFaceFlags faceMask, uint32_t reference, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetStencilReference(m_commandBuffer, static_cast<VkStencilFaceFlags>(faceMask), reference); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindDescriptorSets(VULKAN_HPP_NAMESPACE::PipelineBindPoint pipelineBindPoint, |
| VULKAN_HPP_NAMESPACE::PipelineLayout layout, |
| uint32_t firstSet, |
| uint32_t descriptorSetCount, |
| const VULKAN_HPP_NAMESPACE::DescriptorSet *pDescriptorSets, |
| uint32_t dynamicOffsetCount, |
| const uint32_t *pDynamicOffsets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindDescriptorSets(m_commandBuffer, |
| static_cast<VkPipelineBindPoint>(pipelineBindPoint), |
| static_cast<VkPipelineLayout>(layout), |
| firstSet, |
| descriptorSetCount, |
| reinterpret_cast<const VkDescriptorSet *>(pDescriptorSets), |
| dynamicOffsetCount, |
| pDynamicOffsets); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindDescriptorSets(VULKAN_HPP_NAMESPACE::PipelineBindPoint pipelineBindPoint, |
| VULKAN_HPP_NAMESPACE::PipelineLayout layout, |
| uint32_t firstSet, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DescriptorSet> const &descriptorSets, |
| ArrayProxy<const uint32_t> const &dynamicOffsets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBindDescriptorSets(m_commandBuffer, |
| static_cast<VkPipelineBindPoint>(pipelineBindPoint), |
| static_cast<VkPipelineLayout>(layout), |
| firstSet, |
| descriptorSets.size(), |
| reinterpret_cast<const VkDescriptorSet *>(descriptorSets.data()), |
| dynamicOffsets.size(), |
| dynamicOffsets.data()); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindIndexBuffer(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::IndexType indexType, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindIndexBuffer(m_commandBuffer, static_cast<VkBuffer>(buffer), static_cast<VkDeviceSize>(offset), static_cast<VkIndexType>(indexType)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindVertexBuffers(uint32_t firstBinding, |
| uint32_t bindingCount, |
| const VULKAN_HPP_NAMESPACE::Buffer *pBuffers, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pOffsets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindVertexBuffers( |
| m_commandBuffer, firstBinding, bindingCount, reinterpret_cast<const VkBuffer *>(pBuffers), reinterpret_cast<const VkDeviceSize *>(pOffsets)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindVertexBuffers(uint32_t firstBinding, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer> const &buffers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &offsets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(buffers.size() == offsets.size()); |
| # else |
| if(buffers.size() != offsets.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers: buffers.size() != offsets.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdBindVertexBuffers(m_commandBuffer, |
| firstBinding, |
| buffers.size(), |
| reinterpret_cast<const VkBuffer *>(buffers.data()), |
| reinterpret_cast<const VkDeviceSize *>(offsets.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::draw( |
| uint32_t vertexCount, uint32_t instanceCount, uint32_t firstVertex, uint32_t firstInstance, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDraw(m_commandBuffer, vertexCount, instanceCount, firstVertex, firstInstance); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndexed(uint32_t indexCount, |
| uint32_t instanceCount, |
| uint32_t firstIndex, |
| int32_t vertexOffset, |
| uint32_t firstInstance, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndexed(m_commandBuffer, indexCount, instanceCount, firstIndex, vertexOffset, firstInstance); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndirect(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| uint32_t drawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndirect(m_commandBuffer, static_cast<VkBuffer>(buffer), static_cast<VkDeviceSize>(offset), drawCount, stride); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndexedIndirect(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| uint32_t drawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndexedIndirect(m_commandBuffer, static_cast<VkBuffer>(buffer), static_cast<VkDeviceSize>(offset), drawCount, stride); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::dispatch(uint32_t groupCountX, uint32_t groupCountY, uint32_t groupCountZ, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDispatch(m_commandBuffer, groupCountX, groupCountY, groupCountZ); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::dispatchIndirect(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDispatchIndirect(m_commandBuffer, static_cast<VkBuffer>(buffer), static_cast<VkDeviceSize>(offset)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBuffer(VULKAN_HPP_NAMESPACE::Buffer srcBuffer, |
| VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| uint32_t regionCount, |
| const VULKAN_HPP_NAMESPACE::BufferCopy *pRegions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyBuffer(m_commandBuffer, |
| static_cast<VkBuffer>(srcBuffer), |
| static_cast<VkBuffer>(dstBuffer), |
| regionCount, |
| reinterpret_cast<const VkBufferCopy *>(pRegions)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBuffer(VULKAN_HPP_NAMESPACE::Buffer srcBuffer, |
| VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferCopy> const ®ions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyBuffer(m_commandBuffer, |
| static_cast<VkBuffer>(srcBuffer), |
| static_cast<VkBuffer>(dstBuffer), |
| regions.size(), |
| reinterpret_cast<const VkBufferCopy *>(regions.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImage(VULKAN_HPP_NAMESPACE::Image srcImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout srcImageLayout, |
| VULKAN_HPP_NAMESPACE::Image dstImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout dstImageLayout, |
| uint32_t regionCount, |
| const VULKAN_HPP_NAMESPACE::ImageCopy *pRegions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyImage(m_commandBuffer, |
| static_cast<VkImage>(srcImage), |
| static_cast<VkImageLayout>(srcImageLayout), |
| static_cast<VkImage>(dstImage), |
| static_cast<VkImageLayout>(dstImageLayout), |
| regionCount, |
| reinterpret_cast<const VkImageCopy *>(pRegions)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImage(VULKAN_HPP_NAMESPACE::Image srcImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout srcImageLayout, |
| VULKAN_HPP_NAMESPACE::Image dstImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout dstImageLayout, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageCopy> const ®ions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyImage(m_commandBuffer, |
| static_cast<VkImage>(srcImage), |
| static_cast<VkImageLayout>(srcImageLayout), |
| static_cast<VkImage>(dstImage), |
| static_cast<VkImageLayout>(dstImageLayout), |
| regions.size(), |
| reinterpret_cast<const VkImageCopy *>(regions.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::blitImage(VULKAN_HPP_NAMESPACE::Image srcImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout srcImageLayout, |
| VULKAN_HPP_NAMESPACE::Image dstImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout dstImageLayout, |
| uint32_t regionCount, |
| const VULKAN_HPP_NAMESPACE::ImageBlit *pRegions, |
| VULKAN_HPP_NAMESPACE::Filter filter, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBlitImage(m_commandBuffer, |
| static_cast<VkImage>(srcImage), |
| static_cast<VkImageLayout>(srcImageLayout), |
| static_cast<VkImage>(dstImage), |
| static_cast<VkImageLayout>(dstImageLayout), |
| regionCount, |
| reinterpret_cast<const VkImageBlit *>(pRegions), |
| static_cast<VkFilter>(filter)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::blitImage(VULKAN_HPP_NAMESPACE::Image srcImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout srcImageLayout, |
| VULKAN_HPP_NAMESPACE::Image dstImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout dstImageLayout, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageBlit> const ®ions, |
| VULKAN_HPP_NAMESPACE::Filter filter, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBlitImage(m_commandBuffer, |
| static_cast<VkImage>(srcImage), |
| static_cast<VkImageLayout>(srcImageLayout), |
| static_cast<VkImage>(dstImage), |
| static_cast<VkImageLayout>(dstImageLayout), |
| regions.size(), |
| reinterpret_cast<const VkImageBlit *>(regions.data()), |
| static_cast<VkFilter>(filter)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage(VULKAN_HPP_NAMESPACE::Buffer srcBuffer, |
| VULKAN_HPP_NAMESPACE::Image dstImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout dstImageLayout, |
| uint32_t regionCount, |
| const VULKAN_HPP_NAMESPACE::BufferImageCopy *pRegions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyBufferToImage(m_commandBuffer, |
| static_cast<VkBuffer>(srcBuffer), |
| static_cast<VkImage>(dstImage), |
| static_cast<VkImageLayout>(dstImageLayout), |
| regionCount, |
| reinterpret_cast<const VkBufferImageCopy *>(pRegions)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage(VULKAN_HPP_NAMESPACE::Buffer srcBuffer, |
| VULKAN_HPP_NAMESPACE::Image dstImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout dstImageLayout, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferImageCopy> const ®ions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyBufferToImage(m_commandBuffer, |
| static_cast<VkBuffer>(srcBuffer), |
| static_cast<VkImage>(dstImage), |
| static_cast<VkImageLayout>(dstImageLayout), |
| regions.size(), |
| reinterpret_cast<const VkBufferImageCopy *>(regions.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer(VULKAN_HPP_NAMESPACE::Image srcImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout srcImageLayout, |
| VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| uint32_t regionCount, |
| const VULKAN_HPP_NAMESPACE::BufferImageCopy *pRegions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyImageToBuffer(m_commandBuffer, |
| static_cast<VkImage>(srcImage), |
| static_cast<VkImageLayout>(srcImageLayout), |
| static_cast<VkBuffer>(dstBuffer), |
| regionCount, |
| reinterpret_cast<const VkBufferImageCopy *>(pRegions)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer(VULKAN_HPP_NAMESPACE::Image srcImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout srcImageLayout, |
| VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferImageCopy> const ®ions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyImageToBuffer(m_commandBuffer, |
| static_cast<VkImage>(srcImage), |
| static_cast<VkImageLayout>(srcImageLayout), |
| static_cast<VkBuffer>(dstBuffer), |
| regions.size(), |
| reinterpret_cast<const VkBufferImageCopy *>(regions.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::updateBuffer(VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize dstOffset, |
| VULKAN_HPP_NAMESPACE::DeviceSize dataSize, |
| const void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdUpdateBuffer( |
| m_commandBuffer, static_cast<VkBuffer>(dstBuffer), static_cast<VkDeviceSize>(dstOffset), static_cast<VkDeviceSize>(dataSize), pData); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::updateBuffer(VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize dstOffset, |
| ArrayProxy<const DataType> const &data, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdUpdateBuffer(m_commandBuffer, |
| static_cast<VkBuffer>(dstBuffer), |
| static_cast<VkDeviceSize>(dstOffset), |
| data.size() * sizeof(DataType), |
| reinterpret_cast<const void *>(data.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::fillBuffer(VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize dstOffset, |
| VULKAN_HPP_NAMESPACE::DeviceSize size, |
| uint32_t data, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdFillBuffer(m_commandBuffer, static_cast<VkBuffer>(dstBuffer), static_cast<VkDeviceSize>(dstOffset), static_cast<VkDeviceSize>(size), data); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::clearColorImage(VULKAN_HPP_NAMESPACE::Image image, |
| VULKAN_HPP_NAMESPACE::ImageLayout imageLayout, |
| const VULKAN_HPP_NAMESPACE::ClearColorValue *pColor, |
| uint32_t rangeCount, |
| const VULKAN_HPP_NAMESPACE::ImageSubresourceRange *pRanges, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdClearColorImage(m_commandBuffer, |
| static_cast<VkImage>(image), |
| static_cast<VkImageLayout>(imageLayout), |
| reinterpret_cast<const VkClearColorValue *>(pColor), |
| rangeCount, |
| reinterpret_cast<const VkImageSubresourceRange *>(pRanges)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::clearColorImage(VULKAN_HPP_NAMESPACE::Image image, |
| VULKAN_HPP_NAMESPACE::ImageLayout imageLayout, |
| const VULKAN_HPP_NAMESPACE::ClearColorValue &color, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageSubresourceRange> const &ranges, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdClearColorImage(m_commandBuffer, |
| static_cast<VkImage>(image), |
| static_cast<VkImageLayout>(imageLayout), |
| reinterpret_cast<const VkClearColorValue *>(&color), |
| ranges.size(), |
| reinterpret_cast<const VkImageSubresourceRange *>(ranges.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::clearDepthStencilImage(VULKAN_HPP_NAMESPACE::Image image, |
| VULKAN_HPP_NAMESPACE::ImageLayout imageLayout, |
| const VULKAN_HPP_NAMESPACE::ClearDepthStencilValue *pDepthStencil, |
| uint32_t rangeCount, |
| const VULKAN_HPP_NAMESPACE::ImageSubresourceRange *pRanges, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdClearDepthStencilImage(m_commandBuffer, |
| static_cast<VkImage>(image), |
| static_cast<VkImageLayout>(imageLayout), |
| reinterpret_cast<const VkClearDepthStencilValue *>(pDepthStencil), |
| rangeCount, |
| reinterpret_cast<const VkImageSubresourceRange *>(pRanges)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::clearDepthStencilImage(VULKAN_HPP_NAMESPACE::Image image, |
| VULKAN_HPP_NAMESPACE::ImageLayout imageLayout, |
| const VULKAN_HPP_NAMESPACE::ClearDepthStencilValue &depthStencil, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageSubresourceRange> const &ranges, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdClearDepthStencilImage(m_commandBuffer, |
| static_cast<VkImage>(image), |
| static_cast<VkImageLayout>(imageLayout), |
| reinterpret_cast<const VkClearDepthStencilValue *>(&depthStencil), |
| ranges.size(), |
| reinterpret_cast<const VkImageSubresourceRange *>(ranges.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::clearAttachments(uint32_t attachmentCount, |
| const VULKAN_HPP_NAMESPACE::ClearAttachment *pAttachments, |
| uint32_t rectCount, |
| const VULKAN_HPP_NAMESPACE::ClearRect *pRects, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdClearAttachments(m_commandBuffer, |
| attachmentCount, |
| reinterpret_cast<const VkClearAttachment *>(pAttachments), |
| rectCount, |
| reinterpret_cast<const VkClearRect *>(pRects)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::clearAttachments(ArrayProxy<const VULKAN_HPP_NAMESPACE::ClearAttachment> const &attachments, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ClearRect> const &rects, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdClearAttachments(m_commandBuffer, |
| attachments.size(), |
| reinterpret_cast<const VkClearAttachment *>(attachments.data()), |
| rects.size(), |
| reinterpret_cast<const VkClearRect *>(rects.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resolveImage(VULKAN_HPP_NAMESPACE::Image srcImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout srcImageLayout, |
| VULKAN_HPP_NAMESPACE::Image dstImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout dstImageLayout, |
| uint32_t regionCount, |
| const VULKAN_HPP_NAMESPACE::ImageResolve *pRegions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdResolveImage(m_commandBuffer, |
| static_cast<VkImage>(srcImage), |
| static_cast<VkImageLayout>(srcImageLayout), |
| static_cast<VkImage>(dstImage), |
| static_cast<VkImageLayout>(dstImageLayout), |
| regionCount, |
| reinterpret_cast<const VkImageResolve *>(pRegions)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resolveImage(VULKAN_HPP_NAMESPACE::Image srcImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout srcImageLayout, |
| VULKAN_HPP_NAMESPACE::Image dstImage, |
| VULKAN_HPP_NAMESPACE::ImageLayout dstImageLayout, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageResolve> const ®ions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdResolveImage(m_commandBuffer, |
| static_cast<VkImage>(srcImage), |
| static_cast<VkImageLayout>(srcImageLayout), |
| static_cast<VkImage>(dstImage), |
| static_cast<VkImageLayout>(dstImageLayout), |
| regions.size(), |
| reinterpret_cast<const VkImageResolve *>(regions.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setEvent(VULKAN_HPP_NAMESPACE::Event event, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags stageMask, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetEvent(m_commandBuffer, static_cast<VkEvent>(event), static_cast<VkPipelineStageFlags>(stageMask)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resetEvent(VULKAN_HPP_NAMESPACE::Event event, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags stageMask, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdResetEvent(m_commandBuffer, static_cast<VkEvent>(event), static_cast<VkPipelineStageFlags>(stageMask)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::waitEvents(uint32_t eventCount, |
| const VULKAN_HPP_NAMESPACE::Event *pEvents, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags srcStageMask, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags dstStageMask, |
| uint32_t memoryBarrierCount, |
| const VULKAN_HPP_NAMESPACE::MemoryBarrier *pMemoryBarriers, |
| uint32_t bufferMemoryBarrierCount, |
| const VULKAN_HPP_NAMESPACE::BufferMemoryBarrier *pBufferMemoryBarriers, |
| uint32_t imageMemoryBarrierCount, |
| const VULKAN_HPP_NAMESPACE::ImageMemoryBarrier *pImageMemoryBarriers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWaitEvents(m_commandBuffer, |
| eventCount, |
| reinterpret_cast<const VkEvent *>(pEvents), |
| static_cast<VkPipelineStageFlags>(srcStageMask), |
| static_cast<VkPipelineStageFlags>(dstStageMask), |
| memoryBarrierCount, |
| reinterpret_cast<const VkMemoryBarrier *>(pMemoryBarriers), |
| bufferMemoryBarrierCount, |
| reinterpret_cast<const VkBufferMemoryBarrier *>(pBufferMemoryBarriers), |
| imageMemoryBarrierCount, |
| reinterpret_cast<const VkImageMemoryBarrier *>(pImageMemoryBarriers)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::waitEvents(ArrayProxy<const VULKAN_HPP_NAMESPACE::Event> const &events, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags srcStageMask, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags dstStageMask, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::MemoryBarrier> const &memoryBarriers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferMemoryBarrier> const &bufferMemoryBarriers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageMemoryBarrier> const &imageMemoryBarriers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdWaitEvents(m_commandBuffer, |
| events.size(), |
| reinterpret_cast<const VkEvent *>(events.data()), |
| static_cast<VkPipelineStageFlags>(srcStageMask), |
| static_cast<VkPipelineStageFlags>(dstStageMask), |
| memoryBarriers.size(), |
| reinterpret_cast<const VkMemoryBarrier *>(memoryBarriers.data()), |
| bufferMemoryBarriers.size(), |
| reinterpret_cast<const VkBufferMemoryBarrier *>(bufferMemoryBarriers.data()), |
| imageMemoryBarriers.size(), |
| reinterpret_cast<const VkImageMemoryBarrier *>(imageMemoryBarriers.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier(VULKAN_HPP_NAMESPACE::PipelineStageFlags srcStageMask, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags dstStageMask, |
| VULKAN_HPP_NAMESPACE::DependencyFlags dependencyFlags, |
| uint32_t memoryBarrierCount, |
| const VULKAN_HPP_NAMESPACE::MemoryBarrier *pMemoryBarriers, |
| uint32_t bufferMemoryBarrierCount, |
| const VULKAN_HPP_NAMESPACE::BufferMemoryBarrier *pBufferMemoryBarriers, |
| uint32_t imageMemoryBarrierCount, |
| const VULKAN_HPP_NAMESPACE::ImageMemoryBarrier *pImageMemoryBarriers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdPipelineBarrier(m_commandBuffer, |
| static_cast<VkPipelineStageFlags>(srcStageMask), |
| static_cast<VkPipelineStageFlags>(dstStageMask), |
| static_cast<VkDependencyFlags>(dependencyFlags), |
| memoryBarrierCount, |
| reinterpret_cast<const VkMemoryBarrier *>(pMemoryBarriers), |
| bufferMemoryBarrierCount, |
| reinterpret_cast<const VkBufferMemoryBarrier *>(pBufferMemoryBarriers), |
| imageMemoryBarrierCount, |
| reinterpret_cast<const VkImageMemoryBarrier *>(pImageMemoryBarriers)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier(VULKAN_HPP_NAMESPACE::PipelineStageFlags srcStageMask, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags dstStageMask, |
| VULKAN_HPP_NAMESPACE::DependencyFlags dependencyFlags, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::MemoryBarrier> const &memoryBarriers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::BufferMemoryBarrier> const &bufferMemoryBarriers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ImageMemoryBarrier> const &imageMemoryBarriers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdPipelineBarrier(m_commandBuffer, |
| static_cast<VkPipelineStageFlags>(srcStageMask), |
| static_cast<VkPipelineStageFlags>(dstStageMask), |
| static_cast<VkDependencyFlags>(dependencyFlags), |
| memoryBarriers.size(), |
| reinterpret_cast<const VkMemoryBarrier *>(memoryBarriers.data()), |
| bufferMemoryBarriers.size(), |
| reinterpret_cast<const VkBufferMemoryBarrier *>(bufferMemoryBarriers.data()), |
| imageMemoryBarriers.size(), |
| reinterpret_cast<const VkImageMemoryBarrier *>(imageMemoryBarriers.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginQuery(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t query, |
| VULKAN_HPP_NAMESPACE::QueryControlFlags flags, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginQuery(m_commandBuffer, static_cast<VkQueryPool>(queryPool), query, static_cast<VkQueryControlFlags>(flags)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endQuery(VULKAN_HPP_NAMESPACE::QueryPool queryPool, uint32_t query, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndQuery(m_commandBuffer, static_cast<VkQueryPool>(queryPool), query); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resetQueryPool(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| uint32_t queryCount, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdResetQueryPool(m_commandBuffer, static_cast<VkQueryPool>(queryPool), firstQuery, queryCount); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::writeTimestamp(VULKAN_HPP_NAMESPACE::PipelineStageFlagBits pipelineStage, |
| VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t query, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWriteTimestamp(m_commandBuffer, static_cast<VkPipelineStageFlagBits>(pipelineStage), static_cast<VkQueryPool>(queryPool), query); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyQueryPoolResults(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| uint32_t queryCount, |
| VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize dstOffset, |
| VULKAN_HPP_NAMESPACE::DeviceSize stride, |
| VULKAN_HPP_NAMESPACE::QueryResultFlags flags, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyQueryPoolResults(m_commandBuffer, |
| static_cast<VkQueryPool>(queryPool), |
| firstQuery, |
| queryCount, |
| static_cast<VkBuffer>(dstBuffer), |
| static_cast<VkDeviceSize>(dstOffset), |
| static_cast<VkDeviceSize>(stride), |
| static_cast<VkQueryResultFlags>(flags)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pushConstants(VULKAN_HPP_NAMESPACE::PipelineLayout layout, |
| VULKAN_HPP_NAMESPACE::ShaderStageFlags stageFlags, |
| uint32_t offset, |
| uint32_t size, |
| const void *pValues, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdPushConstants(m_commandBuffer, static_cast<VkPipelineLayout>(layout), static_cast<VkShaderStageFlags>(stageFlags), offset, size, pValues); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename ValuesType, typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pushConstants(VULKAN_HPP_NAMESPACE::PipelineLayout layout, |
| VULKAN_HPP_NAMESPACE::ShaderStageFlags stageFlags, |
| uint32_t offset, |
| ArrayProxy<const ValuesType> const &values, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdPushConstants(m_commandBuffer, |
| static_cast<VkPipelineLayout>(layout), |
| static_cast<VkShaderStageFlags>(stageFlags), |
| offset, |
| values.size() * sizeof(ValuesType), |
| reinterpret_cast<const void *>(values.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass(const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo *pRenderPassBegin, |
| VULKAN_HPP_NAMESPACE::SubpassContents contents, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginRenderPass(m_commandBuffer, reinterpret_cast<const VkRenderPassBeginInfo *>(pRenderPassBegin), static_cast<VkSubpassContents>(contents)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass(const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo &renderPassBegin, |
| VULKAN_HPP_NAMESPACE::SubpassContents contents, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBeginRenderPass(m_commandBuffer, reinterpret_cast<const VkRenderPassBeginInfo *>(&renderPassBegin), static_cast<VkSubpassContents>(contents)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::nextSubpass(VULKAN_HPP_NAMESPACE::SubpassContents contents, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdNextSubpass(m_commandBuffer, static_cast<VkSubpassContents>(contents)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endRenderPass(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndRenderPass(m_commandBuffer); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::executeCommands(uint32_t commandBufferCount, |
| const VULKAN_HPP_NAMESPACE::CommandBuffer *pCommandBuffers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdExecuteCommands(m_commandBuffer, commandBufferCount, reinterpret_cast<const VkCommandBuffer *>(pCommandBuffers)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::executeCommands(ArrayProxy<const VULKAN_HPP_NAMESPACE::CommandBuffer> const &commandBuffers, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdExecuteCommands(m_commandBuffer, commandBuffers.size(), reinterpret_cast<const VkCommandBuffer *>(commandBuffers.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_VERSION_1_1 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result enumerateInstanceVersion(uint32_t *pApiVersion, Dispatch const &d) VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkEnumerateInstanceVersion(pApiVersion)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<uint32_t>::type enumerateInstanceVersion(Dispatch const &d) |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint32_t apiVersion; |
| VkResult result = d.vkEnumerateInstanceVersion(&apiVersion); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::enumerateInstanceVersion"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), apiVersion); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::bindBufferMemory2(uint32_t bindInfoCount, |
| const VULKAN_HPP_NAMESPACE::BindBufferMemoryInfo *pBindInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkBindBufferMemory2(m_device, bindInfoCount, reinterpret_cast<const VkBindBufferMemoryInfo *>(pBindInfos))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::bindBufferMemory2(ArrayProxy<const VULKAN_HPP_NAMESPACE::BindBufferMemoryInfo> const &bindInfos, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkBindBufferMemory2(m_device, bindInfos.size(), reinterpret_cast<const VkBindBufferMemoryInfo *>(bindInfos.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::bindBufferMemory2"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::bindImageMemory2(uint32_t bindInfoCount, |
| const VULKAN_HPP_NAMESPACE::BindImageMemoryInfo *pBindInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkBindImageMemory2(m_device, bindInfoCount, reinterpret_cast<const VkBindImageMemoryInfo *>(pBindInfos))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::bindImageMemory2(ArrayProxy<const VULKAN_HPP_NAMESPACE::BindImageMemoryInfo> const &bindInfos, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkBindImageMemory2(m_device, bindInfos.size(), reinterpret_cast<const VkBindImageMemoryInfo *>(bindInfos.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::bindImageMemory2"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getGroupPeerMemoryFeatures(uint32_t heapIndex, |
| uint32_t localDeviceIndex, |
| uint32_t remoteDeviceIndex, |
| VULKAN_HPP_NAMESPACE::PeerMemoryFeatureFlags *pPeerMemoryFeatures, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceGroupPeerMemoryFeatures( |
| m_device, heapIndex, localDeviceIndex, remoteDeviceIndex, reinterpret_cast<VkPeerMemoryFeatureFlags *>(pPeerMemoryFeatures)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PeerMemoryFeatureFlags Device::getGroupPeerMemoryFeatures( |
| uint32_t heapIndex, uint32_t localDeviceIndex, uint32_t remoteDeviceIndex, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PeerMemoryFeatureFlags peerMemoryFeatures; |
| d.vkGetDeviceGroupPeerMemoryFeatures( |
| m_device, heapIndex, localDeviceIndex, remoteDeviceIndex, reinterpret_cast<VkPeerMemoryFeatureFlags *>(&peerMemoryFeatures)); |
| |
| return peerMemoryFeatures; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDeviceMask(uint32_t deviceMask, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDeviceMask(m_commandBuffer, deviceMask); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::dispatchBase(uint32_t baseGroupX, |
| uint32_t baseGroupY, |
| uint32_t baseGroupZ, |
| uint32_t groupCountX, |
| uint32_t groupCountY, |
| uint32_t groupCountZ, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDispatchBase(m_commandBuffer, baseGroupX, baseGroupY, baseGroupZ, groupCountX, groupCountY, groupCountZ); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Instance::enumeratePhysicalDeviceGroups(uint32_t *pPhysicalDeviceGroupCount, |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties *pPhysicalDeviceGroupProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkEnumeratePhysicalDeviceGroups( |
| m_instance, pPhysicalDeviceGroupCount, reinterpret_cast<VkPhysicalDeviceGroupProperties *>(pPhysicalDeviceGroupProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PhysicalDeviceGroupPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator>>::type |
| Instance::enumeratePhysicalDeviceGroups(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator> physicalDeviceGroupProperties; |
| uint32_t physicalDeviceGroupCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceGroups(m_instance, &physicalDeviceGroupCount, nullptr)); |
| if((result == Result::eSuccess) && physicalDeviceGroupCount) |
| { |
| physicalDeviceGroupProperties.resize(physicalDeviceGroupCount); |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceGroups( |
| m_instance, &physicalDeviceGroupCount, reinterpret_cast<VkPhysicalDeviceGroupProperties *>(physicalDeviceGroupProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(physicalDeviceGroupCount <= physicalDeviceGroupProperties.size()); |
| if(physicalDeviceGroupCount < physicalDeviceGroupProperties.size()) |
| { |
| physicalDeviceGroupProperties.resize(physicalDeviceGroupCount); |
| } |
| } |
| return createResultValue(result, physicalDeviceGroupProperties, VULKAN_HPP_NAMESPACE_STRING "::Instance::enumeratePhysicalDeviceGroups"); |
| } |
| |
| template<typename PhysicalDeviceGroupPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PhysicalDeviceGroupProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator>>::type |
| Instance::enumeratePhysicalDeviceGroups(PhysicalDeviceGroupPropertiesAllocator &physicalDeviceGroupPropertiesAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator> physicalDeviceGroupProperties(physicalDeviceGroupPropertiesAllocator); |
| uint32_t physicalDeviceGroupCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceGroups(m_instance, &physicalDeviceGroupCount, nullptr)); |
| if((result == Result::eSuccess) && physicalDeviceGroupCount) |
| { |
| physicalDeviceGroupProperties.resize(physicalDeviceGroupCount); |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceGroups( |
| m_instance, &physicalDeviceGroupCount, reinterpret_cast<VkPhysicalDeviceGroupProperties *>(physicalDeviceGroupProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(physicalDeviceGroupCount <= physicalDeviceGroupProperties.size()); |
| if(physicalDeviceGroupCount < physicalDeviceGroupProperties.size()) |
| { |
| physicalDeviceGroupProperties.resize(physicalDeviceGroupCount); |
| } |
| } |
| return createResultValue(result, physicalDeviceGroupProperties, VULKAN_HPP_NAMESPACE_STRING "::Instance::enumeratePhysicalDeviceGroups"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageMemoryRequirements2(const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetImageMemoryRequirements2( |
| m_device, reinterpret_cast<const VkImageMemoryRequirementsInfo2 *>(pInfo), reinterpret_cast<VkMemoryRequirements2 *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 |
| Device::getImageMemoryRequirements2(const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements; |
| d.vkGetImageMemoryRequirements2( |
| m_device, reinterpret_cast<const VkImageMemoryRequirementsInfo2 *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getImageMemoryRequirements2(const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>(); |
| d.vkGetImageMemoryRequirements2( |
| m_device, reinterpret_cast<const VkImageMemoryRequirementsInfo2 *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getBufferMemoryRequirements2(const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetBufferMemoryRequirements2( |
| m_device, reinterpret_cast<const VkBufferMemoryRequirementsInfo2 *>(pInfo), reinterpret_cast<VkMemoryRequirements2 *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 |
| Device::getBufferMemoryRequirements2(const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements; |
| d.vkGetBufferMemoryRequirements2( |
| m_device, reinterpret_cast<const VkBufferMemoryRequirementsInfo2 *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getBufferMemoryRequirements2(const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>(); |
| d.vkGetBufferMemoryRequirements2( |
| m_device, reinterpret_cast<const VkBufferMemoryRequirementsInfo2 *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageSparseMemoryRequirements2(const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 *pInfo, |
| uint32_t *pSparseMemoryRequirementCount, |
| VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2 *pSparseMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetImageSparseMemoryRequirements2(m_device, |
| reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(pInfo), |
| pSparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(pSparseMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SparseImageMemoryRequirements2Allocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> |
| Device::getImageSparseMemoryRequirements2(const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 &info, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements; |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetImageSparseMemoryRequirements2( |
| m_device, reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(&info), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetImageSparseMemoryRequirements2(m_device, |
| reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(&info), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| |
| template<typename SparseImageMemoryRequirements2Allocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SparseImageMemoryRequirements2>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> |
| Device::getImageSparseMemoryRequirements2(const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 &info, |
| SparseImageMemoryRequirements2Allocator &sparseImageMemoryRequirements2Allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements(sparseImageMemoryRequirements2Allocator); |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetImageSparseMemoryRequirements2( |
| m_device, reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(&info), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetImageSparseMemoryRequirements2(m_device, |
| reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(&info), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getFeatures2(VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 *pFeatures, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceFeatures2(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceFeatures2 *>(pFeatures)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 |
| PhysicalDevice::getFeatures2(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 features; |
| d.vkGetPhysicalDeviceFeatures2(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceFeatures2 *>(&features)); |
| |
| return features; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> PhysicalDevice::getFeatures2(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 &features = structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2>(); |
| d.vkGetPhysicalDeviceFeatures2(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceFeatures2 *>(&features)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getProperties2(VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceProperties2(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceProperties2 *>(pProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 |
| PhysicalDevice::getProperties2(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 properties; |
| d.vkGetPhysicalDeviceProperties2(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceProperties2 *>(&properties)); |
| |
| return properties; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> PhysicalDevice::getProperties2(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 &properties = structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2>(); |
| d.vkGetPhysicalDeviceProperties2(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceProperties2 *>(&properties)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getFormatProperties2(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::FormatProperties2 *pFormatProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceFormatProperties2(m_physicalDevice, static_cast<VkFormat>(format), reinterpret_cast<VkFormatProperties2 *>(pFormatProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::FormatProperties2 |
| PhysicalDevice::getFormatProperties2(VULKAN_HPP_NAMESPACE::Format format, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::FormatProperties2 formatProperties; |
| d.vkGetPhysicalDeviceFormatProperties2(m_physicalDevice, static_cast<VkFormat>(format), reinterpret_cast<VkFormatProperties2 *>(&formatProperties)); |
| |
| return formatProperties; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> PhysicalDevice::getFormatProperties2(VULKAN_HPP_NAMESPACE::Format format, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::FormatProperties2 &formatProperties = structureChain.template get<VULKAN_HPP_NAMESPACE::FormatProperties2>(); |
| d.vkGetPhysicalDeviceFormatProperties2(m_physicalDevice, static_cast<VkFormat>(format), reinterpret_cast<VkFormatProperties2 *>(&formatProperties)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getImageFormatProperties2(const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 *pImageFormatInfo, |
| VULKAN_HPP_NAMESPACE::ImageFormatProperties2 *pImageFormatProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceImageFormatProperties2(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceImageFormatInfo2 *>(pImageFormatInfo), |
| reinterpret_cast<VkImageFormatProperties2 *>(pImageFormatProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::ImageFormatProperties2>::type |
| PhysicalDevice::getImageFormatProperties2(const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 &imageFormatInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ImageFormatProperties2 imageFormatProperties; |
| VkResult result = d.vkGetPhysicalDeviceImageFormatProperties2(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceImageFormatInfo2 *>(&imageFormatInfo), |
| reinterpret_cast<VkImageFormatProperties2 *>(&imageFormatProperties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getImageFormatProperties2"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), imageFormatProperties); |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<StructureChain<X, Y, Z...>>::type |
| PhysicalDevice::getImageFormatProperties2(const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 &imageFormatInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::ImageFormatProperties2 &imageFormatProperties = structureChain.template get<VULKAN_HPP_NAMESPACE::ImageFormatProperties2>(); |
| Result result = |
| static_cast<Result>(d.vkGetPhysicalDeviceImageFormatProperties2(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceImageFormatInfo2 *>(&imageFormatInfo), |
| reinterpret_cast<VkImageFormatProperties2 *>(&imageFormatProperties))); |
| return createResultValue(result, structureChain, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getImageFormatProperties2"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getQueueFamilyProperties2(uint32_t *pQueueFamilyPropertyCount, |
| VULKAN_HPP_NAMESPACE::QueueFamilyProperties2 *pQueueFamilyProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceQueueFamilyProperties2( |
| m_physicalDevice, pQueueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(pQueueFamilyProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename QueueFamilyProperties2Allocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<QueueFamilyProperties2, QueueFamilyProperties2Allocator> |
| PhysicalDevice::getQueueFamilyProperties2(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<QueueFamilyProperties2, QueueFamilyProperties2Allocator> queueFamilyProperties; |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties2(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| queueFamilyProperties.resize(queueFamilyPropertyCount); |
| d.vkGetPhysicalDeviceQueueFamilyProperties2( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| return queueFamilyProperties; |
| } |
| |
| template<typename QueueFamilyProperties2Allocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, QueueFamilyProperties2>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<QueueFamilyProperties2, QueueFamilyProperties2Allocator> |
| PhysicalDevice::getQueueFamilyProperties2(QueueFamilyProperties2Allocator &queueFamilyProperties2Allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<QueueFamilyProperties2, QueueFamilyProperties2Allocator> queueFamilyProperties(queueFamilyProperties2Allocator); |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties2(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| queueFamilyProperties.resize(queueFamilyPropertyCount); |
| d.vkGetPhysicalDeviceQueueFamilyProperties2( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| return queueFamilyProperties; |
| } |
| |
| template<typename StructureChain, typename StructureChainAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<StructureChain, StructureChainAllocator> |
| PhysicalDevice::getQueueFamilyProperties2(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties2(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| std::vector<StructureChain, StructureChainAllocator> returnVector(queueFamilyPropertyCount); |
| std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2> queueFamilyProperties(queueFamilyPropertyCount); |
| for(uint32_t i = 0; i < queueFamilyPropertyCount; i++) |
| { |
| queueFamilyProperties[i].pNext = returnVector[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>().pNext; |
| } |
| d.vkGetPhysicalDeviceQueueFamilyProperties2( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| for(uint32_t i = 0; i < queueFamilyPropertyCount; i++) |
| { |
| returnVector[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>() = queueFamilyProperties[i]; |
| } |
| return returnVector; |
| } |
| |
| template<typename StructureChain, |
| typename StructureChainAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, StructureChain>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<StructureChain, StructureChainAllocator> |
| PhysicalDevice::getQueueFamilyProperties2(StructureChainAllocator &structureChainAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties2(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| std::vector<StructureChain, StructureChainAllocator> returnVector(queueFamilyPropertyCount, structureChainAllocator); |
| std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2> queueFamilyProperties(queueFamilyPropertyCount); |
| for(uint32_t i = 0; i < queueFamilyPropertyCount; i++) |
| { |
| queueFamilyProperties[i].pNext = returnVector[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>().pNext; |
| } |
| d.vkGetPhysicalDeviceQueueFamilyProperties2( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| for(uint32_t i = 0; i < queueFamilyPropertyCount; i++) |
| { |
| returnVector[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>() = queueFamilyProperties[i]; |
| } |
| return returnVector; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getMemoryProperties2(VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 *pMemoryProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceMemoryProperties2(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceMemoryProperties2 *>(pMemoryProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 |
| PhysicalDevice::getMemoryProperties2(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 memoryProperties; |
| d.vkGetPhysicalDeviceMemoryProperties2(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceMemoryProperties2 *>(&memoryProperties)); |
| |
| return memoryProperties; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> PhysicalDevice::getMemoryProperties2(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 &memoryProperties = |
| structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2>(); |
| d.vkGetPhysicalDeviceMemoryProperties2(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceMemoryProperties2 *>(&memoryProperties)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getSparseImageFormatProperties2(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 *pFormatInfo, |
| uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2 *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(pFormatInfo), |
| pPropertyCount, |
| reinterpret_cast<VkSparseImageFormatProperties2 *>(pProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SparseImageFormatProperties2Allocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> |
| PhysicalDevice::getSparseImageFormatProperties2(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 &formatInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> properties; |
| uint32_t propertyCount; |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(&formatInfo), &propertyCount, nullptr); |
| properties.resize(propertyCount); |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(&formatInfo), |
| &propertyCount, |
| reinterpret_cast<VkSparseImageFormatProperties2 *>(properties.data())); |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| return properties; |
| } |
| |
| template<typename SparseImageFormatProperties2Allocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SparseImageFormatProperties2>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> |
| PhysicalDevice::getSparseImageFormatProperties2(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 &formatInfo, |
| SparseImageFormatProperties2Allocator &sparseImageFormatProperties2Allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> properties(sparseImageFormatProperties2Allocator); |
| uint32_t propertyCount; |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(&formatInfo), &propertyCount, nullptr); |
| properties.resize(propertyCount); |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(&formatInfo), |
| &propertyCount, |
| reinterpret_cast<VkSparseImageFormatProperties2 *>(properties.data())); |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| return properties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::trimCommandPool(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| VULKAN_HPP_NAMESPACE::CommandPoolTrimFlags flags, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkTrimCommandPool(m_device, static_cast<VkCommandPool>(commandPool), static_cast<VkCommandPoolTrimFlags>(flags)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getQueue2(const VULKAN_HPP_NAMESPACE::DeviceQueueInfo2 *pQueueInfo, |
| VULKAN_HPP_NAMESPACE::Queue *pQueue, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceQueue2(m_device, reinterpret_cast<const VkDeviceQueueInfo2 *>(pQueueInfo), reinterpret_cast<VkQueue *>(pQueue)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Queue Device::getQueue2(const VULKAN_HPP_NAMESPACE::DeviceQueueInfo2 &queueInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Queue queue; |
| d.vkGetDeviceQueue2(m_device, reinterpret_cast<const VkDeviceQueueInfo2 *>(&queueInfo), reinterpret_cast<VkQueue *>(&queue)); |
| |
| return queue; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createSamplerYcbcrConversion(const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion *pYcbcrConversion, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateSamplerYcbcrConversion(m_device, |
| reinterpret_cast<const VkSamplerYcbcrConversionCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSamplerYcbcrConversion *>(pYcbcrConversion))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion>::type |
| Device::createSamplerYcbcrConversion(const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion; |
| VkResult result = d.vkCreateSamplerYcbcrConversion( |
| m_device, |
| reinterpret_cast<const VkSamplerYcbcrConversionCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSamplerYcbcrConversion *>(&ycbcrConversion)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createSamplerYcbcrConversion"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), ycbcrConversion); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion, Dispatch>>::type |
| Device::createSamplerYcbcrConversionUnique(const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion; |
| Result result = static_cast<Result>(d.vkCreateSamplerYcbcrConversion( |
| m_device, |
| reinterpret_cast<const VkSamplerYcbcrConversionCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSamplerYcbcrConversion *>(&ycbcrConversion))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion, Dispatch>( |
| result, ycbcrConversion, VULKAN_HPP_NAMESPACE_STRING "::Device::createSamplerYcbcrConversionUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySamplerYcbcrConversion(VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySamplerYcbcrConversion( |
| m_device, static_cast<VkSamplerYcbcrConversion>(ycbcrConversion), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySamplerYcbcrConversion(VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySamplerYcbcrConversion( |
| m_device, |
| static_cast<VkSamplerYcbcrConversion>(ycbcrConversion), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySamplerYcbcrConversion( |
| m_device, static_cast<VkSamplerYcbcrConversion>(ycbcrConversion), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySamplerYcbcrConversion( |
| m_device, |
| static_cast<VkSamplerYcbcrConversion>(ycbcrConversion), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createDescriptorUpdateTemplate(const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate *pDescriptorUpdateTemplate, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDescriptorUpdateTemplate(m_device, |
| reinterpret_cast<const VkDescriptorUpdateTemplateCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkDescriptorUpdateTemplate *>(pDescriptorUpdateTemplate))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate>::type |
| Device::createDescriptorUpdateTemplate(const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate; |
| VkResult result = d.vkCreateDescriptorUpdateTemplate( |
| m_device, |
| reinterpret_cast<const VkDescriptorUpdateTemplateCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDescriptorUpdateTemplate *>(&descriptorUpdateTemplate)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createDescriptorUpdateTemplate"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), descriptorUpdateTemplate); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate, Dispatch>>::type |
| Device::createDescriptorUpdateTemplateUnique(const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate; |
| Result result = static_cast<Result>(d.vkCreateDescriptorUpdateTemplate( |
| m_device, |
| reinterpret_cast<const VkDescriptorUpdateTemplateCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDescriptorUpdateTemplate *>(&descriptorUpdateTemplate))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate, Dispatch>( |
| result, descriptorUpdateTemplate, VULKAN_HPP_NAMESPACE_STRING "::Device::createDescriptorUpdateTemplateUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDescriptorUpdateTemplate(VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDescriptorUpdateTemplate( |
| m_device, static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDescriptorUpdateTemplate(VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDescriptorUpdateTemplate( |
| m_device, |
| static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDescriptorUpdateTemplate( |
| m_device, static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDescriptorUpdateTemplate( |
| m_device, |
| static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::updateDescriptorSetWithTemplate(VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet, |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| const void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkUpdateDescriptorSetWithTemplate( |
| m_device, static_cast<VkDescriptorSet>(descriptorSet), static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), pData); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_INLINE void Device::updateDescriptorSetWithTemplate(VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet, |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| DataType const &data, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkUpdateDescriptorSetWithTemplate(m_device, |
| static_cast<VkDescriptorSet>(descriptorSet), |
| static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), |
| reinterpret_cast<const void *>(&data)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getExternalBufferProperties(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalBufferInfo *pExternalBufferInfo, |
| VULKAN_HPP_NAMESPACE::ExternalBufferProperties *pExternalBufferProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceExternalBufferProperties(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalBufferInfo *>(pExternalBufferInfo), |
| reinterpret_cast<VkExternalBufferProperties *>(pExternalBufferProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::ExternalBufferProperties |
| PhysicalDevice::getExternalBufferProperties(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalBufferInfo &externalBufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ExternalBufferProperties externalBufferProperties; |
| d.vkGetPhysicalDeviceExternalBufferProperties(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalBufferInfo *>(&externalBufferInfo), |
| reinterpret_cast<VkExternalBufferProperties *>(&externalBufferProperties)); |
| |
| return externalBufferProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getExternalFenceProperties(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalFenceInfo *pExternalFenceInfo, |
| VULKAN_HPP_NAMESPACE::ExternalFenceProperties *pExternalFenceProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceExternalFenceProperties(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalFenceInfo *>(pExternalFenceInfo), |
| reinterpret_cast<VkExternalFenceProperties *>(pExternalFenceProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::ExternalFenceProperties |
| PhysicalDevice::getExternalFenceProperties(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalFenceInfo &externalFenceInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ExternalFenceProperties externalFenceProperties; |
| d.vkGetPhysicalDeviceExternalFenceProperties(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalFenceInfo *>(&externalFenceInfo), |
| reinterpret_cast<VkExternalFenceProperties *>(&externalFenceProperties)); |
| |
| return externalFenceProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| PhysicalDevice::getExternalSemaphoreProperties(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalSemaphoreInfo *pExternalSemaphoreInfo, |
| VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties *pExternalSemaphoreProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceExternalSemaphoreProperties(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalSemaphoreInfo *>(pExternalSemaphoreInfo), |
| reinterpret_cast<VkExternalSemaphoreProperties *>(pExternalSemaphoreProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties |
| PhysicalDevice::getExternalSemaphoreProperties(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalSemaphoreInfo &externalSemaphoreInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties externalSemaphoreProperties; |
| d.vkGetPhysicalDeviceExternalSemaphoreProperties(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalSemaphoreInfo *>(&externalSemaphoreInfo), |
| reinterpret_cast<VkExternalSemaphoreProperties *>(&externalSemaphoreProperties)); |
| |
| return externalSemaphoreProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getDescriptorSetLayoutSupport(const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo *pCreateInfo, |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport *pSupport, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDescriptorSetLayoutSupport( |
| m_device, reinterpret_cast<const VkDescriptorSetLayoutCreateInfo *>(pCreateInfo), reinterpret_cast<VkDescriptorSetLayoutSupport *>(pSupport)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport |
| Device::getDescriptorSetLayoutSupport(const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo &createInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport support; |
| d.vkGetDescriptorSetLayoutSupport( |
| m_device, reinterpret_cast<const VkDescriptorSetLayoutCreateInfo *>(&createInfo), reinterpret_cast<VkDescriptorSetLayoutSupport *>(&support)); |
| |
| return support; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getDescriptorSetLayoutSupport(const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo &createInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport &support = structureChain.template get<VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport>(); |
| d.vkGetDescriptorSetLayoutSupport( |
| m_device, reinterpret_cast<const VkDescriptorSetLayoutCreateInfo *>(&createInfo), reinterpret_cast<VkDescriptorSetLayoutSupport *>(&support)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_VERSION_1_2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndirectCount(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::Buffer countBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize countBufferOffset, |
| uint32_t maxDrawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndirectCount(m_commandBuffer, |
| static_cast<VkBuffer>(buffer), |
| static_cast<VkDeviceSize>(offset), |
| static_cast<VkBuffer>(countBuffer), |
| static_cast<VkDeviceSize>(countBufferOffset), |
| maxDrawCount, |
| stride); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndexedIndirectCount(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::Buffer countBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize countBufferOffset, |
| uint32_t maxDrawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndexedIndirectCount(m_commandBuffer, |
| static_cast<VkBuffer>(buffer), |
| static_cast<VkDeviceSize>(offset), |
| static_cast<VkBuffer>(countBuffer), |
| static_cast<VkDeviceSize>(countBufferOffset), |
| maxDrawCount, |
| stride); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createRenderPass2(const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::RenderPass *pRenderPass, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateRenderPass2(m_device, |
| reinterpret_cast<const VkRenderPassCreateInfo2 *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkRenderPass *>(pRenderPass))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::RenderPass>::type |
| Device::createRenderPass2(const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::RenderPass renderPass; |
| VkResult result = |
| d.vkCreateRenderPass2(m_device, |
| reinterpret_cast<const VkRenderPassCreateInfo2 *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkRenderPass *>(&renderPass)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createRenderPass2"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), renderPass); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::RenderPass, Dispatch>>::type |
| Device::createRenderPass2Unique(const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::RenderPass renderPass; |
| Result result = static_cast<Result>( |
| d.vkCreateRenderPass2(m_device, |
| reinterpret_cast<const VkRenderPassCreateInfo2 *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkRenderPass *>(&renderPass))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::RenderPass, Dispatch>( |
| result, renderPass, VULKAN_HPP_NAMESPACE_STRING "::Device::createRenderPass2Unique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass2(const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo *pRenderPassBegin, |
| const VULKAN_HPP_NAMESPACE::SubpassBeginInfo *pSubpassBeginInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginRenderPass2( |
| m_commandBuffer, reinterpret_cast<const VkRenderPassBeginInfo *>(pRenderPassBegin), reinterpret_cast<const VkSubpassBeginInfo *>(pSubpassBeginInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass2(const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo &renderPassBegin, |
| const VULKAN_HPP_NAMESPACE::SubpassBeginInfo &subpassBeginInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBeginRenderPass2( |
| m_commandBuffer, reinterpret_cast<const VkRenderPassBeginInfo *>(&renderPassBegin), reinterpret_cast<const VkSubpassBeginInfo *>(&subpassBeginInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::nextSubpass2(const VULKAN_HPP_NAMESPACE::SubpassBeginInfo *pSubpassBeginInfo, |
| const VULKAN_HPP_NAMESPACE::SubpassEndInfo *pSubpassEndInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdNextSubpass2( |
| m_commandBuffer, reinterpret_cast<const VkSubpassBeginInfo *>(pSubpassBeginInfo), reinterpret_cast<const VkSubpassEndInfo *>(pSubpassEndInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::nextSubpass2(const VULKAN_HPP_NAMESPACE::SubpassBeginInfo &subpassBeginInfo, |
| const VULKAN_HPP_NAMESPACE::SubpassEndInfo &subpassEndInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdNextSubpass2( |
| m_commandBuffer, reinterpret_cast<const VkSubpassBeginInfo *>(&subpassBeginInfo), reinterpret_cast<const VkSubpassEndInfo *>(&subpassEndInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endRenderPass2(const VULKAN_HPP_NAMESPACE::SubpassEndInfo *pSubpassEndInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndRenderPass2(m_commandBuffer, reinterpret_cast<const VkSubpassEndInfo *>(pSubpassEndInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endRenderPass2(const VULKAN_HPP_NAMESPACE::SubpassEndInfo &subpassEndInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdEndRenderPass2(m_commandBuffer, reinterpret_cast<const VkSubpassEndInfo *>(&subpassEndInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| Device::resetQueryPool(VULKAN_HPP_NAMESPACE::QueryPool queryPool, uint32_t firstQuery, uint32_t queryCount, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkResetQueryPool(m_device, static_cast<VkQueryPool>(queryPool), firstQuery, queryCount); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getSemaphoreCounterValue(VULKAN_HPP_NAMESPACE::Semaphore semaphore, |
| uint64_t *pValue, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetSemaphoreCounterValue(m_device, static_cast<VkSemaphore>(semaphore), pValue)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<uint64_t>::type Device::getSemaphoreCounterValue(VULKAN_HPP_NAMESPACE::Semaphore semaphore, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint64_t value; |
| VkResult result = d.vkGetSemaphoreCounterValue(m_device, static_cast<VkSemaphore>(semaphore), &value); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getSemaphoreCounterValue"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), value); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::waitSemaphores(const VULKAN_HPP_NAMESPACE::SemaphoreWaitInfo *pWaitInfo, |
| uint64_t timeout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkWaitSemaphores(m_device, reinterpret_cast<const VkSemaphoreWaitInfo *>(pWaitInfo), timeout)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result |
| Device::waitSemaphores(const VULKAN_HPP_NAMESPACE::SemaphoreWaitInfo &waitInfo, uint64_t timeout, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkWaitSemaphores(m_device, reinterpret_cast<const VkSemaphoreWaitInfo *>(&waitInfo), timeout); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::waitSemaphores", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eTimeout }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::signalSemaphore(const VULKAN_HPP_NAMESPACE::SemaphoreSignalInfo *pSignalInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkSignalSemaphore(m_device, reinterpret_cast<const VkSemaphoreSignalInfo *>(pSignalInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::signalSemaphore(const VULKAN_HPP_NAMESPACE::SemaphoreSignalInfo &signalInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkSignalSemaphore(m_device, reinterpret_cast<const VkSemaphoreSignalInfo *>(&signalInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::signalSemaphore"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE DeviceAddress Device::getBufferAddress(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<DeviceAddress>(d.vkGetBufferDeviceAddress(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(pInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::DeviceAddress Device::getBufferAddress(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkDeviceAddress result = d.vkGetBufferDeviceAddress(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(&info)); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::DeviceAddress>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint64_t Device::getBufferOpaqueCaptureAddress(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return d.vkGetBufferOpaqueCaptureAddress(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(pInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint64_t Device::getBufferOpaqueCaptureAddress(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint64_t result = d.vkGetBufferOpaqueCaptureAddress(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(&info)); |
| |
| return result; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint64_t Device::getMemoryOpaqueCaptureAddress(const VULKAN_HPP_NAMESPACE::DeviceMemoryOpaqueCaptureAddressInfo *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return d.vkGetDeviceMemoryOpaqueCaptureAddress(m_device, reinterpret_cast<const VkDeviceMemoryOpaqueCaptureAddressInfo *>(pInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint64_t Device::getMemoryOpaqueCaptureAddress(const VULKAN_HPP_NAMESPACE::DeviceMemoryOpaqueCaptureAddressInfo &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint64_t result = d.vkGetDeviceMemoryOpaqueCaptureAddress(m_device, reinterpret_cast<const VkDeviceMemoryOpaqueCaptureAddressInfo *>(&info)); |
| |
| return result; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_VERSION_1_3 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getToolProperties(uint32_t *pToolCount, |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties *pToolProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPhysicalDeviceToolProperties(m_physicalDevice, pToolCount, reinterpret_cast<VkPhysicalDeviceToolProperties *>(pToolProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PhysicalDeviceToolPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator>>::type |
| PhysicalDevice::getToolProperties(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator> toolProperties; |
| uint32_t toolCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceToolProperties(m_physicalDevice, &toolCount, nullptr)); |
| if((result == Result::eSuccess) && toolCount) |
| { |
| toolProperties.resize(toolCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceToolProperties(m_physicalDevice, &toolCount, reinterpret_cast<VkPhysicalDeviceToolProperties *>(toolProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(toolCount <= toolProperties.size()); |
| if(toolCount < toolProperties.size()) |
| { |
| toolProperties.resize(toolCount); |
| } |
| } |
| return createResultValue(result, toolProperties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getToolProperties"); |
| } |
| |
| template<typename PhysicalDeviceToolPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PhysicalDeviceToolProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator>>::type |
| PhysicalDevice::getToolProperties(PhysicalDeviceToolPropertiesAllocator &physicalDeviceToolPropertiesAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator> toolProperties(physicalDeviceToolPropertiesAllocator); |
| uint32_t toolCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceToolProperties(m_physicalDevice, &toolCount, nullptr)); |
| if((result == Result::eSuccess) && toolCount) |
| { |
| toolProperties.resize(toolCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceToolProperties(m_physicalDevice, &toolCount, reinterpret_cast<VkPhysicalDeviceToolProperties *>(toolProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(toolCount <= toolProperties.size()); |
| if(toolCount < toolProperties.size()) |
| { |
| toolProperties.resize(toolCount); |
| } |
| } |
| return createResultValue(result, toolProperties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getToolProperties"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createPrivateDataSlot(const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot *pPrivateDataSlot, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreatePrivateDataSlot(m_device, |
| reinterpret_cast<const VkPrivateDataSlotCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkPrivateDataSlot *>(pPrivateDataSlot))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::PrivateDataSlot>::type |
| Device::createPrivateDataSlot(const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot; |
| VkResult result = |
| d.vkCreatePrivateDataSlot(m_device, |
| reinterpret_cast<const VkPrivateDataSlotCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPrivateDataSlot *>(&privateDataSlot)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createPrivateDataSlot"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), privateDataSlot); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::PrivateDataSlot, Dispatch>>::type |
| Device::createPrivateDataSlotUnique(const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot; |
| Result result = static_cast<Result>( |
| d.vkCreatePrivateDataSlot(m_device, |
| reinterpret_cast<const VkPrivateDataSlotCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPrivateDataSlot *>(&privateDataSlot))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::PrivateDataSlot, Dispatch>( |
| result, privateDataSlot, VULKAN_HPP_NAMESPACE_STRING "::Device::createPrivateDataSlotUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPrivateDataSlot(VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyPrivateDataSlot(m_device, static_cast<VkPrivateDataSlot>(privateDataSlot), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPrivateDataSlot(VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyPrivateDataSlot( |
| m_device, |
| static_cast<VkPrivateDataSlot>(privateDataSlot), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyPrivateDataSlot(m_device, static_cast<VkPrivateDataSlot>(privateDataSlot), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyPrivateDataSlot( |
| m_device, |
| static_cast<VkPrivateDataSlot>(privateDataSlot), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::setPrivateData(VULKAN_HPP_NAMESPACE::ObjectType objectType, |
| uint64_t objectHandle, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| uint64_t data, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkSetPrivateData(m_device, static_cast<VkObjectType>(objectType), objectHandle, static_cast<VkPrivateDataSlot>(privateDataSlot), data)); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::setPrivateData(VULKAN_HPP_NAMESPACE::ObjectType objectType, |
| uint64_t objectHandle, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| uint64_t data, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = |
| d.vkSetPrivateData(m_device, static_cast<VkObjectType>(objectType), objectHandle, static_cast<VkPrivateDataSlot>(privateDataSlot), data); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::setPrivateData"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getPrivateData(VULKAN_HPP_NAMESPACE::ObjectType objectType, |
| uint64_t objectHandle, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| uint64_t *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPrivateData(m_device, static_cast<VkObjectType>(objectType), objectHandle, static_cast<VkPrivateDataSlot>(privateDataSlot), pData); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE uint64_t Device::getPrivateData(VULKAN_HPP_NAMESPACE::ObjectType objectType, |
| uint64_t objectHandle, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint64_t data; |
| d.vkGetPrivateData(m_device, static_cast<VkObjectType>(objectType), objectHandle, static_cast<VkPrivateDataSlot>(privateDataSlot), &data); |
| |
| return data; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setEvent2(VULKAN_HPP_NAMESPACE::Event event, |
| const VULKAN_HPP_NAMESPACE::DependencyInfo *pDependencyInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetEvent2(m_commandBuffer, static_cast<VkEvent>(event), reinterpret_cast<const VkDependencyInfo *>(pDependencyInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setEvent2(VULKAN_HPP_NAMESPACE::Event event, |
| const VULKAN_HPP_NAMESPACE::DependencyInfo &dependencyInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetEvent2(m_commandBuffer, static_cast<VkEvent>(event), reinterpret_cast<const VkDependencyInfo *>(&dependencyInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resetEvent2(VULKAN_HPP_NAMESPACE::Event event, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags2 stageMask, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdResetEvent2(m_commandBuffer, static_cast<VkEvent>(event), static_cast<VkPipelineStageFlags2>(stageMask)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::waitEvents2(uint32_t eventCount, |
| const VULKAN_HPP_NAMESPACE::Event *pEvents, |
| const VULKAN_HPP_NAMESPACE::DependencyInfo *pDependencyInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWaitEvents2( |
| m_commandBuffer, eventCount, reinterpret_cast<const VkEvent *>(pEvents), reinterpret_cast<const VkDependencyInfo *>(pDependencyInfos)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::waitEvents2(ArrayProxy<const VULKAN_HPP_NAMESPACE::Event> const &events, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DependencyInfo> const &dependencyInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(events.size() == dependencyInfos.size()); |
| # else |
| if(events.size() != dependencyInfos.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::waitEvents2: events.size() != dependencyInfos.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdWaitEvents2(m_commandBuffer, |
| events.size(), |
| reinterpret_cast<const VkEvent *>(events.data()), |
| reinterpret_cast<const VkDependencyInfo *>(dependencyInfos.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier2(const VULKAN_HPP_NAMESPACE::DependencyInfo *pDependencyInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdPipelineBarrier2(m_commandBuffer, reinterpret_cast<const VkDependencyInfo *>(pDependencyInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier2(const VULKAN_HPP_NAMESPACE::DependencyInfo &dependencyInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdPipelineBarrier2(m_commandBuffer, reinterpret_cast<const VkDependencyInfo *>(&dependencyInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::writeTimestamp2(VULKAN_HPP_NAMESPACE::PipelineStageFlags2 stage, |
| VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t query, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWriteTimestamp2(m_commandBuffer, static_cast<VkPipelineStageFlags2>(stage), static_cast<VkQueryPool>(queryPool), query); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Queue::submit2(uint32_t submitCount, |
| const VULKAN_HPP_NAMESPACE::SubmitInfo2 *pSubmits, |
| VULKAN_HPP_NAMESPACE::Fence fence, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkQueueSubmit2(m_queue, submitCount, reinterpret_cast<const VkSubmitInfo2 *>(pSubmits), static_cast<VkFence>(fence))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Queue::submit2(ArrayProxy<const VULKAN_HPP_NAMESPACE::SubmitInfo2> const &submits, VULKAN_HPP_NAMESPACE::Fence fence, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkQueueSubmit2(m_queue, submits.size(), reinterpret_cast<const VkSubmitInfo2 *>(submits.data()), static_cast<VkFence>(fence)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Queue::submit2"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBuffer2(const VULKAN_HPP_NAMESPACE::CopyBufferInfo2 *pCopyBufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyBuffer2(m_commandBuffer, reinterpret_cast<const VkCopyBufferInfo2 *>(pCopyBufferInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBuffer2(const VULKAN_HPP_NAMESPACE::CopyBufferInfo2 ©BufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyBuffer2(m_commandBuffer, reinterpret_cast<const VkCopyBufferInfo2 *>(©BufferInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImage2(const VULKAN_HPP_NAMESPACE::CopyImageInfo2 *pCopyImageInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyImage2(m_commandBuffer, reinterpret_cast<const VkCopyImageInfo2 *>(pCopyImageInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImage2(const VULKAN_HPP_NAMESPACE::CopyImageInfo2 ©ImageInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyImage2(m_commandBuffer, reinterpret_cast<const VkCopyImageInfo2 *>(©ImageInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage2(const VULKAN_HPP_NAMESPACE::CopyBufferToImageInfo2 *pCopyBufferToImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyBufferToImage2(m_commandBuffer, reinterpret_cast<const VkCopyBufferToImageInfo2 *>(pCopyBufferToImageInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage2(const VULKAN_HPP_NAMESPACE::CopyBufferToImageInfo2 ©BufferToImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyBufferToImage2(m_commandBuffer, reinterpret_cast<const VkCopyBufferToImageInfo2 *>(©BufferToImageInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer2(const VULKAN_HPP_NAMESPACE::CopyImageToBufferInfo2 *pCopyImageToBufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyImageToBuffer2(m_commandBuffer, reinterpret_cast<const VkCopyImageToBufferInfo2 *>(pCopyImageToBufferInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer2(const VULKAN_HPP_NAMESPACE::CopyImageToBufferInfo2 ©ImageToBufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyImageToBuffer2(m_commandBuffer, reinterpret_cast<const VkCopyImageToBufferInfo2 *>(©ImageToBufferInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::blitImage2(const VULKAN_HPP_NAMESPACE::BlitImageInfo2 *pBlitImageInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBlitImage2(m_commandBuffer, reinterpret_cast<const VkBlitImageInfo2 *>(pBlitImageInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::blitImage2(const VULKAN_HPP_NAMESPACE::BlitImageInfo2 &blitImageInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBlitImage2(m_commandBuffer, reinterpret_cast<const VkBlitImageInfo2 *>(&blitImageInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resolveImage2(const VULKAN_HPP_NAMESPACE::ResolveImageInfo2 *pResolveImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdResolveImage2(m_commandBuffer, reinterpret_cast<const VkResolveImageInfo2 *>(pResolveImageInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resolveImage2(const VULKAN_HPP_NAMESPACE::ResolveImageInfo2 &resolveImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdResolveImage2(m_commandBuffer, reinterpret_cast<const VkResolveImageInfo2 *>(&resolveImageInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRendering(const VULKAN_HPP_NAMESPACE::RenderingInfo *pRenderingInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginRendering(m_commandBuffer, reinterpret_cast<const VkRenderingInfo *>(pRenderingInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRendering(const VULKAN_HPP_NAMESPACE::RenderingInfo &renderingInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBeginRendering(m_commandBuffer, reinterpret_cast<const VkRenderingInfo *>(&renderingInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endRendering(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndRendering(m_commandBuffer); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setCullMode(VULKAN_HPP_NAMESPACE::CullModeFlags cullMode, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetCullMode(m_commandBuffer, static_cast<VkCullModeFlags>(cullMode)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setFrontFace(VULKAN_HPP_NAMESPACE::FrontFace frontFace, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetFrontFace(m_commandBuffer, static_cast<VkFrontFace>(frontFace)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setPrimitiveTopology(VULKAN_HPP_NAMESPACE::PrimitiveTopology primitiveTopology, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetPrimitiveTopology(m_commandBuffer, static_cast<VkPrimitiveTopology>(primitiveTopology)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setViewportWithCount(uint32_t viewportCount, |
| const VULKAN_HPP_NAMESPACE::Viewport *pViewports, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetViewportWithCount(m_commandBuffer, viewportCount, reinterpret_cast<const VkViewport *>(pViewports)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setViewportWithCount(ArrayProxy<const VULKAN_HPP_NAMESPACE::Viewport> const &viewports, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetViewportWithCount(m_commandBuffer, viewports.size(), reinterpret_cast<const VkViewport *>(viewports.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::setScissorWithCount(uint32_t scissorCount, const VULKAN_HPP_NAMESPACE::Rect2D *pScissors, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetScissorWithCount(m_commandBuffer, scissorCount, reinterpret_cast<const VkRect2D *>(pScissors)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setScissorWithCount(ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D> const &scissors, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetScissorWithCount(m_commandBuffer, scissors.size(), reinterpret_cast<const VkRect2D *>(scissors.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindVertexBuffers2(uint32_t firstBinding, |
| uint32_t bindingCount, |
| const VULKAN_HPP_NAMESPACE::Buffer *pBuffers, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pOffsets, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pSizes, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pStrides, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindVertexBuffers2(m_commandBuffer, |
| firstBinding, |
| bindingCount, |
| reinterpret_cast<const VkBuffer *>(pBuffers), |
| reinterpret_cast<const VkDeviceSize *>(pOffsets), |
| reinterpret_cast<const VkDeviceSize *>(pSizes), |
| reinterpret_cast<const VkDeviceSize *>(pStrides)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindVertexBuffers2(uint32_t firstBinding, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer> const &buffers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &offsets, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &sizes, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &strides, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(buffers.size() == offsets.size()); |
| VULKAN_HPP_ASSERT(sizes.empty() || buffers.size() == sizes.size()); |
| VULKAN_HPP_ASSERT(strides.empty() || buffers.size() == strides.size()); |
| # else |
| if(buffers.size() != offsets.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2: buffers.size() != offsets.size()"); |
| } |
| if(!sizes.empty() && buffers.size() != sizes.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2: buffers.size() != sizes.size()"); |
| } |
| if(!strides.empty() && buffers.size() != strides.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2: buffers.size() != strides.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdBindVertexBuffers2(m_commandBuffer, |
| firstBinding, |
| buffers.size(), |
| reinterpret_cast<const VkBuffer *>(buffers.data()), |
| reinterpret_cast<const VkDeviceSize *>(offsets.data()), |
| reinterpret_cast<const VkDeviceSize *>(sizes.data()), |
| reinterpret_cast<const VkDeviceSize *>(strides.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthTestEnable(VULKAN_HPP_NAMESPACE::Bool32 depthTestEnable, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthTestEnable(m_commandBuffer, static_cast<VkBool32>(depthTestEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthWriteEnable(VULKAN_HPP_NAMESPACE::Bool32 depthWriteEnable, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthWriteEnable(m_commandBuffer, static_cast<VkBool32>(depthWriteEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthCompareOp(VULKAN_HPP_NAMESPACE::CompareOp depthCompareOp, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthCompareOp(m_commandBuffer, static_cast<VkCompareOp>(depthCompareOp)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthBoundsTestEnable(VULKAN_HPP_NAMESPACE::Bool32 depthBoundsTestEnable, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthBoundsTestEnable(m_commandBuffer, static_cast<VkBool32>(depthBoundsTestEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setStencilTestEnable(VULKAN_HPP_NAMESPACE::Bool32 stencilTestEnable, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetStencilTestEnable(m_commandBuffer, static_cast<VkBool32>(stencilTestEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setStencilOp(VULKAN_HPP_NAMESPACE::StencilFaceFlags faceMask, |
| VULKAN_HPP_NAMESPACE::StencilOp failOp, |
| VULKAN_HPP_NAMESPACE::StencilOp passOp, |
| VULKAN_HPP_NAMESPACE::StencilOp depthFailOp, |
| VULKAN_HPP_NAMESPACE::CompareOp compareOp, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetStencilOp(m_commandBuffer, |
| static_cast<VkStencilFaceFlags>(faceMask), |
| static_cast<VkStencilOp>(failOp), |
| static_cast<VkStencilOp>(passOp), |
| static_cast<VkStencilOp>(depthFailOp), |
| static_cast<VkCompareOp>(compareOp)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setRasterizerDiscardEnable(VULKAN_HPP_NAMESPACE::Bool32 rasterizerDiscardEnable, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetRasterizerDiscardEnable(m_commandBuffer, static_cast<VkBool32>(rasterizerDiscardEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthBiasEnable(VULKAN_HPP_NAMESPACE::Bool32 depthBiasEnable, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthBiasEnable(m_commandBuffer, static_cast<VkBool32>(depthBiasEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setPrimitiveRestartEnable(VULKAN_HPP_NAMESPACE::Bool32 primitiveRestartEnable, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetPrimitiveRestartEnable(m_commandBuffer, static_cast<VkBool32>(primitiveRestartEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getBufferMemoryRequirements(const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceBufferMemoryRequirements( |
| m_device, reinterpret_cast<const VkDeviceBufferMemoryRequirements *>(pInfo), reinterpret_cast<VkMemoryRequirements2 *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 |
| Device::getBufferMemoryRequirements(const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements; |
| d.vkGetDeviceBufferMemoryRequirements( |
| m_device, reinterpret_cast<const VkDeviceBufferMemoryRequirements *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getBufferMemoryRequirements(const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>(); |
| d.vkGetDeviceBufferMemoryRequirements( |
| m_device, reinterpret_cast<const VkDeviceBufferMemoryRequirements *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageMemoryRequirements(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceImageMemoryRequirements( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(pInfo), reinterpret_cast<VkMemoryRequirements2 *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 |
| Device::getImageMemoryRequirements(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements; |
| d.vkGetDeviceImageMemoryRequirements( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getImageMemoryRequirements(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>(); |
| d.vkGetDeviceImageMemoryRequirements( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageSparseMemoryRequirements(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements *pInfo, |
| uint32_t *pSparseMemoryRequirementCount, |
| VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2 *pSparseMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceImageSparseMemoryRequirements(m_device, |
| reinterpret_cast<const VkDeviceImageMemoryRequirements *>(pInfo), |
| pSparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(pSparseMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SparseImageMemoryRequirements2Allocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> |
| Device::getImageSparseMemoryRequirements(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements &info, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements; |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetDeviceImageSparseMemoryRequirements( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetDeviceImageSparseMemoryRequirements(m_device, |
| reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| |
| template<typename SparseImageMemoryRequirements2Allocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SparseImageMemoryRequirements2>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> |
| Device::getImageSparseMemoryRequirements(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements &info, |
| SparseImageMemoryRequirements2Allocator &sparseImageMemoryRequirements2Allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements(sparseImageMemoryRequirements2Allocator); |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetDeviceImageSparseMemoryRequirements( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetDeviceImageSparseMemoryRequirements(m_device, |
| reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroySurfaceKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySurfaceKHR(m_instance, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroySurfaceKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySurfaceKHR(m_instance, |
| static_cast<VkSurfaceKHR>(surface), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroy(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySurfaceKHR(m_instance, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroy(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySurfaceKHR(m_instance, |
| static_cast<VkSurfaceKHR>(surface), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getSurfaceSupportKHR(uint32_t queueFamilyIndex, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| VULKAN_HPP_NAMESPACE::Bool32 *pSupported, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceSurfaceSupportKHR( |
| m_physicalDevice, queueFamilyIndex, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<VkBool32 *>(pSupported))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Bool32>::type |
| PhysicalDevice::getSurfaceSupportKHR(uint32_t queueFamilyIndex, VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Bool32 supported; |
| VkResult result = d.vkGetPhysicalDeviceSurfaceSupportKHR( |
| m_physicalDevice, queueFamilyIndex, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<VkBool32 *>(&supported)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfaceSupportKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), supported); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getSurfaceCapabilitiesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| VULKAN_HPP_NAMESPACE::SurfaceCapabilitiesKHR *pSurfaceCapabilities, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceSurfaceCapabilitiesKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<VkSurfaceCapabilitiesKHR *>(pSurfaceCapabilities))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceCapabilitiesKHR>::type |
| PhysicalDevice::getSurfaceCapabilitiesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceCapabilitiesKHR surfaceCapabilities; |
| VkResult result = d.vkGetPhysicalDeviceSurfaceCapabilitiesKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<VkSurfaceCapabilitiesKHR *>(&surfaceCapabilities)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfaceCapabilitiesKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surfaceCapabilities); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getSurfaceFormatsKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| uint32_t *pSurfaceFormatCount, |
| VULKAN_HPP_NAMESPACE::SurfaceFormatKHR *pSurfaceFormats, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormatsKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), pSurfaceFormatCount, reinterpret_cast<VkSurfaceFormatKHR *>(pSurfaceFormats))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SurfaceFormatKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<SurfaceFormatKHR, SurfaceFormatKHRAllocator>>::type |
| PhysicalDevice::getSurfaceFormatsKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SurfaceFormatKHR, SurfaceFormatKHRAllocator> surfaceFormats; |
| uint32_t surfaceFormatCount; |
| Result result; |
| do |
| { |
| result = |
| static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormatsKHR(m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &surfaceFormatCount, nullptr)); |
| if((result == Result::eSuccess) && surfaceFormatCount) |
| { |
| surfaceFormats.resize(surfaceFormatCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormatsKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &surfaceFormatCount, reinterpret_cast<VkSurfaceFormatKHR *>(surfaceFormats.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(surfaceFormatCount <= surfaceFormats.size()); |
| if(surfaceFormatCount < surfaceFormats.size()) |
| { |
| surfaceFormats.resize(surfaceFormatCount); |
| } |
| } |
| return createResultValue(result, surfaceFormats, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfaceFormatsKHR"); |
| } |
| |
| template<typename SurfaceFormatKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SurfaceFormatKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<SurfaceFormatKHR, SurfaceFormatKHRAllocator>>::type |
| PhysicalDevice::getSurfaceFormatsKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| SurfaceFormatKHRAllocator &surfaceFormatKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SurfaceFormatKHR, SurfaceFormatKHRAllocator> surfaceFormats(surfaceFormatKHRAllocator); |
| uint32_t surfaceFormatCount; |
| Result result; |
| do |
| { |
| result = |
| static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormatsKHR(m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &surfaceFormatCount, nullptr)); |
| if((result == Result::eSuccess) && surfaceFormatCount) |
| { |
| surfaceFormats.resize(surfaceFormatCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormatsKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &surfaceFormatCount, reinterpret_cast<VkSurfaceFormatKHR *>(surfaceFormats.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(surfaceFormatCount <= surfaceFormats.size()); |
| if(surfaceFormatCount < surfaceFormats.size()) |
| { |
| surfaceFormats.resize(surfaceFormatCount); |
| } |
| } |
| return createResultValue(result, surfaceFormats, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfaceFormatsKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getSurfacePresentModesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| uint32_t *pPresentModeCount, |
| VULKAN_HPP_NAMESPACE::PresentModeKHR *pPresentModes, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceSurfacePresentModesKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), pPresentModeCount, reinterpret_cast<VkPresentModeKHR *>(pPresentModes))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PresentModeKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PresentModeKHR, PresentModeKHRAllocator>>::type |
| PhysicalDevice::getSurfacePresentModesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PresentModeKHR, PresentModeKHRAllocator> presentModes; |
| uint32_t presentModeCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceSurfacePresentModesKHR(m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &presentModeCount, nullptr)); |
| if((result == Result::eSuccess) && presentModeCount) |
| { |
| presentModes.resize(presentModeCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfacePresentModesKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &presentModeCount, reinterpret_cast<VkPresentModeKHR *>(presentModes.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(presentModeCount <= presentModes.size()); |
| if(presentModeCount < presentModes.size()) |
| { |
| presentModes.resize(presentModeCount); |
| } |
| } |
| return createResultValue(result, presentModes, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfacePresentModesKHR"); |
| } |
| |
| template<typename PresentModeKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PresentModeKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PresentModeKHR, PresentModeKHRAllocator>>::type |
| PhysicalDevice::getSurfacePresentModesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| PresentModeKHRAllocator &presentModeKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PresentModeKHR, PresentModeKHRAllocator> presentModes(presentModeKHRAllocator); |
| uint32_t presentModeCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceSurfacePresentModesKHR(m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &presentModeCount, nullptr)); |
| if((result == Result::eSuccess) && presentModeCount) |
| { |
| presentModes.resize(presentModeCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfacePresentModesKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &presentModeCount, reinterpret_cast<VkPresentModeKHR *>(presentModes.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(presentModeCount <= presentModes.size()); |
| if(presentModeCount < presentModes.size()) |
| { |
| presentModes.resize(presentModeCount); |
| } |
| } |
| return createResultValue(result, presentModes, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfacePresentModesKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_swapchain === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createSwapchainKHR(const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SwapchainKHR *pSwapchain, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateSwapchainKHR(m_device, |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSwapchainKHR *>(pSwapchain))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SwapchainKHR>::type |
| Device::createSwapchainKHR(const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain; |
| VkResult result = |
| d.vkCreateSwapchainKHR(m_device, |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSwapchainKHR *>(&swapchain)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createSwapchainKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), swapchain); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SwapchainKHR, Dispatch>>::type |
| Device::createSwapchainKHRUnique(const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain; |
| Result result = static_cast<Result>( |
| d.vkCreateSwapchainKHR(m_device, |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSwapchainKHR *>(&swapchain))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SwapchainKHR, Dispatch>( |
| result, swapchain, VULKAN_HPP_NAMESPACE_STRING "::Device::createSwapchainKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySwapchainKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySwapchainKHR(m_device, static_cast<VkSwapchainKHR>(swapchain), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySwapchainKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySwapchainKHR(m_device, |
| static_cast<VkSwapchainKHR>(swapchain), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySwapchainKHR(m_device, static_cast<VkSwapchainKHR>(swapchain), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySwapchainKHR(m_device, |
| static_cast<VkSwapchainKHR>(swapchain), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getSwapchainImagesKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| uint32_t *pSwapchainImageCount, |
| VULKAN_HPP_NAMESPACE::Image *pSwapchainImages, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetSwapchainImagesKHR(m_device, static_cast<VkSwapchainKHR>(swapchain), pSwapchainImageCount, reinterpret_cast<VkImage *>(pSwapchainImages))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename ImageAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<Image, ImageAllocator>>::type |
| Device::getSwapchainImagesKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<Image, ImageAllocator> swapchainImages; |
| uint32_t swapchainImageCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetSwapchainImagesKHR(m_device, static_cast<VkSwapchainKHR>(swapchain), &swapchainImageCount, nullptr)); |
| if((result == Result::eSuccess) && swapchainImageCount) |
| { |
| swapchainImages.resize(swapchainImageCount); |
| result = static_cast<Result>(d.vkGetSwapchainImagesKHR( |
| m_device, static_cast<VkSwapchainKHR>(swapchain), &swapchainImageCount, reinterpret_cast<VkImage *>(swapchainImages.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(swapchainImageCount <= swapchainImages.size()); |
| if(swapchainImageCount < swapchainImages.size()) |
| { |
| swapchainImages.resize(swapchainImageCount); |
| } |
| } |
| return createResultValue(result, swapchainImages, VULKAN_HPP_NAMESPACE_STRING "::Device::getSwapchainImagesKHR"); |
| } |
| |
| template<typename ImageAllocator, typename Dispatch, typename B, typename std::enable_if<std::is_same<typename B::value_type, Image>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<Image, ImageAllocator>>::type |
| Device::getSwapchainImagesKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, ImageAllocator &imageAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<Image, ImageAllocator> swapchainImages(imageAllocator); |
| uint32_t swapchainImageCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetSwapchainImagesKHR(m_device, static_cast<VkSwapchainKHR>(swapchain), &swapchainImageCount, nullptr)); |
| if((result == Result::eSuccess) && swapchainImageCount) |
| { |
| swapchainImages.resize(swapchainImageCount); |
| result = static_cast<Result>(d.vkGetSwapchainImagesKHR( |
| m_device, static_cast<VkSwapchainKHR>(swapchain), &swapchainImageCount, reinterpret_cast<VkImage *>(swapchainImages.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(swapchainImageCount <= swapchainImages.size()); |
| if(swapchainImageCount < swapchainImages.size()) |
| { |
| swapchainImages.resize(swapchainImageCount); |
| } |
| } |
| return createResultValue(result, swapchainImages, VULKAN_HPP_NAMESPACE_STRING "::Device::getSwapchainImagesKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::acquireNextImageKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| uint64_t timeout, |
| VULKAN_HPP_NAMESPACE::Semaphore semaphore, |
| VULKAN_HPP_NAMESPACE::Fence fence, |
| uint32_t *pImageIndex, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAcquireNextImageKHR( |
| m_device, static_cast<VkSwapchainKHR>(swapchain), timeout, static_cast<VkSemaphore>(semaphore), static_cast<VkFence>(fence), pImageIndex)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<uint32_t> Device::acquireNextImageKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| uint64_t timeout, |
| VULKAN_HPP_NAMESPACE::Semaphore semaphore, |
| VULKAN_HPP_NAMESPACE::Fence fence, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint32_t imageIndex; |
| VkResult result = d.vkAcquireNextImageKHR( |
| m_device, static_cast<VkSwapchainKHR>(swapchain), timeout, static_cast<VkSemaphore>(semaphore), static_cast<VkFence>(fence), &imageIndex); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::acquireNextImageKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, |
| VULKAN_HPP_NAMESPACE::Result::eTimeout, |
| VULKAN_HPP_NAMESPACE::Result::eNotReady, |
| VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR }); |
| return ResultValue<uint32_t>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), imageIndex); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Queue::presentKHR(const VULKAN_HPP_NAMESPACE::PresentInfoKHR *pPresentInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkQueuePresentKHR(m_queue, reinterpret_cast<const VkPresentInfoKHR *>(pPresentInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result Queue::presentKHR(const VULKAN_HPP_NAMESPACE::PresentInfoKHR &presentInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkQueuePresentKHR(m_queue, reinterpret_cast<const VkPresentInfoKHR *>(&presentInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Queue::presentKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getGroupPresentCapabilitiesKHR( |
| VULKAN_HPP_NAMESPACE::DeviceGroupPresentCapabilitiesKHR *pDeviceGroupPresentCapabilities, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetDeviceGroupPresentCapabilitiesKHR(m_device, reinterpret_cast<VkDeviceGroupPresentCapabilitiesKHR *>(pDeviceGroupPresentCapabilities))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DeviceGroupPresentCapabilitiesKHR>::type |
| Device::getGroupPresentCapabilitiesKHR(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DeviceGroupPresentCapabilitiesKHR deviceGroupPresentCapabilities; |
| VkResult result = |
| d.vkGetDeviceGroupPresentCapabilitiesKHR(m_device, reinterpret_cast<VkDeviceGroupPresentCapabilitiesKHR *>(&deviceGroupPresentCapabilities)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getGroupPresentCapabilitiesKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), deviceGroupPresentCapabilities); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getGroupSurfacePresentModesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| VULKAN_HPP_NAMESPACE::DeviceGroupPresentModeFlagsKHR *pModes, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetDeviceGroupSurfacePresentModesKHR( |
| m_device, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<VkDeviceGroupPresentModeFlagsKHR *>(pModes))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DeviceGroupPresentModeFlagsKHR>::type |
| Device::getGroupSurfacePresentModesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DeviceGroupPresentModeFlagsKHR modes; |
| VkResult result = d.vkGetDeviceGroupSurfacePresentModesKHR( |
| m_device, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<VkDeviceGroupPresentModeFlagsKHR *>(&modes)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getGroupSurfacePresentModesKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), modes); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getPresentRectanglesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| uint32_t *pRectCount, |
| VULKAN_HPP_NAMESPACE::Rect2D *pRects, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPhysicalDevicePresentRectanglesKHR(m_physicalDevice, static_cast<VkSurfaceKHR>(surface), pRectCount, reinterpret_cast<VkRect2D *>(pRects))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Rect2DAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<Rect2D, Rect2DAllocator>>::type |
| PhysicalDevice::getPresentRectanglesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<Rect2D, Rect2DAllocator> rects; |
| uint32_t rectCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDevicePresentRectanglesKHR(m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &rectCount, nullptr)); |
| if((result == Result::eSuccess) && rectCount) |
| { |
| rects.resize(rectCount); |
| result = static_cast<Result>(d.vkGetPhysicalDevicePresentRectanglesKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &rectCount, reinterpret_cast<VkRect2D *>(rects.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(rectCount <= rects.size()); |
| if(rectCount < rects.size()) |
| { |
| rects.resize(rectCount); |
| } |
| } |
| return createResultValue(result, rects, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getPresentRectanglesKHR"); |
| } |
| |
| template<typename Rect2DAllocator, typename Dispatch, typename B, typename std::enable_if<std::is_same<typename B::value_type, Rect2D>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<Rect2D, Rect2DAllocator>>::type |
| PhysicalDevice::getPresentRectanglesKHR(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Rect2DAllocator &rect2DAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<Rect2D, Rect2DAllocator> rects(rect2DAllocator); |
| uint32_t rectCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDevicePresentRectanglesKHR(m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &rectCount, nullptr)); |
| if((result == Result::eSuccess) && rectCount) |
| { |
| rects.resize(rectCount); |
| result = static_cast<Result>(d.vkGetPhysicalDevicePresentRectanglesKHR( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), &rectCount, reinterpret_cast<VkRect2D *>(rects.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(rectCount <= rects.size()); |
| if(rectCount < rects.size()) |
| { |
| rects.resize(rectCount); |
| } |
| } |
| return createResultValue(result, rects, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getPresentRectanglesKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::acquireNextImage2KHR(const VULKAN_HPP_NAMESPACE::AcquireNextImageInfoKHR *pAcquireInfo, |
| uint32_t *pImageIndex, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAcquireNextImage2KHR(m_device, reinterpret_cast<const VkAcquireNextImageInfoKHR *>(pAcquireInfo), pImageIndex)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<uint32_t> Device::acquireNextImage2KHR(const VULKAN_HPP_NAMESPACE::AcquireNextImageInfoKHR &acquireInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint32_t imageIndex; |
| VkResult result = d.vkAcquireNextImage2KHR(m_device, reinterpret_cast<const VkAcquireNextImageInfoKHR *>(&acquireInfo), &imageIndex); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::acquireNextImage2KHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, |
| VULKAN_HPP_NAMESPACE::Result::eTimeout, |
| VULKAN_HPP_NAMESPACE::Result::eNotReady, |
| VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR }); |
| return ResultValue<uint32_t>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), imageIndex); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_display === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getDisplayPropertiesKHR(uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::DisplayPropertiesKHR *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPhysicalDeviceDisplayPropertiesKHR(m_physicalDevice, pPropertyCount, reinterpret_cast<VkDisplayPropertiesKHR *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DisplayPropertiesKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayPropertiesKHR, DisplayPropertiesKHRAllocator>>::type |
| PhysicalDevice::getDisplayPropertiesKHR(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayPropertiesKHR, DisplayPropertiesKHRAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPropertiesKHR(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceDisplayPropertiesKHR(m_physicalDevice, &propertyCount, reinterpret_cast<VkDisplayPropertiesKHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPropertiesKHR"); |
| } |
| |
| template<typename DisplayPropertiesKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, DisplayPropertiesKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayPropertiesKHR, DisplayPropertiesKHRAllocator>>::type |
| PhysicalDevice::getDisplayPropertiesKHR(DisplayPropertiesKHRAllocator &displayPropertiesKHRAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayPropertiesKHR, DisplayPropertiesKHRAllocator> properties(displayPropertiesKHRAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPropertiesKHR(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceDisplayPropertiesKHR(m_physicalDevice, &propertyCount, reinterpret_cast<VkDisplayPropertiesKHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPropertiesKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getDisplayPlanePropertiesKHR(uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::DisplayPlanePropertiesKHR *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR(m_physicalDevice, pPropertyCount, reinterpret_cast<VkDisplayPlanePropertiesKHR *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DisplayPlanePropertiesKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayPlanePropertiesKHR, DisplayPlanePropertiesKHRAllocator>>::type |
| PhysicalDevice::getDisplayPlanePropertiesKHR(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayPlanePropertiesKHR, DisplayPlanePropertiesKHRAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR( |
| m_physicalDevice, &propertyCount, reinterpret_cast<VkDisplayPlanePropertiesKHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPlanePropertiesKHR"); |
| } |
| |
| template<typename DisplayPlanePropertiesKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, DisplayPlanePropertiesKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayPlanePropertiesKHR, DisplayPlanePropertiesKHRAllocator>>::type |
| PhysicalDevice::getDisplayPlanePropertiesKHR(DisplayPlanePropertiesKHRAllocator &displayPlanePropertiesKHRAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayPlanePropertiesKHR, DisplayPlanePropertiesKHRAllocator> properties(displayPlanePropertiesKHRAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPlanePropertiesKHR( |
| m_physicalDevice, &propertyCount, reinterpret_cast<VkDisplayPlanePropertiesKHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPlanePropertiesKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getDisplayPlaneSupportedDisplaysKHR(uint32_t planeIndex, |
| uint32_t *pDisplayCount, |
| VULKAN_HPP_NAMESPACE::DisplayKHR *pDisplays, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetDisplayPlaneSupportedDisplaysKHR(m_physicalDevice, planeIndex, pDisplayCount, reinterpret_cast<VkDisplayKHR *>(pDisplays))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DisplayKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayKHR, DisplayKHRAllocator>>::type |
| PhysicalDevice::getDisplayPlaneSupportedDisplaysKHR(uint32_t planeIndex, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayKHR, DisplayKHRAllocator> displays; |
| uint32_t displayCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetDisplayPlaneSupportedDisplaysKHR(m_physicalDevice, planeIndex, &displayCount, nullptr)); |
| if((result == Result::eSuccess) && displayCount) |
| { |
| displays.resize(displayCount); |
| result = static_cast<Result>( |
| d.vkGetDisplayPlaneSupportedDisplaysKHR(m_physicalDevice, planeIndex, &displayCount, reinterpret_cast<VkDisplayKHR *>(displays.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(displayCount <= displays.size()); |
| if(displayCount < displays.size()) |
| { |
| displays.resize(displayCount); |
| } |
| } |
| return createResultValue(result, displays, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPlaneSupportedDisplaysKHR"); |
| } |
| |
| template<typename DisplayKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, DisplayKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayKHR, DisplayKHRAllocator>>::type |
| PhysicalDevice::getDisplayPlaneSupportedDisplaysKHR(uint32_t planeIndex, DisplayKHRAllocator &displayKHRAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayKHR, DisplayKHRAllocator> displays(displayKHRAllocator); |
| uint32_t displayCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetDisplayPlaneSupportedDisplaysKHR(m_physicalDevice, planeIndex, &displayCount, nullptr)); |
| if((result == Result::eSuccess) && displayCount) |
| { |
| displays.resize(displayCount); |
| result = static_cast<Result>( |
| d.vkGetDisplayPlaneSupportedDisplaysKHR(m_physicalDevice, planeIndex, &displayCount, reinterpret_cast<VkDisplayKHR *>(displays.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(displayCount <= displays.size()); |
| if(displayCount < displays.size()) |
| { |
| displays.resize(displayCount); |
| } |
| } |
| return createResultValue(result, displays, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPlaneSupportedDisplaysKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getDisplayModePropertiesKHR(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::DisplayModePropertiesKHR *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetDisplayModePropertiesKHR( |
| m_physicalDevice, static_cast<VkDisplayKHR>(display), pPropertyCount, reinterpret_cast<VkDisplayModePropertiesKHR *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DisplayModePropertiesKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayModePropertiesKHR, DisplayModePropertiesKHRAllocator>>::type |
| PhysicalDevice::getDisplayModePropertiesKHR(VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayModePropertiesKHR, DisplayModePropertiesKHRAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetDisplayModePropertiesKHR(m_physicalDevice, static_cast<VkDisplayKHR>(display), &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetDisplayModePropertiesKHR( |
| m_physicalDevice, static_cast<VkDisplayKHR>(display), &propertyCount, reinterpret_cast<VkDisplayModePropertiesKHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayModePropertiesKHR"); |
| } |
| |
| template<typename DisplayModePropertiesKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, DisplayModePropertiesKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayModePropertiesKHR, DisplayModePropertiesKHRAllocator>>::type |
| PhysicalDevice::getDisplayModePropertiesKHR(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| DisplayModePropertiesKHRAllocator &displayModePropertiesKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayModePropertiesKHR, DisplayModePropertiesKHRAllocator> properties(displayModePropertiesKHRAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetDisplayModePropertiesKHR(m_physicalDevice, static_cast<VkDisplayKHR>(display), &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetDisplayModePropertiesKHR( |
| m_physicalDevice, static_cast<VkDisplayKHR>(display), &propertyCount, reinterpret_cast<VkDisplayModePropertiesKHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayModePropertiesKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::createDisplayModeKHR(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| const VULKAN_HPP_NAMESPACE::DisplayModeCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::DisplayModeKHR *pMode, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDisplayModeKHR(m_physicalDevice, |
| static_cast<VkDisplayKHR>(display), |
| reinterpret_cast<const VkDisplayModeCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkDisplayModeKHR *>(pMode))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DisplayModeKHR>::type |
| PhysicalDevice::createDisplayModeKHR(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| const VULKAN_HPP_NAMESPACE::DisplayModeCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DisplayModeKHR mode; |
| VkResult result = |
| d.vkCreateDisplayModeKHR(m_physicalDevice, |
| static_cast<VkDisplayKHR>(display), |
| reinterpret_cast<const VkDisplayModeCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDisplayModeKHR *>(&mode)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::createDisplayModeKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), mode); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DisplayModeKHR, Dispatch>>::type |
| PhysicalDevice::createDisplayModeKHRUnique(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| const VULKAN_HPP_NAMESPACE::DisplayModeCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DisplayModeKHR mode; |
| Result result = static_cast<Result>( |
| d.vkCreateDisplayModeKHR(m_physicalDevice, |
| static_cast<VkDisplayKHR>(display), |
| reinterpret_cast<const VkDisplayModeCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDisplayModeKHR *>(&mode))); |
| ObjectDestroy<PhysicalDevice, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DisplayModeKHR, Dispatch>( |
| result, mode, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::createDisplayModeKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getDisplayPlaneCapabilitiesKHR(VULKAN_HPP_NAMESPACE::DisplayModeKHR mode, |
| uint32_t planeIndex, |
| VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilitiesKHR *pCapabilities, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetDisplayPlaneCapabilitiesKHR( |
| m_physicalDevice, static_cast<VkDisplayModeKHR>(mode), planeIndex, reinterpret_cast<VkDisplayPlaneCapabilitiesKHR *>(pCapabilities))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilitiesKHR>::type |
| PhysicalDevice::getDisplayPlaneCapabilitiesKHR(VULKAN_HPP_NAMESPACE::DisplayModeKHR mode, uint32_t planeIndex, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilitiesKHR capabilities; |
| VkResult result = d.vkGetDisplayPlaneCapabilitiesKHR( |
| m_physicalDevice, static_cast<VkDisplayModeKHR>(mode), planeIndex, reinterpret_cast<VkDisplayPlaneCapabilitiesKHR *>(&capabilities)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPlaneCapabilitiesKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), capabilities); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createDisplayPlaneSurfaceKHR(const VULKAN_HPP_NAMESPACE::DisplaySurfaceCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDisplayPlaneSurfaceKHR(m_instance, |
| reinterpret_cast<const VkDisplaySurfaceCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createDisplayPlaneSurfaceKHR(const VULKAN_HPP_NAMESPACE::DisplaySurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = d.vkCreateDisplayPlaneSurfaceKHR( |
| m_instance, |
| reinterpret_cast<const VkDisplaySurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createDisplayPlaneSurfaceKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createDisplayPlaneSurfaceKHRUnique(const VULKAN_HPP_NAMESPACE::DisplaySurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>(d.vkCreateDisplayPlaneSurfaceKHR( |
| m_instance, |
| reinterpret_cast<const VkDisplaySurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createDisplayPlaneSurfaceKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_display_swapchain === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createSharedSwapchainsKHR(uint32_t swapchainCount, |
| const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR *pCreateInfos, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SwapchainKHR *pSwapchains, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateSharedSwapchainsKHR(m_device, |
| swapchainCount, |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(pCreateInfos), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSwapchainKHR *>(pSwapchains))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SwapchainKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR, SwapchainKHRAllocator>>::type |
| Device::createSharedSwapchainsKHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR, SwapchainKHRAllocator> swapchains(createInfos.size()); |
| VkResult result = d.vkCreateSharedSwapchainsKHR( |
| m_device, |
| createInfos.size(), |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSwapchainKHR *>(swapchains.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createSharedSwapchainsKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), swapchains); |
| } |
| |
| template<typename SwapchainKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SwapchainKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR, SwapchainKHRAllocator>>::type |
| Device::createSharedSwapchainsKHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| SwapchainKHRAllocator &swapchainKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR, SwapchainKHRAllocator> swapchains(createInfos.size(), swapchainKHRAllocator); |
| VkResult result = d.vkCreateSharedSwapchainsKHR( |
| m_device, |
| createInfos.size(), |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSwapchainKHR *>(swapchains.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createSharedSwapchainsKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), swapchains); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SwapchainKHR>::type |
| Device::createSharedSwapchainKHR(const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain; |
| VkResult result = d.vkCreateSharedSwapchainsKHR( |
| m_device, |
| 1, |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSwapchainKHR *>(&swapchain)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createSharedSwapchainKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), swapchain); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch, typename SwapchainKHRAllocator> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<UniqueHandle<SwapchainKHR, Dispatch>, SwapchainKHRAllocator>>::type |
| Device::createSharedSwapchainsKHRUnique(ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<SwapchainKHR, Dispatch>, SwapchainKHRAllocator> uniqueSwapchains; |
| std::vector<SwapchainKHR> swapchains(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateSharedSwapchainsKHR( |
| m_device, |
| createInfos.size(), |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSwapchainKHR *>(swapchains.data()))); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| uniqueSwapchains.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniqueSwapchains.push_back(UniqueHandle<SwapchainKHR, Dispatch>(swapchains[i], deleter)); |
| } |
| } |
| return createResultValue(result, std::move(uniqueSwapchains), VULKAN_HPP_NAMESPACE_STRING "::Device::createSharedSwapchainsKHRUnique"); |
| } |
| |
| template<typename Dispatch, |
| typename SwapchainKHRAllocator, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, UniqueHandle<SwapchainKHR, Dispatch>>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<UniqueHandle<SwapchainKHR, Dispatch>, SwapchainKHRAllocator>>::type |
| Device::createSharedSwapchainsKHRUnique(ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| SwapchainKHRAllocator &swapchainKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<SwapchainKHR, Dispatch>, SwapchainKHRAllocator> uniqueSwapchains(swapchainKHRAllocator); |
| std::vector<SwapchainKHR> swapchains(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateSharedSwapchainsKHR( |
| m_device, |
| createInfos.size(), |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSwapchainKHR *>(swapchains.data()))); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| uniqueSwapchains.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniqueSwapchains.push_back(UniqueHandle<SwapchainKHR, Dispatch>(swapchains[i], deleter)); |
| } |
| } |
| return createResultValue(result, std::move(uniqueSwapchains), VULKAN_HPP_NAMESPACE_STRING "::Device::createSharedSwapchainsKHRUnique"); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<SwapchainKHR, Dispatch>>::type |
| Device::createSharedSwapchainKHRUnique(const VULKAN_HPP_NAMESPACE::SwapchainCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| SwapchainKHR swapchain; |
| Result result = static_cast<Result>(d.vkCreateSharedSwapchainsKHR( |
| m_device, |
| 1, |
| reinterpret_cast<const VkSwapchainCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSwapchainKHR *>(&swapchain))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<SwapchainKHR, Dispatch>(result, swapchain, VULKAN_HPP_NAMESPACE_STRING "::Device::createSharedSwapchainKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_XLIB_KHR) |
| //=== VK_KHR_xlib_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createXlibSurfaceKHR(const VULKAN_HPP_NAMESPACE::XlibSurfaceCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateXlibSurfaceKHR(m_instance, |
| reinterpret_cast<const VkXlibSurfaceCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createXlibSurfaceKHR(const VULKAN_HPP_NAMESPACE::XlibSurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = |
| d.vkCreateXlibSurfaceKHR(m_instance, |
| reinterpret_cast<const VkXlibSurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createXlibSurfaceKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createXlibSurfaceKHRUnique(const VULKAN_HPP_NAMESPACE::XlibSurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>( |
| d.vkCreateXlibSurfaceKHR(m_instance, |
| reinterpret_cast<const VkXlibSurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createXlibSurfaceKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Bool32 |
| PhysicalDevice::getXlibPresentationSupportKHR(uint32_t queueFamilyIndex, Display *dpy, VisualID visualID, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Bool32>(d.vkGetPhysicalDeviceXlibPresentationSupportKHR(m_physicalDevice, queueFamilyIndex, dpy, visualID)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Bool32 |
| PhysicalDevice::getXlibPresentationSupportKHR(uint32_t queueFamilyIndex, Display &dpy, VisualID visualID, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkBool32 result = d.vkGetPhysicalDeviceXlibPresentationSupportKHR(m_physicalDevice, queueFamilyIndex, &dpy, visualID); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::Bool32>(result); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_XLIB_KHR*/ |
| |
| #if defined(VK_USE_PLATFORM_XCB_KHR) |
| //=== VK_KHR_xcb_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createXcbSurfaceKHR(const VULKAN_HPP_NAMESPACE::XcbSurfaceCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateXcbSurfaceKHR(m_instance, |
| reinterpret_cast<const VkXcbSurfaceCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createXcbSurfaceKHR(const VULKAN_HPP_NAMESPACE::XcbSurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = |
| d.vkCreateXcbSurfaceKHR(m_instance, |
| reinterpret_cast<const VkXcbSurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createXcbSurfaceKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createXcbSurfaceKHRUnique(const VULKAN_HPP_NAMESPACE::XcbSurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>( |
| d.vkCreateXcbSurfaceKHR(m_instance, |
| reinterpret_cast<const VkXcbSurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createXcbSurfaceKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Bool32 PhysicalDevice::getXcbPresentationSupportKHR(uint32_t queueFamilyIndex, |
| xcb_connection_t *connection, |
| xcb_visualid_t visual_id, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Bool32>(d.vkGetPhysicalDeviceXcbPresentationSupportKHR(m_physicalDevice, queueFamilyIndex, connection, visual_id)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Bool32 PhysicalDevice::getXcbPresentationSupportKHR(uint32_t queueFamilyIndex, |
| xcb_connection_t &connection, |
| xcb_visualid_t visual_id, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkBool32 result = d.vkGetPhysicalDeviceXcbPresentationSupportKHR(m_physicalDevice, queueFamilyIndex, &connection, visual_id); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::Bool32>(result); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_XCB_KHR*/ |
| |
| #if defined(VK_USE_PLATFORM_WAYLAND_KHR) |
| //=== VK_KHR_wayland_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createWaylandSurfaceKHR(const VULKAN_HPP_NAMESPACE::WaylandSurfaceCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateWaylandSurfaceKHR(m_instance, |
| reinterpret_cast<const VkWaylandSurfaceCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createWaylandSurfaceKHR(const VULKAN_HPP_NAMESPACE::WaylandSurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = d.vkCreateWaylandSurfaceKHR( |
| m_instance, |
| reinterpret_cast<const VkWaylandSurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createWaylandSurfaceKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createWaylandSurfaceKHRUnique(const VULKAN_HPP_NAMESPACE::WaylandSurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>(d.vkCreateWaylandSurfaceKHR( |
| m_instance, |
| reinterpret_cast<const VkWaylandSurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createWaylandSurfaceKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Bool32 PhysicalDevice::getWaylandPresentationSupportKHR(uint32_t queueFamilyIndex, |
| struct wl_display *display, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Bool32>(d.vkGetPhysicalDeviceWaylandPresentationSupportKHR(m_physicalDevice, queueFamilyIndex, display)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Bool32 |
| PhysicalDevice::getWaylandPresentationSupportKHR(uint32_t queueFamilyIndex, struct wl_display &display, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkBool32 result = d.vkGetPhysicalDeviceWaylandPresentationSupportKHR(m_physicalDevice, queueFamilyIndex, &display); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::Bool32>(result); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_WAYLAND_KHR*/ |
| |
| #if defined(VK_USE_PLATFORM_ANDROID_KHR) |
| //=== VK_KHR_android_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createAndroidSurfaceKHR(const VULKAN_HPP_NAMESPACE::AndroidSurfaceCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateAndroidSurfaceKHR(m_instance, |
| reinterpret_cast<const VkAndroidSurfaceCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createAndroidSurfaceKHR(const VULKAN_HPP_NAMESPACE::AndroidSurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = d.vkCreateAndroidSurfaceKHR( |
| m_instance, |
| reinterpret_cast<const VkAndroidSurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createAndroidSurfaceKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createAndroidSurfaceKHRUnique(const VULKAN_HPP_NAMESPACE::AndroidSurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>(d.vkCreateAndroidSurfaceKHR( |
| m_instance, |
| reinterpret_cast<const VkAndroidSurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createAndroidSurfaceKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_ANDROID_KHR*/ |
| |
| #if defined(VK_USE_PLATFORM_WIN32_KHR) |
| //=== VK_KHR_win32_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createWin32SurfaceKHR(const VULKAN_HPP_NAMESPACE::Win32SurfaceCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateWin32SurfaceKHR(m_instance, |
| reinterpret_cast<const VkWin32SurfaceCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createWin32SurfaceKHR(const VULKAN_HPP_NAMESPACE::Win32SurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = |
| d.vkCreateWin32SurfaceKHR(m_instance, |
| reinterpret_cast<const VkWin32SurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createWin32SurfaceKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createWin32SurfaceKHRUnique(const VULKAN_HPP_NAMESPACE::Win32SurfaceCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>( |
| d.vkCreateWin32SurfaceKHR(m_instance, |
| reinterpret_cast<const VkWin32SurfaceCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createWin32SurfaceKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Bool32 PhysicalDevice::getWin32PresentationSupportKHR(uint32_t queueFamilyIndex, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Bool32>(d.vkGetPhysicalDeviceWin32PresentationSupportKHR(m_physicalDevice, queueFamilyIndex)); |
| } |
| #endif /*VK_USE_PLATFORM_WIN32_KHR*/ |
| |
| //=== VK_EXT_debug_report === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Instance::createDebugReportCallbackEXT(const VULKAN_HPP_NAMESPACE::DebugReportCallbackCreateInfoEXT *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT *pCallback, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDebugReportCallbackEXT(m_instance, |
| reinterpret_cast<const VkDebugReportCallbackCreateInfoEXT *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkDebugReportCallbackEXT *>(pCallback))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT>::type |
| Instance::createDebugReportCallbackEXT(const VULKAN_HPP_NAMESPACE::DebugReportCallbackCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback; |
| VkResult result = d.vkCreateDebugReportCallbackEXT( |
| m_instance, |
| reinterpret_cast<const VkDebugReportCallbackCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDebugReportCallbackEXT *>(&callback)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createDebugReportCallbackEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), callback); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT, Dispatch>>::type |
| Instance::createDebugReportCallbackEXTUnique(const VULKAN_HPP_NAMESPACE::DebugReportCallbackCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback; |
| Result result = static_cast<Result>(d.vkCreateDebugReportCallbackEXT( |
| m_instance, |
| reinterpret_cast<const VkDebugReportCallbackCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDebugReportCallbackEXT *>(&callback))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT, Dispatch>( |
| result, callback, VULKAN_HPP_NAMESPACE_STRING "::Instance::createDebugReportCallbackEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroyDebugReportCallbackEXT(VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDebugReportCallbackEXT( |
| m_instance, static_cast<VkDebugReportCallbackEXT>(callback), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroyDebugReportCallbackEXT(VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDebugReportCallbackEXT( |
| m_instance, |
| static_cast<VkDebugReportCallbackEXT>(callback), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroy(VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDebugReportCallbackEXT( |
| m_instance, static_cast<VkDebugReportCallbackEXT>(callback), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroy(VULKAN_HPP_NAMESPACE::DebugReportCallbackEXT callback, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDebugReportCallbackEXT( |
| m_instance, |
| static_cast<VkDebugReportCallbackEXT>(callback), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::debugReportMessageEXT(VULKAN_HPP_NAMESPACE::DebugReportFlagsEXT flags, |
| VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT objectType, |
| uint64_t object, |
| size_t location, |
| int32_t messageCode, |
| const char *pLayerPrefix, |
| const char *pMessage, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDebugReportMessageEXT(m_instance, |
| static_cast<VkDebugReportFlagsEXT>(flags), |
| static_cast<VkDebugReportObjectTypeEXT>(objectType), |
| object, |
| location, |
| messageCode, |
| pLayerPrefix, |
| pMessage); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::debugReportMessageEXT(VULKAN_HPP_NAMESPACE::DebugReportFlagsEXT flags, |
| VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT objectType, |
| uint64_t object, |
| size_t location, |
| int32_t messageCode, |
| const std::string &layerPrefix, |
| const std::string &message, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDebugReportMessageEXT(m_instance, |
| static_cast<VkDebugReportFlagsEXT>(flags), |
| static_cast<VkDebugReportObjectTypeEXT>(objectType), |
| object, |
| location, |
| messageCode, |
| layerPrefix.c_str(), |
| message.c_str()); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_debug_marker === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::debugMarkerSetObjectTagEXT(const VULKAN_HPP_NAMESPACE::DebugMarkerObjectTagInfoEXT *pTagInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkDebugMarkerSetObjectTagEXT(m_device, reinterpret_cast<const VkDebugMarkerObjectTagInfoEXT *>(pTagInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::debugMarkerSetObjectTagEXT(const VULKAN_HPP_NAMESPACE::DebugMarkerObjectTagInfoEXT &tagInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkDebugMarkerSetObjectTagEXT(m_device, reinterpret_cast<const VkDebugMarkerObjectTagInfoEXT *>(&tagInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::debugMarkerSetObjectTagEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::debugMarkerSetObjectNameEXT(const VULKAN_HPP_NAMESPACE::DebugMarkerObjectNameInfoEXT *pNameInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkDebugMarkerSetObjectNameEXT(m_device, reinterpret_cast<const VkDebugMarkerObjectNameInfoEXT *>(pNameInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::debugMarkerSetObjectNameEXT(const VULKAN_HPP_NAMESPACE::DebugMarkerObjectNameInfoEXT &nameInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkDebugMarkerSetObjectNameEXT(m_device, reinterpret_cast<const VkDebugMarkerObjectNameInfoEXT *>(&nameInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::debugMarkerSetObjectNameEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::debugMarkerBeginEXT(const VULKAN_HPP_NAMESPACE::DebugMarkerMarkerInfoEXT *pMarkerInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDebugMarkerBeginEXT(m_commandBuffer, reinterpret_cast<const VkDebugMarkerMarkerInfoEXT *>(pMarkerInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::debugMarkerBeginEXT(const VULKAN_HPP_NAMESPACE::DebugMarkerMarkerInfoEXT &markerInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdDebugMarkerBeginEXT(m_commandBuffer, reinterpret_cast<const VkDebugMarkerMarkerInfoEXT *>(&markerInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::debugMarkerEndEXT(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDebugMarkerEndEXT(m_commandBuffer); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::debugMarkerInsertEXT(const VULKAN_HPP_NAMESPACE::DebugMarkerMarkerInfoEXT *pMarkerInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDebugMarkerInsertEXT(m_commandBuffer, reinterpret_cast<const VkDebugMarkerMarkerInfoEXT *>(pMarkerInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::debugMarkerInsertEXT(const VULKAN_HPP_NAMESPACE::DebugMarkerMarkerInfoEXT &markerInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdDebugMarkerInsertEXT(m_commandBuffer, reinterpret_cast<const VkDebugMarkerMarkerInfoEXT *>(&markerInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_ENABLE_BETA_EXTENSIONS) |
| //=== VK_KHR_video_queue === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getVideoCapabilitiesKHR(const VULKAN_HPP_NAMESPACE::VideoProfileKHR *pVideoProfile, |
| VULKAN_HPP_NAMESPACE::VideoCapabilitiesKHR *pCapabilities, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceVideoCapabilitiesKHR( |
| m_physicalDevice, reinterpret_cast<const VkVideoProfileKHR *>(pVideoProfile), reinterpret_cast<VkVideoCapabilitiesKHR *>(pCapabilities))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::VideoCapabilitiesKHR>::type |
| PhysicalDevice::getVideoCapabilitiesKHR(const VULKAN_HPP_NAMESPACE::VideoProfileKHR &videoProfile, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::VideoCapabilitiesKHR capabilities; |
| VkResult result = d.vkGetPhysicalDeviceVideoCapabilitiesKHR( |
| m_physicalDevice, reinterpret_cast<const VkVideoProfileKHR *>(&videoProfile), reinterpret_cast<VkVideoCapabilitiesKHR *>(&capabilities)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getVideoCapabilitiesKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), capabilities); |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<StructureChain<X, Y, Z...>>::type |
| PhysicalDevice::getVideoCapabilitiesKHR(const VULKAN_HPP_NAMESPACE::VideoProfileKHR &videoProfile, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::VideoCapabilitiesKHR &capabilities = structureChain.template get<VULKAN_HPP_NAMESPACE::VideoCapabilitiesKHR>(); |
| Result result = static_cast<Result>(d.vkGetPhysicalDeviceVideoCapabilitiesKHR( |
| m_physicalDevice, reinterpret_cast<const VkVideoProfileKHR *>(&videoProfile), reinterpret_cast<VkVideoCapabilitiesKHR *>(&capabilities))); |
| return createResultValue(result, structureChain, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getVideoCapabilitiesKHR"); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getVideoFormatPropertiesKHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceVideoFormatInfoKHR *pVideoFormatInfo, |
| uint32_t *pVideoFormatPropertyCount, |
| VULKAN_HPP_NAMESPACE::VideoFormatPropertiesKHR *pVideoFormatProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceVideoFormatPropertiesKHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceVideoFormatInfoKHR *>(pVideoFormatInfo), |
| pVideoFormatPropertyCount, |
| reinterpret_cast<VkVideoFormatPropertiesKHR *>(pVideoFormatProperties))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename VideoFormatPropertiesKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VideoFormatPropertiesKHR, VideoFormatPropertiesKHRAllocator>>::type |
| PhysicalDevice::getVideoFormatPropertiesKHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceVideoFormatInfoKHR &videoFormatInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<VideoFormatPropertiesKHR, VideoFormatPropertiesKHRAllocator> videoFormatProperties; |
| uint32_t videoFormatPropertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceVideoFormatPropertiesKHR( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceVideoFormatInfoKHR *>(&videoFormatInfo), &videoFormatPropertyCount, nullptr)); |
| if((result == Result::eSuccess) && videoFormatPropertyCount) |
| { |
| videoFormatProperties.resize(videoFormatPropertyCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceVideoFormatPropertiesKHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceVideoFormatInfoKHR *>(&videoFormatInfo), |
| &videoFormatPropertyCount, |
| reinterpret_cast<VkVideoFormatPropertiesKHR *>(videoFormatProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(videoFormatPropertyCount <= videoFormatProperties.size()); |
| if(videoFormatPropertyCount < videoFormatProperties.size()) |
| { |
| videoFormatProperties.resize(videoFormatPropertyCount); |
| } |
| } |
| return createResultValue(result, videoFormatProperties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getVideoFormatPropertiesKHR"); |
| } |
| |
| template<typename VideoFormatPropertiesKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, VideoFormatPropertiesKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VideoFormatPropertiesKHR, VideoFormatPropertiesKHRAllocator>>::type |
| PhysicalDevice::getVideoFormatPropertiesKHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceVideoFormatInfoKHR &videoFormatInfo, |
| VideoFormatPropertiesKHRAllocator &videoFormatPropertiesKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<VideoFormatPropertiesKHR, VideoFormatPropertiesKHRAllocator> videoFormatProperties(videoFormatPropertiesKHRAllocator); |
| uint32_t videoFormatPropertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceVideoFormatPropertiesKHR( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceVideoFormatInfoKHR *>(&videoFormatInfo), &videoFormatPropertyCount, nullptr)); |
| if((result == Result::eSuccess) && videoFormatPropertyCount) |
| { |
| videoFormatProperties.resize(videoFormatPropertyCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceVideoFormatPropertiesKHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceVideoFormatInfoKHR *>(&videoFormatInfo), |
| &videoFormatPropertyCount, |
| reinterpret_cast<VkVideoFormatPropertiesKHR *>(videoFormatProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(videoFormatPropertyCount <= videoFormatProperties.size()); |
| if(videoFormatPropertyCount < videoFormatProperties.size()) |
| { |
| videoFormatProperties.resize(videoFormatPropertyCount); |
| } |
| } |
| return createResultValue(result, videoFormatProperties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getVideoFormatPropertiesKHR"); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createVideoSessionKHR(const VULKAN_HPP_NAMESPACE::VideoSessionCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::VideoSessionKHR *pVideoSession, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateVideoSessionKHR(m_device, |
| reinterpret_cast<const VkVideoSessionCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkVideoSessionKHR *>(pVideoSession))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::VideoSessionKHR>::type |
| Device::createVideoSessionKHR(const VULKAN_HPP_NAMESPACE::VideoSessionCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession; |
| VkResult result = |
| d.vkCreateVideoSessionKHR(m_device, |
| reinterpret_cast<const VkVideoSessionCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkVideoSessionKHR *>(&videoSession)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createVideoSessionKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), videoSession); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::VideoSessionKHR, Dispatch>>::type |
| Device::createVideoSessionKHRUnique(const VULKAN_HPP_NAMESPACE::VideoSessionCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession; |
| Result result = static_cast<Result>( |
| d.vkCreateVideoSessionKHR(m_device, |
| reinterpret_cast<const VkVideoSessionCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkVideoSessionKHR *>(&videoSession))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::VideoSessionKHR, Dispatch>( |
| result, videoSession, VULKAN_HPP_NAMESPACE_STRING "::Device::createVideoSessionKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyVideoSessionKHR(VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyVideoSessionKHR(m_device, static_cast<VkVideoSessionKHR>(videoSession), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyVideoSessionKHR(VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyVideoSessionKHR( |
| m_device, |
| static_cast<VkVideoSessionKHR>(videoSession), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyVideoSessionKHR(m_device, static_cast<VkVideoSessionKHR>(videoSession), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyVideoSessionKHR( |
| m_device, |
| static_cast<VkVideoSessionKHR>(videoSession), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getVideoSessionMemoryRequirementsKHR(VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, |
| uint32_t *pVideoSessionMemoryRequirementsCount, |
| VULKAN_HPP_NAMESPACE::VideoGetMemoryPropertiesKHR *pVideoSessionMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetVideoSessionMemoryRequirementsKHR(m_device, |
| static_cast<VkVideoSessionKHR>(videoSession), |
| pVideoSessionMemoryRequirementsCount, |
| reinterpret_cast<VkVideoGetMemoryPropertiesKHR *>(pVideoSessionMemoryRequirements))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename VideoGetMemoryPropertiesKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VideoGetMemoryPropertiesKHR, VideoGetMemoryPropertiesKHRAllocator>>::type |
| Device::getVideoSessionMemoryRequirementsKHR(VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<VideoGetMemoryPropertiesKHR, VideoGetMemoryPropertiesKHRAllocator> videoSessionMemoryRequirements; |
| uint32_t videoSessionMemoryRequirementsCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>( |
| d.vkGetVideoSessionMemoryRequirementsKHR(m_device, static_cast<VkVideoSessionKHR>(videoSession), &videoSessionMemoryRequirementsCount, nullptr)); |
| if((result == Result::eSuccess) && videoSessionMemoryRequirementsCount) |
| { |
| videoSessionMemoryRequirements.resize(videoSessionMemoryRequirementsCount); |
| result = static_cast<Result>( |
| d.vkGetVideoSessionMemoryRequirementsKHR(m_device, |
| static_cast<VkVideoSessionKHR>(videoSession), |
| &videoSessionMemoryRequirementsCount, |
| reinterpret_cast<VkVideoGetMemoryPropertiesKHR *>(videoSessionMemoryRequirements.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(videoSessionMemoryRequirementsCount <= videoSessionMemoryRequirements.size()); |
| if(videoSessionMemoryRequirementsCount < videoSessionMemoryRequirements.size()) |
| { |
| videoSessionMemoryRequirements.resize(videoSessionMemoryRequirementsCount); |
| } |
| } |
| return createResultValue(result, videoSessionMemoryRequirements, VULKAN_HPP_NAMESPACE_STRING "::Device::getVideoSessionMemoryRequirementsKHR"); |
| } |
| |
| template<typename VideoGetMemoryPropertiesKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, VideoGetMemoryPropertiesKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<VideoGetMemoryPropertiesKHR, VideoGetMemoryPropertiesKHRAllocator>>::type |
| Device::getVideoSessionMemoryRequirementsKHR(VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, |
| VideoGetMemoryPropertiesKHRAllocator &videoGetMemoryPropertiesKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<VideoGetMemoryPropertiesKHR, VideoGetMemoryPropertiesKHRAllocator> videoSessionMemoryRequirements(videoGetMemoryPropertiesKHRAllocator); |
| uint32_t videoSessionMemoryRequirementsCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>( |
| d.vkGetVideoSessionMemoryRequirementsKHR(m_device, static_cast<VkVideoSessionKHR>(videoSession), &videoSessionMemoryRequirementsCount, nullptr)); |
| if((result == Result::eSuccess) && videoSessionMemoryRequirementsCount) |
| { |
| videoSessionMemoryRequirements.resize(videoSessionMemoryRequirementsCount); |
| result = static_cast<Result>( |
| d.vkGetVideoSessionMemoryRequirementsKHR(m_device, |
| static_cast<VkVideoSessionKHR>(videoSession), |
| &videoSessionMemoryRequirementsCount, |
| reinterpret_cast<VkVideoGetMemoryPropertiesKHR *>(videoSessionMemoryRequirements.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(videoSessionMemoryRequirementsCount <= videoSessionMemoryRequirements.size()); |
| if(videoSessionMemoryRequirementsCount < videoSessionMemoryRequirements.size()) |
| { |
| videoSessionMemoryRequirements.resize(videoSessionMemoryRequirementsCount); |
| } |
| } |
| return createResultValue(result, videoSessionMemoryRequirements, VULKAN_HPP_NAMESPACE_STRING "::Device::getVideoSessionMemoryRequirementsKHR"); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::bindVideoSessionMemoryKHR(VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, |
| uint32_t videoSessionBindMemoryCount, |
| const VULKAN_HPP_NAMESPACE::VideoBindMemoryKHR *pVideoSessionBindMemories, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkBindVideoSessionMemoryKHR(m_device, |
| static_cast<VkVideoSessionKHR>(videoSession), |
| videoSessionBindMemoryCount, |
| reinterpret_cast<const VkVideoBindMemoryKHR *>(pVideoSessionBindMemories))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::bindVideoSessionMemoryKHR(VULKAN_HPP_NAMESPACE::VideoSessionKHR videoSession, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::VideoBindMemoryKHR> const &videoSessionBindMemories, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkBindVideoSessionMemoryKHR(m_device, |
| static_cast<VkVideoSessionKHR>(videoSession), |
| videoSessionBindMemories.size(), |
| reinterpret_cast<const VkVideoBindMemoryKHR *>(videoSessionBindMemories.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::bindVideoSessionMemoryKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createVideoSessionParametersKHR(const VULKAN_HPP_NAMESPACE::VideoSessionParametersCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR *pVideoSessionParameters, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateVideoSessionParametersKHR(m_device, |
| reinterpret_cast<const VkVideoSessionParametersCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkVideoSessionParametersKHR *>(pVideoSessionParameters))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR>::type |
| Device::createVideoSessionParametersKHR(const VULKAN_HPP_NAMESPACE::VideoSessionParametersCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters; |
| VkResult result = d.vkCreateVideoSessionParametersKHR( |
| m_device, |
| reinterpret_cast<const VkVideoSessionParametersCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkVideoSessionParametersKHR *>(&videoSessionParameters)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createVideoSessionParametersKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), videoSessionParameters); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR, Dispatch>>::type |
| Device::createVideoSessionParametersKHRUnique(const VULKAN_HPP_NAMESPACE::VideoSessionParametersCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters; |
| Result result = static_cast<Result>(d.vkCreateVideoSessionParametersKHR( |
| m_device, |
| reinterpret_cast<const VkVideoSessionParametersCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkVideoSessionParametersKHR *>(&videoSessionParameters))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR, Dispatch>( |
| result, videoSessionParameters, VULKAN_HPP_NAMESPACE_STRING "::Device::createVideoSessionParametersKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::updateVideoSessionParametersKHR(VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters, |
| const VULKAN_HPP_NAMESPACE::VideoSessionParametersUpdateInfoKHR *pUpdateInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkUpdateVideoSessionParametersKHR(m_device, |
| static_cast<VkVideoSessionParametersKHR>(videoSessionParameters), |
| reinterpret_cast<const VkVideoSessionParametersUpdateInfoKHR *>(pUpdateInfo))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::updateVideoSessionParametersKHR(VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters, |
| const VULKAN_HPP_NAMESPACE::VideoSessionParametersUpdateInfoKHR &updateInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkUpdateVideoSessionParametersKHR(m_device, |
| static_cast<VkVideoSessionParametersKHR>(videoSessionParameters), |
| reinterpret_cast<const VkVideoSessionParametersUpdateInfoKHR *>(&updateInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::updateVideoSessionParametersKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyVideoSessionParametersKHR(VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyVideoSessionParametersKHR( |
| m_device, static_cast<VkVideoSessionParametersKHR>(videoSessionParameters), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyVideoSessionParametersKHR(VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyVideoSessionParametersKHR( |
| m_device, |
| static_cast<VkVideoSessionParametersKHR>(videoSessionParameters), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyVideoSessionParametersKHR( |
| m_device, static_cast<VkVideoSessionParametersKHR>(videoSessionParameters), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::VideoSessionParametersKHR videoSessionParameters, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyVideoSessionParametersKHR( |
| m_device, |
| static_cast<VkVideoSessionParametersKHR>(videoSessionParameters), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginVideoCodingKHR(const VULKAN_HPP_NAMESPACE::VideoBeginCodingInfoKHR *pBeginInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginVideoCodingKHR(m_commandBuffer, reinterpret_cast<const VkVideoBeginCodingInfoKHR *>(pBeginInfo)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginVideoCodingKHR(const VULKAN_HPP_NAMESPACE::VideoBeginCodingInfoKHR &beginInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBeginVideoCodingKHR(m_commandBuffer, reinterpret_cast<const VkVideoBeginCodingInfoKHR *>(&beginInfo)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endVideoCodingKHR(const VULKAN_HPP_NAMESPACE::VideoEndCodingInfoKHR *pEndCodingInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndVideoCodingKHR(m_commandBuffer, reinterpret_cast<const VkVideoEndCodingInfoKHR *>(pEndCodingInfo)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endVideoCodingKHR(const VULKAN_HPP_NAMESPACE::VideoEndCodingInfoKHR &endCodingInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdEndVideoCodingKHR(m_commandBuffer, reinterpret_cast<const VkVideoEndCodingInfoKHR *>(&endCodingInfo)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::controlVideoCodingKHR(const VULKAN_HPP_NAMESPACE::VideoCodingControlInfoKHR *pCodingControlInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdControlVideoCodingKHR(m_commandBuffer, reinterpret_cast<const VkVideoCodingControlInfoKHR *>(pCodingControlInfo)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::controlVideoCodingKHR(const VULKAN_HPP_NAMESPACE::VideoCodingControlInfoKHR &codingControlInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdControlVideoCodingKHR(m_commandBuffer, reinterpret_cast<const VkVideoCodingControlInfoKHR *>(&codingControlInfo)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_ENABLE_BETA_EXTENSIONS*/ |
| |
| #if defined(VK_ENABLE_BETA_EXTENSIONS) |
| //=== VK_KHR_video_decode_queue === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::decodeVideoKHR(const VULKAN_HPP_NAMESPACE::VideoDecodeInfoKHR *pFrameInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDecodeVideoKHR(m_commandBuffer, reinterpret_cast<const VkVideoDecodeInfoKHR *>(pFrameInfo)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::decodeVideoKHR(const VULKAN_HPP_NAMESPACE::VideoDecodeInfoKHR &frameInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdDecodeVideoKHR(m_commandBuffer, reinterpret_cast<const VkVideoDecodeInfoKHR *>(&frameInfo)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_ENABLE_BETA_EXTENSIONS*/ |
| |
| //=== VK_EXT_transform_feedback === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindTransformFeedbackBuffersEXT(uint32_t firstBinding, |
| uint32_t bindingCount, |
| const VULKAN_HPP_NAMESPACE::Buffer *pBuffers, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pOffsets, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pSizes, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindTransformFeedbackBuffersEXT(m_commandBuffer, |
| firstBinding, |
| bindingCount, |
| reinterpret_cast<const VkBuffer *>(pBuffers), |
| reinterpret_cast<const VkDeviceSize *>(pOffsets), |
| reinterpret_cast<const VkDeviceSize *>(pSizes)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindTransformFeedbackBuffersEXT(uint32_t firstBinding, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer> const &buffers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &offsets, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &sizes, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(buffers.size() == offsets.size()); |
| VULKAN_HPP_ASSERT(sizes.empty() || buffers.size() == sizes.size()); |
| # else |
| if(buffers.size() != offsets.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindTransformFeedbackBuffersEXT: buffers.size() != offsets.size()"); |
| } |
| if(!sizes.empty() && buffers.size() != sizes.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindTransformFeedbackBuffersEXT: buffers.size() != sizes.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdBindTransformFeedbackBuffersEXT(m_commandBuffer, |
| firstBinding, |
| buffers.size(), |
| reinterpret_cast<const VkBuffer *>(buffers.data()), |
| reinterpret_cast<const VkDeviceSize *>(offsets.data()), |
| reinterpret_cast<const VkDeviceSize *>(sizes.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginTransformFeedbackEXT(uint32_t firstCounterBuffer, |
| uint32_t counterBufferCount, |
| const VULKAN_HPP_NAMESPACE::Buffer *pCounterBuffers, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pCounterBufferOffsets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginTransformFeedbackEXT(m_commandBuffer, |
| firstCounterBuffer, |
| counterBufferCount, |
| reinterpret_cast<const VkBuffer *>(pCounterBuffers), |
| reinterpret_cast<const VkDeviceSize *>(pCounterBufferOffsets)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginTransformFeedbackEXT(uint32_t firstCounterBuffer, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer> const &counterBuffers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &counterBufferOffsets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(counterBufferOffsets.empty() || counterBuffers.size() == counterBufferOffsets.size()); |
| # else |
| if(!counterBufferOffsets.empty() && counterBuffers.size() != counterBufferOffsets.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::beginTransformFeedbackEXT: counterBuffers.size() != counterBufferOffsets.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdBeginTransformFeedbackEXT(m_commandBuffer, |
| firstCounterBuffer, |
| counterBuffers.size(), |
| reinterpret_cast<const VkBuffer *>(counterBuffers.data()), |
| reinterpret_cast<const VkDeviceSize *>(counterBufferOffsets.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endTransformFeedbackEXT(uint32_t firstCounterBuffer, |
| uint32_t counterBufferCount, |
| const VULKAN_HPP_NAMESPACE::Buffer *pCounterBuffers, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pCounterBufferOffsets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndTransformFeedbackEXT(m_commandBuffer, |
| firstCounterBuffer, |
| counterBufferCount, |
| reinterpret_cast<const VkBuffer *>(pCounterBuffers), |
| reinterpret_cast<const VkDeviceSize *>(pCounterBufferOffsets)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endTransformFeedbackEXT(uint32_t firstCounterBuffer, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer> const &counterBuffers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &counterBufferOffsets, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(counterBufferOffsets.empty() || counterBuffers.size() == counterBufferOffsets.size()); |
| # else |
| if(!counterBufferOffsets.empty() && counterBuffers.size() != counterBufferOffsets.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::endTransformFeedbackEXT: counterBuffers.size() != counterBufferOffsets.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdEndTransformFeedbackEXT(m_commandBuffer, |
| firstCounterBuffer, |
| counterBuffers.size(), |
| reinterpret_cast<const VkBuffer *>(counterBuffers.data()), |
| reinterpret_cast<const VkDeviceSize *>(counterBufferOffsets.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginQueryIndexedEXT(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t query, |
| VULKAN_HPP_NAMESPACE::QueryControlFlags flags, |
| uint32_t index, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginQueryIndexedEXT(m_commandBuffer, static_cast<VkQueryPool>(queryPool), query, static_cast<VkQueryControlFlags>(flags), index); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::endQueryIndexedEXT(VULKAN_HPP_NAMESPACE::QueryPool queryPool, uint32_t query, uint32_t index, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndQueryIndexedEXT(m_commandBuffer, static_cast<VkQueryPool>(queryPool), query, index); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndirectByteCountEXT(uint32_t instanceCount, |
| uint32_t firstInstance, |
| VULKAN_HPP_NAMESPACE::Buffer counterBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize counterBufferOffset, |
| uint32_t counterOffset, |
| uint32_t vertexStride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndirectByteCountEXT(m_commandBuffer, |
| instanceCount, |
| firstInstance, |
| static_cast<VkBuffer>(counterBuffer), |
| static_cast<VkDeviceSize>(counterBufferOffset), |
| counterOffset, |
| vertexStride); |
| } |
| |
| //=== VK_NVX_binary_import === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createCuModuleNVX(const VULKAN_HPP_NAMESPACE::CuModuleCreateInfoNVX *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::CuModuleNVX *pModule, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateCuModuleNVX(m_device, |
| reinterpret_cast<const VkCuModuleCreateInfoNVX *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkCuModuleNVX *>(pModule))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::CuModuleNVX>::type |
| Device::createCuModuleNVX(const VULKAN_HPP_NAMESPACE::CuModuleCreateInfoNVX &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::CuModuleNVX module; |
| VkResult result = |
| d.vkCreateCuModuleNVX(m_device, |
| reinterpret_cast<const VkCuModuleCreateInfoNVX *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkCuModuleNVX *>(&module)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createCuModuleNVX"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), module); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::CuModuleNVX, Dispatch>>::type |
| Device::createCuModuleNVXUnique(const VULKAN_HPP_NAMESPACE::CuModuleCreateInfoNVX &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::CuModuleNVX module; |
| Result result = static_cast<Result>( |
| d.vkCreateCuModuleNVX(m_device, |
| reinterpret_cast<const VkCuModuleCreateInfoNVX *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkCuModuleNVX *>(&module))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::CuModuleNVX, Dispatch>( |
| result, module, VULKAN_HPP_NAMESPACE_STRING "::Device::createCuModuleNVXUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createCuFunctionNVX(const VULKAN_HPP_NAMESPACE::CuFunctionCreateInfoNVX *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::CuFunctionNVX *pFunction, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateCuFunctionNVX(m_device, |
| reinterpret_cast<const VkCuFunctionCreateInfoNVX *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkCuFunctionNVX *>(pFunction))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::CuFunctionNVX>::type |
| Device::createCuFunctionNVX(const VULKAN_HPP_NAMESPACE::CuFunctionCreateInfoNVX &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::CuFunctionNVX function; |
| VkResult result = |
| d.vkCreateCuFunctionNVX(m_device, |
| reinterpret_cast<const VkCuFunctionCreateInfoNVX *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkCuFunctionNVX *>(&function)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createCuFunctionNVX"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), function); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::CuFunctionNVX, Dispatch>>::type |
| Device::createCuFunctionNVXUnique(const VULKAN_HPP_NAMESPACE::CuFunctionCreateInfoNVX &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::CuFunctionNVX function; |
| Result result = static_cast<Result>( |
| d.vkCreateCuFunctionNVX(m_device, |
| reinterpret_cast<const VkCuFunctionCreateInfoNVX *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkCuFunctionNVX *>(&function))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::CuFunctionNVX, Dispatch>( |
| result, function, VULKAN_HPP_NAMESPACE_STRING "::Device::createCuFunctionNVXUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyCuModuleNVX(VULKAN_HPP_NAMESPACE::CuModuleNVX module, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyCuModuleNVX(m_device, static_cast<VkCuModuleNVX>(module), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyCuModuleNVX(VULKAN_HPP_NAMESPACE::CuModuleNVX module, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyCuModuleNVX(m_device, |
| static_cast<VkCuModuleNVX>(module), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::CuModuleNVX module, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyCuModuleNVX(m_device, static_cast<VkCuModuleNVX>(module), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::CuModuleNVX module, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyCuModuleNVX(m_device, |
| static_cast<VkCuModuleNVX>(module), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyCuFunctionNVX(VULKAN_HPP_NAMESPACE::CuFunctionNVX function, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyCuFunctionNVX(m_device, static_cast<VkCuFunctionNVX>(function), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyCuFunctionNVX(VULKAN_HPP_NAMESPACE::CuFunctionNVX function, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyCuFunctionNVX(m_device, |
| static_cast<VkCuFunctionNVX>(function), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::CuFunctionNVX function, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyCuFunctionNVX(m_device, static_cast<VkCuFunctionNVX>(function), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::CuFunctionNVX function, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyCuFunctionNVX(m_device, |
| static_cast<VkCuFunctionNVX>(function), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::cuLaunchKernelNVX(const VULKAN_HPP_NAMESPACE::CuLaunchInfoNVX *pLaunchInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCuLaunchKernelNVX(m_commandBuffer, reinterpret_cast<const VkCuLaunchInfoNVX *>(pLaunchInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::cuLaunchKernelNVX(const VULKAN_HPP_NAMESPACE::CuLaunchInfoNVX &launchInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCuLaunchKernelNVX(m_commandBuffer, reinterpret_cast<const VkCuLaunchInfoNVX *>(&launchInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NVX_image_view_handle === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint32_t Device::getImageViewHandleNVX(const VULKAN_HPP_NAMESPACE::ImageViewHandleInfoNVX *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return d.vkGetImageViewHandleNVX(m_device, reinterpret_cast<const VkImageViewHandleInfoNVX *>(pInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint32_t Device::getImageViewHandleNVX(const VULKAN_HPP_NAMESPACE::ImageViewHandleInfoNVX &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint32_t result = d.vkGetImageViewHandleNVX(m_device, reinterpret_cast<const VkImageViewHandleInfoNVX *>(&info)); |
| |
| return result; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getImageViewAddressNVX(VULKAN_HPP_NAMESPACE::ImageView imageView, |
| VULKAN_HPP_NAMESPACE::ImageViewAddressPropertiesNVX *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetImageViewAddressNVX(m_device, static_cast<VkImageView>(imageView), reinterpret_cast<VkImageViewAddressPropertiesNVX *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::ImageViewAddressPropertiesNVX>::type |
| Device::getImageViewAddressNVX(VULKAN_HPP_NAMESPACE::ImageView imageView, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ImageViewAddressPropertiesNVX properties; |
| VkResult result = |
| d.vkGetImageViewAddressNVX(m_device, static_cast<VkImageView>(imageView), reinterpret_cast<VkImageViewAddressPropertiesNVX *>(&properties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getImageViewAddressNVX"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), properties); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_AMD_draw_indirect_count === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndirectCountAMD(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::Buffer countBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize countBufferOffset, |
| uint32_t maxDrawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndirectCountAMD(m_commandBuffer, |
| static_cast<VkBuffer>(buffer), |
| static_cast<VkDeviceSize>(offset), |
| static_cast<VkBuffer>(countBuffer), |
| static_cast<VkDeviceSize>(countBufferOffset), |
| maxDrawCount, |
| stride); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndexedIndirectCountAMD(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::Buffer countBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize countBufferOffset, |
| uint32_t maxDrawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndexedIndirectCountAMD(m_commandBuffer, |
| static_cast<VkBuffer>(buffer), |
| static_cast<VkDeviceSize>(offset), |
| static_cast<VkBuffer>(countBuffer), |
| static_cast<VkDeviceSize>(countBufferOffset), |
| maxDrawCount, |
| stride); |
| } |
| |
| //=== VK_AMD_shader_info === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getShaderInfoAMD(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| VULKAN_HPP_NAMESPACE::ShaderStageFlagBits shaderStage, |
| VULKAN_HPP_NAMESPACE::ShaderInfoTypeAMD infoType, |
| size_t *pInfoSize, |
| void *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetShaderInfoAMD(m_device, |
| static_cast<VkPipeline>(pipeline), |
| static_cast<VkShaderStageFlagBits>(shaderStage), |
| static_cast<VkShaderInfoTypeAMD>(infoType), |
| pInfoSize, |
| pInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Uint8_tAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type |
| Device::getShaderInfoAMD(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| VULKAN_HPP_NAMESPACE::ShaderStageFlagBits shaderStage, |
| VULKAN_HPP_NAMESPACE::ShaderInfoTypeAMD infoType, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<uint8_t, Uint8_tAllocator> info; |
| size_t infoSize; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetShaderInfoAMD(m_device, |
| static_cast<VkPipeline>(pipeline), |
| static_cast<VkShaderStageFlagBits>(shaderStage), |
| static_cast<VkShaderInfoTypeAMD>(infoType), |
| &infoSize, |
| nullptr)); |
| if((result == Result::eSuccess) && infoSize) |
| { |
| info.resize(infoSize); |
| result = static_cast<Result>(d.vkGetShaderInfoAMD(m_device, |
| static_cast<VkPipeline>(pipeline), |
| static_cast<VkShaderStageFlagBits>(shaderStage), |
| static_cast<VkShaderInfoTypeAMD>(infoType), |
| &infoSize, |
| reinterpret_cast<void *>(info.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(infoSize <= info.size()); |
| if(infoSize < info.size()) |
| { |
| info.resize(infoSize); |
| } |
| } |
| return createResultValue(result, info, VULKAN_HPP_NAMESPACE_STRING "::Device::getShaderInfoAMD"); |
| } |
| |
| template<typename Uint8_tAllocator, typename Dispatch, typename B, typename std::enable_if<std::is_same<typename B::value_type, uint8_t>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type |
| Device::getShaderInfoAMD(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| VULKAN_HPP_NAMESPACE::ShaderStageFlagBits shaderStage, |
| VULKAN_HPP_NAMESPACE::ShaderInfoTypeAMD infoType, |
| Uint8_tAllocator &uint8_tAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<uint8_t, Uint8_tAllocator> info(uint8_tAllocator); |
| size_t infoSize; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetShaderInfoAMD(m_device, |
| static_cast<VkPipeline>(pipeline), |
| static_cast<VkShaderStageFlagBits>(shaderStage), |
| static_cast<VkShaderInfoTypeAMD>(infoType), |
| &infoSize, |
| nullptr)); |
| if((result == Result::eSuccess) && infoSize) |
| { |
| info.resize(infoSize); |
| result = static_cast<Result>(d.vkGetShaderInfoAMD(m_device, |
| static_cast<VkPipeline>(pipeline), |
| static_cast<VkShaderStageFlagBits>(shaderStage), |
| static_cast<VkShaderInfoTypeAMD>(infoType), |
| &infoSize, |
| reinterpret_cast<void *>(info.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(infoSize <= info.size()); |
| if(infoSize < info.size()) |
| { |
| info.resize(infoSize); |
| } |
| } |
| return createResultValue(result, info, VULKAN_HPP_NAMESPACE_STRING "::Device::getShaderInfoAMD"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_dynamic_rendering === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRenderingKHR(const VULKAN_HPP_NAMESPACE::RenderingInfo *pRenderingInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginRenderingKHR(m_commandBuffer, reinterpret_cast<const VkRenderingInfo *>(pRenderingInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRenderingKHR(const VULKAN_HPP_NAMESPACE::RenderingInfo &renderingInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBeginRenderingKHR(m_commandBuffer, reinterpret_cast<const VkRenderingInfo *>(&renderingInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endRenderingKHR(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndRenderingKHR(m_commandBuffer); |
| } |
| |
| #if defined(VK_USE_PLATFORM_GGP) |
| //=== VK_GGP_stream_descriptor_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Instance::createStreamDescriptorSurfaceGGP(const VULKAN_HPP_NAMESPACE::StreamDescriptorSurfaceCreateInfoGGP *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateStreamDescriptorSurfaceGGP(m_instance, |
| reinterpret_cast<const VkStreamDescriptorSurfaceCreateInfoGGP *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createStreamDescriptorSurfaceGGP(const VULKAN_HPP_NAMESPACE::StreamDescriptorSurfaceCreateInfoGGP &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = d.vkCreateStreamDescriptorSurfaceGGP( |
| m_instance, |
| reinterpret_cast<const VkStreamDescriptorSurfaceCreateInfoGGP *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createStreamDescriptorSurfaceGGP"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createStreamDescriptorSurfaceGGPUnique(const VULKAN_HPP_NAMESPACE::StreamDescriptorSurfaceCreateInfoGGP &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>(d.vkCreateStreamDescriptorSurfaceGGP( |
| m_instance, |
| reinterpret_cast<const VkStreamDescriptorSurfaceCreateInfoGGP *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createStreamDescriptorSurfaceGGPUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_GGP*/ |
| |
| //=== VK_NV_external_memory_capabilities === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getExternalImageFormatPropertiesNV(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::ImageType type, |
| VULKAN_HPP_NAMESPACE::ImageTiling tiling, |
| VULKAN_HPP_NAMESPACE::ImageUsageFlags usage, |
| VULKAN_HPP_NAMESPACE::ImageCreateFlags flags, |
| VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagsNV externalHandleType, |
| VULKAN_HPP_NAMESPACE::ExternalImageFormatPropertiesNV *pExternalImageFormatProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPhysicalDeviceExternalImageFormatPropertiesNV(m_physicalDevice, |
| static_cast<VkFormat>(format), |
| static_cast<VkImageType>(type), |
| static_cast<VkImageTiling>(tiling), |
| static_cast<VkImageUsageFlags>(usage), |
| static_cast<VkImageCreateFlags>(flags), |
| static_cast<VkExternalMemoryHandleTypeFlagsNV>(externalHandleType), |
| reinterpret_cast<VkExternalImageFormatPropertiesNV *>(pExternalImageFormatProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::ExternalImageFormatPropertiesNV>::type |
| PhysicalDevice::getExternalImageFormatPropertiesNV(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::ImageType type, |
| VULKAN_HPP_NAMESPACE::ImageTiling tiling, |
| VULKAN_HPP_NAMESPACE::ImageUsageFlags usage, |
| VULKAN_HPP_NAMESPACE::ImageCreateFlags flags, |
| VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagsNV externalHandleType, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ExternalImageFormatPropertiesNV externalImageFormatProperties; |
| VkResult result = |
| d.vkGetPhysicalDeviceExternalImageFormatPropertiesNV(m_physicalDevice, |
| static_cast<VkFormat>(format), |
| static_cast<VkImageType>(type), |
| static_cast<VkImageTiling>(tiling), |
| static_cast<VkImageUsageFlags>(usage), |
| static_cast<VkImageCreateFlags>(flags), |
| static_cast<VkExternalMemoryHandleTypeFlagsNV>(externalHandleType), |
| reinterpret_cast<VkExternalImageFormatPropertiesNV *>(&externalImageFormatProperties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getExternalImageFormatPropertiesNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), externalImageFormatProperties); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_WIN32_KHR) |
| //=== VK_NV_external_memory_win32 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getMemoryWin32HandleNV(VULKAN_HPP_NAMESPACE::DeviceMemory memory, |
| VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagsNV handleType, |
| HANDLE *pHandle, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetMemoryWin32HandleNV(m_device, static_cast<VkDeviceMemory>(memory), static_cast<VkExternalMemoryHandleTypeFlagsNV>(handleType), pHandle)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<HANDLE>::type Device::getMemoryWin32HandleNV( |
| VULKAN_HPP_NAMESPACE::DeviceMemory memory, VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagsNV handleType, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| HANDLE handle; |
| VkResult result = |
| d.vkGetMemoryWin32HandleNV(m_device, static_cast<VkDeviceMemory>(memory), static_cast<VkExternalMemoryHandleTypeFlagsNV>(handleType), &handle); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryWin32HandleNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), handle); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_WIN32_KHR*/ |
| |
| //=== VK_KHR_get_physical_device_properties2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getFeatures2KHR(VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 *pFeatures, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceFeatures2KHR(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceFeatures2 *>(pFeatures)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 |
| PhysicalDevice::getFeatures2KHR(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 features; |
| d.vkGetPhysicalDeviceFeatures2KHR(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceFeatures2 *>(&features)); |
| |
| return features; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> PhysicalDevice::getFeatures2KHR(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2 &features = structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceFeatures2>(); |
| d.vkGetPhysicalDeviceFeatures2KHR(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceFeatures2 *>(&features)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getProperties2KHR(VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceProperties2KHR(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceProperties2 *>(pProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 |
| PhysicalDevice::getProperties2KHR(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 properties; |
| d.vkGetPhysicalDeviceProperties2KHR(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceProperties2 *>(&properties)); |
| |
| return properties; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> PhysicalDevice::getProperties2KHR(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2 &properties = structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceProperties2>(); |
| d.vkGetPhysicalDeviceProperties2KHR(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceProperties2 *>(&properties)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getFormatProperties2KHR(VULKAN_HPP_NAMESPACE::Format format, |
| VULKAN_HPP_NAMESPACE::FormatProperties2 *pFormatProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceFormatProperties2KHR( |
| m_physicalDevice, static_cast<VkFormat>(format), reinterpret_cast<VkFormatProperties2 *>(pFormatProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::FormatProperties2 |
| PhysicalDevice::getFormatProperties2KHR(VULKAN_HPP_NAMESPACE::Format format, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::FormatProperties2 formatProperties; |
| d.vkGetPhysicalDeviceFormatProperties2KHR( |
| m_physicalDevice, static_cast<VkFormat>(format), reinterpret_cast<VkFormatProperties2 *>(&formatProperties)); |
| |
| return formatProperties; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> PhysicalDevice::getFormatProperties2KHR(VULKAN_HPP_NAMESPACE::Format format, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::FormatProperties2 &formatProperties = structureChain.template get<VULKAN_HPP_NAMESPACE::FormatProperties2>(); |
| d.vkGetPhysicalDeviceFormatProperties2KHR( |
| m_physicalDevice, static_cast<VkFormat>(format), reinterpret_cast<VkFormatProperties2 *>(&formatProperties)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getImageFormatProperties2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 *pImageFormatInfo, |
| VULKAN_HPP_NAMESPACE::ImageFormatProperties2 *pImageFormatProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceImageFormatProperties2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceImageFormatInfo2 *>(pImageFormatInfo), |
| reinterpret_cast<VkImageFormatProperties2 *>(pImageFormatProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::ImageFormatProperties2>::type |
| PhysicalDevice::getImageFormatProperties2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 &imageFormatInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ImageFormatProperties2 imageFormatProperties; |
| VkResult result = d.vkGetPhysicalDeviceImageFormatProperties2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceImageFormatInfo2 *>(&imageFormatInfo), |
| reinterpret_cast<VkImageFormatProperties2 *>(&imageFormatProperties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getImageFormatProperties2KHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), imageFormatProperties); |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<StructureChain<X, Y, Z...>>::type |
| PhysicalDevice::getImageFormatProperties2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceImageFormatInfo2 &imageFormatInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::ImageFormatProperties2 &imageFormatProperties = structureChain.template get<VULKAN_HPP_NAMESPACE::ImageFormatProperties2>(); |
| Result result = |
| static_cast<Result>(d.vkGetPhysicalDeviceImageFormatProperties2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceImageFormatInfo2 *>(&imageFormatInfo), |
| reinterpret_cast<VkImageFormatProperties2 *>(&imageFormatProperties))); |
| return createResultValue(result, structureChain, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getImageFormatProperties2KHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getQueueFamilyProperties2KHR(uint32_t *pQueueFamilyPropertyCount, |
| VULKAN_HPP_NAMESPACE::QueueFamilyProperties2 *pQueueFamilyProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceQueueFamilyProperties2KHR( |
| m_physicalDevice, pQueueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(pQueueFamilyProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename QueueFamilyProperties2Allocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<QueueFamilyProperties2, QueueFamilyProperties2Allocator> |
| PhysicalDevice::getQueueFamilyProperties2KHR(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<QueueFamilyProperties2, QueueFamilyProperties2Allocator> queueFamilyProperties; |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties2KHR(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| queueFamilyProperties.resize(queueFamilyPropertyCount); |
| d.vkGetPhysicalDeviceQueueFamilyProperties2KHR( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| return queueFamilyProperties; |
| } |
| |
| template<typename QueueFamilyProperties2Allocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, QueueFamilyProperties2>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<QueueFamilyProperties2, QueueFamilyProperties2Allocator> |
| PhysicalDevice::getQueueFamilyProperties2KHR(QueueFamilyProperties2Allocator &queueFamilyProperties2Allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<QueueFamilyProperties2, QueueFamilyProperties2Allocator> queueFamilyProperties(queueFamilyProperties2Allocator); |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties2KHR(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| queueFamilyProperties.resize(queueFamilyPropertyCount); |
| d.vkGetPhysicalDeviceQueueFamilyProperties2KHR( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| return queueFamilyProperties; |
| } |
| |
| template<typename StructureChain, typename StructureChainAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<StructureChain, StructureChainAllocator> |
| PhysicalDevice::getQueueFamilyProperties2KHR(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties2KHR(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| std::vector<StructureChain, StructureChainAllocator> returnVector(queueFamilyPropertyCount); |
| std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2> queueFamilyProperties(queueFamilyPropertyCount); |
| for(uint32_t i = 0; i < queueFamilyPropertyCount; i++) |
| { |
| queueFamilyProperties[i].pNext = returnVector[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>().pNext; |
| } |
| d.vkGetPhysicalDeviceQueueFamilyProperties2KHR( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| for(uint32_t i = 0; i < queueFamilyPropertyCount; i++) |
| { |
| returnVector[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>() = queueFamilyProperties[i]; |
| } |
| return returnVector; |
| } |
| |
| template<typename StructureChain, |
| typename StructureChainAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, StructureChain>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<StructureChain, StructureChainAllocator> |
| PhysicalDevice::getQueueFamilyProperties2KHR(StructureChainAllocator &structureChainAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| uint32_t queueFamilyPropertyCount; |
| d.vkGetPhysicalDeviceQueueFamilyProperties2KHR(m_physicalDevice, &queueFamilyPropertyCount, nullptr); |
| std::vector<StructureChain, StructureChainAllocator> returnVector(queueFamilyPropertyCount, structureChainAllocator); |
| std::vector<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2> queueFamilyProperties(queueFamilyPropertyCount); |
| for(uint32_t i = 0; i < queueFamilyPropertyCount; i++) |
| { |
| queueFamilyProperties[i].pNext = returnVector[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>().pNext; |
| } |
| d.vkGetPhysicalDeviceQueueFamilyProperties2KHR( |
| m_physicalDevice, &queueFamilyPropertyCount, reinterpret_cast<VkQueueFamilyProperties2 *>(queueFamilyProperties.data())); |
| VULKAN_HPP_ASSERT(queueFamilyPropertyCount <= queueFamilyProperties.size()); |
| for(uint32_t i = 0; i < queueFamilyPropertyCount; i++) |
| { |
| returnVector[i].template get<VULKAN_HPP_NAMESPACE::QueueFamilyProperties2>() = queueFamilyProperties[i]; |
| } |
| return returnVector; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getMemoryProperties2KHR(VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 *pMemoryProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceMemoryProperties2KHR(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceMemoryProperties2 *>(pMemoryProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 |
| PhysicalDevice::getMemoryProperties2KHR(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 memoryProperties; |
| d.vkGetPhysicalDeviceMemoryProperties2KHR(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceMemoryProperties2 *>(&memoryProperties)); |
| |
| return memoryProperties; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> PhysicalDevice::getMemoryProperties2KHR(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2 &memoryProperties = |
| structureChain.template get<VULKAN_HPP_NAMESPACE::PhysicalDeviceMemoryProperties2>(); |
| d.vkGetPhysicalDeviceMemoryProperties2KHR(m_physicalDevice, reinterpret_cast<VkPhysicalDeviceMemoryProperties2 *>(&memoryProperties)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getSparseImageFormatProperties2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 *pFormatInfo, |
| uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::SparseImageFormatProperties2 *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(pFormatInfo), |
| pPropertyCount, |
| reinterpret_cast<VkSparseImageFormatProperties2 *>(pProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SparseImageFormatProperties2Allocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> |
| PhysicalDevice::getSparseImageFormatProperties2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 &formatInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> properties; |
| uint32_t propertyCount; |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(&formatInfo), &propertyCount, nullptr); |
| properties.resize(propertyCount); |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(&formatInfo), |
| &propertyCount, |
| reinterpret_cast<VkSparseImageFormatProperties2 *>(properties.data())); |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| return properties; |
| } |
| |
| template<typename SparseImageFormatProperties2Allocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SparseImageFormatProperties2>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> |
| PhysicalDevice::getSparseImageFormatProperties2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSparseImageFormatInfo2 &formatInfo, |
| SparseImageFormatProperties2Allocator &sparseImageFormatProperties2Allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageFormatProperties2, SparseImageFormatProperties2Allocator> properties(sparseImageFormatProperties2Allocator); |
| uint32_t propertyCount; |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(&formatInfo), &propertyCount, nullptr); |
| properties.resize(propertyCount); |
| d.vkGetPhysicalDeviceSparseImageFormatProperties2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSparseImageFormatInfo2 *>(&formatInfo), |
| &propertyCount, |
| reinterpret_cast<VkSparseImageFormatProperties2 *>(properties.data())); |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| return properties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_device_group === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getGroupPeerMemoryFeaturesKHR(uint32_t heapIndex, |
| uint32_t localDeviceIndex, |
| uint32_t remoteDeviceIndex, |
| VULKAN_HPP_NAMESPACE::PeerMemoryFeatureFlags *pPeerMemoryFeatures, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceGroupPeerMemoryFeaturesKHR( |
| m_device, heapIndex, localDeviceIndex, remoteDeviceIndex, reinterpret_cast<VkPeerMemoryFeatureFlags *>(pPeerMemoryFeatures)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::PeerMemoryFeatureFlags Device::getGroupPeerMemoryFeaturesKHR( |
| uint32_t heapIndex, uint32_t localDeviceIndex, uint32_t remoteDeviceIndex, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PeerMemoryFeatureFlags peerMemoryFeatures; |
| d.vkGetDeviceGroupPeerMemoryFeaturesKHR( |
| m_device, heapIndex, localDeviceIndex, remoteDeviceIndex, reinterpret_cast<VkPeerMemoryFeatureFlags *>(&peerMemoryFeatures)); |
| |
| return peerMemoryFeatures; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDeviceMaskKHR(uint32_t deviceMask, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDeviceMaskKHR(m_commandBuffer, deviceMask); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::dispatchBaseKHR(uint32_t baseGroupX, |
| uint32_t baseGroupY, |
| uint32_t baseGroupZ, |
| uint32_t groupCountX, |
| uint32_t groupCountY, |
| uint32_t groupCountZ, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDispatchBaseKHR(m_commandBuffer, baseGroupX, baseGroupY, baseGroupZ, groupCountX, groupCountY, groupCountZ); |
| } |
| |
| #if defined(VK_USE_PLATFORM_VI_NN) |
| //=== VK_NN_vi_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createViSurfaceNN(const VULKAN_HPP_NAMESPACE::ViSurfaceCreateInfoNN *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateViSurfaceNN(m_instance, |
| reinterpret_cast<const VkViSurfaceCreateInfoNN *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createViSurfaceNN(const VULKAN_HPP_NAMESPACE::ViSurfaceCreateInfoNN &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = |
| d.vkCreateViSurfaceNN(m_instance, |
| reinterpret_cast<const VkViSurfaceCreateInfoNN *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createViSurfaceNN"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createViSurfaceNNUnique(const VULKAN_HPP_NAMESPACE::ViSurfaceCreateInfoNN &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>( |
| d.vkCreateViSurfaceNN(m_instance, |
| reinterpret_cast<const VkViSurfaceCreateInfoNN *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createViSurfaceNNUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_VI_NN*/ |
| |
| //=== VK_KHR_maintenance1 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::trimCommandPoolKHR(VULKAN_HPP_NAMESPACE::CommandPool commandPool, |
| VULKAN_HPP_NAMESPACE::CommandPoolTrimFlags flags, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkTrimCommandPoolKHR(m_device, static_cast<VkCommandPool>(commandPool), static_cast<VkCommandPoolTrimFlags>(flags)); |
| } |
| |
| //=== VK_KHR_device_group_creation === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Instance::enumeratePhysicalDeviceGroupsKHR(uint32_t *pPhysicalDeviceGroupCount, |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceGroupProperties *pPhysicalDeviceGroupProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkEnumeratePhysicalDeviceGroupsKHR( |
| m_instance, pPhysicalDeviceGroupCount, reinterpret_cast<VkPhysicalDeviceGroupProperties *>(pPhysicalDeviceGroupProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PhysicalDeviceGroupPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator>>::type |
| Instance::enumeratePhysicalDeviceGroupsKHR(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator> physicalDeviceGroupProperties; |
| uint32_t physicalDeviceGroupCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceGroupsKHR(m_instance, &physicalDeviceGroupCount, nullptr)); |
| if((result == Result::eSuccess) && physicalDeviceGroupCount) |
| { |
| physicalDeviceGroupProperties.resize(physicalDeviceGroupCount); |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceGroupsKHR( |
| m_instance, &physicalDeviceGroupCount, reinterpret_cast<VkPhysicalDeviceGroupProperties *>(physicalDeviceGroupProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(physicalDeviceGroupCount <= physicalDeviceGroupProperties.size()); |
| if(physicalDeviceGroupCount < physicalDeviceGroupProperties.size()) |
| { |
| physicalDeviceGroupProperties.resize(physicalDeviceGroupCount); |
| } |
| } |
| return createResultValue(result, physicalDeviceGroupProperties, VULKAN_HPP_NAMESPACE_STRING "::Instance::enumeratePhysicalDeviceGroupsKHR"); |
| } |
| |
| template<typename PhysicalDeviceGroupPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PhysicalDeviceGroupProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator>>::type |
| Instance::enumeratePhysicalDeviceGroupsKHR(PhysicalDeviceGroupPropertiesAllocator &physicalDeviceGroupPropertiesAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceGroupProperties, PhysicalDeviceGroupPropertiesAllocator> physicalDeviceGroupProperties(physicalDeviceGroupPropertiesAllocator); |
| uint32_t physicalDeviceGroupCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceGroupsKHR(m_instance, &physicalDeviceGroupCount, nullptr)); |
| if((result == Result::eSuccess) && physicalDeviceGroupCount) |
| { |
| physicalDeviceGroupProperties.resize(physicalDeviceGroupCount); |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceGroupsKHR( |
| m_instance, &physicalDeviceGroupCount, reinterpret_cast<VkPhysicalDeviceGroupProperties *>(physicalDeviceGroupProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(physicalDeviceGroupCount <= physicalDeviceGroupProperties.size()); |
| if(physicalDeviceGroupCount < physicalDeviceGroupProperties.size()) |
| { |
| physicalDeviceGroupProperties.resize(physicalDeviceGroupCount); |
| } |
| } |
| return createResultValue(result, physicalDeviceGroupProperties, VULKAN_HPP_NAMESPACE_STRING "::Instance::enumeratePhysicalDeviceGroupsKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_external_memory_capabilities === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getExternalBufferPropertiesKHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalBufferInfo *pExternalBufferInfo, |
| VULKAN_HPP_NAMESPACE::ExternalBufferProperties *pExternalBufferProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceExternalBufferPropertiesKHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalBufferInfo *>(pExternalBufferInfo), |
| reinterpret_cast<VkExternalBufferProperties *>(pExternalBufferProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::ExternalBufferProperties |
| PhysicalDevice::getExternalBufferPropertiesKHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalBufferInfo &externalBufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ExternalBufferProperties externalBufferProperties; |
| d.vkGetPhysicalDeviceExternalBufferPropertiesKHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalBufferInfo *>(&externalBufferInfo), |
| reinterpret_cast<VkExternalBufferProperties *>(&externalBufferProperties)); |
| |
| return externalBufferProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_WIN32_KHR) |
| //=== VK_KHR_external_memory_win32 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getMemoryWin32HandleKHR(const VULKAN_HPP_NAMESPACE::MemoryGetWin32HandleInfoKHR *pGetWin32HandleInfo, |
| HANDLE *pHandle, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetMemoryWin32HandleKHR(m_device, reinterpret_cast<const VkMemoryGetWin32HandleInfoKHR *>(pGetWin32HandleInfo), pHandle)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<HANDLE>::type |
| Device::getMemoryWin32HandleKHR(const VULKAN_HPP_NAMESPACE::MemoryGetWin32HandleInfoKHR &getWin32HandleInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| HANDLE handle; |
| VkResult result = d.vkGetMemoryWin32HandleKHR(m_device, reinterpret_cast<const VkMemoryGetWin32HandleInfoKHR *>(&getWin32HandleInfo), &handle); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryWin32HandleKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), handle); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getMemoryWin32HandlePropertiesKHR(VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagBits handleType, |
| HANDLE handle, |
| VULKAN_HPP_NAMESPACE::MemoryWin32HandlePropertiesKHR *pMemoryWin32HandleProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetMemoryWin32HandlePropertiesKHR(m_device, |
| static_cast<VkExternalMemoryHandleTypeFlagBits>(handleType), |
| handle, |
| reinterpret_cast<VkMemoryWin32HandlePropertiesKHR *>(pMemoryWin32HandleProperties))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::MemoryWin32HandlePropertiesKHR>::type |
| Device::getMemoryWin32HandlePropertiesKHR(VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagBits handleType, HANDLE handle, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryWin32HandlePropertiesKHR memoryWin32HandleProperties; |
| VkResult result = d.vkGetMemoryWin32HandlePropertiesKHR(m_device, |
| static_cast<VkExternalMemoryHandleTypeFlagBits>(handleType), |
| handle, |
| reinterpret_cast<VkMemoryWin32HandlePropertiesKHR *>(&memoryWin32HandleProperties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryWin32HandlePropertiesKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), memoryWin32HandleProperties); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_WIN32_KHR*/ |
| |
| //=== VK_KHR_external_memory_fd === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getMemoryFdKHR(const VULKAN_HPP_NAMESPACE::MemoryGetFdInfoKHR *pGetFdInfo, |
| int *pFd, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetMemoryFdKHR(m_device, reinterpret_cast<const VkMemoryGetFdInfoKHR *>(pGetFdInfo), pFd)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<int>::type Device::getMemoryFdKHR(const VULKAN_HPP_NAMESPACE::MemoryGetFdInfoKHR &getFdInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| int fd; |
| VkResult result = d.vkGetMemoryFdKHR(m_device, reinterpret_cast<const VkMemoryGetFdInfoKHR *>(&getFdInfo), &fd); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryFdKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), fd); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getMemoryFdPropertiesKHR(VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagBits handleType, |
| int fd, |
| VULKAN_HPP_NAMESPACE::MemoryFdPropertiesKHR *pMemoryFdProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetMemoryFdPropertiesKHR( |
| m_device, static_cast<VkExternalMemoryHandleTypeFlagBits>(handleType), fd, reinterpret_cast<VkMemoryFdPropertiesKHR *>(pMemoryFdProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::MemoryFdPropertiesKHR>::type |
| Device::getMemoryFdPropertiesKHR(VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagBits handleType, int fd, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryFdPropertiesKHR memoryFdProperties; |
| VkResult result = d.vkGetMemoryFdPropertiesKHR( |
| m_device, static_cast<VkExternalMemoryHandleTypeFlagBits>(handleType), fd, reinterpret_cast<VkMemoryFdPropertiesKHR *>(&memoryFdProperties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryFdPropertiesKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), memoryFdProperties); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_external_semaphore_capabilities === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| PhysicalDevice::getExternalSemaphorePropertiesKHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalSemaphoreInfo *pExternalSemaphoreInfo, |
| VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties *pExternalSemaphoreProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceExternalSemaphorePropertiesKHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalSemaphoreInfo *>(pExternalSemaphoreInfo), |
| reinterpret_cast<VkExternalSemaphoreProperties *>(pExternalSemaphoreProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties |
| PhysicalDevice::getExternalSemaphorePropertiesKHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalSemaphoreInfo &externalSemaphoreInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ExternalSemaphoreProperties externalSemaphoreProperties; |
| d.vkGetPhysicalDeviceExternalSemaphorePropertiesKHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalSemaphoreInfo *>(&externalSemaphoreInfo), |
| reinterpret_cast<VkExternalSemaphoreProperties *>(&externalSemaphoreProperties)); |
| |
| return externalSemaphoreProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_WIN32_KHR) |
| //=== VK_KHR_external_semaphore_win32 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::importSemaphoreWin32HandleKHR( |
| const VULKAN_HPP_NAMESPACE::ImportSemaphoreWin32HandleInfoKHR *pImportSemaphoreWin32HandleInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkImportSemaphoreWin32HandleKHR(m_device, reinterpret_cast<const VkImportSemaphoreWin32HandleInfoKHR *>(pImportSemaphoreWin32HandleInfo))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::importSemaphoreWin32HandleKHR(const VULKAN_HPP_NAMESPACE::ImportSemaphoreWin32HandleInfoKHR &importSemaphoreWin32HandleInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = |
| d.vkImportSemaphoreWin32HandleKHR(m_device, reinterpret_cast<const VkImportSemaphoreWin32HandleInfoKHR *>(&importSemaphoreWin32HandleInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::importSemaphoreWin32HandleKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getSemaphoreWin32HandleKHR( |
| const VULKAN_HPP_NAMESPACE::SemaphoreGetWin32HandleInfoKHR *pGetWin32HandleInfo, HANDLE *pHandle, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetSemaphoreWin32HandleKHR(m_device, reinterpret_cast<const VkSemaphoreGetWin32HandleInfoKHR *>(pGetWin32HandleInfo), pHandle)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<HANDLE>::type |
| Device::getSemaphoreWin32HandleKHR(const VULKAN_HPP_NAMESPACE::SemaphoreGetWin32HandleInfoKHR &getWin32HandleInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| HANDLE handle; |
| VkResult result = d.vkGetSemaphoreWin32HandleKHR(m_device, reinterpret_cast<const VkSemaphoreGetWin32HandleInfoKHR *>(&getWin32HandleInfo), &handle); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getSemaphoreWin32HandleKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), handle); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_WIN32_KHR*/ |
| |
| //=== VK_KHR_external_semaphore_fd === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::importSemaphoreFdKHR(const VULKAN_HPP_NAMESPACE::ImportSemaphoreFdInfoKHR *pImportSemaphoreFdInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkImportSemaphoreFdKHR(m_device, reinterpret_cast<const VkImportSemaphoreFdInfoKHR *>(pImportSemaphoreFdInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::importSemaphoreFdKHR(const VULKAN_HPP_NAMESPACE::ImportSemaphoreFdInfoKHR &importSemaphoreFdInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkImportSemaphoreFdKHR(m_device, reinterpret_cast<const VkImportSemaphoreFdInfoKHR *>(&importSemaphoreFdInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::importSemaphoreFdKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getSemaphoreFdKHR(const VULKAN_HPP_NAMESPACE::SemaphoreGetFdInfoKHR *pGetFdInfo, |
| int *pFd, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetSemaphoreFdKHR(m_device, reinterpret_cast<const VkSemaphoreGetFdInfoKHR *>(pGetFdInfo), pFd)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<int>::type |
| Device::getSemaphoreFdKHR(const VULKAN_HPP_NAMESPACE::SemaphoreGetFdInfoKHR &getFdInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| int fd; |
| VkResult result = d.vkGetSemaphoreFdKHR(m_device, reinterpret_cast<const VkSemaphoreGetFdInfoKHR *>(&getFdInfo), &fd); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getSemaphoreFdKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), fd); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_push_descriptor === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pushDescriptorSetKHR(VULKAN_HPP_NAMESPACE::PipelineBindPoint pipelineBindPoint, |
| VULKAN_HPP_NAMESPACE::PipelineLayout layout, |
| uint32_t set, |
| uint32_t descriptorWriteCount, |
| const VULKAN_HPP_NAMESPACE::WriteDescriptorSet *pDescriptorWrites, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdPushDescriptorSetKHR(m_commandBuffer, |
| static_cast<VkPipelineBindPoint>(pipelineBindPoint), |
| static_cast<VkPipelineLayout>(layout), |
| set, |
| descriptorWriteCount, |
| reinterpret_cast<const VkWriteDescriptorSet *>(pDescriptorWrites)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pushDescriptorSetKHR(VULKAN_HPP_NAMESPACE::PipelineBindPoint pipelineBindPoint, |
| VULKAN_HPP_NAMESPACE::PipelineLayout layout, |
| uint32_t set, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::WriteDescriptorSet> const &descriptorWrites, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdPushDescriptorSetKHR(m_commandBuffer, |
| static_cast<VkPipelineBindPoint>(pipelineBindPoint), |
| static_cast<VkPipelineLayout>(layout), |
| set, |
| descriptorWrites.size(), |
| reinterpret_cast<const VkWriteDescriptorSet *>(descriptorWrites.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pushDescriptorSetWithTemplateKHR(VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| VULKAN_HPP_NAMESPACE::PipelineLayout layout, |
| uint32_t set, |
| const void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdPushDescriptorSetWithTemplateKHR( |
| m_commandBuffer, static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), static_cast<VkPipelineLayout>(layout), set, pData); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pushDescriptorSetWithTemplateKHR(VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| VULKAN_HPP_NAMESPACE::PipelineLayout layout, |
| uint32_t set, |
| DataType const &data, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdPushDescriptorSetWithTemplateKHR(m_commandBuffer, |
| static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), |
| static_cast<VkPipelineLayout>(layout), |
| set, |
| reinterpret_cast<const void *>(&data)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_conditional_rendering === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginConditionalRenderingEXT(const VULKAN_HPP_NAMESPACE::ConditionalRenderingBeginInfoEXT *pConditionalRenderingBegin, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginConditionalRenderingEXT(m_commandBuffer, reinterpret_cast<const VkConditionalRenderingBeginInfoEXT *>(pConditionalRenderingBegin)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginConditionalRenderingEXT(const VULKAN_HPP_NAMESPACE::ConditionalRenderingBeginInfoEXT &conditionalRenderingBegin, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBeginConditionalRenderingEXT(m_commandBuffer, reinterpret_cast<const VkConditionalRenderingBeginInfoEXT *>(&conditionalRenderingBegin)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endConditionalRenderingEXT(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndConditionalRenderingEXT(m_commandBuffer); |
| } |
| |
| //=== VK_KHR_descriptor_update_template === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createDescriptorUpdateTemplateKHR(const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate *pDescriptorUpdateTemplate, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDescriptorUpdateTemplateKHR(m_device, |
| reinterpret_cast<const VkDescriptorUpdateTemplateCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkDescriptorUpdateTemplate *>(pDescriptorUpdateTemplate))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate>::type |
| Device::createDescriptorUpdateTemplateKHR(const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate; |
| VkResult result = d.vkCreateDescriptorUpdateTemplateKHR( |
| m_device, |
| reinterpret_cast<const VkDescriptorUpdateTemplateCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDescriptorUpdateTemplate *>(&descriptorUpdateTemplate)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createDescriptorUpdateTemplateKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), descriptorUpdateTemplate); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate, Dispatch>>::type |
| Device::createDescriptorUpdateTemplateKHRUnique(const VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplateCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate; |
| Result result = static_cast<Result>(d.vkCreateDescriptorUpdateTemplateKHR( |
| m_device, |
| reinterpret_cast<const VkDescriptorUpdateTemplateCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDescriptorUpdateTemplate *>(&descriptorUpdateTemplate))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate, Dispatch>( |
| result, descriptorUpdateTemplate, VULKAN_HPP_NAMESPACE_STRING "::Device::createDescriptorUpdateTemplateKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDescriptorUpdateTemplateKHR(VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDescriptorUpdateTemplateKHR( |
| m_device, static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDescriptorUpdateTemplateKHR(VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDescriptorUpdateTemplateKHR( |
| m_device, |
| static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::updateDescriptorSetWithTemplateKHR(VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet, |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| const void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkUpdateDescriptorSetWithTemplateKHR( |
| m_device, static_cast<VkDescriptorSet>(descriptorSet), static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), pData); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_INLINE void Device::updateDescriptorSetWithTemplateKHR(VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet, |
| VULKAN_HPP_NAMESPACE::DescriptorUpdateTemplate descriptorUpdateTemplate, |
| DataType const &data, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkUpdateDescriptorSetWithTemplateKHR(m_device, |
| static_cast<VkDescriptorSet>(descriptorSet), |
| static_cast<VkDescriptorUpdateTemplate>(descriptorUpdateTemplate), |
| reinterpret_cast<const void *>(&data)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NV_clip_space_w_scaling === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setViewportWScalingNV(uint32_t firstViewport, |
| uint32_t viewportCount, |
| const VULKAN_HPP_NAMESPACE::ViewportWScalingNV *pViewportWScalings, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetViewportWScalingNV(m_commandBuffer, firstViewport, viewportCount, reinterpret_cast<const VkViewportWScalingNV *>(pViewportWScalings)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setViewportWScalingNV(uint32_t firstViewport, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ViewportWScalingNV> const &viewportWScalings, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetViewportWScalingNV( |
| m_commandBuffer, firstViewport, viewportWScalings.size(), reinterpret_cast<const VkViewportWScalingNV *>(viewportWScalings.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_direct_mode_display === |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Result PhysicalDevice::releaseDisplayEXT(VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkReleaseDisplayEXT(m_physicalDevice, static_cast<VkDisplayKHR>(display))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::releaseDisplayEXT(VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkReleaseDisplayEXT(m_physicalDevice, static_cast<VkDisplayKHR>(display)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_XLIB_XRANDR_EXT) |
| //=== VK_EXT_acquire_xlib_display === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::acquireXlibDisplayEXT(Display *dpy, |
| VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAcquireXlibDisplayEXT(m_physicalDevice, dpy, static_cast<VkDisplayKHR>(display))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| PhysicalDevice::acquireXlibDisplayEXT(Display &dpy, VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkAcquireXlibDisplayEXT(m_physicalDevice, &dpy, static_cast<VkDisplayKHR>(display)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::acquireXlibDisplayEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getRandROutputDisplayEXT(Display *dpy, |
| RROutput rrOutput, |
| VULKAN_HPP_NAMESPACE::DisplayKHR *pDisplay, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetRandROutputDisplayEXT(m_physicalDevice, dpy, rrOutput, reinterpret_cast<VkDisplayKHR *>(pDisplay))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DisplayKHR>::type |
| PhysicalDevice::getRandROutputDisplayEXT(Display &dpy, RROutput rrOutput, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DisplayKHR display; |
| VkResult result = d.vkGetRandROutputDisplayEXT(m_physicalDevice, &dpy, rrOutput, reinterpret_cast<VkDisplayKHR *>(&display)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getRandROutputDisplayEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), display); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DisplayKHR, Dispatch>>::type |
| PhysicalDevice::getRandROutputDisplayEXTUnique(Display &dpy, RROutput rrOutput, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DisplayKHR display; |
| Result result = static_cast<Result>(d.vkGetRandROutputDisplayEXT(m_physicalDevice, &dpy, rrOutput, reinterpret_cast<VkDisplayKHR *>(&display))); |
| ObjectRelease<PhysicalDevice, Dispatch> deleter(*this, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DisplayKHR, Dispatch>( |
| result, display, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getRandROutputDisplayEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_XLIB_XRANDR_EXT*/ |
| |
| //=== VK_EXT_display_surface_counter === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getSurfaceCapabilities2EXT(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, |
| VULKAN_HPP_NAMESPACE::SurfaceCapabilities2EXT *pSurfaceCapabilities, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceSurfaceCapabilities2EXT( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<VkSurfaceCapabilities2EXT *>(pSurfaceCapabilities))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceCapabilities2EXT>::type |
| PhysicalDevice::getSurfaceCapabilities2EXT(VULKAN_HPP_NAMESPACE::SurfaceKHR surface, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceCapabilities2EXT surfaceCapabilities; |
| VkResult result = d.vkGetPhysicalDeviceSurfaceCapabilities2EXT( |
| m_physicalDevice, static_cast<VkSurfaceKHR>(surface), reinterpret_cast<VkSurfaceCapabilities2EXT *>(&surfaceCapabilities)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfaceCapabilities2EXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surfaceCapabilities); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_display_control === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::displayPowerControlEXT(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| const VULKAN_HPP_NAMESPACE::DisplayPowerInfoEXT *pDisplayPowerInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkDisplayPowerControlEXT(m_device, static_cast<VkDisplayKHR>(display), reinterpret_cast<const VkDisplayPowerInfoEXT *>(pDisplayPowerInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::displayPowerControlEXT(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| const VULKAN_HPP_NAMESPACE::DisplayPowerInfoEXT &displayPowerInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = |
| d.vkDisplayPowerControlEXT(m_device, static_cast<VkDisplayKHR>(display), reinterpret_cast<const VkDisplayPowerInfoEXT *>(&displayPowerInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::displayPowerControlEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::registerEventEXT(const VULKAN_HPP_NAMESPACE::DeviceEventInfoEXT *pDeviceEventInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Fence *pFence, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkRegisterDeviceEventEXT(m_device, |
| reinterpret_cast<const VkDeviceEventInfoEXT *>(pDeviceEventInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkFence *>(pFence))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Fence>::type |
| Device::registerEventEXT(const VULKAN_HPP_NAMESPACE::DeviceEventInfoEXT &deviceEventInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Fence fence; |
| VkResult result = d.vkRegisterDeviceEventEXT( |
| m_device, |
| reinterpret_cast<const VkDeviceEventInfoEXT *>(&deviceEventInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkFence *>(&fence)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::registerEventEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), fence); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Fence, Dispatch>>::type |
| Device::registerEventEXTUnique(const VULKAN_HPP_NAMESPACE::DeviceEventInfoEXT &deviceEventInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Fence fence; |
| Result result = static_cast<Result>(d.vkRegisterDeviceEventEXT( |
| m_device, |
| reinterpret_cast<const VkDeviceEventInfoEXT *>(&deviceEventInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkFence *>(&fence))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Fence, Dispatch>(result, fence, VULKAN_HPP_NAMESPACE_STRING "::Device::registerEventEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::registerDisplayEventEXT(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| const VULKAN_HPP_NAMESPACE::DisplayEventInfoEXT *pDisplayEventInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Fence *pFence, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkRegisterDisplayEventEXT(m_device, |
| static_cast<VkDisplayKHR>(display), |
| reinterpret_cast<const VkDisplayEventInfoEXT *>(pDisplayEventInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkFence *>(pFence))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::Fence>::type |
| Device::registerDisplayEventEXT(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| const VULKAN_HPP_NAMESPACE::DisplayEventInfoEXT &displayEventInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Fence fence; |
| VkResult result = d.vkRegisterDisplayEventEXT( |
| m_device, |
| static_cast<VkDisplayKHR>(display), |
| reinterpret_cast<const VkDisplayEventInfoEXT *>(&displayEventInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkFence *>(&fence)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::registerDisplayEventEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), fence); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::Fence, Dispatch>>::type |
| Device::registerDisplayEventEXTUnique(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| const VULKAN_HPP_NAMESPACE::DisplayEventInfoEXT &displayEventInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::Fence fence; |
| Result result = static_cast<Result>(d.vkRegisterDisplayEventEXT( |
| m_device, |
| static_cast<VkDisplayKHR>(display), |
| reinterpret_cast<const VkDisplayEventInfoEXT *>(&displayEventInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkFence *>(&fence))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::Fence, Dispatch>( |
| result, fence, VULKAN_HPP_NAMESPACE_STRING "::Device::registerDisplayEventEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getSwapchainCounterEXT(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| VULKAN_HPP_NAMESPACE::SurfaceCounterFlagBitsEXT counter, |
| uint64_t *pCounterValue, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetSwapchainCounterEXT(m_device, static_cast<VkSwapchainKHR>(swapchain), static_cast<VkSurfaceCounterFlagBitsEXT>(counter), pCounterValue)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<uint64_t>::type Device::getSwapchainCounterEXT( |
| VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, VULKAN_HPP_NAMESPACE::SurfaceCounterFlagBitsEXT counter, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint64_t counterValue; |
| VkResult result = |
| d.vkGetSwapchainCounterEXT(m_device, static_cast<VkSwapchainKHR>(swapchain), static_cast<VkSurfaceCounterFlagBitsEXT>(counter), &counterValue); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getSwapchainCounterEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), counterValue); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_GOOGLE_display_timing === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getRefreshCycleDurationGOOGLE(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| VULKAN_HPP_NAMESPACE::RefreshCycleDurationGOOGLE *pDisplayTimingProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetRefreshCycleDurationGOOGLE( |
| m_device, static_cast<VkSwapchainKHR>(swapchain), reinterpret_cast<VkRefreshCycleDurationGOOGLE *>(pDisplayTimingProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::RefreshCycleDurationGOOGLE>::type |
| Device::getRefreshCycleDurationGOOGLE(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::RefreshCycleDurationGOOGLE displayTimingProperties; |
| VkResult result = d.vkGetRefreshCycleDurationGOOGLE( |
| m_device, static_cast<VkSwapchainKHR>(swapchain), reinterpret_cast<VkRefreshCycleDurationGOOGLE *>(&displayTimingProperties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getRefreshCycleDurationGOOGLE"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), displayTimingProperties); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getPastPresentationTimingGOOGLE(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| uint32_t *pPresentationTimingCount, |
| VULKAN_HPP_NAMESPACE::PastPresentationTimingGOOGLE *pPresentationTimings, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPastPresentationTimingGOOGLE(m_device, |
| static_cast<VkSwapchainKHR>(swapchain), |
| pPresentationTimingCount, |
| reinterpret_cast<VkPastPresentationTimingGOOGLE *>(pPresentationTimings))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PastPresentationTimingGOOGLEAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PastPresentationTimingGOOGLE, PastPresentationTimingGOOGLEAllocator>>::type |
| Device::getPastPresentationTimingGOOGLE(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PastPresentationTimingGOOGLE, PastPresentationTimingGOOGLEAllocator> presentationTimings; |
| uint32_t presentationTimingCount; |
| Result result; |
| do |
| { |
| result = |
| static_cast<Result>(d.vkGetPastPresentationTimingGOOGLE(m_device, static_cast<VkSwapchainKHR>(swapchain), &presentationTimingCount, nullptr)); |
| if((result == Result::eSuccess) && presentationTimingCount) |
| { |
| presentationTimings.resize(presentationTimingCount); |
| result = static_cast<Result>(d.vkGetPastPresentationTimingGOOGLE(m_device, |
| static_cast<VkSwapchainKHR>(swapchain), |
| &presentationTimingCount, |
| reinterpret_cast<VkPastPresentationTimingGOOGLE *>(presentationTimings.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(presentationTimingCount <= presentationTimings.size()); |
| if(presentationTimingCount < presentationTimings.size()) |
| { |
| presentationTimings.resize(presentationTimingCount); |
| } |
| } |
| return createResultValue(result, presentationTimings, VULKAN_HPP_NAMESPACE_STRING "::Device::getPastPresentationTimingGOOGLE"); |
| } |
| |
| template<typename PastPresentationTimingGOOGLEAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PastPresentationTimingGOOGLE>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PastPresentationTimingGOOGLE, PastPresentationTimingGOOGLEAllocator>>::type |
| Device::getPastPresentationTimingGOOGLE(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| PastPresentationTimingGOOGLEAllocator &pastPresentationTimingGOOGLEAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PastPresentationTimingGOOGLE, PastPresentationTimingGOOGLEAllocator> presentationTimings(pastPresentationTimingGOOGLEAllocator); |
| uint32_t presentationTimingCount; |
| Result result; |
| do |
| { |
| result = |
| static_cast<Result>(d.vkGetPastPresentationTimingGOOGLE(m_device, static_cast<VkSwapchainKHR>(swapchain), &presentationTimingCount, nullptr)); |
| if((result == Result::eSuccess) && presentationTimingCount) |
| { |
| presentationTimings.resize(presentationTimingCount); |
| result = static_cast<Result>(d.vkGetPastPresentationTimingGOOGLE(m_device, |
| static_cast<VkSwapchainKHR>(swapchain), |
| &presentationTimingCount, |
| reinterpret_cast<VkPastPresentationTimingGOOGLE *>(presentationTimings.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(presentationTimingCount <= presentationTimings.size()); |
| if(presentationTimingCount < presentationTimings.size()) |
| { |
| presentationTimings.resize(presentationTimingCount); |
| } |
| } |
| return createResultValue(result, presentationTimings, VULKAN_HPP_NAMESPACE_STRING "::Device::getPastPresentationTimingGOOGLE"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_discard_rectangles === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDiscardRectangleEXT(uint32_t firstDiscardRectangle, |
| uint32_t discardRectangleCount, |
| const VULKAN_HPP_NAMESPACE::Rect2D *pDiscardRectangles, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDiscardRectangleEXT(m_commandBuffer, firstDiscardRectangle, discardRectangleCount, reinterpret_cast<const VkRect2D *>(pDiscardRectangles)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDiscardRectangleEXT(uint32_t firstDiscardRectangle, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D> const &discardRectangles, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetDiscardRectangleEXT( |
| m_commandBuffer, firstDiscardRectangle, discardRectangles.size(), reinterpret_cast<const VkRect2D *>(discardRectangles.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_hdr_metadata === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::setHdrMetadataEXT(uint32_t swapchainCount, |
| const VULKAN_HPP_NAMESPACE::SwapchainKHR *pSwapchains, |
| const VULKAN_HPP_NAMESPACE::HdrMetadataEXT *pMetadata, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkSetHdrMetadataEXT( |
| m_device, swapchainCount, reinterpret_cast<const VkSwapchainKHR *>(pSwapchains), reinterpret_cast<const VkHdrMetadataEXT *>(pMetadata)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::setHdrMetadataEXT(ArrayProxy<const VULKAN_HPP_NAMESPACE::SwapchainKHR> const &swapchains, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::HdrMetadataEXT> const &metadata, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(swapchains.size() == metadata.size()); |
| # else |
| if(swapchains.size() != metadata.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::Device::setHdrMetadataEXT: swapchains.size() != metadata.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkSetHdrMetadataEXT(m_device, |
| swapchains.size(), |
| reinterpret_cast<const VkSwapchainKHR *>(swapchains.data()), |
| reinterpret_cast<const VkHdrMetadataEXT *>(metadata.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_create_renderpass2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createRenderPass2KHR(const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::RenderPass *pRenderPass, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateRenderPass2KHR(m_device, |
| reinterpret_cast<const VkRenderPassCreateInfo2 *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkRenderPass *>(pRenderPass))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::RenderPass>::type |
| Device::createRenderPass2KHR(const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::RenderPass renderPass; |
| VkResult result = |
| d.vkCreateRenderPass2KHR(m_device, |
| reinterpret_cast<const VkRenderPassCreateInfo2 *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkRenderPass *>(&renderPass)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createRenderPass2KHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), renderPass); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::RenderPass, Dispatch>>::type |
| Device::createRenderPass2KHRUnique(const VULKAN_HPP_NAMESPACE::RenderPassCreateInfo2 &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::RenderPass renderPass; |
| Result result = static_cast<Result>( |
| d.vkCreateRenderPass2KHR(m_device, |
| reinterpret_cast<const VkRenderPassCreateInfo2 *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkRenderPass *>(&renderPass))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::RenderPass, Dispatch>( |
| result, renderPass, VULKAN_HPP_NAMESPACE_STRING "::Device::createRenderPass2KHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass2KHR(const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo *pRenderPassBegin, |
| const VULKAN_HPP_NAMESPACE::SubpassBeginInfo *pSubpassBeginInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginRenderPass2KHR( |
| m_commandBuffer, reinterpret_cast<const VkRenderPassBeginInfo *>(pRenderPassBegin), reinterpret_cast<const VkSubpassBeginInfo *>(pSubpassBeginInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginRenderPass2KHR(const VULKAN_HPP_NAMESPACE::RenderPassBeginInfo &renderPassBegin, |
| const VULKAN_HPP_NAMESPACE::SubpassBeginInfo &subpassBeginInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBeginRenderPass2KHR( |
| m_commandBuffer, reinterpret_cast<const VkRenderPassBeginInfo *>(&renderPassBegin), reinterpret_cast<const VkSubpassBeginInfo *>(&subpassBeginInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::nextSubpass2KHR(const VULKAN_HPP_NAMESPACE::SubpassBeginInfo *pSubpassBeginInfo, |
| const VULKAN_HPP_NAMESPACE::SubpassEndInfo *pSubpassEndInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdNextSubpass2KHR( |
| m_commandBuffer, reinterpret_cast<const VkSubpassBeginInfo *>(pSubpassBeginInfo), reinterpret_cast<const VkSubpassEndInfo *>(pSubpassEndInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::nextSubpass2KHR(const VULKAN_HPP_NAMESPACE::SubpassBeginInfo &subpassBeginInfo, |
| const VULKAN_HPP_NAMESPACE::SubpassEndInfo &subpassEndInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdNextSubpass2KHR( |
| m_commandBuffer, reinterpret_cast<const VkSubpassBeginInfo *>(&subpassBeginInfo), reinterpret_cast<const VkSubpassEndInfo *>(&subpassEndInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endRenderPass2KHR(const VULKAN_HPP_NAMESPACE::SubpassEndInfo *pSubpassEndInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndRenderPass2KHR(m_commandBuffer, reinterpret_cast<const VkSubpassEndInfo *>(pSubpassEndInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endRenderPass2KHR(const VULKAN_HPP_NAMESPACE::SubpassEndInfo &subpassEndInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdEndRenderPass2KHR(m_commandBuffer, reinterpret_cast<const VkSubpassEndInfo *>(&subpassEndInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_shared_presentable_image === |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getSwapchainStatusKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetSwapchainStatusKHR(m_device, static_cast<VkSwapchainKHR>(swapchain))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result Device::getSwapchainStatusKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkGetSwapchainStatusKHR(m_device, static_cast<VkSwapchainKHR>(swapchain)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::getSwapchainStatusKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_external_fence_capabilities === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getExternalFencePropertiesKHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalFenceInfo *pExternalFenceInfo, |
| VULKAN_HPP_NAMESPACE::ExternalFenceProperties *pExternalFenceProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceExternalFencePropertiesKHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalFenceInfo *>(pExternalFenceInfo), |
| reinterpret_cast<VkExternalFenceProperties *>(pExternalFenceProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::ExternalFenceProperties |
| PhysicalDevice::getExternalFencePropertiesKHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceExternalFenceInfo &externalFenceInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ExternalFenceProperties externalFenceProperties; |
| d.vkGetPhysicalDeviceExternalFencePropertiesKHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceExternalFenceInfo *>(&externalFenceInfo), |
| reinterpret_cast<VkExternalFenceProperties *>(&externalFenceProperties)); |
| |
| return externalFenceProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_WIN32_KHR) |
| //=== VK_KHR_external_fence_win32 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::importFenceWin32HandleKHR( |
| const VULKAN_HPP_NAMESPACE::ImportFenceWin32HandleInfoKHR *pImportFenceWin32HandleInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkImportFenceWin32HandleKHR(m_device, reinterpret_cast<const VkImportFenceWin32HandleInfoKHR *>(pImportFenceWin32HandleInfo))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::importFenceWin32HandleKHR(const VULKAN_HPP_NAMESPACE::ImportFenceWin32HandleInfoKHR &importFenceWin32HandleInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkImportFenceWin32HandleKHR(m_device, reinterpret_cast<const VkImportFenceWin32HandleInfoKHR *>(&importFenceWin32HandleInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::importFenceWin32HandleKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getFenceWin32HandleKHR(const VULKAN_HPP_NAMESPACE::FenceGetWin32HandleInfoKHR *pGetWin32HandleInfo, |
| HANDLE *pHandle, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetFenceWin32HandleKHR(m_device, reinterpret_cast<const VkFenceGetWin32HandleInfoKHR *>(pGetWin32HandleInfo), pHandle)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<HANDLE>::type |
| Device::getFenceWin32HandleKHR(const VULKAN_HPP_NAMESPACE::FenceGetWin32HandleInfoKHR &getWin32HandleInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| HANDLE handle; |
| VkResult result = d.vkGetFenceWin32HandleKHR(m_device, reinterpret_cast<const VkFenceGetWin32HandleInfoKHR *>(&getWin32HandleInfo), &handle); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getFenceWin32HandleKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), handle); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_WIN32_KHR*/ |
| |
| //=== VK_KHR_external_fence_fd === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::importFenceFdKHR(const VULKAN_HPP_NAMESPACE::ImportFenceFdInfoKHR *pImportFenceFdInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkImportFenceFdKHR(m_device, reinterpret_cast<const VkImportFenceFdInfoKHR *>(pImportFenceFdInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::importFenceFdKHR(const VULKAN_HPP_NAMESPACE::ImportFenceFdInfoKHR &importFenceFdInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkImportFenceFdKHR(m_device, reinterpret_cast<const VkImportFenceFdInfoKHR *>(&importFenceFdInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::importFenceFdKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getFenceFdKHR(const VULKAN_HPP_NAMESPACE::FenceGetFdInfoKHR *pGetFdInfo, |
| int *pFd, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetFenceFdKHR(m_device, reinterpret_cast<const VkFenceGetFdInfoKHR *>(pGetFdInfo), pFd)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<int>::type Device::getFenceFdKHR(const VULKAN_HPP_NAMESPACE::FenceGetFdInfoKHR &getFdInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| int fd; |
| VkResult result = d.vkGetFenceFdKHR(m_device, reinterpret_cast<const VkFenceGetFdInfoKHR *>(&getFdInfo), &fd); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getFenceFdKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), fd); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_performance_query === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::enumerateQueueFamilyPerformanceQueryCountersKHR(uint32_t queueFamilyIndex, |
| uint32_t *pCounterCount, |
| VULKAN_HPP_NAMESPACE::PerformanceCounterKHR *pCounters, |
| VULKAN_HPP_NAMESPACE::PerformanceCounterDescriptionKHR *pCounterDescriptions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR(m_physicalDevice, |
| queueFamilyIndex, |
| pCounterCount, |
| reinterpret_cast<VkPerformanceCounterKHR *>(pCounters), |
| reinterpret_cast<VkPerformanceCounterDescriptionKHR *>(pCounterDescriptions))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PerformanceCounterKHRAllocator, typename PerformanceCounterDescriptionKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE |
| typename ResultValueType<std::pair<std::vector<PerformanceCounterKHR, PerformanceCounterKHRAllocator>, |
| std::vector<PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator>>>::type |
| PhysicalDevice::enumerateQueueFamilyPerformanceQueryCountersKHR(uint32_t queueFamilyIndex, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::pair<std::vector<PerformanceCounterKHR, PerformanceCounterKHRAllocator>, |
| std::vector<PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator>> |
| data; |
| std::vector<PerformanceCounterKHR, PerformanceCounterKHRAllocator> &counters = data.first; |
| std::vector<PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator> &counterDescriptions = data.second; |
| uint32_t counterCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>( |
| d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR(m_physicalDevice, queueFamilyIndex, &counterCount, nullptr, nullptr)); |
| if((result == Result::eSuccess) && counterCount) |
| { |
| counters.resize(counterCount); |
| counterDescriptions.resize(counterCount); |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR( |
| m_physicalDevice, |
| queueFamilyIndex, |
| &counterCount, |
| reinterpret_cast<VkPerformanceCounterKHR *>(counters.data()), |
| reinterpret_cast<VkPerformanceCounterDescriptionKHR *>(counterDescriptions.data()))); |
| VULKAN_HPP_ASSERT(counterCount <= counters.size()); |
| } |
| } while(result == Result::eIncomplete); |
| if((result == Result::eSuccess) && (counterCount < counters.size())) |
| { |
| counters.resize(counterCount); |
| counterDescriptions.resize(counterCount); |
| } |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::enumerateQueueFamilyPerformanceQueryCountersKHR"); |
| } |
| |
| template<typename PerformanceCounterKHRAllocator, |
| typename PerformanceCounterDescriptionKHRAllocator, |
| typename Dispatch, |
| typename B1, |
| typename B2, |
| typename std::enable_if<std::is_same<typename B1::value_type, PerformanceCounterKHR>::value && |
| std::is_same<typename B2::value_type, PerformanceCounterDescriptionKHR>::value, |
| int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE |
| typename ResultValueType<std::pair<std::vector<PerformanceCounterKHR, PerformanceCounterKHRAllocator>, |
| std::vector<PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator>>>::type |
| PhysicalDevice::enumerateQueueFamilyPerformanceQueryCountersKHR(uint32_t queueFamilyIndex, |
| PerformanceCounterKHRAllocator &performanceCounterKHRAllocator, |
| PerformanceCounterDescriptionKHRAllocator &performanceCounterDescriptionKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::pair<std::vector<PerformanceCounterKHR, PerformanceCounterKHRAllocator>, |
| std::vector<PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator>> |
| data( |
| std::piecewise_construct, std::forward_as_tuple(performanceCounterKHRAllocator), std::forward_as_tuple(performanceCounterDescriptionKHRAllocator)); |
| std::vector<PerformanceCounterKHR, PerformanceCounterKHRAllocator> &counters = data.first; |
| std::vector<PerformanceCounterDescriptionKHR, PerformanceCounterDescriptionKHRAllocator> &counterDescriptions = data.second; |
| uint32_t counterCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>( |
| d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR(m_physicalDevice, queueFamilyIndex, &counterCount, nullptr, nullptr)); |
| if((result == Result::eSuccess) && counterCount) |
| { |
| counters.resize(counterCount); |
| counterDescriptions.resize(counterCount); |
| result = static_cast<Result>(d.vkEnumeratePhysicalDeviceQueueFamilyPerformanceQueryCountersKHR( |
| m_physicalDevice, |
| queueFamilyIndex, |
| &counterCount, |
| reinterpret_cast<VkPerformanceCounterKHR *>(counters.data()), |
| reinterpret_cast<VkPerformanceCounterDescriptionKHR *>(counterDescriptions.data()))); |
| VULKAN_HPP_ASSERT(counterCount <= counters.size()); |
| } |
| } while(result == Result::eIncomplete); |
| if((result == Result::eSuccess) && (counterCount < counters.size())) |
| { |
| counters.resize(counterCount); |
| counterDescriptions.resize(counterCount); |
| } |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::enumerateQueueFamilyPerformanceQueryCountersKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| PhysicalDevice::getQueueFamilyPerformanceQueryPassesKHR(const VULKAN_HPP_NAMESPACE::QueryPoolPerformanceCreateInfoKHR *pPerformanceQueryCreateInfo, |
| uint32_t *pNumPasses, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceQueueFamilyPerformanceQueryPassesKHR( |
| m_physicalDevice, reinterpret_cast<const VkQueryPoolPerformanceCreateInfoKHR *>(pPerformanceQueryCreateInfo), pNumPasses); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE uint32_t PhysicalDevice::getQueueFamilyPerformanceQueryPassesKHR( |
| const VULKAN_HPP_NAMESPACE::QueryPoolPerformanceCreateInfoKHR &performanceQueryCreateInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint32_t numPasses; |
| d.vkGetPhysicalDeviceQueueFamilyPerformanceQueryPassesKHR( |
| m_physicalDevice, reinterpret_cast<const VkQueryPoolPerformanceCreateInfoKHR *>(&performanceQueryCreateInfo), &numPasses); |
| |
| return numPasses; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::acquireProfilingLockKHR(const VULKAN_HPP_NAMESPACE::AcquireProfilingLockInfoKHR *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAcquireProfilingLockKHR(m_device, reinterpret_cast<const VkAcquireProfilingLockInfoKHR *>(pInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::acquireProfilingLockKHR(const VULKAN_HPP_NAMESPACE::AcquireProfilingLockInfoKHR &info, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkAcquireProfilingLockKHR(m_device, reinterpret_cast<const VkAcquireProfilingLockInfoKHR *>(&info)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::acquireProfilingLockKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::releaseProfilingLockKHR(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkReleaseProfilingLockKHR(m_device); |
| } |
| |
| //=== VK_KHR_get_surface_capabilities2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getSurfaceCapabilities2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR *pSurfaceInfo, |
| VULKAN_HPP_NAMESPACE::SurfaceCapabilities2KHR *pSurfaceCapabilities, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceSurfaceCapabilities2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(pSurfaceInfo), |
| reinterpret_cast<VkSurfaceCapabilities2KHR *>(pSurfaceCapabilities))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceCapabilities2KHR>::type |
| PhysicalDevice::getSurfaceCapabilities2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR &surfaceInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceCapabilities2KHR surfaceCapabilities; |
| VkResult result = d.vkGetPhysicalDeviceSurfaceCapabilities2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), |
| reinterpret_cast<VkSurfaceCapabilities2KHR *>(&surfaceCapabilities)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfaceCapabilities2KHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surfaceCapabilities); |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<StructureChain<X, Y, Z...>>::type |
| PhysicalDevice::getSurfaceCapabilities2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR &surfaceInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::SurfaceCapabilities2KHR &surfaceCapabilities = structureChain.template get<VULKAN_HPP_NAMESPACE::SurfaceCapabilities2KHR>(); |
| Result result = |
| static_cast<Result>(d.vkGetPhysicalDeviceSurfaceCapabilities2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), |
| reinterpret_cast<VkSurfaceCapabilities2KHR *>(&surfaceCapabilities))); |
| return createResultValue(result, structureChain, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfaceCapabilities2KHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getSurfaceFormats2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR *pSurfaceInfo, |
| uint32_t *pSurfaceFormatCount, |
| VULKAN_HPP_NAMESPACE::SurfaceFormat2KHR *pSurfaceFormats, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormats2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(pSurfaceInfo), |
| pSurfaceFormatCount, |
| reinterpret_cast<VkSurfaceFormat2KHR *>(pSurfaceFormats))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SurfaceFormat2KHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<SurfaceFormat2KHR, SurfaceFormat2KHRAllocator>>::type |
| PhysicalDevice::getSurfaceFormats2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR &surfaceInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SurfaceFormat2KHR, SurfaceFormat2KHRAllocator> surfaceFormats; |
| uint32_t surfaceFormatCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormats2KHR( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), &surfaceFormatCount, nullptr)); |
| if((result == Result::eSuccess) && surfaceFormatCount) |
| { |
| surfaceFormats.resize(surfaceFormatCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormats2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), |
| &surfaceFormatCount, |
| reinterpret_cast<VkSurfaceFormat2KHR *>(surfaceFormats.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(surfaceFormatCount <= surfaceFormats.size()); |
| if(surfaceFormatCount < surfaceFormats.size()) |
| { |
| surfaceFormats.resize(surfaceFormatCount); |
| } |
| } |
| return createResultValue(result, surfaceFormats, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfaceFormats2KHR"); |
| } |
| |
| template<typename SurfaceFormat2KHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SurfaceFormat2KHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<SurfaceFormat2KHR, SurfaceFormat2KHRAllocator>>::type |
| PhysicalDevice::getSurfaceFormats2KHR(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR &surfaceInfo, |
| SurfaceFormat2KHRAllocator &surfaceFormat2KHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SurfaceFormat2KHR, SurfaceFormat2KHRAllocator> surfaceFormats(surfaceFormat2KHRAllocator); |
| uint32_t surfaceFormatCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormats2KHR( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), &surfaceFormatCount, nullptr)); |
| if((result == Result::eSuccess) && surfaceFormatCount) |
| { |
| surfaceFormats.resize(surfaceFormatCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfaceFormats2KHR(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), |
| &surfaceFormatCount, |
| reinterpret_cast<VkSurfaceFormat2KHR *>(surfaceFormats.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(surfaceFormatCount <= surfaceFormats.size()); |
| if(surfaceFormatCount < surfaceFormats.size()) |
| { |
| surfaceFormats.resize(surfaceFormatCount); |
| } |
| } |
| return createResultValue(result, surfaceFormats, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfaceFormats2KHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_get_display_properties2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getDisplayProperties2KHR(uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::DisplayProperties2KHR *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPhysicalDeviceDisplayProperties2KHR(m_physicalDevice, pPropertyCount, reinterpret_cast<VkDisplayProperties2KHR *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DisplayProperties2KHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayProperties2KHR, DisplayProperties2KHRAllocator>>::type |
| PhysicalDevice::getDisplayProperties2KHR(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayProperties2KHR, DisplayProperties2KHRAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayProperties2KHR(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceDisplayProperties2KHR(m_physicalDevice, &propertyCount, reinterpret_cast<VkDisplayProperties2KHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayProperties2KHR"); |
| } |
| |
| template<typename DisplayProperties2KHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, DisplayProperties2KHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayProperties2KHR, DisplayProperties2KHRAllocator>>::type |
| PhysicalDevice::getDisplayProperties2KHR(DisplayProperties2KHRAllocator &displayProperties2KHRAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayProperties2KHR, DisplayProperties2KHRAllocator> properties(displayProperties2KHRAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayProperties2KHR(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceDisplayProperties2KHR(m_physicalDevice, &propertyCount, reinterpret_cast<VkDisplayProperties2KHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayProperties2KHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getDisplayPlaneProperties2KHR(uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::DisplayPlaneProperties2KHR *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPhysicalDeviceDisplayPlaneProperties2KHR(m_physicalDevice, pPropertyCount, reinterpret_cast<VkDisplayPlaneProperties2KHR *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DisplayPlaneProperties2KHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayPlaneProperties2KHR, DisplayPlaneProperties2KHRAllocator>>::type |
| PhysicalDevice::getDisplayPlaneProperties2KHR(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayPlaneProperties2KHR, DisplayPlaneProperties2KHRAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPlaneProperties2KHR(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPlaneProperties2KHR( |
| m_physicalDevice, &propertyCount, reinterpret_cast<VkDisplayPlaneProperties2KHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPlaneProperties2KHR"); |
| } |
| |
| template<typename DisplayPlaneProperties2KHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, DisplayPlaneProperties2KHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayPlaneProperties2KHR, DisplayPlaneProperties2KHRAllocator>>::type |
| PhysicalDevice::getDisplayPlaneProperties2KHR(DisplayPlaneProperties2KHRAllocator &displayPlaneProperties2KHRAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayPlaneProperties2KHR, DisplayPlaneProperties2KHRAllocator> properties(displayPlaneProperties2KHRAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPlaneProperties2KHR(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceDisplayPlaneProperties2KHR( |
| m_physicalDevice, &propertyCount, reinterpret_cast<VkDisplayPlaneProperties2KHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPlaneProperties2KHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getDisplayModeProperties2KHR(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| uint32_t *pPropertyCount, |
| VULKAN_HPP_NAMESPACE::DisplayModeProperties2KHR *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetDisplayModeProperties2KHR( |
| m_physicalDevice, static_cast<VkDisplayKHR>(display), pPropertyCount, reinterpret_cast<VkDisplayModeProperties2KHR *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DisplayModeProperties2KHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayModeProperties2KHR, DisplayModeProperties2KHRAllocator>>::type |
| PhysicalDevice::getDisplayModeProperties2KHR(VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayModeProperties2KHR, DisplayModeProperties2KHRAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetDisplayModeProperties2KHR(m_physicalDevice, static_cast<VkDisplayKHR>(display), &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetDisplayModeProperties2KHR( |
| m_physicalDevice, static_cast<VkDisplayKHR>(display), &propertyCount, reinterpret_cast<VkDisplayModeProperties2KHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayModeProperties2KHR"); |
| } |
| |
| template<typename DisplayModeProperties2KHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, DisplayModeProperties2KHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DisplayModeProperties2KHR, DisplayModeProperties2KHRAllocator>>::type |
| PhysicalDevice::getDisplayModeProperties2KHR(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| DisplayModeProperties2KHRAllocator &displayModeProperties2KHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<DisplayModeProperties2KHR, DisplayModeProperties2KHRAllocator> properties(displayModeProperties2KHRAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetDisplayModeProperties2KHR(m_physicalDevice, static_cast<VkDisplayKHR>(display), &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetDisplayModeProperties2KHR( |
| m_physicalDevice, static_cast<VkDisplayKHR>(display), &propertyCount, reinterpret_cast<VkDisplayModeProperties2KHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayModeProperties2KHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getDisplayPlaneCapabilities2KHR(const VULKAN_HPP_NAMESPACE::DisplayPlaneInfo2KHR *pDisplayPlaneInfo, |
| VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilities2KHR *pCapabilities, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetDisplayPlaneCapabilities2KHR(m_physicalDevice, |
| reinterpret_cast<const VkDisplayPlaneInfo2KHR *>(pDisplayPlaneInfo), |
| reinterpret_cast<VkDisplayPlaneCapabilities2KHR *>(pCapabilities))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilities2KHR>::type |
| PhysicalDevice::getDisplayPlaneCapabilities2KHR(const VULKAN_HPP_NAMESPACE::DisplayPlaneInfo2KHR &displayPlaneInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DisplayPlaneCapabilities2KHR capabilities; |
| VkResult result = d.vkGetDisplayPlaneCapabilities2KHR(m_physicalDevice, |
| reinterpret_cast<const VkDisplayPlaneInfo2KHR *>(&displayPlaneInfo), |
| reinterpret_cast<VkDisplayPlaneCapabilities2KHR *>(&capabilities)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDisplayPlaneCapabilities2KHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), capabilities); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_IOS_MVK) |
| //=== VK_MVK_ios_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createIOSSurfaceMVK(const VULKAN_HPP_NAMESPACE::IOSSurfaceCreateInfoMVK *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateIOSSurfaceMVK(m_instance, |
| reinterpret_cast<const VkIOSSurfaceCreateInfoMVK *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createIOSSurfaceMVK(const VULKAN_HPP_NAMESPACE::IOSSurfaceCreateInfoMVK &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = |
| d.vkCreateIOSSurfaceMVK(m_instance, |
| reinterpret_cast<const VkIOSSurfaceCreateInfoMVK *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createIOSSurfaceMVK"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createIOSSurfaceMVKUnique(const VULKAN_HPP_NAMESPACE::IOSSurfaceCreateInfoMVK &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>( |
| d.vkCreateIOSSurfaceMVK(m_instance, |
| reinterpret_cast<const VkIOSSurfaceCreateInfoMVK *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createIOSSurfaceMVKUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_IOS_MVK*/ |
| |
| #if defined(VK_USE_PLATFORM_MACOS_MVK) |
| //=== VK_MVK_macos_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createMacOSSurfaceMVK(const VULKAN_HPP_NAMESPACE::MacOSSurfaceCreateInfoMVK *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateMacOSSurfaceMVK(m_instance, |
| reinterpret_cast<const VkMacOSSurfaceCreateInfoMVK *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createMacOSSurfaceMVK(const VULKAN_HPP_NAMESPACE::MacOSSurfaceCreateInfoMVK &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = |
| d.vkCreateMacOSSurfaceMVK(m_instance, |
| reinterpret_cast<const VkMacOSSurfaceCreateInfoMVK *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createMacOSSurfaceMVK"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createMacOSSurfaceMVKUnique(const VULKAN_HPP_NAMESPACE::MacOSSurfaceCreateInfoMVK &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>( |
| d.vkCreateMacOSSurfaceMVK(m_instance, |
| reinterpret_cast<const VkMacOSSurfaceCreateInfoMVK *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createMacOSSurfaceMVKUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_MACOS_MVK*/ |
| |
| //=== VK_EXT_debug_utils === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::setDebugUtilsObjectNameEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsObjectNameInfoEXT *pNameInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkSetDebugUtilsObjectNameEXT(m_device, reinterpret_cast<const VkDebugUtilsObjectNameInfoEXT *>(pNameInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::setDebugUtilsObjectNameEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsObjectNameInfoEXT &nameInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkSetDebugUtilsObjectNameEXT(m_device, reinterpret_cast<const VkDebugUtilsObjectNameInfoEXT *>(&nameInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::setDebugUtilsObjectNameEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::setDebugUtilsObjectTagEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsObjectTagInfoEXT *pTagInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkSetDebugUtilsObjectTagEXT(m_device, reinterpret_cast<const VkDebugUtilsObjectTagInfoEXT *>(pTagInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::setDebugUtilsObjectTagEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsObjectTagInfoEXT &tagInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkSetDebugUtilsObjectTagEXT(m_device, reinterpret_cast<const VkDebugUtilsObjectTagInfoEXT *>(&tagInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::setDebugUtilsObjectTagEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Queue::beginDebugUtilsLabelEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT *pLabelInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkQueueBeginDebugUtilsLabelEXT(m_queue, reinterpret_cast<const VkDebugUtilsLabelEXT *>(pLabelInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Queue::beginDebugUtilsLabelEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT &labelInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkQueueBeginDebugUtilsLabelEXT(m_queue, reinterpret_cast<const VkDebugUtilsLabelEXT *>(&labelInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Queue::endDebugUtilsLabelEXT(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkQueueEndDebugUtilsLabelEXT(m_queue); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Queue::insertDebugUtilsLabelEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT *pLabelInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkQueueInsertDebugUtilsLabelEXT(m_queue, reinterpret_cast<const VkDebugUtilsLabelEXT *>(pLabelInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Queue::insertDebugUtilsLabelEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT &labelInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkQueueInsertDebugUtilsLabelEXT(m_queue, reinterpret_cast<const VkDebugUtilsLabelEXT *>(&labelInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginDebugUtilsLabelEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT *pLabelInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBeginDebugUtilsLabelEXT(m_commandBuffer, reinterpret_cast<const VkDebugUtilsLabelEXT *>(pLabelInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::beginDebugUtilsLabelEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT &labelInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBeginDebugUtilsLabelEXT(m_commandBuffer, reinterpret_cast<const VkDebugUtilsLabelEXT *>(&labelInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::endDebugUtilsLabelEXT(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEndDebugUtilsLabelEXT(m_commandBuffer); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::insertDebugUtilsLabelEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT *pLabelInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdInsertDebugUtilsLabelEXT(m_commandBuffer, reinterpret_cast<const VkDebugUtilsLabelEXT *>(pLabelInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::insertDebugUtilsLabelEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsLabelEXT &labelInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdInsertDebugUtilsLabelEXT(m_commandBuffer, reinterpret_cast<const VkDebugUtilsLabelEXT *>(&labelInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Instance::createDebugUtilsMessengerEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCreateInfoEXT *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT *pMessenger, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDebugUtilsMessengerEXT(m_instance, |
| reinterpret_cast<const VkDebugUtilsMessengerCreateInfoEXT *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkDebugUtilsMessengerEXT *>(pMessenger))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT>::type |
| Instance::createDebugUtilsMessengerEXT(const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger; |
| VkResult result = d.vkCreateDebugUtilsMessengerEXT( |
| m_instance, |
| reinterpret_cast<const VkDebugUtilsMessengerCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDebugUtilsMessengerEXT *>(&messenger)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createDebugUtilsMessengerEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), messenger); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT, Dispatch>>::type |
| Instance::createDebugUtilsMessengerEXTUnique(const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger; |
| Result result = static_cast<Result>(d.vkCreateDebugUtilsMessengerEXT( |
| m_instance, |
| reinterpret_cast<const VkDebugUtilsMessengerCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDebugUtilsMessengerEXT *>(&messenger))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT, Dispatch>( |
| result, messenger, VULKAN_HPP_NAMESPACE_STRING "::Instance::createDebugUtilsMessengerEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroyDebugUtilsMessengerEXT(VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDebugUtilsMessengerEXT( |
| m_instance, static_cast<VkDebugUtilsMessengerEXT>(messenger), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroyDebugUtilsMessengerEXT(VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDebugUtilsMessengerEXT( |
| m_instance, |
| static_cast<VkDebugUtilsMessengerEXT>(messenger), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroy(VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDebugUtilsMessengerEXT( |
| m_instance, static_cast<VkDebugUtilsMessengerEXT>(messenger), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::destroy(VULKAN_HPP_NAMESPACE::DebugUtilsMessengerEXT messenger, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDebugUtilsMessengerEXT( |
| m_instance, |
| static_cast<VkDebugUtilsMessengerEXT>(messenger), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::submitDebugUtilsMessageEXT(VULKAN_HPP_NAMESPACE::DebugUtilsMessageSeverityFlagBitsEXT messageSeverity, |
| VULKAN_HPP_NAMESPACE::DebugUtilsMessageTypeFlagsEXT messageTypes, |
| const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCallbackDataEXT *pCallbackData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkSubmitDebugUtilsMessageEXT(m_instance, |
| static_cast<VkDebugUtilsMessageSeverityFlagBitsEXT>(messageSeverity), |
| static_cast<VkDebugUtilsMessageTypeFlagsEXT>(messageTypes), |
| reinterpret_cast<const VkDebugUtilsMessengerCallbackDataEXT *>(pCallbackData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Instance::submitDebugUtilsMessageEXT(VULKAN_HPP_NAMESPACE::DebugUtilsMessageSeverityFlagBitsEXT messageSeverity, |
| VULKAN_HPP_NAMESPACE::DebugUtilsMessageTypeFlagsEXT messageTypes, |
| const VULKAN_HPP_NAMESPACE::DebugUtilsMessengerCallbackDataEXT &callbackData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkSubmitDebugUtilsMessageEXT(m_instance, |
| static_cast<VkDebugUtilsMessageSeverityFlagBitsEXT>(messageSeverity), |
| static_cast<VkDebugUtilsMessageTypeFlagsEXT>(messageTypes), |
| reinterpret_cast<const VkDebugUtilsMessengerCallbackDataEXT *>(&callbackData)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_ANDROID_KHR) |
| //=== VK_ANDROID_external_memory_android_hardware_buffer === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getAndroidHardwareBufferPropertiesANDROID(const struct AHardwareBuffer *buffer, |
| VULKAN_HPP_NAMESPACE::AndroidHardwareBufferPropertiesANDROID *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetAndroidHardwareBufferPropertiesANDROID(m_device, buffer, reinterpret_cast<VkAndroidHardwareBufferPropertiesANDROID *>(pProperties))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::AndroidHardwareBufferPropertiesANDROID>::type |
| Device::getAndroidHardwareBufferPropertiesANDROID(const struct AHardwareBuffer &buffer, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::AndroidHardwareBufferPropertiesANDROID properties; |
| VkResult result = |
| d.vkGetAndroidHardwareBufferPropertiesANDROID(m_device, &buffer, reinterpret_cast<VkAndroidHardwareBufferPropertiesANDROID *>(&properties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getAndroidHardwareBufferPropertiesANDROID"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), properties); |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<StructureChain<X, Y, Z...>>::type |
| Device::getAndroidHardwareBufferPropertiesANDROID(const struct AHardwareBuffer &buffer, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::AndroidHardwareBufferPropertiesANDROID &properties = |
| structureChain.template get<VULKAN_HPP_NAMESPACE::AndroidHardwareBufferPropertiesANDROID>(); |
| Result result = static_cast<Result>( |
| d.vkGetAndroidHardwareBufferPropertiesANDROID(m_device, &buffer, reinterpret_cast<VkAndroidHardwareBufferPropertiesANDROID *>(&properties))); |
| return createResultValue(result, structureChain, VULKAN_HPP_NAMESPACE_STRING "::Device::getAndroidHardwareBufferPropertiesANDROID"); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getMemoryAndroidHardwareBufferANDROID(const VULKAN_HPP_NAMESPACE::MemoryGetAndroidHardwareBufferInfoANDROID *pInfo, |
| struct AHardwareBuffer **pBuffer, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetMemoryAndroidHardwareBufferANDROID(m_device, reinterpret_cast<const VkMemoryGetAndroidHardwareBufferInfoANDROID *>(pInfo), pBuffer)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<struct AHardwareBuffer *>::type |
| Device::getMemoryAndroidHardwareBufferANDROID(const VULKAN_HPP_NAMESPACE::MemoryGetAndroidHardwareBufferInfoANDROID &info, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| struct AHardwareBuffer *buffer; |
| VkResult result = |
| d.vkGetMemoryAndroidHardwareBufferANDROID(m_device, reinterpret_cast<const VkMemoryGetAndroidHardwareBufferInfoANDROID *>(&info), &buffer); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryAndroidHardwareBufferANDROID"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), buffer); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_ANDROID_KHR*/ |
| |
| //=== VK_EXT_sample_locations === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setSampleLocationsEXT(const VULKAN_HPP_NAMESPACE::SampleLocationsInfoEXT *pSampleLocationsInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetSampleLocationsEXT(m_commandBuffer, reinterpret_cast<const VkSampleLocationsInfoEXT *>(pSampleLocationsInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setSampleLocationsEXT(const VULKAN_HPP_NAMESPACE::SampleLocationsInfoEXT &sampleLocationsInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetSampleLocationsEXT(m_commandBuffer, reinterpret_cast<const VkSampleLocationsInfoEXT *>(&sampleLocationsInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void PhysicalDevice::getMultisamplePropertiesEXT(VULKAN_HPP_NAMESPACE::SampleCountFlagBits samples, |
| VULKAN_HPP_NAMESPACE::MultisamplePropertiesEXT *pMultisampleProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPhysicalDeviceMultisamplePropertiesEXT( |
| m_physicalDevice, static_cast<VkSampleCountFlagBits>(samples), reinterpret_cast<VkMultisamplePropertiesEXT *>(pMultisampleProperties)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MultisamplePropertiesEXT |
| PhysicalDevice::getMultisamplePropertiesEXT(VULKAN_HPP_NAMESPACE::SampleCountFlagBits samples, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MultisamplePropertiesEXT multisampleProperties; |
| d.vkGetPhysicalDeviceMultisamplePropertiesEXT( |
| m_physicalDevice, static_cast<VkSampleCountFlagBits>(samples), reinterpret_cast<VkMultisamplePropertiesEXT *>(&multisampleProperties)); |
| |
| return multisampleProperties; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_get_memory_requirements2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageMemoryRequirements2KHR(const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetImageMemoryRequirements2KHR( |
| m_device, reinterpret_cast<const VkImageMemoryRequirementsInfo2 *>(pInfo), reinterpret_cast<VkMemoryRequirements2 *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 |
| Device::getImageMemoryRequirements2KHR(const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements; |
| d.vkGetImageMemoryRequirements2KHR( |
| m_device, reinterpret_cast<const VkImageMemoryRequirementsInfo2 *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getImageMemoryRequirements2KHR(const VULKAN_HPP_NAMESPACE::ImageMemoryRequirementsInfo2 &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>(); |
| d.vkGetImageMemoryRequirements2KHR( |
| m_device, reinterpret_cast<const VkImageMemoryRequirementsInfo2 *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getBufferMemoryRequirements2KHR(const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetBufferMemoryRequirements2KHR( |
| m_device, reinterpret_cast<const VkBufferMemoryRequirementsInfo2 *>(pInfo), reinterpret_cast<VkMemoryRequirements2 *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 |
| Device::getBufferMemoryRequirements2KHR(const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements; |
| d.vkGetBufferMemoryRequirements2KHR( |
| m_device, reinterpret_cast<const VkBufferMemoryRequirementsInfo2 *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getBufferMemoryRequirements2KHR(const VULKAN_HPP_NAMESPACE::BufferMemoryRequirementsInfo2 &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>(); |
| d.vkGetBufferMemoryRequirements2KHR( |
| m_device, reinterpret_cast<const VkBufferMemoryRequirementsInfo2 *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageSparseMemoryRequirements2KHR(const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 *pInfo, |
| uint32_t *pSparseMemoryRequirementCount, |
| VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2 *pSparseMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetImageSparseMemoryRequirements2KHR(m_device, |
| reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(pInfo), |
| pSparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(pSparseMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SparseImageMemoryRequirements2Allocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> |
| Device::getImageSparseMemoryRequirements2KHR(const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 &info, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements; |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetImageSparseMemoryRequirements2KHR( |
| m_device, reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(&info), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetImageSparseMemoryRequirements2KHR(m_device, |
| reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(&info), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| |
| template<typename SparseImageMemoryRequirements2Allocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SparseImageMemoryRequirements2>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> |
| Device::getImageSparseMemoryRequirements2KHR(const VULKAN_HPP_NAMESPACE::ImageSparseMemoryRequirementsInfo2 &info, |
| SparseImageMemoryRequirements2Allocator &sparseImageMemoryRequirements2Allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements(sparseImageMemoryRequirements2Allocator); |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetImageSparseMemoryRequirements2KHR( |
| m_device, reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(&info), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetImageSparseMemoryRequirements2KHR(m_device, |
| reinterpret_cast<const VkImageSparseMemoryRequirementsInfo2 *>(&info), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_acceleration_structure === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createAccelerationStructureKHR(const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoKHR *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::AccelerationStructureKHR *pAccelerationStructure, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateAccelerationStructureKHR(m_device, |
| reinterpret_cast<const VkAccelerationStructureCreateInfoKHR *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkAccelerationStructureKHR *>(pAccelerationStructure))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::AccelerationStructureKHR>::type |
| Device::createAccelerationStructureKHR(const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure; |
| VkResult result = d.vkCreateAccelerationStructureKHR( |
| m_device, |
| reinterpret_cast<const VkAccelerationStructureCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkAccelerationStructureKHR *>(&accelerationStructure)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createAccelerationStructureKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), accelerationStructure); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::AccelerationStructureKHR, Dispatch>>::type |
| Device::createAccelerationStructureKHRUnique(const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure; |
| Result result = static_cast<Result>(d.vkCreateAccelerationStructureKHR( |
| m_device, |
| reinterpret_cast<const VkAccelerationStructureCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkAccelerationStructureKHR *>(&accelerationStructure))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::AccelerationStructureKHR, Dispatch>( |
| result, accelerationStructure, VULKAN_HPP_NAMESPACE_STRING "::Device::createAccelerationStructureKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyAccelerationStructureKHR(VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyAccelerationStructureKHR( |
| m_device, static_cast<VkAccelerationStructureKHR>(accelerationStructure), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyAccelerationStructureKHR(VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyAccelerationStructureKHR( |
| m_device, |
| static_cast<VkAccelerationStructureKHR>(accelerationStructure), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyAccelerationStructureKHR( |
| m_device, static_cast<VkAccelerationStructureKHR>(accelerationStructure), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::AccelerationStructureKHR accelerationStructure, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyAccelerationStructureKHR( |
| m_device, |
| static_cast<VkAccelerationStructureKHR>(accelerationStructure), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::buildAccelerationStructuresKHR(uint32_t infoCount, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR *pInfos, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildRangeInfoKHR *const *ppBuildRangeInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBuildAccelerationStructuresKHR(m_commandBuffer, |
| infoCount, |
| reinterpret_cast<const VkAccelerationStructureBuildGeometryInfoKHR *>(pInfos), |
| reinterpret_cast<const VkAccelerationStructureBuildRangeInfoKHR *const *>(ppBuildRangeInfos)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::buildAccelerationStructuresKHR( |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR> const &infos, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildRangeInfoKHR *const> const &pBuildRangeInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(infos.size() == pBuildRangeInfos.size()); |
| # else |
| if(infos.size() != pBuildRangeInfos.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::buildAccelerationStructuresKHR: infos.size() != pBuildRangeInfos.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdBuildAccelerationStructuresKHR(m_commandBuffer, |
| infos.size(), |
| reinterpret_cast<const VkAccelerationStructureBuildGeometryInfoKHR *>(infos.data()), |
| reinterpret_cast<const VkAccelerationStructureBuildRangeInfoKHR *const *>(pBuildRangeInfos.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::buildAccelerationStructuresIndirectKHR(uint32_t infoCount, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR *pInfos, |
| const VULKAN_HPP_NAMESPACE::DeviceAddress *pIndirectDeviceAddresses, |
| const uint32_t *pIndirectStrides, |
| const uint32_t *const *ppMaxPrimitiveCounts, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBuildAccelerationStructuresIndirectKHR(m_commandBuffer, |
| infoCount, |
| reinterpret_cast<const VkAccelerationStructureBuildGeometryInfoKHR *>(pInfos), |
| reinterpret_cast<const VkDeviceAddress *>(pIndirectDeviceAddresses), |
| pIndirectStrides, |
| ppMaxPrimitiveCounts); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::buildAccelerationStructuresIndirectKHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR> const &infos, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceAddress> const &indirectDeviceAddresses, |
| ArrayProxy<const uint32_t> const &indirectStrides, |
| ArrayProxy<const uint32_t *const> const &pMaxPrimitiveCounts, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(infos.size() == indirectDeviceAddresses.size()); |
| VULKAN_HPP_ASSERT(infos.size() == indirectStrides.size()); |
| VULKAN_HPP_ASSERT(infos.size() == pMaxPrimitiveCounts.size()); |
| # else |
| if(infos.size() != indirectDeviceAddresses.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::buildAccelerationStructuresIndirectKHR: infos.size() != indirectDeviceAddresses.size()"); |
| } |
| if(infos.size() != indirectStrides.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::buildAccelerationStructuresIndirectKHR: infos.size() != indirectStrides.size()"); |
| } |
| if(infos.size() != pMaxPrimitiveCounts.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::buildAccelerationStructuresIndirectKHR: infos.size() != pMaxPrimitiveCounts.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdBuildAccelerationStructuresIndirectKHR(m_commandBuffer, |
| infos.size(), |
| reinterpret_cast<const VkAccelerationStructureBuildGeometryInfoKHR *>(infos.data()), |
| reinterpret_cast<const VkDeviceAddress *>(indirectDeviceAddresses.data()), |
| indirectStrides.data(), |
| pMaxPrimitiveCounts.data()); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::buildAccelerationStructuresKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| uint32_t infoCount, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR *pInfos, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildRangeInfoKHR *const *ppBuildRangeInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkBuildAccelerationStructuresKHR(m_device, |
| static_cast<VkDeferredOperationKHR>(deferredOperation), |
| infoCount, |
| reinterpret_cast<const VkAccelerationStructureBuildGeometryInfoKHR *>(pInfos), |
| reinterpret_cast<const VkAccelerationStructureBuildRangeInfoKHR *const *>(ppBuildRangeInfos))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result |
| Device::buildAccelerationStructuresKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR> const &infos, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildRangeInfoKHR *const> const &pBuildRangeInfos, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(infos.size() == pBuildRangeInfos.size()); |
| # else |
| if(infos.size() != pBuildRangeInfos.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::Device::buildAccelerationStructuresKHR: infos.size() != pBuildRangeInfos.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| VkResult result = |
| d.vkBuildAccelerationStructuresKHR(m_device, |
| static_cast<VkDeferredOperationKHR>(deferredOperation), |
| infos.size(), |
| reinterpret_cast<const VkAccelerationStructureBuildGeometryInfoKHR *>(infos.data()), |
| reinterpret_cast<const VkAccelerationStructureBuildRangeInfoKHR *const *>(pBuildRangeInfos.data())); |
| resultCheck( |
| static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::buildAccelerationStructuresKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::copyAccelerationStructureKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureInfoKHR *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCopyAccelerationStructureKHR( |
| m_device, static_cast<VkDeferredOperationKHR>(deferredOperation), reinterpret_cast<const VkCopyAccelerationStructureInfoKHR *>(pInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result |
| Device::copyAccelerationStructureKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureInfoKHR &info, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkCopyAccelerationStructureKHR( |
| m_device, static_cast<VkDeferredOperationKHR>(deferredOperation), reinterpret_cast<const VkCopyAccelerationStructureInfoKHR *>(&info)); |
| resultCheck( |
| static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::copyAccelerationStructureKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::copyAccelerationStructureToMemoryKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureToMemoryInfoKHR *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCopyAccelerationStructureToMemoryKHR( |
| m_device, static_cast<VkDeferredOperationKHR>(deferredOperation), reinterpret_cast<const VkCopyAccelerationStructureToMemoryInfoKHR *>(pInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result |
| Device::copyAccelerationStructureToMemoryKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureToMemoryInfoKHR &info, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkCopyAccelerationStructureToMemoryKHR( |
| m_device, static_cast<VkDeferredOperationKHR>(deferredOperation), reinterpret_cast<const VkCopyAccelerationStructureToMemoryInfoKHR *>(&info)); |
| resultCheck( |
| static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::copyAccelerationStructureToMemoryKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::copyMemoryToAccelerationStructureKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| const VULKAN_HPP_NAMESPACE::CopyMemoryToAccelerationStructureInfoKHR *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCopyMemoryToAccelerationStructureKHR( |
| m_device, static_cast<VkDeferredOperationKHR>(deferredOperation), reinterpret_cast<const VkCopyMemoryToAccelerationStructureInfoKHR *>(pInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result |
| Device::copyMemoryToAccelerationStructureKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| const VULKAN_HPP_NAMESPACE::CopyMemoryToAccelerationStructureInfoKHR &info, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkCopyMemoryToAccelerationStructureKHR( |
| m_device, static_cast<VkDeferredOperationKHR>(deferredOperation), reinterpret_cast<const VkCopyMemoryToAccelerationStructureInfoKHR *>(&info)); |
| resultCheck( |
| static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::copyMemoryToAccelerationStructureKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::writeAccelerationStructuresPropertiesKHR(uint32_t accelerationStructureCount, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR *pAccelerationStructures, |
| VULKAN_HPP_NAMESPACE::QueryType queryType, |
| size_t dataSize, |
| void *pData, |
| size_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkWriteAccelerationStructuresPropertiesKHR(m_device, |
| accelerationStructureCount, |
| reinterpret_cast<const VkAccelerationStructureKHR *>(pAccelerationStructures), |
| static_cast<VkQueryType>(queryType), |
| dataSize, |
| pData, |
| stride)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename DataTypeAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DataType, DataTypeAllocator>>::type |
| Device::writeAccelerationStructuresPropertiesKHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR> const &accelerationStructures, |
| VULKAN_HPP_NAMESPACE::QueryType queryType, |
| size_t dataSize, |
| size_t stride, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_ASSERT(dataSize % sizeof(DataType) == 0); |
| std::vector<DataType, DataTypeAllocator> data(dataSize / sizeof(DataType)); |
| Result result = |
| static_cast<Result>(d.vkWriteAccelerationStructuresPropertiesKHR(m_device, |
| accelerationStructures.size(), |
| reinterpret_cast<const VkAccelerationStructureKHR *>(accelerationStructures.data()), |
| static_cast<VkQueryType>(queryType), |
| data.size() * sizeof(DataType), |
| reinterpret_cast<void *>(data.data()), |
| stride)); |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::writeAccelerationStructuresPropertiesKHR"); |
| } |
| |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<DataType>::type |
| Device::writeAccelerationStructuresPropertyKHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR> const &accelerationStructures, |
| VULKAN_HPP_NAMESPACE::QueryType queryType, |
| size_t stride, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| DataType data; |
| VkResult result = d.vkWriteAccelerationStructuresPropertiesKHR(m_device, |
| accelerationStructures.size(), |
| reinterpret_cast<const VkAccelerationStructureKHR *>(accelerationStructures.data()), |
| static_cast<VkQueryType>(queryType), |
| sizeof(DataType), |
| reinterpret_cast<void *>(&data), |
| stride); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::writeAccelerationStructuresPropertyKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), data); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureKHR(const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureInfoKHR *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyAccelerationStructureKHR(m_commandBuffer, reinterpret_cast<const VkCopyAccelerationStructureInfoKHR *>(pInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureKHR(const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureInfoKHR &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyAccelerationStructureKHR(m_commandBuffer, reinterpret_cast<const VkCopyAccelerationStructureInfoKHR *>(&info)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureToMemoryKHR(const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureToMemoryInfoKHR *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyAccelerationStructureToMemoryKHR(m_commandBuffer, reinterpret_cast<const VkCopyAccelerationStructureToMemoryInfoKHR *>(pInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureToMemoryKHR(const VULKAN_HPP_NAMESPACE::CopyAccelerationStructureToMemoryInfoKHR &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyAccelerationStructureToMemoryKHR(m_commandBuffer, reinterpret_cast<const VkCopyAccelerationStructureToMemoryInfoKHR *>(&info)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyMemoryToAccelerationStructureKHR(const VULKAN_HPP_NAMESPACE::CopyMemoryToAccelerationStructureInfoKHR *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyMemoryToAccelerationStructureKHR(m_commandBuffer, reinterpret_cast<const VkCopyMemoryToAccelerationStructureInfoKHR *>(pInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyMemoryToAccelerationStructureKHR(const VULKAN_HPP_NAMESPACE::CopyMemoryToAccelerationStructureInfoKHR &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyMemoryToAccelerationStructureKHR(m_commandBuffer, reinterpret_cast<const VkCopyMemoryToAccelerationStructureInfoKHR *>(&info)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE DeviceAddress Device::getAccelerationStructureAddressKHR(const VULKAN_HPP_NAMESPACE::AccelerationStructureDeviceAddressInfoKHR *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<DeviceAddress>( |
| d.vkGetAccelerationStructureDeviceAddressKHR(m_device, reinterpret_cast<const VkAccelerationStructureDeviceAddressInfoKHR *>(pInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::DeviceAddress |
| Device::getAccelerationStructureAddressKHR(const VULKAN_HPP_NAMESPACE::AccelerationStructureDeviceAddressInfoKHR &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkDeviceAddress result = |
| d.vkGetAccelerationStructureDeviceAddressKHR(m_device, reinterpret_cast<const VkAccelerationStructureDeviceAddressInfoKHR *>(&info)); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::DeviceAddress>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::writeAccelerationStructuresPropertiesKHR(uint32_t accelerationStructureCount, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR *pAccelerationStructures, |
| VULKAN_HPP_NAMESPACE::QueryType queryType, |
| VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWriteAccelerationStructuresPropertiesKHR(m_commandBuffer, |
| accelerationStructureCount, |
| reinterpret_cast<const VkAccelerationStructureKHR *>(pAccelerationStructures), |
| static_cast<VkQueryType>(queryType), |
| static_cast<VkQueryPool>(queryPool), |
| firstQuery); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::writeAccelerationStructuresPropertiesKHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureKHR> const &accelerationStructures, |
| VULKAN_HPP_NAMESPACE::QueryType queryType, |
| VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdWriteAccelerationStructuresPropertiesKHR(m_commandBuffer, |
| accelerationStructures.size(), |
| reinterpret_cast<const VkAccelerationStructureKHR *>(accelerationStructures.data()), |
| static_cast<VkQueryType>(queryType), |
| static_cast<VkQueryPool>(queryPool), |
| firstQuery); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getAccelerationStructureCompatibilityKHR(const VULKAN_HPP_NAMESPACE::AccelerationStructureVersionInfoKHR *pVersionInfo, |
| VULKAN_HPP_NAMESPACE::AccelerationStructureCompatibilityKHR *pCompatibility, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceAccelerationStructureCompatibilityKHR(m_device, |
| reinterpret_cast<const VkAccelerationStructureVersionInfoKHR *>(pVersionInfo), |
| reinterpret_cast<VkAccelerationStructureCompatibilityKHR *>(pCompatibility)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::AccelerationStructureCompatibilityKHR |
| Device::getAccelerationStructureCompatibilityKHR(const VULKAN_HPP_NAMESPACE::AccelerationStructureVersionInfoKHR &versionInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::AccelerationStructureCompatibilityKHR compatibility; |
| d.vkGetDeviceAccelerationStructureCompatibilityKHR(m_device, |
| reinterpret_cast<const VkAccelerationStructureVersionInfoKHR *>(&versionInfo), |
| reinterpret_cast<VkAccelerationStructureCompatibilityKHR *>(&compatibility)); |
| |
| return compatibility; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getAccelerationStructureBuildSizesKHR(VULKAN_HPP_NAMESPACE::AccelerationStructureBuildTypeKHR buildType, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR *pBuildInfo, |
| const uint32_t *pMaxPrimitiveCounts, |
| VULKAN_HPP_NAMESPACE::AccelerationStructureBuildSizesInfoKHR *pSizeInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetAccelerationStructureBuildSizesKHR(m_device, |
| static_cast<VkAccelerationStructureBuildTypeKHR>(buildType), |
| reinterpret_cast<const VkAccelerationStructureBuildGeometryInfoKHR *>(pBuildInfo), |
| pMaxPrimitiveCounts, |
| reinterpret_cast<VkAccelerationStructureBuildSizesInfoKHR *>(pSizeInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::AccelerationStructureBuildSizesInfoKHR |
| Device::getAccelerationStructureBuildSizesKHR(VULKAN_HPP_NAMESPACE::AccelerationStructureBuildTypeKHR buildType, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureBuildGeometryInfoKHR &buildInfo, |
| ArrayProxy<const uint32_t> const &maxPrimitiveCounts, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(maxPrimitiveCounts.size() == buildInfo.geometryCount); |
| # else |
| if(maxPrimitiveCounts.size() != buildInfo.geometryCount) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::Device::getAccelerationStructureBuildSizesKHR: maxPrimitiveCounts.size() != buildInfo.geometryCount"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| VULKAN_HPP_NAMESPACE::AccelerationStructureBuildSizesInfoKHR sizeInfo; |
| d.vkGetAccelerationStructureBuildSizesKHR(m_device, |
| static_cast<VkAccelerationStructureBuildTypeKHR>(buildType), |
| reinterpret_cast<const VkAccelerationStructureBuildGeometryInfoKHR *>(&buildInfo), |
| maxPrimitiveCounts.data(), |
| reinterpret_cast<VkAccelerationStructureBuildSizesInfoKHR *>(&sizeInfo)); |
| |
| return sizeInfo; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_sampler_ycbcr_conversion === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createSamplerYcbcrConversionKHR(const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion *pYcbcrConversion, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateSamplerYcbcrConversionKHR(m_device, |
| reinterpret_cast<const VkSamplerYcbcrConversionCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSamplerYcbcrConversion *>(pYcbcrConversion))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion>::type |
| Device::createSamplerYcbcrConversionKHR(const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion; |
| VkResult result = d.vkCreateSamplerYcbcrConversionKHR( |
| m_device, |
| reinterpret_cast<const VkSamplerYcbcrConversionCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSamplerYcbcrConversion *>(&ycbcrConversion)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createSamplerYcbcrConversionKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), ycbcrConversion); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion, Dispatch>>::type |
| Device::createSamplerYcbcrConversionKHRUnique(const VULKAN_HPP_NAMESPACE::SamplerYcbcrConversionCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion; |
| Result result = static_cast<Result>(d.vkCreateSamplerYcbcrConversionKHR( |
| m_device, |
| reinterpret_cast<const VkSamplerYcbcrConversionCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSamplerYcbcrConversion *>(&ycbcrConversion))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion, Dispatch>( |
| result, ycbcrConversion, VULKAN_HPP_NAMESPACE_STRING "::Device::createSamplerYcbcrConversionKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySamplerYcbcrConversionKHR(VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroySamplerYcbcrConversionKHR( |
| m_device, static_cast<VkSamplerYcbcrConversion>(ycbcrConversion), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroySamplerYcbcrConversionKHR(VULKAN_HPP_NAMESPACE::SamplerYcbcrConversion ycbcrConversion, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroySamplerYcbcrConversionKHR( |
| m_device, |
| static_cast<VkSamplerYcbcrConversion>(ycbcrConversion), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_bind_memory2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::bindBufferMemory2KHR(uint32_t bindInfoCount, |
| const VULKAN_HPP_NAMESPACE::BindBufferMemoryInfo *pBindInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkBindBufferMemory2KHR(m_device, bindInfoCount, reinterpret_cast<const VkBindBufferMemoryInfo *>(pBindInfos))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::bindBufferMemory2KHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::BindBufferMemoryInfo> const &bindInfos, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkBindBufferMemory2KHR(m_device, bindInfos.size(), reinterpret_cast<const VkBindBufferMemoryInfo *>(bindInfos.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::bindBufferMemory2KHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::bindImageMemory2KHR(uint32_t bindInfoCount, |
| const VULKAN_HPP_NAMESPACE::BindImageMemoryInfo *pBindInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkBindImageMemory2KHR(m_device, bindInfoCount, reinterpret_cast<const VkBindImageMemoryInfo *>(pBindInfos))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::bindImageMemory2KHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::BindImageMemoryInfo> const &bindInfos, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkBindImageMemory2KHR(m_device, bindInfos.size(), reinterpret_cast<const VkBindImageMemoryInfo *>(bindInfos.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::bindImageMemory2KHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_image_drm_format_modifier === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getImageDrmFormatModifierPropertiesEXT( |
| VULKAN_HPP_NAMESPACE::Image image, VULKAN_HPP_NAMESPACE::ImageDrmFormatModifierPropertiesEXT *pProperties, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetImageDrmFormatModifierPropertiesEXT( |
| m_device, static_cast<VkImage>(image), reinterpret_cast<VkImageDrmFormatModifierPropertiesEXT *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::ImageDrmFormatModifierPropertiesEXT>::type |
| Device::getImageDrmFormatModifierPropertiesEXT(VULKAN_HPP_NAMESPACE::Image image, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ImageDrmFormatModifierPropertiesEXT properties; |
| VkResult result = d.vkGetImageDrmFormatModifierPropertiesEXT( |
| m_device, static_cast<VkImage>(image), reinterpret_cast<VkImageDrmFormatModifierPropertiesEXT *>(&properties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getImageDrmFormatModifierPropertiesEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), properties); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_validation_cache === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createValidationCacheEXT(const VULKAN_HPP_NAMESPACE::ValidationCacheCreateInfoEXT *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::ValidationCacheEXT *pValidationCache, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateValidationCacheEXT(m_device, |
| reinterpret_cast<const VkValidationCacheCreateInfoEXT *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkValidationCacheEXT *>(pValidationCache))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::ValidationCacheEXT>::type |
| Device::createValidationCacheEXT(const VULKAN_HPP_NAMESPACE::ValidationCacheCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache; |
| VkResult result = d.vkCreateValidationCacheEXT( |
| m_device, |
| reinterpret_cast<const VkValidationCacheCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkValidationCacheEXT *>(&validationCache)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createValidationCacheEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), validationCache); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::ValidationCacheEXT, Dispatch>>::type |
| Device::createValidationCacheEXTUnique(const VULKAN_HPP_NAMESPACE::ValidationCacheCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache; |
| Result result = static_cast<Result>(d.vkCreateValidationCacheEXT( |
| m_device, |
| reinterpret_cast<const VkValidationCacheCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkValidationCacheEXT *>(&validationCache))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::ValidationCacheEXT, Dispatch>( |
| result, validationCache, VULKAN_HPP_NAMESPACE_STRING "::Device::createValidationCacheEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyValidationCacheEXT(VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyValidationCacheEXT( |
| m_device, static_cast<VkValidationCacheEXT>(validationCache), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyValidationCacheEXT(VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyValidationCacheEXT( |
| m_device, |
| static_cast<VkValidationCacheEXT>(validationCache), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyValidationCacheEXT( |
| m_device, static_cast<VkValidationCacheEXT>(validationCache), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyValidationCacheEXT( |
| m_device, |
| static_cast<VkValidationCacheEXT>(validationCache), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::mergeValidationCachesEXT(VULKAN_HPP_NAMESPACE::ValidationCacheEXT dstCache, |
| uint32_t srcCacheCount, |
| const VULKAN_HPP_NAMESPACE::ValidationCacheEXT *pSrcCaches, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkMergeValidationCachesEXT( |
| m_device, static_cast<VkValidationCacheEXT>(dstCache), srcCacheCount, reinterpret_cast<const VkValidationCacheEXT *>(pSrcCaches))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::mergeValidationCachesEXT( |
| VULKAN_HPP_NAMESPACE::ValidationCacheEXT dstCache, ArrayProxy<const VULKAN_HPP_NAMESPACE::ValidationCacheEXT> const &srcCaches, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkMergeValidationCachesEXT( |
| m_device, static_cast<VkValidationCacheEXT>(dstCache), srcCaches.size(), reinterpret_cast<const VkValidationCacheEXT *>(srcCaches.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::mergeValidationCachesEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getValidationCacheDataEXT(VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache, |
| size_t *pDataSize, |
| void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetValidationCacheDataEXT(m_device, static_cast<VkValidationCacheEXT>(validationCache), pDataSize, pData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Uint8_tAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type |
| Device::getValidationCacheDataEXT(VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<uint8_t, Uint8_tAllocator> data; |
| size_t dataSize; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetValidationCacheDataEXT(m_device, static_cast<VkValidationCacheEXT>(validationCache), &dataSize, nullptr)); |
| if((result == Result::eSuccess) && dataSize) |
| { |
| data.resize(dataSize); |
| result = static_cast<Result>( |
| d.vkGetValidationCacheDataEXT(m_device, static_cast<VkValidationCacheEXT>(validationCache), &dataSize, reinterpret_cast<void *>(data.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(dataSize <= data.size()); |
| if(dataSize < data.size()) |
| { |
| data.resize(dataSize); |
| } |
| } |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getValidationCacheDataEXT"); |
| } |
| |
| template<typename Uint8_tAllocator, typename Dispatch, typename B, typename std::enable_if<std::is_same<typename B::value_type, uint8_t>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<uint8_t, Uint8_tAllocator>>::type |
| Device::getValidationCacheDataEXT(VULKAN_HPP_NAMESPACE::ValidationCacheEXT validationCache, Uint8_tAllocator &uint8_tAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<uint8_t, Uint8_tAllocator> data(uint8_tAllocator); |
| size_t dataSize; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetValidationCacheDataEXT(m_device, static_cast<VkValidationCacheEXT>(validationCache), &dataSize, nullptr)); |
| if((result == Result::eSuccess) && dataSize) |
| { |
| data.resize(dataSize); |
| result = static_cast<Result>( |
| d.vkGetValidationCacheDataEXT(m_device, static_cast<VkValidationCacheEXT>(validationCache), &dataSize, reinterpret_cast<void *>(data.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(dataSize <= data.size()); |
| if(dataSize < data.size()) |
| { |
| data.resize(dataSize); |
| } |
| } |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getValidationCacheDataEXT"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NV_shading_rate_image === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindShadingRateImageNV(VULKAN_HPP_NAMESPACE::ImageView imageView, |
| VULKAN_HPP_NAMESPACE::ImageLayout imageLayout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindShadingRateImageNV(m_commandBuffer, static_cast<VkImageView>(imageView), static_cast<VkImageLayout>(imageLayout)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setViewportShadingRatePaletteNV(uint32_t firstViewport, |
| uint32_t viewportCount, |
| const VULKAN_HPP_NAMESPACE::ShadingRatePaletteNV *pShadingRatePalettes, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetViewportShadingRatePaletteNV( |
| m_commandBuffer, firstViewport, viewportCount, reinterpret_cast<const VkShadingRatePaletteNV *>(pShadingRatePalettes)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::setViewportShadingRatePaletteNV(uint32_t firstViewport, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::ShadingRatePaletteNV> const &shadingRatePalettes, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetViewportShadingRatePaletteNV( |
| m_commandBuffer, firstViewport, shadingRatePalettes.size(), reinterpret_cast<const VkShadingRatePaletteNV *>(shadingRatePalettes.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setCoarseSampleOrderNV(VULKAN_HPP_NAMESPACE::CoarseSampleOrderTypeNV sampleOrderType, |
| uint32_t customSampleOrderCount, |
| const VULKAN_HPP_NAMESPACE::CoarseSampleOrderCustomNV *pCustomSampleOrders, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetCoarseSampleOrderNV(m_commandBuffer, |
| static_cast<VkCoarseSampleOrderTypeNV>(sampleOrderType), |
| customSampleOrderCount, |
| reinterpret_cast<const VkCoarseSampleOrderCustomNV *>(pCustomSampleOrders)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setCoarseSampleOrderNV(VULKAN_HPP_NAMESPACE::CoarseSampleOrderTypeNV sampleOrderType, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::CoarseSampleOrderCustomNV> const &customSampleOrders, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetCoarseSampleOrderNV(m_commandBuffer, |
| static_cast<VkCoarseSampleOrderTypeNV>(sampleOrderType), |
| customSampleOrders.size(), |
| reinterpret_cast<const VkCoarseSampleOrderCustomNV *>(customSampleOrders.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NV_ray_tracing === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createAccelerationStructureNV(const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoNV *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::AccelerationStructureNV *pAccelerationStructure, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateAccelerationStructureNV(m_device, |
| reinterpret_cast<const VkAccelerationStructureCreateInfoNV *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkAccelerationStructureNV *>(pAccelerationStructure))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::AccelerationStructureNV>::type |
| Device::createAccelerationStructureNV(const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoNV &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure; |
| VkResult result = d.vkCreateAccelerationStructureNV( |
| m_device, |
| reinterpret_cast<const VkAccelerationStructureCreateInfoNV *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkAccelerationStructureNV *>(&accelerationStructure)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createAccelerationStructureNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), accelerationStructure); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::AccelerationStructureNV, Dispatch>>::type |
| Device::createAccelerationStructureNVUnique(const VULKAN_HPP_NAMESPACE::AccelerationStructureCreateInfoNV &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure; |
| Result result = static_cast<Result>(d.vkCreateAccelerationStructureNV( |
| m_device, |
| reinterpret_cast<const VkAccelerationStructureCreateInfoNV *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkAccelerationStructureNV *>(&accelerationStructure))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::AccelerationStructureNV, Dispatch>( |
| result, accelerationStructure, VULKAN_HPP_NAMESPACE_STRING "::Device::createAccelerationStructureNVUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyAccelerationStructureNV(VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyAccelerationStructureNV( |
| m_device, static_cast<VkAccelerationStructureNV>(accelerationStructure), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyAccelerationStructureNV(VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyAccelerationStructureNV( |
| m_device, |
| static_cast<VkAccelerationStructureNV>(accelerationStructure), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyAccelerationStructureNV( |
| m_device, static_cast<VkAccelerationStructureNV>(accelerationStructure), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyAccelerationStructureNV( |
| m_device, |
| static_cast<VkAccelerationStructureNV>(accelerationStructure), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| Device::getAccelerationStructureMemoryRequirementsNV(const VULKAN_HPP_NAMESPACE::AccelerationStructureMemoryRequirementsInfoNV *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2KHR *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetAccelerationStructureMemoryRequirementsNV(m_device, |
| reinterpret_cast<const VkAccelerationStructureMemoryRequirementsInfoNV *>(pInfo), |
| reinterpret_cast<VkMemoryRequirements2KHR *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2KHR |
| Device::getAccelerationStructureMemoryRequirementsNV(const VULKAN_HPP_NAMESPACE::AccelerationStructureMemoryRequirementsInfoNV &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2KHR memoryRequirements; |
| d.vkGetAccelerationStructureMemoryRequirementsNV(m_device, |
| reinterpret_cast<const VkAccelerationStructureMemoryRequirementsInfoNV *>(&info), |
| reinterpret_cast<VkMemoryRequirements2KHR *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getAccelerationStructureMemoryRequirementsNV(const VULKAN_HPP_NAMESPACE::AccelerationStructureMemoryRequirementsInfoNV &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2KHR &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2KHR>(); |
| d.vkGetAccelerationStructureMemoryRequirementsNV(m_device, |
| reinterpret_cast<const VkAccelerationStructureMemoryRequirementsInfoNV *>(&info), |
| reinterpret_cast<VkMemoryRequirements2KHR *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::bindAccelerationStructureMemoryNV( |
| uint32_t bindInfoCount, const VULKAN_HPP_NAMESPACE::BindAccelerationStructureMemoryInfoNV *pBindInfos, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkBindAccelerationStructureMemoryNV(m_device, bindInfoCount, reinterpret_cast<const VkBindAccelerationStructureMemoryInfoNV *>(pBindInfos))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::bindAccelerationStructureMemoryNV(ArrayProxy<const VULKAN_HPP_NAMESPACE::BindAccelerationStructureMemoryInfoNV> const &bindInfos, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkBindAccelerationStructureMemoryNV( |
| m_device, bindInfos.size(), reinterpret_cast<const VkBindAccelerationStructureMemoryInfoNV *>(bindInfos.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::bindAccelerationStructureMemoryNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::buildAccelerationStructureNV(const VULKAN_HPP_NAMESPACE::AccelerationStructureInfoNV *pInfo, |
| VULKAN_HPP_NAMESPACE::Buffer instanceData, |
| VULKAN_HPP_NAMESPACE::DeviceSize instanceOffset, |
| VULKAN_HPP_NAMESPACE::Bool32 update, |
| VULKAN_HPP_NAMESPACE::AccelerationStructureNV dst, |
| VULKAN_HPP_NAMESPACE::AccelerationStructureNV src, |
| VULKAN_HPP_NAMESPACE::Buffer scratch, |
| VULKAN_HPP_NAMESPACE::DeviceSize scratchOffset, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBuildAccelerationStructureNV(m_commandBuffer, |
| reinterpret_cast<const VkAccelerationStructureInfoNV *>(pInfo), |
| static_cast<VkBuffer>(instanceData), |
| static_cast<VkDeviceSize>(instanceOffset), |
| static_cast<VkBool32>(update), |
| static_cast<VkAccelerationStructureNV>(dst), |
| static_cast<VkAccelerationStructureNV>(src), |
| static_cast<VkBuffer>(scratch), |
| static_cast<VkDeviceSize>(scratchOffset)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::buildAccelerationStructureNV(const VULKAN_HPP_NAMESPACE::AccelerationStructureInfoNV &info, |
| VULKAN_HPP_NAMESPACE::Buffer instanceData, |
| VULKAN_HPP_NAMESPACE::DeviceSize instanceOffset, |
| VULKAN_HPP_NAMESPACE::Bool32 update, |
| VULKAN_HPP_NAMESPACE::AccelerationStructureNV dst, |
| VULKAN_HPP_NAMESPACE::AccelerationStructureNV src, |
| VULKAN_HPP_NAMESPACE::Buffer scratch, |
| VULKAN_HPP_NAMESPACE::DeviceSize scratchOffset, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBuildAccelerationStructureNV(m_commandBuffer, |
| reinterpret_cast<const VkAccelerationStructureInfoNV *>(&info), |
| static_cast<VkBuffer>(instanceData), |
| static_cast<VkDeviceSize>(instanceOffset), |
| static_cast<VkBool32>(update), |
| static_cast<VkAccelerationStructureNV>(dst), |
| static_cast<VkAccelerationStructureNV>(src), |
| static_cast<VkBuffer>(scratch), |
| static_cast<VkDeviceSize>(scratchOffset)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyAccelerationStructureNV(VULKAN_HPP_NAMESPACE::AccelerationStructureNV dst, |
| VULKAN_HPP_NAMESPACE::AccelerationStructureNV src, |
| VULKAN_HPP_NAMESPACE::CopyAccelerationStructureModeKHR mode, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyAccelerationStructureNV(m_commandBuffer, |
| static_cast<VkAccelerationStructureNV>(dst), |
| static_cast<VkAccelerationStructureNV>(src), |
| static_cast<VkCopyAccelerationStructureModeKHR>(mode)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::traceRaysNV(VULKAN_HPP_NAMESPACE::Buffer raygenShaderBindingTableBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize raygenShaderBindingOffset, |
| VULKAN_HPP_NAMESPACE::Buffer missShaderBindingTableBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize missShaderBindingOffset, |
| VULKAN_HPP_NAMESPACE::DeviceSize missShaderBindingStride, |
| VULKAN_HPP_NAMESPACE::Buffer hitShaderBindingTableBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize hitShaderBindingOffset, |
| VULKAN_HPP_NAMESPACE::DeviceSize hitShaderBindingStride, |
| VULKAN_HPP_NAMESPACE::Buffer callableShaderBindingTableBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize callableShaderBindingOffset, |
| VULKAN_HPP_NAMESPACE::DeviceSize callableShaderBindingStride, |
| uint32_t width, |
| uint32_t height, |
| uint32_t depth, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdTraceRaysNV(m_commandBuffer, |
| static_cast<VkBuffer>(raygenShaderBindingTableBuffer), |
| static_cast<VkDeviceSize>(raygenShaderBindingOffset), |
| static_cast<VkBuffer>(missShaderBindingTableBuffer), |
| static_cast<VkDeviceSize>(missShaderBindingOffset), |
| static_cast<VkDeviceSize>(missShaderBindingStride), |
| static_cast<VkBuffer>(hitShaderBindingTableBuffer), |
| static_cast<VkDeviceSize>(hitShaderBindingOffset), |
| static_cast<VkDeviceSize>(hitShaderBindingStride), |
| static_cast<VkBuffer>(callableShaderBindingTableBuffer), |
| static_cast<VkDeviceSize>(callableShaderBindingOffset), |
| static_cast<VkDeviceSize>(callableShaderBindingStride), |
| width, |
| height, |
| depth); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createRayTracingPipelinesNV(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| uint32_t createInfoCount, |
| const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV *pCreateInfos, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Pipeline *pPipelines, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateRayTracingPipelinesNV(m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfoCount, |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoNV *>(pCreateInfos), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkPipeline *>(pPipelines))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PipelineAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>> |
| Device::createRayTracingPipelinesNV(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines(createInfos.size()); |
| VkResult result = d.vkCreateRayTracingPipelinesNV( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoNV *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelinesNV", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelines); |
| } |
| |
| template<typename PipelineAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, Pipeline>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>> |
| Device::createRayTracingPipelinesNV(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| PipelineAllocator &pipelineAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines(createInfos.size(), pipelineAllocator); |
| VkResult result = d.vkCreateRayTracingPipelinesNV( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoNV *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelinesNV", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelines); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<VULKAN_HPP_NAMESPACE::Pipeline> |
| Device::createRayTracingPipelineNV(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline; |
| VkResult result = d.vkCreateRayTracingPipelinesNV( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| 1, |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoNV *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(&pipeline)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelineNV", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<VULKAN_HPP_NAMESPACE::Pipeline>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipeline); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch, typename PipelineAllocator> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator>> |
| Device::createRayTracingPipelinesNVUnique(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator> uniquePipelines; |
| std::vector<Pipeline> pipelines(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateRayTracingPipelinesNV( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoNV *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data()))); |
| if((result == VULKAN_HPP_NAMESPACE::Result::eSuccess) || (result == VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT)) |
| { |
| uniquePipelines.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniquePipelines.push_back(UniqueHandle<Pipeline, Dispatch>(pipelines[i], deleter)); |
| } |
| } |
| return createResultValue(result, |
| std::move(uniquePipelines), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelinesNVUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| } |
| |
| template<typename Dispatch, |
| typename PipelineAllocator, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, UniqueHandle<Pipeline, Dispatch>>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator>> |
| Device::createRayTracingPipelinesNVUnique(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| PipelineAllocator &pipelineAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator> uniquePipelines(pipelineAllocator); |
| std::vector<Pipeline> pipelines(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateRayTracingPipelinesNV( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoNV *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data()))); |
| if((result == VULKAN_HPP_NAMESPACE::Result::eSuccess) || (result == VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT)) |
| { |
| uniquePipelines.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniquePipelines.push_back(UniqueHandle<Pipeline, Dispatch>(pipelines[i], deleter)); |
| } |
| } |
| return createResultValue(result, |
| std::move(uniquePipelines), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelinesNVUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<UniqueHandle<Pipeline, Dispatch>> |
| Device::createRayTracingPipelineNVUnique(VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoNV &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| Pipeline pipeline; |
| Result result = static_cast<Result>(d.vkCreateRayTracingPipelinesNV( |
| m_device, |
| static_cast<VkPipelineCache>(pipelineCache), |
| 1, |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoNV *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(&pipeline))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<Pipeline, Dispatch>(result, |
| pipeline, |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelineNVUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }, |
| deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getRayTracingShaderGroupHandlesNV(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| uint32_t firstGroup, |
| uint32_t groupCount, |
| size_t dataSize, |
| void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetRayTracingShaderGroupHandlesNV(m_device, static_cast<VkPipeline>(pipeline), firstGroup, groupCount, dataSize, pData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename DataTypeAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DataType, DataTypeAllocator>>::type Device::getRayTracingShaderGroupHandlesNV( |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount, size_t dataSize, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_ASSERT(dataSize % sizeof(DataType) == 0); |
| std::vector<DataType, DataTypeAllocator> data(dataSize / sizeof(DataType)); |
| Result result = static_cast<Result>(d.vkGetRayTracingShaderGroupHandlesNV( |
| m_device, static_cast<VkPipeline>(pipeline), firstGroup, groupCount, data.size() * sizeof(DataType), reinterpret_cast<void *>(data.data()))); |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getRayTracingShaderGroupHandlesNV"); |
| } |
| |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<DataType>::type |
| Device::getRayTracingShaderGroupHandleNV(VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| DataType data; |
| VkResult result = d.vkGetRayTracingShaderGroupHandlesNV( |
| m_device, static_cast<VkPipeline>(pipeline), firstGroup, groupCount, sizeof(DataType), reinterpret_cast<void *>(&data)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getRayTracingShaderGroupHandleNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), data); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getAccelerationStructureHandleNV(VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure, |
| size_t dataSize, |
| void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetAccelerationStructureHandleNV(m_device, static_cast<VkAccelerationStructureNV>(accelerationStructure), dataSize, pData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename DataTypeAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DataType, DataTypeAllocator>>::type |
| Device::getAccelerationStructureHandleNV(VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure, size_t dataSize, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_ASSERT(dataSize % sizeof(DataType) == 0); |
| std::vector<DataType, DataTypeAllocator> data(dataSize / sizeof(DataType)); |
| Result result = static_cast<Result>(d.vkGetAccelerationStructureHandleNV( |
| m_device, static_cast<VkAccelerationStructureNV>(accelerationStructure), data.size() * sizeof(DataType), reinterpret_cast<void *>(data.data()))); |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getAccelerationStructureHandleNV"); |
| } |
| |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<DataType>::type |
| Device::getAccelerationStructureHandleNV(VULKAN_HPP_NAMESPACE::AccelerationStructureNV accelerationStructure, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| DataType data; |
| VkResult result = d.vkGetAccelerationStructureHandleNV( |
| m_device, static_cast<VkAccelerationStructureNV>(accelerationStructure), sizeof(DataType), reinterpret_cast<void *>(&data)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getAccelerationStructureHandleNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), data); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::writeAccelerationStructuresPropertiesNV(uint32_t accelerationStructureCount, |
| const VULKAN_HPP_NAMESPACE::AccelerationStructureNV *pAccelerationStructures, |
| VULKAN_HPP_NAMESPACE::QueryType queryType, |
| VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWriteAccelerationStructuresPropertiesNV(m_commandBuffer, |
| accelerationStructureCount, |
| reinterpret_cast<const VkAccelerationStructureNV *>(pAccelerationStructures), |
| static_cast<VkQueryType>(queryType), |
| static_cast<VkQueryPool>(queryPool), |
| firstQuery); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::writeAccelerationStructuresPropertiesNV(ArrayProxy<const VULKAN_HPP_NAMESPACE::AccelerationStructureNV> const &accelerationStructures, |
| VULKAN_HPP_NAMESPACE::QueryType queryType, |
| VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdWriteAccelerationStructuresPropertiesNV(m_commandBuffer, |
| accelerationStructures.size(), |
| reinterpret_cast<const VkAccelerationStructureNV *>(accelerationStructures.data()), |
| static_cast<VkQueryType>(queryType), |
| static_cast<VkQueryPool>(queryPool), |
| firstQuery); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::compileDeferredNV(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| uint32_t shader, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCompileDeferredNV(m_device, static_cast<VkPipeline>(pipeline), shader)); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::compileDeferredNV(VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t shader, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkCompileDeferredNV(m_device, static_cast<VkPipeline>(pipeline), shader); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::compileDeferredNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_maintenance3 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getDescriptorSetLayoutSupportKHR(const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo *pCreateInfo, |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport *pSupport, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDescriptorSetLayoutSupportKHR( |
| m_device, reinterpret_cast<const VkDescriptorSetLayoutCreateInfo *>(pCreateInfo), reinterpret_cast<VkDescriptorSetLayoutSupport *>(pSupport)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport |
| Device::getDescriptorSetLayoutSupportKHR(const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo &createInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport support; |
| d.vkGetDescriptorSetLayoutSupportKHR( |
| m_device, reinterpret_cast<const VkDescriptorSetLayoutCreateInfo *>(&createInfo), reinterpret_cast<VkDescriptorSetLayoutSupport *>(&support)); |
| |
| return support; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getDescriptorSetLayoutSupportKHR(const VULKAN_HPP_NAMESPACE::DescriptorSetLayoutCreateInfo &createInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport &support = structureChain.template get<VULKAN_HPP_NAMESPACE::DescriptorSetLayoutSupport>(); |
| d.vkGetDescriptorSetLayoutSupportKHR( |
| m_device, reinterpret_cast<const VkDescriptorSetLayoutCreateInfo *>(&createInfo), reinterpret_cast<VkDescriptorSetLayoutSupport *>(&support)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_draw_indirect_count === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndirectCountKHR(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::Buffer countBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize countBufferOffset, |
| uint32_t maxDrawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndirectCountKHR(m_commandBuffer, |
| static_cast<VkBuffer>(buffer), |
| static_cast<VkDeviceSize>(offset), |
| static_cast<VkBuffer>(countBuffer), |
| static_cast<VkDeviceSize>(countBufferOffset), |
| maxDrawCount, |
| stride); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawIndexedIndirectCountKHR(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::Buffer countBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize countBufferOffset, |
| uint32_t maxDrawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawIndexedIndirectCountKHR(m_commandBuffer, |
| static_cast<VkBuffer>(buffer), |
| static_cast<VkDeviceSize>(offset), |
| static_cast<VkBuffer>(countBuffer), |
| static_cast<VkDeviceSize>(countBufferOffset), |
| maxDrawCount, |
| stride); |
| } |
| |
| //=== VK_EXT_external_memory_host === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getMemoryHostPointerPropertiesEXT(VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagBits handleType, |
| const void *pHostPointer, |
| VULKAN_HPP_NAMESPACE::MemoryHostPointerPropertiesEXT *pMemoryHostPointerProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetMemoryHostPointerPropertiesEXT(m_device, |
| static_cast<VkExternalMemoryHandleTypeFlagBits>(handleType), |
| pHostPointer, |
| reinterpret_cast<VkMemoryHostPointerPropertiesEXT *>(pMemoryHostPointerProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::MemoryHostPointerPropertiesEXT>::type |
| Device::getMemoryHostPointerPropertiesEXT(VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagBits handleType, |
| const void *pHostPointer, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryHostPointerPropertiesEXT memoryHostPointerProperties; |
| VkResult result = d.vkGetMemoryHostPointerPropertiesEXT(m_device, |
| static_cast<VkExternalMemoryHandleTypeFlagBits>(handleType), |
| pHostPointer, |
| reinterpret_cast<VkMemoryHostPointerPropertiesEXT *>(&memoryHostPointerProperties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryHostPointerPropertiesEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), memoryHostPointerProperties); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_AMD_buffer_marker === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::writeBufferMarkerAMD(VULKAN_HPP_NAMESPACE::PipelineStageFlagBits pipelineStage, |
| VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize dstOffset, |
| uint32_t marker, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWriteBufferMarkerAMD(m_commandBuffer, |
| static_cast<VkPipelineStageFlagBits>(pipelineStage), |
| static_cast<VkBuffer>(dstBuffer), |
| static_cast<VkDeviceSize>(dstOffset), |
| marker); |
| } |
| |
| //=== VK_EXT_calibrated_timestamps === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getCalibrateableTimeDomainsEXT(uint32_t *pTimeDomainCount, |
| VULKAN_HPP_NAMESPACE::TimeDomainEXT *pTimeDomains, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT(m_physicalDevice, pTimeDomainCount, reinterpret_cast<VkTimeDomainEXT *>(pTimeDomains))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename TimeDomainEXTAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<TimeDomainEXT, TimeDomainEXTAllocator>>::type |
| PhysicalDevice::getCalibrateableTimeDomainsEXT(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<TimeDomainEXT, TimeDomainEXTAllocator> timeDomains; |
| uint32_t timeDomainCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT(m_physicalDevice, &timeDomainCount, nullptr)); |
| if((result == Result::eSuccess) && timeDomainCount) |
| { |
| timeDomains.resize(timeDomainCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT(m_physicalDevice, &timeDomainCount, reinterpret_cast<VkTimeDomainEXT *>(timeDomains.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(timeDomainCount <= timeDomains.size()); |
| if(timeDomainCount < timeDomains.size()) |
| { |
| timeDomains.resize(timeDomainCount); |
| } |
| } |
| return createResultValue(result, timeDomains, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getCalibrateableTimeDomainsEXT"); |
| } |
| |
| template<typename TimeDomainEXTAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, TimeDomainEXT>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<TimeDomainEXT, TimeDomainEXTAllocator>>::type |
| PhysicalDevice::getCalibrateableTimeDomainsEXT(TimeDomainEXTAllocator &timeDomainEXTAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<TimeDomainEXT, TimeDomainEXTAllocator> timeDomains(timeDomainEXTAllocator); |
| uint32_t timeDomainCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT(m_physicalDevice, &timeDomainCount, nullptr)); |
| if((result == Result::eSuccess) && timeDomainCount) |
| { |
| timeDomains.resize(timeDomainCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceCalibrateableTimeDomainsEXT(m_physicalDevice, &timeDomainCount, reinterpret_cast<VkTimeDomainEXT *>(timeDomains.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(timeDomainCount <= timeDomains.size()); |
| if(timeDomainCount < timeDomains.size()) |
| { |
| timeDomains.resize(timeDomainCount); |
| } |
| } |
| return createResultValue(result, timeDomains, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getCalibrateableTimeDomainsEXT"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getCalibratedTimestampsEXT(uint32_t timestampCount, |
| const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoEXT *pTimestampInfos, |
| uint64_t *pTimestamps, |
| uint64_t *pMaxDeviation, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetCalibratedTimestampsEXT( |
| m_device, timestampCount, reinterpret_cast<const VkCalibratedTimestampInfoEXT *>(pTimestampInfos), pTimestamps, pMaxDeviation)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Uint64_tAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::pair<std::vector<uint64_t, Uint64_tAllocator>, uint64_t>>::type |
| Device::getCalibratedTimestampsEXT(ArrayProxy<const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoEXT> const ×tampInfos, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::pair<std::vector<uint64_t, Uint64_tAllocator>, uint64_t> data( |
| std::piecewise_construct, std::forward_as_tuple(timestampInfos.size()), std::forward_as_tuple(0)); |
| std::vector<uint64_t, Uint64_tAllocator> ×tamps = data.first; |
| uint64_t &maxDeviation = data.second; |
| Result result = static_cast<Result>(d.vkGetCalibratedTimestampsEXT( |
| m_device, timestampInfos.size(), reinterpret_cast<const VkCalibratedTimestampInfoEXT *>(timestampInfos.data()), timestamps.data(), &maxDeviation)); |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getCalibratedTimestampsEXT"); |
| } |
| |
| template<typename Uint64_tAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, uint64_t>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::pair<std::vector<uint64_t, Uint64_tAllocator>, uint64_t>>::type |
| Device::getCalibratedTimestampsEXT(ArrayProxy<const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoEXT> const ×tampInfos, |
| Uint64_tAllocator &uint64_tAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::pair<std::vector<uint64_t, Uint64_tAllocator>, uint64_t> data( |
| std::piecewise_construct, std::forward_as_tuple(timestampInfos.size(), uint64_tAllocator), std::forward_as_tuple(0)); |
| std::vector<uint64_t, Uint64_tAllocator> ×tamps = data.first; |
| uint64_t &maxDeviation = data.second; |
| Result result = static_cast<Result>(d.vkGetCalibratedTimestampsEXT( |
| m_device, timestampInfos.size(), reinterpret_cast<const VkCalibratedTimestampInfoEXT *>(timestampInfos.data()), timestamps.data(), &maxDeviation)); |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getCalibratedTimestampsEXT"); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::pair<uint64_t, uint64_t>>::type |
| Device::getCalibratedTimestampEXT(const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoEXT ×tampInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::pair<uint64_t, uint64_t> data; |
| uint64_t ×tamp = data.first; |
| uint64_t &maxDeviation = data.second; |
| Result result = static_cast<Result>( |
| d.vkGetCalibratedTimestampsEXT(m_device, 1, reinterpret_cast<const VkCalibratedTimestampInfoEXT *>(×tampInfo), ×tamp, &maxDeviation)); |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getCalibratedTimestampEXT"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NV_mesh_shader === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawMeshTasksNV(uint32_t taskCount, uint32_t firstTask, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawMeshTasksNV(m_commandBuffer, taskCount, firstTask); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawMeshTasksIndirectNV(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| uint32_t drawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawMeshTasksIndirectNV(m_commandBuffer, static_cast<VkBuffer>(buffer), static_cast<VkDeviceSize>(offset), drawCount, stride); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawMeshTasksIndirectCountNV(VULKAN_HPP_NAMESPACE::Buffer buffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize offset, |
| VULKAN_HPP_NAMESPACE::Buffer countBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize countBufferOffset, |
| uint32_t maxDrawCount, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawMeshTasksIndirectCountNV(m_commandBuffer, |
| static_cast<VkBuffer>(buffer), |
| static_cast<VkDeviceSize>(offset), |
| static_cast<VkBuffer>(countBuffer), |
| static_cast<VkDeviceSize>(countBufferOffset), |
| maxDrawCount, |
| stride); |
| } |
| |
| //=== VK_NV_scissor_exclusive === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setExclusiveScissorNV(uint32_t firstExclusiveScissor, |
| uint32_t exclusiveScissorCount, |
| const VULKAN_HPP_NAMESPACE::Rect2D *pExclusiveScissors, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetExclusiveScissorNV(m_commandBuffer, firstExclusiveScissor, exclusiveScissorCount, reinterpret_cast<const VkRect2D *>(pExclusiveScissors)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setExclusiveScissorNV(uint32_t firstExclusiveScissor, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D> const &exclusiveScissors, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetExclusiveScissorNV( |
| m_commandBuffer, firstExclusiveScissor, exclusiveScissors.size(), reinterpret_cast<const VkRect2D *>(exclusiveScissors.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NV_device_diagnostic_checkpoints === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setCheckpointNV(const void *pCheckpointMarker, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetCheckpointNV(m_commandBuffer, pCheckpointMarker); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename CheckpointMarkerType, typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setCheckpointNV(CheckpointMarkerType const &checkpointMarker, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetCheckpointNV(m_commandBuffer, reinterpret_cast<const void *>(&checkpointMarker)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Queue::getCheckpointDataNV(uint32_t *pCheckpointDataCount, |
| VULKAN_HPP_NAMESPACE::CheckpointDataNV *pCheckpointData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetQueueCheckpointDataNV(m_queue, pCheckpointDataCount, reinterpret_cast<VkCheckpointDataNV *>(pCheckpointData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename CheckpointDataNVAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<CheckpointDataNV, CheckpointDataNVAllocator> Queue::getCheckpointDataNV(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<CheckpointDataNV, CheckpointDataNVAllocator> checkpointData; |
| uint32_t checkpointDataCount; |
| d.vkGetQueueCheckpointDataNV(m_queue, &checkpointDataCount, nullptr); |
| checkpointData.resize(checkpointDataCount); |
| d.vkGetQueueCheckpointDataNV(m_queue, &checkpointDataCount, reinterpret_cast<VkCheckpointDataNV *>(checkpointData.data())); |
| VULKAN_HPP_ASSERT(checkpointDataCount <= checkpointData.size()); |
| return checkpointData; |
| } |
| |
| template<typename CheckpointDataNVAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, CheckpointDataNV>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<CheckpointDataNV, CheckpointDataNVAllocator> |
| Queue::getCheckpointDataNV(CheckpointDataNVAllocator &checkpointDataNVAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<CheckpointDataNV, CheckpointDataNVAllocator> checkpointData(checkpointDataNVAllocator); |
| uint32_t checkpointDataCount; |
| d.vkGetQueueCheckpointDataNV(m_queue, &checkpointDataCount, nullptr); |
| checkpointData.resize(checkpointDataCount); |
| d.vkGetQueueCheckpointDataNV(m_queue, &checkpointDataCount, reinterpret_cast<VkCheckpointDataNV *>(checkpointData.data())); |
| VULKAN_HPP_ASSERT(checkpointDataCount <= checkpointData.size()); |
| return checkpointData; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_timeline_semaphore === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getSemaphoreCounterValueKHR(VULKAN_HPP_NAMESPACE::Semaphore semaphore, |
| uint64_t *pValue, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetSemaphoreCounterValueKHR(m_device, static_cast<VkSemaphore>(semaphore), pValue)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<uint64_t>::type |
| Device::getSemaphoreCounterValueKHR(VULKAN_HPP_NAMESPACE::Semaphore semaphore, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint64_t value; |
| VkResult result = d.vkGetSemaphoreCounterValueKHR(m_device, static_cast<VkSemaphore>(semaphore), &value); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getSemaphoreCounterValueKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), value); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::waitSemaphoresKHR(const VULKAN_HPP_NAMESPACE::SemaphoreWaitInfo *pWaitInfo, |
| uint64_t timeout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkWaitSemaphoresKHR(m_device, reinterpret_cast<const VkSemaphoreWaitInfo *>(pWaitInfo), timeout)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result |
| Device::waitSemaphoresKHR(const VULKAN_HPP_NAMESPACE::SemaphoreWaitInfo &waitInfo, uint64_t timeout, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkWaitSemaphoresKHR(m_device, reinterpret_cast<const VkSemaphoreWaitInfo *>(&waitInfo), timeout); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::waitSemaphoresKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eTimeout }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::signalSemaphoreKHR(const VULKAN_HPP_NAMESPACE::SemaphoreSignalInfo *pSignalInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkSignalSemaphoreKHR(m_device, reinterpret_cast<const VkSemaphoreSignalInfo *>(pSignalInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::signalSemaphoreKHR(const VULKAN_HPP_NAMESPACE::SemaphoreSignalInfo &signalInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkSignalSemaphoreKHR(m_device, reinterpret_cast<const VkSemaphoreSignalInfo *>(&signalInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::signalSemaphoreKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_INTEL_performance_query === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::initializePerformanceApiINTEL( |
| const VULKAN_HPP_NAMESPACE::InitializePerformanceApiInfoINTEL *pInitializeInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkInitializePerformanceApiINTEL(m_device, reinterpret_cast<const VkInitializePerformanceApiInfoINTEL *>(pInitializeInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::initializePerformanceApiINTEL(const VULKAN_HPP_NAMESPACE::InitializePerformanceApiInfoINTEL &initializeInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkInitializePerformanceApiINTEL(m_device, reinterpret_cast<const VkInitializePerformanceApiInfoINTEL *>(&initializeInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::initializePerformanceApiINTEL"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::uninitializePerformanceApiINTEL(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkUninitializePerformanceApiINTEL(m_device); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result CommandBuffer::setPerformanceMarkerINTEL(const VULKAN_HPP_NAMESPACE::PerformanceMarkerInfoINTEL *pMarkerInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCmdSetPerformanceMarkerINTEL(m_commandBuffer, reinterpret_cast<const VkPerformanceMarkerInfoINTEL *>(pMarkerInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| CommandBuffer::setPerformanceMarkerINTEL(const VULKAN_HPP_NAMESPACE::PerformanceMarkerInfoINTEL &markerInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkCmdSetPerformanceMarkerINTEL(m_commandBuffer, reinterpret_cast<const VkPerformanceMarkerInfoINTEL *>(&markerInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::setPerformanceMarkerINTEL"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result CommandBuffer::setPerformanceStreamMarkerINTEL( |
| const VULKAN_HPP_NAMESPACE::PerformanceStreamMarkerInfoINTEL *pMarkerInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkCmdSetPerformanceStreamMarkerINTEL(m_commandBuffer, reinterpret_cast<const VkPerformanceStreamMarkerInfoINTEL *>(pMarkerInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| CommandBuffer::setPerformanceStreamMarkerINTEL(const VULKAN_HPP_NAMESPACE::PerformanceStreamMarkerInfoINTEL &markerInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkCmdSetPerformanceStreamMarkerINTEL(m_commandBuffer, reinterpret_cast<const VkPerformanceStreamMarkerInfoINTEL *>(&markerInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::setPerformanceStreamMarkerINTEL"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result CommandBuffer::setPerformanceOverrideINTEL( |
| const VULKAN_HPP_NAMESPACE::PerformanceOverrideInfoINTEL *pOverrideInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkCmdSetPerformanceOverrideINTEL(m_commandBuffer, reinterpret_cast<const VkPerformanceOverrideInfoINTEL *>(pOverrideInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| CommandBuffer::setPerformanceOverrideINTEL(const VULKAN_HPP_NAMESPACE::PerformanceOverrideInfoINTEL &overrideInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkCmdSetPerformanceOverrideINTEL(m_commandBuffer, reinterpret_cast<const VkPerformanceOverrideInfoINTEL *>(&overrideInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::setPerformanceOverrideINTEL"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::acquirePerformanceConfigurationINTEL(const VULKAN_HPP_NAMESPACE::PerformanceConfigurationAcquireInfoINTEL *pAcquireInfo, |
| VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL *pConfiguration, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAcquirePerformanceConfigurationINTEL(m_device, |
| reinterpret_cast<const VkPerformanceConfigurationAcquireInfoINTEL *>(pAcquireInfo), |
| reinterpret_cast<VkPerformanceConfigurationINTEL *>(pConfiguration))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL>::type |
| Device::acquirePerformanceConfigurationINTEL(const VULKAN_HPP_NAMESPACE::PerformanceConfigurationAcquireInfoINTEL &acquireInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration; |
| VkResult result = d.vkAcquirePerformanceConfigurationINTEL(m_device, |
| reinterpret_cast<const VkPerformanceConfigurationAcquireInfoINTEL *>(&acquireInfo), |
| reinterpret_cast<VkPerformanceConfigurationINTEL *>(&configuration)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::acquirePerformanceConfigurationINTEL"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), configuration); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL, Dispatch>>::type |
| Device::acquirePerformanceConfigurationINTELUnique(const VULKAN_HPP_NAMESPACE::PerformanceConfigurationAcquireInfoINTEL &acquireInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration; |
| Result result = |
| static_cast<Result>(d.vkAcquirePerformanceConfigurationINTEL(m_device, |
| reinterpret_cast<const VkPerformanceConfigurationAcquireInfoINTEL *>(&acquireInfo), |
| reinterpret_cast<VkPerformanceConfigurationINTEL *>(&configuration))); |
| ObjectRelease<Device, Dispatch> deleter(*this, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL, Dispatch>( |
| result, configuration, VULKAN_HPP_NAMESPACE_STRING "::Device::acquirePerformanceConfigurationINTELUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::releasePerformanceConfigurationINTEL(VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkReleasePerformanceConfigurationINTEL(m_device, static_cast<VkPerformanceConfigurationINTEL>(configuration))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::releasePerformanceConfigurationINTEL(VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkReleasePerformanceConfigurationINTEL(m_device, static_cast<VkPerformanceConfigurationINTEL>(configuration)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::releasePerformanceConfigurationINTEL"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::release(VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkReleasePerformanceConfigurationINTEL(m_device, static_cast<VkPerformanceConfigurationINTEL>(configuration))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::release(VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkReleasePerformanceConfigurationINTEL(m_device, static_cast<VkPerformanceConfigurationINTEL>(configuration)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::release"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Queue::setPerformanceConfigurationINTEL(VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkQueueSetPerformanceConfigurationINTEL(m_queue, static_cast<VkPerformanceConfigurationINTEL>(configuration))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Queue::setPerformanceConfigurationINTEL(VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkQueueSetPerformanceConfigurationINTEL(m_queue, static_cast<VkPerformanceConfigurationINTEL>(configuration)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Queue::setPerformanceConfigurationINTEL"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getPerformanceParameterINTEL(VULKAN_HPP_NAMESPACE::PerformanceParameterTypeINTEL parameter, |
| VULKAN_HPP_NAMESPACE::PerformanceValueINTEL *pValue, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPerformanceParameterINTEL( |
| m_device, static_cast<VkPerformanceParameterTypeINTEL>(parameter), reinterpret_cast<VkPerformanceValueINTEL *>(pValue))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::PerformanceValueINTEL>::type |
| Device::getPerformanceParameterINTEL(VULKAN_HPP_NAMESPACE::PerformanceParameterTypeINTEL parameter, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PerformanceValueINTEL value; |
| VkResult result = d.vkGetPerformanceParameterINTEL( |
| m_device, static_cast<VkPerformanceParameterTypeINTEL>(parameter), reinterpret_cast<VkPerformanceValueINTEL *>(&value)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getPerformanceParameterINTEL"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), value); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_AMD_display_native_hdr === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::setLocalDimmingAMD(VULKAN_HPP_NAMESPACE::SwapchainKHR swapChain, |
| VULKAN_HPP_NAMESPACE::Bool32 localDimmingEnable, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkSetLocalDimmingAMD(m_device, static_cast<VkSwapchainKHR>(swapChain), static_cast<VkBool32>(localDimmingEnable)); |
| } |
| |
| #if defined(VK_USE_PLATFORM_FUCHSIA) |
| //=== VK_FUCHSIA_imagepipe_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Instance::createImagePipeSurfaceFUCHSIA(const VULKAN_HPP_NAMESPACE::ImagePipeSurfaceCreateInfoFUCHSIA *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateImagePipeSurfaceFUCHSIA(m_instance, |
| reinterpret_cast<const VkImagePipeSurfaceCreateInfoFUCHSIA *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createImagePipeSurfaceFUCHSIA(const VULKAN_HPP_NAMESPACE::ImagePipeSurfaceCreateInfoFUCHSIA &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = d.vkCreateImagePipeSurfaceFUCHSIA( |
| m_instance, |
| reinterpret_cast<const VkImagePipeSurfaceCreateInfoFUCHSIA *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createImagePipeSurfaceFUCHSIA"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createImagePipeSurfaceFUCHSIAUnique(const VULKAN_HPP_NAMESPACE::ImagePipeSurfaceCreateInfoFUCHSIA &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>(d.vkCreateImagePipeSurfaceFUCHSIA( |
| m_instance, |
| reinterpret_cast<const VkImagePipeSurfaceCreateInfoFUCHSIA *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createImagePipeSurfaceFUCHSIAUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_FUCHSIA*/ |
| |
| #if defined(VK_USE_PLATFORM_METAL_EXT) |
| //=== VK_EXT_metal_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createMetalSurfaceEXT(const VULKAN_HPP_NAMESPACE::MetalSurfaceCreateInfoEXT *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateMetalSurfaceEXT(m_instance, |
| reinterpret_cast<const VkMetalSurfaceCreateInfoEXT *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createMetalSurfaceEXT(const VULKAN_HPP_NAMESPACE::MetalSurfaceCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = |
| d.vkCreateMetalSurfaceEXT(m_instance, |
| reinterpret_cast<const VkMetalSurfaceCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createMetalSurfaceEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createMetalSurfaceEXTUnique(const VULKAN_HPP_NAMESPACE::MetalSurfaceCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>( |
| d.vkCreateMetalSurfaceEXT(m_instance, |
| reinterpret_cast<const VkMetalSurfaceCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createMetalSurfaceEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_METAL_EXT*/ |
| |
| //=== VK_KHR_fragment_shading_rate === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getFragmentShadingRatesKHR(uint32_t *pFragmentShadingRateCount, |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceFragmentShadingRateKHR *pFragmentShadingRates, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceFragmentShadingRatesKHR( |
| m_physicalDevice, pFragmentShadingRateCount, reinterpret_cast<VkPhysicalDeviceFragmentShadingRateKHR *>(pFragmentShadingRates))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PhysicalDeviceFragmentShadingRateKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE |
| typename ResultValueType<std::vector<PhysicalDeviceFragmentShadingRateKHR, PhysicalDeviceFragmentShadingRateKHRAllocator>>::type |
| PhysicalDevice::getFragmentShadingRatesKHR(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceFragmentShadingRateKHR, PhysicalDeviceFragmentShadingRateKHRAllocator> fragmentShadingRates; |
| uint32_t fragmentShadingRateCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceFragmentShadingRatesKHR(m_physicalDevice, &fragmentShadingRateCount, nullptr)); |
| if((result == Result::eSuccess) && fragmentShadingRateCount) |
| { |
| fragmentShadingRates.resize(fragmentShadingRateCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceFragmentShadingRatesKHR( |
| m_physicalDevice, &fragmentShadingRateCount, reinterpret_cast<VkPhysicalDeviceFragmentShadingRateKHR *>(fragmentShadingRates.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(fragmentShadingRateCount <= fragmentShadingRates.size()); |
| if(fragmentShadingRateCount < fragmentShadingRates.size()) |
| { |
| fragmentShadingRates.resize(fragmentShadingRateCount); |
| } |
| } |
| return createResultValue(result, fragmentShadingRates, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getFragmentShadingRatesKHR"); |
| } |
| |
| template<typename PhysicalDeviceFragmentShadingRateKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PhysicalDeviceFragmentShadingRateKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE |
| typename ResultValueType<std::vector<PhysicalDeviceFragmentShadingRateKHR, PhysicalDeviceFragmentShadingRateKHRAllocator>>::type |
| PhysicalDevice::getFragmentShadingRatesKHR(PhysicalDeviceFragmentShadingRateKHRAllocator &physicalDeviceFragmentShadingRateKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceFragmentShadingRateKHR, PhysicalDeviceFragmentShadingRateKHRAllocator> fragmentShadingRates( |
| physicalDeviceFragmentShadingRateKHRAllocator); |
| uint32_t fragmentShadingRateCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceFragmentShadingRatesKHR(m_physicalDevice, &fragmentShadingRateCount, nullptr)); |
| if((result == Result::eSuccess) && fragmentShadingRateCount) |
| { |
| fragmentShadingRates.resize(fragmentShadingRateCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceFragmentShadingRatesKHR( |
| m_physicalDevice, &fragmentShadingRateCount, reinterpret_cast<VkPhysicalDeviceFragmentShadingRateKHR *>(fragmentShadingRates.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(fragmentShadingRateCount <= fragmentShadingRates.size()); |
| if(fragmentShadingRateCount < fragmentShadingRates.size()) |
| { |
| fragmentShadingRates.resize(fragmentShadingRateCount); |
| } |
| } |
| return createResultValue(result, fragmentShadingRates, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getFragmentShadingRatesKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setFragmentShadingRateKHR(const VULKAN_HPP_NAMESPACE::Extent2D *pFragmentSize, |
| const VULKAN_HPP_NAMESPACE::FragmentShadingRateCombinerOpKHR combinerOps[2], |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetFragmentShadingRateKHR( |
| m_commandBuffer, reinterpret_cast<const VkExtent2D *>(pFragmentSize), reinterpret_cast<const VkFragmentShadingRateCombinerOpKHR *>(combinerOps)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setFragmentShadingRateKHR(const VULKAN_HPP_NAMESPACE::Extent2D &fragmentSize, |
| const VULKAN_HPP_NAMESPACE::FragmentShadingRateCombinerOpKHR combinerOps[2], |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetFragmentShadingRateKHR( |
| m_commandBuffer, reinterpret_cast<const VkExtent2D *>(&fragmentSize), reinterpret_cast<const VkFragmentShadingRateCombinerOpKHR *>(combinerOps)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_buffer_device_address === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE DeviceAddress Device::getBufferAddressEXT(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<DeviceAddress>(d.vkGetBufferDeviceAddressEXT(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(pInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::DeviceAddress Device::getBufferAddressEXT(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkDeviceAddress result = d.vkGetBufferDeviceAddressEXT(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(&info)); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::DeviceAddress>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_tooling_info === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getToolPropertiesEXT(uint32_t *pToolCount, |
| VULKAN_HPP_NAMESPACE::PhysicalDeviceToolProperties *pToolProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPhysicalDeviceToolPropertiesEXT(m_physicalDevice, pToolCount, reinterpret_cast<VkPhysicalDeviceToolProperties *>(pToolProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PhysicalDeviceToolPropertiesAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator>>::type |
| PhysicalDevice::getToolPropertiesEXT(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator> toolProperties; |
| uint32_t toolCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceToolPropertiesEXT(m_physicalDevice, &toolCount, nullptr)); |
| if((result == Result::eSuccess) && toolCount) |
| { |
| toolProperties.resize(toolCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceToolPropertiesEXT(m_physicalDevice, &toolCount, reinterpret_cast<VkPhysicalDeviceToolProperties *>(toolProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(toolCount <= toolProperties.size()); |
| if(toolCount < toolProperties.size()) |
| { |
| toolProperties.resize(toolCount); |
| } |
| } |
| return createResultValue(result, toolProperties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getToolPropertiesEXT"); |
| } |
| |
| template<typename PhysicalDeviceToolPropertiesAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PhysicalDeviceToolProperties>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator>>::type |
| PhysicalDevice::getToolPropertiesEXT(PhysicalDeviceToolPropertiesAllocator &physicalDeviceToolPropertiesAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PhysicalDeviceToolProperties, PhysicalDeviceToolPropertiesAllocator> toolProperties(physicalDeviceToolPropertiesAllocator); |
| uint32_t toolCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceToolPropertiesEXT(m_physicalDevice, &toolCount, nullptr)); |
| if((result == Result::eSuccess) && toolCount) |
| { |
| toolProperties.resize(toolCount); |
| result = static_cast<Result>( |
| d.vkGetPhysicalDeviceToolPropertiesEXT(m_physicalDevice, &toolCount, reinterpret_cast<VkPhysicalDeviceToolProperties *>(toolProperties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(toolCount <= toolProperties.size()); |
| if(toolCount < toolProperties.size()) |
| { |
| toolProperties.resize(toolCount); |
| } |
| } |
| return createResultValue(result, toolProperties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getToolPropertiesEXT"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_present_wait === |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::waitForPresentKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| uint64_t presentId, |
| uint64_t timeout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkWaitForPresentKHR(m_device, static_cast<VkSwapchainKHR>(swapchain), presentId, timeout)); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result |
| Device::waitForPresentKHR(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, uint64_t presentId, uint64_t timeout, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkWaitForPresentKHR(m_device, static_cast<VkSwapchainKHR>(swapchain), presentId, timeout); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::waitForPresentKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eTimeout, VULKAN_HPP_NAMESPACE::Result::eSuboptimalKHR }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NV_cooperative_matrix === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getCooperativeMatrixPropertiesNV( |
| uint32_t *pPropertyCount, VULKAN_HPP_NAMESPACE::CooperativeMatrixPropertiesNV *pProperties, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceCooperativeMatrixPropertiesNV( |
| m_physicalDevice, pPropertyCount, reinterpret_cast<VkCooperativeMatrixPropertiesNV *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename CooperativeMatrixPropertiesNVAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<CooperativeMatrixPropertiesNV, CooperativeMatrixPropertiesNVAllocator>>::type |
| PhysicalDevice::getCooperativeMatrixPropertiesNV(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<CooperativeMatrixPropertiesNV, CooperativeMatrixPropertiesNVAllocator> properties; |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceCooperativeMatrixPropertiesNV(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceCooperativeMatrixPropertiesNV( |
| m_physicalDevice, &propertyCount, reinterpret_cast<VkCooperativeMatrixPropertiesNV *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getCooperativeMatrixPropertiesNV"); |
| } |
| |
| template<typename CooperativeMatrixPropertiesNVAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, CooperativeMatrixPropertiesNV>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<CooperativeMatrixPropertiesNV, CooperativeMatrixPropertiesNVAllocator>>::type |
| PhysicalDevice::getCooperativeMatrixPropertiesNV(CooperativeMatrixPropertiesNVAllocator &cooperativeMatrixPropertiesNVAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<CooperativeMatrixPropertiesNV, CooperativeMatrixPropertiesNVAllocator> properties(cooperativeMatrixPropertiesNVAllocator); |
| uint32_t propertyCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceCooperativeMatrixPropertiesNV(m_physicalDevice, &propertyCount, nullptr)); |
| if((result == Result::eSuccess) && propertyCount) |
| { |
| properties.resize(propertyCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceCooperativeMatrixPropertiesNV( |
| m_physicalDevice, &propertyCount, reinterpret_cast<VkCooperativeMatrixPropertiesNV *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(propertyCount <= properties.size()); |
| if(propertyCount < properties.size()) |
| { |
| properties.resize(propertyCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getCooperativeMatrixPropertiesNV"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NV_coverage_reduction_mode === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getSupportedFramebufferMixedSamplesCombinationsNV( |
| uint32_t *pCombinationCount, VULKAN_HPP_NAMESPACE::FramebufferMixedSamplesCombinationNV *pCombinations, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV( |
| m_physicalDevice, pCombinationCount, reinterpret_cast<VkFramebufferMixedSamplesCombinationNV *>(pCombinations))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename FramebufferMixedSamplesCombinationNVAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE |
| typename ResultValueType<std::vector<FramebufferMixedSamplesCombinationNV, FramebufferMixedSamplesCombinationNVAllocator>>::type |
| PhysicalDevice::getSupportedFramebufferMixedSamplesCombinationsNV(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<FramebufferMixedSamplesCombinationNV, FramebufferMixedSamplesCombinationNVAllocator> combinations; |
| uint32_t combinationCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV(m_physicalDevice, &combinationCount, nullptr)); |
| if((result == Result::eSuccess) && combinationCount) |
| { |
| combinations.resize(combinationCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV( |
| m_physicalDevice, &combinationCount, reinterpret_cast<VkFramebufferMixedSamplesCombinationNV *>(combinations.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(combinationCount <= combinations.size()); |
| if(combinationCount < combinations.size()) |
| { |
| combinations.resize(combinationCount); |
| } |
| } |
| return createResultValue(result, combinations, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSupportedFramebufferMixedSamplesCombinationsNV"); |
| } |
| |
| template<typename FramebufferMixedSamplesCombinationNVAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, FramebufferMixedSamplesCombinationNV>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE |
| typename ResultValueType<std::vector<FramebufferMixedSamplesCombinationNV, FramebufferMixedSamplesCombinationNVAllocator>>::type |
| PhysicalDevice::getSupportedFramebufferMixedSamplesCombinationsNV( |
| FramebufferMixedSamplesCombinationNVAllocator &framebufferMixedSamplesCombinationNVAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<FramebufferMixedSamplesCombinationNV, FramebufferMixedSamplesCombinationNVAllocator> combinations( |
| framebufferMixedSamplesCombinationNVAllocator); |
| uint32_t combinationCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV(m_physicalDevice, &combinationCount, nullptr)); |
| if((result == Result::eSuccess) && combinationCount) |
| { |
| combinations.resize(combinationCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSupportedFramebufferMixedSamplesCombinationsNV( |
| m_physicalDevice, &combinationCount, reinterpret_cast<VkFramebufferMixedSamplesCombinationNV *>(combinations.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(combinationCount <= combinations.size()); |
| if(combinationCount < combinations.size()) |
| { |
| combinations.resize(combinationCount); |
| } |
| } |
| return createResultValue(result, combinations, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSupportedFramebufferMixedSamplesCombinationsNV"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_WIN32_KHR) |
| //=== VK_EXT_full_screen_exclusive === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| PhysicalDevice::getSurfacePresentModes2EXT(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR *pSurfaceInfo, |
| uint32_t *pPresentModeCount, |
| VULKAN_HPP_NAMESPACE::PresentModeKHR *pPresentModes, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPhysicalDeviceSurfacePresentModes2EXT(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(pSurfaceInfo), |
| pPresentModeCount, |
| reinterpret_cast<VkPresentModeKHR *>(pPresentModes))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PresentModeKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PresentModeKHR, PresentModeKHRAllocator>>::type |
| PhysicalDevice::getSurfacePresentModes2EXT(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR &surfaceInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PresentModeKHR, PresentModeKHRAllocator> presentModes; |
| uint32_t presentModeCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfacePresentModes2EXT( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), &presentModeCount, nullptr)); |
| if((result == Result::eSuccess) && presentModeCount) |
| { |
| presentModes.resize(presentModeCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfacePresentModes2EXT(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), |
| &presentModeCount, |
| reinterpret_cast<VkPresentModeKHR *>(presentModes.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(presentModeCount <= presentModes.size()); |
| if(presentModeCount < presentModes.size()) |
| { |
| presentModes.resize(presentModeCount); |
| } |
| } |
| return createResultValue(result, presentModes, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfacePresentModes2EXT"); |
| } |
| |
| template<typename PresentModeKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PresentModeKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PresentModeKHR, PresentModeKHRAllocator>>::type |
| PhysicalDevice::getSurfacePresentModes2EXT(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR &surfaceInfo, |
| PresentModeKHRAllocator &presentModeKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PresentModeKHR, PresentModeKHRAllocator> presentModes(presentModeKHRAllocator); |
| uint32_t presentModeCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfacePresentModes2EXT( |
| m_physicalDevice, reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), &presentModeCount, nullptr)); |
| if((result == Result::eSuccess) && presentModeCount) |
| { |
| presentModes.resize(presentModeCount); |
| result = static_cast<Result>(d.vkGetPhysicalDeviceSurfacePresentModes2EXT(m_physicalDevice, |
| reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), |
| &presentModeCount, |
| reinterpret_cast<VkPresentModeKHR *>(presentModes.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(presentModeCount <= presentModes.size()); |
| if(presentModeCount < presentModes.size()) |
| { |
| presentModes.resize(presentModeCount); |
| } |
| } |
| return createResultValue(result, presentModes, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getSurfacePresentModes2EXT"); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| # ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::acquireFullScreenExclusiveModeEXT(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAcquireFullScreenExclusiveModeEXT(m_device, static_cast<VkSwapchainKHR>(swapchain))); |
| } |
| # else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::acquireFullScreenExclusiveModeEXT(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkAcquireFullScreenExclusiveModeEXT(m_device, static_cast<VkSwapchainKHR>(swapchain)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::acquireFullScreenExclusiveModeEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| # ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::releaseFullScreenExclusiveModeEXT(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkReleaseFullScreenExclusiveModeEXT(m_device, static_cast<VkSwapchainKHR>(swapchain))); |
| } |
| # else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::releaseFullScreenExclusiveModeEXT(VULKAN_HPP_NAMESPACE::SwapchainKHR swapchain, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkReleaseFullScreenExclusiveModeEXT(m_device, static_cast<VkSwapchainKHR>(swapchain)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::releaseFullScreenExclusiveModeEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getGroupSurfacePresentModes2EXT(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR *pSurfaceInfo, |
| VULKAN_HPP_NAMESPACE::DeviceGroupPresentModeFlagsKHR *pModes, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetDeviceGroupSurfacePresentModes2EXT( |
| m_device, reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(pSurfaceInfo), reinterpret_cast<VkDeviceGroupPresentModeFlagsKHR *>(pModes))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DeviceGroupPresentModeFlagsKHR>::type |
| Device::getGroupSurfacePresentModes2EXT(const VULKAN_HPP_NAMESPACE::PhysicalDeviceSurfaceInfo2KHR &surfaceInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DeviceGroupPresentModeFlagsKHR modes; |
| VkResult result = d.vkGetDeviceGroupSurfacePresentModes2EXT( |
| m_device, reinterpret_cast<const VkPhysicalDeviceSurfaceInfo2KHR *>(&surfaceInfo), reinterpret_cast<VkDeviceGroupPresentModeFlagsKHR *>(&modes)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getGroupSurfacePresentModes2EXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), modes); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_WIN32_KHR*/ |
| |
| //=== VK_EXT_headless_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createHeadlessSurfaceEXT(const VULKAN_HPP_NAMESPACE::HeadlessSurfaceCreateInfoEXT *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateHeadlessSurfaceEXT(m_instance, |
| reinterpret_cast<const VkHeadlessSurfaceCreateInfoEXT *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createHeadlessSurfaceEXT(const VULKAN_HPP_NAMESPACE::HeadlessSurfaceCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = d.vkCreateHeadlessSurfaceEXT( |
| m_instance, |
| reinterpret_cast<const VkHeadlessSurfaceCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createHeadlessSurfaceEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createHeadlessSurfaceEXTUnique(const VULKAN_HPP_NAMESPACE::HeadlessSurfaceCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>(d.vkCreateHeadlessSurfaceEXT( |
| m_instance, |
| reinterpret_cast<const VkHeadlessSurfaceCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createHeadlessSurfaceEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_buffer_device_address === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE DeviceAddress Device::getBufferAddressKHR(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<DeviceAddress>(d.vkGetBufferDeviceAddressKHR(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(pInfo))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::DeviceAddress Device::getBufferAddressKHR(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkDeviceAddress result = d.vkGetBufferDeviceAddressKHR(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(&info)); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::DeviceAddress>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint64_t Device::getBufferOpaqueCaptureAddressKHR(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return d.vkGetBufferOpaqueCaptureAddressKHR(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(pInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint64_t Device::getBufferOpaqueCaptureAddressKHR(const VULKAN_HPP_NAMESPACE::BufferDeviceAddressInfo &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint64_t result = d.vkGetBufferOpaqueCaptureAddressKHR(m_device, reinterpret_cast<const VkBufferDeviceAddressInfo *>(&info)); |
| |
| return result; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint64_t Device::getMemoryOpaqueCaptureAddressKHR(const VULKAN_HPP_NAMESPACE::DeviceMemoryOpaqueCaptureAddressInfo *pInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return d.vkGetDeviceMemoryOpaqueCaptureAddressKHR(m_device, reinterpret_cast<const VkDeviceMemoryOpaqueCaptureAddressInfo *>(pInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint64_t Device::getMemoryOpaqueCaptureAddressKHR(const VULKAN_HPP_NAMESPACE::DeviceMemoryOpaqueCaptureAddressInfo &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint64_t result = d.vkGetDeviceMemoryOpaqueCaptureAddressKHR(m_device, reinterpret_cast<const VkDeviceMemoryOpaqueCaptureAddressInfo *>(&info)); |
| |
| return result; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_line_rasterization === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::setLineStippleEXT(uint32_t lineStippleFactor, uint16_t lineStipplePattern, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetLineStippleEXT(m_commandBuffer, lineStippleFactor, lineStipplePattern); |
| } |
| |
| //=== VK_EXT_host_query_reset === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::resetQueryPoolEXT(VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t firstQuery, |
| uint32_t queryCount, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkResetQueryPoolEXT(m_device, static_cast<VkQueryPool>(queryPool), firstQuery, queryCount); |
| } |
| |
| //=== VK_EXT_extended_dynamic_state === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setCullModeEXT(VULKAN_HPP_NAMESPACE::CullModeFlags cullMode, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetCullModeEXT(m_commandBuffer, static_cast<VkCullModeFlags>(cullMode)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setFrontFaceEXT(VULKAN_HPP_NAMESPACE::FrontFace frontFace, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetFrontFaceEXT(m_commandBuffer, static_cast<VkFrontFace>(frontFace)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setPrimitiveTopologyEXT(VULKAN_HPP_NAMESPACE::PrimitiveTopology primitiveTopology, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetPrimitiveTopologyEXT(m_commandBuffer, static_cast<VkPrimitiveTopology>(primitiveTopology)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setViewportWithCountEXT(uint32_t viewportCount, |
| const VULKAN_HPP_NAMESPACE::Viewport *pViewports, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetViewportWithCountEXT(m_commandBuffer, viewportCount, reinterpret_cast<const VkViewport *>(pViewports)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setViewportWithCountEXT(ArrayProxy<const VULKAN_HPP_NAMESPACE::Viewport> const &viewports, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetViewportWithCountEXT(m_commandBuffer, viewports.size(), reinterpret_cast<const VkViewport *>(viewports.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::setScissorWithCountEXT(uint32_t scissorCount, const VULKAN_HPP_NAMESPACE::Rect2D *pScissors, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetScissorWithCountEXT(m_commandBuffer, scissorCount, reinterpret_cast<const VkRect2D *>(pScissors)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setScissorWithCountEXT(ArrayProxy<const VULKAN_HPP_NAMESPACE::Rect2D> const &scissors, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetScissorWithCountEXT(m_commandBuffer, scissors.size(), reinterpret_cast<const VkRect2D *>(scissors.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindVertexBuffers2EXT(uint32_t firstBinding, |
| uint32_t bindingCount, |
| const VULKAN_HPP_NAMESPACE::Buffer *pBuffers, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pOffsets, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pSizes, |
| const VULKAN_HPP_NAMESPACE::DeviceSize *pStrides, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindVertexBuffers2EXT(m_commandBuffer, |
| firstBinding, |
| bindingCount, |
| reinterpret_cast<const VkBuffer *>(pBuffers), |
| reinterpret_cast<const VkDeviceSize *>(pOffsets), |
| reinterpret_cast<const VkDeviceSize *>(pSizes), |
| reinterpret_cast<const VkDeviceSize *>(pStrides)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindVertexBuffers2EXT(uint32_t firstBinding, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::Buffer> const &buffers, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &offsets, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &sizes, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DeviceSize> const &strides, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(buffers.size() == offsets.size()); |
| VULKAN_HPP_ASSERT(sizes.empty() || buffers.size() == sizes.size()); |
| VULKAN_HPP_ASSERT(strides.empty() || buffers.size() == strides.size()); |
| # else |
| if(buffers.size() != offsets.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2EXT: buffers.size() != offsets.size()"); |
| } |
| if(!sizes.empty() && buffers.size() != sizes.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2EXT: buffers.size() != sizes.size()"); |
| } |
| if(!strides.empty() && buffers.size() != strides.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2EXT: buffers.size() != strides.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdBindVertexBuffers2EXT(m_commandBuffer, |
| firstBinding, |
| buffers.size(), |
| reinterpret_cast<const VkBuffer *>(buffers.data()), |
| reinterpret_cast<const VkDeviceSize *>(offsets.data()), |
| reinterpret_cast<const VkDeviceSize *>(sizes.data()), |
| reinterpret_cast<const VkDeviceSize *>(strides.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthTestEnableEXT(VULKAN_HPP_NAMESPACE::Bool32 depthTestEnable, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthTestEnableEXT(m_commandBuffer, static_cast<VkBool32>(depthTestEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthWriteEnableEXT(VULKAN_HPP_NAMESPACE::Bool32 depthWriteEnable, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthWriteEnableEXT(m_commandBuffer, static_cast<VkBool32>(depthWriteEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthCompareOpEXT(VULKAN_HPP_NAMESPACE::CompareOp depthCompareOp, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthCompareOpEXT(m_commandBuffer, static_cast<VkCompareOp>(depthCompareOp)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthBoundsTestEnableEXT(VULKAN_HPP_NAMESPACE::Bool32 depthBoundsTestEnable, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthBoundsTestEnableEXT(m_commandBuffer, static_cast<VkBool32>(depthBoundsTestEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setStencilTestEnableEXT(VULKAN_HPP_NAMESPACE::Bool32 stencilTestEnable, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetStencilTestEnableEXT(m_commandBuffer, static_cast<VkBool32>(stencilTestEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setStencilOpEXT(VULKAN_HPP_NAMESPACE::StencilFaceFlags faceMask, |
| VULKAN_HPP_NAMESPACE::StencilOp failOp, |
| VULKAN_HPP_NAMESPACE::StencilOp passOp, |
| VULKAN_HPP_NAMESPACE::StencilOp depthFailOp, |
| VULKAN_HPP_NAMESPACE::CompareOp compareOp, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetStencilOpEXT(m_commandBuffer, |
| static_cast<VkStencilFaceFlags>(faceMask), |
| static_cast<VkStencilOp>(failOp), |
| static_cast<VkStencilOp>(passOp), |
| static_cast<VkStencilOp>(depthFailOp), |
| static_cast<VkCompareOp>(compareOp)); |
| } |
| |
| //=== VK_KHR_deferred_host_operations === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createDeferredOperationKHR(const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::DeferredOperationKHR *pDeferredOperation, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDeferredOperationKHR( |
| m_device, reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), reinterpret_cast<VkDeferredOperationKHR *>(pDeferredOperation))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DeferredOperationKHR>::type |
| Device::createDeferredOperationKHR(Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation; |
| VkResult result = d.vkCreateDeferredOperationKHR( |
| m_device, |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDeferredOperationKHR *>(&deferredOperation)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createDeferredOperationKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), deferredOperation); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DeferredOperationKHR, Dispatch>>::type |
| Device::createDeferredOperationKHRUnique(Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation; |
| Result result = static_cast<Result>(d.vkCreateDeferredOperationKHR( |
| m_device, |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkDeferredOperationKHR *>(&deferredOperation))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DeferredOperationKHR, Dispatch>( |
| result, deferredOperation, VULKAN_HPP_NAMESPACE_STRING "::Device::createDeferredOperationKHRUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDeferredOperationKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDeferredOperationKHR( |
| m_device, static_cast<VkDeferredOperationKHR>(operation), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyDeferredOperationKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDeferredOperationKHR( |
| m_device, |
| static_cast<VkDeferredOperationKHR>(operation), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyDeferredOperationKHR( |
| m_device, static_cast<VkDeferredOperationKHR>(operation), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyDeferredOperationKHR( |
| m_device, |
| static_cast<VkDeferredOperationKHR>(operation), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE uint32_t Device::getDeferredOperationMaxConcurrencyKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return d.vkGetDeferredOperationMaxConcurrencyKHR(m_device, static_cast<VkDeferredOperationKHR>(operation)); |
| } |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getDeferredOperationResultKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetDeferredOperationResultKHR(m_device, static_cast<VkDeferredOperationKHR>(operation))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result |
| Device::getDeferredOperationResultKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkGetDeferredOperationResultKHR(m_device, static_cast<VkDeferredOperationKHR>(operation)); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::deferredOperationJoinKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkDeferredOperationJoinKHR(m_device, static_cast<VkDeferredOperationKHR>(operation))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Result Device::deferredOperationJoinKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR operation, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkDeferredOperationJoinKHR(m_device, static_cast<VkDeferredOperationKHR>(operation)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::deferredOperationJoinKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eThreadDoneKHR, VULKAN_HPP_NAMESPACE::Result::eThreadIdleKHR }); |
| return static_cast<VULKAN_HPP_NAMESPACE::Result>(result); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_KHR_pipeline_executable_properties === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getPipelineExecutablePropertiesKHR(const VULKAN_HPP_NAMESPACE::PipelineInfoKHR *pPipelineInfo, |
| uint32_t *pExecutableCount, |
| VULKAN_HPP_NAMESPACE::PipelineExecutablePropertiesKHR *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPipelineExecutablePropertiesKHR(m_device, |
| reinterpret_cast<const VkPipelineInfoKHR *>(pPipelineInfo), |
| pExecutableCount, |
| reinterpret_cast<VkPipelineExecutablePropertiesKHR *>(pProperties))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PipelineExecutablePropertiesKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PipelineExecutablePropertiesKHR, PipelineExecutablePropertiesKHRAllocator>>::type |
| Device::getPipelineExecutablePropertiesKHR(const VULKAN_HPP_NAMESPACE::PipelineInfoKHR &pipelineInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PipelineExecutablePropertiesKHR, PipelineExecutablePropertiesKHRAllocator> properties; |
| uint32_t executableCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>( |
| d.vkGetPipelineExecutablePropertiesKHR(m_device, reinterpret_cast<const VkPipelineInfoKHR *>(&pipelineInfo), &executableCount, nullptr)); |
| if((result == Result::eSuccess) && executableCount) |
| { |
| properties.resize(executableCount); |
| result = static_cast<Result>(d.vkGetPipelineExecutablePropertiesKHR(m_device, |
| reinterpret_cast<const VkPipelineInfoKHR *>(&pipelineInfo), |
| &executableCount, |
| reinterpret_cast<VkPipelineExecutablePropertiesKHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(executableCount <= properties.size()); |
| if(executableCount < properties.size()) |
| { |
| properties.resize(executableCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::Device::getPipelineExecutablePropertiesKHR"); |
| } |
| |
| template<typename PipelineExecutablePropertiesKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PipelineExecutablePropertiesKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PipelineExecutablePropertiesKHR, PipelineExecutablePropertiesKHRAllocator>>::type |
| Device::getPipelineExecutablePropertiesKHR(const VULKAN_HPP_NAMESPACE::PipelineInfoKHR &pipelineInfo, |
| PipelineExecutablePropertiesKHRAllocator &pipelineExecutablePropertiesKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PipelineExecutablePropertiesKHR, PipelineExecutablePropertiesKHRAllocator> properties(pipelineExecutablePropertiesKHRAllocator); |
| uint32_t executableCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>( |
| d.vkGetPipelineExecutablePropertiesKHR(m_device, reinterpret_cast<const VkPipelineInfoKHR *>(&pipelineInfo), &executableCount, nullptr)); |
| if((result == Result::eSuccess) && executableCount) |
| { |
| properties.resize(executableCount); |
| result = static_cast<Result>(d.vkGetPipelineExecutablePropertiesKHR(m_device, |
| reinterpret_cast<const VkPipelineInfoKHR *>(&pipelineInfo), |
| &executableCount, |
| reinterpret_cast<VkPipelineExecutablePropertiesKHR *>(properties.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(executableCount <= properties.size()); |
| if(executableCount < properties.size()) |
| { |
| properties.resize(executableCount); |
| } |
| } |
| return createResultValue(result, properties, VULKAN_HPP_NAMESPACE_STRING "::Device::getPipelineExecutablePropertiesKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getPipelineExecutableStatisticsKHR(const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR *pExecutableInfo, |
| uint32_t *pStatisticCount, |
| VULKAN_HPP_NAMESPACE::PipelineExecutableStatisticKHR *pStatistics, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetPipelineExecutableStatisticsKHR(m_device, |
| reinterpret_cast<const VkPipelineExecutableInfoKHR *>(pExecutableInfo), |
| pStatisticCount, |
| reinterpret_cast<VkPipelineExecutableStatisticKHR *>(pStatistics))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PipelineExecutableStatisticKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PipelineExecutableStatisticKHR, PipelineExecutableStatisticKHRAllocator>>::type |
| Device::getPipelineExecutableStatisticsKHR(const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR &executableInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PipelineExecutableStatisticKHR, PipelineExecutableStatisticKHRAllocator> statistics; |
| uint32_t statisticCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPipelineExecutableStatisticsKHR( |
| m_device, reinterpret_cast<const VkPipelineExecutableInfoKHR *>(&executableInfo), &statisticCount, nullptr)); |
| if((result == Result::eSuccess) && statisticCount) |
| { |
| statistics.resize(statisticCount); |
| result = static_cast<Result>(d.vkGetPipelineExecutableStatisticsKHR(m_device, |
| reinterpret_cast<const VkPipelineExecutableInfoKHR *>(&executableInfo), |
| &statisticCount, |
| reinterpret_cast<VkPipelineExecutableStatisticKHR *>(statistics.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(statisticCount <= statistics.size()); |
| if(statisticCount < statistics.size()) |
| { |
| statistics.resize(statisticCount); |
| } |
| } |
| return createResultValue(result, statistics, VULKAN_HPP_NAMESPACE_STRING "::Device::getPipelineExecutableStatisticsKHR"); |
| } |
| |
| template<typename PipelineExecutableStatisticKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PipelineExecutableStatisticKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<PipelineExecutableStatisticKHR, PipelineExecutableStatisticKHRAllocator>>::type |
| Device::getPipelineExecutableStatisticsKHR(const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR &executableInfo, |
| PipelineExecutableStatisticKHRAllocator &pipelineExecutableStatisticKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PipelineExecutableStatisticKHR, PipelineExecutableStatisticKHRAllocator> statistics(pipelineExecutableStatisticKHRAllocator); |
| uint32_t statisticCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPipelineExecutableStatisticsKHR( |
| m_device, reinterpret_cast<const VkPipelineExecutableInfoKHR *>(&executableInfo), &statisticCount, nullptr)); |
| if((result == Result::eSuccess) && statisticCount) |
| { |
| statistics.resize(statisticCount); |
| result = static_cast<Result>(d.vkGetPipelineExecutableStatisticsKHR(m_device, |
| reinterpret_cast<const VkPipelineExecutableInfoKHR *>(&executableInfo), |
| &statisticCount, |
| reinterpret_cast<VkPipelineExecutableStatisticKHR *>(statistics.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(statisticCount <= statistics.size()); |
| if(statisticCount < statistics.size()) |
| { |
| statistics.resize(statisticCount); |
| } |
| } |
| return createResultValue(result, statistics, VULKAN_HPP_NAMESPACE_STRING "::Device::getPipelineExecutableStatisticsKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getPipelineExecutableInternalRepresentationsKHR(const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR *pExecutableInfo, |
| uint32_t *pInternalRepresentationCount, |
| VULKAN_HPP_NAMESPACE::PipelineExecutableInternalRepresentationKHR *pInternalRepresentations, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetPipelineExecutableInternalRepresentationsKHR(m_device, |
| reinterpret_cast<const VkPipelineExecutableInfoKHR *>(pExecutableInfo), |
| pInternalRepresentationCount, |
| reinterpret_cast<VkPipelineExecutableInternalRepresentationKHR *>(pInternalRepresentations))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PipelineExecutableInternalRepresentationKHRAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE |
| typename ResultValueType<std::vector<PipelineExecutableInternalRepresentationKHR, PipelineExecutableInternalRepresentationKHRAllocator>>::type |
| Device::getPipelineExecutableInternalRepresentationsKHR(const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR &executableInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PipelineExecutableInternalRepresentationKHR, PipelineExecutableInternalRepresentationKHRAllocator> internalRepresentations; |
| uint32_t internalRepresentationCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPipelineExecutableInternalRepresentationsKHR( |
| m_device, reinterpret_cast<const VkPipelineExecutableInfoKHR *>(&executableInfo), &internalRepresentationCount, nullptr)); |
| if((result == Result::eSuccess) && internalRepresentationCount) |
| { |
| internalRepresentations.resize(internalRepresentationCount); |
| result = static_cast<Result>(d.vkGetPipelineExecutableInternalRepresentationsKHR( |
| m_device, |
| reinterpret_cast<const VkPipelineExecutableInfoKHR *>(&executableInfo), |
| &internalRepresentationCount, |
| reinterpret_cast<VkPipelineExecutableInternalRepresentationKHR *>(internalRepresentations.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(internalRepresentationCount <= internalRepresentations.size()); |
| if(internalRepresentationCount < internalRepresentations.size()) |
| { |
| internalRepresentations.resize(internalRepresentationCount); |
| } |
| } |
| return createResultValue(result, internalRepresentations, VULKAN_HPP_NAMESPACE_STRING "::Device::getPipelineExecutableInternalRepresentationsKHR"); |
| } |
| |
| template<typename PipelineExecutableInternalRepresentationKHRAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, PipelineExecutableInternalRepresentationKHR>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE |
| typename ResultValueType<std::vector<PipelineExecutableInternalRepresentationKHR, PipelineExecutableInternalRepresentationKHRAllocator>>::type |
| Device::getPipelineExecutableInternalRepresentationsKHR( |
| const VULKAN_HPP_NAMESPACE::PipelineExecutableInfoKHR &executableInfo, |
| PipelineExecutableInternalRepresentationKHRAllocator &pipelineExecutableInternalRepresentationKHRAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<PipelineExecutableInternalRepresentationKHR, PipelineExecutableInternalRepresentationKHRAllocator> internalRepresentations( |
| pipelineExecutableInternalRepresentationKHRAllocator); |
| uint32_t internalRepresentationCount; |
| Result result; |
| do |
| { |
| result = static_cast<Result>(d.vkGetPipelineExecutableInternalRepresentationsKHR( |
| m_device, reinterpret_cast<const VkPipelineExecutableInfoKHR *>(&executableInfo), &internalRepresentationCount, nullptr)); |
| if((result == Result::eSuccess) && internalRepresentationCount) |
| { |
| internalRepresentations.resize(internalRepresentationCount); |
| result = static_cast<Result>(d.vkGetPipelineExecutableInternalRepresentationsKHR( |
| m_device, |
| reinterpret_cast<const VkPipelineExecutableInfoKHR *>(&executableInfo), |
| &internalRepresentationCount, |
| reinterpret_cast<VkPipelineExecutableInternalRepresentationKHR *>(internalRepresentations.data()))); |
| } |
| } while(result == Result::eIncomplete); |
| if(result == VULKAN_HPP_NAMESPACE::Result::eSuccess) |
| { |
| VULKAN_HPP_ASSERT(internalRepresentationCount <= internalRepresentations.size()); |
| if(internalRepresentationCount < internalRepresentations.size()) |
| { |
| internalRepresentations.resize(internalRepresentationCount); |
| } |
| } |
| return createResultValue(result, internalRepresentations, VULKAN_HPP_NAMESPACE_STRING "::Device::getPipelineExecutableInternalRepresentationsKHR"); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NV_device_generated_commands === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getGeneratedCommandsMemoryRequirementsNV(const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoNV *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetGeneratedCommandsMemoryRequirementsNV(m_device, |
| reinterpret_cast<const VkGeneratedCommandsMemoryRequirementsInfoNV *>(pInfo), |
| reinterpret_cast<VkMemoryRequirements2 *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 |
| Device::getGeneratedCommandsMemoryRequirementsNV(const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoNV &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements; |
| d.vkGetGeneratedCommandsMemoryRequirementsNV(m_device, |
| reinterpret_cast<const VkGeneratedCommandsMemoryRequirementsInfoNV *>(&info), |
| reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getGeneratedCommandsMemoryRequirementsNV(const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoNV &info, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>(); |
| d.vkGetGeneratedCommandsMemoryRequirementsNV(m_device, |
| reinterpret_cast<const VkGeneratedCommandsMemoryRequirementsInfoNV *>(&info), |
| reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::preprocessGeneratedCommandsNV(const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoNV *pGeneratedCommandsInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdPreprocessGeneratedCommandsNV(m_commandBuffer, reinterpret_cast<const VkGeneratedCommandsInfoNV *>(pGeneratedCommandsInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::preprocessGeneratedCommandsNV(const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoNV &generatedCommandsInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdPreprocessGeneratedCommandsNV(m_commandBuffer, reinterpret_cast<const VkGeneratedCommandsInfoNV *>(&generatedCommandsInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::executeGeneratedCommandsNV(VULKAN_HPP_NAMESPACE::Bool32 isPreprocessed, |
| const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoNV *pGeneratedCommandsInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdExecuteGeneratedCommandsNV( |
| m_commandBuffer, static_cast<VkBool32>(isPreprocessed), reinterpret_cast<const VkGeneratedCommandsInfoNV *>(pGeneratedCommandsInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::executeGeneratedCommandsNV(VULKAN_HPP_NAMESPACE::Bool32 isPreprocessed, |
| const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoNV &generatedCommandsInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdExecuteGeneratedCommandsNV( |
| m_commandBuffer, static_cast<VkBool32>(isPreprocessed), reinterpret_cast<const VkGeneratedCommandsInfoNV *>(&generatedCommandsInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindPipelineShaderGroupNV(VULKAN_HPP_NAMESPACE::PipelineBindPoint pipelineBindPoint, |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| uint32_t groupIndex, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindPipelineShaderGroupNV(m_commandBuffer, static_cast<VkPipelineBindPoint>(pipelineBindPoint), static_cast<VkPipeline>(pipeline), groupIndex); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createIndirectCommandsLayoutNV(const VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutCreateInfoNV *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV *pIndirectCommandsLayout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateIndirectCommandsLayoutNV(m_device, |
| reinterpret_cast<const VkIndirectCommandsLayoutCreateInfoNV *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkIndirectCommandsLayoutNV *>(pIndirectCommandsLayout))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV>::type |
| Device::createIndirectCommandsLayoutNV(const VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutCreateInfoNV &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout; |
| VkResult result = d.vkCreateIndirectCommandsLayoutNV( |
| m_device, |
| reinterpret_cast<const VkIndirectCommandsLayoutCreateInfoNV *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkIndirectCommandsLayoutNV *>(&indirectCommandsLayout)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createIndirectCommandsLayoutNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), indirectCommandsLayout); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV, Dispatch>>::type |
| Device::createIndirectCommandsLayoutNVUnique(const VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutCreateInfoNV &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout; |
| Result result = static_cast<Result>(d.vkCreateIndirectCommandsLayoutNV( |
| m_device, |
| reinterpret_cast<const VkIndirectCommandsLayoutCreateInfoNV *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkIndirectCommandsLayoutNV *>(&indirectCommandsLayout))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV, Dispatch>( |
| result, indirectCommandsLayout, VULKAN_HPP_NAMESPACE_STRING "::Device::createIndirectCommandsLayoutNVUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyIndirectCommandsLayoutNV(VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyIndirectCommandsLayoutNV( |
| m_device, static_cast<VkIndirectCommandsLayoutNV>(indirectCommandsLayout), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyIndirectCommandsLayoutNV(VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyIndirectCommandsLayoutNV( |
| m_device, |
| static_cast<VkIndirectCommandsLayoutNV>(indirectCommandsLayout), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyIndirectCommandsLayoutNV( |
| m_device, static_cast<VkIndirectCommandsLayoutNV>(indirectCommandsLayout), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutNV indirectCommandsLayout, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyIndirectCommandsLayoutNV( |
| m_device, |
| static_cast<VkIndirectCommandsLayoutNV>(indirectCommandsLayout), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_acquire_drm_display === |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::acquireDrmDisplayEXT(int32_t drmFd, |
| VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAcquireDrmDisplayEXT(m_physicalDevice, drmFd, static_cast<VkDisplayKHR>(display))); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| PhysicalDevice::acquireDrmDisplayEXT(int32_t drmFd, VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkAcquireDrmDisplayEXT(m_physicalDevice, drmFd, static_cast<VkDisplayKHR>(display)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::acquireDrmDisplayEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getDrmDisplayEXT(int32_t drmFd, |
| uint32_t connectorId, |
| VULKAN_HPP_NAMESPACE::DisplayKHR *display, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetDrmDisplayEXT(m_physicalDevice, drmFd, connectorId, reinterpret_cast<VkDisplayKHR *>(display))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DisplayKHR>::type |
| PhysicalDevice::getDrmDisplayEXT(int32_t drmFd, uint32_t connectorId, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DisplayKHR display; |
| VkResult result = d.vkGetDrmDisplayEXT(m_physicalDevice, drmFd, connectorId, reinterpret_cast<VkDisplayKHR *>(&display)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDrmDisplayEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), display); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DisplayKHR, Dispatch>>::type |
| PhysicalDevice::getDrmDisplayEXTUnique(int32_t drmFd, uint32_t connectorId, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DisplayKHR display; |
| Result result = static_cast<Result>(d.vkGetDrmDisplayEXT(m_physicalDevice, drmFd, connectorId, reinterpret_cast<VkDisplayKHR *>(&display))); |
| ObjectRelease<PhysicalDevice, Dispatch> deleter(*this, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DisplayKHR, Dispatch>( |
| result, display, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getDrmDisplayEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_private_data === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::createPrivateDataSlotEXT(const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot *pPrivateDataSlot, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreatePrivateDataSlotEXT(m_device, |
| reinterpret_cast<const VkPrivateDataSlotCreateInfo *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkPrivateDataSlot *>(pPrivateDataSlot))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::PrivateDataSlot>::type |
| Device::createPrivateDataSlotEXT(const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot; |
| VkResult result = d.vkCreatePrivateDataSlotEXT( |
| m_device, |
| reinterpret_cast<const VkPrivateDataSlotCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPrivateDataSlot *>(&privateDataSlot)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createPrivateDataSlotEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), privateDataSlot); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::PrivateDataSlot, Dispatch>>::type |
| Device::createPrivateDataSlotEXTUnique(const VULKAN_HPP_NAMESPACE::PrivateDataSlotCreateInfo &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot; |
| Result result = static_cast<Result>(d.vkCreatePrivateDataSlotEXT( |
| m_device, |
| reinterpret_cast<const VkPrivateDataSlotCreateInfo *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPrivateDataSlot *>(&privateDataSlot))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::PrivateDataSlot, Dispatch>( |
| result, privateDataSlot, VULKAN_HPP_NAMESPACE_STRING "::Device::createPrivateDataSlotEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPrivateDataSlotEXT(VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyPrivateDataSlotEXT(m_device, static_cast<VkPrivateDataSlot>(privateDataSlot), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyPrivateDataSlotEXT(VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyPrivateDataSlotEXT( |
| m_device, |
| static_cast<VkPrivateDataSlot>(privateDataSlot), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::setPrivateDataEXT(VULKAN_HPP_NAMESPACE::ObjectType objectType, |
| uint64_t objectHandle, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| uint64_t data, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkSetPrivateDataEXT(m_device, static_cast<VkObjectType>(objectType), objectHandle, static_cast<VkPrivateDataSlot>(privateDataSlot), data)); |
| } |
| #else |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE typename ResultValueType<void>::type Device::setPrivateDataEXT(VULKAN_HPP_NAMESPACE::ObjectType objectType, |
| uint64_t objectHandle, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| uint64_t data, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = |
| d.vkSetPrivateDataEXT(m_device, static_cast<VkObjectType>(objectType), objectHandle, static_cast<VkPrivateDataSlot>(privateDataSlot), data); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::setPrivateDataEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getPrivateDataEXT(VULKAN_HPP_NAMESPACE::ObjectType objectType, |
| uint64_t objectHandle, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| uint64_t *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetPrivateDataEXT(m_device, static_cast<VkObjectType>(objectType), objectHandle, static_cast<VkPrivateDataSlot>(privateDataSlot), pData); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE uint64_t Device::getPrivateDataEXT(VULKAN_HPP_NAMESPACE::ObjectType objectType, |
| uint64_t objectHandle, |
| VULKAN_HPP_NAMESPACE::PrivateDataSlot privateDataSlot, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| uint64_t data; |
| d.vkGetPrivateDataEXT(m_device, static_cast<VkObjectType>(objectType), objectHandle, static_cast<VkPrivateDataSlot>(privateDataSlot), &data); |
| |
| return data; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_ENABLE_BETA_EXTENSIONS) |
| //=== VK_KHR_video_encode_queue === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::encodeVideoKHR(const VULKAN_HPP_NAMESPACE::VideoEncodeInfoKHR *pEncodeInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdEncodeVideoKHR(m_commandBuffer, reinterpret_cast<const VkVideoEncodeInfoKHR *>(pEncodeInfo)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::encodeVideoKHR(const VULKAN_HPP_NAMESPACE::VideoEncodeInfoKHR &encodeInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdEncodeVideoKHR(m_commandBuffer, reinterpret_cast<const VkVideoEncodeInfoKHR *>(&encodeInfo)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_ENABLE_BETA_EXTENSIONS*/ |
| |
| //=== VK_KHR_synchronization2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setEvent2KHR(VULKAN_HPP_NAMESPACE::Event event, |
| const VULKAN_HPP_NAMESPACE::DependencyInfo *pDependencyInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetEvent2KHR(m_commandBuffer, static_cast<VkEvent>(event), reinterpret_cast<const VkDependencyInfo *>(pDependencyInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setEvent2KHR(VULKAN_HPP_NAMESPACE::Event event, |
| const VULKAN_HPP_NAMESPACE::DependencyInfo &dependencyInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetEvent2KHR(m_commandBuffer, static_cast<VkEvent>(event), reinterpret_cast<const VkDependencyInfo *>(&dependencyInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resetEvent2KHR(VULKAN_HPP_NAMESPACE::Event event, |
| VULKAN_HPP_NAMESPACE::PipelineStageFlags2 stageMask, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdResetEvent2KHR(m_commandBuffer, static_cast<VkEvent>(event), static_cast<VkPipelineStageFlags2>(stageMask)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::waitEvents2KHR(uint32_t eventCount, |
| const VULKAN_HPP_NAMESPACE::Event *pEvents, |
| const VULKAN_HPP_NAMESPACE::DependencyInfo *pDependencyInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWaitEvents2KHR( |
| m_commandBuffer, eventCount, reinterpret_cast<const VkEvent *>(pEvents), reinterpret_cast<const VkDependencyInfo *>(pDependencyInfos)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::waitEvents2KHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::Event> const &events, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::DependencyInfo> const &dependencyInfos, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT_WHEN_NO_EXCEPTIONS |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| # ifdef VULKAN_HPP_NO_EXCEPTIONS |
| VULKAN_HPP_ASSERT(events.size() == dependencyInfos.size()); |
| # else |
| if(events.size() != dependencyInfos.size()) |
| { |
| throw LogicError(VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::waitEvents2KHR: events.size() != dependencyInfos.size()"); |
| } |
| # endif /*VULKAN_HPP_NO_EXCEPTIONS*/ |
| |
| d.vkCmdWaitEvents2KHR(m_commandBuffer, |
| events.size(), |
| reinterpret_cast<const VkEvent *>(events.data()), |
| reinterpret_cast<const VkDependencyInfo *>(dependencyInfos.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier2KHR(const VULKAN_HPP_NAMESPACE::DependencyInfo *pDependencyInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdPipelineBarrier2KHR(m_commandBuffer, reinterpret_cast<const VkDependencyInfo *>(pDependencyInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::pipelineBarrier2KHR(const VULKAN_HPP_NAMESPACE::DependencyInfo &dependencyInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdPipelineBarrier2KHR(m_commandBuffer, reinterpret_cast<const VkDependencyInfo *>(&dependencyInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::writeTimestamp2KHR(VULKAN_HPP_NAMESPACE::PipelineStageFlags2 stage, |
| VULKAN_HPP_NAMESPACE::QueryPool queryPool, |
| uint32_t query, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWriteTimestamp2KHR(m_commandBuffer, static_cast<VkPipelineStageFlags2>(stage), static_cast<VkQueryPool>(queryPool), query); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Queue::submit2KHR(uint32_t submitCount, |
| const VULKAN_HPP_NAMESPACE::SubmitInfo2 *pSubmits, |
| VULKAN_HPP_NAMESPACE::Fence fence, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkQueueSubmit2KHR(m_queue, submitCount, reinterpret_cast<const VkSubmitInfo2 *>(pSubmits), static_cast<VkFence>(fence))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Queue::submit2KHR(ArrayProxy<const VULKAN_HPP_NAMESPACE::SubmitInfo2> const &submits, VULKAN_HPP_NAMESPACE::Fence fence, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkQueueSubmit2KHR(m_queue, submits.size(), reinterpret_cast<const VkSubmitInfo2 *>(submits.data()), static_cast<VkFence>(fence)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Queue::submit2KHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::writeBufferMarker2AMD(VULKAN_HPP_NAMESPACE::PipelineStageFlags2 stage, |
| VULKAN_HPP_NAMESPACE::Buffer dstBuffer, |
| VULKAN_HPP_NAMESPACE::DeviceSize dstOffset, |
| uint32_t marker, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdWriteBufferMarker2AMD( |
| m_commandBuffer, static_cast<VkPipelineStageFlags2>(stage), static_cast<VkBuffer>(dstBuffer), static_cast<VkDeviceSize>(dstOffset), marker); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Queue::getCheckpointData2NV(uint32_t *pCheckpointDataCount, |
| VULKAN_HPP_NAMESPACE::CheckpointData2NV *pCheckpointData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetQueueCheckpointData2NV(m_queue, pCheckpointDataCount, reinterpret_cast<VkCheckpointData2NV *>(pCheckpointData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename CheckpointData2NVAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<CheckpointData2NV, CheckpointData2NVAllocator> Queue::getCheckpointData2NV(Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<CheckpointData2NV, CheckpointData2NVAllocator> checkpointData; |
| uint32_t checkpointDataCount; |
| d.vkGetQueueCheckpointData2NV(m_queue, &checkpointDataCount, nullptr); |
| checkpointData.resize(checkpointDataCount); |
| d.vkGetQueueCheckpointData2NV(m_queue, &checkpointDataCount, reinterpret_cast<VkCheckpointData2NV *>(checkpointData.data())); |
| VULKAN_HPP_ASSERT(checkpointDataCount <= checkpointData.size()); |
| return checkpointData; |
| } |
| |
| template<typename CheckpointData2NVAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, CheckpointData2NV>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<CheckpointData2NV, CheckpointData2NVAllocator> |
| Queue::getCheckpointData2NV(CheckpointData2NVAllocator &checkpointData2NVAllocator, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<CheckpointData2NV, CheckpointData2NVAllocator> checkpointData(checkpointData2NVAllocator); |
| uint32_t checkpointDataCount; |
| d.vkGetQueueCheckpointData2NV(m_queue, &checkpointDataCount, nullptr); |
| checkpointData.resize(checkpointDataCount); |
| d.vkGetQueueCheckpointData2NV(m_queue, &checkpointDataCount, reinterpret_cast<VkCheckpointData2NV *>(checkpointData.data())); |
| VULKAN_HPP_ASSERT(checkpointDataCount <= checkpointData.size()); |
| return checkpointData; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_NV_fragment_shading_rate_enums === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setFragmentShadingRateEnumNV(VULKAN_HPP_NAMESPACE::FragmentShadingRateNV shadingRate, |
| const VULKAN_HPP_NAMESPACE::FragmentShadingRateCombinerOpKHR combinerOps[2], |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetFragmentShadingRateEnumNV( |
| m_commandBuffer, static_cast<VkFragmentShadingRateNV>(shadingRate), reinterpret_cast<const VkFragmentShadingRateCombinerOpKHR *>(combinerOps)); |
| } |
| |
| //=== VK_KHR_copy_commands2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBuffer2KHR(const VULKAN_HPP_NAMESPACE::CopyBufferInfo2 *pCopyBufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyBuffer2KHR(m_commandBuffer, reinterpret_cast<const VkCopyBufferInfo2 *>(pCopyBufferInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBuffer2KHR(const VULKAN_HPP_NAMESPACE::CopyBufferInfo2 ©BufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyBuffer2KHR(m_commandBuffer, reinterpret_cast<const VkCopyBufferInfo2 *>(©BufferInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImage2KHR(const VULKAN_HPP_NAMESPACE::CopyImageInfo2 *pCopyImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyImage2KHR(m_commandBuffer, reinterpret_cast<const VkCopyImageInfo2 *>(pCopyImageInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImage2KHR(const VULKAN_HPP_NAMESPACE::CopyImageInfo2 ©ImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyImage2KHR(m_commandBuffer, reinterpret_cast<const VkCopyImageInfo2 *>(©ImageInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage2KHR(const VULKAN_HPP_NAMESPACE::CopyBufferToImageInfo2 *pCopyBufferToImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyBufferToImage2KHR(m_commandBuffer, reinterpret_cast<const VkCopyBufferToImageInfo2 *>(pCopyBufferToImageInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyBufferToImage2KHR(const VULKAN_HPP_NAMESPACE::CopyBufferToImageInfo2 ©BufferToImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyBufferToImage2KHR(m_commandBuffer, reinterpret_cast<const VkCopyBufferToImageInfo2 *>(©BufferToImageInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer2KHR(const VULKAN_HPP_NAMESPACE::CopyImageToBufferInfo2 *pCopyImageToBufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdCopyImageToBuffer2KHR(m_commandBuffer, reinterpret_cast<const VkCopyImageToBufferInfo2 *>(pCopyImageToBufferInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::copyImageToBuffer2KHR(const VULKAN_HPP_NAMESPACE::CopyImageToBufferInfo2 ©ImageToBufferInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdCopyImageToBuffer2KHR(m_commandBuffer, reinterpret_cast<const VkCopyImageToBufferInfo2 *>(©ImageToBufferInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::blitImage2KHR(const VULKAN_HPP_NAMESPACE::BlitImageInfo2 *pBlitImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBlitImage2KHR(m_commandBuffer, reinterpret_cast<const VkBlitImageInfo2 *>(pBlitImageInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::blitImage2KHR(const VULKAN_HPP_NAMESPACE::BlitImageInfo2 &blitImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdBlitImage2KHR(m_commandBuffer, reinterpret_cast<const VkBlitImageInfo2 *>(&blitImageInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resolveImage2KHR(const VULKAN_HPP_NAMESPACE::ResolveImageInfo2 *pResolveImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdResolveImage2KHR(m_commandBuffer, reinterpret_cast<const VkResolveImageInfo2 *>(pResolveImageInfo)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::resolveImage2KHR(const VULKAN_HPP_NAMESPACE::ResolveImageInfo2 &resolveImageInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdResolveImage2KHR(m_commandBuffer, reinterpret_cast<const VkResolveImageInfo2 *>(&resolveImageInfo)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_WIN32_KHR) |
| //=== VK_NV_acquire_winrt_display === |
| |
| # ifdef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::acquireWinrtDisplayNV(VULKAN_HPP_NAMESPACE::DisplayKHR display, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkAcquireWinrtDisplayNV(m_physicalDevice, static_cast<VkDisplayKHR>(display))); |
| } |
| # else |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| PhysicalDevice::acquireWinrtDisplayNV(VULKAN_HPP_NAMESPACE::DisplayKHR display, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkAcquireWinrtDisplayNV(m_physicalDevice, static_cast<VkDisplayKHR>(display)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::acquireWinrtDisplayNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result PhysicalDevice::getWinrtDisplayNV(uint32_t deviceRelativeId, |
| VULKAN_HPP_NAMESPACE::DisplayKHR *pDisplay, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetWinrtDisplayNV(m_physicalDevice, deviceRelativeId, reinterpret_cast<VkDisplayKHR *>(pDisplay))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::DisplayKHR>::type |
| PhysicalDevice::getWinrtDisplayNV(uint32_t deviceRelativeId, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DisplayKHR display; |
| VkResult result = d.vkGetWinrtDisplayNV(m_physicalDevice, deviceRelativeId, reinterpret_cast<VkDisplayKHR *>(&display)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getWinrtDisplayNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), display); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::DisplayKHR, Dispatch>>::type |
| PhysicalDevice::getWinrtDisplayNVUnique(uint32_t deviceRelativeId, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::DisplayKHR display; |
| Result result = static_cast<Result>(d.vkGetWinrtDisplayNV(m_physicalDevice, deviceRelativeId, reinterpret_cast<VkDisplayKHR *>(&display))); |
| ObjectRelease<PhysicalDevice, Dispatch> deleter(*this, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::DisplayKHR, Dispatch>( |
| result, display, VULKAN_HPP_NAMESPACE_STRING "::PhysicalDevice::getWinrtDisplayNVUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_WIN32_KHR*/ |
| |
| #if defined(VK_USE_PLATFORM_DIRECTFB_EXT) |
| //=== VK_EXT_directfb_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createDirectFBSurfaceEXT(const VULKAN_HPP_NAMESPACE::DirectFBSurfaceCreateInfoEXT *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateDirectFBSurfaceEXT(m_instance, |
| reinterpret_cast<const VkDirectFBSurfaceCreateInfoEXT *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createDirectFBSurfaceEXT(const VULKAN_HPP_NAMESPACE::DirectFBSurfaceCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = d.vkCreateDirectFBSurfaceEXT( |
| m_instance, |
| reinterpret_cast<const VkDirectFBSurfaceCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createDirectFBSurfaceEXT"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createDirectFBSurfaceEXTUnique(const VULKAN_HPP_NAMESPACE::DirectFBSurfaceCreateInfoEXT &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>(d.vkCreateDirectFBSurfaceEXT( |
| m_instance, |
| reinterpret_cast<const VkDirectFBSurfaceCreateInfoEXT *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createDirectFBSurfaceEXTUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Bool32 PhysicalDevice::getDirectFBPresentationSupportEXT(uint32_t queueFamilyIndex, |
| IDirectFB *dfb, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Bool32>(d.vkGetPhysicalDeviceDirectFBPresentationSupportEXT(m_physicalDevice, queueFamilyIndex, dfb)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Bool32 |
| PhysicalDevice::getDirectFBPresentationSupportEXT(uint32_t queueFamilyIndex, IDirectFB &dfb, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkBool32 result = d.vkGetPhysicalDeviceDirectFBPresentationSupportEXT(m_physicalDevice, queueFamilyIndex, &dfb); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::Bool32>(result); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_DIRECTFB_EXT*/ |
| |
| //=== VK_KHR_ray_tracing_pipeline === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::traceRaysKHR(const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR *pRaygenShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR *pMissShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR *pHitShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR *pCallableShaderBindingTable, |
| uint32_t width, |
| uint32_t height, |
| uint32_t depth, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdTraceRaysKHR(m_commandBuffer, |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(pRaygenShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(pMissShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(pHitShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(pCallableShaderBindingTable), |
| width, |
| height, |
| depth); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::traceRaysKHR(const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR &raygenShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR &missShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR &hitShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR &callableShaderBindingTable, |
| uint32_t width, |
| uint32_t height, |
| uint32_t depth, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdTraceRaysKHR(m_commandBuffer, |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(&raygenShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(&missShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(&hitShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(&callableShaderBindingTable), |
| width, |
| height, |
| depth); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createRayTracingPipelinesKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| uint32_t createInfoCount, |
| const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR *pCreateInfos, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::Pipeline *pPipelines, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateRayTracingPipelinesKHR(m_device, |
| static_cast<VkDeferredOperationKHR>(deferredOperation), |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfoCount, |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoKHR *>(pCreateInfos), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkPipeline *>(pPipelines))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename PipelineAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>> |
| Device::createRayTracingPipelinesKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines(createInfos.size()); |
| VkResult result = d.vkCreateRayTracingPipelinesKHR( |
| m_device, |
| static_cast<VkDeferredOperationKHR>(deferredOperation), |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoKHR *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelinesKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, |
| VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelines); |
| } |
| |
| template<typename PipelineAllocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, Pipeline>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>> |
| Device::createRayTracingPipelinesKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| PipelineAllocator &pipelineAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines(createInfos.size(), pipelineAllocator); |
| VkResult result = d.vkCreateRayTracingPipelinesKHR( |
| m_device, |
| static_cast<VkDeferredOperationKHR>(deferredOperation), |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoKHR *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data())); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelinesKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, |
| VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator>>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipelines); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<VULKAN_HPP_NAMESPACE::Pipeline> |
| Device::createRayTracingPipelineKHR(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline; |
| VkResult result = d.vkCreateRayTracingPipelinesKHR( |
| m_device, |
| static_cast<VkDeferredOperationKHR>(deferredOperation), |
| static_cast<VkPipelineCache>(pipelineCache), |
| 1, |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(&pipeline)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelineKHR", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, |
| VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| return ResultValue<VULKAN_HPP_NAMESPACE::Pipeline>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), pipeline); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch, typename PipelineAllocator> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator>> |
| Device::createRayTracingPipelinesKHRUnique(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator> uniquePipelines; |
| std::vector<Pipeline> pipelines(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateRayTracingPipelinesKHR( |
| m_device, |
| static_cast<VkDeferredOperationKHR>(deferredOperation), |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoKHR *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data()))); |
| if((result == VULKAN_HPP_NAMESPACE::Result::eSuccess) || (result == VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR) || |
| (result == VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR) || (result == VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT)) |
| { |
| uniquePipelines.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniquePipelines.push_back(UniqueHandle<Pipeline, Dispatch>(pipelines[i], deleter)); |
| } |
| } |
| return createResultValue(result, |
| std::move(uniquePipelines), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelinesKHRUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, |
| VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| } |
| |
| template<typename Dispatch, |
| typename PipelineAllocator, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, UniqueHandle<Pipeline, Dispatch>>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator>> |
| Device::createRayTracingPipelinesKHRUnique(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR> const &createInfos, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| PipelineAllocator &pipelineAllocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<UniqueHandle<Pipeline, Dispatch>, PipelineAllocator> uniquePipelines(pipelineAllocator); |
| std::vector<Pipeline> pipelines(createInfos.size()); |
| Result result = static_cast<Result>(d.vkCreateRayTracingPipelinesKHR( |
| m_device, |
| static_cast<VkDeferredOperationKHR>(deferredOperation), |
| static_cast<VkPipelineCache>(pipelineCache), |
| createInfos.size(), |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoKHR *>(createInfos.data()), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(pipelines.data()))); |
| if((result == VULKAN_HPP_NAMESPACE::Result::eSuccess) || (result == VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR) || |
| (result == VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR) || (result == VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT)) |
| { |
| uniquePipelines.reserve(createInfos.size()); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| for(size_t i = 0; i < createInfos.size(); i++) |
| { |
| uniquePipelines.push_back(UniqueHandle<Pipeline, Dispatch>(pipelines[i], deleter)); |
| } |
| } |
| return createResultValue(result, |
| std::move(uniquePipelines), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelinesKHRUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, |
| VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<UniqueHandle<Pipeline, Dispatch>> |
| Device::createRayTracingPipelineKHRUnique(VULKAN_HPP_NAMESPACE::DeferredOperationKHR deferredOperation, |
| VULKAN_HPP_NAMESPACE::PipelineCache pipelineCache, |
| const VULKAN_HPP_NAMESPACE::RayTracingPipelineCreateInfoKHR &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| Pipeline pipeline; |
| Result result = static_cast<Result>(d.vkCreateRayTracingPipelinesKHR( |
| m_device, |
| static_cast<VkDeferredOperationKHR>(deferredOperation), |
| static_cast<VkPipelineCache>(pipelineCache), |
| 1, |
| reinterpret_cast<const VkRayTracingPipelineCreateInfoKHR *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkPipeline *>(&pipeline))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<Pipeline, Dispatch>(result, |
| pipeline, |
| VULKAN_HPP_NAMESPACE_STRING "::Device::createRayTracingPipelineKHRUnique", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, |
| VULKAN_HPP_NAMESPACE::Result::eOperationDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::eOperationNotDeferredKHR, |
| VULKAN_HPP_NAMESPACE::Result::ePipelineCompileRequiredEXT }, |
| deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getRayTracingShaderGroupHandlesKHR(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| uint32_t firstGroup, |
| uint32_t groupCount, |
| size_t dataSize, |
| void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetRayTracingShaderGroupHandlesKHR(m_device, static_cast<VkPipeline>(pipeline), firstGroup, groupCount, dataSize, pData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename DataTypeAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DataType, DataTypeAllocator>>::type Device::getRayTracingShaderGroupHandlesKHR( |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount, size_t dataSize, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_ASSERT(dataSize % sizeof(DataType) == 0); |
| std::vector<DataType, DataTypeAllocator> data(dataSize / sizeof(DataType)); |
| Result result = static_cast<Result>(d.vkGetRayTracingShaderGroupHandlesKHR( |
| m_device, static_cast<VkPipeline>(pipeline), firstGroup, groupCount, data.size() * sizeof(DataType), reinterpret_cast<void *>(data.data()))); |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getRayTracingShaderGroupHandlesKHR"); |
| } |
| |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<DataType>::type |
| Device::getRayTracingShaderGroupHandleKHR(VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| DataType data; |
| VkResult result = d.vkGetRayTracingShaderGroupHandlesKHR( |
| m_device, static_cast<VkPipeline>(pipeline), firstGroup, groupCount, sizeof(DataType), reinterpret_cast<void *>(&data)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getRayTracingShaderGroupHandleKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), data); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getRayTracingCaptureReplayShaderGroupHandlesKHR(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| uint32_t firstGroup, |
| uint32_t groupCount, |
| size_t dataSize, |
| void *pData, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetRayTracingCaptureReplayShaderGroupHandlesKHR(m_device, static_cast<VkPipeline>(pipeline), firstGroup, groupCount, dataSize, pData)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename DataType, typename DataTypeAllocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<std::vector<DataType, DataTypeAllocator>>::type |
| Device::getRayTracingCaptureReplayShaderGroupHandlesKHR( |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount, size_t dataSize, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_ASSERT(dataSize % sizeof(DataType) == 0); |
| std::vector<DataType, DataTypeAllocator> data(dataSize / sizeof(DataType)); |
| Result result = static_cast<Result>(d.vkGetRayTracingCaptureReplayShaderGroupHandlesKHR( |
| m_device, static_cast<VkPipeline>(pipeline), firstGroup, groupCount, data.size() * sizeof(DataType), reinterpret_cast<void *>(data.data()))); |
| return createResultValue(result, data, VULKAN_HPP_NAMESPACE_STRING "::Device::getRayTracingCaptureReplayShaderGroupHandlesKHR"); |
| } |
| |
| template<typename DataType, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<DataType>::type Device::getRayTracingCaptureReplayShaderGroupHandleKHR( |
| VULKAN_HPP_NAMESPACE::Pipeline pipeline, uint32_t firstGroup, uint32_t groupCount, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| DataType data; |
| VkResult result = d.vkGetRayTracingCaptureReplayShaderGroupHandlesKHR( |
| m_device, static_cast<VkPipeline>(pipeline), firstGroup, groupCount, sizeof(DataType), reinterpret_cast<void *>(&data)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getRayTracingCaptureReplayShaderGroupHandleKHR"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), data); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::traceRaysIndirectKHR(const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR *pRaygenShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR *pMissShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR *pHitShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR *pCallableShaderBindingTable, |
| VULKAN_HPP_NAMESPACE::DeviceAddress indirectDeviceAddress, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdTraceRaysIndirectKHR(m_commandBuffer, |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(pRaygenShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(pMissShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(pHitShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(pCallableShaderBindingTable), |
| static_cast<VkDeviceAddress>(indirectDeviceAddress)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::traceRaysIndirectKHR(const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR &raygenShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR &missShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR &hitShaderBindingTable, |
| const VULKAN_HPP_NAMESPACE::StridedDeviceAddressRegionKHR &callableShaderBindingTable, |
| VULKAN_HPP_NAMESPACE::DeviceAddress indirectDeviceAddress, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdTraceRaysIndirectKHR(m_commandBuffer, |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(&raygenShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(&missShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(&hitShaderBindingTable), |
| reinterpret_cast<const VkStridedDeviceAddressRegionKHR *>(&callableShaderBindingTable), |
| static_cast<VkDeviceAddress>(indirectDeviceAddress)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE DeviceSize Device::getRayTracingShaderGroupStackSizeKHR(VULKAN_HPP_NAMESPACE::Pipeline pipeline, |
| uint32_t group, |
| VULKAN_HPP_NAMESPACE::ShaderGroupShaderKHR groupShader, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<DeviceSize>( |
| d.vkGetRayTracingShaderGroupStackSizeKHR(m_device, static_cast<VkPipeline>(pipeline), group, static_cast<VkShaderGroupShaderKHR>(groupShader))); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setRayTracingPipelineStackSizeKHR(uint32_t pipelineStackSize, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetRayTracingPipelineStackSizeKHR(m_commandBuffer, pipelineStackSize); |
| } |
| |
| //=== VK_EXT_vertex_input_dynamic_state === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setVertexInputEXT(uint32_t vertexBindingDescriptionCount, |
| const VULKAN_HPP_NAMESPACE::VertexInputBindingDescription2EXT *pVertexBindingDescriptions, |
| uint32_t vertexAttributeDescriptionCount, |
| const VULKAN_HPP_NAMESPACE::VertexInputAttributeDescription2EXT *pVertexAttributeDescriptions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetVertexInputEXT(m_commandBuffer, |
| vertexBindingDescriptionCount, |
| reinterpret_cast<const VkVertexInputBindingDescription2EXT *>(pVertexBindingDescriptions), |
| vertexAttributeDescriptionCount, |
| reinterpret_cast<const VkVertexInputAttributeDescription2EXT *>(pVertexAttributeDescriptions)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| CommandBuffer::setVertexInputEXT(ArrayProxy<const VULKAN_HPP_NAMESPACE::VertexInputBindingDescription2EXT> const &vertexBindingDescriptions, |
| ArrayProxy<const VULKAN_HPP_NAMESPACE::VertexInputAttributeDescription2EXT> const &vertexAttributeDescriptions, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetVertexInputEXT(m_commandBuffer, |
| vertexBindingDescriptions.size(), |
| reinterpret_cast<const VkVertexInputBindingDescription2EXT *>(vertexBindingDescriptions.data()), |
| vertexAttributeDescriptions.size(), |
| reinterpret_cast<const VkVertexInputAttributeDescription2EXT *>(vertexAttributeDescriptions.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| #if defined(VK_USE_PLATFORM_FUCHSIA) |
| //=== VK_FUCHSIA_external_memory === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getMemoryZirconHandleFUCHSIA(const VULKAN_HPP_NAMESPACE::MemoryGetZirconHandleInfoFUCHSIA *pGetZirconHandleInfo, |
| zx_handle_t *pZirconHandle, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetMemoryZirconHandleFUCHSIA(m_device, reinterpret_cast<const VkMemoryGetZirconHandleInfoFUCHSIA *>(pGetZirconHandleInfo), pZirconHandle)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<zx_handle_t>::type |
| Device::getMemoryZirconHandleFUCHSIA(const VULKAN_HPP_NAMESPACE::MemoryGetZirconHandleInfoFUCHSIA &getZirconHandleInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| zx_handle_t zirconHandle; |
| VkResult result = |
| d.vkGetMemoryZirconHandleFUCHSIA(m_device, reinterpret_cast<const VkMemoryGetZirconHandleInfoFUCHSIA *>(&getZirconHandleInfo), &zirconHandle); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryZirconHandleFUCHSIA"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), zirconHandle); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getMemoryZirconHandlePropertiesFUCHSIA(VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagBits handleType, |
| zx_handle_t zirconHandle, |
| VULKAN_HPP_NAMESPACE::MemoryZirconHandlePropertiesFUCHSIA *pMemoryZirconHandleProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetMemoryZirconHandlePropertiesFUCHSIA(m_device, |
| static_cast<VkExternalMemoryHandleTypeFlagBits>(handleType), |
| zirconHandle, |
| reinterpret_cast<VkMemoryZirconHandlePropertiesFUCHSIA *>(pMemoryZirconHandleProperties))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::MemoryZirconHandlePropertiesFUCHSIA>::type |
| Device::getMemoryZirconHandlePropertiesFUCHSIA(VULKAN_HPP_NAMESPACE::ExternalMemoryHandleTypeFlagBits handleType, |
| zx_handle_t zirconHandle, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryZirconHandlePropertiesFUCHSIA memoryZirconHandleProperties; |
| VkResult result = d.vkGetMemoryZirconHandlePropertiesFUCHSIA(m_device, |
| static_cast<VkExternalMemoryHandleTypeFlagBits>(handleType), |
| zirconHandle, |
| reinterpret_cast<VkMemoryZirconHandlePropertiesFUCHSIA *>(&memoryZirconHandleProperties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryZirconHandlePropertiesFUCHSIA"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), memoryZirconHandleProperties); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_FUCHSIA*/ |
| |
| #if defined(VK_USE_PLATFORM_FUCHSIA) |
| //=== VK_FUCHSIA_external_semaphore === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::importSemaphoreZirconHandleFUCHSIA( |
| const VULKAN_HPP_NAMESPACE::ImportSemaphoreZirconHandleInfoFUCHSIA *pImportSemaphoreZirconHandleInfo, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkImportSemaphoreZirconHandleFUCHSIA( |
| m_device, reinterpret_cast<const VkImportSemaphoreZirconHandleInfoFUCHSIA *>(pImportSemaphoreZirconHandleInfo))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::importSemaphoreZirconHandleFUCHSIA(const VULKAN_HPP_NAMESPACE::ImportSemaphoreZirconHandleInfoFUCHSIA &importSemaphoreZirconHandleInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkImportSemaphoreZirconHandleFUCHSIA( |
| m_device, reinterpret_cast<const VkImportSemaphoreZirconHandleInfoFUCHSIA *>(&importSemaphoreZirconHandleInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::importSemaphoreZirconHandleFUCHSIA"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getSemaphoreZirconHandleFUCHSIA(const VULKAN_HPP_NAMESPACE::SemaphoreGetZirconHandleInfoFUCHSIA *pGetZirconHandleInfo, |
| zx_handle_t *pZirconHandle, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>( |
| d.vkGetSemaphoreZirconHandleFUCHSIA(m_device, reinterpret_cast<const VkSemaphoreGetZirconHandleInfoFUCHSIA *>(pGetZirconHandleInfo), pZirconHandle)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<zx_handle_t>::type |
| Device::getSemaphoreZirconHandleFUCHSIA(const VULKAN_HPP_NAMESPACE::SemaphoreGetZirconHandleInfoFUCHSIA &getZirconHandleInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| zx_handle_t zirconHandle; |
| VkResult result = |
| d.vkGetSemaphoreZirconHandleFUCHSIA(m_device, reinterpret_cast<const VkSemaphoreGetZirconHandleInfoFUCHSIA *>(&getZirconHandleInfo), &zirconHandle); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getSemaphoreZirconHandleFUCHSIA"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), zirconHandle); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_FUCHSIA*/ |
| |
| #if defined(VK_USE_PLATFORM_FUCHSIA) |
| //=== VK_FUCHSIA_buffer_collection === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::createBufferCollectionFUCHSIA(const VULKAN_HPP_NAMESPACE::BufferCollectionCreateInfoFUCHSIA *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA *pCollection, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateBufferCollectionFUCHSIA(m_device, |
| reinterpret_cast<const VkBufferCollectionCreateInfoFUCHSIA *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkBufferCollectionFUCHSIA *>(pCollection))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA>::type |
| Device::createBufferCollectionFUCHSIA(const VULKAN_HPP_NAMESPACE::BufferCollectionCreateInfoFUCHSIA &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection; |
| VkResult result = d.vkCreateBufferCollectionFUCHSIA( |
| m_device, |
| reinterpret_cast<const VkBufferCollectionCreateInfoFUCHSIA *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkBufferCollectionFUCHSIA *>(&collection)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::createBufferCollectionFUCHSIA"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), collection); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA, Dispatch>>::type |
| Device::createBufferCollectionFUCHSIAUnique(const VULKAN_HPP_NAMESPACE::BufferCollectionCreateInfoFUCHSIA &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection; |
| Result result = static_cast<Result>(d.vkCreateBufferCollectionFUCHSIA( |
| m_device, |
| reinterpret_cast<const VkBufferCollectionCreateInfoFUCHSIA *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkBufferCollectionFUCHSIA *>(&collection))); |
| ObjectDestroy<Device, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA, Dispatch>( |
| result, collection, VULKAN_HPP_NAMESPACE_STRING "::Device::createBufferCollectionFUCHSIAUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::setBufferCollectionImageConstraintsFUCHSIA(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, |
| const VULKAN_HPP_NAMESPACE::ImageConstraintsInfoFUCHSIA *pImageConstraintsInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkSetBufferCollectionImageConstraintsFUCHSIA( |
| m_device, static_cast<VkBufferCollectionFUCHSIA>(collection), reinterpret_cast<const VkImageConstraintsInfoFUCHSIA *>(pImageConstraintsInfo))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::setBufferCollectionImageConstraintsFUCHSIA(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, |
| const VULKAN_HPP_NAMESPACE::ImageConstraintsInfoFUCHSIA &imageConstraintsInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkSetBufferCollectionImageConstraintsFUCHSIA( |
| m_device, static_cast<VkBufferCollectionFUCHSIA>(collection), reinterpret_cast<const VkImageConstraintsInfoFUCHSIA *>(&imageConstraintsInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::setBufferCollectionImageConstraintsFUCHSIA"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::setBufferCollectionBufferConstraintsFUCHSIA(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, |
| const VULKAN_HPP_NAMESPACE::BufferConstraintsInfoFUCHSIA *pBufferConstraintsInfo, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkSetBufferCollectionBufferConstraintsFUCHSIA( |
| m_device, static_cast<VkBufferCollectionFUCHSIA>(collection), reinterpret_cast<const VkBufferConstraintsInfoFUCHSIA *>(pBufferConstraintsInfo))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD_WHEN_NO_EXCEPTIONS VULKAN_HPP_INLINE typename ResultValueType<void>::type |
| Device::setBufferCollectionBufferConstraintsFUCHSIA(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, |
| const VULKAN_HPP_NAMESPACE::BufferConstraintsInfoFUCHSIA &bufferConstraintsInfo, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkResult result = d.vkSetBufferCollectionBufferConstraintsFUCHSIA( |
| m_device, static_cast<VkBufferCollectionFUCHSIA>(collection), reinterpret_cast<const VkBufferConstraintsInfoFUCHSIA *>(&bufferConstraintsInfo)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::setBufferCollectionBufferConstraintsFUCHSIA"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result)); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyBufferCollectionFUCHSIA(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyBufferCollectionFUCHSIA( |
| m_device, static_cast<VkBufferCollectionFUCHSIA>(collection), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroyBufferCollectionFUCHSIA(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyBufferCollectionFUCHSIA( |
| m_device, |
| static_cast<VkBufferCollectionFUCHSIA>(collection), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkDestroyBufferCollectionFUCHSIA( |
| m_device, static_cast<VkBufferCollectionFUCHSIA>(collection), reinterpret_cast<const VkAllocationCallbacks *>(pAllocator)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::destroy(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkDestroyBufferCollectionFUCHSIA( |
| m_device, |
| static_cast<VkBufferCollectionFUCHSIA>(collection), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator))); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getBufferCollectionPropertiesFUCHSIA(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, |
| VULKAN_HPP_NAMESPACE::BufferCollectionPropertiesFUCHSIA *pProperties, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetBufferCollectionPropertiesFUCHSIA( |
| m_device, static_cast<VkBufferCollectionFUCHSIA>(collection), reinterpret_cast<VkBufferCollectionPropertiesFUCHSIA *>(pProperties))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::BufferCollectionPropertiesFUCHSIA>::type |
| Device::getBufferCollectionPropertiesFUCHSIA(VULKAN_HPP_NAMESPACE::BufferCollectionFUCHSIA collection, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::BufferCollectionPropertiesFUCHSIA properties; |
| VkResult result = d.vkGetBufferCollectionPropertiesFUCHSIA( |
| m_device, static_cast<VkBufferCollectionFUCHSIA>(collection), reinterpret_cast<VkBufferCollectionPropertiesFUCHSIA *>(&properties)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getBufferCollectionPropertiesFUCHSIA"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), properties); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_FUCHSIA*/ |
| |
| //=== VK_HUAWEI_subpass_shading === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Device::getSubpassShadingMaxWorkgroupSizeHUAWEI(VULKAN_HPP_NAMESPACE::RenderPass renderpass, |
| VULKAN_HPP_NAMESPACE::Extent2D *pMaxWorkgroupSize, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetDeviceSubpassShadingMaxWorkgroupSizeHUAWEI( |
| m_device, static_cast<VkRenderPass>(renderpass), reinterpret_cast<VkExtent2D *>(pMaxWorkgroupSize))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE ResultValue<VULKAN_HPP_NAMESPACE::Extent2D> |
| Device::getSubpassShadingMaxWorkgroupSizeHUAWEI(VULKAN_HPP_NAMESPACE::RenderPass renderpass, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::Extent2D maxWorkgroupSize; |
| VkResult result = d.vkGetDeviceSubpassShadingMaxWorkgroupSizeHUAWEI( |
| m_device, static_cast<VkRenderPass>(renderpass), reinterpret_cast<VkExtent2D *>(&maxWorkgroupSize)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), |
| VULKAN_HPP_NAMESPACE_STRING "::Device::getSubpassShadingMaxWorkgroupSizeHUAWEI", |
| { VULKAN_HPP_NAMESPACE::Result::eSuccess, VULKAN_HPP_NAMESPACE::Result::eIncomplete }); |
| return ResultValue<VULKAN_HPP_NAMESPACE::Extent2D>(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), maxWorkgroupSize); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::subpassShadingHUAWEI(Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSubpassShadingHUAWEI(m_commandBuffer); |
| } |
| |
| //=== VK_HUAWEI_invocation_mask === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::bindInvocationMaskHUAWEI(VULKAN_HPP_NAMESPACE::ImageView imageView, |
| VULKAN_HPP_NAMESPACE::ImageLayout imageLayout, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdBindInvocationMaskHUAWEI(m_commandBuffer, static_cast<VkImageView>(imageView), static_cast<VkImageLayout>(imageLayout)); |
| } |
| |
| //=== VK_NV_external_memory_rdma === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result |
| Device::getMemoryRemoteAddressNV(const VULKAN_HPP_NAMESPACE::MemoryGetRemoteAddressInfoNV *pMemoryGetRemoteAddressInfo, |
| VULKAN_HPP_NAMESPACE::RemoteAddressNV *pAddress, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkGetMemoryRemoteAddressNV( |
| m_device, reinterpret_cast<const VkMemoryGetRemoteAddressInfoNV *>(pMemoryGetRemoteAddressInfo), reinterpret_cast<VkRemoteAddressNV *>(pAddress))); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::RemoteAddressNV>::type |
| Device::getMemoryRemoteAddressNV(const VULKAN_HPP_NAMESPACE::MemoryGetRemoteAddressInfoNV &memoryGetRemoteAddressInfo, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::RemoteAddressNV address; |
| VkResult result = d.vkGetMemoryRemoteAddressNV( |
| m_device, reinterpret_cast<const VkMemoryGetRemoteAddressInfoNV *>(&memoryGetRemoteAddressInfo), reinterpret_cast<VkRemoteAddressNV *>(&address)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Device::getMemoryRemoteAddressNV"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), address); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_extended_dynamic_state2 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setPatchControlPointsEXT(uint32_t patchControlPoints, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetPatchControlPointsEXT(m_commandBuffer, patchControlPoints); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setRasterizerDiscardEnableEXT(VULKAN_HPP_NAMESPACE::Bool32 rasterizerDiscardEnable, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetRasterizerDiscardEnableEXT(m_commandBuffer, static_cast<VkBool32>(rasterizerDiscardEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setDepthBiasEnableEXT(VULKAN_HPP_NAMESPACE::Bool32 depthBiasEnable, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetDepthBiasEnableEXT(m_commandBuffer, static_cast<VkBool32>(depthBiasEnable)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setLogicOpEXT(VULKAN_HPP_NAMESPACE::LogicOp logicOp, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetLogicOpEXT(m_commandBuffer, static_cast<VkLogicOp>(logicOp)); |
| } |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setPrimitiveRestartEnableEXT(VULKAN_HPP_NAMESPACE::Bool32 primitiveRestartEnable, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetPrimitiveRestartEnableEXT(m_commandBuffer, static_cast<VkBool32>(primitiveRestartEnable)); |
| } |
| |
| #if defined(VK_USE_PLATFORM_SCREEN_QNX) |
| //=== VK_QNX_screen_surface === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE Result Instance::createScreenSurfaceQNX(const VULKAN_HPP_NAMESPACE::ScreenSurfaceCreateInfoQNX *pCreateInfo, |
| const VULKAN_HPP_NAMESPACE::AllocationCallbacks *pAllocator, |
| VULKAN_HPP_NAMESPACE::SurfaceKHR *pSurface, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Result>(d.vkCreateScreenSurfaceQNX(m_instance, |
| reinterpret_cast<const VkScreenSurfaceCreateInfoQNX *>(pCreateInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(pAllocator), |
| reinterpret_cast<VkSurfaceKHR *>(pSurface))); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<VULKAN_HPP_NAMESPACE::SurfaceKHR>::type |
| Instance::createScreenSurfaceQNX(const VULKAN_HPP_NAMESPACE::ScreenSurfaceCreateInfoQNX &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| VkResult result = d.vkCreateScreenSurfaceQNX( |
| m_instance, |
| reinterpret_cast<const VkScreenSurfaceCreateInfoQNX *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface)); |
| resultCheck(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), VULKAN_HPP_NAMESPACE_STRING "::Instance::createScreenSurfaceQNX"); |
| return createResultValueType(static_cast<VULKAN_HPP_NAMESPACE::Result>(result), surface); |
| } |
| |
| # ifndef VULKAN_HPP_NO_SMART_HANDLE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE typename ResultValueType<UniqueHandle<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>>::type |
| Instance::createScreenSurfaceQNXUnique(const VULKAN_HPP_NAMESPACE::ScreenSurfaceCreateInfoQNX &createInfo, |
| Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| VULKAN_HPP_NAMESPACE::SurfaceKHR surface; |
| Result result = static_cast<Result>(d.vkCreateScreenSurfaceQNX( |
| m_instance, |
| reinterpret_cast<const VkScreenSurfaceCreateInfoQNX *>(&createInfo), |
| reinterpret_cast<const VkAllocationCallbacks *>(static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>(allocator)), |
| reinterpret_cast<VkSurfaceKHR *>(&surface))); |
| ObjectDestroy<Instance, Dispatch> deleter(*this, allocator, d); |
| return createResultValue<VULKAN_HPP_NAMESPACE::SurfaceKHR, Dispatch>( |
| result, surface, VULKAN_HPP_NAMESPACE_STRING "::Instance::createScreenSurfaceQNXUnique", deleter); |
| } |
| # endif /*VULKAN_HPP_NO_SMART_HANDLE*/ |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE Bool32 PhysicalDevice::getScreenPresentationSupportQNX(uint32_t queueFamilyIndex, |
| struct _screen_window *window, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| return static_cast<Bool32>(d.vkGetPhysicalDeviceScreenPresentationSupportQNX(m_physicalDevice, queueFamilyIndex, window)); |
| } |
| |
| # ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::Bool32 |
| PhysicalDevice::getScreenPresentationSupportQNX(uint32_t queueFamilyIndex, struct _screen_window &window, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VkBool32 result = d.vkGetPhysicalDeviceScreenPresentationSupportQNX(m_physicalDevice, queueFamilyIndex, &window); |
| |
| return static_cast<VULKAN_HPP_NAMESPACE::Bool32>(result); |
| } |
| # endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| #endif /*VK_USE_PLATFORM_SCREEN_QNX*/ |
| |
| //=== VK_EXT_color_write_enable === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setColorWriteEnableEXT(uint32_t attachmentCount, |
| const VULKAN_HPP_NAMESPACE::Bool32 *pColorWriteEnables, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdSetColorWriteEnableEXT(m_commandBuffer, attachmentCount, reinterpret_cast<const VkBool32 *>(pColorWriteEnables)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::setColorWriteEnableEXT(ArrayProxy<const VULKAN_HPP_NAMESPACE::Bool32> const &colorWriteEnables, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdSetColorWriteEnableEXT(m_commandBuffer, colorWriteEnables.size(), reinterpret_cast<const VkBool32 *>(colorWriteEnables.data())); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_multi_draw === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawMultiEXT(uint32_t drawCount, |
| const VULKAN_HPP_NAMESPACE::MultiDrawInfoEXT *pVertexInfo, |
| uint32_t instanceCount, |
| uint32_t firstInstance, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawMultiEXT(m_commandBuffer, drawCount, reinterpret_cast<const VkMultiDrawInfoEXT *>(pVertexInfo), instanceCount, firstInstance, stride); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawMultiEXT(ArrayProxy<const VULKAN_HPP_NAMESPACE::MultiDrawInfoEXT> const &vertexInfo, |
| uint32_t instanceCount, |
| uint32_t firstInstance, |
| uint32_t stride, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdDrawMultiEXT( |
| m_commandBuffer, vertexInfo.size(), reinterpret_cast<const VkMultiDrawInfoEXT *>(vertexInfo.data()), instanceCount, firstInstance, stride); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawMultiIndexedEXT(uint32_t drawCount, |
| const VULKAN_HPP_NAMESPACE::MultiDrawIndexedInfoEXT *pIndexInfo, |
| uint32_t instanceCount, |
| uint32_t firstInstance, |
| uint32_t stride, |
| const int32_t *pVertexOffset, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkCmdDrawMultiIndexedEXT( |
| m_commandBuffer, drawCount, reinterpret_cast<const VkMultiDrawIndexedInfoEXT *>(pIndexInfo), instanceCount, firstInstance, stride, pVertexOffset); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void CommandBuffer::drawMultiIndexedEXT(ArrayProxy<const VULKAN_HPP_NAMESPACE::MultiDrawIndexedInfoEXT> const &indexInfo, |
| uint32_t instanceCount, |
| uint32_t firstInstance, |
| uint32_t stride, |
| Optional<const int32_t> vertexOffset, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| d.vkCmdDrawMultiIndexedEXT(m_commandBuffer, |
| indexInfo.size(), |
| reinterpret_cast<const VkMultiDrawIndexedInfoEXT *>(indexInfo.data()), |
| instanceCount, |
| firstInstance, |
| stride, |
| static_cast<const int32_t *>(vertexOffset)); |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_EXT_pageable_device_local_memory === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::setMemoryPriorityEXT(VULKAN_HPP_NAMESPACE::DeviceMemory memory, float priority, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkSetDeviceMemoryPriorityEXT(m_device, static_cast<VkDeviceMemory>(memory), priority); |
| } |
| |
| //=== VK_KHR_maintenance4 === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getBufferMemoryRequirementsKHR(const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceBufferMemoryRequirementsKHR( |
| m_device, reinterpret_cast<const VkDeviceBufferMemoryRequirements *>(pInfo), reinterpret_cast<VkMemoryRequirements2 *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 |
| Device::getBufferMemoryRequirementsKHR(const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements; |
| d.vkGetDeviceBufferMemoryRequirementsKHR( |
| m_device, reinterpret_cast<const VkDeviceBufferMemoryRequirements *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getBufferMemoryRequirementsKHR(const VULKAN_HPP_NAMESPACE::DeviceBufferMemoryRequirements &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>(); |
| d.vkGetDeviceBufferMemoryRequirementsKHR( |
| m_device, reinterpret_cast<const VkDeviceBufferMemoryRequirements *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageMemoryRequirementsKHR(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements *pInfo, |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 *pMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceImageMemoryRequirementsKHR( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(pInfo), reinterpret_cast<VkMemoryRequirements2 *>(pMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 |
| Device::getImageMemoryRequirementsKHR(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements; |
| d.vkGetDeviceImageMemoryRequirementsKHR( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return memoryRequirements; |
| } |
| |
| template<typename X, typename Y, typename... Z, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE StructureChain<X, Y, Z...> |
| Device::getImageMemoryRequirementsKHR(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements &info, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| StructureChain<X, Y, Z...> structureChain; |
| VULKAN_HPP_NAMESPACE::MemoryRequirements2 &memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>(); |
| d.vkGetDeviceImageMemoryRequirementsKHR( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), reinterpret_cast<VkMemoryRequirements2 *>(&memoryRequirements)); |
| |
| return structureChain; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getImageSparseMemoryRequirementsKHR(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements *pInfo, |
| uint32_t *pSparseMemoryRequirementCount, |
| VULKAN_HPP_NAMESPACE::SparseImageMemoryRequirements2 *pSparseMemoryRequirements, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDeviceImageSparseMemoryRequirementsKHR(m_device, |
| reinterpret_cast<const VkDeviceImageMemoryRequirements *>(pInfo), |
| pSparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(pSparseMemoryRequirements)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename SparseImageMemoryRequirements2Allocator, typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> |
| Device::getImageSparseMemoryRequirementsKHR(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements &info, Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements; |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetDeviceImageSparseMemoryRequirementsKHR( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetDeviceImageSparseMemoryRequirementsKHR(m_device, |
| reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| |
| template<typename SparseImageMemoryRequirements2Allocator, |
| typename Dispatch, |
| typename B, |
| typename std::enable_if<std::is_same<typename B::value_type, SparseImageMemoryRequirements2>::value, int>::type> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> |
| Device::getImageSparseMemoryRequirementsKHR(const VULKAN_HPP_NAMESPACE::DeviceImageMemoryRequirements &info, |
| SparseImageMemoryRequirements2Allocator &sparseImageMemoryRequirements2Allocator, |
| Dispatch const &d) const |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| std::vector<SparseImageMemoryRequirements2, SparseImageMemoryRequirements2Allocator> sparseMemoryRequirements(sparseImageMemoryRequirements2Allocator); |
| uint32_t sparseMemoryRequirementCount; |
| d.vkGetDeviceImageSparseMemoryRequirementsKHR( |
| m_device, reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), &sparseMemoryRequirementCount, nullptr); |
| sparseMemoryRequirements.resize(sparseMemoryRequirementCount); |
| d.vkGetDeviceImageSparseMemoryRequirementsKHR(m_device, |
| reinterpret_cast<const VkDeviceImageMemoryRequirements *>(&info), |
| &sparseMemoryRequirementCount, |
| reinterpret_cast<VkSparseImageMemoryRequirements2 *>(sparseMemoryRequirements.data())); |
| VULKAN_HPP_ASSERT(sparseMemoryRequirementCount <= sparseMemoryRequirements.size()); |
| return sparseMemoryRequirements; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| //=== VK_VALVE_descriptor_set_host_mapping === |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void Device::getDescriptorSetLayoutHostMappingInfoVALVE(const VULKAN_HPP_NAMESPACE::DescriptorSetBindingReferenceVALVE *pBindingReference, |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayoutHostMappingInfoVALVE *pHostMapping, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDescriptorSetLayoutHostMappingInfoVALVE(m_device, |
| reinterpret_cast<const VkDescriptorSetBindingReferenceVALVE *>(pBindingReference), |
| reinterpret_cast<VkDescriptorSetLayoutHostMappingInfoVALVE *>(pHostMapping)); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::DescriptorSetLayoutHostMappingInfoVALVE |
| Device::getDescriptorSetLayoutHostMappingInfoVALVE(const VULKAN_HPP_NAMESPACE::DescriptorSetBindingReferenceVALVE &bindingReference, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| VULKAN_HPP_NAMESPACE::DescriptorSetLayoutHostMappingInfoVALVE hostMapping; |
| d.vkGetDescriptorSetLayoutHostMappingInfoVALVE(m_device, |
| reinterpret_cast<const VkDescriptorSetBindingReferenceVALVE *>(&bindingReference), |
| reinterpret_cast<VkDescriptorSetLayoutHostMappingInfoVALVE *>(&hostMapping)); |
| |
| return hostMapping; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| template<typename Dispatch> |
| VULKAN_HPP_INLINE void |
| Device::getDescriptorSetHostMappingVALVE(VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet, void **ppData, Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| d.vkGetDescriptorSetHostMappingVALVE(m_device, static_cast<VkDescriptorSet>(descriptorSet), ppData); |
| } |
| |
| #ifndef VULKAN_HPP_DISABLE_ENHANCED_MODE |
| template<typename Dispatch> |
| VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE void *Device::getDescriptorSetHostMappingVALVE(VULKAN_HPP_NAMESPACE::DescriptorSet descriptorSet, |
| Dispatch const &d) const VULKAN_HPP_NOEXCEPT |
| { |
| VULKAN_HPP_ASSERT(d.getVkHeaderVersion() == VK_HEADER_VERSION); |
| |
| void *pData; |
| d.vkGetDescriptorSetHostMappingVALVE(m_device, static_cast<VkDescriptorSet>(descriptorSet), &pData); |
| |
| return pData; |
| } |
| #endif /*VULKAN_HPP_DISABLE_ENHANCED_MODE*/ |
| |
| } // namespace VULKAN_HPP_NAMESPACE |
| #endif |