Lines Matching defs:size
117 VULKAN_HPP_ASSERT( physicalDeviceCount <= physicalDevices.size() );
118 if ( physicalDeviceCount < physicalDevices.size() )
148 VULKAN_HPP_ASSERT( physicalDeviceCount <= physicalDevices.size() );
149 if ( physicalDeviceCount < physicalDevices.size() )
293 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
294 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
317 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
318 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
490 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
491 if ( propertyCount < properties.size() )
524 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
525 if ( propertyCount < properties.size() )
566 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
567 if ( propertyCount < properties.size() )
600 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
601 if ( propertyCount < properties.size() )
639 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
640 if ( propertyCount < properties.size() )
670 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
671 if ( propertyCount < properties.size() )
709 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
710 if ( propertyCount < properties.size() )
740 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
741 if ( propertyCount < properties.size() )
789 d.vkQueueSubmit( m_queue, submits.size(), reinterpret_cast<const VkSubmitInfo *>( submits.data() ), static_cast<VkFence>( fence ) ) );
941 VULKAN_HPP_NAMESPACE::DeviceSize size,
950 static_cast<VkDeviceSize>( size ),
959 VULKAN_HPP_NAMESPACE::DeviceSize size,
969 static_cast<VkDeviceSize>( size ),
1003 d.vkFlushMappedMemoryRanges( m_device, memoryRanges.size(), reinterpret_cast<const VkMappedMemoryRange *>( memoryRanges.data() ) ) );
1029 d.vkInvalidateMappedMemoryRanges( m_device, memoryRanges.size(), reinterpret_cast<const VkMappedMemoryRange *>( memoryRanges.data() ) ) );
1186 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
1187 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
1215 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
1216 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
1277 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
1278 if ( propertyCount < properties.size() )
1320 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
1321 if ( propertyCount < properties.size() )
1348 d.vkQueueBindSparse( m_queue, bindInfo.size(), reinterpret_cast<const VkBindSparseInfo *>( bindInfo.data() ), static_cast<VkFence>( fence ) ) );
1470 static_cast<VULKAN_HPP_NAMESPACE::Result>( d.vkResetFences( m_device, fences.size(), reinterpret_cast<const VkFence *>( fences.data() ) ) );
1521 d.vkWaitForFences( m_device, fences.size(), reinterpret_cast<const VkFence *>( fences.data() ), static_cast<VkBool32>( waitAll ), timeout ) );
1933 data.size() * sizeof( DataType ),
2636 VULKAN_HPP_ASSERT( dataSize <= data.size() );
2637 if ( dataSize < data.size() )
2668 VULKAN_HPP_ASSERT( dataSize <= data.size() );
2669 if ( dataSize < data.size() )
2698 m_device, static_cast<VkPipelineCache>( dstCache ), srcCaches.size(), reinterpret_cast<const VkPipelineCache *>( srcCaches.data() ) ) );
2732 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
2736 createInfos.size(),
2760 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
2764 createInfos.size(),
2809 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
2813 createInfos.size(),
2821 uniquePipelines.reserve( createInfos.size() );
2844 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
2848 createInfos.size(),
2856 uniquePipelines.reserve( createInfos.size() );
2921 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
2925 createInfos.size(),
2949 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
2953 createInfos.size(),
2998 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
3002 createInfos.size(),
3010 uniquePipelines.reserve( createInfos.size() );
3033 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
3037 createInfos.size(),
3045 uniquePipelines.reserve( createInfos.size() );
3673 m_device, static_cast<VkDescriptorPool>( descriptorPool ), descriptorSets.size(), reinterpret_cast<const VkDescriptorSet *>( descriptorSets.data() ) );
3697 m_device, static_cast<VkDescriptorPool>( descriptorPool ), descriptorSets.size(), reinterpret_cast<const VkDescriptorSet *>( descriptorSets.data() ) );
3726 descriptorWrites.size(),
3728 descriptorCopies.size(),
4200 m_device, static_cast<VkCommandPool>( commandPool ), commandBuffers.size(), reinterpret_cast<const VkCommandBuffer *>( commandBuffers.data() ) );
4224 m_device, static_cast<VkCommandPool>( commandPool ), commandBuffers.size(), reinterpret_cast<const VkCommandBuffer *>( commandBuffers.data() ) );
4320 d.vkCmdSetViewport( m_commandBuffer, firstViewport, viewports.size(), reinterpret_cast<const VkViewport *>( viewports.data() ) );
4342 d.vkCmdSetScissor( m_commandBuffer, firstScissor, scissors.size(), reinterpret_cast<const VkRect2D *>( scissors.data() ) );
4435 descriptorSets.size(),
4437 dynamicOffsets.size(),
4473 VULKAN_HPP_ASSERT( buffers.size() == offsets.size() );
4475 if ( buffers.size() != offsets.size() )
4477 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers: buffers.size() != offsets.size()" );
4483 buffers.size(),
4575 regions.size(),
4615 regions.size(),
4658 regions.size(),
4695 regions.size(),
4731 regions.size(),
4760 data.size() * sizeof( DataType ),
4768 VULKAN_HPP_NAMESPACE::DeviceSize size,
4773 d.vkCmdFillBuffer( m_commandBuffer, static_cast<VkBuffer>( dstBuffer ), static_cast<VkDeviceSize>( dstOffset ), static_cast<VkDeviceSize>( size ), data );
4807 ranges.size(),
4844 ranges.size(),
4873 attachments.size(),
4875 rects.size(),
4915 regions.size(),
4979 events.size(),
4983 memoryBarriers.size(),
4985 bufferMemoryBarriers.size(),
4987 imageMemoryBarriers.size(),
5034 memoryBarriers.size(),
5036 bufferMemoryBarriers.size(),
5038 imageMemoryBarriers.size(),
5105 uint32_t size,
5110 d.vkCmdPushConstants( m_commandBuffer, static_cast<VkPipelineLayout>( layout ), static_cast<VkShaderStageFlags>( stageFlags ), offset, size, pValues );
5127 values.size() * sizeof( ValuesType ),
5183 d.vkCmdExecuteCommands( m_commandBuffer, commandBuffers.size(), reinterpret_cast<const VkCommandBuffer *>( commandBuffers.data() ) );
5227 d.vkBindBufferMemory2( m_device, bindInfos.size(), reinterpret_cast<const VkBindBufferMemoryInfo *>( bindInfos.data() ) ) );
5251 d.vkBindImageMemory2( m_device, bindInfos.size(), reinterpret_cast<const VkBindImageMemoryInfo *>( bindInfos.data() ) ) );
5338 VULKAN_HPP_ASSERT( physicalDeviceGroupCount <= physicalDeviceGroupProperties.size() );
5339 if ( physicalDeviceGroupCount < physicalDeviceGroupProperties.size() )
5371 VULKAN_HPP_ASSERT( physicalDeviceGroupCount <= physicalDeviceGroupProperties.size() );
5372 if ( physicalDeviceGroupCount < physicalDeviceGroupProperties.size() )
5488 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
5489 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
5518 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
5519 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
5703 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
5704 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
5727 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
5728 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
5754 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
5755 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
5789 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
5790 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
5868 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
5869 if ( propertyCount < properties.size() )
5897 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
5898 if ( propertyCount < properties.size() )
6645 VULKAN_HPP_ASSERT( toolCount <= toolProperties.size() );
6646 if ( toolCount < toolProperties.size() )
6678 VULKAN_HPP_ASSERT( toolCount <= toolProperties.size() );
6679 if ( toolCount < toolProperties.size() )
6897 VULKAN_HPP_ASSERT( events.size() == dependencyInfos.size() );
6899 if ( events.size() != dependencyInfos.size() )
6901 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::waitEvents2: events.size() != dependencyInfos.size()" );
6906 events.size(),
6959 d.vkQueueSubmit2( m_queue, submits.size(), reinterpret_cast<const VkSubmitInfo2 *>( submits.data() ), static_cast<VkFence>( fence ) ) );
7140 d.vkCmdSetViewportWithCount( m_commandBuffer, viewports.size(), reinterpret_cast<const VkViewport *>( viewports.data() ) );
7159 d.vkCmdSetScissorWithCount( m_commandBuffer, scissors.size(), reinterpret_cast<const VkRect2D *>( scissors.data() ) );
7193 VULKAN_HPP_ASSERT( buffers.size() == offsets.size() );
7194 VULKAN_HPP_ASSERT( sizes.empty() || buffers.size() == sizes.size() );
7195 VULKAN_HPP_ASSERT( strides.empty() || buffers.size() == strides.size() );
7197 if ( buffers.size() != offsets.size() )
7199 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2: buffers.size() != offsets.size()" );
7201 if ( !sizes.empty() && buffers.size() != sizes.size() )
7203 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2: buffers.size() != sizes.size()" );
7205 if ( !strides.empty() && buffers.size() != strides.size() )
7207 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2: buffers.size() != strides.size()" );
7213 buffers.size(),
7405 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
7406 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
7435 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
7436 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
7578 VULKAN_HPP_ASSERT( surfaceFormatCount <= surfaceFormats.size() );
7579 if ( surfaceFormatCount < surfaceFormats.size() )
7612 VULKAN_HPP_ASSERT( surfaceFormatCount <= surfaceFormats.size() );
7613 if ( surfaceFormatCount < surfaceFormats.size() )
7654 VULKAN_HPP_ASSERT( presentModeCount <= presentModes.size() );
7655 if ( presentModeCount < presentModes.size() )
7688 VULKAN_HPP_ASSERT( presentModeCount <= presentModes.size() );
7689 if ( presentModeCount < presentModes.size() )
7834 VULKAN_HPP_ASSERT( swapchainImageCount <= swapchainImages.size() );
7835 if ( swapchainImageCount < swapchainImages.size() )
7866 VULKAN_HPP_ASSERT( swapchainImageCount <= swapchainImages.size() );
7867 if ( swapchainImageCount < swapchainImages.size() )
8020 VULKAN_HPP_ASSERT( rectCount <= rects.size() );
8021 if ( rectCount < rects.size() )
8052 VULKAN_HPP_ASSERT( rectCount <= rects.size() );
8053 if ( rectCount < rects.size() )
8124 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
8125 if ( propertyCount < properties.size() )
8155 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
8156 if ( propertyCount < properties.size() )
8196 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
8197 if ( propertyCount < properties.size() )
8228 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
8229 if ( propertyCount < properties.size() )
8269 VULKAN_HPP_ASSERT( displayCount <= displays.size() );
8270 if ( displayCount < displays.size() )
8300 VULKAN_HPP_ASSERT( displayCount <= displays.size() );
8301 if ( displayCount < displays.size() )
8343 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
8344 if ( propertyCount < properties.size() )
8378 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
8379 if ( propertyCount < properties.size() )
8559 std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR, SwapchainKHRAllocator> swapchains( createInfos.size() );
8562 createInfos.size(),
8583 std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR, SwapchainKHRAllocator> swapchains( createInfos.size(), swapchainKHRAllocator );
8586 createInfos.size(),
8625 std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR> swapchains( createInfos.size() );
8628 createInfos.size(),
8634 uniqueSwapchains.reserve( createInfos.size() );
8656 std::vector<VULKAN_HPP_NAMESPACE::SwapchainKHR> swapchains( createInfos.size() );
8659 createInfos.size(),
8665 uniqueSwapchains.reserve( createInfos.size() );
9470 VULKAN_HPP_ASSERT( videoFormatPropertyCount <= videoFormatProperties.size() );
9471 if ( videoFormatPropertyCount < videoFormatProperties.size() )
9508 VULKAN_HPP_ASSERT( videoFormatPropertyCount <= videoFormatProperties.size() );
9509 if ( videoFormatPropertyCount < videoFormatProperties.size() )
9661 VULKAN_HPP_ASSERT( memoryRequirementsCount <= memoryRequirements.size() );
9662 if ( memoryRequirementsCount < memoryRequirements.size() )
9700 VULKAN_HPP_ASSERT( memoryRequirementsCount <= memoryRequirements.size() );
9701 if ( memoryRequirementsCount < memoryRequirements.size() )
9735 bindSessionMemoryInfos.size(),
9990 VULKAN_HPP_ASSERT( buffers.size() == offsets.size() );
9991 VULKAN_HPP_ASSERT( sizes.empty() || buffers.size() == sizes.size() );
9993 if ( buffers.size() != offsets.size() )
9995 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindTransformFeedbackBuffersEXT: buffers.size() != offsets.size()" );
9997 if ( !sizes.empty() && buffers.size() != sizes.size() )
9999 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindTransformFeedbackBuffersEXT: buffers.size() != sizes.size()" );
10005 buffers.size(),
10037 VULKAN_HPP_ASSERT( counterBufferOffsets.empty() || counterBuffers.size() == counterBufferOffsets.size() );
10039 if ( !counterBufferOffsets.empty() && counterBuffers.size() != counterBufferOffsets.size() )
10041 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::beginTransformFeedbackEXT: counterBuffers.size() != counterBufferOffsets.size()" );
10047 counterBuffers.size(),
10078 VULKAN_HPP_ASSERT( counterBufferOffsets.empty() || counterBuffers.size() == counterBufferOffsets.size() );
10080 if ( !counterBufferOffsets.empty() && counterBuffers.size() != counterBufferOffsets.size() )
10082 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::endTransformFeedbackEXT: counterBuffers.size() != counterBufferOffsets.size()" );
10088 counterBuffers.size(),
10498 VULKAN_HPP_ASSERT( infoSize <= info.size() );
10499 if ( infoSize < info.size() )
10542 VULKAN_HPP_ASSERT( infoSize <= info.size() );
10543 if ( infoSize < info.size() )
10907 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
10908 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
10931 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
10932 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
10958 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
10959 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
10993 VULKAN_HPP_ASSERT( queueFamilyPropertyCount <= queueFamilyProperties.size() );
10994 if ( queueFamilyPropertyCount < queueFamilyProperties.size() )
11073 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
11074 if ( propertyCount < properties.size() )
11102 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
11103 if ( propertyCount < properties.size() )
11266 VULKAN_HPP_ASSERT( physicalDeviceGroupCount <= physicalDeviceGroupProperties.size() );
11267 if ( physicalDeviceGroupCount < physicalDeviceGroupProperties.size() )
11299 VULKAN_HPP_ASSERT( physicalDeviceGroupCount <= physicalDeviceGroupProperties.size() );
11300 if ( physicalDeviceGroupCount < physicalDeviceGroupProperties.size() )
11624 descriptorWrites.size(),
11821 m_commandBuffer, firstViewport, viewportWScalings.size(), reinterpret_cast<const VkViewportWScalingNV *>( viewportWScalings.data() ) );
12186 VULKAN_HPP_ASSERT( presentationTimingCount <= presentationTimings.size() );
12187 if ( presentationTimingCount < presentationTimings.size() )
12225 VULKAN_HPP_ASSERT( presentationTimingCount <= presentationTimings.size() );
12226 if ( presentationTimingCount < presentationTimings.size() )
12255 m_commandBuffer, firstDiscardRectangle, discardRectangles.size(), reinterpret_cast<const VkRect2D *>( discardRectangles.data() ) );
12296 VULKAN_HPP_ASSERT( swapchains.size() == metadata.size() );
12298 if ( swapchains.size() != metadata.size() )
12300 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::Device::setHdrMetadataEXT: swapchains.size() != metadata.size()" );
12305 swapchains.size(),
12647 VULKAN_HPP_ASSERT( counterCount <= counters.size() );
12648 if ( counterCount < counters.size() )
12699 VULKAN_HPP_ASSERT( counterCount <= counters.size() );
12700 if ( counterCount < counters.size() )
12852 VULKAN_HPP_ASSERT( surfaceFormatCount <= surfaceFormats.size() );
12853 if ( surfaceFormatCount < surfaceFormats.size() )
12889 VULKAN_HPP_ASSERT( surfaceFormatCount <= surfaceFormats.size() );
12890 if ( surfaceFormatCount < surfaceFormats.size() )
12927 VULKAN_HPP_ASSERT( surfaceFormatCount <= surfaceFormats.size() );
12928 if ( surfaceFormatCount < surfaceFormats.size() )
12975 VULKAN_HPP_ASSERT( surfaceFormatCount <= surfaceFormats.size() );
12976 if ( surfaceFormatCount < surfaceFormats.size() )
13022 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
13023 if ( propertyCount < properties.size() )
13054 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
13055 if ( propertyCount < properties.size() )
13095 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
13096 if ( propertyCount < properties.size() )
13127 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
13128 if ( propertyCount < properties.size() )
13170 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
13171 if ( propertyCount < properties.size() )
13205 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
13206 if ( propertyCount < properties.size() )
13743 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
13747 createInfos.size(),
13771 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
13775 createInfos.size(),
13821 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
13825 createInfos.size(),
13833 uniquePipelines.reserve( createInfos.size() );
13857 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
13861 createInfos.size(),
13869 uniquePipelines.reserve( createInfos.size() );
14180 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
14181 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
14210 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
14211 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
14352 VULKAN_HPP_ASSERT( infos.size() == pBuildRangeInfos.size() );
14354 if ( infos.size() != pBuildRangeInfos.size() )
14356 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::buildAccelerationStructuresKHR: infos.size() != pBuildRangeInfos.size()" );
14361 infos.size(),
14395 VULKAN_HPP_ASSERT( infos.size() == indirectDeviceAddresses.size() );
14396 VULKAN_HPP_ASSERT( infos.size() == indirectStrides.size() );
14397 VULKAN_HPP_ASSERT( infos.size() == pMaxPrimitiveCounts.size() );
14399 if ( infos.size() != indirectDeviceAddresses.size() )
14401 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::buildAccelerationStructuresIndirectKHR: infos.size() != indirectDeviceAddresses.size()" );
14403 if ( infos.size() != indirectStrides.size() )
14405 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::buildAccelerationStructuresIndirectKHR: infos.size() != indirectStrides.size()" );
14407 if ( infos.size() != pMaxPrimitiveCounts.size() )
14409 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::buildAccelerationStructuresIndirectKHR: infos.size() != pMaxPrimitiveCounts.size()" );
14414 infos.size(),
14449 VULKAN_HPP_ASSERT( infos.size() == pBuildRangeInfos.size() );
14451 if ( infos.size() != pBuildRangeInfos.size() )
14453 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::Device::buildAccelerationStructuresKHR: infos.size() != pBuildRangeInfos.size()" );
14460 infos.size(),
14600 accelerationStructures.size(),
14603 data.size() * sizeof( DataType ),
14623 accelerationStructures.size(),
14746 accelerationStructures.size(),
14807 VULKAN_HPP_ASSERT( maxPrimitiveCounts.size() == buildInfo.geometryCount );
14809 if ( maxPrimitiveCounts.size() != buildInfo.geometryCount )
14811 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::Device::getAccelerationStructureBuildSizesKHR: maxPrimitiveCounts.size() != buildInfo.geometryCount" );
14904 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
14909 createInfos.size(),
14937 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
14942 createInfos.size(),
14997 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
15002 createInfos.size(),
15013 uniquePipelines.reserve( createInfos.size() );
15038 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
15043 createInfos.size(),
15054 uniquePipelines.reserve( createInfos.size() );
15120 m_device, static_cast<VkPipeline>( pipeline ), firstGroup, groupCount, data.size() * sizeof( DataType ), reinterpret_cast<void *>( data.data() ) ) );
15165 m_device, static_cast<VkPipeline>( pipeline ), firstGroup, groupCount, data.size() * sizeof( DataType ), reinterpret_cast<void *>( data.data() ) ) );
15345 d.vkBindBufferMemory2KHR( m_device, bindInfos.size(), reinterpret_cast<const VkBindBufferMemoryInfo *>( bindInfos.data() ) ) );
15369 d.vkBindImageMemory2KHR( m_device, bindInfos.size(), reinterpret_cast<const VkBindImageMemoryInfo *>( bindInfos.data() ) ) );
15532 m_device, static_cast<VkValidationCacheEXT>( dstCache ), srcCaches.size(), reinterpret_cast<const VkValidationCacheEXT *>( srcCaches.data() ) ) );
15571 VULKAN_HPP_ASSERT( dataSize <= data.size() );
15572 if ( dataSize < data.size() )
15603 VULKAN_HPP_ASSERT( dataSize <= data.size() );
15604 if ( dataSize < data.size() )
15644 m_commandBuffer, firstViewport, shadingRatePalettes.size(), reinterpret_cast<const VkShadingRatePaletteNV *>( shadingRatePalettes.data() ) );
15672 customSampleOrders.size(),
15849 m_device, bindInfos.size(), reinterpret_cast<const VkBindAccelerationStructureMemoryInfoNV *>( bindInfos.data() ) ) );
15980 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size() );
15984 createInfos.size(),
16008 std::vector<VULKAN_HPP_NAMESPACE::Pipeline, PipelineAllocator> pipelines( createInfos.size(), pipelineAllocator );
16012 createInfos.size(),
16057 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
16061 createInfos.size(),
16069 uniquePipelines.reserve( createInfos.size() );
16092 std::vector<VULKAN_HPP_NAMESPACE::Pipeline> pipelines( createInfos.size() );
16096 createInfos.size(),
16104 uniquePipelines.reserve( createInfos.size() );
16165 m_device, static_cast<VkPipeline>( pipeline ), firstGroup, groupCount, data.size() * sizeof( DataType ), reinterpret_cast<void *>( data.data() ) ) );
16207 m_device, static_cast<VkAccelerationStructureNV>( accelerationStructure ), data.size() * sizeof( DataType ), reinterpret_cast<void *>( data.data() ) ) );
16257 accelerationStructures.size(),
16459 VULKAN_HPP_ASSERT( timeDomainCount <= timeDomains.size() );
16460 if ( timeDomainCount < timeDomains.size() )
16490 VULKAN_HPP_ASSERT( timeDomainCount <= timeDomains.size() );
16491 if ( timeDomainCount < timeDomains.size() )
16520 std::piecewise_construct, std::forward_as_tuple( timestampInfos.size() ), std::forward_as_tuple( 0 ) );
16524 m_device, timestampInfos.size(), reinterpret_cast<const VkCalibratedTimestampInfoKHR *>( timestampInfos.data() ), timestamps.data(), &maxDeviation ) );
16542 std::piecewise_construct, std::forward_as_tuple( timestampInfos.size(), uint64_tAllocator ), std::forward_as_tuple( 0 ) );
16546 m_device, timestampInfos.size(), reinterpret_cast<const VkCalibratedTimestampInfoKHR *>( timestampInfos.data() ), timestamps.data(), &maxDeviation ) );
16631 m_commandBuffer, firstExclusiveScissor, exclusiveScissorEnables.size(), reinterpret_cast<const VkBool32 *>( exclusiveScissorEnables.data() ) );
16654 m_commandBuffer, firstExclusiveScissor, exclusiveScissors.size(), reinterpret_cast<const VkRect2D *>( exclusiveScissors.data() ) );
16699 VULKAN_HPP_ASSERT( checkpointDataCount <= checkpointData.size() );
16700 if ( checkpointDataCount < checkpointData.size() )
16722 VULKAN_HPP_ASSERT( checkpointDataCount <= checkpointData.size() );
16723 if ( checkpointDataCount < checkpointData.size() )
17223 VULKAN_HPP_ASSERT( fragmentShadingRateCount <= fragmentShadingRates.size() );
17224 if ( fragmentShadingRateCount < fragmentShadingRates.size() )
17258 VULKAN_HPP_ASSERT( fragmentShadingRateCount <= fragmentShadingRates.size() );
17259 if ( fragmentShadingRateCount < fragmentShadingRates.size() )
17347 VULKAN_HPP_ASSERT( toolCount <= toolProperties.size() );
17348 if ( toolCount < toolProperties.size() )
17380 VULKAN_HPP_ASSERT( toolCount <= toolProperties.size() );
17381 if ( toolCount < toolProperties.size() )
17451 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
17452 if ( propertyCount < properties.size() )
17485 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
17486 if ( propertyCount < properties.size() )
17528 VULKAN_HPP_ASSERT( combinationCount <= combinations.size() );
17529 if ( combinationCount < combinations.size() )
17563 VULKAN_HPP_ASSERT( combinationCount <= combinations.size() );
17564 if ( combinationCount < combinations.size() )
17614 VULKAN_HPP_ASSERT( presentModeCount <= presentModes.size() );
17615 if ( presentModeCount < presentModes.size() )
17651 VULKAN_HPP_ASSERT( presentModeCount <= presentModes.size() );
17652 if ( presentModeCount < presentModes.size() )
17919 d.vkCmdSetViewportWithCountEXT( m_commandBuffer, viewports.size(), reinterpret_cast<const VkViewport *>( viewports.data() ) );
17938 d.vkCmdSetScissorWithCountEXT( m_commandBuffer, scissors.size(), reinterpret_cast<const VkRect2D *>( scissors.data() ) );
17972 VULKAN_HPP_ASSERT( buffers.size() == offsets.size() );
17973 VULKAN_HPP_ASSERT( sizes.empty() || buffers.size() == sizes.size() );
17974 VULKAN_HPP_ASSERT( strides.empty() || buffers.size() == strides.size() );
17976 if ( buffers.size() != offsets.size() )
17978 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2EXT: buffers.size() != offsets.size()" );
17980 if ( !sizes.empty() && buffers.size() != sizes.size() )
17982 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2EXT: buffers.size() != sizes.size()" );
17984 if ( !strides.empty() && buffers.size() != strides.size() )
17986 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindVertexBuffers2EXT: buffers.size() != strides.size()" );
17992 buffers.size(),
18248 VULKAN_HPP_ASSERT( executableCount <= properties.size() );
18249 if ( executableCount < properties.size() )
18287 VULKAN_HPP_ASSERT( executableCount <= properties.size() );
18288 if ( executableCount < properties.size() )
18336 VULKAN_HPP_ASSERT( statisticCount <= statistics.size() );
18337 if ( statisticCount < statistics.size() )
18375 VULKAN_HPP_ASSERT( statisticCount <= statistics.size() );
18376 if ( statisticCount < statistics.size() )
18426 VULKAN_HPP_ASSERT( internalRepresentationCount <= internalRepresentations.size() );
18427 if ( internalRepresentationCount < internalRepresentations.size() )
18466 VULKAN_HPP_ASSERT( internalRepresentationCount <= internalRepresentations.size() );
18467 if ( internalRepresentationCount < internalRepresentations.size() )
18565 d.vkTransitionImageLayoutEXT( m_device, transitions.size(), reinterpret_cast<const VkHostImageLayoutTransitionInfoEXT *>( transitions.data() ) ) );
19468 VULKAN_HPP_ASSERT( cacheSize <= cacheData.size() );
19469 if ( cacheSize < cacheData.size() )
19499 VULKAN_HPP_ASSERT( cacheSize <= cacheData.size() );
19500 if ( cacheSize < cacheData.size() )
19766 VULKAN_HPP_ASSERT( events.size() == dependencyInfos.size() );
19768 if ( events.size() != dependencyInfos.size() )
19770 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::waitEvents2KHR: events.size() != dependencyInfos.size()" );
19775 events.size(),
19829 d.vkQueueSubmit2KHR( m_queue, submits.size(), reinterpret_cast<const VkSubmitInfo2 *>( submits.data() ), static_cast<VkFence>( fence ) ) );
19870 VULKAN_HPP_ASSERT( checkpointDataCount <= checkpointData.size() );
19871 if ( checkpointDataCount < checkpointData.size() )
19893 VULKAN_HPP_ASSERT( checkpointDataCount <= checkpointData.size() );
19894 if ( checkpointDataCount < checkpointData.size() )
20004 d.vkCmdBindDescriptorBuffersEXT( m_commandBuffer, bindingInfos.size(), reinterpret_cast<const VkDescriptorBufferBindingInfoEXT *>( bindingInfos.data() ) );
20038 VULKAN_HPP_ASSERT( bufferIndices.size() == offsets.size() );
20040 if ( bufferIndices.size() != offsets.size() )
20042 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::setDescriptorBufferOffsetsEXT: bufferIndices.size() != offsets.size()" );
20050 bufferIndices.size(),
20571 vertexBindingDescriptions.size(),
20573 vertexAttributeDescriptions.size(),
21139 d.vkCmdSetColorWriteEnableEXT( m_commandBuffer, colorWriteEnables.size(), reinterpret_cast<const VkBool32 *>( colorWriteEnables.data() ) );
21177 vertexInfo.size(),
21211 indexInfo.size(),
21340 d.vkCmdBuildMicromapsEXT( m_commandBuffer, infos.size(), reinterpret_cast<const VkMicromapBuildInfoEXT *>( infos.data() ) );
21365 m_device, static_cast<VkDeferredOperationKHR>( deferredOperation ), infos.size(), reinterpret_cast<const VkMicromapBuildInfoEXT *>( infos.data() ) ) );
21489 micromaps.size(),
21492 data.size() * sizeof( DataType ),
21512 micromaps.size(),
21608 micromaps.size(),
21813 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
21814 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
21843 VULKAN_HPP_ASSERT( sparseMemoryRequirementCount <= sparseMemoryRequirements.size() );
21844 if ( sparseMemoryRequirementCount < sparseMemoryRequirements.size() )
21949 imageSubresources.size(),
21977 m_commandBuffer, decompressMemoryRegions.size(), reinterpret_cast<const VkDecompressMemoryRegionNV *>( decompressMemoryRegions.data() ) );
22115 VULKAN_HPP_ASSERT( sampleMask.size() == ( static_cast<uint32_t>( samples ) + 31 ) / 32 );
22117 if ( sampleMask.size() != ( static_cast<uint32_t>( samples ) + 31 ) / 32 )
22119 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::setSampleMaskEXT: sampleMask.size() != ( static_cast<uint32_t>( samples ) + 31 ) / 32" );
22167 d.vkCmdSetColorBlendEnableEXT( m_commandBuffer, firstAttachment, colorBlendEnables.size(), reinterpret_cast<const VkBool32 *>( colorBlendEnables.data() ) );
22192 m_commandBuffer, firstAttachment, colorBlendEquations.size(), reinterpret_cast<const VkColorBlendEquationEXT *>( colorBlendEquations.data() ) );
22216 m_commandBuffer, firstAttachment, colorWriteMasks.size(), reinterpret_cast<const VkColorComponentFlags *>( colorWriteMasks.data() ) );
22280 m_commandBuffer, firstAttachment, colorBlendAdvanced.size(), reinterpret_cast<const VkColorBlendAdvancedEXT *>( colorBlendAdvanced.data() ) );
22343 m_commandBuffer, firstViewport, viewportSwizzles.size(), reinterpret_cast<const VkViewportSwizzleNV *>( viewportSwizzles.data() ) );
22394 d.vkCmdSetCoverageModulationTableNV( m_commandBuffer, coverageModulationTable.size(), coverageModulationTable.data() );
22517 VULKAN_HPP_ASSERT( formatCount <= imageFormatProperties.size() );
22518 if ( formatCount < imageFormatProperties.size() )
22556 VULKAN_HPP_ASSERT( formatCount <= imageFormatProperties.size() );
22557 if ( formatCount < imageFormatProperties.size() )
22735 VULKAN_HPP_NAMESPACE::DeviceSize size,
22743 static_cast<VkDeviceSize>( size ),
22883 std::vector<VULKAN_HPP_NAMESPACE::ShaderEXT, ShaderEXTAllocator> shaders( createInfos.size() );
22886 createInfos.size(),
22907 std::vector<VULKAN_HPP_NAMESPACE::ShaderEXT, ShaderEXTAllocator> shaders( createInfos.size(), shaderEXTAllocator );
22910 createInfos.size(),
22949 std::vector<VULKAN_HPP_NAMESPACE::ShaderEXT> shaders( createInfos.size() );
22952 createInfos.size(),
22958 uniqueShaders.reserve( createInfos.size() );
22980 std::vector<VULKAN_HPP_NAMESPACE::ShaderEXT> shaders( createInfos.size() );
22983 createInfos.size(),
22989 uniqueShaders.reserve( createInfos.size() );
23096 VULKAN_HPP_ASSERT( dataSize <= data.size() );
23097 if ( dataSize < data.size() )
23127 VULKAN_HPP_ASSERT( dataSize <= data.size() );
23128 if ( dataSize < data.size() )
23155 VULKAN_HPP_ASSERT( stages.size() == shaders.size() );
23157 if ( stages.size() != shaders.size() )
23159 throw LogicError( VULKAN_HPP_NAMESPACE_STRING "::CommandBuffer::bindShadersEXT: stages.size() != shaders.size()" );
23164 stages.size(),
23205 VULKAN_HPP_ASSERT( propertiesCount <= properties.size() );
23206 if ( propertiesCount < properties.size() )
23239 VULKAN_HPP_ASSERT( propertiesCount <= properties.size() );
23240 if ( propertiesCount < properties.size() )
23422 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
23423 if ( propertyCount < properties.size() )
23456 VULKAN_HPP_ASSERT( propertyCount <= properties.size() );
23457 if ( propertyCount < properties.size() )
23552 VULKAN_HPP_ASSERT( timeDomainCount <= timeDomains.size() );
23553 if ( timeDomainCount < timeDomains.size() )
23583 VULKAN_HPP_ASSERT( timeDomainCount <= timeDomains.size() );
23584 if ( timeDomainCount < timeDomains.size() )
23613 std::piecewise_construct, std::forward_as_tuple( timestampInfos.size() ), std::forward_as_tuple( 0 ) );
23617 m_device, timestampInfos.size(), reinterpret_cast<const VkCalibratedTimestampInfoKHR *>( timestampInfos.data() ), timestamps.data(), &maxDeviation ) );
23635 std::piecewise_construct, std::forward_as_tuple( timestampInfos.size(), uint64_tAllocator ), std::forward_as_tuple( 0 ) );
23639 m_device, timestampInfos.size(), reinterpret_cast<const VkCalibratedTimestampInfoKHR *>( timestampInfos.data() ), timestamps.data(), &maxDeviation ) );