Lines Matching refs:vkDevice

138 	const VkDevice			vkDevice			= context.getDevice();
153 Move<VkBuffer> vertexBuffer = createBuffer(vk, vkDevice, &vertexBufferParams);
155 *pAlloc = context.getDefaultAllocator().allocate(getBufferMemoryRequirements(vk, vkDevice, *vertexBuffer), MemoryRequirement::HostVisible);
156 VK_CHECK(vk.bindBufferMemory(vkDevice, *vertexBuffer, (*pAlloc)->getMemory(), (*pAlloc)->getOffset()));
170 const VkDevice vkDevice = context.getDevice();
192 Move<VkImage> image = createImage(vk, vkDevice, &colorImageParams);
194 *pAlloc = context.getDefaultAllocator().allocate(getImageMemoryRequirements(vk, vkDevice, *image), MemoryRequirement::Any);
195 VK_CHECK(vk.bindImageMemory(vkDevice, *image, (*pAlloc)->getMemory(), (*pAlloc)->getOffset()));
244 const VkDevice vkDevice = m_context.getDevice();
248 m_cmdPool = createCommandPool(vk, vkDevice, VK_COMMAND_POOL_CREATE_TRANSIENT_BIT, queueFamilyIndex);
251 m_cmdBuffer = allocateCommandBuffer(vk, vkDevice, *m_cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY);
264 m_cache = createPipelineCache(vk, vkDevice, &pipelineCacheCreateInfo);
275 const VkDevice vkDevice = m_context.getDevice();
280 submitCommandsAndWait(vk, vkDevice, queue, m_cmdBuffer.get());
480 const VkDevice vkDevice = m_context.getDevice();
489 flushAlloc(vk, vkDevice, *m_vertexBufferMemory);
493 m_renderPassFramebuffer[PIPELINE_CACHE_NDX_NO_CACHE] = RenderPassWrapper(m_param->getPipelineConstructionType(), vk, vkDevice, m_colorFormat, m_depthFormat);
494 m_renderPassFramebuffer[PIPELINE_CACHE_NDX_CACHED] = RenderPassWrapper(m_param->getPipelineConstructionType(), vk, vkDevice, m_colorFormat, m_depthFormat);
577 m_colorAttachmentView[PIPELINE_CACHE_NDX_NO_CACHE] = createImageView(vk, vkDevice, &colorAttachmentViewParams);
580 m_colorAttachmentView[PIPELINE_CACHE_NDX_CACHED] = createImageView(vk, vkDevice, &colorAttachmentViewParams);
597 m_depthAttachmentView = createImageView(vk, vkDevice, &depthAttachmentViewParams);
625 m_renderPassFramebuffer[PIPELINE_CACHE_NDX_NO_CACHE].createFramebuffer(vk, vkDevice, &framebufferParams, images);
630 m_renderPassFramebuffer[PIPELINE_CACHE_NDX_CACHED].createFramebuffer(vk, vkDevice, &framebufferParams, images);
646 m_pipelineLayout = PipelineLayoutWrapper(m_param->getPipelineConstructionType(), vk, vkDevice, &pipelineLayoutParams);
725 const VkDevice vkDevice = m_context.getDevice();
728 auto createModule = [&vk, vkDevice, &postfix](Context& context, std::string shaderName)
730 return ShaderWrapper(vk, vkDevice, context.getBinaryCollection().get(shaderName + postfix), 0);
821 const VkDevice vkDevice = m_context.getDevice();
826 vkDevice,
834 vkDevice,
927 const VkDevice vkDevice = m_context.getDevice();
940 flushAlloc(vk, vkDevice, *m_inputBufferAlloc);
952 flushAlloc(vk, vkDevice, *m_outputBufferAlloc[ndx]);
959 const VkDevice vkDevice = m_context.getDevice();
967 m_descriptorSetLayout[ndx] = descLayoutBuilder.build(vk, vkDevice);
975 vkDevice,
988 m_descriptorSet[ndx] = allocateDescriptorSet(vk, vkDevice, &descriptorSetAllocInfo);
998 builder.update(vk, vkDevice);
1004 const VkDevice vkDevice = m_context.getDevice();
1015 m_computeShaderModule = createShaderModule(vk, vkDevice, &shaderModuleCreateInfo);
1021 const VkDevice vkDevice = m_context.getDevice();
1035 m_pipelineLayout[ndx] = createPipelineLayout(vk, vkDevice, &pipelineLayoutCreateInfo);
1059 m_pipeline[ndx] = createComputePipeline(vk, vkDevice, *m_cache, &pipelineCreateInfo);
1102 const VkDevice vkDevice = m_context.getDevice();
1105 invalidateAlloc(vk, vkDevice, *m_outputBufferAlloc[PIPELINE_CACHE_NDX_NO_CACHE]);
1107 invalidateAlloc(vk, vkDevice, *m_outputBufferAlloc[PIPELINE_CACHE_NDX_CACHED]);
1159 const VkDevice vkDevice = m_context.getDevice();
1165 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, DE_NULL));
1169 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, (void*)m_data));
1179 m_newCache = createPipelineCache(vk, vkDevice, &pipelineCacheCreateInfo);
1229 const VkDevice vkDevice = m_context.getDevice();
1234 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, DE_NULL));
1243 if (vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, (void*)m_data) != VK_INCOMPLETE)
1254 m_newCache = createPipelineCache(vk, vkDevice, &pipelineCacheCreateInfo);
1364 const VkDevice vkDevice = m_context.getDevice();
1367 m_cacheMerged = createPipelineCache(vki, vk, physicalDevice, vkDevice, mergeCacheParam->destCacheType);
1378 *pipelineCachePtr = createPipelineCache(vki, vk, physicalDevice, vkDevice, mergeCacheParam->srcCacheTypes[sourceIdx]);
1386 VK_CHECK(vk.mergePipelineCaches(vkDevice, *m_cacheMerged, static_cast<deUint32>(sourceCaches.size()), &sourceCaches[0]));
1519 const VkDevice vkDevice = m_context.getDevice();
1525 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, DE_NULL));
1532 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, (void*)m_data));
1592 const VkDevice vkDevice = m_context.getDevice();
1600 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, DE_NULL));
1610 if (vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, (void*)m_data) != VK_INCOMPLETE)
1623 if (vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, (void*)m_data) != VK_INCOMPLETE)
1679 const VkDevice vkDevice = m_context.getDevice();
1687 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, DE_NULL));
1693 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, (void*)m_data));
1706 const Unique<VkPipelineCache> pipelineCache (createPipelineCache(vk, vkDevice, &pipelineCacheCreateInfo));
1757 const VkDevice vkDevice = m_context.getDevice();
1765 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, DE_NULL));
1771 VK_CHECK(vk.getPipelineCacheData(vkDevice, *m_cache, (deUintptr*)&dataSize, (void*)m_data));
1800 const Unique<VkPipelineCache> pipelineCache (createPipelineCache(vk, vkDevice, &pipelineCacheCreateInfo));