/external/skia/src/gpu/vk/ |
H A D | GrVkResourceProvider.cpp | 66 GrVkCommandBuffer* cmdBuffer = GrVkCommandBuffer::Create(fGpu, fGpu->cmdPool()); local 67 fActiveCommandBuffers.push_back(cmdBuffer); 68 cmdBuffer->ref(); 69 return cmdBuffer;
|
H A D | GrVkCommandBuffer.h | 169 explicit GrVkCommandBuffer(VkCommandBuffer cmdBuffer) argument 171 , fCmdBuffer(cmdBuffer)
|
H A D | GrVkCommandBuffer.cpp | 27 VkCommandBuffer cmdBuffer; local 30 &cmdBuffer)); 34 return new GrVkCommandBuffer(cmdBuffer);
|
/external/deqp/external/vulkancts/modules/vulkan/query_pool/ |
H A D | vktQueryPoolOcclusionTests.cpp | 372 vk::Unique<vk::VkCommandBuffer> cmdBuffer (vk::allocateCommandBuffer(vk, device, &cmdBufferAllocateInfo)); 375 vk.beginCommandBuffer(*cmdBuffer, &beginInfo); 377 transition2DImage(vk, *cmdBuffer, m_stateObjects->m_colorAttachmentImage->object(), vk::VK_IMAGE_ASPECT_COLOR_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, vk::VK_IMAGE_LAYOUT_GENERAL, 0, vk::VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT); 378 transition2DImage(vk, *cmdBuffer, m_stateObjects->m_DepthImage->object(), vk::VK_IMAGE_ASPECT_DEPTH_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, vk::VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL, 0, vk::VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT); 391 vk.cmdResetQueryPool(*cmdBuffer, m_queryPool, 0, NUM_QUERIES_IN_POOL); 393 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBegin, vk::VK_SUBPASS_CONTENTS_INLINE); 395 vk.cmdBindPipeline(*cmdBuffer, vk::VK_PIPELINE_BIND_POINT_GRAPHICS, *m_stateObjects->m_pipeline); 399 vk.cmdBindVertexBuffers(*cmdBuffer, 0, 1, &vertexBuffer, &vertexBufferOffset); 401 vk.cmdBeginQuery(*cmdBuffer, m_queryPool, QUERY_INDEX_CAPTURE_EMPTY, m_testVector.queryControlFlags); 402 vk.cmdEndQuery(*cmdBuffer, m_queryPoo [all...] |
H A D | vktQueryPoolImageObjectUtil.hpp | 254 vk::VkCommandBuffer cmdBuffer, 262 void initialTransitionColor2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout); 264 void initialTransitionDepth2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout); 266 void initialTransitionStencil2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout); 268 void initialTransitionDepthStencil2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout);
|
/external/deqp/external/vulkancts/modules/vulkan/image/ |
H A D | vktImageLoadStoreTests.cpp | 357 void commandImageWriteBarrierBetweenShaderInvocations (Context& context, const VkCommandBuffer cmdBuffer, const VkImage image, const Texture& texture) argument 367 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 0, (const VkBufferMemoryBarrier*)DE_NULL, 1, &shaderWriteBarrier); 370 void commandBufferWriteBarrierBeforeHostRead (Context& context, const VkCommandBuffer cmdBuffer, const VkBuffer buffer, const VkDeviceSize bufferSizeBytes) argument 378 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &shaderWriteBarrier, 0, (const VkImageMemoryBarrier*)DE_NULL); 383 const VkCommandBuffer cmdBuffer, 403 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 0, (const VkBufferMemoryBarrier*)DE_NULL, 1, &prepareForTransferBarrier); 404 vk.cmdCopyImageToBuffer(cmdBuffer, image, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL, buffer, 1u, ©Region); 405 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, ©Barrier, 0, (const VkImageMemoryBarrier*)DE_NULL); 525 virtual void commandBeforeCompute (const VkCommandBuffer cmdBuffer) = 0; 526 virtual void commandBetweenShaderInvocations (const VkCommandBuffer cmdBuffer) 382 commandCopyImageToBuffer(Context& context, const VkCommandBuffer cmdBuffer, const VkImage image, const VkBuffer buffer, const VkDeviceSize bufferSizeBytes, const Texture& texture) argument 751 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument 772 commandBeforeCompute(const VkCommandBuffer cmdBuffer) argument 790 commandBetweenShaderInvocations(const VkCommandBuffer cmdBuffer) argument 795 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument 848 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument 862 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument 1119 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument 1138 commandBeforeCompute(const VkCommandBuffer cmdBuffer) argument 1176 commandBetweenShaderInvocations(const VkCommandBuffer cmdBuffer) argument 1181 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument 1253 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument 1268 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument [all...] |
H A D | vktImageAtomicOperationTests.cpp | 407 virtual void commandsBeforeCompute (const VkCommandBuffer cmdBuffer) const = 0; 408 virtual void commandsAfterCompute (const VkCommandBuffer cmdBuffer) const = 0; 517 const Unique<VkCommandBuffer> cmdBuffer(makeCommandBuffer(deviceInterface, device, *cmdPool)); 519 beginCommandBuffer(deviceInterface, *cmdBuffer); 521 deviceInterface.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline); 522 deviceInterface.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &m_descriptorSet.get(), 0u, DE_NULL); 539 deviceInterface.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, DE_FALSE, 0u, DE_NULL, 1u, &inputBufferPostHostWriteBarrier, 1u, &resultImagePreCopyBarrier); 543 deviceInterface.cmdCopyBufferToImage(*cmdBuffer, *inputBuffer, m_resultImage->get(), VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1u, &bufferImageCopyParams); 553 deviceInterface.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, DE_FALSE, 0u, DE_NULL, 0u, DE_NULL, 1u, &resultImagePostCopyBarrier); 555 commandsBeforeCompute(*cmdBuffer); [all...] |
H A D | vktImageQualifiersTests.cpp | 323 virtual void commandsBeforeCompute (const VkCommandBuffer cmdBuffer, 326 virtual void commandsAfterCompute (const VkCommandBuffer cmdBuffer, 380 const Unique<VkCommandBuffer> cmdBuffer(makeCommandBuffer(deviceInterface, device, *cmdPool)); 383 beginCommandBuffer(deviceInterface, *cmdBuffer); 385 deviceInterface.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline); 386 deviceInterface.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &m_descriptorSet.get(), 0u, DE_NULL); 388 commandsBeforeCompute(*cmdBuffer, bufferSizeInBytes); 391 deviceInterface.cmdDispatch(*cmdBuffer, numGroups.x(), numGroups.y(), numGroups.z()); 393 commandsAfterCompute(*cmdBuffer, bufferSizeInBytes); 395 endCommandBuffer(deviceInterface, *cmdBuffer); 555 commandsBeforeCompute(const VkCommandBuffer cmdBuffer, const VkDeviceSize bufferSizeInBytes) const argument 573 commandsAfterCompute(const VkCommandBuffer cmdBuffer, const VkDeviceSize bufferSizeInBytes) const argument 666 commandsAfterCompute(const VkCommandBuffer cmdBuffer, const VkDeviceSize bufferSizeInBytes) const argument [all...] |
H A D | vktImageSizeTests.cpp | 257 virtual void commandBeforeCompute (const VkCommandBuffer cmdBuffer) = 0; 307 const Unique<VkCommandBuffer> cmdBuffer(makeCommandBuffer(vk, device, *cmdPool)); 309 beginCommandBuffer(vk, *cmdBuffer); 311 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline); 312 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &descriptorSet, 0u, DE_NULL); 314 commandBeforeCompute(*cmdBuffer); 315 vk.cmdDispatch(*cmdBuffer, 1, 1, 1); 316 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &shaderWriteBarrier, 0, (const VkImageMemoryBarrier*)DE_NULL); 318 endCommandBuffer(vk, *cmdBuffer); 320 submitCommandsAndWait(vk, device, queue, *cmdBuffer); 399 commandBeforeCompute(const VkCommandBuffer cmdBuffer) argument [all...] |
H A D | vktImageTestsUtil.hpp | 164 const vk::VkCommandBuffer cmdBuffer); 167 const vk::VkCommandBuffer cmdBuffer); 172 const vk::VkCommandBuffer cmdBuffer);
|
/external/deqp/external/vulkancts/modules/vulkan/compute/ |
H A D | vktComputeBasicComputeShaderTests.cpp | 231 const Unique<VkCommandBuffer> cmdBuffer(makeCommandBuffer(vk, device, *cmdPool)); 235 beginCommandBuffer(vk, *cmdBuffer); 237 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline); 238 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL); 240 vk.cmdDispatch(*cmdBuffer, m_workSize.x(), m_workSize.y(), m_workSize.z()); 242 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &computeFinishBarrier, 0, (const VkImageMemoryBarrier*)DE_NULL); 244 endCommandBuffer(vk, *cmdBuffer); 248 submitCommandsAndWait(vk, device, queue, *cmdBuffer); 402 const Unique<VkCommandBuffer> cmdBuffer(makeCommandBuffer(vk, device, *cmdPool)); 406 beginCommandBuffer(vk, *cmdBuffer); [all...] |
H A D | vktComputeShaderBuiltinVarTests.cpp | 443 const Unique<VkCommandBuffer> cmdBuffer(makeCommandBuffer(m_vki, m_device, *cmdPool)); 446 beginCommandBuffer(m_vki, *cmdBuffer); 448 m_vki.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline); 461 m_vki.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL); 464 m_vki.cmdDispatch(*cmdBuffer, subCase.numWorkGroups()[0], subCase.numWorkGroups()[1], subCase.numWorkGroups()[2]); 466 m_vki.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 472 endCommandBuffer(m_vki, *cmdBuffer); 475 submitCommandsAndWait(m_vki, m_device, m_queue, *cmdBuffer);
|
H A D | vktComputeTestsUtil.hpp | 140 const vk::VkCommandBuffer cmdBuffer); 143 const vk::VkCommandBuffer cmdBuffer); 148 const vk::VkCommandBuffer cmdBuffer);
|
H A D | vktComputeIndirectComputeDispatchTests.cpp | 268 const vk::Unique<vk::VkCommandBuffer> cmdBuffer(makeCommandBuffer(m_device_interface, m_device, *cmdPool)); 271 beginCommandBuffer(m_device_interface, *cmdBuffer); 278 fillIndirectBufferData(*cmdBuffer, indirectBuffer); 281 m_device_interface.cmdBindPipeline(*cmdBuffer, vk::VK_PIPELINE_BIND_POINT_COMPUTE, *computePipeline); 302 m_device_interface.cmdBindDescriptorSets(*cmdBuffer, vk::VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &(**descriptorSets[cmdNdx]), 0u, DE_NULL); 305 m_device_interface.cmdDispatchIndirect(*cmdBuffer, *indirectBuffer, m_dispatchCommands[cmdNdx].m_offset); 311 m_device_interface.cmdPipelineBarrier(*cmdBuffer, vk::VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, vk::VK_PIPELINE_STAGE_HOST_BIT, (vk::VkDependencyFlags)0, 317 endCommandBuffer(m_device_interface, *cmdBuffer); 320 submitCommandsAndWait(m_device_interface, m_device, m_queue, *cmdBuffer);
|
/external/deqp/external/vulkancts/modules/vulkan/pipeline/ |
H A D | vktPipelineEarlyFragmentTests.cpp | 936 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 948 beginCommandBuffer(vk, *cmdBuffer); 962 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_HOST_BIT, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT, 0u, 967 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor, 0.5f, 0u); 969 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline); 970 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL); 971 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset); 975 commandClearStencilAttachment(vk, *cmdBuffer, makeOffset2D(0, 0), makeExtent2D(renderSize.x()/2, renderSize.y()), 1u); 977 vk.cmdDraw(*cmdBuffer, numVertices, 1u, 0u, 0u); 978 endRenderPass(vk, *cmdBuffer); [all...] |
H A D | vktPipelineImageUtil.cpp | 179 Move<VkCommandBuffer> cmdBuffer; 225 cmdBuffer = allocateCommandBuffer(vk, device, &cmdBufferAllocateInfo); 295 VK_CHECK(vk.beginCommandBuffer(*cmdBuffer, &cmdBufferBeginInfo)); 296 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 0, (const VkBufferMemoryBarrier*)DE_NULL, 1, &imageBarrier); 297 vk.cmdCopyImageToBuffer(*cmdBuffer, image, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL, *buffer, 1, ©Region); 298 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &bufferBarrier, 0, (const VkImageMemoryBarrier*)DE_NULL); 299 VK_CHECK(vk.endCommandBuffer(*cmdBuffer)); 309 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers; 336 Move<VkCommandBuffer> cmdBuffer; local 383 cmdBuffer [all...] |
/external/deqp/external/vulkancts/modules/vulkan/spirv_assembly/ |
H A D | vktSpvAsmComputeShaderCase.cpp | 359 Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vkdi, device, &cmdBufferCreateInfo)); 371 VK_CHECK(vkdi.beginCommandBuffer(*cmdBuffer, &cmdBufferBeginInfo)); 372 vkdi.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *computePipeline); 373 vkdi.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0, 1, &descriptorSet.get(), 0, DE_NULL); 374 vkdi.cmdDispatch(*cmdBuffer, numWorkGroups.x(), numWorkGroups.y(), numWorkGroups.z()); 375 VK_CHECK(vkdi.endCommandBuffer(*cmdBuffer)); 395 &cmdBuffer.get(),
|
/external/deqp/external/vulkancts/modules/vulkan/draw/ |
H A D | vktDrawImageObjectUtil.hpp | 263 void transition2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageAspectFlags aspectMask, vk::VkImageLayout oldLayout, vk::VkImageLayout newLayout); 265 void initialTransitionColor2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout); 267 void initialTransitionDepth2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout); 269 void initialTransitionStencil2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout); 271 void initialTransitionDepthStencil2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout);
|
H A D | vktDrawImageObjectUtil.cpp | 879 vk::VkCommandBuffer cmdBuffer, 901 vk.cmdPipelineBarrier(cmdBuffer, vk::VK_PIPELINE_STAGE_ALL_GRAPHICS_BIT, vk::VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, (vk::VkDependencyFlags)0, 907 void initialTransitionColor2DImage (const vk::DeviceInterface &vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout) argument 909 transition2DImage(vk, cmdBuffer, image, vk::VK_IMAGE_ASPECT_COLOR_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, layout); 912 void initialTransitionDepth2DImage (const vk::DeviceInterface &vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout) argument 914 transition2DImage(vk, cmdBuffer, image, vk::VK_IMAGE_ASPECT_DEPTH_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, layout); 917 void initialTransitionStencil2DImage (const vk::DeviceInterface &vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout) argument 919 transition2DImage(vk, cmdBuffer, image, vk::VK_IMAGE_ASPECT_STENCIL_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, layout); 922 void initialTransitionDepthStencil2DImage (const vk::DeviceInterface &vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout) argument 924 transition2DImage(vk, cmdBuffer, imag 878 transition2DImage(const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageAspectFlags aspectMask, vk::VkImageLayout oldLayout, vk::VkImageLayout newLayout) argument [all...] |
/external/deqp/external/vulkancts/modules/vulkan/dynamic_state/ |
H A D | vktDynamicStateImageObjectUtil.hpp | 265 vk::VkCommandBuffer cmdBuffer, 273 void initialTransitionColor2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout); 275 void initialTransitionDepth2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout); 277 void initialTransitionStencil2DImage (const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout); 280 vk::VkCommandBuffer cmdBuffer,
|
H A D | vktDynamicStateImageObjectUtil.cpp | 879 vk::VkCommandBuffer cmdBuffer, 903 vk.cmdPipelineBarrier(cmdBuffer, vk::VK_PIPELINE_STAGE_ALL_GRAPHICS_BIT, vk::VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, (vk::VkDependencyFlags)0, 909 void initialTransitionColor2DImage (const vk::DeviceInterface &vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout) argument 911 transition2DImage(vk, cmdBuffer, image, vk::VK_IMAGE_ASPECT_COLOR_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, layout); 914 void initialTransitionDepth2DImage (const vk::DeviceInterface &vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout) argument 916 transition2DImage(vk, cmdBuffer, image, vk::VK_IMAGE_ASPECT_DEPTH_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, layout); 919 void initialTransitionStencil2DImage (const vk::DeviceInterface &vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout) argument 921 transition2DImage(vk, cmdBuffer, image, vk::VK_IMAGE_ASPECT_STENCIL_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, layout); 925 vk::VkCommandBuffer cmdBuffer, 931 transition2DImage(vk, cmdBuffer, imag 878 transition2DImage(const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageAspectFlags aspectMask, vk::VkImageLayout oldLayout, vk::VkImageLayout newLayout, vk::VkAccessFlags srcAccessMask, vk::VkAccessFlags dstAccessMask) argument 924 initialTransitionDepthStencil2DImage(const vk::DeviceInterface& vk, vk::VkCommandBuffer cmdBuffer, vk::VkImage image, vk::VkImageLayout layout, vk::VkAccessFlags srcAccessMask, vk::VkAccessFlags dstAccessMask) argument [all...] |
/external/deqp/external/vulkancts/modules/vulkan/sparse_resources/ |
H A D | vktSparseResourcesTestsUtil.hpp | 179 const vk::VkCommandBuffer cmdBuffer); 182 const vk::VkCommandBuffer cmdBuffer); 186 const vk::VkCommandBuffer cmdBuffer, 196 const vk::VkCommandBuffer cmdBuffer,
|
/external/deqp/external/vulkancts/modules/vulkan/shaderexecutor/ |
H A D | vktShaderExecutor.cpp | 731 Move<VkCommandBuffer> cmdBuffer; local 1236 cmdBuffer = allocateCommandBuffer(vk, vkDevice, &cmdBufferParams); 1238 VK_CHECK(vk.beginCommandBuffer(*cmdBuffer, &cmdBufferBeginInfo)); 1240 vk.cmdPipelineBarrier(*cmdBuffer, vk::VK_PIPELINE_STAGE_ALL_GRAPHICS_BIT, vk::VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, (VkDependencyFlags)0, 1244 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE); 1246 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *graphicsPipeline); 1247 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipelineLayout, 0u, 1u, &*descriptorSet, 0u, DE_NULL); 1259 vk.cmdBindVertexBuffers(*cmdBuffer, 0, numberOfVertexAttributes, &buffers[0], &offsets[0]); 1260 vk.cmdDraw(*cmdBuffer, (deUint32)positions.size(), 1u, 0u, 0u); 1262 vk.cmdEndRenderPass(*cmdBuffer); 2084 Move<VkCommandBuffer> cmdBuffer; local 2213 Move<VkCommandBuffer> cmdBuffer; local 3157 Move<VkCommandBuffer> cmdBuffer; local [all...] |
/external/deqp/external/vulkancts/modules/vulkan/api/ |
H A D | vktApiCopiesAndBlittingTests.cpp | 232 Move<VkCommandBuffer> cmdBuffer; local 265 cmdBuffer = allocateCommandBuffer(vk, vkDevice, &cmdBufferAllocateInfo); 366 VK_CHECK(vk.beginCommandBuffer(*cmdBuffer, &cmdBufferBeginInfo)); 367 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_HOST_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &preBufferBarrier, 1, &preImageBarrier); 368 vk.cmdCopyBufferToImage(*cmdBuffer, *buffer, image, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1u, ©Region); 369 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 0, (const VkBufferMemoryBarrier*)DE_NULL, 1, &postImageBarrier); 370 VK_CHECK(vk.endCommandBuffer(*cmdBuffer)); 380 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers; 447 Move<VkCommandBuffer> cmdBuffer; local 484 cmdBuffer [all...] |
/external/deqp/external/vulkancts/modules/vulkan/shaderrender/ |
H A D | vktShaderRender.cpp | 851 Move<VkCommandBuffer> cmdBuffer = allocateCommandBuffer(vk, vkDevice, &cmdBufferParams); local 853 VK_CHECK(vk.beginCommandBuffer(*cmdBuffer, &cmdBufferBeginInfo)); 862 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, (VkDependencyFlags)0, 899 vk.cmdCopyImage(*cmdBuffer, srcImage, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL, dstImage, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1, &imageCopy); 905 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, (VkDependencyFlags)0, 910 VK_CHECK(vk.endCommandBuffer(*cmdBuffer)); 927 &cmdBuffer.get(), 1656 const Move<VkCommandBuffer> cmdBuffer = allocateCommandBuffer(vk, vkDevice, &cmdBufferParams); local 1680 &cmdBuffer.get(), 1685 VK_CHECK(vk.beginCommandBuffer(*cmdBuffer, [all...] |