/external/deqp/external/vulkancts/modules/vulkan/geometry/ |
H A D | vktGeometryBasicClass.cpp | 89 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 132 beginCommandBuffer(vk, *cmdBuffer); 141 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u, 152 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor); 155 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline); 159 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, DE_LENGTH_OF_ARRAY(buffers), buffers, offsets); 162 bindDescriptorSets(vk, device, memAlloc, *cmdBuffer, *pipelineLayout); 164 drawCommand (*cmdBuffer); 165 endRenderPass(vk, *cmdBuffer); 174 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI 210 drawCommand(const VkCommandBuffer& cmdBuffer) argument [all...] |
H A D | vktGeometryBasicClass.hpp | 51 const vk::VkCommandBuffer& /*cmdBuffer*/, 53 virtual void drawCommand (const vk::VkCommandBuffer& cmdBuffer);
|
/external/deqp/external/vulkancts/modules/vulkan/query_pool/ |
H A D | vktQueryPoolOcclusionTests.cpp | 370 vk::Unique<vk::VkCommandBuffer> cmdBuffer (vk::allocateCommandBuffer(vk, device, *cmdPool, vk::VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 373 vk.beginCommandBuffer(*cmdBuffer, &beginInfo); 375 transition2DImage(vk, *cmdBuffer, m_stateObjects->m_colorAttachmentImage->object(), vk::VK_IMAGE_ASPECT_COLOR_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, vk::VK_IMAGE_LAYOUT_GENERAL, 0, vk::VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT); 376 transition2DImage(vk, *cmdBuffer, m_stateObjects->m_DepthImage->object(), vk::VK_IMAGE_ASPECT_DEPTH_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, vk::VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL, 0, vk::VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT); 389 vk.cmdResetQueryPool(*cmdBuffer, m_queryPool, 0, NUM_QUERIES_IN_POOL); 391 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBegin, vk::VK_SUBPASS_CONTENTS_INLINE); 393 vk.cmdBindPipeline(*cmdBuffer, vk::VK_PIPELINE_BIND_POINT_GRAPHICS, *m_stateObjects->m_pipeline); 397 vk.cmdBindVertexBuffers(*cmdBuffer, 0, 1, &vertexBuffer, &vertexBufferOffset); 399 vk.cmdBeginQuery(*cmdBuffer, m_queryPool, QUERY_INDEX_CAPTURE_EMPTY, m_testVector.queryControlFlags); 400 vk.cmdEndQuery(*cmdBuffer, m_queryPoo [all...] |
/external/deqp/external/vulkancts/modules/vulkan/image/ |
H A D | vktImageCompressionTranscodingSupport.cpp | 398 void copyDataToImage (const VkCommandBuffer& cmdBuffer, 402 virtual void executeShader (const VkCommandBuffer& cmdBuffer, 406 bool copyResultAndCompare (const VkCommandBuffer& cmdBuffer, 415 bool decompressImage (const VkCommandBuffer& cmdBuffer, 434 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 488 copyDataToImage(*cmdBuffer, imageData[compressedNdx], mipMapSizes, true); 491 copyDataToImage(*cmdBuffer, imageData[1], mipMapSizes, false); 525 executeShader(*cmdBuffer, *descriptorSetLayout, *descriptorPool, imageData); 536 if (!copyResultAndCompare(*cmdBuffer, imageData[resultImageNdx].getImage(imageNdx), offset, size)) 542 if (!decompressImage(*cmdBuffer, imageDat 547 copyDataToImage(const VkCommandBuffer& cmdBuffer, ImageData& imageData, const vector<UVec3>& mipMapSizes, const bool isCompressed) argument 616 executeShader(const VkCommandBuffer& cmdBuffer, const VkDescriptorSetLayout& descriptorSetLayout, const VkDescriptorPool& descriptorPool, vector<ImageData>& imageData) argument 721 copyResultAndCompare(const VkCommandBuffer& cmdBuffer, const VkImage& uncompressed, const VkDeviceSize offset, const UVec3& size) argument 879 decompressImage(const VkCommandBuffer& cmdBuffer, vector<ImageData>& imageData, const vector<UVec3>& mipMapSizes) argument 1188 executeShader(const VkCommandBuffer& cmdBuffer, const VkDescriptorSetLayout& descriptorSetLayout, const VkDescriptorPool& descriptorPool, vector<ImageData>& imageData) argument [all...] |
H A D | vktImageLoadStoreTests.cpp | 258 void commandImageWriteBarrierBetweenShaderInvocations (Context& context, const VkCommandBuffer cmdBuffer, const VkImage image, const Texture& texture) argument 268 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 0, (const VkBufferMemoryBarrier*)DE_NULL, 1, &shaderWriteBarrier); 271 void commandBufferWriteBarrierBeforeHostRead (Context& context, const VkCommandBuffer cmdBuffer, const VkBuffer buffer, const VkDeviceSize bufferSizeBytes) argument 279 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &shaderWriteBarrier, 0, (const VkImageMemoryBarrier*)DE_NULL); 284 const VkCommandBuffer cmdBuffer, 304 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 0, (const VkBufferMemoryBarrier*)DE_NULL, 1, &prepareForTransferBarrier); 305 vk.cmdCopyImageToBuffer(cmdBuffer, image, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL, buffer, 1u, ©Region); 306 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, ©Barrier, 0, (const VkImageMemoryBarrier*)DE_NULL); 414 virtual void commandBeforeCompute (const VkCommandBuffer cmdBuffer) = 0; 415 virtual void commandBetweenShaderInvocations (const VkCommandBuffer cmdBuffer) 283 commandCopyImageToBuffer(Context& context, const VkCommandBuffer cmdBuffer, const VkImage image, const VkBuffer buffer, const VkDeviceSize bufferSizeBytes, const Texture& texture) argument 640 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument 661 commandBeforeCompute(const VkCommandBuffer cmdBuffer) argument 679 commandBetweenShaderInvocations(const VkCommandBuffer cmdBuffer) argument 684 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument 737 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument 751 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument 1008 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument 1027 commandBeforeCompute(const VkCommandBuffer cmdBuffer) argument 1065 commandBetweenShaderInvocations(const VkCommandBuffer cmdBuffer) argument 1070 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument 1142 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument 1157 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument [all...] |
H A D | vktImageMultisampleLoadStoreTests.cpp | 322 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 333 beginCommandBuffer(vk, *cmdBuffer); 334 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline); 343 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, (VkDependencyFlags)0, 358 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &descriptorSet, 0u, DE_NULL); 359 vk.cmdDispatch(*cmdBuffer, workSize.x(), workSize.y(), workSize.z()); 362 endCommandBuffer(vk, *cmdBuffer); 363 submitCommandsAndWait(vk, device, queue, *cmdBuffer); 371 beginCommandBuffer(vk, *cmdBuffer); 372 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUT [all...] |
H A D | vktImageAtomicOperationTests.cpp | 407 virtual void commandsBeforeCompute (const VkCommandBuffer cmdBuffer) const = 0; 408 virtual void commandsAfterCompute (const VkCommandBuffer cmdBuffer) const = 0; 517 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(deviceInterface, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 519 beginCommandBuffer(deviceInterface, *cmdBuffer); 521 deviceInterface.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline); 522 deviceInterface.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &m_descriptorSet.get(), 0u, DE_NULL); 539 deviceInterface.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, DE_FALSE, 0u, DE_NULL, 1u, &inputBufferPostHostWriteBarrier, 1u, &resultImagePreCopyBarrier); 543 deviceInterface.cmdCopyBufferToImage(*cmdBuffer, *inputBuffer, m_resultImage->get(), VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1u, &bufferImageCopyParams); 553 deviceInterface.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, DE_FALSE, 0u, DE_NULL, 0u, DE_NULL, 1u, &resultImagePostCopyBarrier); 555 commandsBeforeCompute(*cmdBuffer); [all...] |
H A D | vktImageQualifiersTests.cpp | 323 virtual void commandsBeforeCompute (const VkCommandBuffer cmdBuffer, 326 virtual void commandsAfterCompute (const VkCommandBuffer cmdBuffer, 380 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(deviceInterface, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 383 beginCommandBuffer(deviceInterface, *cmdBuffer); 385 deviceInterface.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline); 386 deviceInterface.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &m_descriptorSet.get(), 0u, DE_NULL); 388 commandsBeforeCompute(*cmdBuffer, bufferSizeInBytes); 391 deviceInterface.cmdDispatch(*cmdBuffer, numGroups.x(), numGroups.y(), numGroups.z()); 393 commandsAfterCompute(*cmdBuffer, bufferSizeInBytes); 395 endCommandBuffer(deviceInterface, *cmdBuffer); 555 commandsBeforeCompute(const VkCommandBuffer cmdBuffer, const VkDeviceSize bufferSizeInBytes) const argument 573 commandsAfterCompute(const VkCommandBuffer cmdBuffer, const VkDeviceSize bufferSizeInBytes) const argument 666 commandsAfterCompute(const VkCommandBuffer cmdBuffer, const VkDeviceSize bufferSizeInBytes) const argument [all...] |
/external/deqp/external/vulkancts/modules/vulkan/synchronization/ |
H A D | vktSynchronizationBasicSemaphoreTests.cpp | 54 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 72 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers; 83 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers; 90 VK_CHECK(vk.beginCommandBuffer(*cmdBuffer, &info)); 91 endCommandBuffer(vk, *cmdBuffer); 135 Move<VkCommandBuffer> cmdBuffer[COUNT]; local 201 cmdBuffer[FIRST] = (makeCommandBuffer(vk, *logicalDevice, *cmdPool[FIRST])); 202 cmdBuffer[SECOND] = (makeCommandBuffer(vk, *logicalDevice, *cmdPool[SECOND])); 210 submitInfo[FIRST].pCommandBuffers = &cmdBuffer[FIRST].get(); 220 submitInfo[SECOND].pCommandBuffers = &cmdBuffer[SECON [all...] |
H A D | vktSynchronizationBasicFenceTests.cpp | 50 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 66 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers; 80 beginCommandBuffer(vk, *cmdBuffer); 81 endCommandBuffer(vk, *cmdBuffer); 108 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 132 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers; 138 VK_CHECK(vk.beginCommandBuffer(*cmdBuffer, &info)); 139 endCommandBuffer(vk, *cmdBuffer);
|
H A D | vktSynchronizationBasicEventTests.cpp | 87 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 96 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers; 102 beginCommandBuffer(vk, *cmdBuffer); 103 vk.cmdSetEvent(*cmdBuffer, *event, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT); 104 endCommandBuffer(vk, *cmdBuffer); 112 beginCommandBuffer(vk, *cmdBuffer); 113 vk.cmdResetEvent(*cmdBuffer, *event, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT); 114 endCommandBuffer(vk, *cmdBuffer); 133 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 142 &cmdBuffer [all...] |
H A D | vktSynchronizationOperation.cpp | 83 virtual void recordCommands (OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) = 0; 500 void recordCommands (const VkCommandBuffer cmdBuffer) argument 505 vk.cmdFillBuffer(cmdBuffer, m_resource.getBuffer().handle, m_resource.getBuffer().offset, m_resource.getBuffer().size, m_fillValue); 507 vk.cmdUpdateBuffer(cmdBuffer, m_resource.getBuffer().handle, m_resource.getBuffer().offset, m_resource.getBuffer().size, reinterpret_cast<deUint32*>(&m_data[0])); 612 void recordCommands (const VkCommandBuffer cmdBuffer) argument 619 vk.cmdCopyBuffer(cmdBuffer, m_resource.getBuffer().handle, **m_hostBuffer, 1u, ©Region); 623 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0u, DE_NULL, 1u, &barrier, 0u, DE_NULL); 626 vk.cmdCopyBuffer(cmdBuffer, **m_hostBuffer, m_resource.getBuffer().handle, 1u, ©Region); 692 virtual void recordCopyCommand (const VkCommandBuffer cmdBuffer) = 0; 722 void recordCommands (const VkCommandBuffer cmdBuffer) argument 856 recordCopyCommand(const VkCommandBuffer cmdBuffer) argument 900 recordCopyCommand(const VkCommandBuffer cmdBuffer) argument 1025 recordCommands(OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) argument 1106 recordCommands(OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) argument 1205 recordCommands(const VkCommandBuffer cmdBuffer) argument 1352 recordCommands(const VkCommandBuffer cmdBuffer) argument 1787 recordCommands(const VkCommandBuffer cmdBuffer) argument 1853 recordCommands(const VkCommandBuffer cmdBuffer) argument 1990 recordCommands(const VkCommandBuffer cmdBuffer) argument 2065 recordCommands(const VkCommandBuffer cmdBuffer) argument 2165 recordCommands(const VkCommandBuffer cmdBuffer) argument 2340 recordCommands(const VkCommandBuffer cmdBuffer) argument 2615 recordCommands(const VkCommandBuffer cmdBuffer) argument 2792 recordCommands(OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) argument 2876 recordCommands(OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) argument 2943 recordCommands(const VkCommandBuffer cmdBuffer) argument 3035 recordCommands(const VkCommandBuffer cmdBuffer) argument 3258 recordCommands(const VkCommandBuffer cmdBuffer) argument [all...] |
H A D | vktSynchronizationOperationSingleQueueTests.cpp | 83 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 88 beginCommandBuffer(vk, *cmdBuffer); 90 m_writeOp->recordCommands(*cmdBuffer); 91 vk.cmdSetEvent(*cmdBuffer, *event, writeSync.stageMask); 97 vk.cmdWaitEvents(*cmdBuffer, 1u, &event.get(), writeSync.stageMask, readSync.stageMask, 0u, DE_NULL, 1u, &barrier, 0u, DE_NULL); 103 vk.cmdWaitEvents(*cmdBuffer, 1u, &event.get(), writeSync.stageMask, readSync.stageMask, 0u, DE_NULL, 0u, DE_NULL, 1u, &barrier); 106 m_readOp->recordCommands(*cmdBuffer); 108 endCommandBuffer(vk, *cmdBuffer); 109 submitCommandsAndWait(vk, device, queue, *cmdBuffer); 138 const Move<VkCommandBuffer> cmdBuffer (makeCommandBuffe [all...] |
/external/deqp/external/vulkancts/modules/vulkan/draw/ |
H A D | vktDrawNegativeViewportHeightTests.cpp | 252 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 258 vk.beginCommandBuffer(*cmdBuffer, &beginInfo); 261 vk.cmdSetViewport(*cmdBuffer, 0u, 1u, &viewport); 267 initialTransitionColor2DImage(vk, *cmdBuffer, m_colorTargetImage->object(), VK_IMAGE_LAYOUT_GENERAL); 268 vk.cmdClearColorImage(*cmdBuffer, m_colorTargetImage->object(), VK_IMAGE_LAYOUT_GENERAL, &clearColor, 1, &subresourceRange); 279 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT, 0, 1, &memBarrier, 0, DE_NULL, 0, DE_NULL); 285 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBegin, VK_SUBPASS_CONTENTS_INLINE); 291 vk.cmdBindVertexBuffers(*cmdBuffer, 0, 1, &buffer, &offset); 294 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *m_pipeline); 295 vk.cmdDraw(*cmdBuffer, [all...] |
/external/deqp/external/vulkancts/modules/vulkan/compute/ |
H A D | vktComputeBasicComputeShaderTests.cpp | 231 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 235 beginCommandBuffer(vk, *cmdBuffer); 237 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline); 238 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL); 240 vk.cmdDispatch(*cmdBuffer, m_workSize.x(), m_workSize.y(), m_workSize.z()); 242 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &computeFinishBarrier, 0, (const VkImageMemoryBarrier*)DE_NULL); 244 endCommandBuffer(vk, *cmdBuffer); 248 submitCommandsAndWait(vk, device, queue, *cmdBuffer); 402 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 406 beginCommandBuffer(vk, *cmdBuffer); [all...] |
/external/deqp/external/vulkancts/modules/vulkan/tessellation/ |
H A D | vktTessellationGeometryPointSizeTests.cpp | 334 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer (vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 358 beginCommandBuffer(vk, *cmdBuffer); 366 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u, 378 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor); 381 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline); 383 vk.cmdDraw(*cmdBuffer, 1u, 1u, 0u, 0u); 384 endRenderPass(vk, *cmdBuffer); 393 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, 0u, 398 vk.cmdCopyImageToBuffer(*cmdBuffer, *colorAttachmentImage, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL, *colorBuffer, 1u, ©Region); 404 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BI [all...] |
H A D | vktTessellationPrimitiveDiscardTests.cpp | 470 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 491 beginCommandBuffer(vk, *cmdBuffer); 500 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u, 512 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor); 515 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline); 516 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL); 519 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset); 522 vk.cmdDraw(*cmdBuffer, static_cast<deUint32>(attributes.size()), 1u, 0u, 0u); 523 endRenderPass(vk, *cmdBuffer); 532 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI [all...] |
H A D | vktTessellationCommonEdgeTests.cpp | 368 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer (vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 387 beginCommandBuffer(vk, *cmdBuffer); 396 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u, 408 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor); 411 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline); 415 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, DE_LENGTH_OF_ARRAY(buffers), buffers, offsets); 417 vk.cmdBindIndexBuffer(*cmdBuffer, *vertexBuffer, vertexIndicesOffset, VK_INDEX_TYPE_UINT16); 420 vk.cmdDrawIndexed(*cmdBuffer, static_cast<deUint32>(gridIndices.size()), 1u, 0u, 0, 0u); 421 endRenderPass(vk, *cmdBuffer); 430 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI [all...] |
H A D | vktTessellationGeometryPassthroughTests.cpp | 493 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 531 beginCommandBuffer(vk, *cmdBuffer); 544 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT | VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u, 556 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor); 559 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline); 562 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset); 566 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL); 568 vk.cmdDraw(*cmdBuffer, static_cast<deUint32>(m_params.vertices.size()), 1u, 0u, 0u); 569 endRenderPass(vk, *cmdBuffer); 578 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI [all...] |
H A D | vktTessellationMiscDrawTests.cpp | 203 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 235 beginCommandBuffer(vk, *cmdBuffer); 248 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT | VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u, 260 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor); 263 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline); 264 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL); 267 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset); 271 vk.cmdDraw(*cmdBuffer, inPatchSize, 1u, 0u, 0u); 272 endRenderPass(vk, *cmdBuffer); 281 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI [all...] |
H A D | vktTessellationWindingTests.cpp | 404 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 413 beginCommandBuffer(vk, *cmdBuffer); 426 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT | VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u, 438 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor); 450 vk.cmdSetViewport(*cmdBuffer, 0, 1, &viewport); 457 vk.cmdSetScissor(*cmdBuffer, 0, 1, &scissor); 459 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, testCases[caseNdx].pipeline); 462 vk.cmdDraw(*cmdBuffer, 1u, 1u, 0u, 0u); 463 endRenderPass(vk, *cmdBuffer); 472 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI [all...] |
/external/deqp/external/vulkancts/modules/vulkan/ycbcr/ |
H A D | vktYCbCrUtil.cpp | 311 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vkd, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 328 VK_CHECK(vkd.beginCommandBuffer(*cmdBuffer, &beginInfo)); 346 vkd.cmdPipelineBarrier(*cmdBuffer, 379 vkd.cmdCopyBufferToImage(*cmdBuffer, **stagingBuffers[planeNdx], image, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1u, ©); 397 vkd.cmdPipelineBarrier(*cmdBuffer, 409 VK_CHECK(vkd.endCommandBuffer(*cmdBuffer)); 421 &*cmdBuffer, 442 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vkd, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY)); 485 VK_CHECK(vkd.beginCommandBuffer(*cmdBuffer, &beginInfo)); 504 vkd.cmdPipelineBarrier(*cmdBuffer, [all...] |
/external/skia/src/gpu/vk/ |
H A D | GrVkCommandBuffer.h | 122 GrVkCommandBuffer(VkCommandBuffer cmdBuffer, const GrVkRenderPass* rp = VK_NULL_HANDLE) argument 125 , fCmdBuffer(cmdBuffer) 277 explicit GrVkPrimaryCommandBuffer(VkCommandBuffer cmdBuffer) argument 278 : INHERITED(cmdBuffer) 306 explicit GrVkSecondaryCommandBuffer(VkCommandBuffer cmdBuffer) argument 307 : INHERITED(cmdBuffer) {
|
/external/deqp/external/vulkancts/modules/vulkan/pipeline/ |
H A D | vktPipelineRenderToImageTests.cpp | 973 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 975 beginCommandBuffer(vk, *cmdBuffer); 997 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u, 1000 VK_CHECK(vk.endCommandBuffer(*cmdBuffer)); 1001 submitCommandsAndWait(vk, device, queue, *cmdBuffer); 1040 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 1042 beginCommandBuffer(vk, *cmdBuffer); 1066 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE); 1067 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset); 1074 vk.cmdNextSubpass(*cmdBuffer, VK_SUBPASS_CONTENTS_INLIN [all...] |
H A D | vktPipelineMultisampleImageTests.cpp | 776 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool)); 814 beginCommandBuffer(vk, *cmdBuffer); 827 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE); 831 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset); 837 vk.cmdNextSubpass(*cmdBuffer, VK_SUBPASS_CONTENTS_INLINE); 839 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, **pipelines[layerNdx]); 841 vk.cmdDraw(*cmdBuffer, static_cast<deUint32>(vertices.size()), 1u, 0u, 0u); 844 vk.cmdEndRenderPass(*cmdBuffer); 846 VK_CHECK(vk.endCommandBuffer(*cmdBuffer)); 847 submitCommandsAndWait(vk, device, queue, *cmdBuffer); [all...] |