Searched refs:cmdBuffer (Results 1 - 25 of 73) sorted by relevance

123

/external/deqp/external/vulkancts/modules/vulkan/geometry/
H A DvktGeometryBasicClass.cpp89 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
132 beginCommandBuffer(vk, *cmdBuffer);
141 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u,
152 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor);
155 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline);
159 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, DE_LENGTH_OF_ARRAY(buffers), buffers, offsets);
162 bindDescriptorSets(vk, device, memAlloc, *cmdBuffer, *pipelineLayout);
164 drawCommand (*cmdBuffer);
165 endRenderPass(vk, *cmdBuffer);
174 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI
210 drawCommand(const VkCommandBuffer& cmdBuffer) argument
[all...]
H A DvktGeometryBasicClass.hpp51 const vk::VkCommandBuffer& /*cmdBuffer*/,
53 virtual void drawCommand (const vk::VkCommandBuffer& cmdBuffer);
/external/deqp/external/vulkancts/modules/vulkan/query_pool/
H A DvktQueryPoolOcclusionTests.cpp370 vk::Unique<vk::VkCommandBuffer> cmdBuffer (vk::allocateCommandBuffer(vk, device, *cmdPool, vk::VK_COMMAND_BUFFER_LEVEL_PRIMARY));
373 vk.beginCommandBuffer(*cmdBuffer, &beginInfo);
375 transition2DImage(vk, *cmdBuffer, m_stateObjects->m_colorAttachmentImage->object(), vk::VK_IMAGE_ASPECT_COLOR_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, vk::VK_IMAGE_LAYOUT_GENERAL, 0, vk::VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT);
376 transition2DImage(vk, *cmdBuffer, m_stateObjects->m_DepthImage->object(), vk::VK_IMAGE_ASPECT_DEPTH_BIT, vk::VK_IMAGE_LAYOUT_UNDEFINED, vk::VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL, 0, vk::VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT);
389 vk.cmdResetQueryPool(*cmdBuffer, m_queryPool, 0, NUM_QUERIES_IN_POOL);
391 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBegin, vk::VK_SUBPASS_CONTENTS_INLINE);
393 vk.cmdBindPipeline(*cmdBuffer, vk::VK_PIPELINE_BIND_POINT_GRAPHICS, *m_stateObjects->m_pipeline);
397 vk.cmdBindVertexBuffers(*cmdBuffer, 0, 1, &vertexBuffer, &vertexBufferOffset);
399 vk.cmdBeginQuery(*cmdBuffer, m_queryPool, QUERY_INDEX_CAPTURE_EMPTY, m_testVector.queryControlFlags);
400 vk.cmdEndQuery(*cmdBuffer, m_queryPoo
[all...]
/external/deqp/external/vulkancts/modules/vulkan/image/
H A DvktImageCompressionTranscodingSupport.cpp398 void copyDataToImage (const VkCommandBuffer& cmdBuffer,
402 virtual void executeShader (const VkCommandBuffer& cmdBuffer,
406 bool copyResultAndCompare (const VkCommandBuffer& cmdBuffer,
415 bool decompressImage (const VkCommandBuffer& cmdBuffer,
434 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
488 copyDataToImage(*cmdBuffer, imageData[compressedNdx], mipMapSizes, true);
491 copyDataToImage(*cmdBuffer, imageData[1], mipMapSizes, false);
525 executeShader(*cmdBuffer, *descriptorSetLayout, *descriptorPool, imageData);
536 if (!copyResultAndCompare(*cmdBuffer, imageData[resultImageNdx].getImage(imageNdx), offset, size))
542 if (!decompressImage(*cmdBuffer, imageDat
547 copyDataToImage(const VkCommandBuffer& cmdBuffer, ImageData& imageData, const vector<UVec3>& mipMapSizes, const bool isCompressed) argument
616 executeShader(const VkCommandBuffer& cmdBuffer, const VkDescriptorSetLayout& descriptorSetLayout, const VkDescriptorPool& descriptorPool, vector<ImageData>& imageData) argument
721 copyResultAndCompare(const VkCommandBuffer& cmdBuffer, const VkImage& uncompressed, const VkDeviceSize offset, const UVec3& size) argument
879 decompressImage(const VkCommandBuffer& cmdBuffer, vector<ImageData>& imageData, const vector<UVec3>& mipMapSizes) argument
1188 executeShader(const VkCommandBuffer& cmdBuffer, const VkDescriptorSetLayout& descriptorSetLayout, const VkDescriptorPool& descriptorPool, vector<ImageData>& imageData) argument
[all...]
H A DvktImageLoadStoreTests.cpp258 void commandImageWriteBarrierBetweenShaderInvocations (Context& context, const VkCommandBuffer cmdBuffer, const VkImage image, const Texture& texture) argument
268 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 0, (const VkBufferMemoryBarrier*)DE_NULL, 1, &shaderWriteBarrier);
271 void commandBufferWriteBarrierBeforeHostRead (Context& context, const VkCommandBuffer cmdBuffer, const VkBuffer buffer, const VkDeviceSize bufferSizeBytes) argument
279 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &shaderWriteBarrier, 0, (const VkImageMemoryBarrier*)DE_NULL);
284 const VkCommandBuffer cmdBuffer,
304 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 0, (const VkBufferMemoryBarrier*)DE_NULL, 1, &prepareForTransferBarrier);
305 vk.cmdCopyImageToBuffer(cmdBuffer, image, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL, buffer, 1u, &copyRegion);
306 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &copyBarrier, 0, (const VkImageMemoryBarrier*)DE_NULL);
414 virtual void commandBeforeCompute (const VkCommandBuffer cmdBuffer) = 0;
415 virtual void commandBetweenShaderInvocations (const VkCommandBuffer cmdBuffer)
283 commandCopyImageToBuffer(Context& context, const VkCommandBuffer cmdBuffer, const VkImage image, const VkBuffer buffer, const VkDeviceSize bufferSizeBytes, const Texture& texture) argument
640 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument
661 commandBeforeCompute(const VkCommandBuffer cmdBuffer) argument
679 commandBetweenShaderInvocations(const VkCommandBuffer cmdBuffer) argument
684 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument
737 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument
751 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument
1008 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument
1027 commandBeforeCompute(const VkCommandBuffer cmdBuffer) argument
1065 commandBetweenShaderInvocations(const VkCommandBuffer cmdBuffer) argument
1070 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument
1142 commandBindDescriptorsForLayer(const VkCommandBuffer cmdBuffer, const VkPipelineLayout pipelineLayout, const int layerNdx) argument
1157 commandAfterCompute(const VkCommandBuffer cmdBuffer) argument
[all...]
H A DvktImageMultisampleLoadStoreTests.cpp322 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
333 beginCommandBuffer(vk, *cmdBuffer);
334 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline);
343 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, (VkDependencyFlags)0,
358 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &descriptorSet, 0u, DE_NULL);
359 vk.cmdDispatch(*cmdBuffer, workSize.x(), workSize.y(), workSize.z());
362 endCommandBuffer(vk, *cmdBuffer);
363 submitCommandsAndWait(vk, device, queue, *cmdBuffer);
371 beginCommandBuffer(vk, *cmdBuffer);
372 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUT
[all...]
H A DvktImageAtomicOperationTests.cpp407 virtual void commandsBeforeCompute (const VkCommandBuffer cmdBuffer) const = 0;
408 virtual void commandsAfterCompute (const VkCommandBuffer cmdBuffer) const = 0;
517 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(deviceInterface, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
519 beginCommandBuffer(deviceInterface, *cmdBuffer);
521 deviceInterface.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline);
522 deviceInterface.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &m_descriptorSet.get(), 0u, DE_NULL);
539 deviceInterface.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, DE_FALSE, 0u, DE_NULL, 1u, &inputBufferPostHostWriteBarrier, 1u, &resultImagePreCopyBarrier);
543 deviceInterface.cmdCopyBufferToImage(*cmdBuffer, *inputBuffer, m_resultImage->get(), VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1u, &bufferImageCopyParams);
553 deviceInterface.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, DE_FALSE, 0u, DE_NULL, 0u, DE_NULL, 1u, &resultImagePostCopyBarrier);
555 commandsBeforeCompute(*cmdBuffer);
[all...]
H A DvktImageQualifiersTests.cpp323 virtual void commandsBeforeCompute (const VkCommandBuffer cmdBuffer,
326 virtual void commandsAfterCompute (const VkCommandBuffer cmdBuffer,
380 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(deviceInterface, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
383 beginCommandBuffer(deviceInterface, *cmdBuffer);
385 deviceInterface.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline);
386 deviceInterface.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &m_descriptorSet.get(), 0u, DE_NULL);
388 commandsBeforeCompute(*cmdBuffer, bufferSizeInBytes);
391 deviceInterface.cmdDispatch(*cmdBuffer, numGroups.x(), numGroups.y(), numGroups.z());
393 commandsAfterCompute(*cmdBuffer, bufferSizeInBytes);
395 endCommandBuffer(deviceInterface, *cmdBuffer);
555 commandsBeforeCompute(const VkCommandBuffer cmdBuffer, const VkDeviceSize bufferSizeInBytes) const argument
573 commandsAfterCompute(const VkCommandBuffer cmdBuffer, const VkDeviceSize bufferSizeInBytes) const argument
666 commandsAfterCompute(const VkCommandBuffer cmdBuffer, const VkDeviceSize bufferSizeInBytes) const argument
[all...]
/external/deqp/external/vulkancts/modules/vulkan/synchronization/
H A DvktSynchronizationBasicSemaphoreTests.cpp54 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool));
72 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers;
83 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers;
90 VK_CHECK(vk.beginCommandBuffer(*cmdBuffer, &info));
91 endCommandBuffer(vk, *cmdBuffer);
135 Move<VkCommandBuffer> cmdBuffer[COUNT]; local
201 cmdBuffer[FIRST] = (makeCommandBuffer(vk, *logicalDevice, *cmdPool[FIRST]));
202 cmdBuffer[SECOND] = (makeCommandBuffer(vk, *logicalDevice, *cmdPool[SECOND]));
210 submitInfo[FIRST].pCommandBuffers = &cmdBuffer[FIRST].get();
220 submitInfo[SECOND].pCommandBuffers = &cmdBuffer[SECON
[all...]
H A DvktSynchronizationBasicFenceTests.cpp50 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool));
66 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers;
80 beginCommandBuffer(vk, *cmdBuffer);
81 endCommandBuffer(vk, *cmdBuffer);
108 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool));
132 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers;
138 VK_CHECK(vk.beginCommandBuffer(*cmdBuffer, &info));
139 endCommandBuffer(vk, *cmdBuffer);
H A DvktSynchronizationBasicEventTests.cpp87 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool));
96 &cmdBuffer.get(), // const VkCommandBuffer* pCommandBuffers;
102 beginCommandBuffer(vk, *cmdBuffer);
103 vk.cmdSetEvent(*cmdBuffer, *event, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT);
104 endCommandBuffer(vk, *cmdBuffer);
112 beginCommandBuffer(vk, *cmdBuffer);
113 vk.cmdResetEvent(*cmdBuffer, *event, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT);
114 endCommandBuffer(vk, *cmdBuffer);
133 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool));
142 &cmdBuffer
[all...]
H A DvktSynchronizationOperation.cpp83 virtual void recordCommands (OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) = 0;
500 void recordCommands (const VkCommandBuffer cmdBuffer) argument
505 vk.cmdFillBuffer(cmdBuffer, m_resource.getBuffer().handle, m_resource.getBuffer().offset, m_resource.getBuffer().size, m_fillValue);
507 vk.cmdUpdateBuffer(cmdBuffer, m_resource.getBuffer().handle, m_resource.getBuffer().offset, m_resource.getBuffer().size, reinterpret_cast<deUint32*>(&m_data[0]));
612 void recordCommands (const VkCommandBuffer cmdBuffer) argument
619 vk.cmdCopyBuffer(cmdBuffer, m_resource.getBuffer().handle, **m_hostBuffer, 1u, &copyRegion);
623 vk.cmdPipelineBarrier(cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0u, DE_NULL, 1u, &barrier, 0u, DE_NULL);
626 vk.cmdCopyBuffer(cmdBuffer, **m_hostBuffer, m_resource.getBuffer().handle, 1u, &copyRegion);
692 virtual void recordCopyCommand (const VkCommandBuffer cmdBuffer) = 0;
722 void recordCommands (const VkCommandBuffer cmdBuffer) argument
856 recordCopyCommand(const VkCommandBuffer cmdBuffer) argument
900 recordCopyCommand(const VkCommandBuffer cmdBuffer) argument
1025 recordCommands(OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) argument
1106 recordCommands(OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) argument
1205 recordCommands(const VkCommandBuffer cmdBuffer) argument
1352 recordCommands(const VkCommandBuffer cmdBuffer) argument
1787 recordCommands(const VkCommandBuffer cmdBuffer) argument
1853 recordCommands(const VkCommandBuffer cmdBuffer) argument
1990 recordCommands(const VkCommandBuffer cmdBuffer) argument
2065 recordCommands(const VkCommandBuffer cmdBuffer) argument
2165 recordCommands(const VkCommandBuffer cmdBuffer) argument
2340 recordCommands(const VkCommandBuffer cmdBuffer) argument
2615 recordCommands(const VkCommandBuffer cmdBuffer) argument
2792 recordCommands(OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) argument
2876 recordCommands(OperationContext& context, const VkCommandBuffer cmdBuffer, const VkDescriptorSet descriptorSet) argument
2943 recordCommands(const VkCommandBuffer cmdBuffer) argument
3035 recordCommands(const VkCommandBuffer cmdBuffer) argument
3258 recordCommands(const VkCommandBuffer cmdBuffer) argument
[all...]
H A DvktSynchronizationOperationSingleQueueTests.cpp83 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool));
88 beginCommandBuffer(vk, *cmdBuffer);
90 m_writeOp->recordCommands(*cmdBuffer);
91 vk.cmdSetEvent(*cmdBuffer, *event, writeSync.stageMask);
97 vk.cmdWaitEvents(*cmdBuffer, 1u, &event.get(), writeSync.stageMask, readSync.stageMask, 0u, DE_NULL, 1u, &barrier, 0u, DE_NULL);
103 vk.cmdWaitEvents(*cmdBuffer, 1u, &event.get(), writeSync.stageMask, readSync.stageMask, 0u, DE_NULL, 0u, DE_NULL, 1u, &barrier);
106 m_readOp->recordCommands(*cmdBuffer);
108 endCommandBuffer(vk, *cmdBuffer);
109 submitCommandsAndWait(vk, device, queue, *cmdBuffer);
138 const Move<VkCommandBuffer> cmdBuffer (makeCommandBuffe
[all...]
/external/deqp/external/vulkancts/modules/vulkan/draw/
H A DvktDrawNegativeViewportHeightTests.cpp252 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
258 vk.beginCommandBuffer(*cmdBuffer, &beginInfo);
261 vk.cmdSetViewport(*cmdBuffer, 0u, 1u, &viewport);
267 initialTransitionColor2DImage(vk, *cmdBuffer, m_colorTargetImage->object(), VK_IMAGE_LAYOUT_GENERAL);
268 vk.cmdClearColorImage(*cmdBuffer, m_colorTargetImage->object(), VK_IMAGE_LAYOUT_GENERAL, &clearColor, 1, &subresourceRange);
279 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT, 0, 1, &memBarrier, 0, DE_NULL, 0, DE_NULL);
285 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBegin, VK_SUBPASS_CONTENTS_INLINE);
291 vk.cmdBindVertexBuffers(*cmdBuffer, 0, 1, &buffer, &offset);
294 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *m_pipeline);
295 vk.cmdDraw(*cmdBuffer,
[all...]
/external/deqp/external/vulkancts/modules/vulkan/compute/
H A DvktComputeBasicComputeShaderTests.cpp231 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
235 beginCommandBuffer(vk, *cmdBuffer);
237 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipeline);
238 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_COMPUTE, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL);
240 vk.cmdDispatch(*cmdBuffer, m_workSize.x(), m_workSize.y(), m_workSize.z());
242 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COMPUTE_SHADER_BIT, VK_PIPELINE_STAGE_HOST_BIT, (VkDependencyFlags)0, 0, (const VkMemoryBarrier*)DE_NULL, 1, &computeFinishBarrier, 0, (const VkImageMemoryBarrier*)DE_NULL);
244 endCommandBuffer(vk, *cmdBuffer);
248 submitCommandsAndWait(vk, device, queue, *cmdBuffer);
402 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
406 beginCommandBuffer(vk, *cmdBuffer);
[all...]
/external/deqp/external/vulkancts/modules/vulkan/tessellation/
H A DvktTessellationGeometryPointSizeTests.cpp334 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer (vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
358 beginCommandBuffer(vk, *cmdBuffer);
366 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u,
378 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor);
381 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline);
383 vk.cmdDraw(*cmdBuffer, 1u, 1u, 0u, 0u);
384 endRenderPass(vk, *cmdBuffer);
393 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT, VK_PIPELINE_STAGE_TRANSFER_BIT, 0u,
398 vk.cmdCopyImageToBuffer(*cmdBuffer, *colorAttachmentImage, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL, *colorBuffer, 1u, &copyRegion);
404 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TRANSFER_BI
[all...]
H A DvktTessellationPrimitiveDiscardTests.cpp470 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
491 beginCommandBuffer(vk, *cmdBuffer);
500 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u,
512 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor);
515 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline);
516 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL);
519 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset);
522 vk.cmdDraw(*cmdBuffer, static_cast<deUint32>(attributes.size()), 1u, 0u, 0u);
523 endRenderPass(vk, *cmdBuffer);
532 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI
[all...]
H A DvktTessellationCommonEdgeTests.cpp368 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer (vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
387 beginCommandBuffer(vk, *cmdBuffer);
396 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u,
408 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor);
411 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline);
415 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, DE_LENGTH_OF_ARRAY(buffers), buffers, offsets);
417 vk.cmdBindIndexBuffer(*cmdBuffer, *vertexBuffer, vertexIndicesOffset, VK_INDEX_TYPE_UINT16);
420 vk.cmdDrawIndexed(*cmdBuffer, static_cast<deUint32>(gridIndices.size()), 1u, 0u, 0, 0u);
421 endRenderPass(vk, *cmdBuffer);
430 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI
[all...]
H A DvktTessellationGeometryPassthroughTests.cpp493 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
531 beginCommandBuffer(vk, *cmdBuffer);
544 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT | VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u,
556 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor);
559 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline);
562 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset);
566 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL);
568 vk.cmdDraw(*cmdBuffer, static_cast<deUint32>(m_params.vertices.size()), 1u, 0u, 0u);
569 endRenderPass(vk, *cmdBuffer);
578 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI
[all...]
H A DvktTessellationMiscDrawTests.cpp203 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
235 beginCommandBuffer(vk, *cmdBuffer);
248 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT | VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u,
260 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor);
263 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipeline);
264 vk.cmdBindDescriptorSets(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, *pipelineLayout, 0u, 1u, &descriptorSet.get(), 0u, DE_NULL);
267 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset);
271 vk.cmdDraw(*cmdBuffer, inPatchSize, 1u, 0u, 0u);
272 endRenderPass(vk, *cmdBuffer);
281 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI
[all...]
H A DvktTessellationWindingTests.cpp404 const Unique<VkCommandBuffer> cmdBuffer(allocateCommandBuffer(vk, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
413 beginCommandBuffer(vk, *cmdBuffer);
426 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT | VK_PIPELINE_STAGE_TRANSFER_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u,
438 beginRenderPass(vk, *cmdBuffer, *renderPass, *framebuffer, renderArea, clearColor);
450 vk.cmdSetViewport(*cmdBuffer, 0, 1, &viewport);
457 vk.cmdSetScissor(*cmdBuffer, 0, 1, &scissor);
459 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, testCases[caseNdx].pipeline);
462 vk.cmdDraw(*cmdBuffer, 1u, 1u, 0u, 0u);
463 endRenderPass(vk, *cmdBuffer);
472 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BI
[all...]
/external/deqp/external/vulkancts/modules/vulkan/ycbcr/
H A DvktYCbCrUtil.cpp311 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vkd, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
328 VK_CHECK(vkd.beginCommandBuffer(*cmdBuffer, &beginInfo));
346 vkd.cmdPipelineBarrier(*cmdBuffer,
379 vkd.cmdCopyBufferToImage(*cmdBuffer, **stagingBuffers[planeNdx], image, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1u, &copy);
397 vkd.cmdPipelineBarrier(*cmdBuffer,
409 VK_CHECK(vkd.endCommandBuffer(*cmdBuffer));
421 &*cmdBuffer,
442 const Unique<VkCommandBuffer> cmdBuffer (allocateCommandBuffer(vkd, device, *cmdPool, VK_COMMAND_BUFFER_LEVEL_PRIMARY));
485 VK_CHECK(vkd.beginCommandBuffer(*cmdBuffer, &beginInfo));
504 vkd.cmdPipelineBarrier(*cmdBuffer,
[all...]
/external/skia/src/gpu/vk/
H A DGrVkCommandBuffer.h122 GrVkCommandBuffer(VkCommandBuffer cmdBuffer, const GrVkRenderPass* rp = VK_NULL_HANDLE) argument
125 , fCmdBuffer(cmdBuffer)
277 explicit GrVkPrimaryCommandBuffer(VkCommandBuffer cmdBuffer) argument
278 : INHERITED(cmdBuffer)
306 explicit GrVkSecondaryCommandBuffer(VkCommandBuffer cmdBuffer) argument
307 : INHERITED(cmdBuffer) {
/external/deqp/external/vulkancts/modules/vulkan/pipeline/
H A DvktPipelineRenderToImageTests.cpp973 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool));
975 beginCommandBuffer(vk, *cmdBuffer);
997 vk.cmdPipelineBarrier(*cmdBuffer, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, 0u,
1000 VK_CHECK(vk.endCommandBuffer(*cmdBuffer));
1001 submitCommandsAndWait(vk, device, queue, *cmdBuffer);
1040 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool));
1042 beginCommandBuffer(vk, *cmdBuffer);
1066 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
1067 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset);
1074 vk.cmdNextSubpass(*cmdBuffer, VK_SUBPASS_CONTENTS_INLIN
[all...]
H A DvktPipelineMultisampleImageTests.cpp776 const Unique<VkCommandBuffer> cmdBuffer (makeCommandBuffer(vk, device, *cmdPool));
814 beginCommandBuffer(vk, *cmdBuffer);
827 vk.cmdBeginRenderPass(*cmdBuffer, &renderPassBeginInfo, VK_SUBPASS_CONTENTS_INLINE);
831 vk.cmdBindVertexBuffers(*cmdBuffer, 0u, 1u, &vertexBuffer.get(), &vertexBufferOffset);
837 vk.cmdNextSubpass(*cmdBuffer, VK_SUBPASS_CONTENTS_INLINE);
839 vk.cmdBindPipeline(*cmdBuffer, VK_PIPELINE_BIND_POINT_GRAPHICS, **pipelines[layerNdx]);
841 vk.cmdDraw(*cmdBuffer, static_cast<deUint32>(vertices.size()), 1u, 0u, 0u);
844 vk.cmdEndRenderPass(*cmdBuffer);
846 VK_CHECK(vk.endCommandBuffer(*cmdBuffer));
847 submitCommandsAndWait(vk, device, queue, *cmdBuffer);
[all...]

Completed in 1792 milliseconds

123