Searched refs:gpu (Results 1 - 25 of 158) sorted by relevance

1234567

/external/skia/src/gpu/
H A DGrSemaphore.h21 void resetGpu(const GrGpu* gpu) { fGpu = gpu; } argument
24 explicit GrSemaphore(const GrGpu* gpu) : fGpu(gpu) {} argument
H A DGrOpFlushState.cpp13 GrOpFlushState::GrOpFlushState(GrGpu* gpu, GrResourceProvider* resourceProvider) argument
14 : fGpu(gpu)
17 , fVertexPool(gpu)
18 , fIndexPool(gpu)
/external/skia/src/gpu/vk/
H A DGrVkSemaphore.cpp18 sk_sp<GrVkSemaphore> GrVkSemaphore::Make(const GrVkGpu* gpu) { argument
25 GR_VK_CALL_ERRCHECK(gpu->vkInterface(),
26 CreateSemaphore(gpu->device(), &createInfo, nullptr, &semaphore));
28 return sk_sp<GrVkSemaphore>(new GrVkSemaphore(gpu, semaphore));
31 GrVkSemaphore::GrVkSemaphore(const GrVkGpu* gpu, VkSemaphore semaphore) : INHERITED(gpu) { argument
43 void GrVkSemaphore::Resource::freeGPUData(const GrVkGpu* gpu) const {
44 GR_VK_CALL(gpu->vkInterface(),
45 DestroySemaphore(gpu->device(), fSemaphore, nullptr));
H A DGrVkUniformBuffer.h18 static GrVkUniformBuffer* Create(GrVkGpu* gpu, size_t size);
19 static const GrVkResource* CreateResource(GrVkGpu* gpu, size_t size);
22 void* map(GrVkGpu* gpu) { argument
23 return this->vkMap(gpu);
25 void unmap(GrVkGpu* gpu) { argument
26 this->vkUnmap(gpu);
30 bool updateData(GrVkGpu* gpu, const void* src, size_t srcSizeInBytes, argument
32 return this->vkUpdateData(gpu, src, srcSizeInBytes, createdNewBuffer);
34 void release(const GrVkGpu* gpu) { this->vkRelease(gpu); } argument
51 GrVkUniformBuffer(GrVkGpu* gpu, const GrVkBuffer::Desc& desc, const GrVkUniformBuffer::Resource* resource) argument
[all...]
H A DGrVkDescriptorSet.cpp23 void GrVkDescriptorSet::freeGPUData(const GrVkGpu* gpu) const {
24 fPool->unref(gpu);
27 void GrVkDescriptorSet::onRecycle(GrVkGpu* gpu) const {
28 gpu->resourceProvider().recycleDescriptorSet(this, fHandle);
H A DGrVkUniformBuffer.cpp13 GrVkUniformBuffer* GrVkUniformBuffer::Create(GrVkGpu* gpu, size_t size) { argument
19 resource = gpu->resourceProvider().findOrCreateStandardUniformBufferResource();
21 resource = CreateResource(gpu, size);
31 GrVkUniformBuffer* buffer = new GrVkUniformBuffer(gpu, desc,
36 resource->unref(gpu);
42 const GrVkResource* GrVkUniformBuffer::CreateResource(GrVkGpu* gpu, size_t size) { argument
62 err = VK_CALL(gpu, CreateBuffer(gpu->device(), &bufInfo, nullptr, &buffer));
67 if (!GrVkMemory::AllocAndBindBufferMemory(gpu,
77 VK_CALL(gpu, DestroyBuffe
85 createResource(GrVkGpu* gpu, const GrVkBuffer::Desc& descriptor) argument
[all...]
H A DGrVkDescriptorSetManager.h29 GrVkDescriptorSetManager(GrVkGpu* gpu,
33 GrVkDescriptorSetManager(GrVkGpu* gpu,
40 void release(const GrVkGpu* gpu);
44 const GrVkDescriptorSet* getDescriptorSet(GrVkGpu* gpu, const Handle& handle);
54 DescriptorPoolManager(VkDescriptorType type, GrVkGpu* gpu,
56 DescriptorPoolManager(VkDescriptorType type, GrVkGpu* gpu,
65 void getNewDescriptorSet(GrVkGpu* gpu, VkDescriptorSet* ds);
67 void freeGPUResources(const GrVkGpu* gpu);
84 void init(GrVkGpu* gpu, VkDescriptorType type, const GrVkUniformHandler* uniformHandler,
87 void getNewPool(GrVkGpu* gpu);
[all...]
H A DGrVkTextureRenderTarget.h38 bool updateForMipmap(GrVkGpu* gpu, const GrVkImageInfo& newInfo);
52 GrVkTextureRenderTarget(GrVkGpu* gpu, argument
60 : GrSurface(gpu, desc)
62 , GrVkTexture(gpu, desc, info, texView, GrVkImage::kNot_Wrapped)
63 , GrVkRenderTarget(gpu, desc, info, msaaInfo, colorAttachmentView,
68 GrVkTextureRenderTarget(GrVkGpu* gpu, argument
74 : GrSurface(gpu, desc)
76 , GrVkTexture(gpu, desc, info, texView, GrVkImage::kNot_Wrapped)
77 , GrVkRenderTarget(gpu, desc, info, colorAttachmentView, GrVkImage::kNot_Wrapped) {
80 GrVkTextureRenderTarget(GrVkGpu* gpu, argument
96 GrVkTextureRenderTarget(GrVkGpu* gpu, const GrSurfaceDesc& desc, const GrVkImageInfo& info, const GrVkImageView* texView, const GrVkImageView* colorAttachmentView, GrVkImage::Wrapped wrapped) argument
[all...]
H A DGrVkImageView.cpp12 const GrVkImageView* GrVkImageView::Create(const GrVkGpu* gpu, VkImage image, VkFormat format, argument
34 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreateImageView(gpu->device(), &viewInfo,
43 void GrVkImageView::freeGPUData(const GrVkGpu* gpu) const {
44 GR_VK_CALL(gpu->vkInterface(), DestroyImageView(gpu->device(), fImageView, nullptr));
H A DGrVkDescriptorPool.cpp14 GrVkDescriptorPool::GrVkDescriptorPool(const GrVkGpu* gpu, VkDescriptorType type, uint32_t count) argument
33 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), CreateDescriptorPool(gpu->device(),
43 void GrVkDescriptorPool::reset(const GrVkGpu* gpu) { argument
44 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), ResetDescriptorPool(gpu->device(), fDescPool, 0));
47 void GrVkDescriptorPool::freeGPUData(const GrVkGpu* gpu) const {
50 GR_VK_CALL(gpu->vkInterface(), DestroyDescriptorPool(gpu->device(), fDescPool, nullptr));
H A DGrVkCommandBuffer.cpp37 void GrVkCommandBuffer::freeGPUData(const GrVkGpu* gpu) const {
40 fTrackedResources[i]->unref(gpu);
44 fTrackedRecycledResources[i]->recycle(const_cast<GrVkGpu*>(gpu));
47 GR_VK_CALL(gpu->vkInterface(), FreeCommandBuffers(gpu->device(), gpu->cmdPool(),
50 this->onFreeGPUData(gpu);
64 void GrVkCommandBuffer::reset(GrVkGpu* gpu) { argument
67 fTrackedResources[i]->unref(gpu);
70 fTrackedRecycledResources[i]->recycle(const_cast<GrVkGpu*>(gpu));
98 pipelineBarrier(const GrVkGpu* gpu, VkPipelineStageFlags srcStageMask, VkPipelineStageFlags dstStageMask, bool byRegion, BarrierType barrierType, void* barrier) const argument
148 clearAttachments(const GrVkGpu* gpu, int numAttachments, const VkClearAttachment* attachments, int numRects, const VkClearRect* clearRects) const argument
173 bindDescriptorSets(const GrVkGpu* gpu, GrVkPipelineState* pipelineState, VkPipelineLayout layout, uint32_t firstSet, uint32_t setCount, const VkDescriptorSet* descriptorSets, uint32_t dynamicOffsetCount, const uint32_t* dynamicOffsets) argument
193 bindDescriptorSets(const GrVkGpu* gpu, const SkTArray<const GrVkRecycledResource*>& recycled, const SkTArray<const GrVkResource*>& resources, VkPipelineLayout layout, uint32_t firstSet, uint32_t setCount, const VkDescriptorSet* descriptorSets, uint32_t dynamicOffsetCount, const uint32_t* dynamicOffsets) argument
219 bindPipeline(const GrVkGpu* gpu, const GrVkPipeline* pipeline) argument
227 drawIndexed(const GrVkGpu* gpu, uint32_t indexCount, uint32_t instanceCount, uint32_t firstIndex, int32_t vertexOffset, uint32_t firstInstance) const argument
243 draw(const GrVkGpu* gpu, uint32_t vertexCount, uint32_t instanceCount, uint32_t firstVertex, uint32_t firstInstance) const argument
257 setViewport(const GrVkGpu* gpu, uint32_t firstViewport, uint32_t viewportCount, const VkViewport* viewports) argument
272 setScissor(const GrVkGpu* gpu, uint32_t firstScissor, uint32_t scissorCount, const VkRect2D* scissors) argument
287 setBlendConstants(const GrVkGpu* gpu, const float blendConstants[4]) argument
304 Create(const GrVkGpu* gpu, VkCommandPool cmdPool) argument
324 begin(const GrVkGpu* gpu) argument
338 end(const GrVkGpu* gpu) argument
346 beginRenderPass(const GrVkGpu* gpu, const GrVkRenderPass* renderPass, const VkClearValue* clearValues, const GrVkRenderTarget& target, const SkIRect& bounds, bool forSecondaryCB) argument
379 endRenderPass(const GrVkGpu* gpu) argument
386 executeCommands(const GrVkGpu* gpu, GrVkSecondaryCommandBuffer* buffer) argument
401 submitToQueue( const GrVkGpu* gpu, VkQueue queue, GrVkGpu::SyncQueue sync, const GrVkSemaphore::Resource* signalSemaphore, SkTArray<const GrVkSemaphore::Resource*>& waitSemaphores) argument
490 onReset(GrVkGpu* gpu) argument
497 copyImage(const GrVkGpu* gpu, GrVkImage* srcImage, VkImageLayout srcLayout, GrVkImage* dstImage, VkImageLayout dstLayout, uint32_t copyRegionCount, const VkImageCopy* copyRegions) argument
517 blitImage(const GrVkGpu* gpu, const GrVkResource* srcResource, VkImage srcImage, VkImageLayout srcLayout, const GrVkResource* dstResource, VkImage dstImage, VkImageLayout dstLayout, uint32_t blitRegionCount, const VkImageBlit* blitRegions, VkFilter filter) argument
541 copyImageToBuffer(const GrVkGpu* gpu, GrVkImage* srcImage, VkImageLayout srcLayout, GrVkTransferBuffer* dstBuffer, uint32_t copyRegionCount, const VkBufferImageCopy* copyRegions) argument
559 copyBufferToImage(const GrVkGpu* gpu, GrVkTransferBuffer* srcBuffer, GrVkImage* dstImage, VkImageLayout dstLayout, uint32_t copyRegionCount, const VkBufferImageCopy* copyRegions) argument
577 updateBuffer(GrVkGpu* gpu, GrVkBuffer* dstBuffer, VkDeviceSize dstOffset, VkDeviceSize dataSize, const void* data) argument
596 clearColorImage(const GrVkGpu* gpu, GrVkImage* image, const VkClearColorValue* color, uint32_t subRangeCount, const VkImageSubresourceRange* subRanges) argument
612 clearDepthStencilImage(const GrVkGpu* gpu, GrVkImage* image, const VkClearDepthStencilValue* color, uint32_t subRangeCount, const VkImageSubresourceRange* subRanges) argument
628 resolveImage(GrVkGpu* gpu, const GrVkImage& srcImage, const GrVkImage& dstImage, uint32_t regionCount, const VkImageResolve* regions) argument
660 Create(const GrVkGpu* gpu, VkCommandPool cmdPool) argument
681 begin(const GrVkGpu* gpu, const GrVkFramebuffer* framebuffer, const GrVkRenderPass* compatibleRenderPass) argument
711 end(const GrVkGpu* gpu) argument
[all...]
H A DGrVkBuffer.h35 void addMemoryBarrier(const GrVkGpu* gpu,
72 void freeGPUData(const GrVkGpu* gpu) const override;
74 void onRecycle(GrVkGpu* gpu) const override { this->unref(gpu); }
80 static const Resource* Create(const GrVkGpu* gpu,
87 void* vkMap(GrVkGpu* gpu) { argument
88 this->internalMap(gpu, fDesc.fSizeInBytes);
91 void vkUnmap(GrVkGpu* gpu) { this->internalUnmap(gpu, this->size()); } argument
95 bool vkUpdateData(GrVkGpu* gpu, cons
102 createResource(GrVkGpu* gpu, const Desc& descriptor) argument
[all...]
H A DGrVkCommandBuffer.h36 void pipelineBarrier(const GrVkGpu* gpu,
43 void bindVertexBuffer(GrVkGpu* gpu, GrVkVertexBuffer* vbuffer) { argument
49 GR_VK_CALL(gpu->vkInterface(), CmdBindVertexBuffers(fCmdBuffer,
60 void bindIndexBuffer(GrVkGpu* gpu, GrVkIndexBuffer* ibuffer) { argument
65 GR_VK_CALL(gpu->vkInterface(), CmdBindIndexBuffer(fCmdBuffer,
75 void bindPipeline(const GrVkGpu* gpu, const GrVkPipeline* pipeline);
77 void bindDescriptorSets(const GrVkGpu* gpu,
86 void bindDescriptorSets(const GrVkGpu* gpu,
96 void setViewport(const GrVkGpu* gpu,
101 void setScissor(const GrVkGpu* gpu,
178 onReset(GrVkGpu* gpu) argument
258 blitImage(const GrVkGpu* gpu, const GrVkImage& srcImage, const GrVkImage& dstImage, uint32_t blitRegionCount, const VkImageBlit* blitRegions, VkFilter filter) argument
[all...]
H A DGrVkRenderTarget.cpp24 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, argument
32 : GrSurface(gpu, desc)
35 , GrRenderTarget(gpu, desc)
42 this->createFramebuffer(gpu);
48 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, argument
55 : GrSurface(gpu, desc)
58 , GrRenderTarget(gpu, desc)
65 this->createFramebuffer(gpu);
70 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, argument
76 : GrSurface(gpu, des
91 GrVkRenderTarget(GrVkGpu* gpu, const GrSurfaceDesc& desc, const GrVkImageInfo& info, const GrVkImageView* colorAttachmentView, GrVkImage::Wrapped wrapped) argument
109 Create(GrVkGpu* gpu, SkBudgeted budgeted, const GrSurfaceDesc& desc, const GrVkImageInfo& info, GrVkImage::Wrapped wrapped) argument
179 CreateNewRenderTarget(GrVkGpu* gpu, SkBudgeted budgeted, const GrSurfaceDesc& desc, const GrVkImage::ImageDesc& imageDesc) argument
199 MakeWrappedRenderTarget(GrVkGpu* gpu, const GrSurfaceDesc& desc, const GrVkImageInfo* info) argument
214 createFramebuffer(GrVkGpu* gpu) argument
280 GrVkGpu* gpu = this->getVkGpu(); local
[all...]
H A DGrVkBuffer.cpp21 const GrVkBuffer::Resource* GrVkBuffer::Create(const GrVkGpu* gpu, const Desc& desc) { argument
57 err = VK_CALL(gpu, CreateBuffer(gpu->device(), &bufInfo, nullptr, &buffer));
62 if (!GrVkMemory::AllocAndBindBufferMemory(gpu,
72 VK_CALL(gpu, DestroyBuffer(gpu->device(), buffer, nullptr));
73 GrVkMemory::FreeBufferMemory(gpu, desc.fType, alloc);
80 void GrVkBuffer::addMemoryBarrier(const GrVkGpu* gpu, argument
99 gpu->addBufferMemoryBarrier(srcStageMask, dstStageMask, byRegion, &bufferMemoryBarrier);
102 void GrVkBuffer::Resource::freeGPUData(const GrVkGpu* gpu) cons
109 vkRelease(const GrVkGpu* gpu) argument
144 internalMap(GrVkGpu* gpu, size_t size, bool* createdNewBuffer) argument
184 internalUnmap(GrVkGpu* gpu, size_t size) argument
208 vkUpdateData(GrVkGpu* gpu, const void* src, size_t srcSizeInBytes, bool* createdNewBuffer) argument
[all...]
H A DGrVkImage.cpp28 void GrVkImage::setImageLayout(const GrVkGpu* gpu, VkImageLayout newLayout, argument
62 gpu->addImageMemoryBarrier(srcStageMask, dstStageMask, byRegion, &imageMemoryBarrier);
67 bool GrVkImage::InitImageInfo(const GrVkGpu* gpu, const ImageDesc& imageDesc, GrVkImageInfo* info) { argument
109 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), CreateImage(gpu->device(), &imageCreateInfo, nullptr,
112 if (!GrVkMemory::AllocAndBindImageMemory(gpu, image, isLinear, &alloc)) {
113 VK_CALL(gpu, DestroyImage(gpu->device(), image, nullptr));
126 void GrVkImage::DestroyImageInfo(const GrVkGpu* gpu, GrVkImageInfo* info) { argument
127 VK_CALL(gpu, DestroyImag
141 releaseImage(const GrVkGpu* gpu) argument
[all...]
H A DGrVkCopyManager.cpp35 bool GrVkCopyManager::createCopyProgram(GrVkGpu* gpu) { argument
36 const GrShaderCaps* shaderCaps = gpu->caps()->shaderCaps();
77 if (!GrCompileVkShaderModule(gpu, vertShaderText.c_str(), VK_SHADER_STAGE_VERTEX_BIT,
79 this->destroyResources(gpu);
84 if (!GrCompileVkShaderModule(gpu, fragShaderText.c_str(), VK_SHADER_STAGE_FRAGMENT_BIT,
86 this->destroyResources(gpu);
93 GrVkResourceProvider& resourceProvider = gpu->resourceProvider();
115 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreatePipelineLayout(gpu->device(),
120 this->destroyResources(gpu);
141 copySurfaceAsDraw(GrVkGpu* gpu, GrSurface* dst, GrSurface* src, const SkIRect& srcRect, const SkIPoint& dstPoint) argument
381 destroyResources(GrVkGpu* gpu) argument
[all...]
H A DGrVkCopyManager.h29 bool copySurfaceAsDraw(GrVkGpu* gpu,
35 void destroyResources(GrVkGpu* gpu);
39 bool createCopyProgram(GrVkGpu* gpu);
H A DGrVkDescriptorPool.h24 GrVkDescriptorPool(const GrVkGpu* gpu, VkDescriptorType type, uint32_t count);
28 void reset(const GrVkGpu* gpu);
42 void freeGPUData(const GrVkGpu* gpu) const override;
/external/autotest/client/site_tests/graphics_dEQP/scripts/
H A Dfailure_matrix.py62 def get_problem_count(dict, gpu):
63 if gpu in dict:
64 if not dict[gpu]:
67 for status in dict[gpu]:
69 count = count + len((dict[gpu])[status])
76 if status in dict[gpu]:
77 count = count + len((dict[gpu])[status])
79 print 'Warning: %s not found in dict!' % gpu
85 for gpu in dict:
87 if status in dict[gpu]
[all...]
/external/skia/src/gpu/gl/
H A DGrGLSemaphore.h17 static sk_sp<GrGLSemaphore> Make(const GrGLGpu* gpu) { argument
18 return sk_sp<GrGLSemaphore>(new GrGLSemaphore(gpu));
31 GrGLSemaphore(const GrGLGpu* gpu) : INHERITED(gpu), fSync(0) {} argument
H A DGrGLTextureRenderTarget.h29 GrGLTextureRenderTarget(GrGLGpu* gpu, argument
35 : GrSurface(gpu, desc)
36 , GrGLTexture(gpu, desc, texIDDesc, wasMipMapDataProvided)
37 , GrGLRenderTarget(gpu, desc, rtIDDesc) {
45 static sk_sp<GrGLTextureRenderTarget> MakeWrapped(GrGLGpu* gpu, const GrSurfaceDesc& desc,
61 GrGLTextureRenderTarget(GrGLGpu* gpu, argument
66 : GrSurface(gpu, desc)
67 , GrGLTexture(gpu, desc, texIDDesc, wasMipMapDataProvided)
68 , GrGLRenderTarget(gpu, desc, rtIDDesc) {
H A DGrGLTexture.cpp21 const GrGLGpu* gpu) {
23 SkASSERT(gpu->caps()->shaderCaps()->externalTextureSupport());
27 SkASSERT(gpu->glCaps().rectangleTextureSupport());
53 GrGLTexture::GrGLTexture(GrGLGpu* gpu, SkBudgeted budgeted, const GrSurfaceDesc& desc, argument
55 : GrSurface(gpu, desc)
56 , INHERITED(gpu, desc, sampler_type(idDesc, desc.fConfig, gpu),
62 GrGLTexture::GrGLTexture(GrGLGpu* gpu, SkBudgeted budgeted, const GrSurfaceDesc& desc, argument
65 : GrSurface(gpu, desc)
66 , INHERITED(gpu, des
20 sampler_type(const GrGLTexture::IDDesc& idDesc, GrPixelConfig config, const GrGLGpu* gpu) argument
73 GrGLTexture(GrGLGpu* gpu, Wrapped, const GrSurfaceDesc& desc, const IDDesc& idDesc) argument
81 GrGLTexture(GrGLGpu* gpu, const GrSurfaceDesc& desc, const IDDesc& idDesc, bool wasMipMapDataProvided) argument
149 MakeWrapped(GrGLGpu* gpu, const GrSurfaceDesc& desc, const IDDesc& idDesc) argument
[all...]
H A DGrGLPathRange.cpp13 GrGLPathRange::GrGLPathRange(GrGLGpu* gpu, PathGenerator* pathGenerator, const GrStyle& style) argument
14 : INHERITED(gpu, pathGenerator),
16 fBasePathID(gpu->glPathRendering()->genPaths(this->getNumPaths())),
22 GrGLPathRange::GrGLPathRange(GrGLGpu* gpu, argument
27 : INHERITED(gpu, numPaths),
54 GrGLGpu* gpu = static_cast<GrGLGpu*>(this->getGpu()); local
55 if (nullptr == gpu) {
61 GR_GL_CALL_RET(gpu->glInterface(), isPath, IsPath(fBasePathID + index)));
65 GrGLPath::InitPathObjectEmptyPath(gpu, fBasePathID + index);
67 GrGLPath::InitPathObjectPathData(gpu, fBasePathI
[all...]
/external/chromium-trace/catapult/telemetry/telemetry/internal/platform/
H A Dsystem_info_unittest.py16 'gpu': {
25 self.assertTrue(isinstance(info.gpu, gpu_info.GPUInfo))
27 self.assertTrue(len(info.gpu.devices) == 1)
28 self.assertTrue(isinstance(info.gpu.devices[0], gpu_device.GPUDevice))
29 self.assertEquals(info.gpu.devices[0].vendor_id, 1000)
30 self.assertEquals(info.gpu.devices[0].device_id, 2000)
31 self.assertEquals(info.gpu.devices[0].vendor_string, 'a')
32 self.assertEquals(info.gpu.devices[0].device_string, 'b')
37 'gpu': {

Completed in 310 milliseconds

1234567