/external/skia/src/gpu/ |
H A D | GrSemaphore.h | 21 void resetGpu(const GrGpu* gpu) { fGpu = gpu; } argument 24 explicit GrSemaphore(const GrGpu* gpu) : fGpu(gpu) {} argument
|
H A D | GrOpFlushState.cpp | 13 GrOpFlushState::GrOpFlushState(GrGpu* gpu, GrResourceProvider* resourceProvider) argument 14 : fGpu(gpu) 17 , fVertexPool(gpu) 18 , fIndexPool(gpu)
|
/external/skia/src/gpu/vk/ |
H A D | GrVkSemaphore.cpp | 18 sk_sp<GrVkSemaphore> GrVkSemaphore::Make(const GrVkGpu* gpu) { argument 25 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), 26 CreateSemaphore(gpu->device(), &createInfo, nullptr, &semaphore)); 28 return sk_sp<GrVkSemaphore>(new GrVkSemaphore(gpu, semaphore)); 31 GrVkSemaphore::GrVkSemaphore(const GrVkGpu* gpu, VkSemaphore semaphore) : INHERITED(gpu) { argument 43 void GrVkSemaphore::Resource::freeGPUData(const GrVkGpu* gpu) const { 44 GR_VK_CALL(gpu->vkInterface(), 45 DestroySemaphore(gpu->device(), fSemaphore, nullptr));
|
H A D | GrVkUniformBuffer.h | 18 static GrVkUniformBuffer* Create(GrVkGpu* gpu, size_t size); 19 static const GrVkResource* CreateResource(GrVkGpu* gpu, size_t size); 22 void* map(GrVkGpu* gpu) { argument 23 return this->vkMap(gpu); 25 void unmap(GrVkGpu* gpu) { argument 26 this->vkUnmap(gpu); 30 bool updateData(GrVkGpu* gpu, const void* src, size_t srcSizeInBytes, argument 32 return this->vkUpdateData(gpu, src, srcSizeInBytes, createdNewBuffer); 34 void release(const GrVkGpu* gpu) { this->vkRelease(gpu); } argument 51 GrVkUniformBuffer(GrVkGpu* gpu, const GrVkBuffer::Desc& desc, const GrVkUniformBuffer::Resource* resource) argument [all...] |
H A D | GrVkDescriptorSet.cpp | 23 void GrVkDescriptorSet::freeGPUData(const GrVkGpu* gpu) const { 24 fPool->unref(gpu); 27 void GrVkDescriptorSet::onRecycle(GrVkGpu* gpu) const { 28 gpu->resourceProvider().recycleDescriptorSet(this, fHandle);
|
H A D | GrVkUniformBuffer.cpp | 13 GrVkUniformBuffer* GrVkUniformBuffer::Create(GrVkGpu* gpu, size_t size) { argument 19 resource = gpu->resourceProvider().findOrCreateStandardUniformBufferResource(); 21 resource = CreateResource(gpu, size); 31 GrVkUniformBuffer* buffer = new GrVkUniformBuffer(gpu, desc, 36 resource->unref(gpu); 42 const GrVkResource* GrVkUniformBuffer::CreateResource(GrVkGpu* gpu, size_t size) { argument 62 err = VK_CALL(gpu, CreateBuffer(gpu->device(), &bufInfo, nullptr, &buffer)); 67 if (!GrVkMemory::AllocAndBindBufferMemory(gpu, 77 VK_CALL(gpu, DestroyBuffe 85 createResource(GrVkGpu* gpu, const GrVkBuffer::Desc& descriptor) argument [all...] |
H A D | GrVkDescriptorSetManager.h | 29 GrVkDescriptorSetManager(GrVkGpu* gpu, 33 GrVkDescriptorSetManager(GrVkGpu* gpu, 40 void release(const GrVkGpu* gpu); 44 const GrVkDescriptorSet* getDescriptorSet(GrVkGpu* gpu, const Handle& handle); 54 DescriptorPoolManager(VkDescriptorType type, GrVkGpu* gpu, 56 DescriptorPoolManager(VkDescriptorType type, GrVkGpu* gpu, 65 void getNewDescriptorSet(GrVkGpu* gpu, VkDescriptorSet* ds); 67 void freeGPUResources(const GrVkGpu* gpu); 84 void init(GrVkGpu* gpu, VkDescriptorType type, const GrVkUniformHandler* uniformHandler, 87 void getNewPool(GrVkGpu* gpu); [all...] |
H A D | GrVkTextureRenderTarget.h | 38 bool updateForMipmap(GrVkGpu* gpu, const GrVkImageInfo& newInfo); 52 GrVkTextureRenderTarget(GrVkGpu* gpu, argument 60 : GrSurface(gpu, desc) 62 , GrVkTexture(gpu, desc, info, texView, GrVkImage::kNot_Wrapped) 63 , GrVkRenderTarget(gpu, desc, info, msaaInfo, colorAttachmentView, 68 GrVkTextureRenderTarget(GrVkGpu* gpu, argument 74 : GrSurface(gpu, desc) 76 , GrVkTexture(gpu, desc, info, texView, GrVkImage::kNot_Wrapped) 77 , GrVkRenderTarget(gpu, desc, info, colorAttachmentView, GrVkImage::kNot_Wrapped) { 80 GrVkTextureRenderTarget(GrVkGpu* gpu, argument 96 GrVkTextureRenderTarget(GrVkGpu* gpu, const GrSurfaceDesc& desc, const GrVkImageInfo& info, const GrVkImageView* texView, const GrVkImageView* colorAttachmentView, GrVkImage::Wrapped wrapped) argument [all...] |
H A D | GrVkImageView.cpp | 12 const GrVkImageView* GrVkImageView::Create(const GrVkGpu* gpu, VkImage image, VkFormat format, argument 34 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreateImageView(gpu->device(), &viewInfo, 43 void GrVkImageView::freeGPUData(const GrVkGpu* gpu) const { 44 GR_VK_CALL(gpu->vkInterface(), DestroyImageView(gpu->device(), fImageView, nullptr));
|
H A D | GrVkDescriptorPool.cpp | 14 GrVkDescriptorPool::GrVkDescriptorPool(const GrVkGpu* gpu, VkDescriptorType type, uint32_t count) argument 33 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), CreateDescriptorPool(gpu->device(), 43 void GrVkDescriptorPool::reset(const GrVkGpu* gpu) { argument 44 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), ResetDescriptorPool(gpu->device(), fDescPool, 0)); 47 void GrVkDescriptorPool::freeGPUData(const GrVkGpu* gpu) const { 50 GR_VK_CALL(gpu->vkInterface(), DestroyDescriptorPool(gpu->device(), fDescPool, nullptr));
|
H A D | GrVkCommandBuffer.cpp | 37 void GrVkCommandBuffer::freeGPUData(const GrVkGpu* gpu) const { 40 fTrackedResources[i]->unref(gpu); 44 fTrackedRecycledResources[i]->recycle(const_cast<GrVkGpu*>(gpu)); 47 GR_VK_CALL(gpu->vkInterface(), FreeCommandBuffers(gpu->device(), gpu->cmdPool(), 50 this->onFreeGPUData(gpu); 64 void GrVkCommandBuffer::reset(GrVkGpu* gpu) { argument 67 fTrackedResources[i]->unref(gpu); 70 fTrackedRecycledResources[i]->recycle(const_cast<GrVkGpu*>(gpu)); 98 pipelineBarrier(const GrVkGpu* gpu, VkPipelineStageFlags srcStageMask, VkPipelineStageFlags dstStageMask, bool byRegion, BarrierType barrierType, void* barrier) const argument 148 clearAttachments(const GrVkGpu* gpu, int numAttachments, const VkClearAttachment* attachments, int numRects, const VkClearRect* clearRects) const argument 173 bindDescriptorSets(const GrVkGpu* gpu, GrVkPipelineState* pipelineState, VkPipelineLayout layout, uint32_t firstSet, uint32_t setCount, const VkDescriptorSet* descriptorSets, uint32_t dynamicOffsetCount, const uint32_t* dynamicOffsets) argument 193 bindDescriptorSets(const GrVkGpu* gpu, const SkTArray<const GrVkRecycledResource*>& recycled, const SkTArray<const GrVkResource*>& resources, VkPipelineLayout layout, uint32_t firstSet, uint32_t setCount, const VkDescriptorSet* descriptorSets, uint32_t dynamicOffsetCount, const uint32_t* dynamicOffsets) argument 219 bindPipeline(const GrVkGpu* gpu, const GrVkPipeline* pipeline) argument 227 drawIndexed(const GrVkGpu* gpu, uint32_t indexCount, uint32_t instanceCount, uint32_t firstIndex, int32_t vertexOffset, uint32_t firstInstance) const argument 243 draw(const GrVkGpu* gpu, uint32_t vertexCount, uint32_t instanceCount, uint32_t firstVertex, uint32_t firstInstance) const argument 257 setViewport(const GrVkGpu* gpu, uint32_t firstViewport, uint32_t viewportCount, const VkViewport* viewports) argument 272 setScissor(const GrVkGpu* gpu, uint32_t firstScissor, uint32_t scissorCount, const VkRect2D* scissors) argument 287 setBlendConstants(const GrVkGpu* gpu, const float blendConstants[4]) argument 304 Create(const GrVkGpu* gpu, VkCommandPool cmdPool) argument 324 begin(const GrVkGpu* gpu) argument 338 end(const GrVkGpu* gpu) argument 346 beginRenderPass(const GrVkGpu* gpu, const GrVkRenderPass* renderPass, const VkClearValue* clearValues, const GrVkRenderTarget& target, const SkIRect& bounds, bool forSecondaryCB) argument 379 endRenderPass(const GrVkGpu* gpu) argument 386 executeCommands(const GrVkGpu* gpu, GrVkSecondaryCommandBuffer* buffer) argument 401 submitToQueue( const GrVkGpu* gpu, VkQueue queue, GrVkGpu::SyncQueue sync, const GrVkSemaphore::Resource* signalSemaphore, SkTArray<const GrVkSemaphore::Resource*>& waitSemaphores) argument 490 onReset(GrVkGpu* gpu) argument 497 copyImage(const GrVkGpu* gpu, GrVkImage* srcImage, VkImageLayout srcLayout, GrVkImage* dstImage, VkImageLayout dstLayout, uint32_t copyRegionCount, const VkImageCopy* copyRegions) argument 517 blitImage(const GrVkGpu* gpu, const GrVkResource* srcResource, VkImage srcImage, VkImageLayout srcLayout, const GrVkResource* dstResource, VkImage dstImage, VkImageLayout dstLayout, uint32_t blitRegionCount, const VkImageBlit* blitRegions, VkFilter filter) argument 541 copyImageToBuffer(const GrVkGpu* gpu, GrVkImage* srcImage, VkImageLayout srcLayout, GrVkTransferBuffer* dstBuffer, uint32_t copyRegionCount, const VkBufferImageCopy* copyRegions) argument 559 copyBufferToImage(const GrVkGpu* gpu, GrVkTransferBuffer* srcBuffer, GrVkImage* dstImage, VkImageLayout dstLayout, uint32_t copyRegionCount, const VkBufferImageCopy* copyRegions) argument 577 updateBuffer(GrVkGpu* gpu, GrVkBuffer* dstBuffer, VkDeviceSize dstOffset, VkDeviceSize dataSize, const void* data) argument 596 clearColorImage(const GrVkGpu* gpu, GrVkImage* image, const VkClearColorValue* color, uint32_t subRangeCount, const VkImageSubresourceRange* subRanges) argument 612 clearDepthStencilImage(const GrVkGpu* gpu, GrVkImage* image, const VkClearDepthStencilValue* color, uint32_t subRangeCount, const VkImageSubresourceRange* subRanges) argument 628 resolveImage(GrVkGpu* gpu, const GrVkImage& srcImage, const GrVkImage& dstImage, uint32_t regionCount, const VkImageResolve* regions) argument 660 Create(const GrVkGpu* gpu, VkCommandPool cmdPool) argument 681 begin(const GrVkGpu* gpu, const GrVkFramebuffer* framebuffer, const GrVkRenderPass* compatibleRenderPass) argument 711 end(const GrVkGpu* gpu) argument [all...] |
H A D | GrVkBuffer.h | 35 void addMemoryBarrier(const GrVkGpu* gpu, 72 void freeGPUData(const GrVkGpu* gpu) const override; 74 void onRecycle(GrVkGpu* gpu) const override { this->unref(gpu); } 80 static const Resource* Create(const GrVkGpu* gpu, 87 void* vkMap(GrVkGpu* gpu) { argument 88 this->internalMap(gpu, fDesc.fSizeInBytes); 91 void vkUnmap(GrVkGpu* gpu) { this->internalUnmap(gpu, this->size()); } argument 95 bool vkUpdateData(GrVkGpu* gpu, cons 102 createResource(GrVkGpu* gpu, const Desc& descriptor) argument [all...] |
H A D | GrVkCommandBuffer.h | 36 void pipelineBarrier(const GrVkGpu* gpu, 43 void bindVertexBuffer(GrVkGpu* gpu, GrVkVertexBuffer* vbuffer) { argument 49 GR_VK_CALL(gpu->vkInterface(), CmdBindVertexBuffers(fCmdBuffer, 60 void bindIndexBuffer(GrVkGpu* gpu, GrVkIndexBuffer* ibuffer) { argument 65 GR_VK_CALL(gpu->vkInterface(), CmdBindIndexBuffer(fCmdBuffer, 75 void bindPipeline(const GrVkGpu* gpu, const GrVkPipeline* pipeline); 77 void bindDescriptorSets(const GrVkGpu* gpu, 86 void bindDescriptorSets(const GrVkGpu* gpu, 96 void setViewport(const GrVkGpu* gpu, 101 void setScissor(const GrVkGpu* gpu, 178 onReset(GrVkGpu* gpu) argument 258 blitImage(const GrVkGpu* gpu, const GrVkImage& srcImage, const GrVkImage& dstImage, uint32_t blitRegionCount, const VkImageBlit* blitRegions, VkFilter filter) argument [all...] |
H A D | GrVkRenderTarget.cpp | 24 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, argument 32 : GrSurface(gpu, desc) 35 , GrRenderTarget(gpu, desc) 42 this->createFramebuffer(gpu); 48 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, argument 55 : GrSurface(gpu, desc) 58 , GrRenderTarget(gpu, desc) 65 this->createFramebuffer(gpu); 70 GrVkRenderTarget::GrVkRenderTarget(GrVkGpu* gpu, argument 76 : GrSurface(gpu, des 91 GrVkRenderTarget(GrVkGpu* gpu, const GrSurfaceDesc& desc, const GrVkImageInfo& info, const GrVkImageView* colorAttachmentView, GrVkImage::Wrapped wrapped) argument 109 Create(GrVkGpu* gpu, SkBudgeted budgeted, const GrSurfaceDesc& desc, const GrVkImageInfo& info, GrVkImage::Wrapped wrapped) argument 179 CreateNewRenderTarget(GrVkGpu* gpu, SkBudgeted budgeted, const GrSurfaceDesc& desc, const GrVkImage::ImageDesc& imageDesc) argument 199 MakeWrappedRenderTarget(GrVkGpu* gpu, const GrSurfaceDesc& desc, const GrVkImageInfo* info) argument 214 createFramebuffer(GrVkGpu* gpu) argument 280 GrVkGpu* gpu = this->getVkGpu(); local [all...] |
H A D | GrVkBuffer.cpp | 21 const GrVkBuffer::Resource* GrVkBuffer::Create(const GrVkGpu* gpu, const Desc& desc) { argument 57 err = VK_CALL(gpu, CreateBuffer(gpu->device(), &bufInfo, nullptr, &buffer)); 62 if (!GrVkMemory::AllocAndBindBufferMemory(gpu, 72 VK_CALL(gpu, DestroyBuffer(gpu->device(), buffer, nullptr)); 73 GrVkMemory::FreeBufferMemory(gpu, desc.fType, alloc); 80 void GrVkBuffer::addMemoryBarrier(const GrVkGpu* gpu, argument 99 gpu->addBufferMemoryBarrier(srcStageMask, dstStageMask, byRegion, &bufferMemoryBarrier); 102 void GrVkBuffer::Resource::freeGPUData(const GrVkGpu* gpu) cons 109 vkRelease(const GrVkGpu* gpu) argument 144 internalMap(GrVkGpu* gpu, size_t size, bool* createdNewBuffer) argument 184 internalUnmap(GrVkGpu* gpu, size_t size) argument 208 vkUpdateData(GrVkGpu* gpu, const void* src, size_t srcSizeInBytes, bool* createdNewBuffer) argument [all...] |
H A D | GrVkImage.cpp | 28 void GrVkImage::setImageLayout(const GrVkGpu* gpu, VkImageLayout newLayout, argument 62 gpu->addImageMemoryBarrier(srcStageMask, dstStageMask, byRegion, &imageMemoryBarrier); 67 bool GrVkImage::InitImageInfo(const GrVkGpu* gpu, const ImageDesc& imageDesc, GrVkImageInfo* info) { argument 109 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), CreateImage(gpu->device(), &imageCreateInfo, nullptr, 112 if (!GrVkMemory::AllocAndBindImageMemory(gpu, image, isLinear, &alloc)) { 113 VK_CALL(gpu, DestroyImage(gpu->device(), image, nullptr)); 126 void GrVkImage::DestroyImageInfo(const GrVkGpu* gpu, GrVkImageInfo* info) { argument 127 VK_CALL(gpu, DestroyImag 141 releaseImage(const GrVkGpu* gpu) argument [all...] |
H A D | GrVkCopyManager.cpp | 35 bool GrVkCopyManager::createCopyProgram(GrVkGpu* gpu) { argument 36 const GrShaderCaps* shaderCaps = gpu->caps()->shaderCaps(); 77 if (!GrCompileVkShaderModule(gpu, vertShaderText.c_str(), VK_SHADER_STAGE_VERTEX_BIT, 79 this->destroyResources(gpu); 84 if (!GrCompileVkShaderModule(gpu, fragShaderText.c_str(), VK_SHADER_STAGE_FRAGMENT_BIT, 86 this->destroyResources(gpu); 93 GrVkResourceProvider& resourceProvider = gpu->resourceProvider(); 115 VkResult err = GR_VK_CALL(gpu->vkInterface(), CreatePipelineLayout(gpu->device(), 120 this->destroyResources(gpu); 141 copySurfaceAsDraw(GrVkGpu* gpu, GrSurface* dst, GrSurface* src, const SkIRect& srcRect, const SkIPoint& dstPoint) argument 381 destroyResources(GrVkGpu* gpu) argument [all...] |
H A D | GrVkCopyManager.h | 29 bool copySurfaceAsDraw(GrVkGpu* gpu, 35 void destroyResources(GrVkGpu* gpu); 39 bool createCopyProgram(GrVkGpu* gpu);
|
H A D | GrVkDescriptorPool.h | 24 GrVkDescriptorPool(const GrVkGpu* gpu, VkDescriptorType type, uint32_t count); 28 void reset(const GrVkGpu* gpu); 42 void freeGPUData(const GrVkGpu* gpu) const override;
|
/external/autotest/client/site_tests/graphics_dEQP/scripts/ |
H A D | failure_matrix.py | 62 def get_problem_count(dict, gpu): 63 if gpu in dict: 64 if not dict[gpu]: 67 for status in dict[gpu]: 69 count = count + len((dict[gpu])[status]) 76 if status in dict[gpu]: 77 count = count + len((dict[gpu])[status]) 79 print 'Warning: %s not found in dict!' % gpu 85 for gpu in dict: 87 if status in dict[gpu] [all...] |
/external/skia/src/gpu/gl/ |
H A D | GrGLSemaphore.h | 17 static sk_sp<GrGLSemaphore> Make(const GrGLGpu* gpu) { argument 18 return sk_sp<GrGLSemaphore>(new GrGLSemaphore(gpu)); 31 GrGLSemaphore(const GrGLGpu* gpu) : INHERITED(gpu), fSync(0) {} argument
|
H A D | GrGLTextureRenderTarget.h | 29 GrGLTextureRenderTarget(GrGLGpu* gpu, argument 35 : GrSurface(gpu, desc) 36 , GrGLTexture(gpu, desc, texIDDesc, wasMipMapDataProvided) 37 , GrGLRenderTarget(gpu, desc, rtIDDesc) { 45 static sk_sp<GrGLTextureRenderTarget> MakeWrapped(GrGLGpu* gpu, const GrSurfaceDesc& desc, 61 GrGLTextureRenderTarget(GrGLGpu* gpu, argument 66 : GrSurface(gpu, desc) 67 , GrGLTexture(gpu, desc, texIDDesc, wasMipMapDataProvided) 68 , GrGLRenderTarget(gpu, desc, rtIDDesc) {
|
H A D | GrGLTexture.cpp | 21 const GrGLGpu* gpu) { 23 SkASSERT(gpu->caps()->shaderCaps()->externalTextureSupport()); 27 SkASSERT(gpu->glCaps().rectangleTextureSupport()); 53 GrGLTexture::GrGLTexture(GrGLGpu* gpu, SkBudgeted budgeted, const GrSurfaceDesc& desc, argument 55 : GrSurface(gpu, desc) 56 , INHERITED(gpu, desc, sampler_type(idDesc, desc.fConfig, gpu), 62 GrGLTexture::GrGLTexture(GrGLGpu* gpu, SkBudgeted budgeted, const GrSurfaceDesc& desc, argument 65 : GrSurface(gpu, desc) 66 , INHERITED(gpu, des 20 sampler_type(const GrGLTexture::IDDesc& idDesc, GrPixelConfig config, const GrGLGpu* gpu) argument 73 GrGLTexture(GrGLGpu* gpu, Wrapped, const GrSurfaceDesc& desc, const IDDesc& idDesc) argument 81 GrGLTexture(GrGLGpu* gpu, const GrSurfaceDesc& desc, const IDDesc& idDesc, bool wasMipMapDataProvided) argument 149 MakeWrapped(GrGLGpu* gpu, const GrSurfaceDesc& desc, const IDDesc& idDesc) argument [all...] |
H A D | GrGLPathRange.cpp | 13 GrGLPathRange::GrGLPathRange(GrGLGpu* gpu, PathGenerator* pathGenerator, const GrStyle& style) argument 14 : INHERITED(gpu, pathGenerator), 16 fBasePathID(gpu->glPathRendering()->genPaths(this->getNumPaths())), 22 GrGLPathRange::GrGLPathRange(GrGLGpu* gpu, argument 27 : INHERITED(gpu, numPaths), 54 GrGLGpu* gpu = static_cast<GrGLGpu*>(this->getGpu()); local 55 if (nullptr == gpu) { 61 GR_GL_CALL_RET(gpu->glInterface(), isPath, IsPath(fBasePathID + index))); 65 GrGLPath::InitPathObjectEmptyPath(gpu, fBasePathID + index); 67 GrGLPath::InitPathObjectPathData(gpu, fBasePathI [all...] |
/external/chromium-trace/catapult/telemetry/telemetry/internal/platform/ |
H A D | system_info_unittest.py | 16 'gpu': { 25 self.assertTrue(isinstance(info.gpu, gpu_info.GPUInfo)) 27 self.assertTrue(len(info.gpu.devices) == 1) 28 self.assertTrue(isinstance(info.gpu.devices[0], gpu_device.GPUDevice)) 29 self.assertEquals(info.gpu.devices[0].vendor_id, 1000) 30 self.assertEquals(info.gpu.devices[0].device_id, 2000) 31 self.assertEquals(info.gpu.devices[0].vendor_string, 'a') 32 self.assertEquals(info.gpu.devices[0].device_string, 'b') 37 'gpu': {
|