/external/chromium_org/mojo/gles2/ |
H A D | gles2_context.h | 10 #include "gpu/command_buffer/client/gles2_implementation.h" 16 namespace gpu { namespace 37 gpu::gles2::GLES2Interface* interface() const { 40 gpu::ContextSupport* context_support() const { return implementation_.get(); } 46 scoped_ptr<gpu::gles2::GLES2CmdHelper> gles2_helper_; 47 scoped_ptr<gpu::TransferBuffer> transfer_buffer_; 48 scoped_ptr<gpu::gles2::GLES2Implementation> implementation_;
|
/external/chromium_org/cc/output/ |
H A D | geometry_binding.h | 14 namespace gpu { namespace 24 GeometryBinding(gpu::gles2::GLES2Interface* gl, 38 gpu::gles2::GLES2Interface* gl_;
|
H A D | context_provider.h | 11 #include "gpu/command_buffer/common/capabilities.h" 15 namespace gpu { namespace 31 virtual gpu::gles2::GLES2Interface* ContextGL() = 0; 32 virtual gpu::ContextSupport* ContextSupport() = 0; 36 gpu::Capabilities gpu; member in struct:cc::ContextProvider::Capabilities 53 // Delete all cached gpu resources.
|
/external/chromium_org/third_party/skia/src/gpu/ |
H A D | GrAARectRenderer.h | 42 void fillAARect(GrGpu* gpu, argument 49 this->shaderFillAlignedAARect(gpu, target, 52 this->shaderFillAARect(gpu, target, 56 this->geometryFillAARect(gpu, target, rect, combinedMatrix, devRect); 60 void strokeAARect(GrGpu* gpu, 68 void fillAANestedRects(GrGpu* gpu, 78 GrIndexBuffer* aaFillRectIndexBuffer(GrGpu* gpu); 81 GrIndexBuffer* aaStrokeRectIndexBuffer(GrGpu* gpu, bool miterStroke); 83 void geometryFillAARect(GrGpu* gpu, 89 void shaderFillAARect(GrGpu* gpu, [all...] |
H A D | GrIndexBuffer.h | 27 GrIndexBuffer(GrGpu* gpu, bool isWrapped, size_t gpuMemorySize, bool dynamic, bool cpuBacked) argument 28 : INHERITED(gpu, isWrapped, gpuMemorySize, dynamic, cpuBacked) {}
|
H A D | GrVertexBuffer.h | 18 GrVertexBuffer(GrGpu* gpu, bool isWrapped, size_t gpuMemorySize, bool dynamic, bool cpuBacked) argument 19 : INHERITED(gpu, isWrapped, gpuMemorySize, dynamic, cpuBacked) {}
|
H A D | GrGpuResource.cpp | 24 static inline GrResourceCache2* get_resource_cache2(GrGpu* gpu) { argument 25 SkASSERT(gpu); 26 SkASSERT(gpu->getContext()); 27 SkASSERT(gpu->getContext()->getResourceCache2()); 28 return gpu->getContext()->getResourceCache2(); 31 GrGpuResource::GrGpuResource(GrGpu* gpu, bool isWrapped) argument 32 : fGpu(gpu)
|
/external/skia/src/gpu/ |
H A D | GrAARectRenderer.h | 42 void fillAARect(GrGpu* gpu, argument 50 this->shaderFillAlignedAARect(gpu, target, 53 this->shaderFillAARect(gpu, target, 57 this->geometryFillAARect(gpu, target, 63 void strokeAARect(GrGpu* gpu, 72 void fillAANestedRects(GrGpu* gpu, 83 GrIndexBuffer* aaFillRectIndexBuffer(GrGpu* gpu); 86 GrIndexBuffer* aaStrokeRectIndexBuffer(GrGpu* gpu, bool miterStroke); 90 void geometryFillAARect(GrGpu* gpu, 97 void shaderFillAARect(GrGpu* gpu, [all...] |
H A D | GrIndexBuffer.h | 27 GrIndexBuffer(GrGpu* gpu, bool isWrapped, size_t gpuMemorySize, bool dynamic, bool cpuBacked) argument 28 : INHERITED(gpu, isWrapped, gpuMemorySize, dynamic, cpuBacked) {}
|
H A D | GrVertexBuffer.h | 18 GrVertexBuffer(GrGpu* gpu, bool isWrapped, size_t gpuMemorySize, bool dynamic, bool cpuBacked) argument 19 : INHERITED(gpu, isWrapped, gpuMemorySize, dynamic, cpuBacked) {}
|
/external/chromium_org/gpu/command_buffer/client/ |
H A D | gles2_interface_stub.h | 8 #include "gpu/command_buffer/client/gles2_interface.h" 10 namespace gpu { namespace 22 #include "gpu/command_buffer/client/gles2_interface_stub_autogen.h" 26 } // namespace gpu
|
H A D | program_info_manager_unittest.cc | 7 #include "gpu/command_buffer/client/program_info_manager.h" 12 namespace gpu { namespace 30 } // namespace gpu
|
H A D | gles2_lib.h | 10 #include "gpu/command_buffer/client/gles2_c_lib_export.h" 11 #include "gpu/command_buffer/client/gles2_interface.h" 29 GLES2_C_LIB_EXPORT gpu::gles2::GLES2Interface* GetGLContext(); 32 GLES2_C_LIB_EXPORT void SetGLContext(gpu::gles2::GLES2Interface* impl);
|
/external/chromium_org/third_party/skia/src/gpu/gl/ |
H A D | GrGLPathRange.cpp | 14 GrGLPathRange::GrGLPathRange(GrGpuGL* gpu, PathGenerator* pathGenerator, const SkStrokeRec& stroke) argument 15 : INHERITED(gpu, pathGenerator, stroke), 16 fBasePathID(gpu->glPathRendering()->genPaths(this->getNumPaths())), 21 GrGLPathRange::GrGLPathRange(GrGpuGL* gpu, argument 26 : INHERITED(gpu, numPaths, stroke), 37 GrGpuGL* gpu = static_cast<GrGpuGL*>(this->getGpu()); local 38 if (NULL == gpu) { 45 GR_GL_CALL_RET(gpu->glInterface(), isPath, IsPath(fBasePathID + index))); 48 GrGLPath::InitPathObject(gpu, fBasePathID + index, skPath, this->getStroke());
|
H A D | GrGLBufferImpl.cpp | 23 GrGLBufferImpl::GrGLBufferImpl(GrGpuGL* gpu, const Desc& desc, GrGLenum bufferType) argument 38 void GrGLBufferImpl::release(GrGpuGL* gpu) { argument 45 GL_CALL(gpu, DeleteBuffers(1, &fDesc.fID)); 47 gpu->notifyVertexBufferDelete(fDesc.fID); 50 gpu->notifyIndexBufferDelete(fDesc.fID); 68 void GrGLBufferImpl::bind(GrGpuGL* gpu) const { 71 gpu->bindVertexBuffer(fDesc.fID); 74 gpu->bindIndexBufferAndDefaultVertexArray(fDesc.fID); 79 void* GrGLBufferImpl::map(GrGpuGL* gpu) { argument 85 switch (gpu 136 unmap(GrGpuGL* gpu) argument 163 updateData(GrGpuGL* gpu, const void* src, size_t srcSizeInBytes) argument [all...] |
H A D | GrGLTexture.cpp | 15 void GrGLTexture::init(GrGpuGL* gpu, argument 34 fRenderTarget.reset(SkNEW_ARGS(GrGLRenderTarget, (gpu, *rtDesc, vp, fTexIDObj, this))); 39 GrGLTexture::GrGLTexture(GrGpuGL* gpu, argument 41 : INHERITED(gpu, textureDesc.fIsWrapped, textureDesc) { 42 this->init(gpu, textureDesc, NULL); 45 GrGLTexture::GrGLTexture(GrGpuGL* gpu, argument 48 : INHERITED(gpu, textureDesc.fIsWrapped, textureDesc) { 49 this->init(gpu, textureDesc, &rtDesc);
|
H A D | GrGLBufferImpl.h | 36 void release(GrGpuGL* gpu); 41 void bind(GrGpuGL* gpu) const; 43 void* map(GrGpuGL* gpu); 44 void unmap(GrGpuGL* gpu); 46 bool updateData(GrGpuGL* gpu, const void* src, size_t srcSizeInBytes);
|
/external/skia/src/gpu/gl/ |
H A D | GrGLBufferImpl.cpp | 23 GrGLBufferImpl::GrGLBufferImpl(GrGpuGL* gpu, const Desc& desc, GrGLenum bufferType) argument 38 void GrGLBufferImpl::release(GrGpuGL* gpu) { argument 45 GL_CALL(gpu, DeleteBuffers(1, &fDesc.fID)); 47 gpu->notifyVertexBufferDelete(fDesc.fID); 50 gpu->notifyIndexBufferDelete(fDesc.fID); 68 void GrGLBufferImpl::bind(GrGpuGL* gpu) const { 71 gpu->bindVertexBuffer(fDesc.fID); 74 gpu->bindIndexBufferAndDefaultVertexArray(fDesc.fID); 79 void* GrGLBufferImpl::map(GrGpuGL* gpu) { argument 85 switch (gpu 136 unmap(GrGpuGL* gpu) argument 163 updateData(GrGpuGL* gpu, const void* src, size_t srcSizeInBytes) argument [all...] |
H A D | GrGLTexture.cpp | 15 void GrGLTexture::init(GrGpuGL* gpu, argument 34 fRenderTarget.reset(SkNEW_ARGS(GrGLRenderTarget, (gpu, *rtDesc, vp, fTexIDObj, this))); 38 GrGLTexture::GrGLTexture(GrGpuGL* gpu, argument 40 : INHERITED(gpu, textureDesc.fIsWrapped, textureDesc) { 41 this->init(gpu, textureDesc, NULL); 44 GrGLTexture::GrGLTexture(GrGpuGL* gpu, argument 47 : INHERITED(gpu, textureDesc.fIsWrapped, textureDesc) { 48 this->init(gpu, textureDesc, &rtDesc);
|
H A D | GrGLBufferImpl.h | 36 void release(GrGpuGL* gpu); 41 void bind(GrGpuGL* gpu) const; 43 void* map(GrGpuGL* gpu); 44 void unmap(GrGpuGL* gpu); 46 bool updateData(GrGpuGL* gpu, const void* src, size_t srcSizeInBytes);
|
/external/chromium_org/content/browser/android/in_process/ |
H A D | synchronous_compositor_factory_impl.cc | 10 #include "content/renderer/gpu/frame_swap_message_queue.h" 11 #include "gpu/command_buffer/client/gl_in_process_context.h" 12 #include "gpu/command_buffer/common/gles2_cmd_utils.h" 16 #include "webkit/common/gpu/context_provider_in_process.h" 17 #include "webkit/common/gpu/webgraphicscontext3d_in_process_command_buffer_impl.h" 19 using webkit::gpu::ContextProviderWebContext; 36 using webkit::gpu::WebGraphicsContext3DInProcessCommandBufferImpl; 37 using webkit::gpu::WebGraphicsContext3DImpl; 39 scoped_ptr<gpu::GLInProcessContext> CreateOffscreenContext( 43 gpu [all...] |
/external/chromium_org/gpu/ipc/ |
H A D | gpu_command_buffer_traits.h | 9 #include "gpu/command_buffer/common/command_buffer.h" 10 #include "gpu/gpu_export.h" 12 namespace gpu { namespace 20 struct GPU_EXPORT ParamTraits<gpu::CommandBuffer::State> { 21 typedef gpu::CommandBuffer::State param_type; 28 struct GPU_EXPORT ParamTraits<gpu::Mailbox> { 29 typedef gpu::Mailbox param_type; 36 struct GPU_EXPORT ParamTraits<gpu::MailboxHolder> { 37 typedef gpu::MailboxHolder param_type;
|
/external/chromium_org/mojo/services/gles2/ |
H A D | command_buffer_type_conversions.cc | 12 TypeConverter<CommandBufferStatePtr, gpu::CommandBuffer::State>::Convert( 13 const gpu::CommandBuffer::State& input) { 25 gpu::CommandBuffer::State 26 TypeConverter<gpu::CommandBuffer::State, CommandBufferStatePtr>::Convert( 28 gpu::CommandBuffer::State state; 33 state.error = static_cast<gpu::error::Error>(input->error); 35 static_cast<gpu::error::ContextLostReason>(input->context_lost_reason);
|
/external/chromium_org/tools/telemetry/telemetry/core/ |
H A D | system_info_unittest.py | 15 'gpu': { 24 self.assertTrue(isinstance(info.gpu, gpu_info.GPUInfo)) 26 self.assertTrue(len(info.gpu.devices) == 1) 27 self.assertTrue(isinstance(info.gpu.devices[0], gpu_device.GPUDevice)) 28 self.assertEquals(info.gpu.devices[0].vendor_id, 1000) 29 self.assertEquals(info.gpu.devices[0].device_id, 2000) 30 self.assertEquals(info.gpu.devices[0].vendor_string, 'a') 31 self.assertEquals(info.gpu.devices[0].device_string, 'b') 36 'gpu': {
|
/external/chromium_org/content/renderer/ |
H A D | webgraphicscontext3d_provider_impl.h | 14 namespace gpu { namespace in namespace:webkit 17 } // namespace gpu 25 scoped_refptr<webkit::gpu::ContextProviderWebContext> provider); 33 scoped_refptr<webkit::gpu::ContextProviderWebContext> provider_;
|