/drivers/gpu/drm/armada/ |
H A D | armada_fbdev.c | 41 struct armada_gem_object *obj; local 54 obj = armada_gem_alloc_private_object(dev, size); 55 if (!obj) { 60 ret = armada_gem_linear_back(dev, obj); 62 drm_gem_object_unreference_unlocked(&obj->obj); 66 ptr = armada_gem_map_object(dev, obj); 68 drm_gem_object_unreference_unlocked(&obj->obj); 72 dfb = armada_framebuffer_create(dev, &mode, obj); [all...] |
/drivers/gpu/drm/nouveau/core/include/core/ |
H A D | client.h | 21 nv_client(void *obj) argument 24 if (unlikely(!nv_iclass(obj, NV_CLIENT_CLASS))) 25 nv_assert("BAD CAST -> NvClient, %08x", nv_hclass(obj)); 27 return obj; 31 nouveau_client(void *obj) argument 33 struct nouveau_object *client = nv_object(obj); 49 const char *nouveau_client_name(void *obj);
|
H A D | gpuobj.h | 28 nv_gpuobj(void *obj) argument 31 if (unlikely(!nv_iclass(obj, NV_GPUOBJ_CLASS))) 32 nv_assert("BAD CAST -> NvGpuObj, %08x", nv_hclass(obj)); 34 return obj; 60 nouveau_gpuobj_ref(struct nouveau_gpuobj *obj, struct nouveau_gpuobj **ref) argument 62 nouveau_object_ref(&obj->base, (struct nouveau_object **)ref);
|
H A D | parent.h | 30 nv_parent(void *obj) argument 33 if (unlikely(!(nv_iclass(obj, NV_PARENT_CLASS)))) 34 nv_assert("BAD CAST -> NvParent, %08x", nv_hclass(obj)); 36 return obj;
|
H A D | engine.h | 23 nv_engine(void *obj) argument 26 if (unlikely(!nv_iclass(obj, NV_ENGINE_CLASS))) 27 nv_assert("BAD CAST -> NvEngine, %08x", nv_hclass(obj)); 29 return obj;
|
/drivers/gpu/drm/nouveau/core/include/subdev/ |
H A D | timer.h | 43 nouveau_timer(void *obj) argument 45 return (void *)nv_device(obj)->subdev[NVDEV_SUBDEV_TIMER];
|
H A D | volt.h | 30 nouveau_volt(void *obj) argument 32 return (void *)nv_device(obj)->subdev[NVDEV_SUBDEV_VOLT];
|
/drivers/gpu/drm/nouveau/ |
H A D | nouveau_prime.c | 31 struct sg_table *nouveau_gem_prime_get_sg_table(struct drm_gem_object *obj) argument 33 struct nouveau_bo *nvbo = nouveau_gem_object(obj); 39 void *nouveau_gem_prime_vmap(struct drm_gem_object *obj) argument 41 struct nouveau_bo *nvbo = nouveau_gem_object(obj); 52 void nouveau_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr) argument 54 struct nouveau_bo *nvbo = nouveau_gem_object(obj); 90 int nouveau_gem_prime_pin(struct drm_gem_object *obj) argument 92 struct nouveau_bo *nvbo = nouveau_gem_object(obj); 103 void nouveau_gem_prime_unpin(struct drm_gem_object *obj) argument 105 struct nouveau_bo *nvbo = nouveau_gem_object(obj); 110 nouveau_gem_prime_res_obj(struct drm_gem_object *obj) argument [all...] |
/drivers/gpu/drm/radeon/ |
H A D | radeon_prime.c | 32 struct sg_table *radeon_gem_prime_get_sg_table(struct drm_gem_object *obj) argument 34 struct radeon_bo *bo = gem_to_radeon_bo(obj); 40 void *radeon_gem_prime_vmap(struct drm_gem_object *obj) argument 42 struct radeon_bo *bo = gem_to_radeon_bo(obj); 53 void radeon_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr) argument 55 struct radeon_bo *bo = gem_to_radeon_bo(obj); 83 int radeon_gem_prime_pin(struct drm_gem_object *obj) argument 85 struct radeon_bo *bo = gem_to_radeon_bo(obj); 98 void radeon_gem_prime_unpin(struct drm_gem_object *obj) argument 100 struct radeon_bo *bo = gem_to_radeon_bo(obj); 112 radeon_gem_prime_res_obj(struct drm_gem_object *obj) argument [all...] |
/drivers/infiniband/hw/ipath/ |
H A D | ipath_mmap.c | 56 vfree(ip->obj); 115 ret = remap_vmalloc_range(vma, ip->obj, 0); 134 void *obj) { 153 ip->obj = obj; 162 u32 size, void *obj) { 173 ip->obj = obj; 131 ipath_create_mmap_info(struct ipath_ibdev *dev, u32 size, struct ib_ucontext *context, void *obj) argument 160 ipath_update_mmap_info(struct ipath_ibdev *dev, struct ipath_mmap_info *ip, u32 size, void *obj) argument
|
/drivers/infiniband/hw/qib/ |
H A D | qib_mmap.c | 56 vfree(ip->obj); 115 ret = remap_vmalloc_range(vma, ip->obj, 0); 134 void *obj) { 153 ip->obj = obj; 161 u32 size, void *obj) 173 ip->obj = obj; 131 qib_create_mmap_info(struct qib_ibdev *dev, u32 size, struct ib_ucontext *context, void *obj) argument 160 qib_update_mmap_info(struct qib_ibdev *dev, struct qib_mmap_info *ip, u32 size, void *obj) argument
|
/drivers/acpi/ |
H A D | processor_core.c | 111 union acpi_object *obj; local 121 obj = buffer.pointer; 122 if (obj->type != ACPI_TYPE_BUFFER || 123 obj->buffer.length < sizeof(struct acpi_subtable_header)) { 127 header = (struct acpi_subtable_header *)obj->buffer.pointer;
|
H A D | processor_pdc.c | 76 union acpi_object *obj; local 84 obj = kmalloc(sizeof(union acpi_object), GFP_KERNEL); 85 if (!obj) { 92 kfree(obj); 99 obj->type = ACPI_TYPE_BUFFER; 100 obj->buffer.length = 12; 101 obj->buffer.pointer = (u8 *) buf; 103 obj_list->pointer = obj; 126 union acpi_object *obj; local 129 obj [all...] |
/drivers/gpu/drm/ |
H A D | drm_info.c | 223 struct drm_gem_object *obj = ptr; local 227 obj->name, obj->size, 228 obj->handle_count, 229 atomic_read(&obj->refcount.refcount));
|
/drivers/gpu/drm/gma500/ |
H A D | gem.c | 32 void psb_gem_free_object(struct drm_gem_object *obj) argument 34 struct gtt_range *gtt = container_of(obj, struct gtt_range, gem); 37 drm_gem_free_mmap_offset(obj); 38 drm_gem_object_release(obj); 41 psb_gtt_free_range(obj->dev, gtt); 63 struct drm_gem_object *obj; local 68 obj = drm_gem_object_lookup(dev, file, handle); 69 if (obj == NULL) { 76 ret = drm_gem_create_mmap_offset(obj); 79 *offset = drm_vma_node_offset_addr(&obj 176 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/i915/ |
H A D | i915_gem_evict.c | 173 drm_gem_object_reference(&vma->obj->base); 181 struct drm_gem_object *obj; local 186 obj = &vma->obj->base; 191 drm_gem_object_unreference(obj);
|
H A D | intel_acpi.c | 96 union acpi_object *obj = &pkg->package.elements[i]; local 97 union acpi_object *connector_id = &obj->package.elements[0]; 98 union acpi_object *info = &obj->package.elements[1];
|
/drivers/gpu/drm/nouveau/core/engine/crypt/ |
H A D | nv84.c | 50 struct nouveau_gpuobj *obj; local 54 16, 16, 0, &obj); 55 *pobject = nv_object(obj); 59 nv_wo32(obj, 0x00, nv_mclass(obj)); 60 nv_wo32(obj, 0x04, 0x00000000); 61 nv_wo32(obj, 0x08, 0x00000000); 62 nv_wo32(obj, 0x0c, 0x00000000);
|
/drivers/gpu/drm/nouveau/core/include/engine/ |
H A D | graph.h | 36 nouveau_graph(void *obj) argument 38 return (void *)nv_device(obj)->subdev[NVDEV_ENGINE_GR]; 79 bool nv04_graph_idle(void *obj);
|
/drivers/gpu/drm/omapdrm/ |
H A D | omap_gem_dmabuf.c | 28 struct drm_gem_object *obj = attachment->dmabuf->priv; local 40 ret = omap_gem_get_paddr(obj, &paddr, true); 49 sg_dma_len(sg->sgl) = obj->size; 50 sg_set_page(sg->sgl, pfn_to_page(PFN_DOWN(paddr)), obj->size, 0); 54 omap_gem_dma_sync(obj, dir); 65 struct drm_gem_object *obj = attachment->dmabuf->priv; local 66 omap_gem_put_paddr(obj); 73 struct drm_gem_object *obj = buffer->priv; local 77 drm_gem_object_unreference_unlocked(obj); 84 struct drm_gem_object *obj local 99 struct drm_gem_object *obj = buffer->priv; local 107 struct drm_gem_object *obj = buffer->priv; local 123 struct drm_gem_object *obj = buffer->priv; local 133 struct drm_gem_object *obj = buffer->priv; local 142 struct drm_gem_object *obj = buffer->priv; local 171 omap_gem_prime_export(struct drm_device *dev, struct drm_gem_object *obj, int flags) argument 180 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/qxl/ |
H A D | qxl_gem.c | 43 struct drm_gem_object **obj) 48 *obj = NULL; 60 *obj = &qbo->gem_base; 98 int qxl_gem_object_open(struct drm_gem_object *obj, struct drm_file *file_priv) argument 103 void qxl_gem_object_close(struct drm_gem_object *obj, argument 39 qxl_gem_object_create(struct qxl_device *qdev, int size, int alignment, int initial_domain, bool discardable, bool kernel, struct qxl_surface *surf, struct drm_gem_object **obj) argument
|
/drivers/gpu/drm/via/ |
H A D | via_mm.c | 184 struct via_memblock *obj; local 187 obj = idr_find(&dev_priv->object_idr, mem->index); 188 if (obj == NULL) { 194 list_del(&obj->owner_list); 195 drm_mm_remove_node(&obj->mm_node); 196 kfree(obj);
|
/drivers/platform/x86/ |
H A D | dell-wmi-aio.c | 89 union acpi_object *obj; local 99 obj = (union acpi_object *)response.pointer; 100 if (obj) { 103 switch (obj->type) { 106 scancode = obj->integer.value; 111 if (dell_wmi_aio_event_check(obj->buffer.pointer, 112 obj->buffer.length)) { 114 obj->buffer.pointer; 119 if (obj->buffer.pointer && 120 obj [all...] |
/drivers/staging/lustre/lustre/llite/ |
H A D | vvp_object.c | 60 struct ccc_object *obj = lu2ccc(o); local 61 struct inode *inode = obj->cob_inode; 65 list_empty(&obj->cob_pending_list) ? "-" : "+", 66 obj->cob_transient_pages, atomic_read(&obj->cob_mmap_cnt), 78 static int vvp_attr_get(const struct lu_env *env, struct cl_object *obj, argument 81 struct inode *inode = ccc_object_inode(obj); 100 static int vvp_attr_set(const struct lu_env *env, struct cl_object *obj, argument 103 struct inode *inode = ccc_object_inode(obj); 123 static int vvp_conf_set(const struct lu_env *env, struct cl_object *obj, argument 187 struct cl_object *obj = lli->lli_clob; local [all...] |
/drivers/staging/lustre/lustre/lov/ |
H A D | lovsub_object.c | 55 int lovsub_object_init(const struct lu_env *env, struct lu_object *obj, argument 58 struct lovsub_device *dev = lu2lovsub_dev(obj->lo_dev); 65 below = under->ld_ops->ldo_object_alloc(env, obj->lo_header, under); 67 lu_object_add(obj, below); 68 cl_object_page_init(lu2cl(obj), sizeof(struct lovsub_page)); 76 static void lovsub_object_free(const struct lu_env *env, struct lu_object *obj) argument 78 struct lovsub_object *los = lu2lovsub(obj); 92 lu_object_fini(obj); 98 lu_printer_t p, const struct lu_object *obj) 100 struct lovsub_object *los = lu2lovsub(obj); 97 lovsub_object_print(const struct lu_env *env, void *cookie, lu_printer_t p, const struct lu_object *obj) argument 105 lovsub_attr_set(const struct lu_env *env, struct cl_object *obj, const struct cl_attr *attr, unsigned valid) argument 114 lovsub_object_glimpse(const struct lu_env *env, const struct cl_object *obj, struct ost_lvb *lvb) argument 146 struct lu_object *obj; local [all...] |