/drivers/staging/lustre/lustre/lclient/ |
H A D | lcommon_misc.c | 127 int cl_get_grouplock(struct cl_object *obj, unsigned long gid, int nonblock, argument 143 io->ci_obj = obj; 156 descr->cld_obj = obj;
|
/drivers/staging/lustre/lustre/lov/ |
H A D | lov_page.c | 157 int lov_page_init_raid0(const struct lu_env *env, struct cl_object *obj, argument 160 struct lov_object *loo = cl2lov(obj); 166 struct lov_page *lpg = cl_object_page_slice(obj, page); 172 offset = cl_offset(obj, page->cp_index); 180 cl_page_slice_add(page, &lpg->lps_cl, obj, &lov_page_ops); 217 int lov_page_init_empty(const struct lu_env *env, struct cl_object *obj, argument 220 struct lov_page *lpg = cl_object_page_slice(obj, page); 223 cl_page_slice_add(page, &lpg->lps_cl, obj, &lov_empty_page_ops); 225 memset(addr, 0, cl_page_size(obj));
|
H A D | lovsub_dev.c | 69 const struct cl_object *obj, 74 subobj = cl2lovsub(obj); 67 lovsub_req_attr_set(const struct lu_env *env, const struct cl_req_slice *slice, const struct cl_object *obj, struct cl_req_attr *attr, u64 flags) argument
|
/drivers/staging/lustre/lustre/osc/ |
H A D | osc_object.c | 60 static struct osc_object *lu2osc(const struct lu_object *obj) argument 62 LINVRNT(osc_is_object(obj)); 63 return container_of0(obj, struct osc_object, oo_cl.co_lu); 72 static int osc_object_init(const struct lu_env *env, struct lu_object *obj, argument 75 struct osc_object *osc = lu2osc(obj); 98 cl_object_page_init(lu2cl(obj), sizeof(struct osc_page)); 103 static void osc_object_free(const struct lu_env *env, struct lu_object *obj) argument 105 struct osc_object *osc = lu2osc(obj); 124 lu_object_fini(obj); 137 lu_printer_t p, const struct lu_object *obj) 136 osc_object_print(const struct lu_env *env, void *cookie, lu_printer_t p, const struct lu_object *obj) argument 153 osc_attr_get(const struct lu_env *env, struct cl_object *obj, struct cl_attr *attr) argument 163 osc_attr_set(const struct lu_env *env, struct cl_object *obj, const struct cl_attr *attr, unsigned valid) argument 187 osc_object_glimpse(const struct lu_env *env, const struct cl_object *obj, struct ost_lvb *lvb) argument 198 osc_object_set_contended(struct osc_object *obj) argument 205 osc_object_clear_contended(struct osc_object *obj) argument 210 osc_object_is_contended(struct osc_object *obj) argument 259 struct lu_object *obj; local [all...] |
/drivers/staging/unisys/visorutil/ |
H A D | procobjecttree.c | 226 MYPROCOBJECT *obj = NULL, *rc = NULL; local 233 obj = kzalloc(sizeof(MYPROCOBJECT), GFP_KERNEL | __GFP_NORETRY); 234 if (obj == NULL) { 238 obj->type = type; 239 obj->context = context; 241 obj->name = NULL; 242 obj->procDir = type->procDir; 244 obj->namesize = strlen(name)+1; 245 obj->name = kmalloc(obj 301 visor_proc_DestroyObject(MYPROCOBJECT *obj) argument [all...] |
/drivers/acpi/ |
H A D | fan.c | 115 union acpi_object *obj; local 125 obj = buffer.pointer; 126 if (!obj || obj->type != ACPI_TYPE_PACKAGE || 127 obj->package.count != 3 || 128 obj->package.elements[1].type != ACPI_TYPE_INTEGER) { 134 control = obj->package.elements[1].integer.value; 148 kfree(obj); 242 union acpi_object *obj; local 249 obj 278 union acpi_object *obj; local [all...] |
/drivers/char/tpm/ |
H A D | tpm_ppi.c | 31 union acpi_object *obj; local 38 obj = acpi_evaluate_dsm_typed(handle, tpm_ppi_uuid, 41 if (obj) { 42 strlcpy(tpm_ppi_version, obj->string.pointer, 44 ACPI_FREE(obj); 70 union acpi_object *obj; local 72 obj = tpm_eval_dsm(TPM_PPI_FN_GETREQ, ACPI_TYPE_PACKAGE, NULL); 73 if (!obj) 82 if (obj->package.count == 2 && 83 obj 104 union acpi_object *obj, tmp; local 154 union acpi_object *obj = NULL; local 197 union acpi_object *obj, *ret_obj; local 256 union acpi_object *obj, tmp; local [all...] |
/drivers/gpu/drm/armada/ |
H A D | armada_gem.c | 19 struct armada_gem_object *obj = drm_to_armada_gem(vma->vm_private_data); local 21 unsigned long pfn = obj->phys_addr >> PAGE_SHIFT; 50 void armada_gem_free_object(struct drm_gem_object *obj) argument 52 struct armada_gem_object *dobj = drm_to_armada_gem(obj); 54 DRM_DEBUG_DRIVER("release obj %p\n", dobj); 56 drm_gem_free_mmap_offset(&dobj->obj); 60 unsigned int order = get_order(dobj->obj.size); 70 if (dobj->obj.import_attach) { 72 dma_buf_unmap_attachment(dobj->obj.import_attach, dobj->sgt, 74 drm_prime_gem_destroy(&dobj->obj, NUL 83 armada_gem_linear_back(struct drm_device *dev, struct armada_gem_object *obj) argument 194 struct armada_gem_object *obj; local 213 struct armada_gem_object *obj; local 273 struct armada_gem_object *obj; local 424 struct drm_gem_object *obj = attach->dmabuf->priv; local 493 struct drm_gem_object *obj = attach->dmabuf->priv; local 538 armada_gem_prime_export(struct drm_device *dev, struct drm_gem_object *obj, int flags) argument 552 struct drm_gem_object *obj = buf->priv; local [all...] |
/drivers/gpu/drm/ast/ |
H A D | ast_fb.c | 50 struct drm_gem_object *obj; local 60 obj = afbdev->afb.obj; 61 bo = gem_to_ast_bo(obj); 310 if (afb->obj) { 311 drm_gem_object_unreference_unlocked(afb->obj); 312 afb->obj = NULL;
|
H A D | ast_main.c | 298 if (ast_fb->obj) 299 drm_gem_object_unreference_unlocked(ast_fb->obj); 313 struct drm_gem_object *obj) 318 ast_fb->obj = obj; 332 struct drm_gem_object *obj; local 336 obj = drm_gem_object_lookup(dev, filp, mode_cmd->handles[0]); 337 if (obj == NULL) 342 drm_gem_object_unreference_unlocked(obj); 346 ret = ast_framebuffer_init(dev, ast_fb, mode_cmd, obj); 310 ast_framebuffer_init(struct drm_device *dev, struct ast_framebuffer *ast_fb, struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_object *obj) argument 495 ast_gem_create(struct drm_device *dev, u32 size, bool iskernel, struct drm_gem_object **obj) argument 555 ast_gem_free_object(struct drm_gem_object *obj) argument 573 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/cirrus/ |
H A D | cirrus_fbdev.c | 24 struct drm_gem_object *obj; local 34 obj = afbdev->gfb.obj; 35 bo = gem_to_cirrus_bo(obj); 279 if (gfb->obj) { 280 drm_gem_object_unreference_unlocked(gfb->obj); 281 gfb->obj = NULL;
|
H A D | cirrus_main.c | 20 if (cirrus_fb->obj) 21 drm_gem_object_unreference_unlocked(cirrus_fb->obj); 33 struct drm_gem_object *obj) 38 gfb->obj = obj; 52 struct drm_gem_object *obj; local 62 obj = drm_gem_object_lookup(dev, filp, mode_cmd->handles[0]); 63 if (obj == NULL) 68 drm_gem_object_unreference_unlocked(obj); 72 ret = cirrus_framebuffer_init(dev, cirrus_fb, mode_cmd, obj); 30 cirrus_framebuffer_init(struct drm_device *dev, struct cirrus_framebuffer *gfb, struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_object *obj) argument 210 cirrus_gem_create(struct drm_device *dev, u32 size, bool iskernel, struct drm_gem_object **obj) argument 270 cirrus_gem_free_object(struct drm_gem_object *obj) argument 289 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/ |
H A D | drm_fb_cma_helper.c | 30 struct drm_gem_cma_object *obj[4]; member in struct:drm_fb_cma 54 if (fb_cma->obj[i]) 55 drm_gem_object_unreference_unlocked(&fb_cma->obj[i]->base); 68 &fb_cma->obj[0]->base, handle); 77 struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_cma_object **obj, 91 fb_cma->obj[i] = obj[i]; 114 struct drm_gem_object *obj; local 128 obj = drm_gem_object_lookup(dev, file_priv, mode_cmd->handles[i]); 129 if (!obj) { 76 drm_fb_cma_alloc(struct drm_device *dev, struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_cma_object **obj, unsigned int num_planes) argument 253 struct drm_gem_cma_object *obj; local [all...] |
H A D | drm_gem.c | 127 * @obj: drm_gem_object to initialize 134 struct drm_gem_object *obj, size_t size) 138 drm_gem_private_object_init(dev, obj, size); 144 obj->filp = filp; 153 * @obj: drm_gem_object to initialize 161 struct drm_gem_object *obj, size_t size) 165 obj->dev = dev; 166 obj->filp = NULL; 168 kref_init(&obj->refcount); 169 obj 133 drm_gem_object_init(struct drm_device *dev, struct drm_gem_object *obj, size_t size) argument 160 drm_gem_private_object_init(struct drm_device *dev, struct drm_gem_object *obj, size_t size) argument 176 drm_gem_remove_prime_handles(struct drm_gem_object *obj, struct drm_file *filp) argument 200 drm_gem_object_handle_free(struct drm_gem_object *obj) argument 211 drm_gem_object_exported_dma_buf_free(struct drm_gem_object *obj) argument 221 drm_gem_object_handle_unreference_unlocked(struct drm_gem_object *obj) argument 254 struct drm_gem_object *obj; local 319 drm_gem_handle_create_tail(struct drm_file *file_priv, struct drm_gem_object *obj, u32 *handlep) argument 375 drm_gem_handle_create(struct drm_file *file_priv, struct drm_gem_object *obj, u32 *handlep) argument 393 drm_gem_free_mmap_offset(struct drm_gem_object *obj) argument 416 drm_gem_create_mmap_offset_size(struct drm_gem_object *obj, size_t size) argument 436 drm_gem_create_mmap_offset(struct drm_gem_object *obj) argument 463 drm_gem_get_pages(struct drm_gem_object *obj) argument 517 drm_gem_put_pages(struct drm_gem_object *obj, struct page **pages, bool dirty, bool accessed) argument 550 struct drm_gem_object *obj; local 608 struct drm_gem_object *obj; local 660 struct drm_gem_object *obj; local 711 struct drm_gem_object *obj = ptr; local 744 drm_gem_object_release(struct drm_gem_object *obj) argument 767 struct drm_gem_object *obj = (struct drm_gem_object *) kref; local 779 struct drm_gem_object *obj = vma->vm_private_data; local 791 struct drm_gem_object *obj = vma->vm_private_data; local 827 drm_gem_mmap_obj(struct drm_gem_object *obj, unsigned long obj_size, struct vm_area_struct *vma) argument 878 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_gem.c | 56 static void update_vm_cache_attr(struct exynos_drm_gem_obj *obj, argument 59 DRM_DEBUG_KMS("flags = 0x%x\n", obj->flags); 62 if (obj->flags & EXYNOS_BO_CACHABLE) 64 else if (obj->flags & EXYNOS_BO_WC) 79 static int exynos_drm_gem_map_buf(struct drm_gem_object *obj, argument 84 struct exynos_drm_gem_obj *exynos_gem_obj = to_exynos_gem_obj(obj); 110 static int exynos_drm_gem_handle_create(struct drm_gem_object *obj, argument 117 * allocate a id of idr table where the obj is registered 120 ret = drm_gem_handle_create(file_priv, obj, handle); 127 drm_gem_object_unreference_unlocked(obj); 134 struct drm_gem_object *obj; local 171 struct drm_gem_object *obj; local 191 struct drm_gem_object *obj; local 287 struct drm_gem_object *obj; local 304 struct drm_gem_object *obj; local 359 struct drm_gem_object *obj; local 504 exynos_drm_gem_free_object(struct drm_gem_object *obj) argument 563 struct drm_gem_object *obj; local 597 struct drm_gem_object *obj = vma->vm_private_data; local 621 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/i915/ |
H A D | i915_gem_context.c | 154 struct drm_i915_gem_object *obj; local 157 obj = i915_gem_alloc_object(dev, size); 158 if (obj == NULL) 170 ret = i915_gem_object_set_cache_level(obj, I915_CACHE_L3_LLC); 173 drm_gem_object_unreference(&obj->base); 178 return obj; 197 struct drm_i915_gem_object *obj = local 199 if (IS_ERR(obj)) { 200 ret = PTR_ERR(obj); 203 ctx->legacy_hw_ctx.rcs_state = obj; [all...] |
H A D | i915_gem_dmabuf.c | 38 struct drm_i915_gem_object *obj = dma_buf_to_obj(attachment->dmabuf); local 43 ret = i915_mutex_lock_interruptible(obj->base.dev); 47 ret = i915_gem_object_get_pages(obj); 51 i915_gem_object_pin_pages(obj); 60 ret = sg_alloc_table(st, obj->pages->nents, GFP_KERNEL); 64 src = obj->pages->sgl; 66 for (i = 0; i < obj->pages->nents; i++) { 77 mutex_unlock(&obj->base.dev->struct_mutex); 85 i915_gem_object_unpin_pages(obj); 87 mutex_unlock(&obj 96 struct drm_i915_gem_object *obj = dma_buf_to_obj(attachment->dmabuf); local 111 struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf); local 162 struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf); local 201 struct drm_i915_gem_object *obj = dma_buf_to_obj(dma_buf); local 232 struct drm_i915_gem_object *obj = to_intel_bo(gem_obj); local 244 i915_gem_object_get_pages_dmabuf(struct drm_i915_gem_object *obj) argument 257 i915_gem_object_put_pages_dmabuf(struct drm_i915_gem_object *obj) argument 273 struct drm_i915_gem_object *obj; local [all...] |
H A D | intel_fbdev.c | 66 ret = i915_gem_object_set_to_gtt_domain(ifbdev->fb->obj, 96 struct drm_i915_gem_object *obj; local 113 obj = i915_gem_object_create_stolen(dev, size); 114 if (obj == NULL) 115 obj = i915_gem_alloc_object(dev, size); 116 if (!obj) { 123 ret = intel_pin_and_fence_fb_obj(dev, obj, NULL); 125 DRM_ERROR("failed to pin obj: %d\n", ret); 129 fb = __intel_framebuffer_create(dev, &mode_cmd, obj); 140 i915_gem_object_ggtt_unpin(obj); 157 struct drm_i915_gem_object *obj; local [all...] |
/drivers/gpu/drm/mgag200/ |
H A D | mgag200_fb.c | 26 struct drm_gem_object *obj; local 36 obj = mfbdev->mfb.obj; 37 bo = gem_to_mga_bo(obj); 264 if (mfb->obj) { 265 drm_gem_object_unreference_unlocked(mfb->obj); 266 mfb->obj = NULL;
|
H A D | mgag200_main.c | 20 if (mga_fb->obj) 21 drm_gem_object_unreference_unlocked(mga_fb->obj); 33 struct drm_gem_object *obj) 38 gfb->obj = obj; 52 struct drm_gem_object *obj; local 56 obj = drm_gem_object_lookup(dev, filp, mode_cmd->handles[0]); 57 if (obj == NULL) 62 drm_gem_object_unreference_unlocked(obj); 66 ret = mgag200_framebuffer_init(dev, mga_fb, mode_cmd, obj); 30 mgag200_framebuffer_init(struct drm_device *dev, struct mga_framebuffer *gfb, struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_object *obj) argument 268 mgag200_gem_create(struct drm_device *dev, u32 size, bool iskernel, struct drm_gem_object **obj) argument 328 mgag200_gem_free_object(struct drm_gem_object *obj) argument 347 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/msm/mdp/mdp4/ |
H A D | mdp4_plane.c | 82 struct drm_mode_object *obj) 81 mdp4_plane_install_properties(struct drm_plane *plane, struct drm_mode_object *obj) argument
|
/drivers/gpu/drm/msm/mdp/mdp5/ |
H A D | mdp5_plane.c | 100 struct drm_mode_object *obj) 99 mdp5_plane_install_properties(struct drm_plane *plane, struct drm_mode_object *obj) argument
|
/drivers/gpu/drm/msm/ |
H A D | msm_gem_submit.c | 68 struct drm_gem_object *obj; local 92 obj = idr_find(&file->object_idr, submit_bo.handle); 93 if (!obj) { 99 msm_obj = to_msm_bo(obj); 108 drm_gem_object_reference(obj); 110 submit->bos[i].obj = msm_obj; 124 struct msm_gem_object *msm_obj = submit->bos[i].obj; 147 struct msm_gem_object *msm_obj = submit->bos[i].obj; 200 struct msm_gem_object *msm_obj = submit->bos[contended].obj; 215 struct msm_gem_object **obj, uint32_ 214 submit_bo(struct msm_gem_submit *submit, uint32_t idx, struct msm_gem_object **obj, uint32_t *iova, bool *valid) argument 234 submit_reloc(struct msm_gem_submit *submit, struct msm_gem_object *obj, uint32_t offset, uint32_t nr_relocs, uint64_t relocs) argument [all...] |
H A D | msm_gpu.c | 432 struct msm_gem_object *obj; local 434 obj = list_first_entry(&gpu->active_list, 437 if ((obj->read_fence <= fence) && 438 (obj->write_fence <= fence)) { 440 msm_gem_move_to_inactive(&obj->base); 441 msm_gem_put_iova(&obj->base, gpu->id); 442 drm_gem_object_unreference(&obj->base); 486 struct msm_gem_object *msm_obj = submit->bos[i].obj;
|
H A D | msm_rd.c | 312 struct msm_gem_object *obj = submit->bos[idx].obj; local 313 const char *buf = msm_gem_vaddr_locked(&obj->base);
|