/drivers/acpi/ |
H A D | processor_perflib.c | 233 union acpi_object obj = { 0 }; local 254 obj = pct->package.elements[0]; 256 if ((obj.type != ACPI_TYPE_BUFFER) 257 || (obj.buffer.length < sizeof(struct acpi_pct_register)) 258 || (obj.buffer.pointer == NULL)) { 263 memcpy(&pr->performance->control_register, obj.buffer.pointer, 270 obj = pct->package.elements[1]; 272 if ((obj.type != ACPI_TYPE_BUFFER) 273 || (obj.buffer.length < sizeof(struct acpi_pct_register)) 274 || (obj [all...] |
H A D | utils.c | 579 union acpi_object obj = { .type = ACPI_TYPE_INTEGER }; local 580 struct acpi_object_list arg_list = { .count = 1, .pointer = &obj, }; 582 obj.integer.value = arg; 699 union acpi_object *obj; local 704 obj = acpi_evaluate_dsm(handle, uuid, rev, 0, NULL); 705 if (!obj) 709 if (obj->type == ACPI_TYPE_INTEGER) 710 mask = obj->integer.value; 711 else if (obj->type == ACPI_TYPE_BUFFER) 712 for (i = 0; i < obj [all...] |
/drivers/base/ |
H A D | node.c | 328 struct device *obj; local 333 obj = get_cpu_device(cpu); 334 if (!obj) 338 &obj->kobj, 339 kobject_name(&obj->kobj)); 343 return sysfs_create_link(&obj->kobj, 350 struct device *obj; local 355 obj = get_cpu_device(cpu); 356 if (!obj) 360 kobject_name(&obj [all...] |
/drivers/block/ |
H A D | osdblk.c | 115 struct osd_obj_id obj; /* OSD partition, obj id */ member in struct:osdblk_device 137 const struct osd_obj_id *obj) 139 osd_sec_init_nosec_doall_caps(cred_a, obj, false, true); 211 osd_req_get_attributes(or, &osdev->obj); 350 osd_req_flush_object(or, &osdev->obj, 353 osd_req_write(or, &osdev->obj, blk_rq_pos(rq) * 512ULL, 356 osd_req_read(or, &osdev->obj, blk_rq_pos(rq) * 512ULL, 488 osdev->obj.partition, 489 osdev->obj 136 osdblk_make_credential(u8 cred_a[OSD_CAP_LEN], const struct osd_obj_id *obj) argument [all...] |
/drivers/extcon/ |
H A D | extcon-class.c | 400 struct extcon_specific_cable_nb *obj = container_of(nb, local 404 if ((val & (1 << obj->cable_index)) != 405 (edev->state & (1 << obj->cable_index))) { 408 obj->previous_value = val; 410 if (val & (1 << obj->cable_index)) 413 return obj->user_nb->notifier_call(obj->user_nb, 424 * @obj: an empty extcon_specific_cable_nb object to be returned. 443 int extcon_register_interest(struct extcon_specific_cable_nb *obj, argument 447 if (!obj || !cable_nam 496 extcon_unregister_interest(struct extcon_specific_cable_nb *obj) argument [all...] |
/drivers/gpu/drm/armada/ |
H A D | armada_overlay.c | 160 struct armada_gem_object *obj = drm_fb_obj(fb); local 183 sy = obj->dev_addr + fb->offsets[0] + src_y * fb->pitches[0] + 185 su = obj->dev_addr + fb->offsets[1] + src_y * fb->pitches[1] + 187 sv = obj->dev_addr + fb->offsets[2] + src_y * fb->pitches[2] +
|
/drivers/gpu/drm/bochs/ |
H A D | bochs.h | 55 struct drm_gem_object *obj; member in struct:bochs_framebuffer 142 struct drm_gem_object **obj); 143 int bochs_gem_init_object(struct drm_gem_object *obj); 144 void bochs_gem_free_object(struct drm_gem_object *obj); 153 struct drm_gem_object *obj);
|
H A D | bochs_mm.c | 389 struct drm_gem_object **obj) 394 *obj = NULL; 406 *obj = &bochsbo->gem; 446 void bochs_gem_free_object(struct drm_gem_object *obj) argument 448 struct bochs_bo *bochs_bo = gem_to_bochs_bo(obj); 456 struct drm_gem_object *obj; local 461 obj = drm_gem_object_lookup(dev, file, handle); 462 if (obj == NULL) { 467 bo = gem_to_bochs_bo(obj); 470 drm_gem_object_unreference(obj); 388 bochs_gem_create(struct drm_device *dev, u32 size, bool iskernel, struct drm_gem_object **obj) argument 493 bochs_framebuffer_init(struct drm_device *dev, struct bochs_framebuffer *gfb, struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_object *obj) argument 515 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/cirrus/ |
H A D | cirrus_mode.c | 134 struct drm_gem_object *obj; local 143 obj = cirrus_fb->obj; 144 bo = gem_to_cirrus_bo(obj); 153 obj = cirrus_fb->obj; 154 bo = gem_to_cirrus_bo(obj);
|
/drivers/gpu/drm/ |
H A D | drm_prime.c | 123 struct drm_gem_object *obj = dma_buf->priv; local 124 struct drm_device *dev = obj->dev; 136 return dev->driver->gem_prime_pin(obj); 143 struct drm_gem_object *obj = dma_buf->priv; local 144 struct drm_device *dev = obj->dev; 148 dev->driver->gem_prime_unpin(obj); 184 struct drm_gem_object *obj = attach->dmabuf->priv; local 201 sgt = obj->dev->driver->gem_prime_get_sg_table(obj); 233 struct drm_gem_object *obj local 242 struct drm_gem_object *obj = dma_buf->priv; local 250 struct drm_gem_object *obj = dma_buf->priv; local 282 struct drm_gem_object *obj = dma_buf->priv; local 339 drm_gem_prime_export(struct drm_device *dev, struct drm_gem_object *obj, int flags) argument 352 export_and_register_object(struct drm_device *dev, struct drm_gem_object *obj, uint32_t flags) argument 403 struct drm_gem_object *obj; local 498 struct drm_gem_object *obj; local 565 struct drm_gem_object *obj; local 746 drm_prime_gem_destroy(struct drm_gem_object *obj, struct sg_table *sg) argument [all...] |
/drivers/gpu/drm/gma500/ |
H A D | gma_display.c | 345 struct drm_gem_object *obj; local 380 obj = drm_gem_object_lookup(dev, file_priv, handle); 381 if (!obj) { 386 if (obj->size < width * height * 4) { 392 gt = container_of(obj, struct gtt_range, gem); 447 gma_crtc->cursor_obj = obj; 453 drm_gem_object_unreference(obj);
|
/drivers/gpu/drm/i915/ |
H A D | i915_gem_userptr.c | 61 struct drm_i915_gem_object *obj; member in struct:i915_mmu_object 65 static unsigned long cancel_userptr(struct drm_i915_gem_object *obj) argument 67 struct drm_device *dev = obj->base.dev; 72 obj->userptr.work = NULL; 74 if (obj->pages != NULL) { 82 list_for_each_entry_safe(vma, tmp, &obj->vma_list, vma_link) { 86 WARN_ON(i915_gem_object_put_pages(obj)); 91 end = obj->userptr.ptr + obj->base.size; 93 drm_gem_object_unreference(&obj 110 struct drm_i915_gem_object *obj; local 141 struct drm_i915_gem_object *obj = NULL; local 222 struct drm_i915_gem_object *obj; local 279 i915_gem_userptr_release__mmu_notifier(struct drm_i915_gem_object *obj) argument 316 i915_gem_userptr_init__mmu_notifier(struct drm_i915_gem_object *obj, unsigned flags) argument 366 i915_gem_userptr_release__mmu_notifier(struct drm_i915_gem_object *obj) argument 371 i915_gem_userptr_init__mmu_notifier(struct drm_i915_gem_object *obj, unsigned flags) argument 405 i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj) argument 473 i915_gem_userptr_release__mm_struct(struct drm_i915_gem_object *obj) argument 486 struct drm_i915_gem_object *obj; member in struct:get_pages_work 533 struct drm_i915_gem_object *obj = work->obj; local 588 i915_gem_userptr_get_pages(struct drm_i915_gem_object *obj) argument 690 i915_gem_userptr_put_pages(struct drm_i915_gem_object *obj) argument 715 i915_gem_userptr_release(struct drm_i915_gem_object *obj) argument 722 i915_gem_userptr_dmabuf_export(struct drm_i915_gem_object *obj) argument 774 struct drm_i915_gem_object *obj; local [all...] |
/drivers/gpu/drm/msm/mdp/mdp5/ |
H A D | mdp5_crtc.c | 367 struct drm_gem_object *obj; local 375 obj = msm_framebuffer_bo(new_fb, 0); 383 return msm_gem_queue_inactive_cb(obj, &mdp5_crtc->pageflip_cb);
|
/drivers/gpu/drm/msm/ |
H A D | msm_gem.c | 27 static dma_addr_t physaddr(struct drm_gem_object *obj) argument 29 struct msm_gem_object *msm_obj = to_msm_bo(obj); 30 struct msm_drm_private *priv = obj->dev->dev_private; 36 static struct page **get_pages_vram(struct drm_gem_object *obj, argument 39 struct msm_gem_object *msm_obj = to_msm_bo(obj); 40 struct msm_drm_private *priv = obj->dev->dev_private; 56 paddr = physaddr(obj); 66 static struct page **get_pages(struct drm_gem_object *obj) argument 68 struct msm_gem_object *msm_obj = to_msm_bo(obj); 71 struct drm_device *dev = obj 105 put_pages(struct drm_gem_object *obj) argument 130 msm_gem_get_pages(struct drm_gem_object *obj) argument 140 msm_gem_put_pages(struct drm_gem_object *obj) argument 145 msm_gem_mmap_obj(struct drm_gem_object *obj, struct vm_area_struct *vma) argument 189 struct drm_gem_object *obj = vma->vm_private_data; local 243 mmap_offset(struct drm_gem_object *obj) argument 261 msm_gem_mmap_offset(struct drm_gem_object *obj) argument 277 msm_gem_get_iova_locked(struct drm_gem_object *obj, int id, uint32_t *iova) argument 312 msm_gem_get_iova(struct drm_gem_object *obj, int id, uint32_t *iova) argument 331 msm_gem_put_iova(struct drm_gem_object *obj, int id) argument 353 struct drm_gem_object *obj; local 371 msm_gem_vaddr_locked(struct drm_gem_object *obj) argument 385 msm_gem_vaddr(struct drm_gem_object *obj) argument 397 msm_gem_queue_inactive_cb(struct drm_gem_object *obj, struct msm_fence_cb *cb) argument 419 msm_gem_move_to_active(struct drm_gem_object *obj, struct msm_gpu *gpu, bool write, uint32_t fence) argument 432 msm_gem_move_to_inactive(struct drm_gem_object *obj) argument 447 msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, struct timespec *timeout) argument 472 msm_gem_cpu_fini(struct drm_gem_object *obj) argument 479 msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m) argument 500 struct drm_gem_object *obj = &msm_obj->base; local 511 msm_gem_free_object(struct drm_gem_object *obj) argument 563 struct drm_gem_object *obj; local 585 msm_gem_new_impl(struct drm_device *dev, uint32_t size, uint32_t flags, struct drm_gem_object **obj) argument 631 struct drm_gem_object *obj = NULL; local 663 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/nouveau/core/include/core/ |
H A D | object.h | 31 nv_object(void *obj) argument 34 if (likely(obj)) { 35 struct nouveau_object *object = obj; 40 return obj; 105 nv_ofuncs(void *obj) argument 107 return nv_oclass(obj)->ofuncs; 120 nv_exec(void *obj, u32 mthd, void *data, u32 size) argument 122 struct nouveau_omthds *method = nv_oclass(obj)->omthds; 126 return method->call(obj, mthd, data, size); 134 nv_call(void *obj, u3 argument 140 nv_ro08(void *obj, u64 addr) argument 148 nv_ro16(void *obj, u64 addr) argument 156 nv_ro32(void *obj, u64 addr) argument 164 nv_wo08(void *obj, u64 addr, u8 data) argument 171 nv_wo16(void *obj, u64 addr, u16 data) argument 178 nv_wo32(void *obj, u64 addr, u32 data) argument 185 nv_mo32(void *obj, u64 addr, u32 mask, u32 data) argument 193 nv_memcmp(void *obj, u32 addr, const char *str, u32 len) argument [all...] |
/drivers/gpu/drm/nouveau/core/include/subdev/ |
H A D | clock.h | 105 nouveau_clock(void *obj) argument 107 return (void *)nv_device(obj)->subdev[NVDEV_SUBDEV_CLOCK];
|
H A D | fb.h | 76 nouveau_fb(void *obj) argument 79 if (nv_iclass(obj, NV_SUBDEV_CLASS) && 80 nv_subidx(obj) == NVDEV_SUBDEV_FB) 81 return obj; 83 return (void *)nv_device(obj)->subdev[NVDEV_SUBDEV_FB];
|
/drivers/gpu/drm/omapdrm/ |
H A D | omap_drv.c | 378 struct drm_gem_object *obj; local 383 obj = drm_gem_object_lookup(dev, file_priv, args->handle); 384 if (!obj) 387 ret = omap_gem_op_sync(obj, args->op); 390 ret = omap_gem_op_start(obj, args->op); 392 drm_gem_object_unreference_unlocked(obj); 401 struct drm_gem_object *obj; local 406 obj = drm_gem_object_lookup(dev, file_priv, args->handle); 407 if (!obj) 414 ret = omap_gem_op_finish(obj, arg 425 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/ttm/ |
H A D | ttm_object.c | 108 * @obj: Base object this ref object is referencing. 126 struct ttm_base_object *obj; member in struct:ttm_ref_object 241 base = drm_hash_entry(hash, struct ttm_ref_object, hash)->obj; 300 if (unlikely(base != ref->obj)) 304 * Verify that the ref->obj pointer was actually valid! 359 ref->obj = base; 391 struct ttm_base_object *base = ref->obj; 404 ttm_base_object_unref(&ref->obj);
|
/drivers/i2c/busses/ |
H A D | i2c-scmi.c | 91 union acpi_object *obj; local 187 obj = pkg->package.elements; 193 if (obj == NULL || obj->type != ACPI_TYPE_INTEGER) { 199 result = obj->integer.value; 224 obj = pkg->package.elements + 1; 225 if (obj->type != ACPI_TYPE_INTEGER) { 231 len = obj->integer.value; 232 obj = pkg->package.elements + 2; 237 if (obj 299 union acpi_object *obj; local [all...] |
/drivers/infiniband/hw/mthca/ |
H A D | mthca_memfree.c | 222 int mthca_table_get(struct mthca_dev *dev, struct mthca_icm_table *table, int obj) argument 224 int i = (obj & (table->num_obj - 1)) * table->obj_size / MTHCA_TABLE_CHUNK_SIZE; 257 void mthca_table_put(struct mthca_dev *dev, struct mthca_icm_table *table, int obj) argument 264 i = (obj & (table->num_obj - 1)) * table->obj_size / MTHCA_TABLE_CHUNK_SIZE; 278 void *mthca_table_find(struct mthca_icm_table *table, int obj, dma_addr_t *dma_handle) argument 290 idx = (obj & (table->num_obj - 1)) * table->obj_size;
|
/drivers/infiniband/hw/usnic/ |
H A D | usnic_ib_qp_grp.c | 62 int usnic_ib_qp_grp_dump_rows(void *obj, char *buf, int buf_sz) argument 64 struct usnic_ib_qp_grp *qp_grp = obj; 66 if (obj) {
|
/drivers/leds/ |
H A D | dell-led.c | 81 union acpi_object *obj; local 94 obj = output.pointer; 95 if (!obj) 98 if (obj->type != ACPI_TYPE_BUFFER) 101 bios_return = (struct app_wmi_args *)obj->buffer.pointer; 110 kfree(obj); 206 union acpi_object *obj; local 231 obj = output.pointer; 233 if (!obj) 235 else if (obj [all...] |
/drivers/staging/android/ |
H A D | sync.c | 40 struct sync_timeline *obj; local 45 obj = kzalloc(size, GFP_KERNEL); 46 if (obj == NULL) 49 kref_init(&obj->kref); 50 obj->ops = ops; 51 obj->context = fence_context_alloc(1); 52 strlcpy(obj->name, name, sizeof(obj->name)); 54 INIT_LIST_HEAD(&obj->child_list_head); 55 INIT_LIST_HEAD(&obj 66 struct sync_timeline *obj = local 77 sync_timeline_get(struct sync_timeline *obj) argument 82 sync_timeline_put(struct sync_timeline *obj) argument 87 sync_timeline_destroy(struct sync_timeline *obj) argument 104 sync_timeline_signal(struct sync_timeline *obj) argument 124 sync_pt_create(struct sync_timeline *obj, int size) argument [all...] |
/drivers/staging/lustre/lustre/llite/ |
H A D | rw26.c | 78 struct cl_object *obj; local 94 obj = ll_i2info(inode)->lli_clob; 95 if (obj != NULL) { 96 page = cl_vmpage_page(vmpage, obj); 121 struct cl_object *obj; local 134 obj = ll_i2info(mapping->host)->lli_clob; 135 if (obj == NULL) 151 page = cl_vmpage_page(vmpage, obj); 168 struct vvp_object *obj = cl_inode2vvp(vmpage->mapping->host); 174 LASSERT(&obj 234 struct cl_object *obj = io->ci_obj; local 369 struct ccc_object *obj = cl_inode2ccc(inode); local [all...] |