/drivers/staging/lustre/lustre/lov/ |
H A D | lov_lock.c | 953 struct lov_object *obj = cl2lov(slice->cls_obj); local 994 lov->lls_sub[0].sub_stripe, lov->lls_nr, lov_r0(obj)->lo_nr, 1122 int lov_lock_init_raid0(const struct lu_env *env, struct cl_object *obj, argument 1130 cl_lock_slice_add(lock, &lck->lls_cl, obj, &lov_lock_ops); 1157 int lov_lock_init_empty(const struct lu_env *env, struct cl_object *obj, argument 1165 cl_lock_slice_add(lock, &lck->lls_cl, obj, &lov_empty_lock_ops);
|
/drivers/thermal/int340x_thermal/ |
H A D | int3403_thermal.c | 110 struct int3403_sensor *obj = priv->priv; local 112 if (priv->type != INT3403_TYPE_SENSOR || !obj) 115 if (trip == obj->crit_trip_id) 116 *temp = obj->crit_temp; 117 else if (trip == obj->psv_trip_id) 118 *temp = obj->psv_temp; 125 *temp = obj->thresholds[trip]; 135 struct int3403_sensor *obj = priv->priv; local 138 if (trip == obj->crit_trip_id) 151 struct int3403_sensor *obj local 190 struct int3403_sensor *obj; local 243 struct int3403_sensor *obj; local 303 struct int3403_sensor *obj = priv->priv; local 314 struct int3403_cdev *obj = priv->priv; local 358 struct int3403_cdev *obj; local 392 struct int3403_cdev *obj = priv->priv; local [all...] |
/drivers/acpi/ |
H A D | processor_throttling.c | 433 union acpi_object obj = { 0 }; local 456 obj = ptc->package.elements[0]; 458 if ((obj.type != ACPI_TYPE_BUFFER) 459 || (obj.buffer.length < sizeof(struct acpi_ptc_register)) 460 || (obj.buffer.pointer == NULL)) { 466 memcpy(&pr->throttling.control_register, obj.buffer.pointer, 473 obj = ptc->package.elements[1]; 475 if ((obj.type != ACPI_TYPE_BUFFER) 476 || (obj.buffer.length < sizeof(struct acpi_ptc_register)) 477 || (obj [all...] |
/drivers/char/agp/ |
H A D | hp-agp.c | 478 zx1_gart_probe (acpi_handle obj, u32 depth, void *context, void **ret) argument 486 status = hp_acpi_csr_space(obj, &lba_hpa, &length); 491 handle = obj;
|
/drivers/gpu/drm/cirrus/ |
H A D | cirrus_drv.h | 122 struct drm_gem_object *obj; member in struct:cirrus_framebuffer 196 void cirrus_gem_free_object(struct drm_gem_object *obj); 203 struct drm_gem_object **obj); 211 struct drm_gem_object *obj);
|
/drivers/gpu/drm/gma500/ |
H A D | framebuffer.c | 505 struct drm_gem_object *obj; local 511 obj = drm_gem_object_lookup(dev, filp, cmd->handles[0]); 512 if (obj == NULL) 516 r = container_of(obj, struct gtt_range, gem);
|
/drivers/gpu/drm/i915/ |
H A D | i915_cmd_parser.c | 805 static u32 *vmap_batch(struct drm_i915_gem_object *obj) argument 812 pages = drm_malloc_ab(obj->base.size >> PAGE_SHIFT, sizeof(*pages)); 819 for_each_sg_page(obj->pages->sgl, &sg_iter, obj->pages->nents, 0) {
|
H A D | i915_gem_gtt.h | 123 struct drm_i915_gem_object *obj; member in struct:i915_vma 298 int __must_check i915_gem_gtt_prepare_object(struct drm_i915_gem_object *obj); 299 void i915_gem_gtt_finish_object(struct drm_i915_gem_object *obj);
|
H A D | intel_sprite.c | 144 struct drm_i915_gem_object *obj, int crtc_x, int crtc_y, 218 if (obj->tiling_mode != I915_TILING_NONE) 235 obj->tiling_mode, 255 if (obj->tiling_mode != I915_TILING_NONE) 262 I915_WRITE(SPSURF(pipe, plane), i915_gem_obj_ggtt_offset(obj) + 354 struct drm_i915_gem_object *obj, int crtc_x, int crtc_y, 408 if (obj->tiling_mode != I915_TILING_NONE) 436 intel_gen4_compute_page_offset(&x, &y, obj->tiling_mode, 463 else if (obj->tiling_mode != I915_TILING_NONE) 473 i915_gem_obj_ggtt_offset(obj) 142 vlv_update_plane(struct drm_plane *dplane, struct drm_crtc *crtc, struct drm_framebuffer *fb, struct drm_i915_gem_object *obj, int crtc_x, int crtc_y, unsigned int crtc_w, unsigned int crtc_h, uint32_t x, uint32_t y, uint32_t src_w, uint32_t src_h) argument 352 ivb_update_plane(struct drm_plane *plane, struct drm_crtc *crtc, struct drm_framebuffer *fb, struct drm_i915_gem_object *obj, int crtc_x, int crtc_y, unsigned int crtc_w, unsigned int crtc_h, uint32_t x, uint32_t y, uint32_t src_w, uint32_t src_h) argument 572 ilk_update_plane(struct drm_plane *plane, struct drm_crtc *crtc, struct drm_framebuffer *fb, struct drm_i915_gem_object *obj, int crtc_x, int crtc_y, unsigned int crtc_w, unsigned int crtc_h, uint32_t x, uint32_t y, uint32_t src_w, uint32_t src_h) argument 859 struct drm_i915_gem_object *obj = intel_fb->obj; local [all...] |
/drivers/gpu/drm/msm/mdp/mdp4/ |
H A D | mdp4_crtc.c | 456 struct drm_gem_object *obj; local 464 obj = msm_framebuffer_bo(new_fb, 0); 472 return msm_gem_queue_inactive_cb(obj, &mdp4_crtc->pageflip_cb); 504 /* take a obj ref + iova ref when we start scanning out: */ 522 /* and drop the iova ref + obj rev when done scanning out: */
|
/drivers/gpu/drm/msm/ |
H A D | msm_drv.c | 699 struct drm_gem_object *obj; local 707 obj = drm_gem_object_lookup(dev, file, args->handle); 708 if (!obj) 711 ret = msm_gem_cpu_prep(obj, args->op, &TS(args->timeout)); 713 drm_gem_object_unreference_unlocked(obj); 722 struct drm_gem_object *obj; local 725 obj = drm_gem_object_lookup(dev, file, args->handle); 726 if (!obj) 729 ret = msm_gem_cpu_fini(obj); 731 drm_gem_object_unreference_unlocked(obj); 740 struct drm_gem_object *obj; local [all...] |
/drivers/gpu/drm/nouveau/core/engine/graph/ |
H A D | ctxnv40.c | 554 struct nouveau_gpuobj *obj = ctx->data; local 587 nv_wo32(obj, offset * 4, 0x3f800000); 591 nv_wo32(obj, (offset + b0_offset + i) * 4, 0x00000001); 593 nv_wo32(obj, (offset + b1_offset + i) * 4, 0x3f800000);
|
H A D | nv04.c | 961 struct nouveau_gpuobj *obj; local 965 16, 16, 0, &obj); 966 *pobject = nv_object(obj); 970 nv_wo32(obj, 0x00, nv_mclass(obj)); 972 nv_mo32(obj, 0x00, 0x00080000, 0x00080000); 974 nv_wo32(obj, 0x04, 0x00000000); 975 nv_wo32(obj, 0x08, 0x00000000); 976 nv_wo32(obj, 0x0c, 0x00000000); 1199 nv04_graph_idle(void *obj) argument [all...] |
H A D | nv50.c | 68 struct nouveau_gpuobj *obj; local 72 16, 16, 0, &obj); 73 *pobject = nv_object(obj); 77 nv_wo32(obj, 0x00, nv_mclass(obj)); 78 nv_wo32(obj, 0x04, 0x00000000); 79 nv_wo32(obj, 0x08, 0x00000000); 80 nv_wo32(obj, 0x0c, 0x00000000);
|
/drivers/gpu/drm/nouveau/core/include/core/ |
H A D | device.h | 108 nv_device(void *obj) argument 110 struct nouveau_object *object = nv_object(obj); 130 nouveau_subdev(void *obj, int sub) argument 132 if (nv_device(obj)->subdev[sub]) 133 return nv_subdev(nv_device(obj)->subdev[sub]); 138 nouveau_engine(void *obj, int sub) argument 140 struct nouveau_subdev *subdev = nouveau_subdev(obj, sub);
|
/drivers/gpu/drm/nouveau/ |
H A D | nouveau_acpi.c | 77 union acpi_object *obj; local 90 obj = acpi_evaluate_dsm_typed(handle, nouveau_op_dsm_muid, 0x00000100, 92 if (!obj) { 96 if (obj->buffer.length == 4) { 97 *result |= obj->buffer.pointer[0]; 98 *result |= (obj->buffer.pointer[1] << 8); 99 *result |= (obj->buffer.pointer[2] << 16); 100 *result |= (obj->buffer.pointer[3] << 24); 102 ACPI_FREE(obj); 134 union acpi_object *obj; local 351 union acpi_object rom_arg_elements[2], *obj; local [all...] |
/drivers/gpu/drm/radeon/ |
H A D | radeon_legacy_crtc.c | 376 struct drm_gem_object *obj; local 423 obj = radeon_fb->obj; 424 rbo = gem_to_radeon_bo(obj); 451 old_rbo = gem_to_radeon_bo(to_radeon_framebuffer(fb)->obj); 560 rbo = gem_to_radeon_bo(radeon_fb->obj); 1096 rbo = gem_to_radeon_bo(radeon_fb->obj);
|
/drivers/gpu/drm/udl/ |
H A D | udl_fb.c | 163 if (!fb->obj->vmapping) { 164 ret = udl_gem_vmap(fb->obj); 169 if (!fb->obj->vmapping) { 228 (char *) fb->obj->vmapping, 411 if (ufb->obj->base.import_attach) { 412 ret = dma_buf_begin_cpu_access(ufb->obj->base.import_attach->dmabuf, 413 0, ufb->obj->base.size, 427 if (ufb->obj->base.import_attach) { 428 dma_buf_end_cpu_access(ufb->obj->base.import_attach->dmabuf, 429 0, ufb->obj 457 udl_framebuffer_init(struct drm_device *dev, struct udl_framebuffer *ufb, struct drm_mode_fb_cmd2 *mode_cmd, struct udl_gem_object *obj) argument 482 struct udl_gem_object *obj; local 638 struct drm_gem_object *obj; local [all...] |
/drivers/iommu/ |
H A D | omap-iommu.h | 74 int (*enable)(struct omap_iommu *obj); 75 void (*disable)(struct omap_iommu *obj); 76 void (*set_twl)(struct omap_iommu *obj, bool on); 77 u32 (*fault_isr)(struct omap_iommu *obj, u32 *ra); 79 void (*tlb_read_cr)(struct omap_iommu *obj, struct cr_regs *cr); 80 void (*tlb_load_cr)(struct omap_iommu *obj, struct cr_regs *cr); 82 struct cr_regs *(*alloc_cr)(struct omap_iommu *obj, 87 ssize_t (*dump_cr)(struct omap_iommu *obj, struct cr_regs *cr, 92 void (*save_ctx)(struct omap_iommu *obj); 93 void (*restore_ctx)(struct omap_iommu *obj); 219 iommu_read_reg(struct omap_iommu *obj, size_t offs) argument 224 iommu_write_reg(struct omap_iommu *obj, u32 val, size_t offs) argument [all...] |
/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_lan_hmc.c | 103 struct i40e_hmc_obj_info *obj, *full_obj; local 127 obj = &hw->hmc.hmc_obj[I40E_HMC_LAN_TX]; 128 obj->max_cnt = rd32(hw, I40E_GLHMC_LANQMAX); 129 obj->cnt = txq_num; 130 obj->base = 0; 132 obj->size = (u64)1 << size_exp; 135 if (txq_num > obj->max_cnt) { 138 txq_num, obj->max_cnt, ret_code); 143 full_obj->max_cnt += obj->max_cnt; 144 full_obj->cnt += obj 467 struct i40e_hmc_obj_info *obj; local [all...] |
/drivers/platform/x86/ |
H A D | alienware-wmi.c | 447 union acpi_object *obj; local 463 obj = (union acpi_object *)output.pointer; 464 if (obj && obj->type == ACPI_TYPE_INTEGER) 465 *out_data = (u32) obj->integer.value;
|
H A D | wmi.c | 488 union acpi_object *obj; local 497 obj = (union acpi_object *)response.pointer; 499 if (!obj) 503 switch(obj->type) { 505 pr_cont("BUFFER_TYPE - length %d\n", obj->buffer.length); 508 pr_cont("STRING_TYPE - %s\n", obj->string.pointer); 511 pr_cont("INTEGER_TYPE - %llu\n", obj->integer.value); 514 pr_cont("PACKAGE_TYPE - %d elements\n", obj->package.count); 517 pr_cont("object type 0x%X\n", obj->type); 519 kfree(obj); 768 union acpi_object *obj; local [all...] |
/drivers/staging/lustre/include/linux/lnet/ |
H A D | lib-lnet.h | 207 lnet_freelist_free(lnet_freelist_t *fl, void *obj) argument 210 lnet_freeobj_t *o = list_entry(obj, lnet_freeobj_t, fo_contents);
|
/drivers/staging/lustre/lustre/llite/ |
H A D | lloop.c | 191 struct cl_object *obj = ll_i2info(inode)->lli_clob; local 209 io->ci_obj = obj; 210 ret = cl_io_init(env, io, CIT_MISC, obj);
|
H A D | vvp_io.c | 119 struct cl_object *obj = io->ci_obj; local 122 CLOBINVRNT(env, obj, ccc_object_invariant(obj)); 126 PFID(lu_object_fid(&obj->co_lu)), 136 rc = ll_layout_restore(ccc_object_inode(obj)); 161 ll_layout_refresh(ccc_object_inode(obj), &gen); 166 PFID(lu_object_fid(&obj->co_lu)), 171 ll_i2info(ccc_object_inode(obj))->lli_flags &= 420 struct cl_object *obj = io->ci_obj; local 425 cl_object_attr_lock(obj); 484 struct cl_object *obj = io->ci_obj; local 575 struct cl_object *obj = io->ci_obj; local 658 struct cl_object *obj = io->ci_obj; local 825 struct cl_object *obj = slice->cpl_obj; local 900 vvp_io_prepare_partial(const struct lu_env *env, struct cl_io *io, struct cl_object *obj, struct cl_page *pg, struct ccc_page *cp, unsigned from, unsigned to) argument 946 struct cl_object *obj = slice->cpl_obj; local 980 struct cl_object *obj = slice->cpl_obj; local 1139 vvp_io_init(const struct lu_env *env, struct cl_object *obj, struct cl_io *io) argument [all...] |