/drivers/video/ |
H A D | videomode.c | 15 struct videomode *vm) 17 vm->pixelclock = dt->pixelclock.typ; 18 vm->hactive = dt->hactive.typ; 19 vm->hfront_porch = dt->hfront_porch.typ; 20 vm->hback_porch = dt->hback_porch.typ; 21 vm->hsync_len = dt->hsync_len.typ; 23 vm->vactive = dt->vactive.typ; 24 vm->vfront_porch = dt->vfront_porch.typ; 25 vm->vback_porch = dt->vback_porch.typ; 26 vm 14 videomode_from_timing(const struct display_timing *dt, struct videomode *vm) argument 32 videomode_from_timings(const struct display_timings *disp, struct videomode *vm, unsigned int index) argument [all...] |
H A D | of_videomode.c | 19 * @vm - set to return value 26 * specified by index into *vm. This function should only be used, if 31 int of_get_videomode(struct device_node *np, struct videomode *vm, argument 46 ret = videomode_from_timings(disp, vm, index);
|
/drivers/gpu/drm/nouveau/core/subdev/vm/ |
H A D | nv04.h | 4 #include <subdev/vm.h> 8 struct nouveau_vm *vm; member in struct:nv04_vmmgr_priv
|
H A D | base.c | 29 #include <subdev/vm.h> 34 struct nouveau_vm *vm = vma->vm; local 35 struct nouveau_vmmgr *vmm = vm->vmm; 40 u32 pde = (offset >> vmm->pgt_bits) - vm->fpde; 51 struct nouveau_gpuobj *pgt = vm->pgt[pde].obj[big]; 72 vmm->flush(vm); 79 struct nouveau_vm *vm = vma->vm; local 80 struct nouveau_vmmgr *vmm = vm 137 struct nouveau_vm *vm = vma->vm; local 186 struct nouveau_vm *vm = vma->vm; local 225 nouveau_vm_unmap_pgt(struct nouveau_vm *vm, int big, u32 fpde, u32 lpde) argument 252 nouveau_vm_map_pgt(struct nouveau_vm *vm, u32 pde, u32 type) argument 289 nouveau_vm_get(struct nouveau_vm *vm, u64 size, u32 page_shift, u32 access, struct nouveau_vma *vma) argument 339 struct nouveau_vm *vm = vma->vm; local 360 struct nouveau_vm *vm; local 402 nouveau_vm_link(struct nouveau_vm *vm, struct nouveau_gpuobj *pgd) argument 426 nouveau_vm_unlink(struct nouveau_vm *vm, struct nouveau_gpuobj *mpgd) argument 452 struct nouveau_vm *vm = container_of(kref, typeof(*vm), refcount); local [all...] |
H A D | nv04.c | 64 nv04_vm_flush(struct nouveau_vm *vm) argument 109 &priv->vm); 116 &priv->vm->pgt[0].obj[0]); 117 dma = priv->vm->pgt[0].obj[0]; 118 priv->vm->pgt[0].refcount[0] = 1; 131 if (priv->vm) { 132 nouveau_gpuobj_ref(NULL, &priv->vm->pgt[0].obj[0]); 133 nouveau_vm_ref(NULL, &priv->vm, NULL);
|
H A D | nv41.c | 29 #include <subdev/vm.h> 68 nv41_vm_flush(struct nouveau_vm *vm) argument 70 struct nv04_vmmgr_priv *priv = (void *)vm->vmm; 118 &priv->vm); 125 &priv->vm->pgt[0].obj[0]); 126 priv->vm->pgt[0].refcount[0] = 1; 137 struct nouveau_gpuobj *dma = priv->vm->pgt[0].obj[0];
|
H A D | nv44.c | 29 #include <subdev/vm.h> 90 struct nv04_vmmgr_priv *priv = (void *)vma->vm->vmm; 143 nv44_vm_flush(struct nouveau_vm *vm) argument 145 struct nv04_vmmgr_priv *priv = (void *)vm->vmm; 195 &priv->vm); 202 &priv->vm->pgt[0].obj[0]); 203 priv->vm->pgt[0].refcount[0] = 1; 214 struct nouveau_gpuobj *gart = priv->vm->pgt[0].obj[0];
|
H A D | nv50.c | 31 #include <subdev/vm.h> 89 if (nouveau_fb(vma->vm->vmm)->ram->stolen) { 90 phys += nouveau_fb(vma->vm->vmm)->ram->stolen; 152 nv50_vm_flush(struct nouveau_vm *vm) argument 154 struct nv50_vmmgr_priv *priv = (void *)vm->vmm; 163 if (!atomic_read(&vm->engref[i])) 188 nv_error(priv, "vm flush timeout: engine %d\n", vme); 212 ret = nouveau_vmmgr_create(parent, engine, oclass, "VM", "vm", &priv);
|
/drivers/video/fbdev/omap2/dss/ |
H A D | display.c | 270 void videomode_to_omap_video_timings(const struct videomode *vm, argument 275 ovt->pixelclock = vm->pixelclock; 276 ovt->x_res = vm->hactive; 277 ovt->hbp = vm->hback_porch; 278 ovt->hfp = vm->hfront_porch; 279 ovt->hsw = vm->hsync_len; 280 ovt->y_res = vm->vactive; 281 ovt->vbp = vm->vback_porch; 282 ovt->vfp = vm->vfront_porch; 283 ovt->vsw = vm 302 omap_video_timings_to_videomode(const struct omap_video_timings *ovt, struct videomode *vm) argument [all...] |
/drivers/gpu/drm/radeon/ |
H A D | radeon_vm.c | 78 * radeon_vm_manager_init - init the vm manager 82 * Init the vm manager (cayman+). 100 * radeon_vm_manager_fini - tear down the vm manager 120 * radeon_vm_get_bos - add the vm BOs to a validation list 122 * @vm: vm providing the BOs 129 struct radeon_vm *vm, 135 list = drm_malloc_ab(vm->max_pde_used + 2, 140 /* add the vm page table to the list */ 142 list[0].robj = vm 128 radeon_vm_get_bos(struct radeon_device *rdev, struct radeon_vm *vm, struct list_head *head) argument 181 radeon_vm_grab_id(struct radeon_device *rdev, struct radeon_vm *vm, int ring) argument 236 radeon_vm_flush(struct radeon_device *rdev, struct radeon_vm *vm, int ring) argument 262 radeon_vm_fence(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_fence *fence) argument 292 radeon_vm_bo_find(struct radeon_vm *vm, struct radeon_bo *bo) argument 318 radeon_vm_bo_add(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_bo *bo) argument 460 struct radeon_vm *vm = bo_va->vm; local 638 radeon_vm_update_page_directory(struct radeon_device *rdev, struct radeon_vm *vm) argument 811 radeon_vm_update_ptes(struct radeon_device *rdev, struct radeon_vm *vm, struct radeon_ib *ib, uint64_t start, uint64_t end, uint64_t dst, uint32_t flags) argument 882 struct radeon_vm *vm = bo_va->vm; local 996 radeon_vm_clear_freed(struct radeon_device *rdev, struct radeon_vm *vm) argument 1024 radeon_vm_clear_invalids(struct radeon_device *rdev, struct radeon_vm *vm) argument 1051 struct radeon_vm *vm = bo_va->vm; local 1101 radeon_vm_init(struct radeon_device *rdev, struct radeon_vm *vm) argument 1154 radeon_vm_fini(struct radeon_device *rdev, struct radeon_vm *vm) argument [all...] |
H A D | radeon_ib.c | 56 struct radeon_ib *ib, struct radeon_vm *vm, 75 ib->vm = vm; 76 if (vm) { 144 /* grab a vm id if necessary */ 145 if (ib->vm) { 147 vm_id_fence = radeon_vm_grab_id(rdev, ib->vm, ib->ring); 159 if (ib->vm) 160 radeon_vm_flush(rdev, ib->vm, ib->ring); 177 if (ib->vm) 55 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib *ib, struct radeon_vm *vm, unsigned size) argument [all...] |
H A D | si_dma.c | 188 void si_dma_vm_flush(struct radeon_device *rdev, int ridx, struct radeon_vm *vm) argument 192 if (vm == NULL) 196 if (vm->id < 8) { 197 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + (vm->id << 2)) >> 2)); 199 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT8_PAGE_TABLE_BASE_ADDR + ((vm->id - 8) << 2)) >> 2)); 201 radeon_ring_write(ring, vm->pd_gpu_addr >> 12); 211 radeon_ring_write(ring, 1 << vm->id);
|
H A D | ni_dma.c | 143 radeon_ring_write(ring, DMA_IB_PACKET(DMA_PACKET_INDIRECT_BUFFER, ib->vm ? ib->vm->id : 0, 0)); 449 void cayman_dma_vm_flush(struct radeon_device *rdev, int ridx, struct radeon_vm *vm) argument 453 if (vm == NULL) 457 radeon_ring_write(ring, (0xf << 16) | ((VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + (vm->id << 2)) >> 2)); 458 radeon_ring_write(ring, vm->pd_gpu_addr >> 12); 468 radeon_ring_write(ring, 1 << vm->id);
|
H A D | radeon_cs.c | 196 p->vm_bos = radeon_vm_get_bos(p->rdev, p->ib.vm, 497 struct radeon_vm *vm) 503 r = radeon_vm_update_page_directory(rdev, vm); 507 r = radeon_vm_clear_freed(rdev, vm); 511 if (vm->ib_bo_va == NULL) { 516 r = radeon_vm_bo_update(rdev, vm->ib_bo_va, 529 bo_va = radeon_vm_bo_find(vm, bo); 531 dev_err(rdev->dev, "bo %p not in vm %p\n", bo, vm); 540 return radeon_vm_clear_invalids(rdev, vm); 496 radeon_bo_vm_update_pte(struct radeon_cs_parser *p, struct radeon_vm *vm) argument 547 struct radeon_vm *vm = &fpriv->vm; local 609 struct radeon_vm *vm = NULL; local [all...] |
/drivers/gpu/drm/i915/ |
H A D | i915_gem_evict.c | 52 * @vm: address space to evict from 69 i915_gem_evict_something(struct drm_device *dev, struct i915_address_space *vm, argument 105 if (start != 0 || end != vm->total) { 106 drm_mm_init_scan_with_range(&vm->mm, min_size, 110 drm_mm_init_scan(&vm->mm, min_size, alignment, cache_level); 114 list_for_each_entry(vma, &vm->inactive_list, mm_list) { 123 list_for_each_entry(vma, &vm->active_list, mm_list) { 198 * i915_gem_evict_vm - Evict all idle vmas from a vm 200 * @vm: Address space to cleanse 203 * This function evicts all idles vmas from a vm 212 i915_gem_evict_vm(struct i915_address_space *vm, bool do_idle) argument 246 struct i915_address_space *vm, *v; local [all...] |
H A D | i915_gem_gtt.c | 246 static void gen8_ppgtt_clear_range(struct i915_address_space *vm, argument 252 container_of(vm, struct i915_hw_ppgtt, base); 289 static void gen8_ppgtt_insert_entries(struct i915_address_space *vm, argument 295 container_of(vm, struct i915_hw_ppgtt, base); 381 static void gen8_ppgtt_cleanup(struct i915_address_space *vm) argument 384 container_of(vm, struct i915_hw_ppgtt, base); 620 struct i915_address_space *vm = &ppgtt->base; local 626 scratch_pte = vm->pte_encode(vm->scratch.addr, I915_CACHE_LLC, true, 0); 631 seq_printf(m, " VM %p (pd_offset %x-%x):\n", vm, 831 gen6_ppgtt_clear_range(struct i915_address_space *vm, uint64_t start, uint64_t length, bool use_scratch) argument 865 gen6_ppgtt_insert_entries(struct i915_address_space *vm, struct sg_table *pages, uint64_t start, enum i915_cache_level cache_level, u32 flags) argument 920 gen6_ppgtt_cleanup(struct i915_address_space *vm) argument 1310 struct i915_address_space *vm; local 1382 gen8_ggtt_insert_entries(struct i915_address_space *vm, struct sg_table *st, uint64_t start, enum i915_cache_level level, u32 unused) argument 1428 gen6_ggtt_insert_entries(struct i915_address_space *vm, struct sg_table *st, uint64_t start, enum i915_cache_level level, u32 flags) argument 1466 gen8_ggtt_clear_range(struct i915_address_space *vm, uint64_t start, uint64_t length, bool use_scratch) argument 1492 gen6_ggtt_clear_range(struct i915_address_space *vm, uint64_t start, uint64_t length, bool use_scratch) argument 1531 i915_ggtt_clear_range(struct i915_address_space *vm, uint64_t start, uint64_t length, bool unused) argument 1742 struct i915_address_space *vm = &dev_priv->gtt.base; local 2037 gen6_gmch_remove(struct i915_address_space *vm) argument 2072 i915_gmch_remove(struct i915_address_space *vm) argument 2132 __i915_gem_vma_create(struct drm_i915_gem_object *obj, struct i915_address_space *vm) argument 2181 i915_gem_obj_lookup_or_create_vma(struct drm_i915_gem_object *obj, struct i915_address_space *vm) argument [all...] |
H A D | i915_trace.h | 117 __field(struct i915_address_space *, vm) 125 __entry->vm = vma->vm; 131 TP_printk("obj=%p, offset=%08x size=%x%s vm=%p", 134 __entry->vm) 143 __field(struct i915_address_space *, vm) 150 __entry->vm = vma->vm; 155 TP_printk("obj=%p, offset=%08x size=%x vm=%p", 156 __entry->obj, __entry->offset, __entry->size, __entry->vm) [all...] |
/drivers/gpu/drm/nouveau/core/subdev/bar/ |
H A D | nvc0.c | 29 #include <subdev/vm.h> 36 struct nouveau_vm *vm; member in struct:nvc0_bar_priv_vm 52 ret = nouveau_vm_get(priv->bar[0].vm, mem->size << 12, 12, flags, vma); 67 ret = nouveau_vm_get(priv->bar[1].vm, mem->size << 12, 88 struct nouveau_vm *vm; local 104 ret = nouveau_vm_new(device, 0, bar_len, 0, &vm); 108 atomic_inc(&vm->engref[NVDEV_SUBDEV_BAR]); 117 &vm->pgt[0].obj[0]); 118 vm->pgt[0].refcount[0] = 1; 123 ret = nouveau_vm_ref(vm, [all...] |
H A D | nv50.c | 29 #include <subdev/vm.h> 113 struct nouveau_vm *vm; local 144 ret = nouveau_vm_new(device, start, limit, start, &vm); 148 atomic_inc(&vm->engref[NVDEV_SUBDEV_BAR]); 152 NVOBJ_FLAG_ZERO_ALLOC, &vm->pgt[0].obj[0]); 153 vm->pgt[0].refcount[0] = 1; 157 ret = nouveau_vm_ref(vm, &priv->bar3_vm, priv->pgd); 158 nouveau_vm_ref(NULL, &vm, NULL); 178 ret = nouveau_vm_new(device, start, limit--, start, &vm); 182 atomic_inc(&vm [all...] |
/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_dpi.c | 32 struct videomode *vm; member in struct:exynos_dpi 67 if (ctx->vm) { 75 drm_display_mode_from_videomode(ctx->vm, mode); 271 struct videomode *vm; local 276 vm = devm_kzalloc(dev, sizeof(*ctx->vm), GFP_KERNEL); 277 if (!vm) 280 ret = of_get_videomode(dn, vm, 0); 282 devm_kfree(dev, vm); 286 ctx->vm [all...] |
/drivers/gpu/drm/nouveau/core/engine/fifo/ |
H A D | nv50.h | 16 struct nouveau_vm *vm; member in struct:nv50_fifo_base
|
/drivers/video/fbdev/core/ |
H A D | fbmon.c | 1387 int fb_videomode_from_videomode(const struct videomode *vm, argument 1392 fbmode->xres = vm->hactive; 1393 fbmode->left_margin = vm->hback_porch; 1394 fbmode->right_margin = vm->hfront_porch; 1395 fbmode->hsync_len = vm->hsync_len; 1397 fbmode->yres = vm->vactive; 1398 fbmode->upper_margin = vm->vback_porch; 1399 fbmode->lower_margin = vm->vfront_porch; 1400 fbmode->vsync_len = vm->vsync_len; 1403 fbmode->pixclock = vm 1458 struct videomode vm; local [all...] |
/drivers/gpu/drm/ |
H A D | drm_modes.c | 578 * drm_display_mode_from_videomode - fill in @dmode using @vm, 579 * @vm: videomode structure to use as source 582 * Fills out @dmode using the display mode specified in @vm. 584 void drm_display_mode_from_videomode(const struct videomode *vm, argument 587 dmode->hdisplay = vm->hactive; 588 dmode->hsync_start = dmode->hdisplay + vm->hfront_porch; 589 dmode->hsync_end = dmode->hsync_start + vm->hsync_len; 590 dmode->htotal = dmode->hsync_end + vm->hback_porch; 592 dmode->vdisplay = vm->vactive; 593 dmode->vsync_start = dmode->vdisplay + vm 635 struct videomode vm; local [all...] |
/drivers/video/fbdev/ |
H A D | sh7760fb.c | 202 struct fb_videomode *vm = par->pd->def_mode; local 212 if (par->rot && (vm->xres > 320)) { 218 hsynp = vm->right_margin + vm->xres; 219 hsynw = vm->hsync_len; 220 htcn = vm->left_margin + hsynp + hsynw; 221 hdcn = vm->xres; 222 vsynp = vm->lower_margin + vm->yres; 223 vsynw = vm [all...] |
/drivers/gpu/drm/nouveau/core/core/ |
H A D | engctx.c | 31 #include <subdev/vm.h> 104 if (client->vm) 105 atomic_inc(&client->vm->engref[nv_engidx(engobj)]); 125 if (client->vm) 126 atomic_dec(&client->vm->engref[nv_engidx(engobj)]);
|