Searched defs:vma (Results 1 - 25 of 219) sorted by relevance

123456789

/drivers/auxdisplay/
H A Dcfag12864bfb.c67 static int cfag12864bfb_mmap(struct fb_info *info, struct vm_area_struct *vma) argument
69 return vm_insert_page(vma, vma->vm_start,
/drivers/char/
H A Duv_mmtimer.c43 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma);
142 * @vma: VMA to map the registers into
147 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma) argument
151 if (vma->vm_end - vma->vm_start != PAGE_SIZE)
154 if (vma->vm_flags & VM_WRITE)
160 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
166 if (remap_pfn_range(vma, vma
[all...]
/drivers/gpu/drm/nouveau/core/include/core/
H A Dengctx.h14 struct nouveau_vma vma; member in struct:nouveau_engctx
/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_ttm_glue.c31 int vmw_mmap(struct file *filp, struct vm_area_struct *vma) argument
36 if (unlikely(vma->vm_pgoff < VMWGFX_FILE_PAGE_OFFSET)) {
43 return ttm_bo_mmap(filp, vma, &dev_priv->bdev);
/drivers/misc/cxl/
H A Dcontext.c101 * Map a per-context mmio space into the given vma.
103 int cxl_context_iomap(struct cxl_context *ctx, struct vm_area_struct *vma) argument
105 u64 len = vma->vm_end - vma->vm_start;
109 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
110 return vm_iomap_memory(vma, ctx->afu->psn_phys, ctx->afu->adapter->ps_size);
126 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
127 return vm_iomap_memory(vma, ct
[all...]
/drivers/video/fbdev/
H A Damba-clcd-versatile.c170 int versatile_clcd_mmap_dma(struct clcd_fb *fb, struct vm_area_struct *vma) argument
172 return dma_mmap_writecombine(&fb->dev->dev, vma,
/drivers/xen/xenfs/
H A Dxenstored.c33 static int xsd_kva_mmap(struct file *file, struct vm_area_struct *vma) argument
35 size_t size = vma->vm_end - vma->vm_start;
37 if ((size > PAGE_SIZE) || (vma->vm_pgoff != 0))
40 if (remap_pfn_range(vma, vma->vm_start,
42 size, vma->vm_page_prot))
/drivers/gpu/drm/exynos/
H A Dexynos_drm_gem.h62 * @vma: a pointer to vm_area.
72 struct vm_area_struct *vma; member in struct:exynos_drm_gem_obj
143 int exynos_drm_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf);
146 int exynos_drm_gem_mmap(struct file *filp, struct vm_area_struct *vma);
148 static inline int vma_is_io(struct vm_area_struct *vma) argument
150 return !!(vma->vm_flags & (VM_IO | VM_PFNMAP));
154 struct vm_area_struct *exynos_gem_get_vma(struct vm_area_struct *vma);
157 void exynos_gem_put_vma(struct vm_area_struct *vma);
163 struct vm_area_struct *vma);
168 struct vm_area_struct *vma);
[all...]
H A Dexynos_drm_dmabuf.c166 struct vm_area_struct *vma)
165 exynos_gem_dmabuf_mmap(struct dma_buf *dma_buf, struct vm_area_struct *vma) argument
H A Dexynos_drm_fbdev.c39 struct vm_area_struct *vma)
48 vma->vm_flags |= VM_IO | VM_DONTEXPAND | VM_DONTDUMP;
50 vm_size = vma->vm_end - vma->vm_start;
55 ret = dma_mmap_attrs(helper->dev->dev, vma, buffer->pages,
38 exynos_drm_fb_mmap(struct fb_info *info, struct vm_area_struct *vma) argument
/drivers/infiniband/hw/ipath/
H A Dipath_mmap.c64 static void ipath_vma_open(struct vm_area_struct *vma) argument
66 struct ipath_mmap_info *ip = vma->vm_private_data;
71 static void ipath_vma_close(struct vm_area_struct *vma) argument
73 struct ipath_mmap_info *ip = vma->vm_private_data;
86 * @vma: the VMA to be initialized
89 int ipath_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) argument
92 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
93 unsigned long size = vma->vm_end - vma->vm_start;
115 ret = remap_vmalloc_range(vma, i
[all...]
/drivers/infiniband/hw/qib/
H A Dqib_mmap.c64 static void qib_vma_open(struct vm_area_struct *vma) argument
66 struct qib_mmap_info *ip = vma->vm_private_data;
71 static void qib_vma_close(struct vm_area_struct *vma) argument
73 struct qib_mmap_info *ip = vma->vm_private_data;
86 * @vma: the VMA to be initialized
89 int qib_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) argument
92 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
93 unsigned long size = vma->vm_end - vma->vm_start;
115 ret = remap_vmalloc_range(vma, i
[all...]
/drivers/xen/xenbus/
H A Dxenbus_dev_backend.c93 static int xenbus_backend_mmap(struct file *file, struct vm_area_struct *vma) argument
95 size_t size = vma->vm_end - vma->vm_start;
100 if ((size > PAGE_SIZE) || (vma->vm_pgoff != 0))
103 if (remap_pfn_range(vma, vma->vm_start,
105 size, vma->vm_page_prot))
/drivers/gpu/drm/gma500/
H A Dgem.c159 * @vma: the VMA of the GEM object
171 * vma->vm_private_data points to the GEM object that is backing this
174 int psb_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf) argument
184 obj = vma->vm_private_data; /* GEM object */
207 page_offset = ((unsigned long) vmf->virtual_address - vma->vm_start)
215 ret = vm_insert_pfn(vma, (unsigned long)vmf->virtual_address, pfn);
/drivers/gpu/drm/i915/
H A Di915_gem_evict.c37 mark_free(struct i915_vma *vma, struct list_head *unwind) argument
39 if (vma->pin_count)
42 if (WARN_ON(!list_empty(&vma->exec_list)))
45 list_add(&vma->exec_list, unwind);
46 return drm_mm_scan_add_block(&vma->node);
63 * This function is used by the object/vma binding code.
75 struct i915_vma *vma; local
114 list_for_each_entry(vma, &vm->inactive_list, mm_list) {
115 if (mark_free(vma, &unwind_list))
123 list_for_each_entry(vma,
214 struct i915_vma *vma, *next; local
[all...]
/drivers/gpu/drm/nouveau/core/subdev/bar/
H A Dbase.c34 struct nouveau_vma vma; member in struct:nouveau_barobj
55 ret = bar->kmap(bar, mem, NV_MEM_ACCESS_RW, &barobj->vma);
60 (u32)barobj->vma.offset, mem->size << 12);
74 if (barobj->vma.node) {
77 bar->unmap(bar, &barobj->vma);
/drivers/gpu/drm/nouveau/core/subdev/vm/
H A Dnv41.c41 nv41_vm_map_sg(struct nouveau_vma *vma, struct nouveau_gpuobj *pgt, argument
/drivers/gpu/drm/nouveau/
H A Dnouveau_chan.h19 struct nouveau_vma vma; member in struct:nouveau_channel::__anon894
H A Dnouveau_dma.c74 if (val < chan->push.vma.offset ||
75 val > chan->push.vma.offset + (chan->dma.max << 2))
78 return (val - chan->push.vma.offset) >> 2;
87 struct nouveau_vma *vma; local
91 vma = nouveau_bo_vma_find(bo, cli->vm);
92 BUG_ON(!vma);
93 offset = vma->offset + delta;
227 OUT_RING(chan, chan->push.vma.offset | 0x20000000);
/drivers/gpu/drm/omapdrm/
H A Domap_gem_dmabuf.c140 struct vm_area_struct *vma)
150 ret = drm_gem_mmap_obj(obj, omap_gem_mmap_size(obj), vma);
155 return omap_gem_mmap_obj(obj, vma);
139 omap_gem_dmabuf_mmap(struct dma_buf *buffer, struct vm_area_struct *vma) argument
/drivers/media/v4l2-core/
H A Dvideobuf2-memops.c27 * @vma: given virtual memory area
36 struct vm_area_struct *vb2_get_vma(struct vm_area_struct *vma) argument
44 if (vma->vm_ops && vma->vm_ops->open)
45 vma->vm_ops->open(vma);
47 if (vma->vm_file)
48 get_file(vma->vm_file);
50 memcpy(vma_copy, vma, sizeof(*vma));
67 vb2_put_vma(struct vm_area_struct *vma) argument
100 struct vm_area_struct *vma; local
146 vb2_common_vm_open(struct vm_area_struct *vma) argument
164 vb2_common_vm_close(struct vm_area_struct *vma) argument
[all...]
/drivers/misc/mic/host/
H A Dmic_fops.c192 mic_mmap(struct file *f, struct vm_area_struct *vma) argument
195 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
196 unsigned long pa, size = vma->vm_end - vma->vm_start, size_rem = size;
203 if (vma->vm_flags & VM_WRITE)
210 err = remap_pfn_range(vma, vma->vm_start + offset,
211 pa >> PAGE_SHIFT, size, vma->vm_page_prot);
215 "%s %d type %d size 0x%lx off 0x%lx pa 0x%lx vma 0x%lx\n",
217 pa, vma
[all...]
/drivers/staging/unisys/visorchipset/
H A Dfile.c38 static int visorchipset_mmap(struct file *file, struct vm_area_struct *vma);
138 visorchipset_mmap(struct file *file, struct vm_area_struct *vma) argument
141 ulong offset = vma->vm_pgoff << PAGE_SHIFT;
152 vma->vm_flags |= VM_IO;
167 if (remap_pfn_range(vma, vma->vm_start,
169 vma->vm_end - vma->vm_start,
171 (vma->vm_page_prot))) {
/drivers/char/agp/
H A Dalpha-agp.c14 static int alpha_core_agp_vm_fault(struct vm_area_struct *vma, argument
22 dma_addr = (unsigned long)vmf->virtual_address - vma->vm_start
/drivers/gpu/drm/
H A Ddrm_gem_cma_helper.c233 struct vm_area_struct *vma)
242 vma->vm_flags &= ~VM_PFNMAP;
243 vma->vm_pgoff = 0;
245 ret = dma_mmap_writecombine(cma_obj->base.dev->dev, vma,
247 vma->vm_end - vma->vm_start);
249 drm_gem_vm_close(vma);
257 int drm_gem_cma_mmap(struct file *filp, struct vm_area_struct *vma) argument
263 ret = drm_gem_mmap(filp, vma);
267 gem_obj = vma
232 drm_gem_cma_mmap_obj(struct drm_gem_cma_object *cma_obj, struct vm_area_struct *vma) argument
342 drm_gem_cma_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) argument
[all...]

Completed in 447 milliseconds

123456789