/drivers/auxdisplay/ |
H A D | cfag12864bfb.c | 67 static int cfag12864bfb_mmap(struct fb_info *info, struct vm_area_struct *vma) argument 69 return vm_insert_page(vma, vma->vm_start,
|
/drivers/char/ |
H A D | uv_mmtimer.c | 43 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma); 142 * @vma: VMA to map the registers into 147 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma) argument 151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) 154 if (vma->vm_flags & VM_WRITE) 160 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); 166 if (remap_pfn_range(vma, vma [all...] |
/drivers/gpu/drm/nouveau/core/include/core/ |
H A D | engctx.h | 14 struct nouveau_vma vma; member in struct:nouveau_engctx
|
/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_ttm_glue.c | 31 int vmw_mmap(struct file *filp, struct vm_area_struct *vma) argument 36 if (unlikely(vma->vm_pgoff < VMWGFX_FILE_PAGE_OFFSET)) { 43 return ttm_bo_mmap(filp, vma, &dev_priv->bdev);
|
/drivers/misc/cxl/ |
H A D | context.c | 101 * Map a per-context mmio space into the given vma. 103 int cxl_context_iomap(struct cxl_context *ctx, struct vm_area_struct *vma) argument 105 u64 len = vma->vm_end - vma->vm_start; 109 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); 110 return vm_iomap_memory(vma, ctx->afu->psn_phys, ctx->afu->adapter->ps_size); 126 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); 127 return vm_iomap_memory(vma, ct [all...] |
/drivers/video/fbdev/ |
H A D | amba-clcd-versatile.c | 170 int versatile_clcd_mmap_dma(struct clcd_fb *fb, struct vm_area_struct *vma) argument 172 return dma_mmap_writecombine(&fb->dev->dev, vma,
|
/drivers/xen/xenfs/ |
H A D | xenstored.c | 33 static int xsd_kva_mmap(struct file *file, struct vm_area_struct *vma) argument 35 size_t size = vma->vm_end - vma->vm_start; 37 if ((size > PAGE_SIZE) || (vma->vm_pgoff != 0)) 40 if (remap_pfn_range(vma, vma->vm_start, 42 size, vma->vm_page_prot))
|
/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_gem.h | 62 * @vma: a pointer to vm_area. 72 struct vm_area_struct *vma; member in struct:exynos_drm_gem_obj 143 int exynos_drm_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf); 146 int exynos_drm_gem_mmap(struct file *filp, struct vm_area_struct *vma); 148 static inline int vma_is_io(struct vm_area_struct *vma) argument 150 return !!(vma->vm_flags & (VM_IO | VM_PFNMAP)); 154 struct vm_area_struct *exynos_gem_get_vma(struct vm_area_struct *vma); 157 void exynos_gem_put_vma(struct vm_area_struct *vma); 163 struct vm_area_struct *vma); 168 struct vm_area_struct *vma); [all...] |
H A D | exynos_drm_dmabuf.c | 166 struct vm_area_struct *vma) 165 exynos_gem_dmabuf_mmap(struct dma_buf *dma_buf, struct vm_area_struct *vma) argument
|
H A D | exynos_drm_fbdev.c | 39 struct vm_area_struct *vma) 48 vma->vm_flags |= VM_IO | VM_DONTEXPAND | VM_DONTDUMP; 50 vm_size = vma->vm_end - vma->vm_start; 55 ret = dma_mmap_attrs(helper->dev->dev, vma, buffer->pages, 38 exynos_drm_fb_mmap(struct fb_info *info, struct vm_area_struct *vma) argument
|
/drivers/infiniband/hw/ipath/ |
H A D | ipath_mmap.c | 64 static void ipath_vma_open(struct vm_area_struct *vma) argument 66 struct ipath_mmap_info *ip = vma->vm_private_data; 71 static void ipath_vma_close(struct vm_area_struct *vma) argument 73 struct ipath_mmap_info *ip = vma->vm_private_data; 86 * @vma: the VMA to be initialized 89 int ipath_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) argument 92 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; 93 unsigned long size = vma->vm_end - vma->vm_start; 115 ret = remap_vmalloc_range(vma, i [all...] |
/drivers/infiniband/hw/qib/ |
H A D | qib_mmap.c | 64 static void qib_vma_open(struct vm_area_struct *vma) argument 66 struct qib_mmap_info *ip = vma->vm_private_data; 71 static void qib_vma_close(struct vm_area_struct *vma) argument 73 struct qib_mmap_info *ip = vma->vm_private_data; 86 * @vma: the VMA to be initialized 89 int qib_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) argument 92 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; 93 unsigned long size = vma->vm_end - vma->vm_start; 115 ret = remap_vmalloc_range(vma, i [all...] |
/drivers/xen/xenbus/ |
H A D | xenbus_dev_backend.c | 93 static int xenbus_backend_mmap(struct file *file, struct vm_area_struct *vma) argument 95 size_t size = vma->vm_end - vma->vm_start; 100 if ((size > PAGE_SIZE) || (vma->vm_pgoff != 0)) 103 if (remap_pfn_range(vma, vma->vm_start, 105 size, vma->vm_page_prot))
|
/drivers/gpu/drm/gma500/ |
H A D | gem.c | 159 * @vma: the VMA of the GEM object 171 * vma->vm_private_data points to the GEM object that is backing this 174 int psb_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf) argument 184 obj = vma->vm_private_data; /* GEM object */ 207 page_offset = ((unsigned long) vmf->virtual_address - vma->vm_start) 215 ret = vm_insert_pfn(vma, (unsigned long)vmf->virtual_address, pfn);
|
/drivers/gpu/drm/i915/ |
H A D | i915_gem_evict.c | 37 mark_free(struct i915_vma *vma, struct list_head *unwind) argument 39 if (vma->pin_count) 42 if (WARN_ON(!list_empty(&vma->exec_list))) 45 list_add(&vma->exec_list, unwind); 46 return drm_mm_scan_add_block(&vma->node); 63 * This function is used by the object/vma binding code. 75 struct i915_vma *vma; local 114 list_for_each_entry(vma, &vm->inactive_list, mm_list) { 115 if (mark_free(vma, &unwind_list)) 123 list_for_each_entry(vma, 214 struct i915_vma *vma, *next; local [all...] |
/drivers/gpu/drm/nouveau/core/subdev/bar/ |
H A D | base.c | 34 struct nouveau_vma vma; member in struct:nouveau_barobj 55 ret = bar->kmap(bar, mem, NV_MEM_ACCESS_RW, &barobj->vma); 60 (u32)barobj->vma.offset, mem->size << 12); 74 if (barobj->vma.node) { 77 bar->unmap(bar, &barobj->vma);
|
/drivers/gpu/drm/nouveau/core/subdev/vm/ |
H A D | nv41.c | 41 nv41_vm_map_sg(struct nouveau_vma *vma, struct nouveau_gpuobj *pgt, argument
|
/drivers/gpu/drm/nouveau/ |
H A D | nouveau_chan.h | 19 struct nouveau_vma vma; member in struct:nouveau_channel::__anon894
|
H A D | nouveau_dma.c | 74 if (val < chan->push.vma.offset || 75 val > chan->push.vma.offset + (chan->dma.max << 2)) 78 return (val - chan->push.vma.offset) >> 2; 87 struct nouveau_vma *vma; local 91 vma = nouveau_bo_vma_find(bo, cli->vm); 92 BUG_ON(!vma); 93 offset = vma->offset + delta; 227 OUT_RING(chan, chan->push.vma.offset | 0x20000000);
|
/drivers/gpu/drm/omapdrm/ |
H A D | omap_gem_dmabuf.c | 140 struct vm_area_struct *vma) 150 ret = drm_gem_mmap_obj(obj, omap_gem_mmap_size(obj), vma); 155 return omap_gem_mmap_obj(obj, vma); 139 omap_gem_dmabuf_mmap(struct dma_buf *buffer, struct vm_area_struct *vma) argument
|
/drivers/media/v4l2-core/ |
H A D | videobuf2-memops.c | 27 * @vma: given virtual memory area 36 struct vm_area_struct *vb2_get_vma(struct vm_area_struct *vma) argument 44 if (vma->vm_ops && vma->vm_ops->open) 45 vma->vm_ops->open(vma); 47 if (vma->vm_file) 48 get_file(vma->vm_file); 50 memcpy(vma_copy, vma, sizeof(*vma)); 67 vb2_put_vma(struct vm_area_struct *vma) argument 100 struct vm_area_struct *vma; local 146 vb2_common_vm_open(struct vm_area_struct *vma) argument 164 vb2_common_vm_close(struct vm_area_struct *vma) argument [all...] |
/drivers/misc/mic/host/ |
H A D | mic_fops.c | 192 mic_mmap(struct file *f, struct vm_area_struct *vma) argument 195 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; 196 unsigned long pa, size = vma->vm_end - vma->vm_start, size_rem = size; 203 if (vma->vm_flags & VM_WRITE) 210 err = remap_pfn_range(vma, vma->vm_start + offset, 211 pa >> PAGE_SHIFT, size, vma->vm_page_prot); 215 "%s %d type %d size 0x%lx off 0x%lx pa 0x%lx vma 0x%lx\n", 217 pa, vma [all...] |
/drivers/staging/unisys/visorchipset/ |
H A D | file.c | 38 static int visorchipset_mmap(struct file *file, struct vm_area_struct *vma); 138 visorchipset_mmap(struct file *file, struct vm_area_struct *vma) argument 141 ulong offset = vma->vm_pgoff << PAGE_SHIFT; 152 vma->vm_flags |= VM_IO; 167 if (remap_pfn_range(vma, vma->vm_start, 169 vma->vm_end - vma->vm_start, 171 (vma->vm_page_prot))) {
|
/drivers/char/agp/ |
H A D | alpha-agp.c | 14 static int alpha_core_agp_vm_fault(struct vm_area_struct *vma, argument 22 dma_addr = (unsigned long)vmf->virtual_address - vma->vm_start
|
/drivers/gpu/drm/ |
H A D | drm_gem_cma_helper.c | 233 struct vm_area_struct *vma) 242 vma->vm_flags &= ~VM_PFNMAP; 243 vma->vm_pgoff = 0; 245 ret = dma_mmap_writecombine(cma_obj->base.dev->dev, vma, 247 vma->vm_end - vma->vm_start); 249 drm_gem_vm_close(vma); 257 int drm_gem_cma_mmap(struct file *filp, struct vm_area_struct *vma) argument 263 ret = drm_gem_mmap(filp, vma); 267 gem_obj = vma 232 drm_gem_cma_mmap_obj(struct drm_gem_cma_object *cma_obj, struct vm_area_struct *vma) argument 342 drm_gem_cma_prime_mmap(struct drm_gem_object *obj, struct vm_area_struct *vma) argument [all...] |