/mm/ |
H A D | mincore.c | 62 static unsigned char mincore_page(struct address_space *mapping, pgoff_t pgoff) argument 75 page = find_get_entry(mapping, pgoff); 85 page = find_get_page(mapping, pgoff); 87 page = find_get_page(mapping, pgoff); 105 pgoff_t pgoff; local 107 pgoff = linear_page_index(vma, addr); 108 for (i = 0; i < nr; i++, pgoff++) 109 vec[i] = mincore_page(vma->vm_file->f_mapping, pgoff); 127 pgoff_t pgoff; local 135 pgoff [all...] |
H A D | fremap.c | 73 unsigned long addr, unsigned long pgoff, pgprot_t prot) 83 ptfile = pgoff_to_pte(pgoff); 103 unsigned long size, pgoff_t pgoff) 109 err = install_file_pte(mm, vma, addr, pgoff, vma->vm_page_prot); 115 pgoff++; 127 * @pgoff: to-be-mapped page of the backing store file 143 unsigned long, prot, unsigned long, pgoff, unsigned long, flags) 168 /* Does pgoff wrap? */ 169 if (pgoff + (size >> PAGE_SHIFT) < pgoff) 72 install_file_pte(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, unsigned long pgoff, pgprot_t prot) argument 102 generic_file_remap_pages(struct vm_area_struct *vma, unsigned long addr, unsigned long size, pgoff_t pgoff) argument [all...] |
H A D | filemap_xip.c | 162 * __xip_sparse_page when found at pgoff. 166 unsigned long pgoff) 186 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) { 189 ((pgoff - vma->vm_pgoff) << PAGE_SHIFT); 236 if (vmf->pgoff >= size) 239 error = mapping->a_ops->get_xip_mem(mapping, vmf->pgoff, 0, 254 error = mapping->a_ops->get_xip_mem(mapping, vmf->pgoff, 1, 259 /* unmap sparse mappings at pgoff from all other vmas */ 260 __xip_unmap(mapping, vmf->pgoff); 165 __xip_unmap(struct address_space * mapping, unsigned long pgoff) argument [all...] |
H A D | mmap.c | 729 unsigned long end, pgoff_t pgoff, struct vm_area_struct *insert) 840 vma->vm_pgoff = pgoff; 1014 * Given a mapping request (addr,end,vm_flags,file,pgoff,anon_name), 1046 pgoff_t pgoff, struct mempolicy *policy, 1074 file, pgoff, anon_name)) { 1081 file, pgoff+pglen, anon_name) && 1102 file, pgoff+pglen, anon_name)) { 1268 unsigned long flags, unsigned long pgoff, 1298 if ((pgoff + (len >> PAGE_SHIFT)) < pgoff) 728 vma_adjust(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff, struct vm_area_struct *insert) argument 1042 vma_merge(struct mm_struct *mm, struct vm_area_struct *prev, unsigned long addr, unsigned long end, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t pgoff, struct mempolicy *policy, const char __user *anon_name) argument 1266 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *populate) argument 1541 mmap_region(struct file *file, unsigned long addr, unsigned long len, vm_flags_t vm_flags, unsigned long pgoff) argument 1922 arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 1958 arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, const unsigned long len, const unsigned long pgoff, const unsigned long flags) argument 2009 get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 2656 pgoff_t pgoff = addr >> PAGE_SHIFT; local 2837 copy_vma(struct vm_area_struct **vmap, unsigned long addr, unsigned long len, pgoff_t pgoff, bool *need_rmap_locks) argument 2958 pgoff_t pgoff; local [all...] |
H A D | nommu.c | 918 unsigned long pgoff, 945 if ((pgoff + (rlen >> PAGE_SHIFT)) < pgoff) 1266 unsigned long pgoff, 1275 kenter(",%lx,%lx,%lx,%lx,%lx", addr, len, prot, flags, pgoff); 1281 ret = validate_mmap_request(file, addr, len, prot, flags, pgoff, 1307 region->vm_pgoff = pgoff; 1311 vma->vm_pgoff = pgoff; 1333 pgend = pgoff + pglen; 1352 if (pgoff > 913 validate_mmap_request(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *_capabilities) argument 1261 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *populate) argument 1862 remap_vmalloc_range(struct vm_area_struct *vma, void *addr, unsigned long pgoff) argument 1877 arch_get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 1997 generic_file_remap_pages(struct vm_area_struct *vma, unsigned long addr, unsigned long size, pgoff_t pgoff) argument [all...] |
H A D | mremap.c | 354 unsigned long pgoff; local 358 pgoff = (addr - vma->vm_start) >> PAGE_SHIFT; 359 pgoff += vma->vm_pgoff; 360 if (pgoff + (new_len >> PAGE_SHIFT) < pgoff)
|
H A D | memory-failure.c | 432 pgoff_t pgoff; local 438 pgoff = page_to_pgoff(page); 447 pgoff, pgoff) { 472 pgoff_t pgoff = page_to_pgoff(page); local 477 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, 478 pgoff) {
|
H A D | madvise.c | 51 pgoff_t pgoff; local 103 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); 105 vma->vm_file, pgoff, vma_policy(vma),
|
H A D | memory.c | 2003 vmf.pgoff = page->index; 2705 pgoff_t pgoff, unsigned int flags, struct page **page) 2711 vmf.pgoff = pgoff; 2839 pte_t *pte, pgoff_t pgoff, unsigned int flags) 2852 pgoff -= off; 2856 * or fault_around_pages() from pgoff, depending what is nearest. 2858 max_pgoff = pgoff - ((start_addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) + 2861 pgoff + nr_pages - 1); 2865 if (++pgoff > max_pgof 2704 __do_fault(struct vm_area_struct *vma, unsigned long address, pgoff_t pgoff, unsigned int flags, struct page **page) argument 2838 do_fault_around(struct vm_area_struct *vma, unsigned long address, pte_t *pte, pgoff_t pgoff, unsigned int flags) argument 2881 do_read_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pmd_t *pmd, pgoff_t pgoff, unsigned int flags, pte_t orig_pte) argument 2922 do_cow_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pmd_t *pmd, pgoff_t pgoff, unsigned int flags, pte_t orig_pte) argument 2971 do_shared_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pmd_t *pmd, pgoff_t pgoff, unsigned int flags, pte_t orig_pte) argument 3039 pgoff_t pgoff = (((address & PAGE_MASK) local 3067 pgoff_t pgoff; local [all...] |
H A D | mprotect.c | 262 pgoff_t pgoff; local 290 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); 292 vma->anon_vma, vma->vm_file, pgoff, vma_policy(vma),
|
H A D | rmap.c | 522 pgoff_t pgoff = page_to_pgoff(page); local 523 return vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT); 1638 pgoff_t pgoff = page_to_pgoff(page); local 1646 anon_vma_interval_tree_foreach(avc, &anon_vma->rb_root, pgoff, pgoff) { 1679 pgoff_t pgoff = page_to_pgoff(page); local 1694 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) {
|
H A D | util.c | 256 unsigned long flag, unsigned long pgoff) 265 ret = do_mmap_pgoff(file, addr, len, prot, flag, pgoff, 254 vm_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flag, unsigned long pgoff) argument
|
H A D | mlock.c | 558 pgoff_t pgoff; local 567 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); 569 vma->vm_file, pgoff, vma_policy(vma),
|
H A D | mempolicy.c | 749 pgoff_t pgoff; local 769 pgoff = vma->vm_pgoff + 772 vma->anon_vma, vma->vm_file, pgoff, 2279 unsigned long pgoff; local 2296 pgoff = vma->vm_pgoff; 2297 pgoff += (addr - vma->vm_start) >> PAGE_SHIFT; 2298 polnid = offset_il_node(pol, vma, pgoff);
|
H A D | filemap.c | 1877 pgoff_t offset = vmf->pgoff; 2004 radix_tree_for_each_slot(slot, &mapping->page_tree, &iter, vmf->pgoff) { 2041 pte = vmf->pte + page->index - vmf->pgoff; 2047 addr = address + (page->index - vmf->pgoff) * PAGE_SIZE;
|
H A D | huge_memory.c | 1857 pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); local 1864 anon_vma_interval_tree_foreach(avc, &anon_vma->rb_root, pgoff, pgoff) { 1889 anon_vma_interval_tree_foreach(avc, &anon_vma->rb_root, pgoff, pgoff) {
|
H A D | migrate.c | 198 /* hugetlbfs does not support remap_pages, so no huge pgoff worries */ 199 pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); local 205 addr = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
|
H A D | memcontrol.c | 5725 pgoff_t pgoff; local 5734 pgoff = linear_page_index(vma, addr); 5736 pgoff = pte_to_pgoff(ptent); 5742 page = find_get_entry(mapping, pgoff); 5750 page = find_get_page(mapping, pgoff); 5752 page = find_get_page(mapping, pgoff);
|
H A D | vmalloc.c | 2168 * @pgoff: number of pages into addr before first page to map 2179 unsigned long pgoff) 2182 addr + (pgoff << PAGE_SHIFT), 2178 remap_vmalloc_range(struct vm_area_struct *vma, void *addr, unsigned long pgoff) argument
|
H A D | hugetlb.c | 2759 pgoff_t pgoff; local 2766 pgoff = ((address - vma->vm_start) >> PAGE_SHIFT) + 2776 vma_interval_tree_foreach(iter_vma, &mapping->i_mmap, pgoff, pgoff) {
|
H A D | shmem.c | 1308 vmf->pgoff >= shmem_falloc->start && 1309 vmf->pgoff < shmem_falloc->next) { 1342 error = shmem_getpage(inode, vmf->pgoff, &vmf->page, SGP_CACHE, &ret);
|