Searched refs:pgoff (Results 1 - 21 of 21) sorted by relevance

/mm/
H A Dmincore.c62 static unsigned char mincore_page(struct address_space *mapping, pgoff_t pgoff) argument
75 page = find_get_entry(mapping, pgoff);
85 page = find_get_page(mapping, pgoff);
87 page = find_get_page(mapping, pgoff);
105 pgoff_t pgoff; local
107 pgoff = linear_page_index(vma, addr);
108 for (i = 0; i < nr; i++, pgoff++)
109 vec[i] = mincore_page(vma->vm_file->f_mapping, pgoff);
127 pgoff_t pgoff; local
135 pgoff
[all...]
H A Dfremap.c73 unsigned long addr, unsigned long pgoff, pgprot_t prot)
83 ptfile = pgoff_to_pte(pgoff);
103 unsigned long size, pgoff_t pgoff)
109 err = install_file_pte(mm, vma, addr, pgoff, vma->vm_page_prot);
115 pgoff++;
127 * @pgoff: to-be-mapped page of the backing store file
143 unsigned long, prot, unsigned long, pgoff, unsigned long, flags)
168 /* Does pgoff wrap? */
169 if (pgoff + (size >> PAGE_SHIFT) < pgoff)
72 install_file_pte(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, unsigned long pgoff, pgprot_t prot) argument
102 generic_file_remap_pages(struct vm_area_struct *vma, unsigned long addr, unsigned long size, pgoff_t pgoff) argument
[all...]
H A Dfilemap_xip.c162 * __xip_sparse_page when found at pgoff.
166 unsigned long pgoff)
186 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) {
189 ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
236 if (vmf->pgoff >= size)
239 error = mapping->a_ops->get_xip_mem(mapping, vmf->pgoff, 0,
254 error = mapping->a_ops->get_xip_mem(mapping, vmf->pgoff, 1,
259 /* unmap sparse mappings at pgoff from all other vmas */
260 __xip_unmap(mapping, vmf->pgoff);
165 __xip_unmap(struct address_space * mapping, unsigned long pgoff) argument
[all...]
H A Dmmap.c729 unsigned long end, pgoff_t pgoff, struct vm_area_struct *insert)
840 vma->vm_pgoff = pgoff;
1014 * Given a mapping request (addr,end,vm_flags,file,pgoff,anon_name),
1046 pgoff_t pgoff, struct mempolicy *policy,
1074 file, pgoff, anon_name)) {
1081 file, pgoff+pglen, anon_name) &&
1102 file, pgoff+pglen, anon_name)) {
1268 unsigned long flags, unsigned long pgoff,
1298 if ((pgoff + (len >> PAGE_SHIFT)) < pgoff)
728 vma_adjust(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgoff_t pgoff, struct vm_area_struct *insert) argument
1042 vma_merge(struct mm_struct *mm, struct vm_area_struct *prev, unsigned long addr, unsigned long end, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t pgoff, struct mempolicy *policy, const char __user *anon_name) argument
1266 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *populate) argument
1541 mmap_region(struct file *file, unsigned long addr, unsigned long len, vm_flags_t vm_flags, unsigned long pgoff) argument
1922 arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument
1958 arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, const unsigned long len, const unsigned long pgoff, const unsigned long flags) argument
2009 get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument
2656 pgoff_t pgoff = addr >> PAGE_SHIFT; local
2837 copy_vma(struct vm_area_struct **vmap, unsigned long addr, unsigned long len, pgoff_t pgoff, bool *need_rmap_locks) argument
2958 pgoff_t pgoff; local
[all...]
H A Dnommu.c918 unsigned long pgoff,
945 if ((pgoff + (rlen >> PAGE_SHIFT)) < pgoff)
1266 unsigned long pgoff,
1275 kenter(",%lx,%lx,%lx,%lx,%lx", addr, len, prot, flags, pgoff);
1281 ret = validate_mmap_request(file, addr, len, prot, flags, pgoff,
1307 region->vm_pgoff = pgoff;
1311 vma->vm_pgoff = pgoff;
1333 pgend = pgoff + pglen;
1352 if (pgoff >
913 validate_mmap_request(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *_capabilities) argument
1261 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *populate) argument
1862 remap_vmalloc_range(struct vm_area_struct *vma, void *addr, unsigned long pgoff) argument
1877 arch_get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument
1997 generic_file_remap_pages(struct vm_area_struct *vma, unsigned long addr, unsigned long size, pgoff_t pgoff) argument
[all...]
H A Dmremap.c354 unsigned long pgoff; local
358 pgoff = (addr - vma->vm_start) >> PAGE_SHIFT;
359 pgoff += vma->vm_pgoff;
360 if (pgoff + (new_len >> PAGE_SHIFT) < pgoff)
H A Dmemory-failure.c432 pgoff_t pgoff; local
438 pgoff = page_to_pgoff(page);
447 pgoff, pgoff) {
472 pgoff_t pgoff = page_to_pgoff(page); local
477 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff,
478 pgoff) {
H A Dmadvise.c51 pgoff_t pgoff; local
103 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT);
105 vma->vm_file, pgoff, vma_policy(vma),
H A Dmemory.c2003 vmf.pgoff = page->index;
2705 pgoff_t pgoff, unsigned int flags, struct page **page)
2711 vmf.pgoff = pgoff;
2839 pte_t *pte, pgoff_t pgoff, unsigned int flags)
2852 pgoff -= off;
2856 * or fault_around_pages() from pgoff, depending what is nearest.
2858 max_pgoff = pgoff - ((start_addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) +
2861 pgoff + nr_pages - 1);
2865 if (++pgoff > max_pgof
2704 __do_fault(struct vm_area_struct *vma, unsigned long address, pgoff_t pgoff, unsigned int flags, struct page **page) argument
2838 do_fault_around(struct vm_area_struct *vma, unsigned long address, pte_t *pte, pgoff_t pgoff, unsigned int flags) argument
2881 do_read_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pmd_t *pmd, pgoff_t pgoff, unsigned int flags, pte_t orig_pte) argument
2922 do_cow_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pmd_t *pmd, pgoff_t pgoff, unsigned int flags, pte_t orig_pte) argument
2971 do_shared_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pmd_t *pmd, pgoff_t pgoff, unsigned int flags, pte_t orig_pte) argument
3039 pgoff_t pgoff = (((address & PAGE_MASK) local
3067 pgoff_t pgoff; local
[all...]
H A Dmprotect.c262 pgoff_t pgoff; local
290 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT);
292 vma->anon_vma, vma->vm_file, pgoff, vma_policy(vma),
H A Drmap.c522 pgoff_t pgoff = page_to_pgoff(page); local
523 return vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
1638 pgoff_t pgoff = page_to_pgoff(page); local
1646 anon_vma_interval_tree_foreach(avc, &anon_vma->rb_root, pgoff, pgoff) {
1679 pgoff_t pgoff = page_to_pgoff(page); local
1694 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) {
H A Dutil.c256 unsigned long flag, unsigned long pgoff)
265 ret = do_mmap_pgoff(file, addr, len, prot, flag, pgoff,
254 vm_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flag, unsigned long pgoff) argument
H A Dmlock.c558 pgoff_t pgoff; local
567 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT);
569 vma->vm_file, pgoff, vma_policy(vma),
H A Dmempolicy.c749 pgoff_t pgoff; local
769 pgoff = vma->vm_pgoff +
772 vma->anon_vma, vma->vm_file, pgoff,
2279 unsigned long pgoff; local
2296 pgoff = vma->vm_pgoff;
2297 pgoff += (addr - vma->vm_start) >> PAGE_SHIFT;
2298 polnid = offset_il_node(pol, vma, pgoff);
H A Dfilemap.c1877 pgoff_t offset = vmf->pgoff;
2004 radix_tree_for_each_slot(slot, &mapping->page_tree, &iter, vmf->pgoff) {
2041 pte = vmf->pte + page->index - vmf->pgoff;
2047 addr = address + (page->index - vmf->pgoff) * PAGE_SIZE;
H A Dhuge_memory.c1857 pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); local
1864 anon_vma_interval_tree_foreach(avc, &anon_vma->rb_root, pgoff, pgoff) {
1889 anon_vma_interval_tree_foreach(avc, &anon_vma->rb_root, pgoff, pgoff) {
H A Dmigrate.c198 /* hugetlbfs does not support remap_pages, so no huge pgoff worries */
199 pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); local
205 addr = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
H A Dmemcontrol.c5725 pgoff_t pgoff; local
5734 pgoff = linear_page_index(vma, addr);
5736 pgoff = pte_to_pgoff(ptent);
5742 page = find_get_entry(mapping, pgoff);
5750 page = find_get_page(mapping, pgoff);
5752 page = find_get_page(mapping, pgoff);
H A Dvmalloc.c2168 * @pgoff: number of pages into addr before first page to map
2179 unsigned long pgoff)
2182 addr + (pgoff << PAGE_SHIFT),
2178 remap_vmalloc_range(struct vm_area_struct *vma, void *addr, unsigned long pgoff) argument
H A Dhugetlb.c2759 pgoff_t pgoff; local
2766 pgoff = ((address - vma->vm_start) >> PAGE_SHIFT) +
2776 vma_interval_tree_foreach(iter_vma, &mapping->i_mmap, pgoff, pgoff) {
H A Dshmem.c1308 vmf->pgoff >= shmem_falloc->start &&
1309 vmf->pgoff < shmem_falloc->next) {
1342 error = shmem_getpage(inode, vmf->pgoff, &vmf->page, SGP_CACHE, &ret);

Completed in 3072 milliseconds