/mm/ |
H A D | mprotect.c | 149 unsigned long oldflags = vma->vm_flags; 204 * vm_flags and vm_page_prot are protected by the mmap_sem 207 vma->vm_flags = newflags; 235 unsigned long vm_flags, nstart, end, tmp, reqprot; local 261 vm_flags = calc_vm_prot_bits(prot); 274 if (!(vma->vm_flags & VM_GROWSDOWN)) 283 if (!(vma->vm_flags & VM_GROWSUP)) 295 newflags = vm_flags | (vma->vm_flags & ~(VM_READ | VM_WRITE | VM_EXEC));
|
H A D | mremap.c | 171 unsigned long vm_flags = vma->vm_flags; local 190 * pages recently unmapped. But leave vma->vm_flags as it was, 194 MADV_UNMERGEABLE, &vm_flags); 218 if (vm_flags & VM_ACCOUNT) { 219 vma->vm_flags &= ~VM_ACCOUNT; 237 vm_stat_account(mm, vma->vm_flags, vma->vm_file, new_len>>PAGE_SHIFT); 248 vma->vm_flags |= VM_ACCOUNT; 250 vma->vm_next->vm_flags |= VM_ACCOUNT; 253 if (vm_flags [all...] |
H A D | rmap.c | 556 } else if (page->mapping && !(vma->vm_flags & VM_NONLINEAR)) { 653 unsigned long *vm_flags) 673 if (vma->vm_flags & VM_LOCKED) { 676 *vm_flags |= VM_LOCKED; 696 if (vma->vm_flags & VM_LOCKED) { 699 *vm_flags |= VM_LOCKED; 726 *vm_flags |= vma->vm_flags; 733 unsigned long *vm_flags) 758 &mapcount, vm_flags); 651 page_referenced_one(struct page *page, struct vm_area_struct *vma, unsigned long address, unsigned int *mapcount, unsigned long *vm_flags) argument 731 page_referenced_anon(struct page *page, struct mem_cgroup *mem_cont, unsigned long *vm_flags) argument 780 page_referenced_file(struct page *page, struct mem_cgroup *mem_cont, unsigned long *vm_flags) argument 845 page_referenced(struct page *page, int is_locked, struct mem_cgroup *mem_cont, unsigned long *vm_flags) argument [all...] |
H A D | huge_memory.c | 599 if (likely(vma->vm_flags & VM_WRITE)) 1431 unsigned long *vm_flags, int advice) 1438 if (*vm_flags & (VM_HUGEPAGE | VM_NO_THP)) 1440 *vm_flags &= ~VM_NOHUGEPAGE; 1441 *vm_flags |= VM_HUGEPAGE; 1454 if (*vm_flags & (VM_NOHUGEPAGE | VM_NO_THP)) 1456 *vm_flags &= ~VM_HUGEPAGE; 1457 *vm_flags |= VM_NOHUGEPAGE; 1595 VM_BUG_ON(is_linear_pfn_mapping(vma) || vma->vm_flags & VM_NO_THP); 1843 if ((!(vma->vm_flags 1430 hugepage_madvise(struct vm_area_struct *vma, unsigned long *vm_flags, int advice) argument [all...] |
H A D | ksm.c | 395 if (!(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma) 429 if (!(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma) 632 if (!(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma) 862 if (!(vma->vm_flags & VM_MERGEABLE)) 899 if ((vma->vm_flags & VM_LOCKED) && kpage && !err) { 1324 if (!(vma->vm_flags & VM_MERGEABLE)) 1456 unsigned long end, int advice, unsigned long *vm_flags) 1466 if (*vm_flags & (VM_MERGEABLE | VM_SHARED | VM_MAYSHARE | 1478 *vm_flags |= VM_MERGEABLE; 1482 if (!(*vm_flags 1455 ksm_madvise(struct vm_area_struct *vma, unsigned long start, unsigned long end, int advice, unsigned long *vm_flags) argument 1590 page_referenced_ksm(struct page *page, struct mem_cgroup *memcg, unsigned long *vm_flags) argument [all...] |
H A D | memory.c | 719 "addr:%p vm_flags:%08lx anon_vma:%p mapping:%p index:%lx\n", 720 (void *)addr, vma->vm_flags, vma->anon_vma, mapping, index); 808 if (vma->vm_flags & (VM_PFNMAP | VM_MIXEDMAP)) 817 if (unlikely(vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP))) { 818 if (vma->vm_flags & VM_MIXEDMAP) { 827 if (!is_cow_mapping(vma->vm_flags)) 859 unsigned long vm_flags = vma->vm_flags; local 882 is_cow_mapping(vm_flags)) { 899 if (is_cow_mapping(vm_flags)) { 1649 unsigned long vm_flags; local [all...] |
H A D | mmap.c | 79 pgprot_t vm_get_page_prot(unsigned long vm_flags) argument 81 return __pgprot(pgprot_val(protection_map[vm_flags & 83 pgprot_val(arch_vm_get_page_prot(vm_flags))); 202 if (vma->vm_flags & VM_DENYWRITE) 204 if (vma->vm_flags & VM_SHARED) 208 if (unlikely(vma->vm_flags & VM_NONLINEAR)) 243 if (vma->vm_flags & VM_EXECUTABLE) 416 if (vma->vm_flags & VM_DENYWRITE) 418 if (vma->vm_flags & VM_SHARED) 422 if (unlikely(vma->vm_flags 658 is_mergeable_vma(struct vm_area_struct *vma, struct file *file, unsigned long vm_flags) argument 697 can_vma_merge_before(struct vm_area_struct *vma, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t vm_pgoff) argument 716 can_vma_merge_after(struct vm_area_struct *vma, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t vm_pgoff) argument 758 vma_merge(struct mm_struct *mm, struct vm_area_struct *prev, unsigned long addr, unsigned long end, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t pgoff, struct mempolicy *policy) argument 956 vm_flags_t vm_flags; local 1161 vm_flags_t vm_flags = vma->vm_flags; local 1189 accountable_mapping(struct file *file, vm_flags_t vm_flags) argument 1201 mmap_region(struct file *file, unsigned long addr, unsigned long len, unsigned long flags, vm_flags_t vm_flags, unsigned long pgoff) argument 2449 install_special_mapping(struct mm_struct *mm, unsigned long addr, unsigned long len, unsigned long vm_flags, struct page **pages) argument [all...] |
H A D | nommu.c | 134 unsigned long vm_flags; local 140 vm_flags = (foll_flags & FOLL_WRITE) ? 142 vm_flags &= (foll_flags & FOLL_FORCE) ? 151 if ((vma->vm_flags & (VM_IO | VM_PFNMAP)) || 152 !(vm_flags & vma->vm_flags)) 207 if (!(vma->vm_flags & (VM_IO | VM_PFNMAP))) 246 vma->vm_flags |= VM_USERMAP; 640 if (region->vm_flags & VM_MAPPED_COPY) { 694 protect_vma(vma, vma->vm_flags); 1069 unsigned long vm_flags; local 1247 unsigned long capabilities, vm_flags, result; local [all...] |
H A D | hugetlb.c | 330 if (!(vma->vm_flags & VM_MAYSHARE)) 339 VM_BUG_ON(vma->vm_flags & VM_MAYSHARE); 348 VM_BUG_ON(vma->vm_flags & VM_MAYSHARE); 364 if (vma->vm_flags & VM_NORESERVE) 367 if (vma->vm_flags & VM_MAYSHARE) { 383 if (!(vma->vm_flags & VM_MAYSHARE)) 390 if (vma->vm_flags & VM_MAYSHARE) 981 if (vma->vm_flags & VM_MAYSHARE) { 1006 if (vma->vm_flags & VM_MAYSHARE) { 2144 cow = (vma->vm_flags 2846 hugetlb_reserve_pages(struct inode *inode, long from, long to, struct vm_area_struct *vma, vm_flags_t vm_flags) argument [all...] |
H A D | vmscan.c | 651 unsigned long vm_flags; local 653 referenced_ptes = page_referenced(page, 1, sc->mem_cgroup, &vm_flags); 664 if (vm_flags & VM_LOCKED) 1538 unsigned long vm_flags; local 1585 if (page_referenced(page, 0, sc->mem_cgroup, &vm_flags)) { 1596 if ((vm_flags & VM_EXEC) && page_is_file_cache(page)) {
|