/mm/ |
H A D | interval_tree.c | 28 /* Insert node immediately after prev in the interval tree */ 30 struct vm_area_struct *prev, 37 VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node); 39 if (!prev->shared.linear.rb.rb_right) { 40 parent = prev; 41 link = &prev->shared.linear.rb.rb_right; 43 parent = rb_entry(prev->shared.linear.rb.rb_right, 29 vma_interval_tree_insert_after(struct vm_area_struct *node, struct vm_area_struct *prev, struct rb_root *root) argument
|
H A D | madvise.c | 46 struct vm_area_struct **prev, 99 *prev = vma; 104 *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma, 107 if (*prev) { 108 vma = *prev; 112 *prev = vma; 220 struct vm_area_struct **prev, 227 *prev = vma; 245 *prev 45 madvise_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) argument 219 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) argument 274 madvise_dontneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) argument 297 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) argument 377 madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) argument 466 struct vm_area_struct *vma, *prev; local [all...] |
H A D | mprotect.c | 339 struct vm_area_struct *vma, *prev; local 372 prev = vma->vm_prev; 391 prev = vma; 414 error = mprotect_fixup(vma, &prev, nstart, tmp, newflags); 419 if (nstart < prev->vm_end) 420 nstart = prev->vm_end; 424 vma = prev->vm_next;
|
H A D | util.c | 142 struct vm_area_struct *prev, struct rb_node *rb_parent) 146 vma->vm_prev = prev; 147 if (prev) { 148 next = prev->vm_next; 149 prev->vm_next = vma; 141 __vma_link_list(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node *rb_parent) argument
|
H A D | mlock.c | 554 static int mlock_fixup(struct vm_area_struct *vma, struct vm_area_struct **prev, argument 568 *prev = vma_merge(mm, *prev, start, end, newflags, vma->anon_vma, 571 if (*prev) { 572 vma = *prev; 609 *prev = vma; 616 struct vm_area_struct * vma, * prev; local 630 prev = vma->vm_prev; 632 prev = vma; 646 error = mlock_fixup(vma, &prev, nstar 775 struct vm_area_struct * vma, * prev = NULL; local [all...] |
H A D | slob.c | 219 slob_t *prev, *cur, *aligned = NULL; local 222 for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) { 236 prev = cur; 243 if (prev) 244 set_slob(prev, slob_units(prev), next); 248 if (prev) 249 set_slob(prev, slob_units(prev), cu 271 struct list_head *prev; local 343 slob_t *prev, *next, *b = (slob_t *)block; local [all...] |
H A D | memory.c | 2598 struct vm_area_struct *prev = vma->vm_prev; local 2606 if (prev && prev->vm_end == address) 2607 return prev->vm_flags & VM_GROWSDOWN ? 0 : -ENOMEM;
|
H A D | mempolicy.c | 662 struct vm_area_struct *vma, *prev; local 667 prev = NULL; 679 if (prev && prev->vm_end < vma->vm_start) 700 prev = vma; 746 struct vm_area_struct *prev; local 757 prev = vma->vm_prev; 759 prev = vma; 761 for (; vma && vma->vm_start < end; prev = vma, vma = next) { 771 prev 2197 struct rb_node *prev = rb_prev(n); local [all...] |
H A D | mmap.c | 61 struct vm_area_struct *vma, struct vm_area_struct *prev, 384 unsigned long prev = 0, pend = 0; local 389 if (vma->vm_start < prev) { 390 pr_emerg("vm_start %lx < prev %lx\n", 391 vma->vm_start, prev); 412 prev = vma->vm_start; 662 struct vm_area_struct *prev, struct rb_node **rb_link, 665 __vma_link_list(mm, vma, prev, rb_parent); 670 struct vm_area_struct *prev, struct rb_node **rb_link, 680 __vma_link(mm, vma, prev, rb_lin 661 __vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) argument 669 vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) argument 696 struct vm_area_struct *prev; local 707 __vma_unlink(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev) argument 1042 vma_merge(struct mm_struct *mm, struct vm_area_struct *prev, unsigned long addr, unsigned long end, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t pgoff, struct mempolicy *policy, const char __user *anon_name) argument 1545 struct vm_area_struct *vma, *prev; local 1778 struct rb_node *prev = &vma->vm_rb; local 1881 struct rb_node *prev = &vma->vm_rb; local 2323 struct vm_area_struct *vma, *prev; local 2338 struct vm_area_struct *prev; local 2401 unmap_region(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long start, unsigned long end) argument 2422 detach_vmas_to_be_unmapped(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long end) argument 2537 struct vm_area_struct *vma, *prev, *last; local 2653 struct vm_area_struct *vma, *prev; local 2803 struct vm_area_struct *prev; local 2844 struct vm_area_struct *new_vma, *prev; local [all...] |
H A D | nommu.c | 708 struct vm_area_struct *pvma, *prev; local 764 prev = NULL; 766 prev = rb_entry(rb_prev, struct vm_area_struct, vm_rb); 768 __vma_link_list(mm, vma, prev, parent);
|
H A D | percpu.c | 335 int prev = chunk->map[i - 1]; local 337 if (!(prev & 1) && prev <= round_down(off, PAGE_SIZE))
|
H A D | swapfile.c | 1283 unsigned int prev, bool frontswap) 1286 unsigned int i = prev; 1297 if (!prev) { 1305 max = prev + 1; 1306 prev = 0; 1662 lh = sis->first_swap_extent.list.prev; /* Highest extent */ 2882 page = list_entry(page->lru.prev, struct page, lru); 2887 page = list_entry(page->lru.prev, struct page, lru); 2907 page = list_entry(page->lru.prev, struct page, lru); 2913 page = list_entry(page->lru.prev, struc 1282 find_next_to_unuse(struct swap_info_struct *si, unsigned int prev, bool frontswap) argument [all...] |
H A D | vmalloc.c | 331 struct vmap_area *prev; local 332 prev = rb_entry(tmp, struct vmap_area, rb_node); 333 list_add_rcu(&va->list, &prev->list); 2260 * pvm_find_next_prev - find the next and prev vmap_area surrounding @end 2265 * Returns: %true if either or both of next and prev are found, 2367 struct vmap_area **vas, *prev, *next; local 2424 if (!pvm_find_next_prev(vmap_area_pcpu_hole, &next, &prev)) { 2428 base = pvm_determine_end(&next, &prev, align) - end; 2432 BUG_ON(prev && prev [all...] |
H A D | slab.c | 2361 p = n->slabs_free.prev; 3664 struct array_cache __percpu *cpu_cache, *prev; local 3671 prev = cachep->cpu_cache; 3680 if (!prev) 3687 struct array_cache *ac = per_cpu_ptr(prev, cpu); 3696 free_percpu(prev);
|
H A D | memcontrol.c | 1184 * @prev: previously returned memcg, NULL on first invocation 1190 * Caller must pass the return value in @prev on subsequent 1199 struct mem_cgroup *prev, 1211 if (prev && !reclaim) 1212 last_visited = prev; 1215 if (prev) 1230 if (prev && reclaim->generation != iter->generation) { 1246 else if (!prev && memcg) 1250 if (prev && !memcg) 1256 if (prev 1198 mem_cgroup_iter(struct mem_cgroup *root, struct mem_cgroup *prev, struct mem_cgroup_reclaim_cookie *reclaim) argument 1267 mem_cgroup_iter_break(struct mem_cgroup *root, struct mem_cgroup *prev) argument [all...] |