Searched defs:prev (Results 1 - 15 of 15) sorted by relevance

/mm/
H A Dinterval_tree.c28 /* Insert node immediately after prev in the interval tree */
30 struct vm_area_struct *prev,
37 VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node);
39 if (!prev->shared.linear.rb.rb_right) {
40 parent = prev;
41 link = &prev->shared.linear.rb.rb_right;
43 parent = rb_entry(prev->shared.linear.rb.rb_right,
29 vma_interval_tree_insert_after(struct vm_area_struct *node, struct vm_area_struct *prev, struct rb_root *root) argument
H A Dmadvise.c46 struct vm_area_struct **prev,
99 *prev = vma;
104 *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma,
107 if (*prev) {
108 vma = *prev;
112 *prev = vma;
220 struct vm_area_struct **prev,
227 *prev = vma;
245 *prev
45 madvise_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) argument
219 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) argument
274 madvise_dontneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) argument
297 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) argument
377 madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) argument
466 struct vm_area_struct *vma, *prev; local
[all...]
H A Dmprotect.c339 struct vm_area_struct *vma, *prev; local
372 prev = vma->vm_prev;
391 prev = vma;
414 error = mprotect_fixup(vma, &prev, nstart, tmp, newflags);
419 if (nstart < prev->vm_end)
420 nstart = prev->vm_end;
424 vma = prev->vm_next;
H A Dutil.c142 struct vm_area_struct *prev, struct rb_node *rb_parent)
146 vma->vm_prev = prev;
147 if (prev) {
148 next = prev->vm_next;
149 prev->vm_next = vma;
141 __vma_link_list(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node *rb_parent) argument
H A Dmlock.c554 static int mlock_fixup(struct vm_area_struct *vma, struct vm_area_struct **prev, argument
568 *prev = vma_merge(mm, *prev, start, end, newflags, vma->anon_vma,
571 if (*prev) {
572 vma = *prev;
609 *prev = vma;
616 struct vm_area_struct * vma, * prev; local
630 prev = vma->vm_prev;
632 prev = vma;
646 error = mlock_fixup(vma, &prev, nstar
775 struct vm_area_struct * vma, * prev = NULL; local
[all...]
H A Dslob.c219 slob_t *prev, *cur, *aligned = NULL; local
222 for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) {
236 prev = cur;
243 if (prev)
244 set_slob(prev, slob_units(prev), next);
248 if (prev)
249 set_slob(prev, slob_units(prev), cu
271 struct list_head *prev; local
343 slob_t *prev, *next, *b = (slob_t *)block; local
[all...]
H A Dmemory.c2598 struct vm_area_struct *prev = vma->vm_prev; local
2606 if (prev && prev->vm_end == address)
2607 return prev->vm_flags & VM_GROWSDOWN ? 0 : -ENOMEM;
H A Dmempolicy.c662 struct vm_area_struct *vma, *prev; local
667 prev = NULL;
679 if (prev && prev->vm_end < vma->vm_start)
700 prev = vma;
746 struct vm_area_struct *prev; local
757 prev = vma->vm_prev;
759 prev = vma;
761 for (; vma && vma->vm_start < end; prev = vma, vma = next) {
771 prev
2197 struct rb_node *prev = rb_prev(n); local
[all...]
H A Dmmap.c61 struct vm_area_struct *vma, struct vm_area_struct *prev,
384 unsigned long prev = 0, pend = 0; local
389 if (vma->vm_start < prev) {
390 pr_emerg("vm_start %lx < prev %lx\n",
391 vma->vm_start, prev);
412 prev = vma->vm_start;
662 struct vm_area_struct *prev, struct rb_node **rb_link,
665 __vma_link_list(mm, vma, prev, rb_parent);
670 struct vm_area_struct *prev, struct rb_node **rb_link,
680 __vma_link(mm, vma, prev, rb_lin
661 __vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) argument
669 vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) argument
696 struct vm_area_struct *prev; local
707 __vma_unlink(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev) argument
1042 vma_merge(struct mm_struct *mm, struct vm_area_struct *prev, unsigned long addr, unsigned long end, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t pgoff, struct mempolicy *policy, const char __user *anon_name) argument
1545 struct vm_area_struct *vma, *prev; local
1778 struct rb_node *prev = &vma->vm_rb; local
1881 struct rb_node *prev = &vma->vm_rb; local
2323 struct vm_area_struct *vma, *prev; local
2338 struct vm_area_struct *prev; local
2401 unmap_region(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long start, unsigned long end) argument
2422 detach_vmas_to_be_unmapped(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long end) argument
2537 struct vm_area_struct *vma, *prev, *last; local
2653 struct vm_area_struct *vma, *prev; local
2803 struct vm_area_struct *prev; local
2844 struct vm_area_struct *new_vma, *prev; local
[all...]
H A Dnommu.c708 struct vm_area_struct *pvma, *prev; local
764 prev = NULL;
766 prev = rb_entry(rb_prev, struct vm_area_struct, vm_rb);
768 __vma_link_list(mm, vma, prev, parent);
H A Dpercpu.c335 int prev = chunk->map[i - 1]; local
337 if (!(prev & 1) && prev <= round_down(off, PAGE_SIZE))
H A Dswapfile.c1283 unsigned int prev, bool frontswap)
1286 unsigned int i = prev;
1297 if (!prev) {
1305 max = prev + 1;
1306 prev = 0;
1662 lh = sis->first_swap_extent.list.prev; /* Highest extent */
2882 page = list_entry(page->lru.prev, struct page, lru);
2887 page = list_entry(page->lru.prev, struct page, lru);
2907 page = list_entry(page->lru.prev, struct page, lru);
2913 page = list_entry(page->lru.prev, struc
1282 find_next_to_unuse(struct swap_info_struct *si, unsigned int prev, bool frontswap) argument
[all...]
H A Dvmalloc.c331 struct vmap_area *prev; local
332 prev = rb_entry(tmp, struct vmap_area, rb_node);
333 list_add_rcu(&va->list, &prev->list);
2260 * pvm_find_next_prev - find the next and prev vmap_area surrounding @end
2265 * Returns: %true if either or both of next and prev are found,
2367 struct vmap_area **vas, *prev, *next; local
2424 if (!pvm_find_next_prev(vmap_area_pcpu_hole, &next, &prev)) {
2428 base = pvm_determine_end(&next, &prev, align) - end;
2432 BUG_ON(prev && prev
[all...]
H A Dslab.c2361 p = n->slabs_free.prev;
3664 struct array_cache __percpu *cpu_cache, *prev; local
3671 prev = cachep->cpu_cache;
3680 if (!prev)
3687 struct array_cache *ac = per_cpu_ptr(prev, cpu);
3696 free_percpu(prev);
H A Dmemcontrol.c1184 * @prev: previously returned memcg, NULL on first invocation
1190 * Caller must pass the return value in @prev on subsequent
1199 struct mem_cgroup *prev,
1211 if (prev && !reclaim)
1212 last_visited = prev;
1215 if (prev)
1230 if (prev && reclaim->generation != iter->generation) {
1246 else if (!prev && memcg)
1250 if (prev && !memcg)
1256 if (prev
1198 mem_cgroup_iter(struct mem_cgroup *root, struct mem_cgroup *prev, struct mem_cgroup_reclaim_cookie *reclaim) argument
1267 mem_cgroup_iter_break(struct mem_cgroup *root, struct mem_cgroup *prev) argument
[all...]

Completed in 1756 milliseconds