/mm/ |
H A D | mlock.c | 50 * the mmap_sem for read, and verify that the vma really is locked 117 * Page must be locked. This is a wrapper for try_to_munlock() 165 * page locked so that we can leave it on the unevictable lru list and not 590 * Keep track of amount of locked VM. 676 int locked = 0; local 688 if (!locked) { 689 locked = 1; 710 ret = __mlock_vma_pages_range(vma, nstart, nend, &locked); 722 if (locked) 729 unsigned long locked; local 848 unsigned long lock_limit, locked; local [all...] |
H A D | mremap.c | 239 unsigned long new_len, unsigned long new_addr, bool *locked) 330 *locked = true; 365 unsigned long locked, lock_limit; local 366 locked = mm->locked_vm << PAGE_SHIFT; 368 locked += new_len - old_len; 369 if (locked > lock_limit && !capable(CAP_IPC_LOCK)) 396 unsigned long new_addr, unsigned long new_len, bool *locked) 446 ret = move_vma(vma, addr, old_len, new_len, new_addr, locked); 484 bool locked = false; local 510 &locked); 237 move_vma(struct vm_area_struct *vma, unsigned long old_addr, unsigned long old_len, unsigned long new_len, unsigned long new_addr, bool *locked) argument 395 mremap_to(unsigned long addr, unsigned long old_len, unsigned long new_addr, unsigned long new_len, bool *locked) argument [all...] |
H A D | compaction.c | 257 * In either case if the lock was locked, it is dropped and not regained. 265 unsigned long flags, bool *locked, struct compact_control *cc) 267 if (*locked) { 269 *locked = false; 348 bool locked = false; local 365 &locked, cc)) 384 if (!locked) { 393 locked = compact_trylock_irqsave(&cc->zone->lock, 395 if (!locked) 439 if (locked) 264 compact_unlock_should_abort(spinlock_t *lock, unsigned long flags, bool *locked, struct compact_control *cc) argument 584 bool locked = false; local [all...] |
H A D | filemap_xip.c | 176 int locked = 0; local 207 if (locked) { 211 locked = 1;
|
H A D | page-writeback.c | 2054 * The page must be locked by the caller and will be unlocked upon return. 2126 * Most callers have locked the page, which pins the address_space in memory. 2183 * page for some reason, it should redirty the locked page via 2241 * Usually, the page _is_ locked, or the caller is a user-space process which 2244 * In other cases, the page should be locked before running set_page_dirty(). 2269 * unfortunate, but it only exists while the page is locked. 2311 * Pages are always locked coming in here, so we get 2332 bool locked; local 2335 memcg = mem_cgroup_begin_page_stat(page, &locked, &memcg_flags); 2360 mem_cgroup_end_page_stat(memcg, locked, memcg_flag 2369 bool locked; local [all...] |
H A D | rmap.c | 387 * the best this function can do is return a locked anon_vma that might 600 * On success returns with pte mapped and locked. 954 * always holds the page locked, except if called from page_dup_rmap, 958 * are initially only visible via the pagetables, and the pte is locked 972 * The caller needs to hold the pte lock, and the page must be locked in 1024 * Page does not have to be locked. 1049 bool locked; local 1051 memcg = mem_cgroup_begin_page_stat(page, &locked, &flags); 1056 mem_cgroup_end_page_stat(memcg, locked, flags); 1063 bool locked; local [all...] |
H A D | mmap.c | 1248 unsigned long locked, lock_limit; local 1252 locked = len >> PAGE_SHIFT; 1253 locked += mm->locked_vm; 1256 if (locked > lock_limit && !capable(CAP_IPC_LOCK)) 2121 unsigned long locked; local 2123 locked = mm->locked_vm + grow; 2126 if (locked > limit && !capable(CAP_IPC_LOCK))
|
H A D | memcontrol.c | 2005 * this subtree of our hierarchy is already locked 2152 bool locked; local 2170 locked = mem_cgroup_oom_trylock(memcg); 2172 if (locked) 2175 if (locked && !memcg->oom_kill_disable) { 2186 if (locked) { 2204 * @locked: &memcg->move_lock slowpath was taken 2211 * memcg = mem_cgroup_begin_page_stat(page, &locked, &flags); 2214 * mem_cgroup_end_page_stat(memcg, locked, flags); 2217 * get away without acquiring the memcg->move_lock (@locked i 2226 mem_cgroup_begin_page_stat(struct page *page, bool *locked, unsigned long *flags) argument 2264 mem_cgroup_end_page_stat(struct mem_cgroup *memcg, bool locked, unsigned long flags) argument [all...] |
H A D | memory.c | 2037 * but allow concurrent faults), with pte both mapped and locked. 2405 * but allow concurrent faults), and pte mapped but not yet locked. 2408 * We return with the mmap_sem locked or unlocked in the same cases 2420 int locked; local 2472 locked = lock_page_or_retry(page, mm, flags); 2475 if (!locked) { 2556 * parallel locked swapcache. 2625 * but allow concurrent faults), and pte mapped but not yet locked. 2821 * not ready to be mapped: not up-to-date, locked, etc. 3058 * but allow concurrent faults), and pte mapped but not yet locked [all...] |