Searched refs:locked (Results 1 - 9 of 9) sorted by relevance

/mm/
H A Dmlock.c50 * the mmap_sem for read, and verify that the vma really is locked
117 * Page must be locked. This is a wrapper for try_to_munlock()
165 * page locked so that we can leave it on the unevictable lru list and not
590 * Keep track of amount of locked VM.
676 int locked = 0; local
688 if (!locked) {
689 locked = 1;
710 ret = __mlock_vma_pages_range(vma, nstart, nend, &locked);
722 if (locked)
729 unsigned long locked; local
848 unsigned long lock_limit, locked; local
[all...]
H A Dmremap.c239 unsigned long new_len, unsigned long new_addr, bool *locked)
330 *locked = true;
365 unsigned long locked, lock_limit; local
366 locked = mm->locked_vm << PAGE_SHIFT;
368 locked += new_len - old_len;
369 if (locked > lock_limit && !capable(CAP_IPC_LOCK))
396 unsigned long new_addr, unsigned long new_len, bool *locked)
446 ret = move_vma(vma, addr, old_len, new_len, new_addr, locked);
484 bool locked = false; local
510 &locked);
237 move_vma(struct vm_area_struct *vma, unsigned long old_addr, unsigned long old_len, unsigned long new_len, unsigned long new_addr, bool *locked) argument
395 mremap_to(unsigned long addr, unsigned long old_len, unsigned long new_addr, unsigned long new_len, bool *locked) argument
[all...]
H A Dcompaction.c257 * In either case if the lock was locked, it is dropped and not regained.
265 unsigned long flags, bool *locked, struct compact_control *cc)
267 if (*locked) {
269 *locked = false;
348 bool locked = false; local
365 &locked, cc))
384 if (!locked) {
393 locked = compact_trylock_irqsave(&cc->zone->lock,
395 if (!locked)
439 if (locked)
264 compact_unlock_should_abort(spinlock_t *lock, unsigned long flags, bool *locked, struct compact_control *cc) argument
584 bool locked = false; local
[all...]
H A Dfilemap_xip.c176 int locked = 0; local
207 if (locked) {
211 locked = 1;
H A Dpage-writeback.c2054 * The page must be locked by the caller and will be unlocked upon return.
2126 * Most callers have locked the page, which pins the address_space in memory.
2183 * page for some reason, it should redirty the locked page via
2241 * Usually, the page _is_ locked, or the caller is a user-space process which
2244 * In other cases, the page should be locked before running set_page_dirty().
2269 * unfortunate, but it only exists while the page is locked.
2311 * Pages are always locked coming in here, so we get
2332 bool locked; local
2335 memcg = mem_cgroup_begin_page_stat(page, &locked, &memcg_flags);
2360 mem_cgroup_end_page_stat(memcg, locked, memcg_flag
2369 bool locked; local
[all...]
H A Drmap.c387 * the best this function can do is return a locked anon_vma that might
600 * On success returns with pte mapped and locked.
954 * always holds the page locked, except if called from page_dup_rmap,
958 * are initially only visible via the pagetables, and the pte is locked
972 * The caller needs to hold the pte lock, and the page must be locked in
1024 * Page does not have to be locked.
1049 bool locked; local
1051 memcg = mem_cgroup_begin_page_stat(page, &locked, &flags);
1056 mem_cgroup_end_page_stat(memcg, locked, flags);
1063 bool locked; local
[all...]
H A Dmmap.c1248 unsigned long locked, lock_limit; local
1252 locked = len >> PAGE_SHIFT;
1253 locked += mm->locked_vm;
1256 if (locked > lock_limit && !capable(CAP_IPC_LOCK))
2121 unsigned long locked; local
2123 locked = mm->locked_vm + grow;
2126 if (locked > limit && !capable(CAP_IPC_LOCK))
H A Dmemcontrol.c2005 * this subtree of our hierarchy is already locked
2152 bool locked; local
2170 locked = mem_cgroup_oom_trylock(memcg);
2172 if (locked)
2175 if (locked && !memcg->oom_kill_disable) {
2186 if (locked) {
2204 * @locked: &memcg->move_lock slowpath was taken
2211 * memcg = mem_cgroup_begin_page_stat(page, &locked, &flags);
2214 * mem_cgroup_end_page_stat(memcg, locked, flags);
2217 * get away without acquiring the memcg->move_lock (@locked i
2226 mem_cgroup_begin_page_stat(struct page *page, bool *locked, unsigned long *flags) argument
2264 mem_cgroup_end_page_stat(struct mem_cgroup *memcg, bool locked, unsigned long flags) argument
[all...]
H A Dmemory.c2037 * but allow concurrent faults), with pte both mapped and locked.
2405 * but allow concurrent faults), and pte mapped but not yet locked.
2408 * We return with the mmap_sem locked or unlocked in the same cases
2420 int locked; local
2472 locked = lock_page_or_retry(page, mm, flags);
2475 if (!locked) {
2556 * parallel locked swapcache.
2625 * but allow concurrent faults), and pte mapped but not yet locked.
2821 * not ready to be mapped: not up-to-date, locked, etc.
3058 * but allow concurrent faults), and pte mapped but not yet locked
[all...]

Completed in 3571 milliseconds