/mm/ |
H A D | mmu_context.c | 26 struct task_struct *tsk = current; local 28 task_lock(tsk); 29 active_mm = tsk->active_mm; 32 tsk->active_mm = mm; 34 tsk->mm = mm; 35 switch_mm(active_mm, mm, tsk); 36 task_unlock(tsk); 53 struct task_struct *tsk = current; local 55 task_lock(tsk); 56 sync_mm_rss(tsk, m [all...] |
H A D | oom_kill.c | 71 * @tsk: task struct of which task to consider 74 * Task eligibility is determined by whether or not a candidate task, @tsk, 78 static bool has_intersects_mems_allowed(struct task_struct *tsk, argument 81 struct task_struct *start = tsk; 86 * If this is a mempolicy constrained oom, tsk's 91 if (mempolicy_nodemask_intersects(tsk, mask)) 96 * check the mems of tsk's cpuset. 98 if (cpuset_mems_allowed_intersects(current, tsk)) 101 } while_each_thread(start, tsk); 106 static bool has_intersects_mems_allowed(struct task_struct *tsk, argument [all...] |
H A D | memory-failure.c | 279 struct task_struct *tsk; member in struct:to_kill 294 static void add_to_kill(struct task_struct *tsk, struct page *p, argument 323 page_to_pfn(p), tsk->comm); 326 get_task_struct(tsk); 327 tk->tsk = tsk; 354 pfn, tk->tsk->comm, tk->tsk->pid); 355 force_sig(SIGKILL, tk->tsk); 364 else if (kill_proc_ao(tk->tsk, t 375 task_early_kill(struct task_struct *tsk) argument 391 struct task_struct *tsk; local 423 struct task_struct *tsk; local [all...] |
H A D | page-writeback.c | 236 void task_dirty_inc(struct task_struct *tsk) argument 238 prop_inc_single(&vm_dirties, &tsk->dirties); 256 static inline void task_dirties_fraction(struct task_struct *tsk, argument 259 prop_fraction_single(&vm_dirties, &tsk->dirties, 277 static unsigned long task_dirty_limit(struct task_struct *tsk, argument 284 task_dirties_fraction(tsk, &numerator, &denominator); 414 struct task_struct *tsk; local 431 tsk = current; 432 if (tsk->flags & PF_LESS_THROTTLE || rt_task(tsk)) { [all...] |
H A D | memory.c | 1596 * @tsk: task_struct of target task 1643 int __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, argument 1776 if (tsk) { 1778 tsk->maj_flt++; 1780 tsk->min_flt++; 1829 * @tsk: the task_struct to use for page fault accounting, or 1854 int fixup_user_fault(struct task_struct *tsk, struct mm_struct *mm, argument 1874 if (tsk) { 1876 tsk->maj_flt++; 1878 tsk 1934 get_user_pages(struct task_struct *tsk, struct mm_struct *mm, unsigned long start, int nr_pages, int write, int force, struct page **pages, struct vm_area_struct **vmas) argument 3767 __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 3846 access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument [all...] |
H A D | mempolicy.c | 419 void mpol_rebind_task(struct task_struct *tsk, const nodemask_t *new, argument 422 mpol_rebind_policy(tsk->mempolicy, new, step); 1737 * If tsk's mempolicy is "default" [NULL], return 'true' to indicate default 1742 * Takes task_lock(tsk) to prevent freeing of its mempolicy. 1744 bool mempolicy_nodemask_intersects(struct task_struct *tsk, argument 1752 task_lock(tsk); 1753 mempolicy = tsk->mempolicy; 1762 * Thus, it's possible for tsk to have allocated memory from 1774 task_unlock(tsk);
|
H A D | nommu.c | 128 int __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, argument 178 int get_user_pages(struct task_struct *tsk, struct mm_struct *mm, argument 189 return __get_user_pages(tsk, mm, start, nr_pages, flags, pages, vmas, 1973 static int __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, argument 2025 int access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument 2032 mm = get_task_mm(tsk); 2036 len = __access_remote_vm(tsk, mm, addr, buf, len, write);
|
H A D | vmscan.c | 2727 struct task_struct *tsk = current; local 2737 set_cpus_allowed_ptr(tsk, cpumask); 2752 tsk->flags |= PF_MEMALLOC | PF_SWAPWRITE | PF_KSWAPD;
|