Searched refs:tsk (Results 1 - 8 of 8) sorted by relevance

/mm/
H A Dmmu_context.c26 struct task_struct *tsk = current; local
28 task_lock(tsk);
29 active_mm = tsk->active_mm;
32 tsk->active_mm = mm;
34 tsk->mm = mm;
35 switch_mm(active_mm, mm, tsk);
36 task_unlock(tsk);
53 struct task_struct *tsk = current; local
55 task_lock(tsk);
56 sync_mm_rss(tsk, m
[all...]
H A Dmemory-failure.c279 struct task_struct *tsk; member in struct:to_kill
294 static void add_to_kill(struct task_struct *tsk, struct page *p, argument
323 page_to_pfn(p), tsk->comm);
326 get_task_struct(tsk);
327 tk->tsk = tsk;
354 pfn, tk->tsk->comm, tk->tsk->pid);
355 force_sig(SIGKILL, tk->tsk);
364 else if (kill_proc_ao(tk->tsk, t
375 task_early_kill(struct task_struct *tsk) argument
391 struct task_struct *tsk; local
423 struct task_struct *tsk; local
[all...]
H A Doom_kill.c71 * @tsk: task struct of which task to consider
74 * Task eligibility is determined by whether or not a candidate task, @tsk,
78 static bool has_intersects_mems_allowed(struct task_struct *tsk, argument
81 struct task_struct *start = tsk;
86 * If this is a mempolicy constrained oom, tsk's
91 if (mempolicy_nodemask_intersects(tsk, mask))
96 * check the mems of tsk's cpuset.
98 if (cpuset_mems_allowed_intersects(current, tsk))
101 } while_each_thread(start, tsk);
106 static bool has_intersects_mems_allowed(struct task_struct *tsk, argument
[all...]
H A Dpage-writeback.c236 void task_dirty_inc(struct task_struct *tsk) argument
238 prop_inc_single(&vm_dirties, &tsk->dirties);
256 static inline void task_dirties_fraction(struct task_struct *tsk, argument
259 prop_fraction_single(&vm_dirties, &tsk->dirties,
277 static unsigned long task_dirty_limit(struct task_struct *tsk, argument
284 task_dirties_fraction(tsk, &numerator, &denominator);
414 struct task_struct *tsk; local
431 tsk = current;
432 if (tsk->flags & PF_LESS_THROTTLE || rt_task(tsk)) {
[all...]
H A Dnommu.c128 int __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, argument
178 int get_user_pages(struct task_struct *tsk, struct mm_struct *mm, argument
189 return __get_user_pages(tsk, mm, start, nr_pages, flags, pages, vmas,
1973 static int __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, argument
2025 int access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument
2032 mm = get_task_mm(tsk);
2036 len = __access_remote_vm(tsk, mm, addr, buf, len, write);
H A Dmemory.c1596 * @tsk: task_struct of target task
1643 int __get_user_pages(struct task_struct *tsk, struct mm_struct *mm, argument
1776 if (tsk) {
1778 tsk->maj_flt++;
1780 tsk->min_flt++;
1829 * @tsk: the task_struct to use for page fault accounting, or
1854 int fixup_user_fault(struct task_struct *tsk, struct mm_struct *mm, argument
1874 if (tsk) {
1876 tsk->maj_flt++;
1878 tsk
1934 get_user_pages(struct task_struct *tsk, struct mm_struct *mm, unsigned long start, int nr_pages, int write, int force, struct page **pages, struct vm_area_struct **vmas) argument
3767 __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument
3846 access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument
[all...]
H A Dmempolicy.c419 void mpol_rebind_task(struct task_struct *tsk, const nodemask_t *new, argument
422 mpol_rebind_policy(tsk->mempolicy, new, step);
1737 * If tsk's mempolicy is "default" [NULL], return 'true' to indicate default
1742 * Takes task_lock(tsk) to prevent freeing of its mempolicy.
1744 bool mempolicy_nodemask_intersects(struct task_struct *tsk, argument
1752 task_lock(tsk);
1753 mempolicy = tsk->mempolicy;
1762 * Thus, it's possible for tsk to have allocated memory from
1774 task_unlock(tsk);
H A Dvmscan.c2727 struct task_struct *tsk = current; local
2737 set_cpus_allowed_ptr(tsk, cpumask);
2752 tsk->flags |= PF_MEMALLOC | PF_SWAPWRITE | PF_KSWAPD;

Completed in 2573 milliseconds