Searched refs:first (Results 1 - 13 of 13) sorted by path
/kernel/ |
H A D | kprobes.c | 29 * exceptions notifier to be first on the priority list. 1699 .priority = 0x7fffffff /* we need to be notified first */ 1796 ri = hlist_entry(rp->free_instances.first,
|
H A D | pid.c | 66 * first use and are never deallocated. This way a low pid_max 277 /* Handle a fork failure of the first process */ 433 struct hlist_node *first; local 434 first = rcu_dereference_check(hlist_first_rcu(&pid->tasks[type]), 436 if (first) 437 result = hlist_entry(first, struct task_struct, pids[(type)].node); 547 * Used by proc to find the first pid that is greater than or equal to nr.
|
H A D | resource.c | 338 * This walks through whole tree and not just first level children 385 * ranges. This walks through whole tree and not just first level children. 745 struct resource *first, *next; local 747 for (;; parent = first) { 748 first = __request_resource(parent, new); 749 if (!first) 750 return first; 752 if (first == parent) 753 return first; 754 if (WARN_ON(first [all...] |
H A D | signal.c | 165 /* Given the mask, find the first available signal that should be serviced. */ 180 * Handle the first word specially: it contains the 181 * synchronous signals that need to be dequeued first. 548 struct sigqueue *q, *first = NULL; local 556 if (first) 558 first = q; 564 if (first) { 566 list_del_init(&first->list); 567 copy_siginfo(info, &first->info); 568 __sigqueue_free(first); [all...] |
H A D | sysctl.c | 1982 int *i, vleft, first = 1, err = 0; local 2025 for (; left && vleft--; i++, first=0) { 2048 if (!first) 2058 if (!write && !first && left && !err) 2065 if (first) 2245 int vleft, first = 1, err = 0; local 2287 for (; left && vleft--; i++, first = 0) { 2307 if (!first) { 2318 if (!write && !first && left && !err) 2325 if (first) 2573 bool first = 1; local [all...] |
H A D | taskstats.c | 220 struct task_struct *tsk, *first; local 229 first = find_task_by_vpid(tgid); 231 if (!first || !lock_task_sighand(first, &flags)) 234 if (first->signal->stats) 235 memcpy(stats, first->signal->stats, sizeof(*stats)); 239 tsk = first; 253 } while_each_thread(first, tsk); 255 unlock_task_sighand(first, &flags);
|
H A D | user.c | 29 .first = 0, 37 .first = 0, 45 .first = 0,
|
H A D | user_namespace.c | 149 u32 first, last, id2; local 157 first = map->extent[idx].first; 158 last = first + map->extent[idx].count - 1; 159 if (id >= first && id <= last && 160 (id2 >= first && id2 <= last)) 165 id = (id - first) + map->extent[idx].lower_first; 175 u32 first, last; local 181 first = map->extent[idx].first; 198 u32 first, last; local [all...] |
/kernel/sched/ |
H A D | core.c | 778 * Iterate task_group tree rooted at *from, calling @down when first entering a 1575 * TIF_NEED_RESCHED remotely (for the first time) will also send 2105 * wake_up_new_task - wake up a newly created task for the first time. 2306 * schedule_tail - first thing a freshly forked thread must call. 3182 * We don't have to worry. Conceptually one call occurs first 3267 * for the first time with its new policy. 5635 struct sched_group *tmp, *first; local 5640 first = sg; 5649 } while (sg != first); 5685 * Also keep a unique ID per domain (we use the first cp 5838 struct sched_group *first = NULL, *last = NULL, *groups = NULL, *sg; local 5938 struct sched_group *first = NULL, *last = NULL; local [all...] |
/kernel/trace/ |
H A D | Kconfig | 162 Its first purpose is to trace the duration of functions and 548 "START". The second string records the cold cache time of the first 557 first=3672 [COLD CACHED] 558 last=632 first=3672 max=632 min=632 avg=316 std=446 std^2=199712 559 last=278 first=3672 max=632 min=278 avg=303 std=316 std^2=100337 560 last=277 first=3672 max=632 min=277 avg=296 std=258 std^2=67064 561 last=273 first=3672 max=632 min=273 avg=292 std=224 std^2=50411 562 last=273 first=3672 max=632 min=273 avg=288 std=200 std^2=40389 563 last=281 first=3672 max=632 min=273 avg=287 std=183 std^2=33666
|
H A D | ftrace.c | 374 int first = *list == &ftrace_list_end; local 376 if (first) 1911 * Need to check removed ops first. 2221 * we need to update to the ftrace_ops_list_func first 2425 * tested first on update. 2619 * correct previous instructions. Making first the NOP 2767 hnd = hhd->first; 3375 if (hhd->first) 3396 if (hhd->first) 4601 * As core pages are first, th [all...] |
H A D | trace.c | 2672 /* Don't print started cpu buffer for the first entry of the trace */ 3705 "\t The first one will disable tracing every time do_fault is hit\n" 3707 "\t The first time do trap is hit and it disables tracing, the\n" 3769 "\t The first disables tracing every time block_unplug is hit.\n" 3770 "\t The second disables tracing the first 3 times block_unplug is hit.\n" 3771 "\t The third enables the kmalloc event the first 3 times block_unplug\n" 4803 /* fill in the size from first enabled cpu */ 5393 /* Force reading ring buffer for first read */ 6337 /* Allocate the first page for all buffers */ 6510 parent = hlist_entry(inode->i_dentry.first, struc [all...] |
H A D | trace_output.c | 79 int i, first = 1; local 89 if (!first && delim) 92 first = 0; 98 if (!first && delim)
|
Completed in 204 milliseconds