/kernel/ |
H A D | Makefile | 136 ifneq ($(wildcard $(obj)/.x509.list),) 137 ifneq ($(shell cat $(obj)/.x509.list),$(X509_CERTIFICATES)) 138 $(info X.509 certificate list changed) 139 $(shell rm $(obj)/.x509.list) 149 $(obj)/x509_certificate_list: $(X509_CERTIFICATES) $(obj)/.x509.list 152 targets += $(obj)/.x509.list 153 $(obj)/.x509.list: 157 clean-files := x509_certificate_list .x509.list
|
H A D | audit.c | 134 /* The audit_freelist is a list of pre-allocated audit buffers (if more 177 struct list_head list; member in struct:audit_buffer 1237 list_add(&ab->list, &audit_freelist); 1252 struct audit_buffer, list); 1253 list_del(&ab->list); 1735 * @record_num: record number to report when handling a list of names
|
H A D | audit.h | 35 * a name dynamically and also add those to the list anchored by names_list. */ 63 struct list_head list; member in struct:audit_entry 84 struct list_head list; /* audit_context->names_list */ member in struct:audit_names 127 * The names_list is the list of all audit_names collected during this 136 struct list_head names_list; /* struct audit_names->list anchor */ 278 extern int audit_add_watch(struct audit_krule *krule, struct list_head **list); 313 #define audit_kill_trees(list) BUG()
|
H A D | audit_tree.c | 17 struct list_head list; member in struct:audit_tree 32 struct list_head list; member in struct:audit_chunk::node 45 * Rules have struct list_head rlist forming a list of rules over 53 * tree.chunks anchors chunk.owners[].list hash_lock 65 * node.index allows to get from node.list to containing chunk. 83 INIT_LIST_HEAD(&tree->list); 153 INIT_LIST_HEAD(&chunk->owners[i].list); 175 struct list_head *list; local 179 list = chunk_hash(entry->i.inode); 180 list_add_rcu(&chunk->hash, list); 186 struct list_head *list = chunk_hash(inode); local 850 audit_kill_trees(struct list_head *list) argument [all...] |
H A D | audit_watch.c | 53 struct list_head wlist; /* entry in parent->watches list */ 212 /* Duplicate the given audit watch. The new watch's rules list is initialized 213 * to an empty list and wlist is undefined. */ 252 audit_log_format(ab, " list=%d res=1", r->listnr); 280 * are on so we need a new watch for the new list */ 294 list_del_rcu(&oentry->list); 298 list_del(&oentry->rule.list); 312 list_add_rcu(&nentry->list, &audit_inode_hash[h]); 313 list_replace(&oentry->rule.list, 314 &nentry->rule.list); 410 audit_add_watch(struct audit_krule *krule, struct list_head **list) argument [all...] |
H A D | auditfilter.c | 176 int __init audit_register_class(int class, unsigned *list) argument 181 while (*list != ~0U) { 182 unsigned n = *list++; 750 * The rlist element is undefined; list manipulations are handled apart from 779 * since we'd have to have rule gone from the list *and* removed 781 * the beginning of list scan. 830 struct list_head *list; local 835 *p = list = &audit_inode_hash[h]; 839 list = &audit_inode_hash[h]; 840 list_for_each_entry(e, list, lis 870 struct list_head *list; local 955 struct list_head *list; local [all...] |
H A D | auditsc.c | 65 #include <linux/list.h> 180 list_for_each_entry(n, &ctx->names_list, list) { 190 * We keep a linked list of fixed-sized (31 pointer) arrays of audit_chunk *; 194 * "empty" becomes (p, p, 31) afterwards. We don't shrink the list (and seriously, 323 list_for_each_entry(n, &ctx->names_list, list) { 347 list_for_each_entry(n, &ctx->names_list, list) { 531 list_for_each_entry(n, &ctx->names_list, list) { 546 list_for_each_entry(n, &ctx->names_list, list) { 559 list_for_each_entry(n, &ctx->names_list, list) { 571 list_for_each_entry(n, &ctx->names_list, list) { 745 audit_filter_syscall(struct task_struct *tsk, struct audit_context *ctx, struct list_head *list) argument 779 struct list_head *list = &audit_inode_hash[h]; local [all...] |
H A D | cgroup.c | 37 #include <linux/list.h> 78 * css_set_rwsem protects task->cgroups pointer, the list of css_set 161 /* The list of hierarchy roots */ 172 * Also, as csses are always appended to the parent's ->children list, it 174 * number order on the list. Protected by cgroup_mutex. 403 /* list of cgrp_cset_links anchored at cgrp->cset_links */ 406 /* list of cgrp_cset_links anchored at css_set->cgrp_links */ 413 * for each subsystem. Also used to anchor the list of css_sets. Not 727 * Always add links to the tail of the list so that the list 3770 pid_t *list; member in struct:cgroup_pidlist 3848 pidlist_uniq(pid_t *list, int length) argument [all...] |
H A D | cpuset.c | 36 #include <linux/list.h> 443 * such as list traversal that depend on the actual address of the 444 * cpuset in the list must use cur below, not trial. 578 * q - a linked-list queue of cpuset pointers, used to implement a 1701 * used, list of ranges of sequential numbers, is variable length, 1703 * gibberish by doing partial reads while a list was changing.
|
H A D | futex.c | 93 * being added to the list is acknowledged when the list is concurrently being 190 * list of 'owned' pi_state instances - these have to be 193 struct list_head list; member in struct:futex_pi_state 208 * @list: priority-sorted list of tasks waiting on this futex 221 * It is considered woken when plist_node_empty(&q->list) || q->lock_ptr == 0. 225 * PI futexes are typically woken before they are removed from the hash list via 229 struct plist_node list; member in struct:futex_q 241 /* list get [all...] |
H A D | futex_compat.c | 20 * Fetch a robust-list pointer. Bit 0 signals PI futexes: 45 * Walk curr->robust_list (very carefully, it's a userspace list!) 48 * We silently return on any sign of list-walking problem. 64 * Fetch the list head (which was registered earlier, via 67 if (fetch_robust_entry(&uentry, &entry, &head->list.next, &pi)) 83 while (entry != (struct robust_list __user *) &head->list) { 85 * Fetch the next entry in the list before calling 91 * A pending lock might already be on the list, so
|
H A D | irq_work.c | 129 static void irq_work_run_list(struct llist_head *list) argument 137 if (llist_empty(list)) 140 llnode = llist_del_all(list);
|
H A D | jump_label.c | 11 #include <linux/list.h>
|
H A D | kexec.c | 18 #include <linux/list.h> 115 * The assembly stub in the control code buffer is passed a linked list 266 /* Initialize the list of control pages */ 269 /* Initialize the list of destination pages */ 272 /* Initialize the list of unusable pages */ 278 static void kimage_free_page_list(struct list_head *list); 483 * In file mode list of segments is prepared by kernel. Copy relevant 484 * data from user space, do error checking, prepare segment list 668 static void kimage_free_page_list(struct list_head *list) argument 672 list_for_each_safe(pos, next, list) { [all...] |
H A D | kprobes.c | 29 * exceptions notifier to be first on the priority list. 89 /* Blacklist -- list of struct kprobe_blacklist_entry */ 100 struct list_head list; member in struct:kprobe_insn_page 154 list_for_each_entry(kip, &c->pages, list) { 190 INIT_LIST_HEAD(&kip->list); 196 list_add(&kip->list, &c->pages); 215 if (!list_is_singular(&kip->list)) { 216 list_del(&kip->list); 232 list_for_each_entry_safe(kip, next, &c->pages, list) { 253 list_for_each_entry(kip, &c->pages, list) { [all...] |
H A D | kthread.c | 38 struct list_head list; member in struct:kthread_create_info 285 list_add_tail(&create->list, &kthread_create_list); 506 struct kthread_create_info, list); 507 list_del_init(&create->list);
|
H A D | latencytop.c | 58 #include <linux/list.h>
|
/kernel/debug/kdb/ |
H A D | kdb_main.c | 1975 list_for_each_entry(mod, kdb_modules, list) { 2258 * list of the active processes. 2810 "Display active task list", 0, KDB_REPEAT_NONE);
|
/kernel/events/ |
H A D | core.c | 1138 * list, group events are kept attached to the group so that 1142 struct list_head *list; local 1147 list = ctx_group_list(event, ctx); 1148 list_add_tail(&event->group_entry, list); 1349 struct list_head *list = NULL; local 1369 list = &event->group_entry; 1374 * to whatever list we are on. 1377 if (list) 1378 list_move_tail(&sibling->group_entry, list); 1515 * remove it from the context list [all...] |
H A D | hw_breakpoint.c | 44 #include <linux/list.h> 115 * The given event must be not on the list.
|
/kernel/gcov/ |
H A D | fs.c | 23 #include <linux/list.h> 32 * @list: list head for child node list 34 * @all: list head for list of all nodes 51 struct list_head list; member in struct:gcov_node 415 INIT_LIST_HEAD(&node->list); 459 list_add(&node->list, &parent->children); 489 list_del(&node->list); [all...] |
/kernel/irq/ |
H A D | generic-chip.c | 305 list_add_tail(&gc->list, &gc_list); 426 list_add_tail(&gc->list, &gc_list); 493 list_del(&gc->list); 532 list_for_each_entry(gc, &gc_list, list) { 549 list_for_each_entry(gc, &gc_list, list) { 569 list_for_each_entry(gc, &gc_list, list) {
|
/kernel/locking/ |
H A D | mutex-debug.c | 34 INIT_LIST_HEAD(&waiter->list); 42 DEBUG_LOCKS_WARN_ON(list_empty(&waiter->list)); 47 DEBUG_LOCKS_WARN_ON(!list_empty(&waiter->list)); 63 DEBUG_LOCKS_WARN_ON(list_empty(&waiter->list)); 68 list_del_init(&waiter->list);
|
H A D | mutex.c | 172 * to waiter list and sleep. 187 list_for_each_entry(cur, &lock->base.wait_list, list) { 524 list_add_tail(&waiter.list, &lock->wait_list); 592 list_for_each_entry(cur, &lock->wait_list, list) { 738 /* get the first entry from the wait-list: */ 741 struct mutex_waiter, list);
|
H A D | mutex.h | 17 __list_del((waiter)->list.prev, (waiter)->list.next)
|