/kernel/ |
H A D | cpu_pm.c | 28 static int cpu_pm_notify(enum cpu_pm_event event, int nr_to_call, int *nr_calls) argument 32 ret = __raw_notifier_call_chain(&cpu_pm_notifier_chain, event, NULL,
|
H A D | hung_task.c | 64 hung_task_panic(struct notifier_block *this, unsigned long event, void *ptr) argument
|
H A D | compat.c | 687 struct sigevent __user *event = NULL; local 692 event = compat_alloc_user_space(sizeof(*event)); 694 copy_to_user(event, &kevent, sizeof(*event))) 698 return sys_timer_create(which_clock, event, created_timer_id); 877 int get_compat_sigevent(struct sigevent *event, argument 880 memset(event, 0, sizeof(*event)); 882 __get_user(event [all...] |
H A D | watchdog.c | 334 /* Callback function for perf event subsystem */ 335 static void watchdog_overflow_callback(struct perf_event *event, argument 340 event->hw.interrupts = 0; 512 /* Enable the perf event */ 530 /* disable the perf event */ 571 struct perf_event *event = per_cpu(watchdog_ev, cpu); local 578 event = ERR_PTR(-ENOENT); 583 if (event && event->state > PERF_EVENT_STATE_OFF) 587 if (event ! 634 struct perf_event *event = per_cpu(watchdog_ev, cpu); local [all...] |
/kernel/events/ |
H A D | callchain.c | 160 perf_callchain(struct perf_event *event, struct pt_regs *regs) argument 165 int kernel = !event->attr.exclude_callchain_kernel; 166 int user = !event->attr.exclude_callchain_user; 197 if (event->ctx->task && event->ctx->task != current)
|
H A D | ring_buffer.c | 23 handle->event->pending_wakeup = 1; 24 irq_work_queue(&handle->event->pending); 29 * event isn't done writing. However since we need to deal with NMIs we 33 * event completes. 105 struct perf_event *event, unsigned int size) 120 if (event->parent) 121 event = event->parent; 123 rb = rcu_dereference(event->rb); 131 handle->event 104 perf_output_begin(struct perf_output_handle *handle, struct perf_event *event, unsigned int size) argument [all...] |
H A D | hw_breakpoint.c | 115 * The given event must be not on the list. 172 * ensure no flexible event can ever be scheduled before a pinned event 553 static void bp_perf_event_destroy(struct perf_event *event) argument 555 release_bp_slot(event);
|
/kernel/gcov/ |
H A D | base.c | 44 * Add new profiling data structure to list and inform event 95 * gcov_enable_events - enable event reporting through gcov_event() 109 /* Perform event callback for previously registered entries. */ 123 static int gcov_module_notifier(struct notifier_block *nb, unsigned long event, argument 130 if (event != MODULE_STATE_GOING)
|
/kernel/trace/ |
H A D | trace_sched_switch.c | 32 struct ring_buffer_event *event; local 35 event = trace_buffer_lock_reserve(buffer, TRACE_CTX, 37 if (!event) 39 entry = ring_buffer_event_data(event); 48 if (!call_filter_check_discard(call, entry, buffer, event)) 49 trace_buffer_unlock_commit(buffer, event, flags, pc); 87 struct ring_buffer_event *event; local 91 event = trace_buffer_lock_reserve(buffer, TRACE_WAKE, 93 if (!event) 95 entry = ring_buffer_event_data(event); [all...] |
H A D | ring_buffer_benchmark.c | 80 struct ring_buffer_event *event; local 84 event = ring_buffer_consume(buffer, cpu, &ts, NULL); 85 if (!event) 88 entry = ring_buffer_event_data(event); 100 struct ring_buffer_event *event; local 116 /* The commit may have missed event flags set, clear them */ 126 event = (void *)&rpage->data[i]; 127 switch (event->type_len) { 130 if (!event->time_delta) 132 inc = event 232 struct ring_buffer_event *event; local [all...] |
H A D | trace_branch.c | 36 struct ring_buffer_event *event; local 61 event = trace_buffer_lock_reserve(buffer, TRACE_BRANCH, 63 if (!event) 66 entry = ring_buffer_event_data(event); 81 if (!call_filter_check_discard(call, entry, buffer, event)) 82 __buffer_unlock_commit(buffer, event); 148 int flags, struct trace_event *event) 147 trace_branch_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
|
H A D | trace_event_perf.c | 2 * trace event based perf event profiling/tracing 21 /* Count the events in use (per event id, not per instance) */ 52 * event, due to issues with page faults while tracing page 215 if (tp_event->event.type == event_id && 331 static int perf_ftrace_function_register(struct perf_event *event) argument 333 struct ftrace_ops *ops = &event->ftrace_ops; 340 static int perf_ftrace_function_unregister(struct perf_event *event) argument 342 struct ftrace_ops *ops = &event->ftrace_ops; 348 static void perf_ftrace_function_enable(struct perf_event *event) argument 353 perf_ftrace_function_disable(struct perf_event *event) argument [all...] |
H A D | trace_mmiotrace.c | 313 struct ring_buffer_event *event; local 317 event = trace_buffer_lock_reserve(buffer, TRACE_MMIO_RW, 319 if (!event) { 323 entry = ring_buffer_event_data(event); 326 if (!call_filter_check_discard(call, entry, buffer, event)) 327 trace_buffer_unlock_commit(buffer, event, 0, pc); 343 struct ring_buffer_event *event; local 347 event = trace_buffer_lock_reserve(buffer, TRACE_MMIO_MAP, 349 if (!event) { 353 entry = ring_buffer_event_data(event); [all...] |
H A D | trace_selftest.c | 26 struct ring_buffer_event *event; local 30 while ((event = ring_buffer_consume(buf->buffer, cpu, NULL, NULL))) { 31 entry = ring_buffer_event_data(event);
|
H A D | trace_syscalls.c | 16 static int syscall_enter_register(struct ftrace_event_call *event, 18 static int syscall_exit_register(struct ftrace_event_call *event, 111 struct trace_event *event) 126 if (entry->enter_event->event.type != ent->type) { 164 struct trace_event *event) 182 if (entry->exit_event->event.type != ent->type) { 308 struct ring_buffer_event *event; local 337 event = trace_buffer_lock_reserve(buffer, 338 sys_data->enter_event->event.type, size, irq_flags, pc); 339 if (!event) 110 print_syscall_enter(struct trace_iterator *iter, int flags, struct trace_event *event) argument 163 print_syscall_exit(struct trace_iterator *iter, int flags, struct trace_event *event) argument 356 struct ring_buffer_event *event; local 708 syscall_enter_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument 736 syscall_exit_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument [all...] |
H A D | trace_events_trigger.c | 2 * trace_events_trigger - trace event triggers 42 * event_triggers_call - Call triggers associated with a trace event 43 * @file: The ftrace_event_file associated with the event 44 * @rec: The trace entry for the event, NULL for unconditional invocation 46 * For each trigger associated with an event, invoke the trigger 53 * in any case until the current event is written, the trigger 94 * event_triggers_post_call - Call 'post_triggers' for a trace event 95 * @file: The ftrace_event_file associated with the event 98 * For each trigger associated with an event, invoke the trigger 311 * Currently we only register event command 1171 const char *event; local [all...] |
H A D | trace_functions_graph.c | 288 struct ring_buffer_event *event; local 295 event = trace_buffer_lock_reserve(buffer, TRACE_GRAPH_ENT, 297 if (!event) 299 entry = ring_buffer_event_data(event); 301 if (!call_filter_check_discard(call, entry, buffer, event)) 302 __buffer_unlock_commit(buffer, event); 403 struct ring_buffer_event *event; local 410 event = trace_buffer_lock_reserve(buffer, TRACE_GRAPH_RET, 412 if (!event) 414 entry = ring_buffer_event_data(event); 625 struct ring_buffer_event *event; local 1248 struct trace_event *event; local [all...] |
H A D | trace_kprobe.c | 28 * Kprobe event core functions 276 const char *event, 303 if (!event || !is_good_name(event)) { 309 tk->tp.call.name = kstrdup(event, GFP_KERNEL); 345 static struct trace_kprobe *find_trace_kprobe(const char *event, argument 351 if (strcmp(ftrace_event_name(&tk->tp.call), event) == 0 && 503 /* Enabled event can not be unregistered */ 525 /* Delete old (same name) event if exist */ 535 /* Register new event */ 275 alloc_trace_kprobe(const char *group, const char *event, void *addr, const char *symbol, unsigned long offs, int nargs, bool is_return) argument 612 char *symbol = NULL, *event = NULL, *group = NULL; local 922 struct ring_buffer_event *event; local 970 struct ring_buffer_event *event; local 1015 print_kprobe_event(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1051 print_kretprobe_event(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1210 kprobe_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument [all...] |
H A D | trace_output.c | 192 struct ftrace_event_call *event; local 198 event = container_of(trace_event, struct ftrace_event_call, event); 201 if (entry->type != event->event.type) { 207 ret = trace_seq_printf(s, "%s: ", ftrace_event_name(event)); 608 * ftrace_find_event - find a registered event 609 * @type: the type of event to look for 611 * Returns an event of type @type otherwise NULL 616 struct trace_event *event; local 684 register_ftrace_event(struct trace_event *event) argument 753 __unregister_ftrace_event(struct trace_event *event) argument 764 unregister_ftrace_event(struct trace_event *event) argument 778 trace_nop_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 788 trace_fn_trace(struct trace_iterator *iter, int flags, struct trace_event *event) argument 816 trace_fn_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 831 trace_fn_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 845 trace_fn_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument 872 trace_graph_ent_trace(struct trace_iterator *iter, int flags, struct trace_event *event) argument 892 trace_graph_ent_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 907 trace_graph_ent_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 921 trace_graph_ent_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument 948 trace_graph_ret_trace(struct trace_iterator *iter, int flags, struct trace_event *event) argument 969 trace_graph_ret_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 987 trace_graph_ret_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1004 trace_graph_ret_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1061 trace_ctx_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1067 trace_wake_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1096 trace_ctx_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1102 trace_wake_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1132 trace_ctx_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1138 trace_wake_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1144 trace_ctxwake_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1188 trace_stack_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1228 trace_user_stack_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1259 trace_bputs_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1285 trace_bputs_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1317 trace_bprint_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1343 trace_bprint_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1374 trace_print_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1394 trace_print_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1437 struct trace_event *event; local [all...] |
H A D | trace_uprobe.c | 50 * uprobe event core functions 239 alloc_trace_uprobe(const char *group, const char *event, int nargs, bool is_ret) argument 243 if (!event || !is_good_name(event)) 254 tu->tp.call.name = kstrdup(event, GFP_KERNEL); 291 static struct trace_uprobe *find_probe_event(const char *event, const char *group) argument 296 if (strcmp(ftrace_event_name(&tu->tp.call), event) == 0 && 325 /* register as an event */ 329 /* delete old event */ 337 pr_warning("Failed to register probe event( 359 char *arg, *event, *group, *filename; local 776 struct ring_buffer_event *event; local 843 print_uprobe_event(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1006 struct perf_event *event; local 1020 uprobe_filter_event(struct trace_uprobe *tu, struct perf_event *event) argument 1025 uprobe_perf_close(struct trace_uprobe *tu, struct perf_event *event) argument 1047 uprobe_perf_open(struct trace_uprobe *tu, struct perf_event *event) argument 1165 trace_uprobe_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument [all...] |
H A D | blktrace.c | 72 struct ring_buffer_event *event = NULL; local 81 event = trace_buffer_lock_reserve(buffer, TRACE_BLK, 84 if (!event) 86 t = ring_buffer_event_data(event); 106 trace_buffer_unlock_commit(buffer, event, 0, pc); 205 struct ring_buffer_event *event = NULL; local 235 event = trace_buffer_lock_reserve(buffer, TRACE_BLK, 238 if (!event) 240 t = ring_buffer_event_data(event); 281 trace_buffer_unlock_commit(buffer, event, 1394 blk_trace_event_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1417 blk_trace_event_print_binary(struct trace_iterator *iter, int flags, struct trace_event *event) argument [all...] |
H A D | trace_events.c | 2 * event tracer 185 id = register_ftrace_event(&call->event); 203 fbuffer->event = 205 event_call->event.type, len, 207 if (!fbuffer->event) 210 fbuffer->entry = ring_buffer_event_data(fbuffer->event); 218 fbuffer->event, fbuffer->entry, 295 * to clear the SOFT_DISABLED flag but leave the event in the 296 * state that it was. That is, if the event was enabled and 298 * is set we do not want the event t 483 __ftrace_set_clr_event_nolock(struct trace_array *tr, const char *match, const char *sub, const char *event, int set) argument 521 __ftrace_set_clr_event(struct trace_array *tr, const char *match, const char *sub, const char *event, int set) argument 535 char *event = NULL, *sub = NULL, *match; local 576 trace_set_clr_event(const char *system, const char *event, int set) argument 1903 find_event_file(struct trace_array *tr, const char *system, const char *event) argument 2066 const char *event; local 2659 struct ring_buffer_event *event; local [all...] |
H A D | trace_events_filter.c | 2 * trace_events_filter - generic event filtering 143 static int filter_pred_##type(struct filter_pred *pred, void *event) \ 145 type *addr = (type *)(event + pred->offset); \ 173 static int filter_pred_##size(struct filter_pred *pred, void *event) \ 175 u##size *addr = (u##size *)(event + pred->offset); \ 199 static int filter_pred_string(struct filter_pred *pred, void *event) argument 201 char *addr = (char *)(event + pred->offset); 212 static int filter_pred_pchar(struct filter_pred *pred, void *event) argument 214 char **addr = (char **)(event + pred->offset); 235 static int filter_pred_strloc(struct filter_pred *pred, void *event) argument 250 filter_pred_none(struct filter_pred *pred, void *event) argument 2043 ftrace_profile_free_filter(struct perf_event *event) argument 2173 ftrace_function_set_filter(struct perf_event *event, struct event_filter *filter) argument 2186 ftrace_function_set_filter(struct perf_event *event, struct event_filter *filter) argument 2193 ftrace_profile_set_filter(struct perf_event *event, int event_id, char *filter_str) argument 2317 test_pred_visited_fn(struct filter_pred *pred, void *event) argument [all...] |
/kernel/locking/ |
H A D | rtmutex-tester.c | 31 int event; member in struct:test_thread_data 51 RTTEST_RESETEVENT = 98, /* 98 Reset event counter */ 66 td->event = atomic_add_return(1, &rttest_event); 96 td->event = atomic_add_return(1, &rttest_event); 98 td->event = atomic_add_return(1, &rttest_event); 109 td->event = atomic_add_return(1, &rttest_event); 111 td->event = atomic_add_return(1, &rttest_event); 120 td->event = atomic_add_return(1, &rttest_event); 122 td->event = atomic_add_return(1, &rttest_event); 136 * This allows us to have finegrained control over the event flo [all...] |
/kernel/time/ |
H A D | posix-timers.c | 499 static struct pid *good_sigevent(sigevent_t * event) argument 503 if ((event->sigev_notify & SIGEV_THREAD_ID ) && 504 (!(rtn = find_task_by_vpid(event->sigev_notify_thread_id)) || 506 (event->sigev_notify & ~SIGEV_THREAD_ID) != SIGEV_SIGNAL)) 509 if (((event->sigev_notify & ~SIGEV_THREAD_ID) != SIGEV_NONE) && 510 ((event->sigev_signo <= 0) || (event->sigev_signo > SIGRTMAX))) 602 sigevent_t event; local 627 if (copy_from_user(&event, timer_event_spec, sizeof (event))) { [all...] |