/kernel/events/ |
H A D | core.c | 127 static bool is_kernel_event(struct perf_event *event) argument 129 return event->owner == EVENT_OWNER_KERNEL; 168 * perf event paranoia level: 180 * max perf event sample rate 316 static u64 perf_event_time(struct perf_event *event); 371 * if there is no cgroup event for the current CPU context. 381 perf_cgroup_match(struct perf_event *event) argument 383 struct perf_event_context *ctx = event->ctx; 386 /* @event doesn't care about cgroup */ 387 if (!event 404 perf_detach_cgroup(struct perf_event *event) argument 410 is_cgroup_event(struct perf_event *event) argument 415 perf_cgroup_event_time(struct perf_event *event) argument 443 update_cgrp_time_from_event(struct perf_event *event) argument 605 perf_cgroup_connect(int fd, struct perf_event *event, struct perf_event_attr *attr, struct perf_event *group_leader) argument 642 perf_cgroup_set_shadow_time(struct perf_event *event, u64 now) argument 650 perf_cgroup_defer_enabled(struct perf_event *event) argument 663 perf_cgroup_mark_enabled(struct perf_event *event, struct perf_event_context *ctx) argument 685 perf_cgroup_match(struct perf_event *event) argument 690 perf_detach_cgroup(struct perf_event *event) argument 693 is_cgroup_event(struct perf_event *event) argument 698 perf_cgroup_event_cgrp_time(struct perf_event *event) argument 703 update_cgrp_time_from_event(struct perf_event *event) argument 721 perf_cgroup_connect(pid_t pid, struct perf_event *event, struct perf_event_attr *attr, struct perf_event *group_leader) argument 740 perf_cgroup_set_shadow_time(struct perf_event *event, u64 now) argument 744 perf_cgroup_event_time(struct perf_event *event) argument 750 perf_cgroup_defer_enabled(struct perf_event *event) argument 755 perf_cgroup_mark_enabled(struct perf_event *event, struct perf_event_context *ctx) argument 928 perf_event_pid(struct perf_event *event, struct task_struct *p) argument 939 perf_event_tid(struct perf_event *event, struct task_struct *p) argument 954 primary_event_id(struct perf_event *event) argument 1055 perf_event_time(struct perf_event *event) argument 1069 update_event_times(struct perf_event *event) argument 1110 struct perf_event *event; local 1118 ctx_group_list(struct perf_event *event, struct perf_event_context *ctx) argument 1131 list_add_event(struct perf_event *event, struct perf_event_context *ctx) argument 1170 perf_event__state_init(struct perf_event *event) argument 1180 perf_event__read_size(struct perf_event *event) argument 1204 perf_event__header_size(struct perf_event *event) argument 1236 perf_event__id_header_size(struct perf_event *event) argument 1263 perf_group_attach(struct perf_event *event) argument 1296 list_del_event(struct perf_event *event, struct perf_event_context *ctx) argument 1346 perf_group_detach(struct perf_event *event) argument 1395 is_orphaned_event(struct perf_event *event) argument 1404 is_orphaned_child(struct perf_event *event) argument 1432 event_filter_match(struct perf_event *event) argument 1439 event_sched_out(struct perf_event *event, struct perf_cpu_context *cpuctx, struct perf_event_context *ctx) argument 1491 struct perf_event *event; local 1507 struct perf_event *event; member in struct:remove_event 1520 struct perf_event *event = re->event; local 1552 perf_remove_from_context(struct perf_event *event, bool detach_group) argument 1608 struct perf_event *event = info; local 1657 perf_event_disable(struct perf_event *event) argument 1700 perf_set_shadow_time(struct perf_event *event, struct perf_event_context *ctx, u64 tstamp) argument 1740 event_sched_in(struct perf_event *event, struct perf_cpu_context *cpuctx, struct perf_event_context *ctx) argument 1806 struct perf_event *event, *partial_group = NULL; local 1873 group_can_go_on(struct perf_event *event, struct perf_cpu_context *cpuctx, int can_add_hw) argument 1901 add_event_to_ctx(struct perf_event *event, struct perf_event_context *ctx) argument 1939 struct perf_event *event = info; local 2004 perf_install_in_context(struct perf_event_context *ctx, struct perf_event *event, int cpu) argument 2060 __perf_event_mark_enabled(struct perf_event *event) argument 2078 struct perf_event *event = info; local 2161 perf_event_enable(struct perf_event *event) argument 2219 perf_event_refresh(struct perf_event *event, int refresh) argument 2238 struct perf_event *event; local 2301 __perf_event_sync_stat(struct perf_event *event, struct perf_event *next_event) argument 2350 struct perf_event *event, *next_event; local 2498 struct perf_event *event; local 2528 struct perf_event *event; local 2718 perf_calculate_period(struct perf_event *event, u64 nsec, u64 count) argument 2794 perf_adjust_period(struct perf_event *event, u64 nsec, u64 count, bool disable) argument 2831 struct perf_event *event; local 2989 event_enable_on_exec(struct perf_event *event, struct perf_event_context *ctx) argument 3011 struct perf_event *event; local 3078 struct perf_event *event = info; local 3103 perf_event_count(struct perf_event *event) argument 3108 perf_event_read(struct perf_event *event) argument 3289 struct perf_event *event; local 3302 unaccount_event_cpu(struct perf_event *event, int cpu) argument 3315 unaccount_event(struct perf_event *event) argument 3338 __free_event(struct perf_event *event) argument 3357 _free_event(struct perf_event *event) argument 3385 free_event(struct perf_event *event) argument 3400 perf_remove_from_owner(struct perf_event *event) argument 3441 put_event(struct perf_event *event) argument 3471 perf_event_release_kernel(struct perf_event *event) argument 3490 struct perf_event *event, *tmp; local 3520 perf_event_read_value(struct perf_event *event, u64 *enabled, u64 *running) argument 3546 perf_event_read_group(struct perf_event *event, u64 read_format, char __user *buf) argument 3596 perf_event_read_one(struct perf_event *event, u64 read_format, char __user *buf) argument 3617 is_event_hup(struct perf_event *event) argument 3634 perf_read_hw(struct perf_event *event, char __user *buf, size_t count) argument 3662 struct perf_event *event = file->private_data; local 3669 struct perf_event *event = file->private_data; local 3690 perf_event_reset(struct perf_event *event) argument 3703 perf_event_for_each_child(struct perf_event *event, void (*func)(struct perf_event *)) argument 3716 perf_event_for_each(struct perf_event *event, void (*func)(struct perf_event *)) argument 3732 perf_event_period(struct perf_event *event, u64 __user *arg) argument 3801 struct perf_event *event = file->private_data; local 3886 struct perf_event *event; local 3898 struct perf_event *event; local 3908 perf_event_index(struct perf_event *event) argument 3919 calc_timer_values(struct perf_event *event, u64 *now, u64 *enabled, u64 *running) argument 3932 perf_event_init_userpage(struct perf_event *event) argument 3961 perf_event_update_userpage(struct perf_event *event) argument 4013 struct perf_event *event = vma->vm_file->private_data; local 4046 ring_buffer_attach(struct perf_event *event, struct ring_buffer *rb) argument 4092 ring_buffer_wakeup(struct perf_event *event) argument 4113 ring_buffer_get(struct perf_event *event) argument 4140 struct perf_event *event = vma->vm_file->private_data; local 4156 struct perf_event *event = vma->vm_file->private_data; local 4243 struct perf_event *event = file->private_data; local 4370 struct perf_event *event = filp->private_data; local 4401 perf_event_wakeup(struct perf_event *event) argument 4413 struct perf_event *event = container_of(entry, local 4574 __perf_event_header__init_id(struct perf_event_header *header, struct perf_sample_data *data, struct perf_event *event) argument 4604 perf_event_header__init_id(struct perf_event_header *header, struct perf_sample_data *data, struct perf_event *event) argument 4636 perf_event__output_id_sample(struct perf_event *event, struct perf_output_handle *handle, struct perf_sample_data *sample) argument 4644 perf_output_read_one(struct perf_output_handle *handle, struct perf_event *event, u64 enabled, u64 running) argument 4670 perf_output_read_group(struct perf_output_handle *handle, struct perf_event *event, u64 enabled, u64 running) argument 4714 perf_output_read(struct perf_output_handle *handle, struct perf_event *event) argument 4738 perf_output_sample(struct perf_output_handle *handle, struct perf_event_header *header, struct perf_sample_data *data, struct perf_event *event) argument 4875 perf_prepare_sample(struct perf_event_header *header, struct perf_sample_data *data, struct perf_event *event, struct pt_regs *regs) argument 4969 perf_event_output(struct perf_event *event, struct perf_sample_data *data, struct pt_regs *regs) argument 5004 perf_event_read_event(struct perf_event *event, struct task_struct *task) argument 5039 struct perf_event *event; local 5106 perf_event_task_match(struct perf_event *event) argument 5113 perf_event_task_output(struct perf_event *event, void *data) argument 5202 perf_event_comm_match(struct perf_event *event) argument 5207 perf_event_comm_output(struct perf_event *event, void *data) argument 5309 perf_event_mmap_match(struct perf_event *event, void *data) argument 5320 perf_event_mmap_output(struct perf_event *event, void *data) argument 5530 perf_log_throttle(struct perf_event *event, int enable) argument 5571 __perf_event_overflow(struct perf_event *event, int throttle, struct perf_sample_data *data, struct pt_regs *regs) argument 5639 perf_event_overflow(struct perf_event *event, struct perf_sample_data *data, struct pt_regs *regs) argument 5671 perf_swevent_set_period(struct perf_event *event) argument 5694 perf_swevent_overflow(struct perf_event *event, u64 overflow, struct perf_sample_data *data, struct pt_regs *regs) argument 5720 perf_swevent_event(struct perf_event *event, u64 nr, struct perf_sample_data *data, struct pt_regs *regs) argument 5749 perf_exclude_event(struct perf_event *event, struct pt_regs *regs) argument 5766 perf_swevent_match(struct perf_event *event, enum perf_type_id type, u32 event_id, struct perf_sample_data *data, struct pt_regs *regs) argument 5814 find_swevent_head(struct swevent_htable *swhash, struct perf_event *event) argument 5839 struct perf_event *event; local 5888 perf_swevent_read(struct perf_event *event) argument 5892 perf_swevent_add(struct perf_event *event, int flags) argument 5920 perf_swevent_del(struct perf_event *event, int flags) argument 5925 perf_swevent_start(struct perf_event *event, int flags) argument 5930 perf_swevent_stop(struct perf_event *event, int flags) argument 5954 swevent_hlist_put_cpu(struct perf_event *event, int cpu) argument 5966 swevent_hlist_put(struct perf_event *event) argument 5974 swevent_hlist_get_cpu(struct perf_event *event, int cpu) argument 5998 swevent_hlist_get(struct perf_event *event) argument 6027 sw_perf_event_destroy(struct perf_event *event) argument 6037 perf_swevent_init(struct perf_event *event) argument 6089 perf_tp_filter_match(struct perf_event *event, struct perf_sample_data *data) argument 6099 perf_tp_event_match(struct perf_event *event, struct perf_sample_data *data, struct pt_regs *regs) argument 6122 struct perf_event *event; local 6166 tp_perf_event_destroy(struct perf_event *event) argument 6171 perf_tp_event_init(struct perf_event *event) argument 6209 perf_event_set_filter(struct perf_event *event, void __user *arg) argument 6227 perf_event_free_filter(struct perf_event *event) argument 6238 perf_event_set_filter(struct perf_event *event, void __user *arg) argument 6243 perf_event_free_filter(struct perf_event *event) argument 6271 struct perf_event *event; local 6296 perf_swevent_start_hrtimer(struct perf_event *event) argument 6318 perf_swevent_cancel_hrtimer(struct perf_event *event) argument 6330 perf_swevent_init_hrtimer(struct perf_event *event) argument 6359 cpu_clock_event_update(struct perf_event *event) argument 6369 cpu_clock_event_start(struct perf_event *event, int flags) argument 6375 cpu_clock_event_stop(struct perf_event *event, int flags) argument 6381 cpu_clock_event_add(struct perf_event *event, int flags) argument 6389 cpu_clock_event_del(struct perf_event *event, int flags) argument 6394 cpu_clock_event_read(struct perf_event *event) argument 6399 cpu_clock_event_init(struct perf_event *event) argument 6433 task_clock_event_update(struct perf_event *event, u64 now) argument 6443 task_clock_event_start(struct perf_event *event, int flags) argument 6449 task_clock_event_stop(struct perf_event *event, int flags) argument 6455 task_clock_event_add(struct perf_event *event, int flags) argument 6463 task_clock_event_del(struct perf_event *event, int flags) argument 6468 task_clock_event_read(struct perf_event *event) argument 6477 task_clock_event_init(struct perf_event *event) argument 6532 perf_event_idx_default(struct perf_event *event) argument 6820 perf_init_event(struct perf_event *event) argument 6865 account_event_cpu(struct perf_event *event, int cpu) argument 6878 account_event(struct perf_event *event) argument 6915 struct perf_event *event; local 7164 perf_event_set_output(struct perf_event *event, struct perf_event *output_event) argument 7224 struct perf_event *event, *sibling; local 7506 struct perf_event *event; local 7550 struct perf_event *event, *tmp; local 7728 struct perf_event *event, *tmp; local 7750 perf_free_event(struct perf_event *event, struct perf_event_context *ctx) argument 7776 struct perf_event *event, *tmp; local 7925 inherit_task_group(struct perf_event *event, struct task_struct *parent, struct perf_event_context *parent_ctx, struct task_struct *child, int ctxn, int *inherited_all) argument 7970 struct perf_event *event; local [all...] |
H A D | ring_buffer.c | 23 handle->event->pending_wakeup = 1; 24 irq_work_queue(&handle->event->pending); 29 * event isn't done writing. However since we need to deal with NMIs we 33 * event completes. 105 struct perf_event *event, unsigned int size) 120 if (event->parent) 121 event = event->parent; 123 rb = rcu_dereference(event->rb); 131 handle->event 104 perf_output_begin(struct perf_output_handle *handle, struct perf_event *event, unsigned int size) argument [all...] |
H A D | callchain.c | 160 perf_callchain(struct perf_event *event, struct pt_regs *regs) argument 165 int kernel = !event->attr.exclude_callchain_kernel; 166 int user = !event->attr.exclude_callchain_user; 197 if (event->ctx->task && event->ctx->task != current)
|
H A D | internal.h | 25 local_t events; /* event limit */ 45 extern void perf_event_wakeup(struct perf_event *event); 50 struct perf_event *event); 52 perf_event__output_id_sample(struct perf_event *event, 150 perf_callchain(struct perf_event *event, struct pt_regs *regs);
|
/kernel/trace/ |
H A D | ring_buffer_benchmark.c | 80 struct ring_buffer_event *event; local 84 event = ring_buffer_consume(buffer, cpu, &ts, NULL); 85 if (!event) 88 entry = ring_buffer_event_data(event); 100 struct ring_buffer_event *event; local 116 /* The commit may have missed event flags set, clear them */ 126 event = (void *)&rpage->data[i]; 127 switch (event->type_len) { 130 if (!event->time_delta) 132 inc = event 232 struct ring_buffer_event *event; local [all...] |
H A D | trace_output.c | 192 struct ftrace_event_call *event; local 198 event = container_of(trace_event, struct ftrace_event_call, event); 201 if (entry->type != event->event.type) { 207 ret = trace_seq_printf(s, "%s: ", ftrace_event_name(event)); 608 * ftrace_find_event - find a registered event 609 * @type: the type of event to look for 611 * Returns an event of type @type otherwise NULL 616 struct trace_event *event; local 684 register_ftrace_event(struct trace_event *event) argument 753 __unregister_ftrace_event(struct trace_event *event) argument 764 unregister_ftrace_event(struct trace_event *event) argument 778 trace_nop_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 788 trace_fn_trace(struct trace_iterator *iter, int flags, struct trace_event *event) argument 816 trace_fn_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 831 trace_fn_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 845 trace_fn_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument 872 trace_graph_ent_trace(struct trace_iterator *iter, int flags, struct trace_event *event) argument 892 trace_graph_ent_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 907 trace_graph_ent_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 921 trace_graph_ent_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument 948 trace_graph_ret_trace(struct trace_iterator *iter, int flags, struct trace_event *event) argument 969 trace_graph_ret_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 987 trace_graph_ret_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1004 trace_graph_ret_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1061 trace_ctx_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1067 trace_wake_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1096 trace_ctx_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1102 trace_wake_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1132 trace_ctx_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1138 trace_wake_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1144 trace_ctxwake_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1188 trace_stack_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1228 trace_user_stack_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1259 trace_bputs_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1285 trace_bputs_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1317 trace_bprint_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1343 trace_bprint_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1374 trace_print_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1394 trace_print_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1437 struct trace_event *event; local [all...] |
H A D | trace_output.h | 30 int flags, struct trace_event *event); 35 extern int __unregister_ftrace_event(struct trace_event *event);
|
H A D | ring_buffer.c | 202 #define skip_time_extend(event) \ 203 ((struct ring_buffer_event *)((char *)event + RB_LEN_TIME_EXTEND)) 205 static inline int rb_null_event(struct ring_buffer_event *event) argument 207 return event->type_len == RINGBUF_TYPE_PADDING && !event->time_delta; 210 static void rb_event_set_padding(struct ring_buffer_event *event) argument 213 event->type_len = RINGBUF_TYPE_PADDING; 214 event->time_delta = 0; 218 rb_event_data_length(struct ring_buffer_event *event) argument 222 if (event 235 rb_event_length(struct ring_buffer_event *event) argument 264 rb_event_ts_length(struct ring_buffer_event *event) argument 286 ring_buffer_event_length(struct ring_buffer_event *event) argument 305 rb_event_data(struct ring_buffer_event *event) argument 321 ring_buffer_event_data(struct ring_buffer_event *event) argument 1892 rb_event_index(struct ring_buffer_event *event) argument 1900 rb_event_is_commit(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument 1993 rb_add_time_stamp(struct ring_buffer_event *event, u64 delta) argument 2022 rb_update_event(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event, unsigned length, int add_timestamp, u64 delta) argument 2209 struct ring_buffer_event event; /* Used only for sizeof array */ local 2229 struct ring_buffer_event *event; local 2409 struct ring_buffer_event *event; local 2461 rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument 2543 struct ring_buffer_event *event; local 2725 struct ring_buffer_event *event; local 2769 rb_update_write_stamp(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument 2796 rb_commit(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument 2829 ring_buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event) argument 2849 rb_event_discard(struct ring_buffer_event *event) argument 2869 rb_decrement_entry(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument 2921 ring_buffer_discard_commit(struct ring_buffer *buffer, struct ring_buffer_event *event) argument 2977 struct ring_buffer_event *event; local 3436 rb_update_read_stamp(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument 3467 rb_update_iter_read_stamp(struct ring_buffer_iter *iter, struct ring_buffer_event *event) argument 3628 struct ring_buffer_event *event; local 3652 struct ring_buffer_event *event; local 3700 struct ring_buffer_event *event; local 3767 struct ring_buffer_event *event; local 3872 struct ring_buffer_event *event; local 3909 struct ring_buffer_event *event; local 3939 struct ring_buffer_event *event = NULL; local 4104 struct ring_buffer_event *event; local 4462 struct ring_buffer_event *event; local 4733 struct ring_buffer_event *event; local 4909 struct ring_buffer_event *event; local [all...] |
H A D | trace_sched_switch.c | 32 struct ring_buffer_event *event; local 35 event = trace_buffer_lock_reserve(buffer, TRACE_CTX, 37 if (!event) 39 entry = ring_buffer_event_data(event); 48 if (!call_filter_check_discard(call, entry, buffer, event)) 49 trace_buffer_unlock_commit(buffer, event, flags, pc); 87 struct ring_buffer_event *event; local 91 event = trace_buffer_lock_reserve(buffer, TRACE_WAKE, 93 if (!event) 95 entry = ring_buffer_event_data(event); [all...] |
H A D | trace_syscalls.c | 16 static int syscall_enter_register(struct ftrace_event_call *event, 18 static int syscall_exit_register(struct ftrace_event_call *event, 111 struct trace_event *event) 126 if (entry->enter_event->event.type != ent->type) { 164 struct trace_event *event) 182 if (entry->exit_event->event.type != ent->type) { 308 struct ring_buffer_event *event; local 337 event = trace_buffer_lock_reserve(buffer, 338 sys_data->enter_event->event.type, size, irq_flags, pc); 339 if (!event) 110 print_syscall_enter(struct trace_iterator *iter, int flags, struct trace_event *event) argument 163 print_syscall_exit(struct trace_iterator *iter, int flags, struct trace_event *event) argument 356 struct ring_buffer_event *event; local 708 syscall_enter_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument 736 syscall_exit_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument [all...] |
H A D | trace_uprobe.c | 50 * uprobe event core functions 239 alloc_trace_uprobe(const char *group, const char *event, int nargs, bool is_ret) argument 243 if (!event || !is_good_name(event)) 254 tu->tp.call.name = kstrdup(event, GFP_KERNEL); 291 static struct trace_uprobe *find_probe_event(const char *event, const char *group) argument 296 if (strcmp(ftrace_event_name(&tu->tp.call), event) == 0 && 325 /* register as an event */ 329 /* delete old event */ 337 pr_warning("Failed to register probe event( 359 char *arg, *event, *group, *filename; local 776 struct ring_buffer_event *event; local 843 print_uprobe_event(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1006 struct perf_event *event; local 1020 uprobe_filter_event(struct trace_uprobe *tu, struct perf_event *event) argument 1025 uprobe_perf_close(struct trace_uprobe *tu, struct perf_event *event) argument 1047 uprobe_perf_open(struct trace_uprobe *tu, struct perf_event *event) argument 1165 trace_uprobe_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument [all...] |
H A D | trace_mmiotrace.c | 313 struct ring_buffer_event *event; local 317 event = trace_buffer_lock_reserve(buffer, TRACE_MMIO_RW, 319 if (!event) { 323 entry = ring_buffer_event_data(event); 326 if (!call_filter_check_discard(call, entry, buffer, event)) 327 trace_buffer_unlock_commit(buffer, event, 0, pc); 343 struct ring_buffer_event *event; local 347 event = trace_buffer_lock_reserve(buffer, TRACE_MMIO_MAP, 349 if (!event) { 353 entry = ring_buffer_event_data(event); [all...] |
H A D | trace_event_perf.c | 2 * trace event based perf event profiling/tracing 21 /* Count the events in use (per event id, not per instance) */ 52 * event, due to issues with page faults while tracing page 215 if (tp_event->event.type == event_id && 331 static int perf_ftrace_function_register(struct perf_event *event) argument 333 struct ftrace_ops *ops = &event->ftrace_ops; 340 static int perf_ftrace_function_unregister(struct perf_event *event) argument 342 struct ftrace_ops *ops = &event->ftrace_ops; 348 static void perf_ftrace_function_enable(struct perf_event *event) argument 353 perf_ftrace_function_disable(struct perf_event *event) argument [all...] |
H A D | trace_kprobe.c | 28 * Kprobe event core functions 276 const char *event, 303 if (!event || !is_good_name(event)) { 309 tk->tp.call.name = kstrdup(event, GFP_KERNEL); 345 static struct trace_kprobe *find_trace_kprobe(const char *event, argument 351 if (strcmp(ftrace_event_name(&tk->tp.call), event) == 0 && 503 /* Enabled event can not be unregistered */ 525 /* Delete old (same name) event if exist */ 535 /* Register new event */ 275 alloc_trace_kprobe(const char *group, const char *event, void *addr, const char *symbol, unsigned long offs, int nargs, bool is_return) argument 612 char *symbol = NULL, *event = NULL, *group = NULL; local 922 struct ring_buffer_event *event; local 970 struct ring_buffer_event *event; local 1015 print_kprobe_event(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1051 print_kretprobe_event(struct trace_iterator *iter, int flags, struct trace_event *event) argument 1210 kprobe_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument [all...] |
H A D | trace_branch.c | 36 struct ring_buffer_event *event; local 61 event = trace_buffer_lock_reserve(buffer, TRACE_BRANCH, 63 if (!event) 66 entry = ring_buffer_event_data(event); 81 if (!call_filter_check_discard(call, entry, buffer, event)) 82 __buffer_unlock_commit(buffer, event); 148 int flags, struct trace_event *event) 147 trace_branch_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
|
H A D | trace_events.c | 2 * event tracer 185 id = register_ftrace_event(&call->event); 203 fbuffer->event = 205 event_call->event.type, len, 207 if (!fbuffer->event) 210 fbuffer->entry = ring_buffer_event_data(fbuffer->event); 218 fbuffer->event, fbuffer->entry, 295 * to clear the SOFT_DISABLED flag but leave the event in the 296 * state that it was. That is, if the event was enabled and 298 * is set we do not want the event t 483 __ftrace_set_clr_event_nolock(struct trace_array *tr, const char *match, const char *sub, const char *event, int set) argument 521 __ftrace_set_clr_event(struct trace_array *tr, const char *match, const char *sub, const char *event, int set) argument 535 char *event = NULL, *sub = NULL, *match; local 576 trace_set_clr_event(const char *system, const char *event, int set) argument 1903 find_event_file(struct trace_array *tr, const char *system, const char *event) argument 2066 const char *event; local 2659 struct ring_buffer_event *event; local [all...] |
H A D | trace.c | 84 * tracing is active, only save the comm when a trace event 252 struct ring_buffer_event *event) 256 ring_buffer_discard_commit(buffer, event); 266 struct ring_buffer_event *event) 270 ring_buffer_discard_commit(buffer, event); 464 struct ring_buffer_event *event; local 483 event = trace_buffer_lock_reserve(buffer, TRACE_PRINT, alloc, 485 if (!event) 488 entry = ring_buffer_event_data(event); 500 __buffer_unlock_commit(buffer, event); 250 filter_check_discard(struct ftrace_event_file *file, void *rec, struct ring_buffer *buffer, struct ring_buffer_event *event) argument 264 call_filter_check_discard(struct ftrace_event_call *call, void *rec, struct ring_buffer *buffer, struct ring_buffer_event *event) argument 514 struct ring_buffer_event *event; local 1623 struct ring_buffer_event *event; local 1637 __buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event) argument 1644 __trace_buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event, unsigned long flags, int pc) argument 1654 trace_buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event, unsigned long flags, int pc) argument 1701 trace_current_buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event, unsigned long flags, int pc) argument 1709 trace_buffer_unlock_commit_regs(struct ring_buffer *buffer, struct ring_buffer_event *event, unsigned long flags, int pc, struct pt_regs *regs) argument 1721 trace_current_buffer_discard_commit(struct ring_buffer *buffer, struct ring_buffer_event *event) argument 1735 struct ring_buffer_event *event; local 1769 struct ring_buffer_event *event; local 1897 struct ring_buffer_event *event; local 2102 struct ring_buffer_event *event; local 2160 struct ring_buffer_event *event; local 2263 struct ring_buffer_event *event; local 2394 struct ring_buffer_event *event; local 2683 struct trace_event *event; local 2716 struct trace_event *event; local 2743 struct trace_event *event; local 2769 struct trace_event *event; local 4918 struct ring_buffer_event *event; local 6676 trace_panic_handler(struct notifier_block *this, unsigned long event, void *unused) argument [all...] |
H A D | trace_functions_graph.c | 288 struct ring_buffer_event *event; local 295 event = trace_buffer_lock_reserve(buffer, TRACE_GRAPH_ENT, 297 if (!event) 299 entry = ring_buffer_event_data(event); 301 if (!call_filter_check_discard(call, entry, buffer, event)) 302 __buffer_unlock_commit(buffer, event); 403 struct ring_buffer_event *event; local 410 event = trace_buffer_lock_reserve(buffer, TRACE_GRAPH_RET, 412 if (!event) 414 entry = ring_buffer_event_data(event); 625 struct ring_buffer_event *event; local 1248 struct trace_event *event; local [all...] |
H A D | trace_events_filter.c | 2 * trace_events_filter - generic event filtering 143 static int filter_pred_##type(struct filter_pred *pred, void *event) \ 145 type *addr = (type *)(event + pred->offset); \ 173 static int filter_pred_##size(struct filter_pred *pred, void *event) \ 175 u##size *addr = (u##size *)(event + pred->offset); \ 199 static int filter_pred_string(struct filter_pred *pred, void *event) argument 201 char *addr = (char *)(event + pred->offset); 212 static int filter_pred_pchar(struct filter_pred *pred, void *event) argument 214 char **addr = (char **)(event + pred->offset); 235 static int filter_pred_strloc(struct filter_pred *pred, void *event) argument 250 filter_pred_none(struct filter_pred *pred, void *event) argument 2043 ftrace_profile_free_filter(struct perf_event *event) argument 2173 ftrace_function_set_filter(struct perf_event *event, struct event_filter *filter) argument 2186 ftrace_function_set_filter(struct perf_event *event, struct event_filter *filter) argument 2193 ftrace_profile_set_filter(struct perf_event *event, int event_id, char *filter_str) argument 2317 test_pred_visited_fn(struct filter_pred *pred, void *event) argument [all...] |
/kernel/ |
H A D | watchdog.c | 334 /* Callback function for perf event subsystem */ 335 static void watchdog_overflow_callback(struct perf_event *event, argument 340 event->hw.interrupts = 0; 512 /* Enable the perf event */ 530 /* disable the perf event */ 571 struct perf_event *event = per_cpu(watchdog_ev, cpu); local 578 event = ERR_PTR(-ENOENT); 583 if (event && event->state > PERF_EVENT_STATE_OFF) 587 if (event ! 634 struct perf_event *event = per_cpu(watchdog_ev, cpu); local [all...] |
H A D | cpu_pm.c | 28 static int cpu_pm_notify(enum cpu_pm_event event, int nr_to_call, int *nr_calls) argument 32 ret = __raw_notifier_call_chain(&cpu_pm_notifier_chain, event, NULL,
|
H A D | compat.c | 687 struct sigevent __user *event = NULL; local 692 event = compat_alloc_user_space(sizeof(*event)); 694 copy_to_user(event, &kevent, sizeof(*event))) 698 return sys_timer_create(which_clock, event, created_timer_id); 877 int get_compat_sigevent(struct sigevent *event, argument 880 memset(event, 0, sizeof(*event)); 882 __get_user(event [all...] |
/kernel/locking/ |
H A D | rtmutex-tester.c | 31 int event; member in struct:test_thread_data 51 RTTEST_RESETEVENT = 98, /* 98 Reset event counter */ 66 td->event = atomic_add_return(1, &rttest_event); 96 td->event = atomic_add_return(1, &rttest_event); 98 td->event = atomic_add_return(1, &rttest_event); 109 td->event = atomic_add_return(1, &rttest_event); 111 td->event = atomic_add_return(1, &rttest_event); 120 td->event = atomic_add_return(1, &rttest_event); 122 td->event = atomic_add_return(1, &rttest_event); 136 * This allows us to have finegrained control over the event flo [all...] |
/kernel/gcov/ |
H A D | base.c | 44 * Add new profiling data structure to list and inform event 95 * gcov_enable_events - enable event reporting through gcov_event() 109 /* Perform event callback for previously registered entries. */ 123 static int gcov_module_notifier(struct notifier_block *nb, unsigned long event, argument 130 if (event != MODULE_STATE_GOING)
|
/kernel/time/ |
H A D | posix-timers.c | 499 static struct pid *good_sigevent(sigevent_t * event) argument 503 if ((event->sigev_notify & SIGEV_THREAD_ID ) && 504 (!(rtn = find_task_by_vpid(event->sigev_notify_thread_id)) || 506 (event->sigev_notify & ~SIGEV_THREAD_ID) != SIGEV_SIGNAL)) 509 if (((event->sigev_notify & ~SIGEV_THREAD_ID) != SIGEV_NONE) && 510 ((event->sigev_signo <= 0) || (event->sigev_signo > SIGRTMAX))) 602 sigevent_t event; local 627 if (copy_from_user(&event, timer_event_spec, sizeof (event))) { [all...] |