Searched refs:event (Results 1 - 25 of 33) sorted by relevance

12

/kernel/events/
H A Dcore.c127 static bool is_kernel_event(struct perf_event *event) argument
129 return event->owner == EVENT_OWNER_KERNEL;
168 * perf event paranoia level:
180 * max perf event sample rate
316 static u64 perf_event_time(struct perf_event *event);
371 * if there is no cgroup event for the current CPU context.
381 perf_cgroup_match(struct perf_event *event) argument
383 struct perf_event_context *ctx = event->ctx;
386 /* @event doesn't care about cgroup */
387 if (!event
404 perf_detach_cgroup(struct perf_event *event) argument
410 is_cgroup_event(struct perf_event *event) argument
415 perf_cgroup_event_time(struct perf_event *event) argument
443 update_cgrp_time_from_event(struct perf_event *event) argument
605 perf_cgroup_connect(int fd, struct perf_event *event, struct perf_event_attr *attr, struct perf_event *group_leader) argument
642 perf_cgroup_set_shadow_time(struct perf_event *event, u64 now) argument
650 perf_cgroup_defer_enabled(struct perf_event *event) argument
663 perf_cgroup_mark_enabled(struct perf_event *event, struct perf_event_context *ctx) argument
685 perf_cgroup_match(struct perf_event *event) argument
690 perf_detach_cgroup(struct perf_event *event) argument
693 is_cgroup_event(struct perf_event *event) argument
698 perf_cgroup_event_cgrp_time(struct perf_event *event) argument
703 update_cgrp_time_from_event(struct perf_event *event) argument
721 perf_cgroup_connect(pid_t pid, struct perf_event *event, struct perf_event_attr *attr, struct perf_event *group_leader) argument
740 perf_cgroup_set_shadow_time(struct perf_event *event, u64 now) argument
744 perf_cgroup_event_time(struct perf_event *event) argument
750 perf_cgroup_defer_enabled(struct perf_event *event) argument
755 perf_cgroup_mark_enabled(struct perf_event *event, struct perf_event_context *ctx) argument
928 perf_event_pid(struct perf_event *event, struct task_struct *p) argument
939 perf_event_tid(struct perf_event *event, struct task_struct *p) argument
954 primary_event_id(struct perf_event *event) argument
1055 perf_event_time(struct perf_event *event) argument
1069 update_event_times(struct perf_event *event) argument
1110 struct perf_event *event; local
1118 ctx_group_list(struct perf_event *event, struct perf_event_context *ctx) argument
1131 list_add_event(struct perf_event *event, struct perf_event_context *ctx) argument
1170 perf_event__state_init(struct perf_event *event) argument
1180 perf_event__read_size(struct perf_event *event) argument
1204 perf_event__header_size(struct perf_event *event) argument
1236 perf_event__id_header_size(struct perf_event *event) argument
1263 perf_group_attach(struct perf_event *event) argument
1296 list_del_event(struct perf_event *event, struct perf_event_context *ctx) argument
1346 perf_group_detach(struct perf_event *event) argument
1395 is_orphaned_event(struct perf_event *event) argument
1404 is_orphaned_child(struct perf_event *event) argument
1432 event_filter_match(struct perf_event *event) argument
1439 event_sched_out(struct perf_event *event, struct perf_cpu_context *cpuctx, struct perf_event_context *ctx) argument
1491 struct perf_event *event; local
1507 struct perf_event *event; member in struct:remove_event
1520 struct perf_event *event = re->event; local
1552 perf_remove_from_context(struct perf_event *event, bool detach_group) argument
1608 struct perf_event *event = info; local
1657 perf_event_disable(struct perf_event *event) argument
1700 perf_set_shadow_time(struct perf_event *event, struct perf_event_context *ctx, u64 tstamp) argument
1740 event_sched_in(struct perf_event *event, struct perf_cpu_context *cpuctx, struct perf_event_context *ctx) argument
1806 struct perf_event *event, *partial_group = NULL; local
1873 group_can_go_on(struct perf_event *event, struct perf_cpu_context *cpuctx, int can_add_hw) argument
1901 add_event_to_ctx(struct perf_event *event, struct perf_event_context *ctx) argument
1939 struct perf_event *event = info; local
2004 perf_install_in_context(struct perf_event_context *ctx, struct perf_event *event, int cpu) argument
2060 __perf_event_mark_enabled(struct perf_event *event) argument
2078 struct perf_event *event = info; local
2161 perf_event_enable(struct perf_event *event) argument
2219 perf_event_refresh(struct perf_event *event, int refresh) argument
2238 struct perf_event *event; local
2301 __perf_event_sync_stat(struct perf_event *event, struct perf_event *next_event) argument
2350 struct perf_event *event, *next_event; local
2498 struct perf_event *event; local
2528 struct perf_event *event; local
2718 perf_calculate_period(struct perf_event *event, u64 nsec, u64 count) argument
2794 perf_adjust_period(struct perf_event *event, u64 nsec, u64 count, bool disable) argument
2831 struct perf_event *event; local
2989 event_enable_on_exec(struct perf_event *event, struct perf_event_context *ctx) argument
3011 struct perf_event *event; local
3078 struct perf_event *event = info; local
3103 perf_event_count(struct perf_event *event) argument
3108 perf_event_read(struct perf_event *event) argument
3289 struct perf_event *event; local
3302 unaccount_event_cpu(struct perf_event *event, int cpu) argument
3315 unaccount_event(struct perf_event *event) argument
3338 __free_event(struct perf_event *event) argument
3357 _free_event(struct perf_event *event) argument
3385 free_event(struct perf_event *event) argument
3400 perf_remove_from_owner(struct perf_event *event) argument
3441 put_event(struct perf_event *event) argument
3471 perf_event_release_kernel(struct perf_event *event) argument
3490 struct perf_event *event, *tmp; local
3520 perf_event_read_value(struct perf_event *event, u64 *enabled, u64 *running) argument
3546 perf_event_read_group(struct perf_event *event, u64 read_format, char __user *buf) argument
3596 perf_event_read_one(struct perf_event *event, u64 read_format, char __user *buf) argument
3617 is_event_hup(struct perf_event *event) argument
3634 perf_read_hw(struct perf_event *event, char __user *buf, size_t count) argument
3662 struct perf_event *event = file->private_data; local
3669 struct perf_event *event = file->private_data; local
3690 perf_event_reset(struct perf_event *event) argument
3703 perf_event_for_each_child(struct perf_event *event, void (*func)(struct perf_event *)) argument
3716 perf_event_for_each(struct perf_event *event, void (*func)(struct perf_event *)) argument
3732 perf_event_period(struct perf_event *event, u64 __user *arg) argument
3801 struct perf_event *event = file->private_data; local
3886 struct perf_event *event; local
3898 struct perf_event *event; local
3908 perf_event_index(struct perf_event *event) argument
3919 calc_timer_values(struct perf_event *event, u64 *now, u64 *enabled, u64 *running) argument
3932 perf_event_init_userpage(struct perf_event *event) argument
3961 perf_event_update_userpage(struct perf_event *event) argument
4013 struct perf_event *event = vma->vm_file->private_data; local
4046 ring_buffer_attach(struct perf_event *event, struct ring_buffer *rb) argument
4092 ring_buffer_wakeup(struct perf_event *event) argument
4113 ring_buffer_get(struct perf_event *event) argument
4140 struct perf_event *event = vma->vm_file->private_data; local
4156 struct perf_event *event = vma->vm_file->private_data; local
4243 struct perf_event *event = file->private_data; local
4370 struct perf_event *event = filp->private_data; local
4401 perf_event_wakeup(struct perf_event *event) argument
4413 struct perf_event *event = container_of(entry, local
4574 __perf_event_header__init_id(struct perf_event_header *header, struct perf_sample_data *data, struct perf_event *event) argument
4604 perf_event_header__init_id(struct perf_event_header *header, struct perf_sample_data *data, struct perf_event *event) argument
4636 perf_event__output_id_sample(struct perf_event *event, struct perf_output_handle *handle, struct perf_sample_data *sample) argument
4644 perf_output_read_one(struct perf_output_handle *handle, struct perf_event *event, u64 enabled, u64 running) argument
4670 perf_output_read_group(struct perf_output_handle *handle, struct perf_event *event, u64 enabled, u64 running) argument
4714 perf_output_read(struct perf_output_handle *handle, struct perf_event *event) argument
4738 perf_output_sample(struct perf_output_handle *handle, struct perf_event_header *header, struct perf_sample_data *data, struct perf_event *event) argument
4875 perf_prepare_sample(struct perf_event_header *header, struct perf_sample_data *data, struct perf_event *event, struct pt_regs *regs) argument
4969 perf_event_output(struct perf_event *event, struct perf_sample_data *data, struct pt_regs *regs) argument
5004 perf_event_read_event(struct perf_event *event, struct task_struct *task) argument
5039 struct perf_event *event; local
5106 perf_event_task_match(struct perf_event *event) argument
5113 perf_event_task_output(struct perf_event *event, void *data) argument
5202 perf_event_comm_match(struct perf_event *event) argument
5207 perf_event_comm_output(struct perf_event *event, void *data) argument
5309 perf_event_mmap_match(struct perf_event *event, void *data) argument
5320 perf_event_mmap_output(struct perf_event *event, void *data) argument
5530 perf_log_throttle(struct perf_event *event, int enable) argument
5571 __perf_event_overflow(struct perf_event *event, int throttle, struct perf_sample_data *data, struct pt_regs *regs) argument
5639 perf_event_overflow(struct perf_event *event, struct perf_sample_data *data, struct pt_regs *regs) argument
5671 perf_swevent_set_period(struct perf_event *event) argument
5694 perf_swevent_overflow(struct perf_event *event, u64 overflow, struct perf_sample_data *data, struct pt_regs *regs) argument
5720 perf_swevent_event(struct perf_event *event, u64 nr, struct perf_sample_data *data, struct pt_regs *regs) argument
5749 perf_exclude_event(struct perf_event *event, struct pt_regs *regs) argument
5766 perf_swevent_match(struct perf_event *event, enum perf_type_id type, u32 event_id, struct perf_sample_data *data, struct pt_regs *regs) argument
5814 find_swevent_head(struct swevent_htable *swhash, struct perf_event *event) argument
5839 struct perf_event *event; local
5888 perf_swevent_read(struct perf_event *event) argument
5892 perf_swevent_add(struct perf_event *event, int flags) argument
5920 perf_swevent_del(struct perf_event *event, int flags) argument
5925 perf_swevent_start(struct perf_event *event, int flags) argument
5930 perf_swevent_stop(struct perf_event *event, int flags) argument
5954 swevent_hlist_put_cpu(struct perf_event *event, int cpu) argument
5966 swevent_hlist_put(struct perf_event *event) argument
5974 swevent_hlist_get_cpu(struct perf_event *event, int cpu) argument
5998 swevent_hlist_get(struct perf_event *event) argument
6027 sw_perf_event_destroy(struct perf_event *event) argument
6037 perf_swevent_init(struct perf_event *event) argument
6089 perf_tp_filter_match(struct perf_event *event, struct perf_sample_data *data) argument
6099 perf_tp_event_match(struct perf_event *event, struct perf_sample_data *data, struct pt_regs *regs) argument
6122 struct perf_event *event; local
6166 tp_perf_event_destroy(struct perf_event *event) argument
6171 perf_tp_event_init(struct perf_event *event) argument
6209 perf_event_set_filter(struct perf_event *event, void __user *arg) argument
6227 perf_event_free_filter(struct perf_event *event) argument
6238 perf_event_set_filter(struct perf_event *event, void __user *arg) argument
6243 perf_event_free_filter(struct perf_event *event) argument
6271 struct perf_event *event; local
6296 perf_swevent_start_hrtimer(struct perf_event *event) argument
6318 perf_swevent_cancel_hrtimer(struct perf_event *event) argument
6330 perf_swevent_init_hrtimer(struct perf_event *event) argument
6359 cpu_clock_event_update(struct perf_event *event) argument
6369 cpu_clock_event_start(struct perf_event *event, int flags) argument
6375 cpu_clock_event_stop(struct perf_event *event, int flags) argument
6381 cpu_clock_event_add(struct perf_event *event, int flags) argument
6389 cpu_clock_event_del(struct perf_event *event, int flags) argument
6394 cpu_clock_event_read(struct perf_event *event) argument
6399 cpu_clock_event_init(struct perf_event *event) argument
6433 task_clock_event_update(struct perf_event *event, u64 now) argument
6443 task_clock_event_start(struct perf_event *event, int flags) argument
6449 task_clock_event_stop(struct perf_event *event, int flags) argument
6455 task_clock_event_add(struct perf_event *event, int flags) argument
6463 task_clock_event_del(struct perf_event *event, int flags) argument
6468 task_clock_event_read(struct perf_event *event) argument
6477 task_clock_event_init(struct perf_event *event) argument
6532 perf_event_idx_default(struct perf_event *event) argument
6820 perf_init_event(struct perf_event *event) argument
6865 account_event_cpu(struct perf_event *event, int cpu) argument
6878 account_event(struct perf_event *event) argument
6915 struct perf_event *event; local
7164 perf_event_set_output(struct perf_event *event, struct perf_event *output_event) argument
7224 struct perf_event *event, *sibling; local
7506 struct perf_event *event; local
7550 struct perf_event *event, *tmp; local
7728 struct perf_event *event, *tmp; local
7750 perf_free_event(struct perf_event *event, struct perf_event_context *ctx) argument
7776 struct perf_event *event, *tmp; local
7925 inherit_task_group(struct perf_event *event, struct task_struct *parent, struct perf_event_context *parent_ctx, struct task_struct *child, int ctxn, int *inherited_all) argument
7970 struct perf_event *event; local
[all...]
H A Dring_buffer.c23 handle->event->pending_wakeup = 1;
24 irq_work_queue(&handle->event->pending);
29 * event isn't done writing. However since we need to deal with NMIs we
33 * event completes.
105 struct perf_event *event, unsigned int size)
120 if (event->parent)
121 event = event->parent;
123 rb = rcu_dereference(event->rb);
131 handle->event
104 perf_output_begin(struct perf_output_handle *handle, struct perf_event *event, unsigned int size) argument
[all...]
H A Dcallchain.c160 perf_callchain(struct perf_event *event, struct pt_regs *regs) argument
165 int kernel = !event->attr.exclude_callchain_kernel;
166 int user = !event->attr.exclude_callchain_user;
197 if (event->ctx->task && event->ctx->task != current)
H A Dinternal.h25 local_t events; /* event limit */
45 extern void perf_event_wakeup(struct perf_event *event);
50 struct perf_event *event);
52 perf_event__output_id_sample(struct perf_event *event,
150 perf_callchain(struct perf_event *event, struct pt_regs *regs);
/kernel/trace/
H A Dring_buffer_benchmark.c80 struct ring_buffer_event *event; local
84 event = ring_buffer_consume(buffer, cpu, &ts, NULL);
85 if (!event)
88 entry = ring_buffer_event_data(event);
100 struct ring_buffer_event *event; local
116 /* The commit may have missed event flags set, clear them */
126 event = (void *)&rpage->data[i];
127 switch (event->type_len) {
130 if (!event->time_delta)
132 inc = event
232 struct ring_buffer_event *event; local
[all...]
H A Dtrace_output.c192 struct ftrace_event_call *event; local
198 event = container_of(trace_event, struct ftrace_event_call, event);
201 if (entry->type != event->event.type) {
207 ret = trace_seq_printf(s, "%s: ", ftrace_event_name(event));
608 * ftrace_find_event - find a registered event
609 * @type: the type of event to look for
611 * Returns an event of type @type otherwise NULL
616 struct trace_event *event; local
684 register_ftrace_event(struct trace_event *event) argument
753 __unregister_ftrace_event(struct trace_event *event) argument
764 unregister_ftrace_event(struct trace_event *event) argument
778 trace_nop_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
788 trace_fn_trace(struct trace_iterator *iter, int flags, struct trace_event *event) argument
816 trace_fn_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument
831 trace_fn_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument
845 trace_fn_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument
872 trace_graph_ent_trace(struct trace_iterator *iter, int flags, struct trace_event *event) argument
892 trace_graph_ent_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument
907 trace_graph_ent_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument
921 trace_graph_ent_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument
948 trace_graph_ret_trace(struct trace_iterator *iter, int flags, struct trace_event *event) argument
969 trace_graph_ret_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument
987 trace_graph_ret_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1004 trace_graph_ret_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1061 trace_ctx_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1067 trace_wake_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1096 trace_ctx_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1102 trace_wake_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1132 trace_ctx_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1138 trace_wake_hex(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1144 trace_ctxwake_bin(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1188 trace_stack_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1228 trace_user_stack_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1259 trace_bputs_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1285 trace_bputs_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1317 trace_bprint_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1343 trace_bprint_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1374 trace_print_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1394 trace_print_raw(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1437 struct trace_event *event; local
[all...]
H A Dtrace_output.h30 int flags, struct trace_event *event);
35 extern int __unregister_ftrace_event(struct trace_event *event);
H A Dring_buffer.c202 #define skip_time_extend(event) \
203 ((struct ring_buffer_event *)((char *)event + RB_LEN_TIME_EXTEND))
205 static inline int rb_null_event(struct ring_buffer_event *event) argument
207 return event->type_len == RINGBUF_TYPE_PADDING && !event->time_delta;
210 static void rb_event_set_padding(struct ring_buffer_event *event) argument
213 event->type_len = RINGBUF_TYPE_PADDING;
214 event->time_delta = 0;
218 rb_event_data_length(struct ring_buffer_event *event) argument
222 if (event
235 rb_event_length(struct ring_buffer_event *event) argument
264 rb_event_ts_length(struct ring_buffer_event *event) argument
286 ring_buffer_event_length(struct ring_buffer_event *event) argument
305 rb_event_data(struct ring_buffer_event *event) argument
321 ring_buffer_event_data(struct ring_buffer_event *event) argument
1892 rb_event_index(struct ring_buffer_event *event) argument
1900 rb_event_is_commit(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument
1993 rb_add_time_stamp(struct ring_buffer_event *event, u64 delta) argument
2022 rb_update_event(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event, unsigned length, int add_timestamp, u64 delta) argument
2209 struct ring_buffer_event event; /* Used only for sizeof array */ local
2229 struct ring_buffer_event *event; local
2409 struct ring_buffer_event *event; local
2461 rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument
2543 struct ring_buffer_event *event; local
2725 struct ring_buffer_event *event; local
2769 rb_update_write_stamp(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument
2796 rb_commit(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument
2829 ring_buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event) argument
2849 rb_event_discard(struct ring_buffer_event *event) argument
2869 rb_decrement_entry(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument
2921 ring_buffer_discard_commit(struct ring_buffer *buffer, struct ring_buffer_event *event) argument
2977 struct ring_buffer_event *event; local
3436 rb_update_read_stamp(struct ring_buffer_per_cpu *cpu_buffer, struct ring_buffer_event *event) argument
3467 rb_update_iter_read_stamp(struct ring_buffer_iter *iter, struct ring_buffer_event *event) argument
3628 struct ring_buffer_event *event; local
3652 struct ring_buffer_event *event; local
3700 struct ring_buffer_event *event; local
3767 struct ring_buffer_event *event; local
3872 struct ring_buffer_event *event; local
3909 struct ring_buffer_event *event; local
3939 struct ring_buffer_event *event = NULL; local
4104 struct ring_buffer_event *event; local
4462 struct ring_buffer_event *event; local
4733 struct ring_buffer_event *event; local
4909 struct ring_buffer_event *event; local
[all...]
H A Dtrace_sched_switch.c32 struct ring_buffer_event *event; local
35 event = trace_buffer_lock_reserve(buffer, TRACE_CTX,
37 if (!event)
39 entry = ring_buffer_event_data(event);
48 if (!call_filter_check_discard(call, entry, buffer, event))
49 trace_buffer_unlock_commit(buffer, event, flags, pc);
87 struct ring_buffer_event *event; local
91 event = trace_buffer_lock_reserve(buffer, TRACE_WAKE,
93 if (!event)
95 entry = ring_buffer_event_data(event);
[all...]
H A Dtrace_syscalls.c16 static int syscall_enter_register(struct ftrace_event_call *event,
18 static int syscall_exit_register(struct ftrace_event_call *event,
111 struct trace_event *event)
126 if (entry->enter_event->event.type != ent->type) {
164 struct trace_event *event)
182 if (entry->exit_event->event.type != ent->type) {
308 struct ring_buffer_event *event; local
337 event = trace_buffer_lock_reserve(buffer,
338 sys_data->enter_event->event.type, size, irq_flags, pc);
339 if (!event)
110 print_syscall_enter(struct trace_iterator *iter, int flags, struct trace_event *event) argument
163 print_syscall_exit(struct trace_iterator *iter, int flags, struct trace_event *event) argument
356 struct ring_buffer_event *event; local
708 syscall_enter_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument
736 syscall_exit_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument
[all...]
H A Dtrace_uprobe.c50 * uprobe event core functions
239 alloc_trace_uprobe(const char *group, const char *event, int nargs, bool is_ret) argument
243 if (!event || !is_good_name(event))
254 tu->tp.call.name = kstrdup(event, GFP_KERNEL);
291 static struct trace_uprobe *find_probe_event(const char *event, const char *group) argument
296 if (strcmp(ftrace_event_name(&tu->tp.call), event) == 0 &&
325 /* register as an event */
329 /* delete old event */
337 pr_warning("Failed to register probe event(
359 char *arg, *event, *group, *filename; local
776 struct ring_buffer_event *event; local
843 print_uprobe_event(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1006 struct perf_event *event; local
1020 uprobe_filter_event(struct trace_uprobe *tu, struct perf_event *event) argument
1025 uprobe_perf_close(struct trace_uprobe *tu, struct perf_event *event) argument
1047 uprobe_perf_open(struct trace_uprobe *tu, struct perf_event *event) argument
1165 trace_uprobe_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument
[all...]
H A Dtrace_mmiotrace.c313 struct ring_buffer_event *event; local
317 event = trace_buffer_lock_reserve(buffer, TRACE_MMIO_RW,
319 if (!event) {
323 entry = ring_buffer_event_data(event);
326 if (!call_filter_check_discard(call, entry, buffer, event))
327 trace_buffer_unlock_commit(buffer, event, 0, pc);
343 struct ring_buffer_event *event; local
347 event = trace_buffer_lock_reserve(buffer, TRACE_MMIO_MAP,
349 if (!event) {
353 entry = ring_buffer_event_data(event);
[all...]
H A Dtrace_event_perf.c2 * trace event based perf event profiling/tracing
21 /* Count the events in use (per event id, not per instance) */
52 * event, due to issues with page faults while tracing page
215 if (tp_event->event.type == event_id &&
331 static int perf_ftrace_function_register(struct perf_event *event) argument
333 struct ftrace_ops *ops = &event->ftrace_ops;
340 static int perf_ftrace_function_unregister(struct perf_event *event) argument
342 struct ftrace_ops *ops = &event->ftrace_ops;
348 static void perf_ftrace_function_enable(struct perf_event *event) argument
353 perf_ftrace_function_disable(struct perf_event *event) argument
[all...]
H A Dtrace_kprobe.c28 * Kprobe event core functions
276 const char *event,
303 if (!event || !is_good_name(event)) {
309 tk->tp.call.name = kstrdup(event, GFP_KERNEL);
345 static struct trace_kprobe *find_trace_kprobe(const char *event, argument
351 if (strcmp(ftrace_event_name(&tk->tp.call), event) == 0 &&
503 /* Enabled event can not be unregistered */
525 /* Delete old (same name) event if exist */
535 /* Register new event */
275 alloc_trace_kprobe(const char *group, const char *event, void *addr, const char *symbol, unsigned long offs, int nargs, bool is_return) argument
612 char *symbol = NULL, *event = NULL, *group = NULL; local
922 struct ring_buffer_event *event; local
970 struct ring_buffer_event *event; local
1015 print_kprobe_event(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1051 print_kretprobe_event(struct trace_iterator *iter, int flags, struct trace_event *event) argument
1210 kprobe_register(struct ftrace_event_call *event, enum trace_reg type, void *data) argument
[all...]
H A Dtrace_branch.c36 struct ring_buffer_event *event; local
61 event = trace_buffer_lock_reserve(buffer, TRACE_BRANCH,
63 if (!event)
66 entry = ring_buffer_event_data(event);
81 if (!call_filter_check_discard(call, entry, buffer, event))
82 __buffer_unlock_commit(buffer, event);
148 int flags, struct trace_event *event)
147 trace_branch_print(struct trace_iterator *iter, int flags, struct trace_event *event) argument
H A Dtrace_events.c2 * event tracer
185 id = register_ftrace_event(&call->event);
203 fbuffer->event =
205 event_call->event.type, len,
207 if (!fbuffer->event)
210 fbuffer->entry = ring_buffer_event_data(fbuffer->event);
218 fbuffer->event, fbuffer->entry,
295 * to clear the SOFT_DISABLED flag but leave the event in the
296 * state that it was. That is, if the event was enabled and
298 * is set we do not want the event t
483 __ftrace_set_clr_event_nolock(struct trace_array *tr, const char *match, const char *sub, const char *event, int set) argument
521 __ftrace_set_clr_event(struct trace_array *tr, const char *match, const char *sub, const char *event, int set) argument
535 char *event = NULL, *sub = NULL, *match; local
576 trace_set_clr_event(const char *system, const char *event, int set) argument
1903 find_event_file(struct trace_array *tr, const char *system, const char *event) argument
2066 const char *event; local
2659 struct ring_buffer_event *event; local
[all...]
H A Dtrace.c84 * tracing is active, only save the comm when a trace event
252 struct ring_buffer_event *event)
256 ring_buffer_discard_commit(buffer, event);
266 struct ring_buffer_event *event)
270 ring_buffer_discard_commit(buffer, event);
464 struct ring_buffer_event *event; local
483 event = trace_buffer_lock_reserve(buffer, TRACE_PRINT, alloc,
485 if (!event)
488 entry = ring_buffer_event_data(event);
500 __buffer_unlock_commit(buffer, event);
250 filter_check_discard(struct ftrace_event_file *file, void *rec, struct ring_buffer *buffer, struct ring_buffer_event *event) argument
264 call_filter_check_discard(struct ftrace_event_call *call, void *rec, struct ring_buffer *buffer, struct ring_buffer_event *event) argument
514 struct ring_buffer_event *event; local
1623 struct ring_buffer_event *event; local
1637 __buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event) argument
1644 __trace_buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event, unsigned long flags, int pc) argument
1654 trace_buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event, unsigned long flags, int pc) argument
1701 trace_current_buffer_unlock_commit(struct ring_buffer *buffer, struct ring_buffer_event *event, unsigned long flags, int pc) argument
1709 trace_buffer_unlock_commit_regs(struct ring_buffer *buffer, struct ring_buffer_event *event, unsigned long flags, int pc, struct pt_regs *regs) argument
1721 trace_current_buffer_discard_commit(struct ring_buffer *buffer, struct ring_buffer_event *event) argument
1735 struct ring_buffer_event *event; local
1769 struct ring_buffer_event *event; local
1897 struct ring_buffer_event *event; local
2102 struct ring_buffer_event *event; local
2160 struct ring_buffer_event *event; local
2263 struct ring_buffer_event *event; local
2394 struct ring_buffer_event *event; local
2683 struct trace_event *event; local
2716 struct trace_event *event; local
2743 struct trace_event *event; local
2769 struct trace_event *event; local
4918 struct ring_buffer_event *event; local
6676 trace_panic_handler(struct notifier_block *this, unsigned long event, void *unused) argument
[all...]
H A Dtrace_functions_graph.c288 struct ring_buffer_event *event; local
295 event = trace_buffer_lock_reserve(buffer, TRACE_GRAPH_ENT,
297 if (!event)
299 entry = ring_buffer_event_data(event);
301 if (!call_filter_check_discard(call, entry, buffer, event))
302 __buffer_unlock_commit(buffer, event);
403 struct ring_buffer_event *event; local
410 event = trace_buffer_lock_reserve(buffer, TRACE_GRAPH_RET,
412 if (!event)
414 entry = ring_buffer_event_data(event);
625 struct ring_buffer_event *event; local
1248 struct trace_event *event; local
[all...]
H A Dtrace_events_filter.c2 * trace_events_filter - generic event filtering
143 static int filter_pred_##type(struct filter_pred *pred, void *event) \
145 type *addr = (type *)(event + pred->offset); \
173 static int filter_pred_##size(struct filter_pred *pred, void *event) \
175 u##size *addr = (u##size *)(event + pred->offset); \
199 static int filter_pred_string(struct filter_pred *pred, void *event) argument
201 char *addr = (char *)(event + pred->offset);
212 static int filter_pred_pchar(struct filter_pred *pred, void *event) argument
214 char **addr = (char **)(event + pred->offset);
235 static int filter_pred_strloc(struct filter_pred *pred, void *event) argument
250 filter_pred_none(struct filter_pred *pred, void *event) argument
2043 ftrace_profile_free_filter(struct perf_event *event) argument
2173 ftrace_function_set_filter(struct perf_event *event, struct event_filter *filter) argument
2186 ftrace_function_set_filter(struct perf_event *event, struct event_filter *filter) argument
2193 ftrace_profile_set_filter(struct perf_event *event, int event_id, char *filter_str) argument
2317 test_pred_visited_fn(struct filter_pred *pred, void *event) argument
[all...]
/kernel/
H A Dwatchdog.c334 /* Callback function for perf event subsystem */
335 static void watchdog_overflow_callback(struct perf_event *event, argument
340 event->hw.interrupts = 0;
512 /* Enable the perf event */
530 /* disable the perf event */
571 struct perf_event *event = per_cpu(watchdog_ev, cpu); local
578 event = ERR_PTR(-ENOENT);
583 if (event && event->state > PERF_EVENT_STATE_OFF)
587 if (event !
634 struct perf_event *event = per_cpu(watchdog_ev, cpu); local
[all...]
H A Dcpu_pm.c28 static int cpu_pm_notify(enum cpu_pm_event event, int nr_to_call, int *nr_calls) argument
32 ret = __raw_notifier_call_chain(&cpu_pm_notifier_chain, event, NULL,
H A Dcompat.c687 struct sigevent __user *event = NULL; local
692 event = compat_alloc_user_space(sizeof(*event));
694 copy_to_user(event, &kevent, sizeof(*event)))
698 return sys_timer_create(which_clock, event, created_timer_id);
877 int get_compat_sigevent(struct sigevent *event, argument
880 memset(event, 0, sizeof(*event));
882 __get_user(event
[all...]
/kernel/locking/
H A Drtmutex-tester.c31 int event; member in struct:test_thread_data
51 RTTEST_RESETEVENT = 98, /* 98 Reset event counter */
66 td->event = atomic_add_return(1, &rttest_event);
96 td->event = atomic_add_return(1, &rttest_event);
98 td->event = atomic_add_return(1, &rttest_event);
109 td->event = atomic_add_return(1, &rttest_event);
111 td->event = atomic_add_return(1, &rttest_event);
120 td->event = atomic_add_return(1, &rttest_event);
122 td->event = atomic_add_return(1, &rttest_event);
136 * This allows us to have finegrained control over the event flo
[all...]
/kernel/gcov/
H A Dbase.c44 * Add new profiling data structure to list and inform event
95 * gcov_enable_events - enable event reporting through gcov_event()
109 /* Perform event callback for previously registered entries. */
123 static int gcov_module_notifier(struct notifier_block *nb, unsigned long event, argument
130 if (event != MODULE_STATE_GOING)
/kernel/time/
H A Dposix-timers.c499 static struct pid *good_sigevent(sigevent_t * event) argument
503 if ((event->sigev_notify & SIGEV_THREAD_ID ) &&
504 (!(rtn = find_task_by_vpid(event->sigev_notify_thread_id)) ||
506 (event->sigev_notify & ~SIGEV_THREAD_ID) != SIGEV_SIGNAL))
509 if (((event->sigev_notify & ~SIGEV_THREAD_ID) != SIGEV_NONE) &&
510 ((event->sigev_signo <= 0) || (event->sigev_signo > SIGRTMAX)))
602 sigevent_t event; local
627 if (copy_from_user(&event, timer_event_spec, sizeof (event))) {
[all...]

Completed in 4585 milliseconds

12