/kernel/time/ |
H A D | sched_clock.c | 109 void __init sched_clock_register(u64 (*read)(void), int bits, argument 133 new_epoch = read(); 139 read_sched_clock = read; 169 pr_debug("Registered %pF as sched_clock source\n", read);
|
/kernel/trace/ |
H A D | ring_buffer_benchmark.c | 33 static unsigned long read; variable 94 read++; 143 read++; 156 read++; 180 read = 0; 318 trace_printk("Read: %ld (by %s)\n", read, 321 trace_printk("Total: %lld\n", entries + overruns + read);
|
H A D | trace_events.c | 597 ssize_t read, ret; local 609 read = trace_get_user(&parser, ubuf, cnt, ppos); 611 if (read >= 0 && trace_parser_loaded((&parser))) { 624 ret = read; 1295 .read = seq_read, 1302 .read = seq_read, 1310 .read = event_enable_read, 1317 .read = seq_read, 1323 .read = event_id_read, 1329 .read [all...] |
H A D | trace.h | 347 * @read: override the default read callback on trace_pipe 366 ssize_t (*read)(struct trace_iterator *iter, member in struct:tracer
|
H A D | ring_buffer.c | 57 * associated with the CPU it is currently executing on. A reader may read 61 * reader page. When a reader has read the entire reader page, this reader 356 unsigned read; /* index for next read */ member in struct:buffer_page 384 * @page: The page to read 468 struct buffer_page *head_page; /* read from head */ 481 unsigned long read; member in struct:ring_buffer_per_cpu 1503 * change read pointer to make sure any read iterators reset 1506 cpu_buffer->read 4468 unsigned int read; local [all...] |
H A D | ftrace.c | 990 .read = ftrace_profile_read, 1096 * it all the time. These are in a read only section such that if 1746 * EINVAL - if what is read at @ip is not what was expected 3720 ssize_t ret, read; local 3737 read = trace_get_user(parser, ubuf, cnt, ppos); 3739 if (read >= 0 && trace_parser_loaded(parser) && 3748 ret = read; 4096 .read = seq_read, 4103 .read = seq_read, 4110 .read 4353 ssize_t read, ret = 0; local [all...] |
H A D | trace.c | 357 * These primitives don't distinguish read-only and read-consume access. 358 * Multi read-only access are also serialized. 863 * Returns number of bytes read. 871 size_t read = 0; local 881 read++; 894 read++; 900 *ppos += read; 901 ret = read; 908 /* read th 5115 unsigned int read; member in struct:ftrace_buffer_info [all...] |
/kernel/ |
H A D | profile.c | 455 .read = seq_read, 477 ssize_t read; local 486 read = 0; 491 buf++; p++; count--; read++; 496 read += count; 497 *ppos += read; 498 return read; 529 .read = read_profile,
|
/kernel/locking/ |
H A D | lockdep.c | 268 if (hlock->read) 1615 * bit 0 - write/read 1624 * keep state, bit flip the direction and strip read. 1645 * Prove that the new dependency does not connect a hardirq-safe-read 1752 * Returns: 0 on deadlock detected, 1 on OK, 2 on recursive read 1756 struct lockdep_map *next_instance, int read) 1772 * Allow read-after-read recursion of the same 1775 if ((read == 2) && prev->read) 1755 check_deadlock(struct task_struct *curr, struct held_lock *next, struct lockdep_map *next_instance, int read) argument 2450 int read = new_bit & 1; local 3047 __lock_acquire(struct lockdep_map *lock, unsigned int subclass, int trylock, int read, int check, int hardirqs_off, struct lockdep_map *nest_lock, unsigned long ip, int references) argument 3588 lock_acquire(struct lockdep_map *lock, unsigned int subclass, int trylock, int read, int check, struct lockdep_map *nest_lock, unsigned long ip) argument [all...] |