Searched defs:read (Results 1 - 9 of 9) sorted by relevance

/kernel/time/
H A Dsched_clock.c109 void __init sched_clock_register(u64 (*read)(void), int bits, argument
133 new_epoch = read();
139 read_sched_clock = read;
169 pr_debug("Registered %pF as sched_clock source\n", read);
/kernel/trace/
H A Dring_buffer_benchmark.c33 static unsigned long read; variable
94 read++;
143 read++;
156 read++;
180 read = 0;
318 trace_printk("Read: %ld (by %s)\n", read,
321 trace_printk("Total: %lld\n", entries + overruns + read);
H A Dtrace_events.c597 ssize_t read, ret; local
609 read = trace_get_user(&parser, ubuf, cnt, ppos);
611 if (read >= 0 && trace_parser_loaded((&parser))) {
624 ret = read;
1295 .read = seq_read,
1302 .read = seq_read,
1310 .read = event_enable_read,
1317 .read = seq_read,
1323 .read = event_id_read,
1329 .read
[all...]
H A Dtrace.h347 * @read: override the default read callback on trace_pipe
366 ssize_t (*read)(struct trace_iterator *iter, member in struct:tracer
H A Dring_buffer.c57 * associated with the CPU it is currently executing on. A reader may read
61 * reader page. When a reader has read the entire reader page, this reader
356 unsigned read; /* index for next read */ member in struct:buffer_page
384 * @page: The page to read
468 struct buffer_page *head_page; /* read from head */
481 unsigned long read; member in struct:ring_buffer_per_cpu
1503 * change read pointer to make sure any read iterators reset
1506 cpu_buffer->read
4468 unsigned int read; local
[all...]
H A Dftrace.c990 .read = ftrace_profile_read,
1096 * it all the time. These are in a read only section such that if
1746 * EINVAL - if what is read at @ip is not what was expected
3720 ssize_t ret, read; local
3737 read = trace_get_user(parser, ubuf, cnt, ppos);
3739 if (read >= 0 && trace_parser_loaded(parser) &&
3748 ret = read;
4096 .read = seq_read,
4103 .read = seq_read,
4110 .read
4353 ssize_t read, ret = 0; local
[all...]
H A Dtrace.c357 * These primitives don't distinguish read-only and read-consume access.
358 * Multi read-only access are also serialized.
863 * Returns number of bytes read.
871 size_t read = 0; local
881 read++;
894 read++;
900 *ppos += read;
901 ret = read;
908 /* read th
5115 unsigned int read; member in struct:ftrace_buffer_info
[all...]
/kernel/
H A Dprofile.c455 .read = seq_read,
477 ssize_t read; local
486 read = 0;
491 buf++; p++; count--; read++;
496 read += count;
497 *ppos += read;
498 return read;
529 .read = read_profile,
/kernel/locking/
H A Dlockdep.c268 if (hlock->read)
1615 * bit 0 - write/read
1624 * keep state, bit flip the direction and strip read.
1645 * Prove that the new dependency does not connect a hardirq-safe-read
1752 * Returns: 0 on deadlock detected, 1 on OK, 2 on recursive read
1756 struct lockdep_map *next_instance, int read)
1772 * Allow read-after-read recursion of the same
1775 if ((read == 2) && prev->read)
1755 check_deadlock(struct task_struct *curr, struct held_lock *next, struct lockdep_map *next_instance, int read) argument
2450 int read = new_bit & 1; local
3047 __lock_acquire(struct lockdep_map *lock, unsigned int subclass, int trylock, int read, int check, int hardirqs_off, struct lockdep_map *nest_lock, unsigned long ip, int references) argument
3588 lock_acquire(struct lockdep_map *lock, unsigned int subclass, int trylock, int read, int check, struct lockdep_map *nest_lock, unsigned long ip) argument
[all...]

Completed in 201 milliseconds