/kernel/irq/ |
H A D | irqdesc.c | 180 static int alloc_descs(unsigned int start, unsigned int cnt, int node, argument 186 for (i = 0; i < cnt; i++) { 201 bitmap_clear(allocated_irqs, start, cnt); 290 static inline int alloc_descs(unsigned int start, unsigned int cnt, int node, argument 295 for (i = 0; i < cnt; i++) { 386 * @cnt: Number of consecutive irqs to free 388 void irq_free_descs(unsigned int from, unsigned int cnt) argument 392 if (from >= nr_irqs || (from + cnt) > nr_irqs) 395 for (i = 0; i < cnt; i++) 399 bitmap_clear(allocated_irqs, from, cnt); 415 __irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node, struct module *owner) argument 468 irq_alloc_hwirqs(int cnt, int node) argument 498 irq_free_hwirqs(unsigned int from, int cnt) argument [all...] |
/kernel/trace/ |
H A D | trace_seq.c | 316 int cnt = 0; local 338 cnt += trace_seq_putmem(s, hex, j); 340 return cnt; 389 * @cnt: The amount to copy 393 * and writes up to @cnt characters or till it reaches the end of 405 int trace_seq_to_user(struct trace_seq *s, char __user *ubuf, int cnt) argument 410 if (!cnt) 417 if (cnt > len) 418 cnt = len; 419 ret = copy_to_user(ubuf, s->buffer + s->readpos, cnt); [all...] |
H A D | trace_kdb.c | 24 int cnt = 0, cpu; local 63 if (!cnt) 65 cnt++; 77 if (!cnt)
|
H A D | trace.c | 868 size_t cnt, loff_t *ppos) 882 cnt--; 890 while (cnt && isspace(ch)) { 895 cnt--; 909 while (cnt && !isspace(ch)) { 920 cnt--; 942 static ssize_t trace_seq_to_buffer(struct trace_seq *s, void *buf, size_t cnt) argument 950 if (cnt > len) 951 cnt = len; 952 memcpy(buf, s->buffer + s->readpos, cnt); 867 trace_get_user(struct trace_parser *parser, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 3602 tracing_trace_options_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 3783 tracing_readme_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 3873 tracing_saved_cmdlines_size_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 3916 tracing_saved_cmdlines_size_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 3946 tracing_saved_tgids_read(struct file *file, char __user *ubuf, size_t cnt, loff_t *ppos) argument 3990 tracing_set_trace_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 4277 tracing_set_trace_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 4310 tracing_nsecs_read(unsigned long *ptr, char __user *ubuf, size_t cnt, loff_t *ppos) argument 4324 tracing_nsecs_write(unsigned long *ptr, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 4340 tracing_thresh_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 4347 tracing_thresh_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 4372 tracing_max_lat_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 4379 tracing_max_lat_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 4543 tracing_read_pipe(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 4783 tracing_entries_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 4831 tracing_entries_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 4859 tracing_total_entries_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 4883 tracing_free_buffer_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 4913 tracing_mark_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *fpos) argument 5066 tracing_clock_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *fpos) argument 5160 tracing_snapshot_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 5714 unsigned long cnt; local 5783 tracing_read_dyn_info(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 6028 trace_options_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 6043 trace_options_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 6080 trace_options_core_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 6095 trace_options_core_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 6192 int cnt; local 6221 int cnt; local 6260 rb_simple_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 6274 rb_simple_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 6764 int cnt = 0, cpu; local [all...] |
H A D | trace_events.c | 592 size_t cnt, loff_t *ppos) 599 if (!cnt) 609 read = trace_get_user(&parser, ubuf, cnt, ppos); 722 event_enable_read(struct file *filp, char __user *ubuf, size_t cnt, argument 748 return simple_read_from_buffer(ubuf, cnt, ppos, buf, strlen(buf)); 752 event_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, argument 759 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 782 *ppos += cnt; 784 return ret ? ret : cnt; 788 system_enable_read(struct file *filp, char __user *ubuf, size_t cnt, argument 591 ftrace_event_write(struct file *file, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 834 system_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 1005 event_id_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 1023 event_filter_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 1055 event_filter_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 1195 subsystem_filter_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 1221 subsystem_filter_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 1252 show_header(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument [all...] |
H A D | trace_mmiotrace.c | 130 unsigned long cnt = atomic_xchg(&dropped_count, 0); local 134 cnt += over - prev_overruns; 136 return cnt; 140 char __user *ubuf, size_t cnt, loff_t *ppos) 169 ret = trace_seq_to_user(s, ubuf, cnt); 139 mmio_read(struct trace_iterator *iter, struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument
|
H A D | trace_selftest.c | 63 unsigned long flags, cnt = 0; local 70 cnt = ring_buffer_entries(buf->buffer); 90 *count = cnt; 183 static int trace_selftest_ops(struct trace_array *tr, int cnt) argument 194 pr_info("Testing dynamic ftrace ops #%d: ", cnt); 219 if (cnt > 1) { 234 if (cnt > 1) { 273 if (cnt > 1) { 301 if (cnt > 1)
|
H A D | ring_buffer_benchmark.c | 223 int cnt = 0; local 249 cnt++; 250 if (consumer && !(cnt % wakeup_interval)) 263 if (cnt % wakeup_interval)
|
H A D | ftrace.c | 167 int cnt = 0; local 173 cnt++; 177 return cnt; 934 size_t cnt, loff_t *ppos) 939 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 950 cnt = ret; 956 cnt = ret; 972 *ppos += cnt; 974 return cnt; 979 size_t cnt, loff_ 933 ftrace_profile_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 978 ftrace_profile_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) argument 2549 int cnt = 0; local 2594 int cnt = ref; local 2641 int cnt; local 2681 int cnt; local 3715 ftrace_regex_write(struct file *file, const char __user *ubuf, size_t cnt, loff_t *ppos, int enable) argument 3754 ftrace_filter_write(struct file *file, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 3761 ftrace_notrace_write(struct file *file, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 4349 ftrace_graph_write(struct file *file, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 5113 ftrace_pid_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument [all...] |
H A D | trace_events_trigger.c | 228 size_t cnt, loff_t *ppos) 234 if (!cnt) 237 if (cnt >= PAGE_SIZE) 244 if (copy_from_user(buf, ubuf, cnt)) { 248 buf[cnt] = '\0'; 265 *ppos += cnt; 266 ret = cnt; 285 size_t cnt, loff_t *ppos) 287 return event_trigger_regex_write(filp, ubuf, cnt, ppos); 226 event_trigger_regex_write(struct file *file, const char __user *ubuf, size_t cnt, loff_t *ppos) argument 284 event_trigger_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument
|
H A D | trace_functions_graph.c | 1545 graph_depth_write(struct file *filp, const char __user *ubuf, size_t cnt, argument 1551 ret = kstrtoul_from_user(ubuf, cnt, 10, &val); 1557 *ppos += cnt; 1559 return cnt; 1563 graph_depth_read(struct file *filp, char __user *ubuf, size_t cnt, argument 1571 return simple_read_from_buffer(ubuf, cnt, ppos, buf, n);
|
H A D | trace_events_filter.c | 126 unsigned int cnt; member in struct:filter_parse_state::__anon58 1044 ps->infix.cnt = strlen(infix_string); 1053 ps->infix.cnt--; 1068 ps->infix.cnt--;
|
H A D | ring_buffer.c | 4711 int cnt; member in struct:rb_test_data 4739 int cnt; local 4742 cnt = data->cnt + (nested ? 27 : 0); 4744 /* Multiply cnt by ~e, to make some unique increment */ 4745 size = (data->cnt * 68 / 25) % (sizeof(rb_string) - 1); 4804 data->cnt++; 4808 usleep_range(((data->cnt % 3) + 1) * 100, 1000); 4855 rb_data[cpu].cnt = cpu;
|
H A D | trace.h | 368 size_t cnt, loff_t *ppos); 906 size_t cnt, loff_t *ppos);
|
/kernel/ |
H A D | softirq.c | 95 void __local_bh_disable_ip(unsigned long ip, unsigned int cnt) argument 109 __preempt_count_add(cnt); 113 if (softirq_count() == (cnt & SOFTIRQ_MASK)) 117 if (preempt_count() == cnt) 123 static void __local_bh_enable(unsigned int cnt) argument 127 if (softirq_count() == (cnt & SOFTIRQ_MASK)) 129 preempt_count_sub(cnt); 144 void __local_bh_enable_ip(unsigned long ip, unsigned int cnt) argument 159 preempt_count_sub(cnt - 1);
|
H A D | ksysfs.c | 117 unsigned long cnt; local 120 if (kstrtoul(buf, 0, &cnt)) 123 ret = crash_shrink_memory(cnt);
|
H A D | cpuset.c | 69 int cnt; /* unprocessed events count */ member in struct:fmeter 1359 #define FM_MAXCNT 1000000 /* limit cnt to avoid overflow */ 1365 fmp->cnt = 0; 1371 /* Internal meter update - process cnt events and update value */ 1385 fmp->val += ((FM_SCALE - FM_COEF) * fmp->cnt) / FM_SCALE; 1386 fmp->cnt = 0; 1389 /* Process any previous ticks, then bump cnt by one (times scale). */ 1394 fmp->cnt = min(FM_MAXCNT, fmp->cnt + FM_SCALE);
|
/kernel/time/ |
H A D | clocksource.c | 914 ssize_t sysfs_get_uname(const char *buf, char *dst, size_t cnt) argument 916 size_t ret = cnt; 919 if (!cnt || cnt >= CS_NAME_LEN) 923 if (buf[cnt-1] == '\n') 924 cnt--; 925 if (cnt > 0) 926 memcpy(dst, buf, cnt); 927 dst[cnt] = 0;
|
H A D | tick-internal.h | 36 extern ssize_t sysfs_get_uname(const char *buf, char *dst, size_t cnt);
|
/kernel/locking/ |
H A D | mutex.c | 951 * @cnt: the atomic which we are to dec 956 int atomic_dec_and_mutex_lock(atomic_t *cnt, struct mutex *lock) argument 959 if (atomic_add_unless(cnt, -1, 1)) 963 if (!atomic_dec_and_test(cnt)) {
|
/kernel/power/ |
H A D | snapshot.c | 1083 unsigned int cnt = 0; local 1087 cnt += zone_page_state(zone, NR_FREE_PAGES); 1089 return cnt; 2154 unsigned int cnt = 0; local 2160 cnt++; 2164 return cnt;
|
/kernel/sched/ |
H A D | core.c | 227 size_t cnt, loff_t *ppos) 234 if (cnt > 63) 235 cnt = 63; 237 if (copy_from_user(&buf, ubuf, cnt)) 240 buf[cnt] = 0; 251 *ppos += cnt; 253 return cnt; 226 sched_feat_write(struct file *filp, const char __user *ubuf, size_t cnt, loff_t *ppos) argument
|