/kernel/sched/ |
H A D | cpupri.h | 14 atomic_t count; member in struct:cpupri_vec
|
/kernel/events/ |
H A D | callchain.c | 100 int count; local 104 count = atomic_inc_return(&nr_callchain_events); 105 if (WARN_ON_ONCE(count < 1)) { 110 if (count > 1) {
|
H A D | hw_breakpoint.c | 121 int count = 0; local 127 count += hw_breakpoint_weight(iter); 130 return count;
|
/kernel/locking/ |
H A D | semaphore.c | 24 * The ->count variable represents how many more tasks can acquire this 58 if (likely(sem->count > 0)) 59 sem->count--; 81 if (likely(sem->count > 0)) 82 sem->count--; 107 if (likely(sem->count > 0)) 108 sem->count--; 133 int count; local 136 count = sem->count [all...] |
H A D | rtmutex-tester.c | 270 * @count: length of buffer 277 const char *buf, size_t count) 288 if (count >= sizeof(cmdbuf)) 292 if (buf[count-1] == '\n') 293 count--; 294 if (count < 1) 297 memcpy(cmdbuf, buf, count); 298 cmdbuf[count] = 0; 331 return count; 276 sysfs_test_command(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) argument
|
H A D | rwsem-xadd.c | 21 * Guide to the rw_semaphore's count field for common values. 49 * Note: writer can attempt to steal lock for this count by adding 50 * ACTIVE_WRITE_BIAS in cmpxchg and checking the old count 56 * the count becomes more than 0 for successful lock acquisition, 61 * checking the count becomes ACTIVE_WRITE_BIAS for successful lock 82 sem->count = RWSEM_UNLOCKED_VALUE; 113 * - the 'active part' of count (&0x0000ffff) reached 0 (but may have changed) 114 * - the 'waiting part' of count (&0xffff0000) is -ve (and will still be so) 160 * of the queue. Note we increment the 'active part' of the count by 208 long count, adjustmen local 251 rwsem_try_write_lock(long count, struct rw_semaphore *sem) argument 274 long old, count = ACCESS_ONCE(sem->count); local 408 long count; local [all...] |
/kernel/ |
H A D | delayacct.c | 50 * accumalator (@total) and @count 52 static void delayacct_end(u64 *start, u64 *total, u32 *count) argument 60 (*count)++;
|
H A D | groups.c | 71 unsigned int count = group_info->ngroups; local 73 for (i = 0; i < count; i++) { 88 unsigned int count = group_info->ngroups; local 90 for (i = 0; i < count; i++) {
|
H A D | latencytop.c | 26 * in the data structure. Both the count, total accumulated latency and maximum 126 latency_record[i].count++; 189 lat.count = 1; 216 mylat->count++; 250 lr->count, lr->time, lr->max); 266 lstats_write(struct file *file, const char __user *buf, size_t count, argument 271 return count;
|
H A D | ksysfs.c | 49 const char *buf, size_t count) 51 if (count+1 > UEVENT_HELPER_PATH_LEN) 53 memcpy(uevent_helper, buf, count); 54 uevent_helper[count] = '\0'; 55 if (count && uevent_helper[count-1] == '\n') 56 uevent_helper[count-1] = '\0'; 57 return count; 70 const char *buf, size_t count) 88 return count; 47 uevent_helper_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) argument 68 profiling_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) argument 113 kexec_crash_size_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) argument 153 rcu_expedited_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) argument 171 notes_read(struct file *filp, struct kobject *kobj, struct bin_attribute *bin_attr, char *buf, loff_t off, size_t count) argument [all...] |
H A D | tracepoint.c | 62 static inline void *allocate_probes(int count) argument 64 struct tp_probes *p = kmalloc(count * sizeof(struct tracepoint_func)
|
H A D | audit_watch.c | 48 atomic_t count; /* reference count */ member in struct:audit_watch 113 atomic_inc(&watch->count); 118 if (atomic_dec_and_test(&watch->count)) { 180 atomic_set(&watch->count, 1);
|
/kernel/irq/ |
H A D | spurious.c | 230 static int count = 100; local 232 if (count > 0) { 233 count--; 337 * count. We just care about the count being
|
H A D | irqdesc.c | 254 int count, i, node = first_online_node; local 262 count = ARRAY_SIZE(irq_desc); 264 for (i = 0; i < count; i++) {
|
H A D | proc.c | 77 const char __user *buffer, size_t count, loff_t *pos) 90 err = cpumask_parselist_user(buffer, count, new_value); 92 err = cpumask_parse_user(buffer, count, new_value); 109 err = irq_select_affinity_usr(irq, new_value) ? -EINVAL : count; 112 err = count; 121 const char __user *buffer, size_t count, loff_t *pos) 123 return write_irq_affinity(0, file, buffer, count, pos); 127 const char __user *buffer, size_t count, loff_t *pos) 129 return write_irq_affinity(1, file, buffer, count, pos); 178 const char __user *buffer, size_t count, loff_ 76 write_irq_affinity(int type, struct file *file, const char __user *buffer, size_t count, loff_t *pos) argument 120 irq_affinity_proc_write(struct file *file, const char __user *buffer, size_t count, loff_t *pos) argument 126 irq_affinity_list_proc_write(struct file *file, const char __user *buffer, size_t count, loff_t *pos) argument 177 default_affinity_write(struct file *file, const char __user *buffer, size_t count, loff_t *ppos) argument [all...] |
H A D | irqdomain.c | 326 irq_hw_number_t hwirq_base, int count) 330 pr_debug("%s(%s, irqbase=%i, hwbase=%i, count=%i)\n", __func__, 331 of_node_full_name(domain->of_node), irq_base, (int)hwirq_base, count); 333 for (i = 0; i < count; i++) { 441 * @count: Number of interrupts to map 455 irq_hw_number_t hwirq_base, int count) 459 ret = irq_alloc_descs(irq_base, irq_base, count, 464 irq_domain_associate_many(domain, irq_base, hwirq_base, count); 573 int count = 0; local 575 count 325 irq_domain_associate_many(struct irq_domain *domain, unsigned int irq_base, irq_hw_number_t hwirq_base, int count) argument 454 irq_create_strict_mappings(struct irq_domain *domain, unsigned int irq_base, irq_hw_number_t hwirq_base, int count) argument [all...] |
/kernel/time/ |
H A D | udelay_test.c | 109 size_t count, loff_t *pos) 116 if (count >= sizeof(lbuf)) 119 if (copy_from_user(lbuf, buf, count)) 121 lbuf[count] = '\0'; 134 return count; 108 udelay_test_write(struct file *file, const char __user *buf, size_t count, loff_t *pos) argument
|
H A D | posix-clock.c | 53 size_t count, loff_t *ppos) 62 err = clk->ops.read(clk, fp->f_flags, buf, count); 52 posix_clock_read(struct file *fp, char __user *buf, size_t count, loff_t *ppos) argument
|
H A D | clockevents.c | 627 ssize_t count = 0; local 632 count = snprintf(buf, PAGE_SIZE, "%s\n", td->evtdev->name); 634 return count; 641 const char *buf, size_t count) 644 ssize_t ret = sysfs_get_uname(buf, name, count); 666 return ret ? ret : count; 639 sysfs_unbind_tick_dev(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) argument
|
H A D | timer_stats.c | 51 * started the timer. We count the number of times an event happens: 70 unsigned long count; member in struct:entry 206 curr->count = 0; 262 entry->count++; 311 entry->count, entry->pid, entry->comm); 314 entry->count, entry->pid, entry->comm); 322 events += entry->count; 360 size_t count, loff_t *offs) 364 if (count != 2 || *offs) 367 if (copy_from_user(ctl, buf, count)) 359 tstats_write(struct file *file, const char __user *buf, size_t count, loff_t *offs) argument [all...] |
/kernel/power/ |
H A D | user.c | 136 size_t count, loff_t *offp) 157 res = simple_read_from_buffer(buf, count, &pg_offp, 169 size_t count, loff_t *offp) 188 buf, count); 135 snapshot_read(struct file *filp, char __user *buf, size_t count, loff_t *offp) argument 168 snapshot_write(struct file *filp, const char __user *buf, size_t count, loff_t *offp) argument
|
H A D | qos.c | 132 size_t count, loff_t *f_pos); 134 size_t count, loff_t *f_pos); 565 size_t count, loff_t *f_pos) 580 return simple_read_from_buffer(buf, count, f_pos, &value, sizeof(s32)); 584 size_t count, loff_t *f_pos) 589 if (count == sizeof(s32)) { 595 ret = kstrtos32_from_user(buf, count, 16, &value); 603 return count; 564 pm_qos_power_read(struct file *filp, char __user *buf, size_t count, loff_t *f_pos) argument 583 pm_qos_power_write(struct file *filp, const char __user *buf, size_t count, loff_t *f_pos) argument
|
/kernel/debug/kdb/ |
H A D | kdb_io.c | 215 int count; local 346 count = kallsyms_symbol_complete(p_tmp, 349 if (tab == 2 && count > 0) { 350 kdb_printf("\n%d symbols are found.", count); 351 if (count > dtab_count) { 352 count = dtab_count; 356 count); 359 for (i = 0; i < count; i++) { 370 } else if (tab != 2 && count > 0) {
|
/kernel/gcov/ |
H A D | gcc_3_4.c | 330 * @count: index into values array 340 unsigned int count; member in struct:gcov_iterator 404 iter->count = 0; 445 /* Advance to next count */ 446 iter->count++; 449 if (iter->count < get_func(iter)->n_ctrs[iter->type]) { 454 get_type(iter)->offset += iter->count; 455 iter->count = 0; 558 values[iter->count + get_type(iter)->offset]);
|
/kernel/trace/ |
H A D | trace_functions.c | 266 unsigned long *count = (long *)data; local 268 if (!*count) 271 if (*count != -1) 272 (*count)--; 359 long count = (long)data; local 363 if (count == -1) 366 seq_printf(m, ":count=%ld\n", count); 451 void *count = (void *)-1; local 476 ret = kstrtoul(number, 0, (unsigned long *)&count); [all...] |