Searched refs:this_cpu (Results 1 - 10 of 10) sorted by relevance
/kernel/trace/ |
H A D | trace_clock.c | 93 int this_cpu; local 98 this_cpu = raw_smp_processor_id(); 99 now = sched_clock_cpu(this_cpu);
|
/kernel/ |
H A D | smp.c | 277 int this_cpu; local 284 this_cpu = get_cpu(); 292 WARN_ON_ONCE(cpu_online(this_cpu) && irqs_disabled() 392 int cpu, next_cpu, this_cpu = smp_processor_id(); local 400 WARN_ON_ONCE(cpu_online(this_cpu) && irqs_disabled() 405 if (cpu == this_cpu) 414 if (next_cpu == this_cpu) 426 cpumask_clear_cpu(this_cpu, cfd->cpumask);
|
H A D | watchdog.c | 354 int this_cpu = smp_processor_id(); local 362 this_cpu); 365 this_cpu);
|
/kernel/sched/ |
H A D | deadline.c | 1176 int this_cpu = smp_processor_id(); local 1214 * Check if this_cpu is to be skipped (i.e., it is 1217 if (!cpumask_test_cpu(this_cpu, later_mask)) 1218 this_cpu = -1; 1225 * If possible, preempting this_cpu is 1228 if (this_cpu != -1 && 1229 cpumask_test_cpu(this_cpu, sched_domain_span(sd))) { 1231 return this_cpu; 1251 if (this_cpu != -1) 1252 return this_cpu; 1417 int this_cpu = this_rq->cpu, ret = 0, cpu; local [all...] |
H A D | rt.c | 1529 int this_cpu = smp_processor_id(); local 1557 if (!cpumask_test_cpu(this_cpu, lowest_mask)) 1558 this_cpu = -1; /* Skip this_cpu opt if not among lowest */ 1566 * "this_cpu" is cheaper to preempt than a 1569 if (this_cpu != -1 && 1570 cpumask_test_cpu(this_cpu, sched_domain_span(sd))) { 1572 return this_cpu; 1590 if (this_cpu != -1) 1591 return this_cpu; 1762 int this_cpu = this_rq->cpu, ret = 0, cpu; local [all...] |
H A D | fair.c | 4308 int idx, this_cpu, prev_cpu; local 4321 this_cpu = smp_processor_id(); 4324 this_load = target_load(this_cpu, idx); 4335 this_load += effective_load(tg, this_cpu, -weight, -weight); 4343 * In low-load situations, where prev_cpu is idle and this_cpu is idle 4349 * task to be woken on this_cpu. 4355 prev_eff_load *= capacity_of(this_cpu); 4359 effective_load(tg, this_cpu, weight, weight); 4383 int this_cpu, int sd_flag) 4403 local_group = cpumask_test_cpu(this_cpu, 4382 find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu, int sd_flag) argument 4439 find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu) argument 6632 load_balance(int this_cpu, struct rq *this_rq, struct sched_domain *sd, enum cpu_idle_type idle, int *continue_balancing) argument 6928 int this_cpu = this_rq->cpu; local 7354 int this_cpu = this_rq->cpu; local 7546 int this_cpu = smp_processor_id(); local [all...] |
H A D | debug.c | 643 unsigned int this_cpu = raw_smp_processor_id(); local 646 t0 = cpu_clock(this_cpu); 647 t1 = cpu_clock(this_cpu);
|
H A D | core.c | 1443 int this_cpu = smp_processor_id(); local 1445 if (cpu == this_cpu) { 1453 for_each_domain(this_cpu, sd) { 1640 bool cpus_share_cache(int this_cpu, int that_cpu) argument 1642 return per_cpu(sd_llc_id, this_cpu) == per_cpu(sd_llc_id, that_cpu);
|
/kernel/time/ |
H A D | hrtimer.c | 202 int this_cpu = smp_processor_id(); local 228 if (cpu != this_cpu && hrtimer_check_target(timer, new_base)) { 229 cpu = this_cpu; 237 if (cpu != this_cpu && hrtimer_check_target(timer, new_base)) { 238 cpu = this_cpu;
|
/kernel/printk/ |
H A D | printk.c | 1630 int this_cpu; local 1646 this_cpu = smp_processor_id(); 1651 if (unlikely(logbuf_cpu == this_cpu)) { 1669 logbuf_cpu = this_cpu;
|
Completed in 154 milliseconds