Searched refs:this_cpu (Results 1 - 10 of 10) sorted by relevance

/kernel/trace/
H A Dtrace_clock.c93 int this_cpu; local
98 this_cpu = raw_smp_processor_id();
99 now = sched_clock_cpu(this_cpu);
/kernel/
H A Dsmp.c277 int this_cpu; local
284 this_cpu = get_cpu();
292 WARN_ON_ONCE(cpu_online(this_cpu) && irqs_disabled()
392 int cpu, next_cpu, this_cpu = smp_processor_id(); local
400 WARN_ON_ONCE(cpu_online(this_cpu) && irqs_disabled()
405 if (cpu == this_cpu)
414 if (next_cpu == this_cpu)
426 cpumask_clear_cpu(this_cpu, cfd->cpumask);
H A Dwatchdog.c354 int this_cpu = smp_processor_id(); local
362 this_cpu);
365 this_cpu);
/kernel/sched/
H A Ddeadline.c1176 int this_cpu = smp_processor_id(); local
1214 * Check if this_cpu is to be skipped (i.e., it is
1217 if (!cpumask_test_cpu(this_cpu, later_mask))
1218 this_cpu = -1;
1225 * If possible, preempting this_cpu is
1228 if (this_cpu != -1 &&
1229 cpumask_test_cpu(this_cpu, sched_domain_span(sd))) {
1231 return this_cpu;
1251 if (this_cpu != -1)
1252 return this_cpu;
1417 int this_cpu = this_rq->cpu, ret = 0, cpu; local
[all...]
H A Drt.c1529 int this_cpu = smp_processor_id(); local
1557 if (!cpumask_test_cpu(this_cpu, lowest_mask))
1558 this_cpu = -1; /* Skip this_cpu opt if not among lowest */
1566 * "this_cpu" is cheaper to preempt than a
1569 if (this_cpu != -1 &&
1570 cpumask_test_cpu(this_cpu, sched_domain_span(sd))) {
1572 return this_cpu;
1590 if (this_cpu != -1)
1591 return this_cpu;
1762 int this_cpu = this_rq->cpu, ret = 0, cpu; local
[all...]
H A Dfair.c4308 int idx, this_cpu, prev_cpu; local
4321 this_cpu = smp_processor_id();
4324 this_load = target_load(this_cpu, idx);
4335 this_load += effective_load(tg, this_cpu, -weight, -weight);
4343 * In low-load situations, where prev_cpu is idle and this_cpu is idle
4349 * task to be woken on this_cpu.
4355 prev_eff_load *= capacity_of(this_cpu);
4359 effective_load(tg, this_cpu, weight, weight);
4383 int this_cpu, int sd_flag)
4403 local_group = cpumask_test_cpu(this_cpu,
4382 find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu, int sd_flag) argument
4439 find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu) argument
6632 load_balance(int this_cpu, struct rq *this_rq, struct sched_domain *sd, enum cpu_idle_type idle, int *continue_balancing) argument
6928 int this_cpu = this_rq->cpu; local
7354 int this_cpu = this_rq->cpu; local
7546 int this_cpu = smp_processor_id(); local
[all...]
H A Ddebug.c643 unsigned int this_cpu = raw_smp_processor_id(); local
646 t0 = cpu_clock(this_cpu);
647 t1 = cpu_clock(this_cpu);
H A Dcore.c1443 int this_cpu = smp_processor_id(); local
1445 if (cpu == this_cpu) {
1453 for_each_domain(this_cpu, sd) {
1640 bool cpus_share_cache(int this_cpu, int that_cpu) argument
1642 return per_cpu(sd_llc_id, this_cpu) == per_cpu(sd_llc_id, that_cpu);
/kernel/time/
H A Dhrtimer.c202 int this_cpu = smp_processor_id(); local
228 if (cpu != this_cpu && hrtimer_check_target(timer, new_base)) {
229 cpu = this_cpu;
237 if (cpu != this_cpu && hrtimer_check_target(timer, new_base)) {
238 cpu = this_cpu;
/kernel/printk/
H A Dprintk.c1630 int this_cpu; local
1646 this_cpu = smp_processor_id();
1651 if (unlikely(logbuf_cpu == this_cpu)) {
1669 logbuf_cpu = this_cpu;

Completed in 154 milliseconds