Searched refs:cpu (Results 1 - 25 of 63) sorted by relevance

123

/include/linux/
H A Dirq_cpustat.h6 * architecture. Some arch (like s390) have per cpu hardware pages and
21 #define __IRQ_STAT(cpu, member) (irq_stat[cpu].member)
29 #define nmi_count(cpu) __IRQ_STAT((cpu), __nmi_count) /* i386 */
H A Dkernel_stat.h51 #define kstat_cpu(cpu) per_cpu(kstat, cpu)
52 #define kcpustat_cpu(cpu) per_cpu(kernel_cpustat, cpu)
67 static inline unsigned int kstat_irqs_cpu(unsigned int irq, int cpu) argument
69 return kstat_cpu(cpu).irqs[irq];
73 extern unsigned int kstat_irqs_cpu(unsigned int irq, int cpu);
88 static inline unsigned int kstat_softirqs_cpu(unsigned int irq, int cpu) argument
90 return kstat_cpu(cpu).softirqs[irq];
100 int cpu; local
114 kstat_cpu_irqs_sum(unsigned int cpu) argument
[all...]
H A Dcpumask.h43 * cpu_possible_mask- has bit 'cpu' set iff cpu is populatable
44 * cpu_present_mask - has bit 'cpu' set iff cpu is populated
45 * cpu_online_mask - has bit 'cpu' set iff cpu available to scheduler
46 * cpu_active_mask - has bit 'cpu' set iff cpu available to migration
89 #define cpu_online(cpu) cpumask_test_cpu((cpu), cpu_online_mas
105 cpumask_check(unsigned int cpu) argument
139 cpumask_any_but(const struct cpumask *mask, unsigned int cpu) argument
255 cpumask_set_cpu(unsigned int cpu, struct cpumask *dstp) argument
265 cpumask_clear_cpu(int cpu, struct cpumask *dstp) argument
287 cpumask_test_and_set_cpu(int cpu, struct cpumask *cpumask) argument
299 cpumask_test_and_clear_cpu(int cpu, struct cpumask *cpumask) argument
737 get_cpu_mask(unsigned int cpu) argument
834 __cpu_set(int cpu, volatile cpumask_t *dstp) argument
840 __cpu_clear(int cpu, volatile cpumask_t *dstp) argument
861 __cpu_test_and_set(int cpu, cpumask_t *addr) argument
[all...]
H A Dtick.h75 extern struct tick_device *tick_get_device(int cpu);
84 extern void tick_cancel_sched_timer(int cpu);
86 static inline void tick_cancel_sched_timer(int cpu) { } argument
102 extern struct tick_sched *tick_get_tick_sched(int cpu);
103 extern void tick_check_idle(int cpu);
106 # define arch_needs_cpu(cpu) (0)
111 static inline void tick_check_idle(int cpu) { } argument
117 static inline void tick_cancel_sched_timer(int cpu) { } argument
120 static inline void tick_check_idle(int cpu) { } argument
129 extern u64 get_cpu_idle_time_us(int cpu, u6
141 get_cpu_idle_time_us(int cpu, u64 *unused) argument
142 get_cpu_iowait_time_us(int cpu, u64 *unused) argument
[all...]
H A Drcutree.h34 extern void rcu_note_context_switch(int cpu);
35 extern int rcu_needs_cpu(int cpu);
43 static inline void rcu_virt_note_context_switch(int cpu) argument
45 rcu_note_context_switch(cpu);
H A Drcutiny.h98 static inline int rcu_needs_cpu(int cpu) argument
109 static inline int rcu_needs_cpu(int cpu) argument
116 static inline void rcu_note_context_switch(int cpu) argument
118 rcu_sched_qs(cpu);
126 static inline void rcu_virt_note_context_switch(int cpu) argument
H A Dring_buffer.h113 ring_buffer_peek(struct ring_buffer *buffer, int cpu, u64 *ts,
116 ring_buffer_consume(struct ring_buffer *buffer, int cpu, u64 *ts,
120 ring_buffer_read_prepare(struct ring_buffer *buffer, int cpu);
134 void ring_buffer_reset_cpu(struct ring_buffer *buffer, int cpu);
139 struct ring_buffer *buffer_b, int cpu);
143 struct ring_buffer *buffer_b, int cpu)
150 int ring_buffer_empty_cpu(struct ring_buffer *buffer, int cpu);
157 void ring_buffer_record_disable_cpu(struct ring_buffer *buffer, int cpu);
158 void ring_buffer_record_enable_cpu(struct ring_buffer *buffer, int cpu);
160 unsigned long ring_buffer_oldest_event_ts(struct ring_buffer *buffer, int cpu);
142 ring_buffer_swap_cpu(struct ring_buffer *buffer_a, struct ring_buffer *buffer_b, int cpu) argument
[all...]
H A Dtopology.h233 static inline int cpu_to_node(int cpu) argument
235 return per_cpu(numa_node, cpu);
247 static inline void set_cpu_numa_node(int cpu, int node) argument
249 per_cpu(numa_node, cpu) = node;
268 * N.B., Do NOT reference the '_numa_mem_' per cpu variable directly.
290 static inline int cpu_to_mem(int cpu) argument
292 return per_cpu(_numa_mem_, cpu);
297 static inline void set_cpu_numa_mem(int cpu, int node) argument
299 per_cpu(_numa_mem_, cpu) = node;
314 static inline int cpu_to_mem(int cpu) argument
[all...]
H A Dcpu_rmap.h47 static inline u16 cpu_rmap_lookup_index(struct cpu_rmap *rmap, unsigned int cpu) argument
49 return rmap->near[cpu].index;
52 static inline void *cpu_rmap_lookup_obj(struct cpu_rmap *rmap, unsigned int cpu) argument
54 return rmap->obj[rmap->near[cpu].index];
H A Dcpufreq.h69 /* /sys/devices/system/cpu/cpufreq: entry point for global variables */
93 unsigned int cpu; /* cpu nr of registered CPU */ member in struct:cpufreq_policy
130 unsigned int cpu; /* cpu nr */ member in struct:cpufreq_freqs
198 unsigned int cpu);
229 unsigned int (*get) (unsigned int cpu);
233 unsigned int cpu);
234 int (*bios_limit) (int cpu, unsigned int *limit);
312 int cpufreq_get_policy(struct cpufreq_policy *policy, unsigned int cpu);
319 cpufreq_get(unsigned int cpu) argument
330 cpufreq_quick_get(unsigned int cpu) argument
334 cpufreq_quick_get_max(unsigned int cpu) argument
[all...]
H A Dnode.h38 extern int register_cpu_under_node(unsigned int cpu, unsigned int nid);
39 extern int unregister_cpu_under_node(unsigned int cpu, unsigned int nid);
58 static inline int register_cpu_under_node(unsigned int cpu, unsigned int nid) argument
62 static inline int unregister_cpu_under_node(unsigned int cpu, unsigned int nid) argument
H A Dstop_machine.h4 #include <linux/cpu.h>
10 * stop_cpu[s]() is simplistic per-cpu maximum priority cpu
15 * Resources for this mechanism are preallocated when a cpu is brought
30 int stop_one_cpu(unsigned int cpu, cpu_stop_fn_t fn, void *arg);
31 void stop_one_cpu_nowait(unsigned int cpu, cpu_stop_fn_t fn, void *arg,
46 static inline int stop_one_cpu(unsigned int cpu, cpu_stop_fn_t fn, void *arg) argument
50 if (cpu == smp_processor_id())
65 static inline void stop_one_cpu_nowait(unsigned int cpu, argument
69 if (cpu
[all...]
H A Dcpu.h2 * include/linux/cpu.h - generic cpu definition
5 * basic 'struct cpu' here, which can be embedded in per-arch
8 * Basic handling of the devices is done in drivers/base/cpu.c
11 * CPUs are exported via sysfs in the class/cpu/devices/
23 struct cpu { struct
29 extern int register_cpu(struct cpu *cpu, int num);
30 extern struct device *get_cpu_device(unsigned cpu);
31 extern bool cpu_is_hotpluggable(unsigned cpu);
[all...]
H A Dkdb.h25 * kdb_initial_cpu is initialized to -1, and is set to the cpu
87 extern int kdb_flags; /* Global flags, see kdb_state for per cpu state */
131 unsigned int cpu = task_thread_info(p)->cpu; local
132 if (cpu > num_possible_cpus())
133 cpu = 0;
134 return cpu;
H A Ddca.h52 int cpu);
70 u8 dca_get_tag(int cpu);
71 u8 dca3_get_tag(struct device *dev, int cpu);
H A Dlglock.h25 #include <linux/cpu.h>
41 #define lg_local_lock_cpu(name, cpu) name##_local_lock_cpu(cpu)
42 #define lg_local_unlock_cpu(name, cpu) name##_local_unlock_cpu(cpu)
67 extern void name##_local_lock_cpu(int cpu); \
68 extern void name##_local_unlock_cpu(int cpu); \
135 void name##_local_lock_cpu(int cpu) { \
139 lock = &per_cpu(name##_lock, cpu); \
144 void name##_local_unlock_cpu(int cpu) { \
[all...]
H A Dvmstat.h15 * Light weight per cpu counter implementation.
52 extern void vm_events_fold_cpu(int cpu);
54 static inline void vm_events_fold_cpu(int cpu) argument
77 static inline void vm_events_fold_cpu(int cpu) argument
88 * Zone based page accounting with per cpu differentials.
132 int cpu; local
133 for_each_online_cpu(cpu)
134 x += per_cpu_ptr(zone->pageset, cpu)->vm_stat_diff[item];
256 static inline void refresh_cpu_vm_stats(int cpu) { } argument
H A Dsmp.h53 extern void smp_send_reschedule(int cpu);
116 void on_each_cpu_cond(bool (*cond_func)(int cpu, void *info),
121 * Mark the boot cpu "online" so that it can call console drivers in
122 * printk() and can access its per-cpu storage.
181 static inline void smp_send_reschedule(int cpu) { } argument
/include/asm-generic/
H A Dtopology.h35 #define cpu_to_node(cpu) ((void)(cpu),0)
41 #define set_cpu_numa_node(cpu, node)
44 #define cpu_to_mem(cpu) ((void)(cpu),0)
71 #define set_cpu_numa_mem(cpu, node)
H A Dpercpu.h26 * means of obtaining the offset to the per cpu variables of the
56 #define per_cpu(var, cpu) \
57 (*SHIFT_PERCPU_PTR(&(var), per_cpu_offset(cpu)))
82 #define per_cpu(var, cpu) (*((void)(cpu), VERIFY_PERCPU_PTR(&(var))))
/include/linux/netfilter/
H A Dxt_cpu.h7 __u32 cpu; member in struct:xt_cpu_info
/include/trace/events/
H A Dworkqueue.h29 * @req_cpu: the requested cpu
49 __field( unsigned int, cpu )
57 __entry->cpu = cwq->gcwq->cpu;
60 TP_printk("work struct=%p function=%pf workqueue=%p req_cpu=%u cpu=%u",
62 __entry->req_cpu, __entry->cpu)
H A Dmce.h26 __field( u32, cpu )
44 __entry->cpu = m->extcpu;
54 __entry->cpu,
/include/xen/
H A Devents.h15 int bind_virq_to_irq(unsigned int virq, unsigned int cpu);
16 int bind_virq_to_irqhandler(unsigned int virq, unsigned int cpu,
21 unsigned int cpu,
47 void xen_send_IPI_one(unsigned int cpu, enum ipi_vector vector);
/include/acpi/
H A Dprocessor.h5 #include <linux/cpu.h>
234 *performance, unsigned int cpu);
238 unsigned int cpu);
250 unsigned int cpu);
251 int acpi_processor_ffh_cstate_probe(unsigned int cpu,
258 *flags, unsigned int cpu)
263 static inline int acpi_processor_ffh_cstate_probe(unsigned int cpu, argument
283 extern int acpi_processor_get_bios_limit(int cpu, unsigned int *limit);
306 static inline int acpi_processor_get_bios_limit(int cpu, unsigned int *limit) argument
323 * Reevaluate whether the T-state is invalid after one cpu i
256 acpi_processor_power_init_bm_check(struct acpi_processor_flags *flags, unsigned int cpu) argument
[all...]

Completed in 3593 milliseconds

123