/include/linux/ |
H A D | irq_cpustat.h | 6 * architecture. Some arch (like s390) have per cpu hardware pages and 21 #define __IRQ_STAT(cpu, member) (irq_stat[cpu].member) 29 #define nmi_count(cpu) __IRQ_STAT((cpu), __nmi_count) /* i386 */
|
H A D | kernel_stat.h | 51 #define kstat_cpu(cpu) per_cpu(kstat, cpu) 52 #define kcpustat_cpu(cpu) per_cpu(kernel_cpustat, cpu) 67 static inline unsigned int kstat_irqs_cpu(unsigned int irq, int cpu) argument 69 return kstat_cpu(cpu).irqs[irq]; 73 extern unsigned int kstat_irqs_cpu(unsigned int irq, int cpu); 88 static inline unsigned int kstat_softirqs_cpu(unsigned int irq, int cpu) argument 90 return kstat_cpu(cpu).softirqs[irq]; 100 int cpu; local 114 kstat_cpu_irqs_sum(unsigned int cpu) argument [all...] |
H A D | cpumask.h | 43 * cpu_possible_mask- has bit 'cpu' set iff cpu is populatable 44 * cpu_present_mask - has bit 'cpu' set iff cpu is populated 45 * cpu_online_mask - has bit 'cpu' set iff cpu available to scheduler 46 * cpu_active_mask - has bit 'cpu' set iff cpu available to migration 89 #define cpu_online(cpu) cpumask_test_cpu((cpu), cpu_online_mas 105 cpumask_check(unsigned int cpu) argument 139 cpumask_any_but(const struct cpumask *mask, unsigned int cpu) argument 255 cpumask_set_cpu(unsigned int cpu, struct cpumask *dstp) argument 265 cpumask_clear_cpu(int cpu, struct cpumask *dstp) argument 287 cpumask_test_and_set_cpu(int cpu, struct cpumask *cpumask) argument 299 cpumask_test_and_clear_cpu(int cpu, struct cpumask *cpumask) argument 737 get_cpu_mask(unsigned int cpu) argument 834 __cpu_set(int cpu, volatile cpumask_t *dstp) argument 840 __cpu_clear(int cpu, volatile cpumask_t *dstp) argument 861 __cpu_test_and_set(int cpu, cpumask_t *addr) argument [all...] |
H A D | tick.h | 75 extern struct tick_device *tick_get_device(int cpu); 84 extern void tick_cancel_sched_timer(int cpu); 86 static inline void tick_cancel_sched_timer(int cpu) { } argument 102 extern struct tick_sched *tick_get_tick_sched(int cpu); 103 extern void tick_check_idle(int cpu); 106 # define arch_needs_cpu(cpu) (0) 111 static inline void tick_check_idle(int cpu) { } argument 117 static inline void tick_cancel_sched_timer(int cpu) { } argument 120 static inline void tick_check_idle(int cpu) { } argument 129 extern u64 get_cpu_idle_time_us(int cpu, u6 141 get_cpu_idle_time_us(int cpu, u64 *unused) argument 142 get_cpu_iowait_time_us(int cpu, u64 *unused) argument [all...] |
H A D | rcutree.h | 34 extern void rcu_note_context_switch(int cpu); 35 extern int rcu_needs_cpu(int cpu); 43 static inline void rcu_virt_note_context_switch(int cpu) argument 45 rcu_note_context_switch(cpu);
|
H A D | rcutiny.h | 98 static inline int rcu_needs_cpu(int cpu) argument 109 static inline int rcu_needs_cpu(int cpu) argument 116 static inline void rcu_note_context_switch(int cpu) argument 118 rcu_sched_qs(cpu); 126 static inline void rcu_virt_note_context_switch(int cpu) argument
|
H A D | ring_buffer.h | 113 ring_buffer_peek(struct ring_buffer *buffer, int cpu, u64 *ts, 116 ring_buffer_consume(struct ring_buffer *buffer, int cpu, u64 *ts, 120 ring_buffer_read_prepare(struct ring_buffer *buffer, int cpu); 134 void ring_buffer_reset_cpu(struct ring_buffer *buffer, int cpu); 139 struct ring_buffer *buffer_b, int cpu); 143 struct ring_buffer *buffer_b, int cpu) 150 int ring_buffer_empty_cpu(struct ring_buffer *buffer, int cpu); 157 void ring_buffer_record_disable_cpu(struct ring_buffer *buffer, int cpu); 158 void ring_buffer_record_enable_cpu(struct ring_buffer *buffer, int cpu); 160 unsigned long ring_buffer_oldest_event_ts(struct ring_buffer *buffer, int cpu); 142 ring_buffer_swap_cpu(struct ring_buffer *buffer_a, struct ring_buffer *buffer_b, int cpu) argument [all...] |
H A D | topology.h | 233 static inline int cpu_to_node(int cpu) argument 235 return per_cpu(numa_node, cpu); 247 static inline void set_cpu_numa_node(int cpu, int node) argument 249 per_cpu(numa_node, cpu) = node; 268 * N.B., Do NOT reference the '_numa_mem_' per cpu variable directly. 290 static inline int cpu_to_mem(int cpu) argument 292 return per_cpu(_numa_mem_, cpu); 297 static inline void set_cpu_numa_mem(int cpu, int node) argument 299 per_cpu(_numa_mem_, cpu) = node; 314 static inline int cpu_to_mem(int cpu) argument [all...] |
H A D | cpu_rmap.h | 47 static inline u16 cpu_rmap_lookup_index(struct cpu_rmap *rmap, unsigned int cpu) argument 49 return rmap->near[cpu].index; 52 static inline void *cpu_rmap_lookup_obj(struct cpu_rmap *rmap, unsigned int cpu) argument 54 return rmap->obj[rmap->near[cpu].index];
|
H A D | cpufreq.h | 69 /* /sys/devices/system/cpu/cpufreq: entry point for global variables */ 93 unsigned int cpu; /* cpu nr of registered CPU */ member in struct:cpufreq_policy 130 unsigned int cpu; /* cpu nr */ member in struct:cpufreq_freqs 198 unsigned int cpu); 229 unsigned int (*get) (unsigned int cpu); 233 unsigned int cpu); 234 int (*bios_limit) (int cpu, unsigned int *limit); 312 int cpufreq_get_policy(struct cpufreq_policy *policy, unsigned int cpu); 319 cpufreq_get(unsigned int cpu) argument 330 cpufreq_quick_get(unsigned int cpu) argument 334 cpufreq_quick_get_max(unsigned int cpu) argument [all...] |
H A D | node.h | 38 extern int register_cpu_under_node(unsigned int cpu, unsigned int nid); 39 extern int unregister_cpu_under_node(unsigned int cpu, unsigned int nid); 58 static inline int register_cpu_under_node(unsigned int cpu, unsigned int nid) argument 62 static inline int unregister_cpu_under_node(unsigned int cpu, unsigned int nid) argument
|
H A D | stop_machine.h | 4 #include <linux/cpu.h> 10 * stop_cpu[s]() is simplistic per-cpu maximum priority cpu 15 * Resources for this mechanism are preallocated when a cpu is brought 30 int stop_one_cpu(unsigned int cpu, cpu_stop_fn_t fn, void *arg); 31 void stop_one_cpu_nowait(unsigned int cpu, cpu_stop_fn_t fn, void *arg, 46 static inline int stop_one_cpu(unsigned int cpu, cpu_stop_fn_t fn, void *arg) argument 50 if (cpu == smp_processor_id()) 65 static inline void stop_one_cpu_nowait(unsigned int cpu, argument 69 if (cpu [all...] |
H A D | cpu.h | 2 * include/linux/cpu.h - generic cpu definition 5 * basic 'struct cpu' here, which can be embedded in per-arch 8 * Basic handling of the devices is done in drivers/base/cpu.c 11 * CPUs are exported via sysfs in the class/cpu/devices/ 23 struct cpu { struct 29 extern int register_cpu(struct cpu *cpu, int num); 30 extern struct device *get_cpu_device(unsigned cpu); 31 extern bool cpu_is_hotpluggable(unsigned cpu); [all...] |
H A D | kdb.h | 25 * kdb_initial_cpu is initialized to -1, and is set to the cpu 87 extern int kdb_flags; /* Global flags, see kdb_state for per cpu state */ 131 unsigned int cpu = task_thread_info(p)->cpu; local 132 if (cpu > num_possible_cpus()) 133 cpu = 0; 134 return cpu;
|
H A D | dca.h | 52 int cpu); 70 u8 dca_get_tag(int cpu); 71 u8 dca3_get_tag(struct device *dev, int cpu);
|
H A D | lglock.h | 25 #include <linux/cpu.h> 41 #define lg_local_lock_cpu(name, cpu) name##_local_lock_cpu(cpu) 42 #define lg_local_unlock_cpu(name, cpu) name##_local_unlock_cpu(cpu) 67 extern void name##_local_lock_cpu(int cpu); \ 68 extern void name##_local_unlock_cpu(int cpu); \ 135 void name##_local_lock_cpu(int cpu) { \ 139 lock = &per_cpu(name##_lock, cpu); \ 144 void name##_local_unlock_cpu(int cpu) { \ [all...] |
H A D | vmstat.h | 15 * Light weight per cpu counter implementation. 52 extern void vm_events_fold_cpu(int cpu); 54 static inline void vm_events_fold_cpu(int cpu) argument 77 static inline void vm_events_fold_cpu(int cpu) argument 88 * Zone based page accounting with per cpu differentials. 132 int cpu; local 133 for_each_online_cpu(cpu) 134 x += per_cpu_ptr(zone->pageset, cpu)->vm_stat_diff[item]; 256 static inline void refresh_cpu_vm_stats(int cpu) { } argument
|
H A D | smp.h | 53 extern void smp_send_reschedule(int cpu); 116 void on_each_cpu_cond(bool (*cond_func)(int cpu, void *info), 121 * Mark the boot cpu "online" so that it can call console drivers in 122 * printk() and can access its per-cpu storage. 181 static inline void smp_send_reschedule(int cpu) { } argument
|
/include/asm-generic/ |
H A D | topology.h | 35 #define cpu_to_node(cpu) ((void)(cpu),0) 41 #define set_cpu_numa_node(cpu, node) 44 #define cpu_to_mem(cpu) ((void)(cpu),0) 71 #define set_cpu_numa_mem(cpu, node)
|
H A D | percpu.h | 26 * means of obtaining the offset to the per cpu variables of the 56 #define per_cpu(var, cpu) \ 57 (*SHIFT_PERCPU_PTR(&(var), per_cpu_offset(cpu))) 82 #define per_cpu(var, cpu) (*((void)(cpu), VERIFY_PERCPU_PTR(&(var))))
|
/include/linux/netfilter/ |
H A D | xt_cpu.h | 7 __u32 cpu; member in struct:xt_cpu_info
|
/include/trace/events/ |
H A D | workqueue.h | 29 * @req_cpu: the requested cpu 49 __field( unsigned int, cpu ) 57 __entry->cpu = cwq->gcwq->cpu; 60 TP_printk("work struct=%p function=%pf workqueue=%p req_cpu=%u cpu=%u", 62 __entry->req_cpu, __entry->cpu)
|
H A D | mce.h | 26 __field( u32, cpu ) 44 __entry->cpu = m->extcpu; 54 __entry->cpu,
|
/include/xen/ |
H A D | events.h | 15 int bind_virq_to_irq(unsigned int virq, unsigned int cpu); 16 int bind_virq_to_irqhandler(unsigned int virq, unsigned int cpu, 21 unsigned int cpu, 47 void xen_send_IPI_one(unsigned int cpu, enum ipi_vector vector);
|
/include/acpi/ |
H A D | processor.h | 5 #include <linux/cpu.h> 234 *performance, unsigned int cpu); 238 unsigned int cpu); 250 unsigned int cpu); 251 int acpi_processor_ffh_cstate_probe(unsigned int cpu, 258 *flags, unsigned int cpu) 263 static inline int acpi_processor_ffh_cstate_probe(unsigned int cpu, argument 283 extern int acpi_processor_get_bios_limit(int cpu, unsigned int *limit); 306 static inline int acpi_processor_get_bios_limit(int cpu, unsigned int *limit) argument 323 * Reevaluate whether the T-state is invalid after one cpu i 256 acpi_processor_power_init_bm_check(struct acpi_processor_flags *flags, unsigned int cpu) argument [all...] |