tick-broadcast.c revision 79bf2bb335b85db25d27421c798595a2fa2a0e82
1/*
2 * linux/kernel/time/tick-broadcast.c
3 *
4 * This file contains functions which emulate a local clock-event
5 * device via a broadcast event source.
6 *
7 * Copyright(C) 2005-2006, Thomas Gleixner <tglx@linutronix.de>
8 * Copyright(C) 2005-2007, Red Hat, Inc., Ingo Molnar
9 * Copyright(C) 2006-2007, Timesys Corp., Thomas Gleixner
10 *
11 * This code is licenced under the GPL version 2. For details see
12 * kernel-base/COPYING.
13 */
14#include <linux/cpu.h>
15#include <linux/err.h>
16#include <linux/hrtimer.h>
17#include <linux/irq.h>
18#include <linux/percpu.h>
19#include <linux/profile.h>
20#include <linux/sched.h>
21#include <linux/tick.h>
22
23#include "tick-internal.h"
24
25/*
26 * Broadcast support for broken x86 hardware, where the local apic
27 * timer stops in C3 state.
28 */
29
30struct tick_device tick_broadcast_device;
31static cpumask_t tick_broadcast_mask;
32static DEFINE_SPINLOCK(tick_broadcast_lock);
33
34/*
35 * Start the device in periodic mode
36 */
37static void tick_broadcast_start_periodic(struct clock_event_device *bc)
38{
39	if (bc && bc->mode == CLOCK_EVT_MODE_SHUTDOWN)
40		tick_setup_periodic(bc, 1);
41}
42
43/*
44 * Check, if the device can be utilized as broadcast device:
45 */
46int tick_check_broadcast_device(struct clock_event_device *dev)
47{
48	if (tick_broadcast_device.evtdev ||
49	    (dev->features & CLOCK_EVT_FEAT_C3STOP))
50		return 0;
51
52	clockevents_exchange_device(NULL, dev);
53	tick_broadcast_device.evtdev = dev;
54	if (!cpus_empty(tick_broadcast_mask))
55		tick_broadcast_start_periodic(dev);
56	return 1;
57}
58
59/*
60 * Check, if the device is the broadcast device
61 */
62int tick_is_broadcast_device(struct clock_event_device *dev)
63{
64	return (dev && tick_broadcast_device.evtdev == dev);
65}
66
67/*
68 * Check, if the device is disfunctional and a place holder, which
69 * needs to be handled by the broadcast device.
70 */
71int tick_device_uses_broadcast(struct clock_event_device *dev, int cpu)
72{
73	unsigned long flags;
74	int ret = 0;
75
76	spin_lock_irqsave(&tick_broadcast_lock, flags);
77
78	/*
79	 * Devices might be registered with both periodic and oneshot
80	 * mode disabled. This signals, that the device needs to be
81	 * operated from the broadcast device and is a placeholder for
82	 * the cpu local device.
83	 */
84	if (!tick_device_is_functional(dev)) {
85		dev->event_handler = tick_handle_periodic;
86		cpu_set(cpu, tick_broadcast_mask);
87		tick_broadcast_start_periodic(tick_broadcast_device.evtdev);
88		ret = 1;
89	}
90
91	spin_unlock_irqrestore(&tick_broadcast_lock, flags);
92	return ret;
93}
94
95/*
96 * Broadcast the event to the cpus, which are set in the mask
97 */
98int tick_do_broadcast(cpumask_t mask)
99{
100	int ret = 0, cpu = smp_processor_id();
101	struct tick_device *td;
102
103	/*
104	 * Check, if the current cpu is in the mask
105	 */
106	if (cpu_isset(cpu, mask)) {
107		cpu_clear(cpu, mask);
108		td = &per_cpu(tick_cpu_device, cpu);
109		td->evtdev->event_handler(td->evtdev);
110		ret = 1;
111	}
112
113	if (!cpus_empty(mask)) {
114		/*
115		 * It might be necessary to actually check whether the devices
116		 * have different broadcast functions. For now, just use the
117		 * one of the first device. This works as long as we have this
118		 * misfeature only on x86 (lapic)
119		 */
120		cpu = first_cpu(mask);
121		td = &per_cpu(tick_cpu_device, cpu);
122		td->evtdev->broadcast(mask);
123		ret = 1;
124	}
125	return ret;
126}
127
128/*
129 * Periodic broadcast:
130 * - invoke the broadcast handlers
131 */
132static void tick_do_periodic_broadcast(void)
133{
134	cpumask_t mask;
135
136	spin_lock(&tick_broadcast_lock);
137
138	cpus_and(mask, cpu_online_map, tick_broadcast_mask);
139	tick_do_broadcast(mask);
140
141	spin_unlock(&tick_broadcast_lock);
142}
143
144/*
145 * Event handler for periodic broadcast ticks
146 */
147static void tick_handle_periodic_broadcast(struct clock_event_device *dev)
148{
149	dev->next_event.tv64 = KTIME_MAX;
150
151	tick_do_periodic_broadcast();
152
153	/*
154	 * The device is in periodic mode. No reprogramming necessary:
155	 */
156	if (dev->mode == CLOCK_EVT_MODE_PERIODIC)
157		return;
158
159	/*
160	 * Setup the next period for devices, which do not have
161	 * periodic mode:
162	 */
163	for (;;) {
164		ktime_t next = ktime_add(dev->next_event, tick_period);
165
166		if (!clockevents_program_event(dev, next, ktime_get()))
167			return;
168		tick_do_periodic_broadcast();
169	}
170}
171
172/*
173 * Powerstate information: The system enters/leaves a state, where
174 * affected devices might stop
175 */
176static void tick_do_broadcast_on_off(void *why)
177{
178	struct clock_event_device *bc, *dev;
179	struct tick_device *td;
180	unsigned long flags, *reason = why;
181	int cpu;
182
183	spin_lock_irqsave(&tick_broadcast_lock, flags);
184
185	cpu = smp_processor_id();
186	td = &per_cpu(tick_cpu_device, cpu);
187	dev = td->evtdev;
188	bc = tick_broadcast_device.evtdev;
189
190	/*
191	 * Is the device in broadcast mode forever or is it not
192	 * affected by the powerstate ?
193	 */
194	if (!dev || !tick_device_is_functional(dev) ||
195	    !(dev->features & CLOCK_EVT_FEAT_C3STOP))
196		goto out;
197
198	if (*reason == CLOCK_EVT_NOTIFY_BROADCAST_ON) {
199		if (!cpu_isset(cpu, tick_broadcast_mask)) {
200			cpu_set(cpu, tick_broadcast_mask);
201			if (td->mode == TICKDEV_MODE_PERIODIC)
202				clockevents_set_mode(dev,
203						     CLOCK_EVT_MODE_SHUTDOWN);
204		}
205	} else {
206		if (cpu_isset(cpu, tick_broadcast_mask)) {
207			cpu_clear(cpu, tick_broadcast_mask);
208			if (td->mode == TICKDEV_MODE_PERIODIC)
209				tick_setup_periodic(dev, 0);
210		}
211	}
212
213	if (cpus_empty(tick_broadcast_mask))
214		clockevents_set_mode(bc, CLOCK_EVT_MODE_SHUTDOWN);
215	else {
216		if (tick_broadcast_device.mode == TICKDEV_MODE_PERIODIC)
217			tick_broadcast_start_periodic(bc);
218		else
219			tick_broadcast_setup_oneshot(bc);
220	}
221out:
222	spin_unlock_irqrestore(&tick_broadcast_lock, flags);
223}
224
225/*
226 * Powerstate information: The system enters/leaves a state, where
227 * affected devices might stop.
228 */
229void tick_broadcast_on_off(unsigned long reason, int *oncpu)
230{
231	int cpu = get_cpu();
232
233	if (cpu == *oncpu)
234		tick_do_broadcast_on_off(&reason);
235	else
236		smp_call_function_single(*oncpu, tick_do_broadcast_on_off,
237					 &reason, 1, 1);
238	put_cpu();
239}
240
241/*
242 * Set the periodic handler depending on broadcast on/off
243 */
244void tick_set_periodic_handler(struct clock_event_device *dev, int broadcast)
245{
246	if (!broadcast)
247		dev->event_handler = tick_handle_periodic;
248	else
249		dev->event_handler = tick_handle_periodic_broadcast;
250}
251
252/*
253 * Remove a CPU from broadcasting
254 */
255void tick_shutdown_broadcast(unsigned int *cpup)
256{
257	struct clock_event_device *bc;
258	unsigned long flags;
259	unsigned int cpu = *cpup;
260
261	spin_lock_irqsave(&tick_broadcast_lock, flags);
262
263	bc = tick_broadcast_device.evtdev;
264	cpu_clear(cpu, tick_broadcast_mask);
265
266	if (tick_broadcast_device.mode == TICKDEV_MODE_PERIODIC) {
267		if (bc && cpus_empty(tick_broadcast_mask))
268			clockevents_set_mode(bc, CLOCK_EVT_MODE_SHUTDOWN);
269	}
270
271	spin_unlock_irqrestore(&tick_broadcast_lock, flags);
272}
273
274#ifdef CONFIG_TICK_ONESHOT
275
276static cpumask_t tick_broadcast_oneshot_mask;
277
278static int tick_broadcast_set_event(ktime_t expires, int force)
279{
280	struct clock_event_device *bc = tick_broadcast_device.evtdev;
281	ktime_t now = ktime_get();
282	int res;
283
284	for(;;) {
285		res = clockevents_program_event(bc, expires, now);
286		if (!res || !force)
287			return res;
288		now = ktime_get();
289		expires = ktime_add(now, ktime_set(0, bc->min_delta_ns));
290	}
291}
292
293/*
294 * Reprogram the broadcast device:
295 *
296 * Called with tick_broadcast_lock held and interrupts disabled.
297 */
298static int tick_broadcast_reprogram(void)
299{
300	ktime_t expires = { .tv64 = KTIME_MAX };
301	struct tick_device *td;
302	int cpu;
303
304	/*
305	 * Find the event which expires next:
306	 */
307	for (cpu = first_cpu(tick_broadcast_oneshot_mask); cpu != NR_CPUS;
308	     cpu = next_cpu(cpu, tick_broadcast_oneshot_mask)) {
309		td = &per_cpu(tick_cpu_device, cpu);
310		if (td->evtdev->next_event.tv64 < expires.tv64)
311			expires = td->evtdev->next_event;
312	}
313
314	if (expires.tv64 == KTIME_MAX)
315		return 0;
316
317	return tick_broadcast_set_event(expires, 0);
318}
319
320/*
321 * Handle oneshot mode broadcasting
322 */
323static void tick_handle_oneshot_broadcast(struct clock_event_device *dev)
324{
325	struct tick_device *td;
326	cpumask_t mask;
327	ktime_t now;
328	int cpu;
329
330	spin_lock(&tick_broadcast_lock);
331again:
332	dev->next_event.tv64 = KTIME_MAX;
333	mask = CPU_MASK_NONE;
334	now = ktime_get();
335	/* Find all expired events */
336	for (cpu = first_cpu(tick_broadcast_oneshot_mask); cpu != NR_CPUS;
337	     cpu = next_cpu(cpu, tick_broadcast_oneshot_mask)) {
338		td = &per_cpu(tick_cpu_device, cpu);
339		if (td->evtdev->next_event.tv64 <= now.tv64)
340			cpu_set(cpu, mask);
341	}
342
343	/*
344	 * Wakeup the cpus which have an expired event. The broadcast
345	 * device is reprogrammed in the return from idle code.
346	 */
347	if (!tick_do_broadcast(mask)) {
348		/*
349		 * The global event did not expire any CPU local
350		 * events. This happens in dyntick mode, as the
351		 * maximum PIT delta is quite small.
352		 */
353		if (tick_broadcast_reprogram())
354			goto again;
355	}
356	spin_unlock(&tick_broadcast_lock);
357}
358
359/*
360 * Powerstate information: The system enters/leaves a state, where
361 * affected devices might stop
362 */
363void tick_broadcast_oneshot_control(unsigned long reason)
364{
365	struct clock_event_device *bc, *dev;
366	struct tick_device *td;
367	unsigned long flags;
368	int cpu;
369
370	spin_lock_irqsave(&tick_broadcast_lock, flags);
371
372	/*
373	 * Periodic mode does not care about the enter/exit of power
374	 * states
375	 */
376	if (tick_broadcast_device.mode == TICKDEV_MODE_PERIODIC)
377		goto out;
378
379	bc = tick_broadcast_device.evtdev;
380	cpu = smp_processor_id();
381	td = &per_cpu(tick_cpu_device, cpu);
382	dev = td->evtdev;
383
384	if (!(dev->features & CLOCK_EVT_FEAT_C3STOP))
385		goto out;
386
387	if (reason == CLOCK_EVT_NOTIFY_BROADCAST_ENTER) {
388		if (!cpu_isset(cpu, tick_broadcast_oneshot_mask)) {
389			cpu_set(cpu, tick_broadcast_oneshot_mask);
390			clockevents_set_mode(dev, CLOCK_EVT_MODE_SHUTDOWN);
391			if (dev->next_event.tv64 < bc->next_event.tv64)
392				tick_broadcast_set_event(dev->next_event, 1);
393		}
394	} else {
395		if (cpu_isset(cpu, tick_broadcast_oneshot_mask)) {
396			cpu_clear(cpu, tick_broadcast_oneshot_mask);
397			clockevents_set_mode(dev, CLOCK_EVT_MODE_ONESHOT);
398			if (dev->next_event.tv64 != KTIME_MAX)
399				tick_program_event(dev->next_event, 1);
400		}
401	}
402
403out:
404	spin_unlock_irqrestore(&tick_broadcast_lock, flags);
405}
406
407/**
408 * tick_broadcast_setup_highres - setup the broadcast device for highres
409 */
410void tick_broadcast_setup_oneshot(struct clock_event_device *bc)
411{
412	if (bc->mode != CLOCK_EVT_MODE_ONESHOT) {
413		bc->event_handler = tick_handle_oneshot_broadcast;
414		clockevents_set_mode(bc, CLOCK_EVT_MODE_ONESHOT);
415		bc->next_event.tv64 = KTIME_MAX;
416	}
417}
418
419/*
420 * Select oneshot operating mode for the broadcast device
421 */
422void tick_broadcast_switch_to_oneshot(void)
423{
424	struct clock_event_device *bc;
425	unsigned long flags;
426
427	spin_lock_irqsave(&tick_broadcast_lock, flags);
428
429	tick_broadcast_device.mode = TICKDEV_MODE_ONESHOT;
430	bc = tick_broadcast_device.evtdev;
431	if (bc)
432		tick_broadcast_setup_oneshot(bc);
433	spin_unlock_irqrestore(&tick_broadcast_lock, flags);
434}
435
436
437/*
438 * Remove a dead CPU from broadcasting
439 */
440void tick_shutdown_broadcast_oneshot(unsigned int *cpup)
441{
442	struct clock_event_device *bc;
443	unsigned long flags;
444	unsigned int cpu = *cpup;
445
446	spin_lock_irqsave(&tick_broadcast_lock, flags);
447
448	bc = tick_broadcast_device.evtdev;
449	cpu_clear(cpu, tick_broadcast_oneshot_mask);
450
451	if (tick_broadcast_device.mode == TICKDEV_MODE_ONESHOT) {
452		if (bc && cpus_empty(tick_broadcast_oneshot_mask))
453			clockevents_set_mode(bc, CLOCK_EVT_MODE_SHUTDOWN);
454	}
455
456	spin_unlock_irqrestore(&tick_broadcast_lock, flags);
457}
458
459#endif
460