/kernel/time/ |
H A D | tick-oneshot.c | 27 int tick_program_event(ktime_t expires, int force) argument 31 return clockevents_program_event(dev, expires, force);
|
H A D | tick-internal.h | 45 extern int tick_program_event(ktime_t expires, int force); 83 static inline int tick_program_event(ktime_t expires, int force) argument
|
H A D | clockevents.c | 237 * @force: program minimum delay if expires can not be set 242 bool force) 264 return force ? clockevents_program_min_delta(dev) : -ETIME; 272 return (rc && force) ? clockevents_program_min_delta(dev) : rc; 241 clockevents_program_event(struct clock_event_device *dev, ktime_t expires, bool force) argument
|
H A D | tick-broadcast.c | 531 ktime_t expires, int force) 538 ret = clockevents_program_event(bc, expires, force); 722 * did not mark ourself in the force mask and 725 * the force mask, then we are going to be 530 tick_broadcast_set_event(struct clock_event_device *bc, int cpu, ktime_t expires, int force) argument
|
/kernel/ |
H A D | res_counter.c | 36 unsigned long val, bool force) 43 if (!force) 54 struct res_counter **limit_fail_at, bool force) 65 r = res_counter_charge_locked(c, val, force); 70 if (!force) 75 if (ret < 0 && !force) { 35 res_counter_charge_locked(struct res_counter *counter, unsigned long val, bool force) argument 53 __res_counter_charge(struct res_counter *counter, unsigned long val, struct res_counter **limit_fail_at, bool force) argument
|
H A D | reboot.c | 391 static int __orderly_poweroff(bool force) argument 409 if (ret && force) { 434 * @force: force poweroff if command execution fails 437 * If the orderly shutdown fails, it will force an immediate shutdown. 439 int orderly_poweroff(bool force) argument 441 if (force) /* do not override the pending "true" */
|
H A D | signal.c | 68 static int sig_task_ignored(struct task_struct *t, int sig, bool force) argument 75 handler == SIG_DFL && !force) 81 static int sig_ignored(struct task_struct *t, int sig, bool force) argument 91 if (!sig_task_ignored(t, sig, force)) 835 static bool prepare_signal(int sig, struct task_struct *p, bool force) argument 897 return !sig_ignored(p, sig, force); 1485 * will force a SIGSEGV. And if the signal that caused
|
H A D | kprobes.c | 629 static void unoptimize_kprobe(struct kprobe *p, bool force) argument 639 if (force && !list_empty(&op->list)) { 658 if (force)
|
/kernel/irq/ |
H A D | spurious.c | 63 static int try_one_irq(int irq, struct irq_desc *desc, bool force) argument 83 if (irqd_irq_disabled(&desc->irq_data) && !force)
|
H A D | internals.h | 26 * IRQTF_FORCED_THREAD - irq action is force threaded 114 const struct cpumask *dest, bool force);
|
H A D | manage.c | 177 bool force) 183 ret = chip->irq_set_affinity(data, mask, force); 196 bool force) 206 ret = irq_do_set_affinity(data, mask, force); 221 int __irq_set_affinity(unsigned int irq, const struct cpumask *mask, bool force) argument 231 ret = irq_set_affinity_locked(irq_desc_get_irq_data(desc), mask, force); 176 irq_do_set_affinity(struct irq_data *data, const struct cpumask *mask, bool force) argument 195 irq_set_affinity_locked(struct irq_data *data, const struct cpumask *mask, bool force) argument
|
/kernel/rcu/ |
H A D | tree_plugin.h | 786 * synchronize_rcu_expedited - Brute-force RCU grace period 847 /* force all RCU readers onto ->blkd_tasks lists. */ 1683 invoke_rcu_core(); /* force nohz to see update. */ 2039 static void wake_nocb_leader(struct rcu_data *rdp, bool force) argument 2045 if (ACCESS_ONCE(rdp_leader->nocb_leader_sleep) || force) { 2799 * Unconditionally force exit from full system-idle state. This is 3002 * Update the sysidle state based on the results of a force-quiescent-state
|
/kernel/sched/ |
H A D | sched.h | 289 extern void __start_cfs_bandwidth(struct cfs_bandwidth *cfs_b, bool force);
|
H A D | fair.c | 294 * enqueued) or force our parent to appear after us when it is 2638 /* we force update consideration on load-balancer moves */ 2652 /* we force update consideration on load-balancer moves */ 3800 void __start_cfs_bandwidth(struct cfs_bandwidth *cfs_b, bool force) argument 3815 if (!force && cfs_b->timer_active) 6381 * a think about bumping its value to force at least one task to be 6581 * ASYM_PACKING needs to force migrate tasks from busy but
|
/kernel/locking/ |
H A D | lockdep.c | 726 register_lock_class(struct lockdep_map *lock, unsigned int subclass, int force) argument 820 if (!subclass || force)
|