/kernel/time/ |
H A D | timekeeping_internal.h | 16 static inline cycle_t clocksource_delta(cycle_t now, cycle_t last, cycle_t mask) argument 18 cycle_t ret = (now - last) & mask; 23 static inline cycle_t clocksource_delta(cycle_t now, cycle_t last, cycle_t mask) argument 25 return (now - last) & mask;
|
H A D | tick-sched.c | 52 static void tick_do_update_jiffies64(ktime_t now) argument 60 delta = ktime_sub(now, last_jiffies_update); 67 delta = ktime_sub(now, last_jiffies_update); 112 static void tick_sched_do_timer(ktime_t now) argument 131 tick_do_update_jiffies64(now); 200 static void tick_nohz_restart_sched_tick(struct tick_sched *ts, ktime_t now); 436 static void tick_nohz_update_jiffies(ktime_t now) argument 440 __this_cpu_write(tick_cpu_sched.idle_waketime, now); 443 tick_do_update_jiffies64(now); 453 update_ts_time_stats(int cpu, struct tick_sched *ts, ktime_t now, u6 argument 471 tick_nohz_stop_idle(struct tick_sched *ts, ktime_t now) argument 473 update_ts_time_stats(smp_processor_id(), ts, now, NULL); local 481 ktime_t now = ktime_get(); local 506 ktime_t now, idle; local 547 ktime_t now, iowait; local 570 tick_nohz_stop_sched_tick(struct tick_sched *ts, ktime_t now, int cpu) argument 795 ktime_t now, expires; local 882 tick_nohz_restart(struct tick_sched *ts, ktime_t now) argument 908 tick_nohz_restart_sched_tick(struct tick_sched *ts, ktime_t now) argument 956 ktime_t now; local 979 tick_nohz_reprogram(struct tick_sched *ts, ktime_t now) argument 992 ktime_t now = ktime_get(); local 1056 tick_nohz_kick_tick(struct tick_sched *ts, ktime_t now) argument 1077 ktime_t now; local 1119 ktime_t now = ktime_get(); local 1155 ktime_t now = ktime_get(); local [all...] |
H A D | timer_list.c | 27 u64 now; member in struct:timer_list_iter 58 int idx, u64 now) 79 (long long)(ktime_to_ns(hrtimer_get_softexpires(timer)) - now), 80 (long long)(ktime_to_ns(hrtimer_get_expires(timer)) - now)); 85 u64 now) 112 print_timer(m, timer, &tmp, i, now); 120 print_base(struct seq_file *m, struct hrtimer_clock_base *base, u64 now) argument 135 print_active_timers(m, base, now); 138 static void print_cpu(struct seq_file *m, int cpu, u64 now) argument 146 print_base(m, cpu_base->clock_base + i, now); 57 print_timer(struct seq_file *m, struct hrtimer *taddr, struct hrtimer *timer, int idx, u64 now) argument 84 print_active_timers(struct seq_file *m, struct hrtimer_clock_base *base, u64 now) argument 257 timer_list_header(struct seq_file *m, u64 now) argument 284 u64 now = ktime_to_ns(ktime_get()); local [all...] |
H A D | alarmtimer.c | 223 ktime_t min, now; local 265 now = rtc_tm_to_ktime(tm); 266 now = ktime_add(now, min); 269 ret = rtc_timer_start(rtc, &rtctimer, now, ktime_set(0, 0)); 339 * @start: time relative to now to run the alarm 404 u64 alarm_forward(struct alarm *alarm, ktime_t now, ktime_t interval) argument 409 delta = ktime_sub(now, alarm->node.expires); 422 if (alarm->node.expires.tv64 > now.tv64) 465 ktime_t now) 464 alarm_handle_timer(struct alarm *alarm, ktime_t now) argument 620 ktime_t now; local 636 alarmtimer_nsleep_wakeup(struct alarm *alarm, ktime_t now) argument 769 ktime_t now = alarm_bases[type].gettime(); local [all...] |
H A D | timekeeping.c | 259 * now = now(tkf->base[idx]); 295 * now = base_mono + clock_delta * slope 325 u64 now; local 330 now = ktime_to_ns(tkr->base_mono) + timekeeping_get_ns(tkr); 333 return now; 424 * nsec = (xtime_sec + wtm_sec) * 1e9 + wtm_nsec + now(); 426 * nsec = base_mono + now(); 697 struct timespec64 now; local 699 getnstimeofday64(&now); 998 struct timespec64 now, boot, tmp; local 1636 struct timespec64 now; local 1652 struct timespec64 now, mono; local [all...] |
H A D | posix-cpu-timers.c | 83 unsigned long long now) 91 if (now < timer->it.cpu.expires) 95 delta = now + incr - timer->it.cpu.expires; 747 unsigned long long now; local 767 cpu_clock_sample(timer->it_clock, p, &now); 788 cpu_timer_sample_group(timer->it_clock, p, &now); 793 if (now < timer->it.cpu.expires) { 795 timer->it.cpu.expires - now, 864 * No need to calculate anything else now. 970 * No need to calculate anything else now 82 bump_cpu_timer(struct k_itimer *timer, unsigned long long now) argument 1007 unsigned long long now; local 1204 unsigned long long now; local [all...] |
H A D | timer.c | 151 /* now that we have rounded, subtract the extra skew again */ 1302 static unsigned long cmp_next_hrtimer_event(unsigned long now, argument 1316 return now + 1; 1336 now += delta; 1337 if (time_before(now, expires)) 1338 return now; 1344 * @now: current time (in jiffies) 1346 unsigned long get_next_timer_interrupt(unsigned long now) argument 1349 unsigned long expires = now + NEXT_TIMER_MAX_DELTA; 1366 if (time_before_eq(expires, now)) [all...] |
H A D | tick-broadcast.c | 214 * switched off now. 273 * have different broadcast functions. For now, just use the 577 ktime_t now, next_event; local 585 now = ktime_get(); 589 if (td->evtdev->next_event.tv64 <= now.tv64) { 692 ktime_t now; local 795 now = ktime_get(); 796 if (dev->next_event.tv64 <= now.tv64) {
|
H A D | hrtimer.c | 528 * T1 expires 50ms from now 529 * T2 expires 5s from now 532 * the hardware to 5s from now. Any hrtimer_start after that 581 * has now objections against negative expiry values. 784 * @now: forward past this time 790 u64 hrtimer_forward(struct hrtimer *timer, ktime_t now, ktime_t interval) argument 795 delta = ktime_sub(now, hrtimer_get_expires(timer)); 808 if (hrtimer_get_expires_tv64(timer) > now.tv64) 1197 static void __run_hrtimer(struct hrtimer *timer, ktime_t *now) argument 1217 trace_hrtimer_expire_entry(timer, now); 1246 ktime_t expires_next, now, entry_time, delta; local [all...] |
H A D | posix-timers.c | 55 * Management arrays for POSIX timers. Timers are now kept in static hash table 59 * This keeps timer ids unique per process, but now they can intersect between 455 ktime_t now = hrtimer_cb_get_time(timer); local 465 * For now we just let timers with an interval 470 * list. By moving now ahead of time we trick 484 now = ktime_add(now, kj); 488 hrtimer_forward(timer, now, 720 * now. 731 ktime_t now, remainin local [all...] |
H A D | ntp.c | 469 struct timespec64 now; local 489 getnstimeofday64(&now); 490 if (abs(now.tv_nsec - (NSEC_PER_SEC / 2)) <= tick_nsec * 5) { 491 struct timespec adjust = timespec64_to_timespec(now); 505 next.tv_nsec = (NSEC_PER_SEC / 2) - now.tv_nsec - (TICK_NSEC / 2); 898 /* ok, now we have a base for frequency calculation */
|
H A D | time.c | 234 struct timespec now = current_kernel_time(); local 235 return timespec_trunc(now, sb->s_time_gran); 338 )*24 + hour /* now have hours */ 339 )*60 + min /* now have minutes */
|
/kernel/trace/ |
H A D | trace_clock.c | 94 u64 now; local 99 now = sched_clock_cpu(this_cpu); 112 * we start ticking with the local clock from now on? 114 if ((s64)(now - trace_clock_struct.prev_time) < 0) 115 now = trace_clock_struct.prev_time + 1; 117 trace_clock_struct.prev_time = now; 124 return now;
|
H A D | blktrace.c | 130 struct timespec now; local 134 getnstimeofday(&now); 135 words[0] = now.tv_sec; 136 words[1] = now.tv_nsec;
|
/kernel/sched/ |
H A D | stats.h | 64 unsigned long long now = rq_clock(rq), delta = 0; local 68 delta = now - t->sched_info.last_queued; 76 * Called when a task finally hits the cpu. We can now calculate how 82 unsigned long long now = rq_clock(rq), delta = 0; local 85 delta = now - t->sched_info.last_queued; 88 t->sched_info.last_arrival = now; 111 * sched_info_queued() to mark that it has now again started waiting on 135 * prev now departs the cpu. It's not interesting to record
|
H A D | clock.c | 190 * update the percpu scd from the raw @now value 197 u64 now, clock, old_clock, min_clock, max_clock; local 201 now = sched_clock(); 202 delta = now - scd->tick_raw; 320 u64 now, now_gtod; local 332 now = sched_clock(); 334 scd->tick_raw = now;
|
H A D | wait.c | 538 unsigned long now = ACCESS_ONCE(jiffies); local 541 if (time_after_eq(now, word->timeout)) 543 schedule_timeout(word->timeout - now); 550 unsigned long now = ACCESS_ONCE(jiffies); local 553 if (time_after_eq(now, word->timeout)) 555 io_schedule_timeout(word->timeout - now);
|
H A D | fair.c | 697 u64 now = rq_clock_task(rq_of(cfs_rq)); local 703 delta_exec = now - curr->exec_start; 707 curr->exec_start = now; 1410 * alternative node to recheck if the tasks is now properly placed. 1564 u64 runtime, delta, now; local 1566 now = p->se.exec_start; 1571 *period = now - p->last_task_numa_placement; 1578 p->last_task_numa_placement = now; 1919 unsigned long migrate, next_scan, now = jiffies; local 1942 mm->numa_next_scan = now 2044 u64 period, now; local 2318 __update_entity_runnable_avg(u64 now, struct sched_avg *sa, int runnable) argument 2541 u64 now; local 2572 u64 now = cfs_rq_clock_task(cfs_rq) >> 20; local 3190 u64 now; local 3760 ktime_t now; local 5592 unsigned long now = jiffies; local 7407 unsigned long now = jiffies; local [all...] |
H A D | deadline.c | 454 ktime_t now, act; local 467 now = hrtimer_cb_get_time(&dl_se->dl_timer); 468 delta = ktime_to_ns(now) - rq_clock(rq); 476 if (ktime_us_delta(act, now) < 0) 649 * Because -- for now -- we share the rt bandwidth, we need to 653 * Account to the root rt group for now. 874 * its budget it needs a replenishment and, since it now is on 903 * a new runtime. This is of little use now, since we 1501 * anytime soon on its runqueue, we try pushing it away now. 1543 * The process used to be able to migrate OR it can now migrat [all...] |
/kernel/power/ |
H A D | suspend_test.c | 74 unsigned long now; local 85 rtc_tm_to_time(&alm.time, &now); 88 rtc_time_to_tm(now + TEST_SUSPEND_SECONDS, &alm.time); 189 /* PM is initialized by now; is that state testable? */ 204 /* RTCs have initialized by now too ... can we use one? */
|
H A D | wakelock.c | 102 ktime_t now; local 107 now = ktime_get(); 113 idle_time_ns = ktime_to_ns(ktime_sub(now, wl->ws.last_time));
|
/kernel/ |
H A D | watchdog.c | 315 unsigned long now = get_timestamp(); local 318 if (time_after(now, touch_ts + get_softlockup_thresh())) 319 return now - touch_ts; 706 * because it will reprogram itself with the new period now. 721 * be much nicer but we do not have an API for that now so
|
H A D | audit.c | 220 unsigned long now; local 230 now = jiffies; 231 elapsed = now - last_check; 233 last_check = now; 256 unsigned long now; local 265 now = jiffies; 266 if (now - last_msg > HZ) { 268 last_msg = now; 469 * are messages in this queue. I don't see the need now, but 642 /* Only support initial user namespace for now [all...] |
/kernel/events/ |
H A D | core.c | 426 u64 now; local 428 now = perf_clock(); 432 info->time += now - info->timestamp; 433 info->timestamp = now; 642 perf_cgroup_set_shadow_time(struct perf_event *event, u64 now) argument 646 event->shadow_ctx_time = now - t->timestamp; 740 perf_cgroup_set_shadow_time(struct perf_event *event, u64 now) argument 1049 u64 now = perf_clock(); local 1051 ctx->time += now - ctx->timestamp; 1052 ctx->timestamp = now; 1808 u64 now = ctx->time; local 2559 u64 now; local 2833 u64 now, period = TICK_NSEC; local 3919 calc_timer_values(struct perf_event *event, u64 *now, u64 *enabled, u64 *running) argument 3952 arch_perf_update_userpage(struct perf_event_mmap_page *userpg, u64 now) argument 3965 u64 enabled, running, now; local 4717 u64 enabled = 0, running = 0, now; local 5604 u64 now = perf_clock(); local 6362 u64 now; local 6433 task_clock_event_update(struct perf_event *event, u64 now) argument 6470 u64 now = perf_clock(); local [all...] |
/kernel/debug/kdb/ |
H A D | kdb_main.c | 804 /* now cp points to a nonzero length search string */ 2492 struct timespec now; local 2507 now = __current_kernel_time(); 2508 kdb_gmtime(&now, &tm);
|