162306a36Sopenharmony_ci// SPDX-License-Identifier: GPL-2.0 262306a36Sopenharmony_ci/* 362306a36Sopenharmony_ci * Copyright(C) 2005-2006, Thomas Gleixner <tglx@linutronix.de> 462306a36Sopenharmony_ci * Copyright(C) 2005-2007, Red Hat, Inc., Ingo Molnar 562306a36Sopenharmony_ci * Copyright(C) 2006-2007 Timesys Corp., Thomas Gleixner 662306a36Sopenharmony_ci * 762306a36Sopenharmony_ci * No idle tick implementation for low and high resolution timers 862306a36Sopenharmony_ci * 962306a36Sopenharmony_ci * Started by: Thomas Gleixner and Ingo Molnar 1062306a36Sopenharmony_ci */ 1162306a36Sopenharmony_ci#include <linux/cpu.h> 1262306a36Sopenharmony_ci#include <linux/err.h> 1362306a36Sopenharmony_ci#include <linux/hrtimer.h> 1462306a36Sopenharmony_ci#include <linux/interrupt.h> 1562306a36Sopenharmony_ci#include <linux/kernel_stat.h> 1662306a36Sopenharmony_ci#include <linux/percpu.h> 1762306a36Sopenharmony_ci#include <linux/nmi.h> 1862306a36Sopenharmony_ci#include <linux/profile.h> 1962306a36Sopenharmony_ci#include <linux/sched/signal.h> 2062306a36Sopenharmony_ci#include <linux/sched/clock.h> 2162306a36Sopenharmony_ci#include <linux/sched/stat.h> 2262306a36Sopenharmony_ci#include <linux/sched/nohz.h> 2362306a36Sopenharmony_ci#include <linux/sched/loadavg.h> 2462306a36Sopenharmony_ci#include <linux/module.h> 2562306a36Sopenharmony_ci#include <linux/irq_work.h> 2662306a36Sopenharmony_ci#include <linux/posix-timers.h> 2762306a36Sopenharmony_ci#include <linux/context_tracking.h> 2862306a36Sopenharmony_ci#include <linux/mm.h> 2962306a36Sopenharmony_ci 3062306a36Sopenharmony_ci#include <asm/irq_regs.h> 3162306a36Sopenharmony_ci 3262306a36Sopenharmony_ci#include "tick-internal.h" 3362306a36Sopenharmony_ci 3462306a36Sopenharmony_ci#include <trace/events/timer.h> 3562306a36Sopenharmony_ci 3662306a36Sopenharmony_ci/* 3762306a36Sopenharmony_ci * Per-CPU nohz control structure 3862306a36Sopenharmony_ci */ 3962306a36Sopenharmony_cistatic DEFINE_PER_CPU(struct tick_sched, tick_cpu_sched); 4062306a36Sopenharmony_ci 4162306a36Sopenharmony_cistruct tick_sched *tick_get_tick_sched(int cpu) 4262306a36Sopenharmony_ci{ 4362306a36Sopenharmony_ci return &per_cpu(tick_cpu_sched, cpu); 4462306a36Sopenharmony_ci} 4562306a36Sopenharmony_ci 4662306a36Sopenharmony_ci#if defined(CONFIG_NO_HZ_COMMON) || defined(CONFIG_HIGH_RES_TIMERS) 4762306a36Sopenharmony_ci/* 4862306a36Sopenharmony_ci * The time, when the last jiffy update happened. Write access must hold 4962306a36Sopenharmony_ci * jiffies_lock and jiffies_seq. tick_nohz_next_event() needs to get a 5062306a36Sopenharmony_ci * consistent view of jiffies and last_jiffies_update. 5162306a36Sopenharmony_ci */ 5262306a36Sopenharmony_cistatic ktime_t last_jiffies_update; 5362306a36Sopenharmony_ci 5462306a36Sopenharmony_ci/* 5562306a36Sopenharmony_ci * Must be called with interrupts disabled ! 5662306a36Sopenharmony_ci */ 5762306a36Sopenharmony_cistatic void tick_do_update_jiffies64(ktime_t now) 5862306a36Sopenharmony_ci{ 5962306a36Sopenharmony_ci unsigned long ticks = 1; 6062306a36Sopenharmony_ci ktime_t delta, nextp; 6162306a36Sopenharmony_ci 6262306a36Sopenharmony_ci /* 6362306a36Sopenharmony_ci * 64bit can do a quick check without holding jiffies lock and 6462306a36Sopenharmony_ci * without looking at the sequence count. The smp_load_acquire() 6562306a36Sopenharmony_ci * pairs with the update done later in this function. 6662306a36Sopenharmony_ci * 6762306a36Sopenharmony_ci * 32bit cannot do that because the store of tick_next_period 6862306a36Sopenharmony_ci * consists of two 32bit stores and the first store could move it 6962306a36Sopenharmony_ci * to a random point in the future. 7062306a36Sopenharmony_ci */ 7162306a36Sopenharmony_ci if (IS_ENABLED(CONFIG_64BIT)) { 7262306a36Sopenharmony_ci if (ktime_before(now, smp_load_acquire(&tick_next_period))) 7362306a36Sopenharmony_ci return; 7462306a36Sopenharmony_ci } else { 7562306a36Sopenharmony_ci unsigned int seq; 7662306a36Sopenharmony_ci 7762306a36Sopenharmony_ci /* 7862306a36Sopenharmony_ci * Avoid contention on jiffies_lock and protect the quick 7962306a36Sopenharmony_ci * check with the sequence count. 8062306a36Sopenharmony_ci */ 8162306a36Sopenharmony_ci do { 8262306a36Sopenharmony_ci seq = read_seqcount_begin(&jiffies_seq); 8362306a36Sopenharmony_ci nextp = tick_next_period; 8462306a36Sopenharmony_ci } while (read_seqcount_retry(&jiffies_seq, seq)); 8562306a36Sopenharmony_ci 8662306a36Sopenharmony_ci if (ktime_before(now, nextp)) 8762306a36Sopenharmony_ci return; 8862306a36Sopenharmony_ci } 8962306a36Sopenharmony_ci 9062306a36Sopenharmony_ci /* Quick check failed, i.e. update is required. */ 9162306a36Sopenharmony_ci raw_spin_lock(&jiffies_lock); 9262306a36Sopenharmony_ci /* 9362306a36Sopenharmony_ci * Reevaluate with the lock held. Another CPU might have done the 9462306a36Sopenharmony_ci * update already. 9562306a36Sopenharmony_ci */ 9662306a36Sopenharmony_ci if (ktime_before(now, tick_next_period)) { 9762306a36Sopenharmony_ci raw_spin_unlock(&jiffies_lock); 9862306a36Sopenharmony_ci return; 9962306a36Sopenharmony_ci } 10062306a36Sopenharmony_ci 10162306a36Sopenharmony_ci write_seqcount_begin(&jiffies_seq); 10262306a36Sopenharmony_ci 10362306a36Sopenharmony_ci delta = ktime_sub(now, tick_next_period); 10462306a36Sopenharmony_ci if (unlikely(delta >= TICK_NSEC)) { 10562306a36Sopenharmony_ci /* Slow path for long idle sleep times */ 10662306a36Sopenharmony_ci s64 incr = TICK_NSEC; 10762306a36Sopenharmony_ci 10862306a36Sopenharmony_ci ticks += ktime_divns(delta, incr); 10962306a36Sopenharmony_ci 11062306a36Sopenharmony_ci last_jiffies_update = ktime_add_ns(last_jiffies_update, 11162306a36Sopenharmony_ci incr * ticks); 11262306a36Sopenharmony_ci } else { 11362306a36Sopenharmony_ci last_jiffies_update = ktime_add_ns(last_jiffies_update, 11462306a36Sopenharmony_ci TICK_NSEC); 11562306a36Sopenharmony_ci } 11662306a36Sopenharmony_ci 11762306a36Sopenharmony_ci /* Advance jiffies to complete the jiffies_seq protected job */ 11862306a36Sopenharmony_ci jiffies_64 += ticks; 11962306a36Sopenharmony_ci 12062306a36Sopenharmony_ci /* 12162306a36Sopenharmony_ci * Keep the tick_next_period variable up to date. 12262306a36Sopenharmony_ci */ 12362306a36Sopenharmony_ci nextp = ktime_add_ns(last_jiffies_update, TICK_NSEC); 12462306a36Sopenharmony_ci 12562306a36Sopenharmony_ci if (IS_ENABLED(CONFIG_64BIT)) { 12662306a36Sopenharmony_ci /* 12762306a36Sopenharmony_ci * Pairs with smp_load_acquire() in the lockless quick 12862306a36Sopenharmony_ci * check above and ensures that the update to jiffies_64 is 12962306a36Sopenharmony_ci * not reordered vs. the store to tick_next_period, neither 13062306a36Sopenharmony_ci * by the compiler nor by the CPU. 13162306a36Sopenharmony_ci */ 13262306a36Sopenharmony_ci smp_store_release(&tick_next_period, nextp); 13362306a36Sopenharmony_ci } else { 13462306a36Sopenharmony_ci /* 13562306a36Sopenharmony_ci * A plain store is good enough on 32bit as the quick check 13662306a36Sopenharmony_ci * above is protected by the sequence count. 13762306a36Sopenharmony_ci */ 13862306a36Sopenharmony_ci tick_next_period = nextp; 13962306a36Sopenharmony_ci } 14062306a36Sopenharmony_ci 14162306a36Sopenharmony_ci /* 14262306a36Sopenharmony_ci * Release the sequence count. calc_global_load() below is not 14362306a36Sopenharmony_ci * protected by it, but jiffies_lock needs to be held to prevent 14462306a36Sopenharmony_ci * concurrent invocations. 14562306a36Sopenharmony_ci */ 14662306a36Sopenharmony_ci write_seqcount_end(&jiffies_seq); 14762306a36Sopenharmony_ci 14862306a36Sopenharmony_ci calc_global_load(); 14962306a36Sopenharmony_ci 15062306a36Sopenharmony_ci raw_spin_unlock(&jiffies_lock); 15162306a36Sopenharmony_ci update_wall_time(); 15262306a36Sopenharmony_ci} 15362306a36Sopenharmony_ci 15462306a36Sopenharmony_ci/* 15562306a36Sopenharmony_ci * Initialize and return retrieve the jiffies update. 15662306a36Sopenharmony_ci */ 15762306a36Sopenharmony_cistatic ktime_t tick_init_jiffy_update(void) 15862306a36Sopenharmony_ci{ 15962306a36Sopenharmony_ci ktime_t period; 16062306a36Sopenharmony_ci 16162306a36Sopenharmony_ci raw_spin_lock(&jiffies_lock); 16262306a36Sopenharmony_ci write_seqcount_begin(&jiffies_seq); 16362306a36Sopenharmony_ci /* Did we start the jiffies update yet ? */ 16462306a36Sopenharmony_ci if (last_jiffies_update == 0) { 16562306a36Sopenharmony_ci u32 rem; 16662306a36Sopenharmony_ci 16762306a36Sopenharmony_ci /* 16862306a36Sopenharmony_ci * Ensure that the tick is aligned to a multiple of 16962306a36Sopenharmony_ci * TICK_NSEC. 17062306a36Sopenharmony_ci */ 17162306a36Sopenharmony_ci div_u64_rem(tick_next_period, TICK_NSEC, &rem); 17262306a36Sopenharmony_ci if (rem) 17362306a36Sopenharmony_ci tick_next_period += TICK_NSEC - rem; 17462306a36Sopenharmony_ci 17562306a36Sopenharmony_ci last_jiffies_update = tick_next_period; 17662306a36Sopenharmony_ci } 17762306a36Sopenharmony_ci period = last_jiffies_update; 17862306a36Sopenharmony_ci write_seqcount_end(&jiffies_seq); 17962306a36Sopenharmony_ci raw_spin_unlock(&jiffies_lock); 18062306a36Sopenharmony_ci return period; 18162306a36Sopenharmony_ci} 18262306a36Sopenharmony_ci 18362306a36Sopenharmony_ci#define MAX_STALLED_JIFFIES 5 18462306a36Sopenharmony_ci 18562306a36Sopenharmony_cistatic void tick_sched_do_timer(struct tick_sched *ts, ktime_t now) 18662306a36Sopenharmony_ci{ 18762306a36Sopenharmony_ci int cpu = smp_processor_id(); 18862306a36Sopenharmony_ci 18962306a36Sopenharmony_ci#ifdef CONFIG_NO_HZ_COMMON 19062306a36Sopenharmony_ci /* 19162306a36Sopenharmony_ci * Check if the do_timer duty was dropped. We don't care about 19262306a36Sopenharmony_ci * concurrency: This happens only when the CPU in charge went 19362306a36Sopenharmony_ci * into a long sleep. If two CPUs happen to assign themselves to 19462306a36Sopenharmony_ci * this duty, then the jiffies update is still serialized by 19562306a36Sopenharmony_ci * jiffies_lock. 19662306a36Sopenharmony_ci * 19762306a36Sopenharmony_ci * If nohz_full is enabled, this should not happen because the 19862306a36Sopenharmony_ci * tick_do_timer_cpu never relinquishes. 19962306a36Sopenharmony_ci */ 20062306a36Sopenharmony_ci if (unlikely(tick_do_timer_cpu == TICK_DO_TIMER_NONE)) { 20162306a36Sopenharmony_ci#ifdef CONFIG_NO_HZ_FULL 20262306a36Sopenharmony_ci WARN_ON_ONCE(tick_nohz_full_running); 20362306a36Sopenharmony_ci#endif 20462306a36Sopenharmony_ci tick_do_timer_cpu = cpu; 20562306a36Sopenharmony_ci } 20662306a36Sopenharmony_ci#endif 20762306a36Sopenharmony_ci 20862306a36Sopenharmony_ci /* Check, if the jiffies need an update */ 20962306a36Sopenharmony_ci if (tick_do_timer_cpu == cpu) 21062306a36Sopenharmony_ci tick_do_update_jiffies64(now); 21162306a36Sopenharmony_ci 21262306a36Sopenharmony_ci /* 21362306a36Sopenharmony_ci * If jiffies update stalled for too long (timekeeper in stop_machine() 21462306a36Sopenharmony_ci * or VMEXIT'ed for several msecs), force an update. 21562306a36Sopenharmony_ci */ 21662306a36Sopenharmony_ci if (ts->last_tick_jiffies != jiffies) { 21762306a36Sopenharmony_ci ts->stalled_jiffies = 0; 21862306a36Sopenharmony_ci ts->last_tick_jiffies = READ_ONCE(jiffies); 21962306a36Sopenharmony_ci } else { 22062306a36Sopenharmony_ci if (++ts->stalled_jiffies == MAX_STALLED_JIFFIES) { 22162306a36Sopenharmony_ci tick_do_update_jiffies64(now); 22262306a36Sopenharmony_ci ts->stalled_jiffies = 0; 22362306a36Sopenharmony_ci ts->last_tick_jiffies = READ_ONCE(jiffies); 22462306a36Sopenharmony_ci } 22562306a36Sopenharmony_ci } 22662306a36Sopenharmony_ci 22762306a36Sopenharmony_ci if (ts->inidle) 22862306a36Sopenharmony_ci ts->got_idle_tick = 1; 22962306a36Sopenharmony_ci} 23062306a36Sopenharmony_ci 23162306a36Sopenharmony_cistatic void tick_sched_handle(struct tick_sched *ts, struct pt_regs *regs) 23262306a36Sopenharmony_ci{ 23362306a36Sopenharmony_ci#ifdef CONFIG_NO_HZ_COMMON 23462306a36Sopenharmony_ci /* 23562306a36Sopenharmony_ci * When we are idle and the tick is stopped, we have to touch 23662306a36Sopenharmony_ci * the watchdog as we might not schedule for a really long 23762306a36Sopenharmony_ci * time. This happens on complete idle SMP systems while 23862306a36Sopenharmony_ci * waiting on the login prompt. We also increment the "start of 23962306a36Sopenharmony_ci * idle" jiffy stamp so the idle accounting adjustment we do 24062306a36Sopenharmony_ci * when we go busy again does not account too much ticks. 24162306a36Sopenharmony_ci */ 24262306a36Sopenharmony_ci if (ts->tick_stopped) { 24362306a36Sopenharmony_ci touch_softlockup_watchdog_sched(); 24462306a36Sopenharmony_ci if (is_idle_task(current)) 24562306a36Sopenharmony_ci ts->idle_jiffies++; 24662306a36Sopenharmony_ci /* 24762306a36Sopenharmony_ci * In case the current tick fired too early past its expected 24862306a36Sopenharmony_ci * expiration, make sure we don't bypass the next clock reprogramming 24962306a36Sopenharmony_ci * to the same deadline. 25062306a36Sopenharmony_ci */ 25162306a36Sopenharmony_ci ts->next_tick = 0; 25262306a36Sopenharmony_ci } 25362306a36Sopenharmony_ci#endif 25462306a36Sopenharmony_ci update_process_times(user_mode(regs)); 25562306a36Sopenharmony_ci profile_tick(CPU_PROFILING); 25662306a36Sopenharmony_ci} 25762306a36Sopenharmony_ci#endif 25862306a36Sopenharmony_ci 25962306a36Sopenharmony_ci#ifdef CONFIG_NO_HZ_FULL 26062306a36Sopenharmony_cicpumask_var_t tick_nohz_full_mask; 26162306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(tick_nohz_full_mask); 26262306a36Sopenharmony_cibool tick_nohz_full_running; 26362306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(tick_nohz_full_running); 26462306a36Sopenharmony_cistatic atomic_t tick_dep_mask; 26562306a36Sopenharmony_ci 26662306a36Sopenharmony_cistatic bool check_tick_dependency(atomic_t *dep) 26762306a36Sopenharmony_ci{ 26862306a36Sopenharmony_ci int val = atomic_read(dep); 26962306a36Sopenharmony_ci 27062306a36Sopenharmony_ci if (val & TICK_DEP_MASK_POSIX_TIMER) { 27162306a36Sopenharmony_ci trace_tick_stop(0, TICK_DEP_MASK_POSIX_TIMER); 27262306a36Sopenharmony_ci return true; 27362306a36Sopenharmony_ci } 27462306a36Sopenharmony_ci 27562306a36Sopenharmony_ci if (val & TICK_DEP_MASK_PERF_EVENTS) { 27662306a36Sopenharmony_ci trace_tick_stop(0, TICK_DEP_MASK_PERF_EVENTS); 27762306a36Sopenharmony_ci return true; 27862306a36Sopenharmony_ci } 27962306a36Sopenharmony_ci 28062306a36Sopenharmony_ci if (val & TICK_DEP_MASK_SCHED) { 28162306a36Sopenharmony_ci trace_tick_stop(0, TICK_DEP_MASK_SCHED); 28262306a36Sopenharmony_ci return true; 28362306a36Sopenharmony_ci } 28462306a36Sopenharmony_ci 28562306a36Sopenharmony_ci if (val & TICK_DEP_MASK_CLOCK_UNSTABLE) { 28662306a36Sopenharmony_ci trace_tick_stop(0, TICK_DEP_MASK_CLOCK_UNSTABLE); 28762306a36Sopenharmony_ci return true; 28862306a36Sopenharmony_ci } 28962306a36Sopenharmony_ci 29062306a36Sopenharmony_ci if (val & TICK_DEP_MASK_RCU) { 29162306a36Sopenharmony_ci trace_tick_stop(0, TICK_DEP_MASK_RCU); 29262306a36Sopenharmony_ci return true; 29362306a36Sopenharmony_ci } 29462306a36Sopenharmony_ci 29562306a36Sopenharmony_ci if (val & TICK_DEP_MASK_RCU_EXP) { 29662306a36Sopenharmony_ci trace_tick_stop(0, TICK_DEP_MASK_RCU_EXP); 29762306a36Sopenharmony_ci return true; 29862306a36Sopenharmony_ci } 29962306a36Sopenharmony_ci 30062306a36Sopenharmony_ci return false; 30162306a36Sopenharmony_ci} 30262306a36Sopenharmony_ci 30362306a36Sopenharmony_cistatic bool can_stop_full_tick(int cpu, struct tick_sched *ts) 30462306a36Sopenharmony_ci{ 30562306a36Sopenharmony_ci lockdep_assert_irqs_disabled(); 30662306a36Sopenharmony_ci 30762306a36Sopenharmony_ci if (unlikely(!cpu_online(cpu))) 30862306a36Sopenharmony_ci return false; 30962306a36Sopenharmony_ci 31062306a36Sopenharmony_ci if (check_tick_dependency(&tick_dep_mask)) 31162306a36Sopenharmony_ci return false; 31262306a36Sopenharmony_ci 31362306a36Sopenharmony_ci if (check_tick_dependency(&ts->tick_dep_mask)) 31462306a36Sopenharmony_ci return false; 31562306a36Sopenharmony_ci 31662306a36Sopenharmony_ci if (check_tick_dependency(¤t->tick_dep_mask)) 31762306a36Sopenharmony_ci return false; 31862306a36Sopenharmony_ci 31962306a36Sopenharmony_ci if (check_tick_dependency(¤t->signal->tick_dep_mask)) 32062306a36Sopenharmony_ci return false; 32162306a36Sopenharmony_ci 32262306a36Sopenharmony_ci return true; 32362306a36Sopenharmony_ci} 32462306a36Sopenharmony_ci 32562306a36Sopenharmony_cistatic void nohz_full_kick_func(struct irq_work *work) 32662306a36Sopenharmony_ci{ 32762306a36Sopenharmony_ci /* Empty, the tick restart happens on tick_nohz_irq_exit() */ 32862306a36Sopenharmony_ci} 32962306a36Sopenharmony_ci 33062306a36Sopenharmony_cistatic DEFINE_PER_CPU(struct irq_work, nohz_full_kick_work) = 33162306a36Sopenharmony_ci IRQ_WORK_INIT_HARD(nohz_full_kick_func); 33262306a36Sopenharmony_ci 33362306a36Sopenharmony_ci/* 33462306a36Sopenharmony_ci * Kick this CPU if it's full dynticks in order to force it to 33562306a36Sopenharmony_ci * re-evaluate its dependency on the tick and restart it if necessary. 33662306a36Sopenharmony_ci * This kick, unlike tick_nohz_full_kick_cpu() and tick_nohz_full_kick_all(), 33762306a36Sopenharmony_ci * is NMI safe. 33862306a36Sopenharmony_ci */ 33962306a36Sopenharmony_cistatic void tick_nohz_full_kick(void) 34062306a36Sopenharmony_ci{ 34162306a36Sopenharmony_ci if (!tick_nohz_full_cpu(smp_processor_id())) 34262306a36Sopenharmony_ci return; 34362306a36Sopenharmony_ci 34462306a36Sopenharmony_ci irq_work_queue(this_cpu_ptr(&nohz_full_kick_work)); 34562306a36Sopenharmony_ci} 34662306a36Sopenharmony_ci 34762306a36Sopenharmony_ci/* 34862306a36Sopenharmony_ci * Kick the CPU if it's full dynticks in order to force it to 34962306a36Sopenharmony_ci * re-evaluate its dependency on the tick and restart it if necessary. 35062306a36Sopenharmony_ci */ 35162306a36Sopenharmony_civoid tick_nohz_full_kick_cpu(int cpu) 35262306a36Sopenharmony_ci{ 35362306a36Sopenharmony_ci if (!tick_nohz_full_cpu(cpu)) 35462306a36Sopenharmony_ci return; 35562306a36Sopenharmony_ci 35662306a36Sopenharmony_ci irq_work_queue_on(&per_cpu(nohz_full_kick_work, cpu), cpu); 35762306a36Sopenharmony_ci} 35862306a36Sopenharmony_ci 35962306a36Sopenharmony_cistatic void tick_nohz_kick_task(struct task_struct *tsk) 36062306a36Sopenharmony_ci{ 36162306a36Sopenharmony_ci int cpu; 36262306a36Sopenharmony_ci 36362306a36Sopenharmony_ci /* 36462306a36Sopenharmony_ci * If the task is not running, run_posix_cpu_timers() 36562306a36Sopenharmony_ci * has nothing to elapse, IPI can then be spared. 36662306a36Sopenharmony_ci * 36762306a36Sopenharmony_ci * activate_task() STORE p->tick_dep_mask 36862306a36Sopenharmony_ci * STORE p->on_rq 36962306a36Sopenharmony_ci * __schedule() (switch to task 'p') smp_mb() (atomic_fetch_or()) 37062306a36Sopenharmony_ci * LOCK rq->lock LOAD p->on_rq 37162306a36Sopenharmony_ci * smp_mb__after_spin_lock() 37262306a36Sopenharmony_ci * tick_nohz_task_switch() 37362306a36Sopenharmony_ci * LOAD p->tick_dep_mask 37462306a36Sopenharmony_ci */ 37562306a36Sopenharmony_ci if (!sched_task_on_rq(tsk)) 37662306a36Sopenharmony_ci return; 37762306a36Sopenharmony_ci 37862306a36Sopenharmony_ci /* 37962306a36Sopenharmony_ci * If the task concurrently migrates to another CPU, 38062306a36Sopenharmony_ci * we guarantee it sees the new tick dependency upon 38162306a36Sopenharmony_ci * schedule. 38262306a36Sopenharmony_ci * 38362306a36Sopenharmony_ci * set_task_cpu(p, cpu); 38462306a36Sopenharmony_ci * STORE p->cpu = @cpu 38562306a36Sopenharmony_ci * __schedule() (switch to task 'p') 38662306a36Sopenharmony_ci * LOCK rq->lock 38762306a36Sopenharmony_ci * smp_mb__after_spin_lock() STORE p->tick_dep_mask 38862306a36Sopenharmony_ci * tick_nohz_task_switch() smp_mb() (atomic_fetch_or()) 38962306a36Sopenharmony_ci * LOAD p->tick_dep_mask LOAD p->cpu 39062306a36Sopenharmony_ci */ 39162306a36Sopenharmony_ci cpu = task_cpu(tsk); 39262306a36Sopenharmony_ci 39362306a36Sopenharmony_ci preempt_disable(); 39462306a36Sopenharmony_ci if (cpu_online(cpu)) 39562306a36Sopenharmony_ci tick_nohz_full_kick_cpu(cpu); 39662306a36Sopenharmony_ci preempt_enable(); 39762306a36Sopenharmony_ci} 39862306a36Sopenharmony_ci 39962306a36Sopenharmony_ci/* 40062306a36Sopenharmony_ci * Kick all full dynticks CPUs in order to force these to re-evaluate 40162306a36Sopenharmony_ci * their dependency on the tick and restart it if necessary. 40262306a36Sopenharmony_ci */ 40362306a36Sopenharmony_cistatic void tick_nohz_full_kick_all(void) 40462306a36Sopenharmony_ci{ 40562306a36Sopenharmony_ci int cpu; 40662306a36Sopenharmony_ci 40762306a36Sopenharmony_ci if (!tick_nohz_full_running) 40862306a36Sopenharmony_ci return; 40962306a36Sopenharmony_ci 41062306a36Sopenharmony_ci preempt_disable(); 41162306a36Sopenharmony_ci for_each_cpu_and(cpu, tick_nohz_full_mask, cpu_online_mask) 41262306a36Sopenharmony_ci tick_nohz_full_kick_cpu(cpu); 41362306a36Sopenharmony_ci preempt_enable(); 41462306a36Sopenharmony_ci} 41562306a36Sopenharmony_ci 41662306a36Sopenharmony_cistatic void tick_nohz_dep_set_all(atomic_t *dep, 41762306a36Sopenharmony_ci enum tick_dep_bits bit) 41862306a36Sopenharmony_ci{ 41962306a36Sopenharmony_ci int prev; 42062306a36Sopenharmony_ci 42162306a36Sopenharmony_ci prev = atomic_fetch_or(BIT(bit), dep); 42262306a36Sopenharmony_ci if (!prev) 42362306a36Sopenharmony_ci tick_nohz_full_kick_all(); 42462306a36Sopenharmony_ci} 42562306a36Sopenharmony_ci 42662306a36Sopenharmony_ci/* 42762306a36Sopenharmony_ci * Set a global tick dependency. Used by perf events that rely on freq and 42862306a36Sopenharmony_ci * by unstable clock. 42962306a36Sopenharmony_ci */ 43062306a36Sopenharmony_civoid tick_nohz_dep_set(enum tick_dep_bits bit) 43162306a36Sopenharmony_ci{ 43262306a36Sopenharmony_ci tick_nohz_dep_set_all(&tick_dep_mask, bit); 43362306a36Sopenharmony_ci} 43462306a36Sopenharmony_ci 43562306a36Sopenharmony_civoid tick_nohz_dep_clear(enum tick_dep_bits bit) 43662306a36Sopenharmony_ci{ 43762306a36Sopenharmony_ci atomic_andnot(BIT(bit), &tick_dep_mask); 43862306a36Sopenharmony_ci} 43962306a36Sopenharmony_ci 44062306a36Sopenharmony_ci/* 44162306a36Sopenharmony_ci * Set per-CPU tick dependency. Used by scheduler and perf events in order to 44262306a36Sopenharmony_ci * manage events throttling. 44362306a36Sopenharmony_ci */ 44462306a36Sopenharmony_civoid tick_nohz_dep_set_cpu(int cpu, enum tick_dep_bits bit) 44562306a36Sopenharmony_ci{ 44662306a36Sopenharmony_ci int prev; 44762306a36Sopenharmony_ci struct tick_sched *ts; 44862306a36Sopenharmony_ci 44962306a36Sopenharmony_ci ts = per_cpu_ptr(&tick_cpu_sched, cpu); 45062306a36Sopenharmony_ci 45162306a36Sopenharmony_ci prev = atomic_fetch_or(BIT(bit), &ts->tick_dep_mask); 45262306a36Sopenharmony_ci if (!prev) { 45362306a36Sopenharmony_ci preempt_disable(); 45462306a36Sopenharmony_ci /* Perf needs local kick that is NMI safe */ 45562306a36Sopenharmony_ci if (cpu == smp_processor_id()) { 45662306a36Sopenharmony_ci tick_nohz_full_kick(); 45762306a36Sopenharmony_ci } else { 45862306a36Sopenharmony_ci /* Remote irq work not NMI-safe */ 45962306a36Sopenharmony_ci if (!WARN_ON_ONCE(in_nmi())) 46062306a36Sopenharmony_ci tick_nohz_full_kick_cpu(cpu); 46162306a36Sopenharmony_ci } 46262306a36Sopenharmony_ci preempt_enable(); 46362306a36Sopenharmony_ci } 46462306a36Sopenharmony_ci} 46562306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(tick_nohz_dep_set_cpu); 46662306a36Sopenharmony_ci 46762306a36Sopenharmony_civoid tick_nohz_dep_clear_cpu(int cpu, enum tick_dep_bits bit) 46862306a36Sopenharmony_ci{ 46962306a36Sopenharmony_ci struct tick_sched *ts = per_cpu_ptr(&tick_cpu_sched, cpu); 47062306a36Sopenharmony_ci 47162306a36Sopenharmony_ci atomic_andnot(BIT(bit), &ts->tick_dep_mask); 47262306a36Sopenharmony_ci} 47362306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(tick_nohz_dep_clear_cpu); 47462306a36Sopenharmony_ci 47562306a36Sopenharmony_ci/* 47662306a36Sopenharmony_ci * Set a per-task tick dependency. RCU need this. Also posix CPU timers 47762306a36Sopenharmony_ci * in order to elapse per task timers. 47862306a36Sopenharmony_ci */ 47962306a36Sopenharmony_civoid tick_nohz_dep_set_task(struct task_struct *tsk, enum tick_dep_bits bit) 48062306a36Sopenharmony_ci{ 48162306a36Sopenharmony_ci if (!atomic_fetch_or(BIT(bit), &tsk->tick_dep_mask)) 48262306a36Sopenharmony_ci tick_nohz_kick_task(tsk); 48362306a36Sopenharmony_ci} 48462306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(tick_nohz_dep_set_task); 48562306a36Sopenharmony_ci 48662306a36Sopenharmony_civoid tick_nohz_dep_clear_task(struct task_struct *tsk, enum tick_dep_bits bit) 48762306a36Sopenharmony_ci{ 48862306a36Sopenharmony_ci atomic_andnot(BIT(bit), &tsk->tick_dep_mask); 48962306a36Sopenharmony_ci} 49062306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(tick_nohz_dep_clear_task); 49162306a36Sopenharmony_ci 49262306a36Sopenharmony_ci/* 49362306a36Sopenharmony_ci * Set a per-taskgroup tick dependency. Posix CPU timers need this in order to elapse 49462306a36Sopenharmony_ci * per process timers. 49562306a36Sopenharmony_ci */ 49662306a36Sopenharmony_civoid tick_nohz_dep_set_signal(struct task_struct *tsk, 49762306a36Sopenharmony_ci enum tick_dep_bits bit) 49862306a36Sopenharmony_ci{ 49962306a36Sopenharmony_ci int prev; 50062306a36Sopenharmony_ci struct signal_struct *sig = tsk->signal; 50162306a36Sopenharmony_ci 50262306a36Sopenharmony_ci prev = atomic_fetch_or(BIT(bit), &sig->tick_dep_mask); 50362306a36Sopenharmony_ci if (!prev) { 50462306a36Sopenharmony_ci struct task_struct *t; 50562306a36Sopenharmony_ci 50662306a36Sopenharmony_ci lockdep_assert_held(&tsk->sighand->siglock); 50762306a36Sopenharmony_ci __for_each_thread(sig, t) 50862306a36Sopenharmony_ci tick_nohz_kick_task(t); 50962306a36Sopenharmony_ci } 51062306a36Sopenharmony_ci} 51162306a36Sopenharmony_ci 51262306a36Sopenharmony_civoid tick_nohz_dep_clear_signal(struct signal_struct *sig, enum tick_dep_bits bit) 51362306a36Sopenharmony_ci{ 51462306a36Sopenharmony_ci atomic_andnot(BIT(bit), &sig->tick_dep_mask); 51562306a36Sopenharmony_ci} 51662306a36Sopenharmony_ci 51762306a36Sopenharmony_ci/* 51862306a36Sopenharmony_ci * Re-evaluate the need for the tick as we switch the current task. 51962306a36Sopenharmony_ci * It might need the tick due to per task/process properties: 52062306a36Sopenharmony_ci * perf events, posix CPU timers, ... 52162306a36Sopenharmony_ci */ 52262306a36Sopenharmony_civoid __tick_nohz_task_switch(void) 52362306a36Sopenharmony_ci{ 52462306a36Sopenharmony_ci struct tick_sched *ts; 52562306a36Sopenharmony_ci 52662306a36Sopenharmony_ci if (!tick_nohz_full_cpu(smp_processor_id())) 52762306a36Sopenharmony_ci return; 52862306a36Sopenharmony_ci 52962306a36Sopenharmony_ci ts = this_cpu_ptr(&tick_cpu_sched); 53062306a36Sopenharmony_ci 53162306a36Sopenharmony_ci if (ts->tick_stopped) { 53262306a36Sopenharmony_ci if (atomic_read(¤t->tick_dep_mask) || 53362306a36Sopenharmony_ci atomic_read(¤t->signal->tick_dep_mask)) 53462306a36Sopenharmony_ci tick_nohz_full_kick(); 53562306a36Sopenharmony_ci } 53662306a36Sopenharmony_ci} 53762306a36Sopenharmony_ci 53862306a36Sopenharmony_ci/* Get the boot-time nohz CPU list from the kernel parameters. */ 53962306a36Sopenharmony_civoid __init tick_nohz_full_setup(cpumask_var_t cpumask) 54062306a36Sopenharmony_ci{ 54162306a36Sopenharmony_ci alloc_bootmem_cpumask_var(&tick_nohz_full_mask); 54262306a36Sopenharmony_ci cpumask_copy(tick_nohz_full_mask, cpumask); 54362306a36Sopenharmony_ci tick_nohz_full_running = true; 54462306a36Sopenharmony_ci} 54562306a36Sopenharmony_ci 54662306a36Sopenharmony_cibool tick_nohz_cpu_hotpluggable(unsigned int cpu) 54762306a36Sopenharmony_ci{ 54862306a36Sopenharmony_ci /* 54962306a36Sopenharmony_ci * The tick_do_timer_cpu CPU handles housekeeping duty (unbound 55062306a36Sopenharmony_ci * timers, workqueues, timekeeping, ...) on behalf of full dynticks 55162306a36Sopenharmony_ci * CPUs. It must remain online when nohz full is enabled. 55262306a36Sopenharmony_ci */ 55362306a36Sopenharmony_ci if (tick_nohz_full_running && tick_do_timer_cpu == cpu) 55462306a36Sopenharmony_ci return false; 55562306a36Sopenharmony_ci return true; 55662306a36Sopenharmony_ci} 55762306a36Sopenharmony_ci 55862306a36Sopenharmony_cistatic int tick_nohz_cpu_down(unsigned int cpu) 55962306a36Sopenharmony_ci{ 56062306a36Sopenharmony_ci return tick_nohz_cpu_hotpluggable(cpu) ? 0 : -EBUSY; 56162306a36Sopenharmony_ci} 56262306a36Sopenharmony_ci 56362306a36Sopenharmony_civoid __init tick_nohz_init(void) 56462306a36Sopenharmony_ci{ 56562306a36Sopenharmony_ci int cpu, ret; 56662306a36Sopenharmony_ci 56762306a36Sopenharmony_ci if (!tick_nohz_full_running) 56862306a36Sopenharmony_ci return; 56962306a36Sopenharmony_ci 57062306a36Sopenharmony_ci /* 57162306a36Sopenharmony_ci * Full dynticks uses irq work to drive the tick rescheduling on safe 57262306a36Sopenharmony_ci * locking contexts. But then we need irq work to raise its own 57362306a36Sopenharmony_ci * interrupts to avoid circular dependency on the tick 57462306a36Sopenharmony_ci */ 57562306a36Sopenharmony_ci if (!arch_irq_work_has_interrupt()) { 57662306a36Sopenharmony_ci pr_warn("NO_HZ: Can't run full dynticks because arch doesn't support irq work self-IPIs\n"); 57762306a36Sopenharmony_ci cpumask_clear(tick_nohz_full_mask); 57862306a36Sopenharmony_ci tick_nohz_full_running = false; 57962306a36Sopenharmony_ci return; 58062306a36Sopenharmony_ci } 58162306a36Sopenharmony_ci 58262306a36Sopenharmony_ci if (IS_ENABLED(CONFIG_PM_SLEEP_SMP) && 58362306a36Sopenharmony_ci !IS_ENABLED(CONFIG_PM_SLEEP_SMP_NONZERO_CPU)) { 58462306a36Sopenharmony_ci cpu = smp_processor_id(); 58562306a36Sopenharmony_ci 58662306a36Sopenharmony_ci if (cpumask_test_cpu(cpu, tick_nohz_full_mask)) { 58762306a36Sopenharmony_ci pr_warn("NO_HZ: Clearing %d from nohz_full range " 58862306a36Sopenharmony_ci "for timekeeping\n", cpu); 58962306a36Sopenharmony_ci cpumask_clear_cpu(cpu, tick_nohz_full_mask); 59062306a36Sopenharmony_ci } 59162306a36Sopenharmony_ci } 59262306a36Sopenharmony_ci 59362306a36Sopenharmony_ci for_each_cpu(cpu, tick_nohz_full_mask) 59462306a36Sopenharmony_ci ct_cpu_track_user(cpu); 59562306a36Sopenharmony_ci 59662306a36Sopenharmony_ci ret = cpuhp_setup_state_nocalls(CPUHP_AP_ONLINE_DYN, 59762306a36Sopenharmony_ci "kernel/nohz:predown", NULL, 59862306a36Sopenharmony_ci tick_nohz_cpu_down); 59962306a36Sopenharmony_ci WARN_ON(ret < 0); 60062306a36Sopenharmony_ci pr_info("NO_HZ: Full dynticks CPUs: %*pbl.\n", 60162306a36Sopenharmony_ci cpumask_pr_args(tick_nohz_full_mask)); 60262306a36Sopenharmony_ci} 60362306a36Sopenharmony_ci#endif 60462306a36Sopenharmony_ci 60562306a36Sopenharmony_ci/* 60662306a36Sopenharmony_ci * NOHZ - aka dynamic tick functionality 60762306a36Sopenharmony_ci */ 60862306a36Sopenharmony_ci#ifdef CONFIG_NO_HZ_COMMON 60962306a36Sopenharmony_ci/* 61062306a36Sopenharmony_ci * NO HZ enabled ? 61162306a36Sopenharmony_ci */ 61262306a36Sopenharmony_cibool tick_nohz_enabled __read_mostly = true; 61362306a36Sopenharmony_ciunsigned long tick_nohz_active __read_mostly; 61462306a36Sopenharmony_ci/* 61562306a36Sopenharmony_ci * Enable / Disable tickless mode 61662306a36Sopenharmony_ci */ 61762306a36Sopenharmony_cistatic int __init setup_tick_nohz(char *str) 61862306a36Sopenharmony_ci{ 61962306a36Sopenharmony_ci return (kstrtobool(str, &tick_nohz_enabled) == 0); 62062306a36Sopenharmony_ci} 62162306a36Sopenharmony_ci 62262306a36Sopenharmony_ci__setup("nohz=", setup_tick_nohz); 62362306a36Sopenharmony_ci 62462306a36Sopenharmony_cibool tick_nohz_tick_stopped(void) 62562306a36Sopenharmony_ci{ 62662306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 62762306a36Sopenharmony_ci 62862306a36Sopenharmony_ci return ts->tick_stopped; 62962306a36Sopenharmony_ci} 63062306a36Sopenharmony_ci 63162306a36Sopenharmony_cibool tick_nohz_tick_stopped_cpu(int cpu) 63262306a36Sopenharmony_ci{ 63362306a36Sopenharmony_ci struct tick_sched *ts = per_cpu_ptr(&tick_cpu_sched, cpu); 63462306a36Sopenharmony_ci 63562306a36Sopenharmony_ci return ts->tick_stopped; 63662306a36Sopenharmony_ci} 63762306a36Sopenharmony_ci 63862306a36Sopenharmony_ci/** 63962306a36Sopenharmony_ci * tick_nohz_update_jiffies - update jiffies when idle was interrupted 64062306a36Sopenharmony_ci * 64162306a36Sopenharmony_ci * Called from interrupt entry when the CPU was idle 64262306a36Sopenharmony_ci * 64362306a36Sopenharmony_ci * In case the sched_tick was stopped on this CPU, we have to check if jiffies 64462306a36Sopenharmony_ci * must be updated. Otherwise an interrupt handler could use a stale jiffy 64562306a36Sopenharmony_ci * value. We do this unconditionally on any CPU, as we don't know whether the 64662306a36Sopenharmony_ci * CPU, which has the update task assigned is in a long sleep. 64762306a36Sopenharmony_ci */ 64862306a36Sopenharmony_cistatic void tick_nohz_update_jiffies(ktime_t now) 64962306a36Sopenharmony_ci{ 65062306a36Sopenharmony_ci unsigned long flags; 65162306a36Sopenharmony_ci 65262306a36Sopenharmony_ci __this_cpu_write(tick_cpu_sched.idle_waketime, now); 65362306a36Sopenharmony_ci 65462306a36Sopenharmony_ci local_irq_save(flags); 65562306a36Sopenharmony_ci tick_do_update_jiffies64(now); 65662306a36Sopenharmony_ci local_irq_restore(flags); 65762306a36Sopenharmony_ci 65862306a36Sopenharmony_ci touch_softlockup_watchdog_sched(); 65962306a36Sopenharmony_ci} 66062306a36Sopenharmony_ci 66162306a36Sopenharmony_cistatic void tick_nohz_stop_idle(struct tick_sched *ts, ktime_t now) 66262306a36Sopenharmony_ci{ 66362306a36Sopenharmony_ci ktime_t delta; 66462306a36Sopenharmony_ci 66562306a36Sopenharmony_ci if (WARN_ON_ONCE(!ts->idle_active)) 66662306a36Sopenharmony_ci return; 66762306a36Sopenharmony_ci 66862306a36Sopenharmony_ci delta = ktime_sub(now, ts->idle_entrytime); 66962306a36Sopenharmony_ci 67062306a36Sopenharmony_ci write_seqcount_begin(&ts->idle_sleeptime_seq); 67162306a36Sopenharmony_ci if (nr_iowait_cpu(smp_processor_id()) > 0) 67262306a36Sopenharmony_ci ts->iowait_sleeptime = ktime_add(ts->iowait_sleeptime, delta); 67362306a36Sopenharmony_ci else 67462306a36Sopenharmony_ci ts->idle_sleeptime = ktime_add(ts->idle_sleeptime, delta); 67562306a36Sopenharmony_ci 67662306a36Sopenharmony_ci ts->idle_entrytime = now; 67762306a36Sopenharmony_ci ts->idle_active = 0; 67862306a36Sopenharmony_ci write_seqcount_end(&ts->idle_sleeptime_seq); 67962306a36Sopenharmony_ci 68062306a36Sopenharmony_ci sched_clock_idle_wakeup_event(); 68162306a36Sopenharmony_ci} 68262306a36Sopenharmony_ci 68362306a36Sopenharmony_cistatic void tick_nohz_start_idle(struct tick_sched *ts) 68462306a36Sopenharmony_ci{ 68562306a36Sopenharmony_ci write_seqcount_begin(&ts->idle_sleeptime_seq); 68662306a36Sopenharmony_ci ts->idle_entrytime = ktime_get(); 68762306a36Sopenharmony_ci ts->idle_active = 1; 68862306a36Sopenharmony_ci write_seqcount_end(&ts->idle_sleeptime_seq); 68962306a36Sopenharmony_ci 69062306a36Sopenharmony_ci sched_clock_idle_sleep_event(); 69162306a36Sopenharmony_ci} 69262306a36Sopenharmony_ci 69362306a36Sopenharmony_cistatic u64 get_cpu_sleep_time_us(struct tick_sched *ts, ktime_t *sleeptime, 69462306a36Sopenharmony_ci bool compute_delta, u64 *last_update_time) 69562306a36Sopenharmony_ci{ 69662306a36Sopenharmony_ci ktime_t now, idle; 69762306a36Sopenharmony_ci unsigned int seq; 69862306a36Sopenharmony_ci 69962306a36Sopenharmony_ci if (!tick_nohz_active) 70062306a36Sopenharmony_ci return -1; 70162306a36Sopenharmony_ci 70262306a36Sopenharmony_ci now = ktime_get(); 70362306a36Sopenharmony_ci if (last_update_time) 70462306a36Sopenharmony_ci *last_update_time = ktime_to_us(now); 70562306a36Sopenharmony_ci 70662306a36Sopenharmony_ci do { 70762306a36Sopenharmony_ci seq = read_seqcount_begin(&ts->idle_sleeptime_seq); 70862306a36Sopenharmony_ci 70962306a36Sopenharmony_ci if (ts->idle_active && compute_delta) { 71062306a36Sopenharmony_ci ktime_t delta = ktime_sub(now, ts->idle_entrytime); 71162306a36Sopenharmony_ci 71262306a36Sopenharmony_ci idle = ktime_add(*sleeptime, delta); 71362306a36Sopenharmony_ci } else { 71462306a36Sopenharmony_ci idle = *sleeptime; 71562306a36Sopenharmony_ci } 71662306a36Sopenharmony_ci } while (read_seqcount_retry(&ts->idle_sleeptime_seq, seq)); 71762306a36Sopenharmony_ci 71862306a36Sopenharmony_ci return ktime_to_us(idle); 71962306a36Sopenharmony_ci 72062306a36Sopenharmony_ci} 72162306a36Sopenharmony_ci 72262306a36Sopenharmony_ci/** 72362306a36Sopenharmony_ci * get_cpu_idle_time_us - get the total idle time of a CPU 72462306a36Sopenharmony_ci * @cpu: CPU number to query 72562306a36Sopenharmony_ci * @last_update_time: variable to store update time in. Do not update 72662306a36Sopenharmony_ci * counters if NULL. 72762306a36Sopenharmony_ci * 72862306a36Sopenharmony_ci * Return the cumulative idle time (since boot) for a given 72962306a36Sopenharmony_ci * CPU, in microseconds. Note this is partially broken due to 73062306a36Sopenharmony_ci * the counter of iowait tasks that can be remotely updated without 73162306a36Sopenharmony_ci * any synchronization. Therefore it is possible to observe backward 73262306a36Sopenharmony_ci * values within two consecutive reads. 73362306a36Sopenharmony_ci * 73462306a36Sopenharmony_ci * This time is measured via accounting rather than sampling, 73562306a36Sopenharmony_ci * and is as accurate as ktime_get() is. 73662306a36Sopenharmony_ci * 73762306a36Sopenharmony_ci * This function returns -1 if NOHZ is not enabled. 73862306a36Sopenharmony_ci */ 73962306a36Sopenharmony_ciu64 get_cpu_idle_time_us(int cpu, u64 *last_update_time) 74062306a36Sopenharmony_ci{ 74162306a36Sopenharmony_ci struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); 74262306a36Sopenharmony_ci 74362306a36Sopenharmony_ci return get_cpu_sleep_time_us(ts, &ts->idle_sleeptime, 74462306a36Sopenharmony_ci !nr_iowait_cpu(cpu), last_update_time); 74562306a36Sopenharmony_ci} 74662306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(get_cpu_idle_time_us); 74762306a36Sopenharmony_ci 74862306a36Sopenharmony_ci/** 74962306a36Sopenharmony_ci * get_cpu_iowait_time_us - get the total iowait time of a CPU 75062306a36Sopenharmony_ci * @cpu: CPU number to query 75162306a36Sopenharmony_ci * @last_update_time: variable to store update time in. Do not update 75262306a36Sopenharmony_ci * counters if NULL. 75362306a36Sopenharmony_ci * 75462306a36Sopenharmony_ci * Return the cumulative iowait time (since boot) for a given 75562306a36Sopenharmony_ci * CPU, in microseconds. Note this is partially broken due to 75662306a36Sopenharmony_ci * the counter of iowait tasks that can be remotely updated without 75762306a36Sopenharmony_ci * any synchronization. Therefore it is possible to observe backward 75862306a36Sopenharmony_ci * values within two consecutive reads. 75962306a36Sopenharmony_ci * 76062306a36Sopenharmony_ci * This time is measured via accounting rather than sampling, 76162306a36Sopenharmony_ci * and is as accurate as ktime_get() is. 76262306a36Sopenharmony_ci * 76362306a36Sopenharmony_ci * This function returns -1 if NOHZ is not enabled. 76462306a36Sopenharmony_ci */ 76562306a36Sopenharmony_ciu64 get_cpu_iowait_time_us(int cpu, u64 *last_update_time) 76662306a36Sopenharmony_ci{ 76762306a36Sopenharmony_ci struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); 76862306a36Sopenharmony_ci 76962306a36Sopenharmony_ci return get_cpu_sleep_time_us(ts, &ts->iowait_sleeptime, 77062306a36Sopenharmony_ci nr_iowait_cpu(cpu), last_update_time); 77162306a36Sopenharmony_ci} 77262306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(get_cpu_iowait_time_us); 77362306a36Sopenharmony_ci 77462306a36Sopenharmony_cistatic void tick_nohz_restart(struct tick_sched *ts, ktime_t now) 77562306a36Sopenharmony_ci{ 77662306a36Sopenharmony_ci hrtimer_cancel(&ts->sched_timer); 77762306a36Sopenharmony_ci hrtimer_set_expires(&ts->sched_timer, ts->last_tick); 77862306a36Sopenharmony_ci 77962306a36Sopenharmony_ci /* Forward the time to expire in the future */ 78062306a36Sopenharmony_ci hrtimer_forward(&ts->sched_timer, now, TICK_NSEC); 78162306a36Sopenharmony_ci 78262306a36Sopenharmony_ci if (ts->nohz_mode == NOHZ_MODE_HIGHRES) { 78362306a36Sopenharmony_ci hrtimer_start_expires(&ts->sched_timer, 78462306a36Sopenharmony_ci HRTIMER_MODE_ABS_PINNED_HARD); 78562306a36Sopenharmony_ci } else { 78662306a36Sopenharmony_ci tick_program_event(hrtimer_get_expires(&ts->sched_timer), 1); 78762306a36Sopenharmony_ci } 78862306a36Sopenharmony_ci 78962306a36Sopenharmony_ci /* 79062306a36Sopenharmony_ci * Reset to make sure next tick stop doesn't get fooled by past 79162306a36Sopenharmony_ci * cached clock deadline. 79262306a36Sopenharmony_ci */ 79362306a36Sopenharmony_ci ts->next_tick = 0; 79462306a36Sopenharmony_ci} 79562306a36Sopenharmony_ci 79662306a36Sopenharmony_cistatic inline bool local_timer_softirq_pending(void) 79762306a36Sopenharmony_ci{ 79862306a36Sopenharmony_ci return local_softirq_pending() & BIT(TIMER_SOFTIRQ); 79962306a36Sopenharmony_ci} 80062306a36Sopenharmony_ci 80162306a36Sopenharmony_cistatic ktime_t tick_nohz_next_event(struct tick_sched *ts, int cpu) 80262306a36Sopenharmony_ci{ 80362306a36Sopenharmony_ci u64 basemono, next_tick, delta, expires; 80462306a36Sopenharmony_ci unsigned long basejiff; 80562306a36Sopenharmony_ci unsigned int seq; 80662306a36Sopenharmony_ci 80762306a36Sopenharmony_ci /* Read jiffies and the time when jiffies were updated last */ 80862306a36Sopenharmony_ci do { 80962306a36Sopenharmony_ci seq = read_seqcount_begin(&jiffies_seq); 81062306a36Sopenharmony_ci basemono = last_jiffies_update; 81162306a36Sopenharmony_ci basejiff = jiffies; 81262306a36Sopenharmony_ci } while (read_seqcount_retry(&jiffies_seq, seq)); 81362306a36Sopenharmony_ci ts->last_jiffies = basejiff; 81462306a36Sopenharmony_ci ts->timer_expires_base = basemono; 81562306a36Sopenharmony_ci 81662306a36Sopenharmony_ci /* 81762306a36Sopenharmony_ci * Keep the periodic tick, when RCU, architecture or irq_work 81862306a36Sopenharmony_ci * requests it. 81962306a36Sopenharmony_ci * Aside of that check whether the local timer softirq is 82062306a36Sopenharmony_ci * pending. If so its a bad idea to call get_next_timer_interrupt() 82162306a36Sopenharmony_ci * because there is an already expired timer, so it will request 82262306a36Sopenharmony_ci * immediate expiry, which rearms the hardware timer with a 82362306a36Sopenharmony_ci * minimal delta which brings us back to this place 82462306a36Sopenharmony_ci * immediately. Lather, rinse and repeat... 82562306a36Sopenharmony_ci */ 82662306a36Sopenharmony_ci if (rcu_needs_cpu() || arch_needs_cpu() || 82762306a36Sopenharmony_ci irq_work_needs_cpu() || local_timer_softirq_pending()) { 82862306a36Sopenharmony_ci next_tick = basemono + TICK_NSEC; 82962306a36Sopenharmony_ci } else { 83062306a36Sopenharmony_ci /* 83162306a36Sopenharmony_ci * Get the next pending timer. If high resolution 83262306a36Sopenharmony_ci * timers are enabled this only takes the timer wheel 83362306a36Sopenharmony_ci * timers into account. If high resolution timers are 83462306a36Sopenharmony_ci * disabled this also looks at the next expiring 83562306a36Sopenharmony_ci * hrtimer. 83662306a36Sopenharmony_ci */ 83762306a36Sopenharmony_ci next_tick = get_next_timer_interrupt(basejiff, basemono); 83862306a36Sopenharmony_ci ts->next_timer = next_tick; 83962306a36Sopenharmony_ci } 84062306a36Sopenharmony_ci 84162306a36Sopenharmony_ci /* 84262306a36Sopenharmony_ci * If the tick is due in the next period, keep it ticking or 84362306a36Sopenharmony_ci * force prod the timer. 84462306a36Sopenharmony_ci */ 84562306a36Sopenharmony_ci delta = next_tick - basemono; 84662306a36Sopenharmony_ci if (delta <= (u64)TICK_NSEC) { 84762306a36Sopenharmony_ci /* 84862306a36Sopenharmony_ci * Tell the timer code that the base is not idle, i.e. undo 84962306a36Sopenharmony_ci * the effect of get_next_timer_interrupt(): 85062306a36Sopenharmony_ci */ 85162306a36Sopenharmony_ci timer_clear_idle(); 85262306a36Sopenharmony_ci /* 85362306a36Sopenharmony_ci * We've not stopped the tick yet, and there's a timer in the 85462306a36Sopenharmony_ci * next period, so no point in stopping it either, bail. 85562306a36Sopenharmony_ci */ 85662306a36Sopenharmony_ci if (!ts->tick_stopped) { 85762306a36Sopenharmony_ci ts->timer_expires = 0; 85862306a36Sopenharmony_ci goto out; 85962306a36Sopenharmony_ci } 86062306a36Sopenharmony_ci } 86162306a36Sopenharmony_ci 86262306a36Sopenharmony_ci /* 86362306a36Sopenharmony_ci * If this CPU is the one which had the do_timer() duty last, we limit 86462306a36Sopenharmony_ci * the sleep time to the timekeeping max_deferment value. 86562306a36Sopenharmony_ci * Otherwise we can sleep as long as we want. 86662306a36Sopenharmony_ci */ 86762306a36Sopenharmony_ci delta = timekeeping_max_deferment(); 86862306a36Sopenharmony_ci if (cpu != tick_do_timer_cpu && 86962306a36Sopenharmony_ci (tick_do_timer_cpu != TICK_DO_TIMER_NONE || !ts->do_timer_last)) 87062306a36Sopenharmony_ci delta = KTIME_MAX; 87162306a36Sopenharmony_ci 87262306a36Sopenharmony_ci /* Calculate the next expiry time */ 87362306a36Sopenharmony_ci if (delta < (KTIME_MAX - basemono)) 87462306a36Sopenharmony_ci expires = basemono + delta; 87562306a36Sopenharmony_ci else 87662306a36Sopenharmony_ci expires = KTIME_MAX; 87762306a36Sopenharmony_ci 87862306a36Sopenharmony_ci ts->timer_expires = min_t(u64, expires, next_tick); 87962306a36Sopenharmony_ci 88062306a36Sopenharmony_ciout: 88162306a36Sopenharmony_ci return ts->timer_expires; 88262306a36Sopenharmony_ci} 88362306a36Sopenharmony_ci 88462306a36Sopenharmony_cistatic void tick_nohz_stop_tick(struct tick_sched *ts, int cpu) 88562306a36Sopenharmony_ci{ 88662306a36Sopenharmony_ci struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); 88762306a36Sopenharmony_ci u64 basemono = ts->timer_expires_base; 88862306a36Sopenharmony_ci u64 expires = ts->timer_expires; 88962306a36Sopenharmony_ci ktime_t tick = expires; 89062306a36Sopenharmony_ci 89162306a36Sopenharmony_ci /* Make sure we won't be trying to stop it twice in a row. */ 89262306a36Sopenharmony_ci ts->timer_expires_base = 0; 89362306a36Sopenharmony_ci 89462306a36Sopenharmony_ci /* 89562306a36Sopenharmony_ci * If this CPU is the one which updates jiffies, then give up 89662306a36Sopenharmony_ci * the assignment and let it be taken by the CPU which runs 89762306a36Sopenharmony_ci * the tick timer next, which might be this CPU as well. If we 89862306a36Sopenharmony_ci * don't drop this here the jiffies might be stale and 89962306a36Sopenharmony_ci * do_timer() never invoked. Keep track of the fact that it 90062306a36Sopenharmony_ci * was the one which had the do_timer() duty last. 90162306a36Sopenharmony_ci */ 90262306a36Sopenharmony_ci if (cpu == tick_do_timer_cpu) { 90362306a36Sopenharmony_ci tick_do_timer_cpu = TICK_DO_TIMER_NONE; 90462306a36Sopenharmony_ci ts->do_timer_last = 1; 90562306a36Sopenharmony_ci } else if (tick_do_timer_cpu != TICK_DO_TIMER_NONE) { 90662306a36Sopenharmony_ci ts->do_timer_last = 0; 90762306a36Sopenharmony_ci } 90862306a36Sopenharmony_ci 90962306a36Sopenharmony_ci /* Skip reprogram of event if its not changed */ 91062306a36Sopenharmony_ci if (ts->tick_stopped && (expires == ts->next_tick)) { 91162306a36Sopenharmony_ci /* Sanity check: make sure clockevent is actually programmed */ 91262306a36Sopenharmony_ci if (tick == KTIME_MAX || ts->next_tick == hrtimer_get_expires(&ts->sched_timer)) 91362306a36Sopenharmony_ci return; 91462306a36Sopenharmony_ci 91562306a36Sopenharmony_ci WARN_ON_ONCE(1); 91662306a36Sopenharmony_ci printk_once("basemono: %llu ts->next_tick: %llu dev->next_event: %llu timer->active: %d timer->expires: %llu\n", 91762306a36Sopenharmony_ci basemono, ts->next_tick, dev->next_event, 91862306a36Sopenharmony_ci hrtimer_active(&ts->sched_timer), hrtimer_get_expires(&ts->sched_timer)); 91962306a36Sopenharmony_ci } 92062306a36Sopenharmony_ci 92162306a36Sopenharmony_ci /* 92262306a36Sopenharmony_ci * nohz_stop_sched_tick can be called several times before 92362306a36Sopenharmony_ci * the nohz_restart_sched_tick is called. This happens when 92462306a36Sopenharmony_ci * interrupts arrive which do not cause a reschedule. In the 92562306a36Sopenharmony_ci * first call we save the current tick time, so we can restart 92662306a36Sopenharmony_ci * the scheduler tick in nohz_restart_sched_tick. 92762306a36Sopenharmony_ci */ 92862306a36Sopenharmony_ci if (!ts->tick_stopped) { 92962306a36Sopenharmony_ci calc_load_nohz_start(); 93062306a36Sopenharmony_ci quiet_vmstat(); 93162306a36Sopenharmony_ci 93262306a36Sopenharmony_ci ts->last_tick = hrtimer_get_expires(&ts->sched_timer); 93362306a36Sopenharmony_ci ts->tick_stopped = 1; 93462306a36Sopenharmony_ci trace_tick_stop(1, TICK_DEP_MASK_NONE); 93562306a36Sopenharmony_ci } 93662306a36Sopenharmony_ci 93762306a36Sopenharmony_ci ts->next_tick = tick; 93862306a36Sopenharmony_ci 93962306a36Sopenharmony_ci /* 94062306a36Sopenharmony_ci * If the expiration time == KTIME_MAX, then we simply stop 94162306a36Sopenharmony_ci * the tick timer. 94262306a36Sopenharmony_ci */ 94362306a36Sopenharmony_ci if (unlikely(expires == KTIME_MAX)) { 94462306a36Sopenharmony_ci if (ts->nohz_mode == NOHZ_MODE_HIGHRES) 94562306a36Sopenharmony_ci hrtimer_cancel(&ts->sched_timer); 94662306a36Sopenharmony_ci else 94762306a36Sopenharmony_ci tick_program_event(KTIME_MAX, 1); 94862306a36Sopenharmony_ci return; 94962306a36Sopenharmony_ci } 95062306a36Sopenharmony_ci 95162306a36Sopenharmony_ci if (ts->nohz_mode == NOHZ_MODE_HIGHRES) { 95262306a36Sopenharmony_ci hrtimer_start(&ts->sched_timer, tick, 95362306a36Sopenharmony_ci HRTIMER_MODE_ABS_PINNED_HARD); 95462306a36Sopenharmony_ci } else { 95562306a36Sopenharmony_ci hrtimer_set_expires(&ts->sched_timer, tick); 95662306a36Sopenharmony_ci tick_program_event(tick, 1); 95762306a36Sopenharmony_ci } 95862306a36Sopenharmony_ci} 95962306a36Sopenharmony_ci 96062306a36Sopenharmony_cistatic void tick_nohz_retain_tick(struct tick_sched *ts) 96162306a36Sopenharmony_ci{ 96262306a36Sopenharmony_ci ts->timer_expires_base = 0; 96362306a36Sopenharmony_ci} 96462306a36Sopenharmony_ci 96562306a36Sopenharmony_ci#ifdef CONFIG_NO_HZ_FULL 96662306a36Sopenharmony_cistatic void tick_nohz_stop_sched_tick(struct tick_sched *ts, int cpu) 96762306a36Sopenharmony_ci{ 96862306a36Sopenharmony_ci if (tick_nohz_next_event(ts, cpu)) 96962306a36Sopenharmony_ci tick_nohz_stop_tick(ts, cpu); 97062306a36Sopenharmony_ci else 97162306a36Sopenharmony_ci tick_nohz_retain_tick(ts); 97262306a36Sopenharmony_ci} 97362306a36Sopenharmony_ci#endif /* CONFIG_NO_HZ_FULL */ 97462306a36Sopenharmony_ci 97562306a36Sopenharmony_cistatic void tick_nohz_restart_sched_tick(struct tick_sched *ts, ktime_t now) 97662306a36Sopenharmony_ci{ 97762306a36Sopenharmony_ci /* Update jiffies first */ 97862306a36Sopenharmony_ci tick_do_update_jiffies64(now); 97962306a36Sopenharmony_ci 98062306a36Sopenharmony_ci /* 98162306a36Sopenharmony_ci * Clear the timer idle flag, so we avoid IPIs on remote queueing and 98262306a36Sopenharmony_ci * the clock forward checks in the enqueue path: 98362306a36Sopenharmony_ci */ 98462306a36Sopenharmony_ci timer_clear_idle(); 98562306a36Sopenharmony_ci 98662306a36Sopenharmony_ci calc_load_nohz_stop(); 98762306a36Sopenharmony_ci touch_softlockup_watchdog_sched(); 98862306a36Sopenharmony_ci /* 98962306a36Sopenharmony_ci * Cancel the scheduled timer and restore the tick 99062306a36Sopenharmony_ci */ 99162306a36Sopenharmony_ci ts->tick_stopped = 0; 99262306a36Sopenharmony_ci tick_nohz_restart(ts, now); 99362306a36Sopenharmony_ci} 99462306a36Sopenharmony_ci 99562306a36Sopenharmony_cistatic void __tick_nohz_full_update_tick(struct tick_sched *ts, 99662306a36Sopenharmony_ci ktime_t now) 99762306a36Sopenharmony_ci{ 99862306a36Sopenharmony_ci#ifdef CONFIG_NO_HZ_FULL 99962306a36Sopenharmony_ci int cpu = smp_processor_id(); 100062306a36Sopenharmony_ci 100162306a36Sopenharmony_ci if (can_stop_full_tick(cpu, ts)) 100262306a36Sopenharmony_ci tick_nohz_stop_sched_tick(ts, cpu); 100362306a36Sopenharmony_ci else if (ts->tick_stopped) 100462306a36Sopenharmony_ci tick_nohz_restart_sched_tick(ts, now); 100562306a36Sopenharmony_ci#endif 100662306a36Sopenharmony_ci} 100762306a36Sopenharmony_ci 100862306a36Sopenharmony_cistatic void tick_nohz_full_update_tick(struct tick_sched *ts) 100962306a36Sopenharmony_ci{ 101062306a36Sopenharmony_ci if (!tick_nohz_full_cpu(smp_processor_id())) 101162306a36Sopenharmony_ci return; 101262306a36Sopenharmony_ci 101362306a36Sopenharmony_ci if (!ts->tick_stopped && ts->nohz_mode == NOHZ_MODE_INACTIVE) 101462306a36Sopenharmony_ci return; 101562306a36Sopenharmony_ci 101662306a36Sopenharmony_ci __tick_nohz_full_update_tick(ts, ktime_get()); 101762306a36Sopenharmony_ci} 101862306a36Sopenharmony_ci 101962306a36Sopenharmony_ci/* 102062306a36Sopenharmony_ci * A pending softirq outside an IRQ (or softirq disabled section) context 102162306a36Sopenharmony_ci * should be waiting for ksoftirqd to handle it. Therefore we shouldn't 102262306a36Sopenharmony_ci * reach here due to the need_resched() early check in can_stop_idle_tick(). 102362306a36Sopenharmony_ci * 102462306a36Sopenharmony_ci * However if we are between CPUHP_AP_SMPBOOT_THREADS and CPU_TEARDOWN_CPU on the 102562306a36Sopenharmony_ci * cpu_down() process, softirqs can still be raised while ksoftirqd is parked, 102662306a36Sopenharmony_ci * triggering the below since wakep_softirqd() is ignored. 102762306a36Sopenharmony_ci * 102862306a36Sopenharmony_ci */ 102962306a36Sopenharmony_cistatic bool report_idle_softirq(void) 103062306a36Sopenharmony_ci{ 103162306a36Sopenharmony_ci static int ratelimit; 103262306a36Sopenharmony_ci unsigned int pending = local_softirq_pending(); 103362306a36Sopenharmony_ci 103462306a36Sopenharmony_ci if (likely(!pending)) 103562306a36Sopenharmony_ci return false; 103662306a36Sopenharmony_ci 103762306a36Sopenharmony_ci /* Some softirqs claim to be safe against hotplug and ksoftirqd parking */ 103862306a36Sopenharmony_ci if (!cpu_active(smp_processor_id())) { 103962306a36Sopenharmony_ci pending &= ~SOFTIRQ_HOTPLUG_SAFE_MASK; 104062306a36Sopenharmony_ci if (!pending) 104162306a36Sopenharmony_ci return false; 104262306a36Sopenharmony_ci } 104362306a36Sopenharmony_ci 104462306a36Sopenharmony_ci if (ratelimit >= 10) 104562306a36Sopenharmony_ci return false; 104662306a36Sopenharmony_ci 104762306a36Sopenharmony_ci /* On RT, softirqs handling may be waiting on some lock */ 104862306a36Sopenharmony_ci if (local_bh_blocked()) 104962306a36Sopenharmony_ci return false; 105062306a36Sopenharmony_ci 105162306a36Sopenharmony_ci pr_warn("NOHZ tick-stop error: local softirq work is pending, handler #%02x!!!\n", 105262306a36Sopenharmony_ci pending); 105362306a36Sopenharmony_ci ratelimit++; 105462306a36Sopenharmony_ci 105562306a36Sopenharmony_ci return true; 105662306a36Sopenharmony_ci} 105762306a36Sopenharmony_ci 105862306a36Sopenharmony_cistatic bool can_stop_idle_tick(int cpu, struct tick_sched *ts) 105962306a36Sopenharmony_ci{ 106062306a36Sopenharmony_ci /* 106162306a36Sopenharmony_ci * If this CPU is offline and it is the one which updates 106262306a36Sopenharmony_ci * jiffies, then give up the assignment and let it be taken by 106362306a36Sopenharmony_ci * the CPU which runs the tick timer next. If we don't drop 106462306a36Sopenharmony_ci * this here the jiffies might be stale and do_timer() never 106562306a36Sopenharmony_ci * invoked. 106662306a36Sopenharmony_ci */ 106762306a36Sopenharmony_ci if (unlikely(!cpu_online(cpu))) { 106862306a36Sopenharmony_ci if (cpu == tick_do_timer_cpu) 106962306a36Sopenharmony_ci tick_do_timer_cpu = TICK_DO_TIMER_NONE; 107062306a36Sopenharmony_ci /* 107162306a36Sopenharmony_ci * Make sure the CPU doesn't get fooled by obsolete tick 107262306a36Sopenharmony_ci * deadline if it comes back online later. 107362306a36Sopenharmony_ci */ 107462306a36Sopenharmony_ci ts->next_tick = 0; 107562306a36Sopenharmony_ci return false; 107662306a36Sopenharmony_ci } 107762306a36Sopenharmony_ci 107862306a36Sopenharmony_ci if (unlikely(ts->nohz_mode == NOHZ_MODE_INACTIVE)) 107962306a36Sopenharmony_ci return false; 108062306a36Sopenharmony_ci 108162306a36Sopenharmony_ci if (need_resched()) 108262306a36Sopenharmony_ci return false; 108362306a36Sopenharmony_ci 108462306a36Sopenharmony_ci if (unlikely(report_idle_softirq())) 108562306a36Sopenharmony_ci return false; 108662306a36Sopenharmony_ci 108762306a36Sopenharmony_ci if (tick_nohz_full_enabled()) { 108862306a36Sopenharmony_ci /* 108962306a36Sopenharmony_ci * Keep the tick alive to guarantee timekeeping progression 109062306a36Sopenharmony_ci * if there are full dynticks CPUs around 109162306a36Sopenharmony_ci */ 109262306a36Sopenharmony_ci if (tick_do_timer_cpu == cpu) 109362306a36Sopenharmony_ci return false; 109462306a36Sopenharmony_ci 109562306a36Sopenharmony_ci /* Should not happen for nohz-full */ 109662306a36Sopenharmony_ci if (WARN_ON_ONCE(tick_do_timer_cpu == TICK_DO_TIMER_NONE)) 109762306a36Sopenharmony_ci return false; 109862306a36Sopenharmony_ci } 109962306a36Sopenharmony_ci 110062306a36Sopenharmony_ci return true; 110162306a36Sopenharmony_ci} 110262306a36Sopenharmony_ci 110362306a36Sopenharmony_ci/** 110462306a36Sopenharmony_ci * tick_nohz_idle_stop_tick - stop the idle tick from the idle task 110562306a36Sopenharmony_ci * 110662306a36Sopenharmony_ci * When the next event is more than a tick into the future, stop the idle tick 110762306a36Sopenharmony_ci */ 110862306a36Sopenharmony_civoid tick_nohz_idle_stop_tick(void) 110962306a36Sopenharmony_ci{ 111062306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 111162306a36Sopenharmony_ci int cpu = smp_processor_id(); 111262306a36Sopenharmony_ci ktime_t expires; 111362306a36Sopenharmony_ci 111462306a36Sopenharmony_ci /* 111562306a36Sopenharmony_ci * If tick_nohz_get_sleep_length() ran tick_nohz_next_event(), the 111662306a36Sopenharmony_ci * tick timer expiration time is known already. 111762306a36Sopenharmony_ci */ 111862306a36Sopenharmony_ci if (ts->timer_expires_base) 111962306a36Sopenharmony_ci expires = ts->timer_expires; 112062306a36Sopenharmony_ci else if (can_stop_idle_tick(cpu, ts)) 112162306a36Sopenharmony_ci expires = tick_nohz_next_event(ts, cpu); 112262306a36Sopenharmony_ci else 112362306a36Sopenharmony_ci return; 112462306a36Sopenharmony_ci 112562306a36Sopenharmony_ci ts->idle_calls++; 112662306a36Sopenharmony_ci 112762306a36Sopenharmony_ci if (expires > 0LL) { 112862306a36Sopenharmony_ci int was_stopped = ts->tick_stopped; 112962306a36Sopenharmony_ci 113062306a36Sopenharmony_ci tick_nohz_stop_tick(ts, cpu); 113162306a36Sopenharmony_ci 113262306a36Sopenharmony_ci ts->idle_sleeps++; 113362306a36Sopenharmony_ci ts->idle_expires = expires; 113462306a36Sopenharmony_ci 113562306a36Sopenharmony_ci if (!was_stopped && ts->tick_stopped) { 113662306a36Sopenharmony_ci ts->idle_jiffies = ts->last_jiffies; 113762306a36Sopenharmony_ci nohz_balance_enter_idle(cpu); 113862306a36Sopenharmony_ci } 113962306a36Sopenharmony_ci } else { 114062306a36Sopenharmony_ci tick_nohz_retain_tick(ts); 114162306a36Sopenharmony_ci } 114262306a36Sopenharmony_ci} 114362306a36Sopenharmony_ci 114462306a36Sopenharmony_civoid tick_nohz_idle_retain_tick(void) 114562306a36Sopenharmony_ci{ 114662306a36Sopenharmony_ci tick_nohz_retain_tick(this_cpu_ptr(&tick_cpu_sched)); 114762306a36Sopenharmony_ci /* 114862306a36Sopenharmony_ci * Undo the effect of get_next_timer_interrupt() called from 114962306a36Sopenharmony_ci * tick_nohz_next_event(). 115062306a36Sopenharmony_ci */ 115162306a36Sopenharmony_ci timer_clear_idle(); 115262306a36Sopenharmony_ci} 115362306a36Sopenharmony_ci 115462306a36Sopenharmony_ci/** 115562306a36Sopenharmony_ci * tick_nohz_idle_enter - prepare for entering idle on the current CPU 115662306a36Sopenharmony_ci * 115762306a36Sopenharmony_ci * Called when we start the idle loop. 115862306a36Sopenharmony_ci */ 115962306a36Sopenharmony_civoid tick_nohz_idle_enter(void) 116062306a36Sopenharmony_ci{ 116162306a36Sopenharmony_ci struct tick_sched *ts; 116262306a36Sopenharmony_ci 116362306a36Sopenharmony_ci lockdep_assert_irqs_enabled(); 116462306a36Sopenharmony_ci 116562306a36Sopenharmony_ci local_irq_disable(); 116662306a36Sopenharmony_ci 116762306a36Sopenharmony_ci ts = this_cpu_ptr(&tick_cpu_sched); 116862306a36Sopenharmony_ci 116962306a36Sopenharmony_ci WARN_ON_ONCE(ts->timer_expires_base); 117062306a36Sopenharmony_ci 117162306a36Sopenharmony_ci ts->inidle = 1; 117262306a36Sopenharmony_ci tick_nohz_start_idle(ts); 117362306a36Sopenharmony_ci 117462306a36Sopenharmony_ci local_irq_enable(); 117562306a36Sopenharmony_ci} 117662306a36Sopenharmony_ci 117762306a36Sopenharmony_ci/** 117862306a36Sopenharmony_ci * tick_nohz_irq_exit - update next tick event from interrupt exit 117962306a36Sopenharmony_ci * 118062306a36Sopenharmony_ci * When an interrupt fires while we are idle and it doesn't cause 118162306a36Sopenharmony_ci * a reschedule, it may still add, modify or delete a timer, enqueue 118262306a36Sopenharmony_ci * an RCU callback, etc... 118362306a36Sopenharmony_ci * So we need to re-calculate and reprogram the next tick event. 118462306a36Sopenharmony_ci */ 118562306a36Sopenharmony_civoid tick_nohz_irq_exit(void) 118662306a36Sopenharmony_ci{ 118762306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 118862306a36Sopenharmony_ci 118962306a36Sopenharmony_ci if (ts->inidle) 119062306a36Sopenharmony_ci tick_nohz_start_idle(ts); 119162306a36Sopenharmony_ci else 119262306a36Sopenharmony_ci tick_nohz_full_update_tick(ts); 119362306a36Sopenharmony_ci} 119462306a36Sopenharmony_ci 119562306a36Sopenharmony_ci/** 119662306a36Sopenharmony_ci * tick_nohz_idle_got_tick - Check whether or not the tick handler has run 119762306a36Sopenharmony_ci */ 119862306a36Sopenharmony_cibool tick_nohz_idle_got_tick(void) 119962306a36Sopenharmony_ci{ 120062306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 120162306a36Sopenharmony_ci 120262306a36Sopenharmony_ci if (ts->got_idle_tick) { 120362306a36Sopenharmony_ci ts->got_idle_tick = 0; 120462306a36Sopenharmony_ci return true; 120562306a36Sopenharmony_ci } 120662306a36Sopenharmony_ci return false; 120762306a36Sopenharmony_ci} 120862306a36Sopenharmony_ci 120962306a36Sopenharmony_ci/** 121062306a36Sopenharmony_ci * tick_nohz_get_next_hrtimer - return the next expiration time for the hrtimer 121162306a36Sopenharmony_ci * or the tick, whatever that expires first. Note that, if the tick has been 121262306a36Sopenharmony_ci * stopped, it returns the next hrtimer. 121362306a36Sopenharmony_ci * 121462306a36Sopenharmony_ci * Called from power state control code with interrupts disabled 121562306a36Sopenharmony_ci */ 121662306a36Sopenharmony_ciktime_t tick_nohz_get_next_hrtimer(void) 121762306a36Sopenharmony_ci{ 121862306a36Sopenharmony_ci return __this_cpu_read(tick_cpu_device.evtdev)->next_event; 121962306a36Sopenharmony_ci} 122062306a36Sopenharmony_ci 122162306a36Sopenharmony_ci/** 122262306a36Sopenharmony_ci * tick_nohz_get_sleep_length - return the expected length of the current sleep 122362306a36Sopenharmony_ci * @delta_next: duration until the next event if the tick cannot be stopped 122462306a36Sopenharmony_ci * 122562306a36Sopenharmony_ci * Called from power state control code with interrupts disabled. 122662306a36Sopenharmony_ci * 122762306a36Sopenharmony_ci * The return value of this function and/or the value returned by it through the 122862306a36Sopenharmony_ci * @delta_next pointer can be negative which must be taken into account by its 122962306a36Sopenharmony_ci * callers. 123062306a36Sopenharmony_ci */ 123162306a36Sopenharmony_ciktime_t tick_nohz_get_sleep_length(ktime_t *delta_next) 123262306a36Sopenharmony_ci{ 123362306a36Sopenharmony_ci struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev); 123462306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 123562306a36Sopenharmony_ci int cpu = smp_processor_id(); 123662306a36Sopenharmony_ci /* 123762306a36Sopenharmony_ci * The idle entry time is expected to be a sufficient approximation of 123862306a36Sopenharmony_ci * the current time at this point. 123962306a36Sopenharmony_ci */ 124062306a36Sopenharmony_ci ktime_t now = ts->idle_entrytime; 124162306a36Sopenharmony_ci ktime_t next_event; 124262306a36Sopenharmony_ci 124362306a36Sopenharmony_ci WARN_ON_ONCE(!ts->inidle); 124462306a36Sopenharmony_ci 124562306a36Sopenharmony_ci *delta_next = ktime_sub(dev->next_event, now); 124662306a36Sopenharmony_ci 124762306a36Sopenharmony_ci if (!can_stop_idle_tick(cpu, ts)) 124862306a36Sopenharmony_ci return *delta_next; 124962306a36Sopenharmony_ci 125062306a36Sopenharmony_ci next_event = tick_nohz_next_event(ts, cpu); 125162306a36Sopenharmony_ci if (!next_event) 125262306a36Sopenharmony_ci return *delta_next; 125362306a36Sopenharmony_ci 125462306a36Sopenharmony_ci /* 125562306a36Sopenharmony_ci * If the next highres timer to expire is earlier than next_event, the 125662306a36Sopenharmony_ci * idle governor needs to know that. 125762306a36Sopenharmony_ci */ 125862306a36Sopenharmony_ci next_event = min_t(u64, next_event, 125962306a36Sopenharmony_ci hrtimer_next_event_without(&ts->sched_timer)); 126062306a36Sopenharmony_ci 126162306a36Sopenharmony_ci return ktime_sub(next_event, now); 126262306a36Sopenharmony_ci} 126362306a36Sopenharmony_ci 126462306a36Sopenharmony_ci/** 126562306a36Sopenharmony_ci * tick_nohz_get_idle_calls_cpu - return the current idle calls counter value 126662306a36Sopenharmony_ci * for a particular CPU. 126762306a36Sopenharmony_ci * 126862306a36Sopenharmony_ci * Called from the schedutil frequency scaling governor in scheduler context. 126962306a36Sopenharmony_ci */ 127062306a36Sopenharmony_ciunsigned long tick_nohz_get_idle_calls_cpu(int cpu) 127162306a36Sopenharmony_ci{ 127262306a36Sopenharmony_ci struct tick_sched *ts = tick_get_tick_sched(cpu); 127362306a36Sopenharmony_ci 127462306a36Sopenharmony_ci return ts->idle_calls; 127562306a36Sopenharmony_ci} 127662306a36Sopenharmony_ci 127762306a36Sopenharmony_ci/** 127862306a36Sopenharmony_ci * tick_nohz_get_idle_calls - return the current idle calls counter value 127962306a36Sopenharmony_ci * 128062306a36Sopenharmony_ci * Called from the schedutil frequency scaling governor in scheduler context. 128162306a36Sopenharmony_ci */ 128262306a36Sopenharmony_ciunsigned long tick_nohz_get_idle_calls(void) 128362306a36Sopenharmony_ci{ 128462306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 128562306a36Sopenharmony_ci 128662306a36Sopenharmony_ci return ts->idle_calls; 128762306a36Sopenharmony_ci} 128862306a36Sopenharmony_ci 128962306a36Sopenharmony_cistatic void tick_nohz_account_idle_time(struct tick_sched *ts, 129062306a36Sopenharmony_ci ktime_t now) 129162306a36Sopenharmony_ci{ 129262306a36Sopenharmony_ci unsigned long ticks; 129362306a36Sopenharmony_ci 129462306a36Sopenharmony_ci ts->idle_exittime = now; 129562306a36Sopenharmony_ci 129662306a36Sopenharmony_ci if (vtime_accounting_enabled_this_cpu()) 129762306a36Sopenharmony_ci return; 129862306a36Sopenharmony_ci /* 129962306a36Sopenharmony_ci * We stopped the tick in idle. Update process times would miss the 130062306a36Sopenharmony_ci * time we slept as update_process_times does only a 1 tick 130162306a36Sopenharmony_ci * accounting. Enforce that this is accounted to idle ! 130262306a36Sopenharmony_ci */ 130362306a36Sopenharmony_ci ticks = jiffies - ts->idle_jiffies; 130462306a36Sopenharmony_ci /* 130562306a36Sopenharmony_ci * We might be one off. Do not randomly account a huge number of ticks! 130662306a36Sopenharmony_ci */ 130762306a36Sopenharmony_ci if (ticks && ticks < LONG_MAX) 130862306a36Sopenharmony_ci account_idle_ticks(ticks); 130962306a36Sopenharmony_ci} 131062306a36Sopenharmony_ci 131162306a36Sopenharmony_civoid tick_nohz_idle_restart_tick(void) 131262306a36Sopenharmony_ci{ 131362306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 131462306a36Sopenharmony_ci 131562306a36Sopenharmony_ci if (ts->tick_stopped) { 131662306a36Sopenharmony_ci ktime_t now = ktime_get(); 131762306a36Sopenharmony_ci tick_nohz_restart_sched_tick(ts, now); 131862306a36Sopenharmony_ci tick_nohz_account_idle_time(ts, now); 131962306a36Sopenharmony_ci } 132062306a36Sopenharmony_ci} 132162306a36Sopenharmony_ci 132262306a36Sopenharmony_cistatic void tick_nohz_idle_update_tick(struct tick_sched *ts, ktime_t now) 132362306a36Sopenharmony_ci{ 132462306a36Sopenharmony_ci if (tick_nohz_full_cpu(smp_processor_id())) 132562306a36Sopenharmony_ci __tick_nohz_full_update_tick(ts, now); 132662306a36Sopenharmony_ci else 132762306a36Sopenharmony_ci tick_nohz_restart_sched_tick(ts, now); 132862306a36Sopenharmony_ci 132962306a36Sopenharmony_ci tick_nohz_account_idle_time(ts, now); 133062306a36Sopenharmony_ci} 133162306a36Sopenharmony_ci 133262306a36Sopenharmony_ci/** 133362306a36Sopenharmony_ci * tick_nohz_idle_exit - restart the idle tick from the idle task 133462306a36Sopenharmony_ci * 133562306a36Sopenharmony_ci * Restart the idle tick when the CPU is woken up from idle 133662306a36Sopenharmony_ci * This also exit the RCU extended quiescent state. The CPU 133762306a36Sopenharmony_ci * can use RCU again after this function is called. 133862306a36Sopenharmony_ci */ 133962306a36Sopenharmony_civoid tick_nohz_idle_exit(void) 134062306a36Sopenharmony_ci{ 134162306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 134262306a36Sopenharmony_ci bool idle_active, tick_stopped; 134362306a36Sopenharmony_ci ktime_t now; 134462306a36Sopenharmony_ci 134562306a36Sopenharmony_ci local_irq_disable(); 134662306a36Sopenharmony_ci 134762306a36Sopenharmony_ci WARN_ON_ONCE(!ts->inidle); 134862306a36Sopenharmony_ci WARN_ON_ONCE(ts->timer_expires_base); 134962306a36Sopenharmony_ci 135062306a36Sopenharmony_ci ts->inidle = 0; 135162306a36Sopenharmony_ci idle_active = ts->idle_active; 135262306a36Sopenharmony_ci tick_stopped = ts->tick_stopped; 135362306a36Sopenharmony_ci 135462306a36Sopenharmony_ci if (idle_active || tick_stopped) 135562306a36Sopenharmony_ci now = ktime_get(); 135662306a36Sopenharmony_ci 135762306a36Sopenharmony_ci if (idle_active) 135862306a36Sopenharmony_ci tick_nohz_stop_idle(ts, now); 135962306a36Sopenharmony_ci 136062306a36Sopenharmony_ci if (tick_stopped) 136162306a36Sopenharmony_ci tick_nohz_idle_update_tick(ts, now); 136262306a36Sopenharmony_ci 136362306a36Sopenharmony_ci local_irq_enable(); 136462306a36Sopenharmony_ci} 136562306a36Sopenharmony_ci 136662306a36Sopenharmony_ci/* 136762306a36Sopenharmony_ci * The nohz low res interrupt handler 136862306a36Sopenharmony_ci */ 136962306a36Sopenharmony_cistatic void tick_nohz_handler(struct clock_event_device *dev) 137062306a36Sopenharmony_ci{ 137162306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 137262306a36Sopenharmony_ci struct pt_regs *regs = get_irq_regs(); 137362306a36Sopenharmony_ci ktime_t now = ktime_get(); 137462306a36Sopenharmony_ci 137562306a36Sopenharmony_ci dev->next_event = KTIME_MAX; 137662306a36Sopenharmony_ci 137762306a36Sopenharmony_ci tick_sched_do_timer(ts, now); 137862306a36Sopenharmony_ci tick_sched_handle(ts, regs); 137962306a36Sopenharmony_ci 138062306a36Sopenharmony_ci if (unlikely(ts->tick_stopped)) { 138162306a36Sopenharmony_ci /* 138262306a36Sopenharmony_ci * The clockevent device is not reprogrammed, so change the 138362306a36Sopenharmony_ci * clock event device to ONESHOT_STOPPED to avoid spurious 138462306a36Sopenharmony_ci * interrupts on devices which might not be truly one shot. 138562306a36Sopenharmony_ci */ 138662306a36Sopenharmony_ci tick_program_event(KTIME_MAX, 1); 138762306a36Sopenharmony_ci return; 138862306a36Sopenharmony_ci } 138962306a36Sopenharmony_ci 139062306a36Sopenharmony_ci hrtimer_forward(&ts->sched_timer, now, TICK_NSEC); 139162306a36Sopenharmony_ci tick_program_event(hrtimer_get_expires(&ts->sched_timer), 1); 139262306a36Sopenharmony_ci} 139362306a36Sopenharmony_ci 139462306a36Sopenharmony_cistatic inline void tick_nohz_activate(struct tick_sched *ts, int mode) 139562306a36Sopenharmony_ci{ 139662306a36Sopenharmony_ci if (!tick_nohz_enabled) 139762306a36Sopenharmony_ci return; 139862306a36Sopenharmony_ci ts->nohz_mode = mode; 139962306a36Sopenharmony_ci /* One update is enough */ 140062306a36Sopenharmony_ci if (!test_and_set_bit(0, &tick_nohz_active)) 140162306a36Sopenharmony_ci timers_update_nohz(); 140262306a36Sopenharmony_ci} 140362306a36Sopenharmony_ci 140462306a36Sopenharmony_ci/** 140562306a36Sopenharmony_ci * tick_nohz_switch_to_nohz - switch to nohz mode 140662306a36Sopenharmony_ci */ 140762306a36Sopenharmony_cistatic void tick_nohz_switch_to_nohz(void) 140862306a36Sopenharmony_ci{ 140962306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 141062306a36Sopenharmony_ci ktime_t next; 141162306a36Sopenharmony_ci 141262306a36Sopenharmony_ci if (!tick_nohz_enabled) 141362306a36Sopenharmony_ci return; 141462306a36Sopenharmony_ci 141562306a36Sopenharmony_ci if (tick_switch_to_oneshot(tick_nohz_handler)) 141662306a36Sopenharmony_ci return; 141762306a36Sopenharmony_ci 141862306a36Sopenharmony_ci /* 141962306a36Sopenharmony_ci * Recycle the hrtimer in ts, so we can share the 142062306a36Sopenharmony_ci * hrtimer_forward with the highres code. 142162306a36Sopenharmony_ci */ 142262306a36Sopenharmony_ci hrtimer_init(&ts->sched_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS_HARD); 142362306a36Sopenharmony_ci /* Get the next period */ 142462306a36Sopenharmony_ci next = tick_init_jiffy_update(); 142562306a36Sopenharmony_ci 142662306a36Sopenharmony_ci hrtimer_set_expires(&ts->sched_timer, next); 142762306a36Sopenharmony_ci hrtimer_forward_now(&ts->sched_timer, TICK_NSEC); 142862306a36Sopenharmony_ci tick_program_event(hrtimer_get_expires(&ts->sched_timer), 1); 142962306a36Sopenharmony_ci tick_nohz_activate(ts, NOHZ_MODE_LOWRES); 143062306a36Sopenharmony_ci} 143162306a36Sopenharmony_ci 143262306a36Sopenharmony_cistatic inline void tick_nohz_irq_enter(void) 143362306a36Sopenharmony_ci{ 143462306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 143562306a36Sopenharmony_ci ktime_t now; 143662306a36Sopenharmony_ci 143762306a36Sopenharmony_ci if (!ts->idle_active && !ts->tick_stopped) 143862306a36Sopenharmony_ci return; 143962306a36Sopenharmony_ci now = ktime_get(); 144062306a36Sopenharmony_ci if (ts->idle_active) 144162306a36Sopenharmony_ci tick_nohz_stop_idle(ts, now); 144262306a36Sopenharmony_ci /* 144362306a36Sopenharmony_ci * If all CPUs are idle. We may need to update a stale jiffies value. 144462306a36Sopenharmony_ci * Note nohz_full is a special case: a timekeeper is guaranteed to stay 144562306a36Sopenharmony_ci * alive but it might be busy looping with interrupts disabled in some 144662306a36Sopenharmony_ci * rare case (typically stop machine). So we must make sure we have a 144762306a36Sopenharmony_ci * last resort. 144862306a36Sopenharmony_ci */ 144962306a36Sopenharmony_ci if (ts->tick_stopped) 145062306a36Sopenharmony_ci tick_nohz_update_jiffies(now); 145162306a36Sopenharmony_ci} 145262306a36Sopenharmony_ci 145362306a36Sopenharmony_ci#else 145462306a36Sopenharmony_ci 145562306a36Sopenharmony_cistatic inline void tick_nohz_switch_to_nohz(void) { } 145662306a36Sopenharmony_cistatic inline void tick_nohz_irq_enter(void) { } 145762306a36Sopenharmony_cistatic inline void tick_nohz_activate(struct tick_sched *ts, int mode) { } 145862306a36Sopenharmony_ci 145962306a36Sopenharmony_ci#endif /* CONFIG_NO_HZ_COMMON */ 146062306a36Sopenharmony_ci 146162306a36Sopenharmony_ci/* 146262306a36Sopenharmony_ci * Called from irq_enter to notify about the possible interruption of idle() 146362306a36Sopenharmony_ci */ 146462306a36Sopenharmony_civoid tick_irq_enter(void) 146562306a36Sopenharmony_ci{ 146662306a36Sopenharmony_ci tick_check_oneshot_broadcast_this_cpu(); 146762306a36Sopenharmony_ci tick_nohz_irq_enter(); 146862306a36Sopenharmony_ci} 146962306a36Sopenharmony_ci 147062306a36Sopenharmony_ci/* 147162306a36Sopenharmony_ci * High resolution timer specific code 147262306a36Sopenharmony_ci */ 147362306a36Sopenharmony_ci#ifdef CONFIG_HIGH_RES_TIMERS 147462306a36Sopenharmony_ci/* 147562306a36Sopenharmony_ci * We rearm the timer until we get disabled by the idle code. 147662306a36Sopenharmony_ci * Called with interrupts disabled. 147762306a36Sopenharmony_ci */ 147862306a36Sopenharmony_cistatic enum hrtimer_restart tick_sched_timer(struct hrtimer *timer) 147962306a36Sopenharmony_ci{ 148062306a36Sopenharmony_ci struct tick_sched *ts = 148162306a36Sopenharmony_ci container_of(timer, struct tick_sched, sched_timer); 148262306a36Sopenharmony_ci struct pt_regs *regs = get_irq_regs(); 148362306a36Sopenharmony_ci ktime_t now = ktime_get(); 148462306a36Sopenharmony_ci 148562306a36Sopenharmony_ci tick_sched_do_timer(ts, now); 148662306a36Sopenharmony_ci 148762306a36Sopenharmony_ci /* 148862306a36Sopenharmony_ci * Do not call, when we are not in irq context and have 148962306a36Sopenharmony_ci * no valid regs pointer 149062306a36Sopenharmony_ci */ 149162306a36Sopenharmony_ci if (regs) 149262306a36Sopenharmony_ci tick_sched_handle(ts, regs); 149362306a36Sopenharmony_ci else 149462306a36Sopenharmony_ci ts->next_tick = 0; 149562306a36Sopenharmony_ci 149662306a36Sopenharmony_ci /* No need to reprogram if we are in idle or full dynticks mode */ 149762306a36Sopenharmony_ci if (unlikely(ts->tick_stopped)) 149862306a36Sopenharmony_ci return HRTIMER_NORESTART; 149962306a36Sopenharmony_ci 150062306a36Sopenharmony_ci hrtimer_forward(timer, now, TICK_NSEC); 150162306a36Sopenharmony_ci 150262306a36Sopenharmony_ci return HRTIMER_RESTART; 150362306a36Sopenharmony_ci} 150462306a36Sopenharmony_ci 150562306a36Sopenharmony_cistatic int sched_skew_tick; 150662306a36Sopenharmony_ci 150762306a36Sopenharmony_cistatic int __init skew_tick(char *str) 150862306a36Sopenharmony_ci{ 150962306a36Sopenharmony_ci get_option(&str, &sched_skew_tick); 151062306a36Sopenharmony_ci 151162306a36Sopenharmony_ci return 0; 151262306a36Sopenharmony_ci} 151362306a36Sopenharmony_ciearly_param("skew_tick", skew_tick); 151462306a36Sopenharmony_ci 151562306a36Sopenharmony_ci/** 151662306a36Sopenharmony_ci * tick_setup_sched_timer - setup the tick emulation timer 151762306a36Sopenharmony_ci */ 151862306a36Sopenharmony_civoid tick_setup_sched_timer(void) 151962306a36Sopenharmony_ci{ 152062306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 152162306a36Sopenharmony_ci ktime_t now = ktime_get(); 152262306a36Sopenharmony_ci 152362306a36Sopenharmony_ci /* 152462306a36Sopenharmony_ci * Emulate tick processing via per-CPU hrtimers: 152562306a36Sopenharmony_ci */ 152662306a36Sopenharmony_ci hrtimer_init(&ts->sched_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS_HARD); 152762306a36Sopenharmony_ci ts->sched_timer.function = tick_sched_timer; 152862306a36Sopenharmony_ci 152962306a36Sopenharmony_ci /* Get the next period (per-CPU) */ 153062306a36Sopenharmony_ci hrtimer_set_expires(&ts->sched_timer, tick_init_jiffy_update()); 153162306a36Sopenharmony_ci 153262306a36Sopenharmony_ci /* Offset the tick to avert jiffies_lock contention. */ 153362306a36Sopenharmony_ci if (sched_skew_tick) { 153462306a36Sopenharmony_ci u64 offset = TICK_NSEC >> 1; 153562306a36Sopenharmony_ci do_div(offset, num_possible_cpus()); 153662306a36Sopenharmony_ci offset *= smp_processor_id(); 153762306a36Sopenharmony_ci hrtimer_add_expires_ns(&ts->sched_timer, offset); 153862306a36Sopenharmony_ci } 153962306a36Sopenharmony_ci 154062306a36Sopenharmony_ci hrtimer_forward(&ts->sched_timer, now, TICK_NSEC); 154162306a36Sopenharmony_ci hrtimer_start_expires(&ts->sched_timer, HRTIMER_MODE_ABS_PINNED_HARD); 154262306a36Sopenharmony_ci tick_nohz_activate(ts, NOHZ_MODE_HIGHRES); 154362306a36Sopenharmony_ci} 154462306a36Sopenharmony_ci#endif /* HIGH_RES_TIMERS */ 154562306a36Sopenharmony_ci 154662306a36Sopenharmony_ci#if defined CONFIG_NO_HZ_COMMON || defined CONFIG_HIGH_RES_TIMERS 154762306a36Sopenharmony_civoid tick_cancel_sched_timer(int cpu) 154862306a36Sopenharmony_ci{ 154962306a36Sopenharmony_ci struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); 155062306a36Sopenharmony_ci ktime_t idle_sleeptime, iowait_sleeptime; 155162306a36Sopenharmony_ci unsigned long idle_calls, idle_sleeps; 155262306a36Sopenharmony_ci 155362306a36Sopenharmony_ci# ifdef CONFIG_HIGH_RES_TIMERS 155462306a36Sopenharmony_ci if (ts->sched_timer.base) 155562306a36Sopenharmony_ci hrtimer_cancel(&ts->sched_timer); 155662306a36Sopenharmony_ci# endif 155762306a36Sopenharmony_ci 155862306a36Sopenharmony_ci idle_sleeptime = ts->idle_sleeptime; 155962306a36Sopenharmony_ci iowait_sleeptime = ts->iowait_sleeptime; 156062306a36Sopenharmony_ci idle_calls = ts->idle_calls; 156162306a36Sopenharmony_ci idle_sleeps = ts->idle_sleeps; 156262306a36Sopenharmony_ci memset(ts, 0, sizeof(*ts)); 156362306a36Sopenharmony_ci ts->idle_sleeptime = idle_sleeptime; 156462306a36Sopenharmony_ci ts->iowait_sleeptime = iowait_sleeptime; 156562306a36Sopenharmony_ci ts->idle_calls = idle_calls; 156662306a36Sopenharmony_ci ts->idle_sleeps = idle_sleeps; 156762306a36Sopenharmony_ci} 156862306a36Sopenharmony_ci#endif 156962306a36Sopenharmony_ci 157062306a36Sopenharmony_ci/* 157162306a36Sopenharmony_ci * Async notification about clocksource changes 157262306a36Sopenharmony_ci */ 157362306a36Sopenharmony_civoid tick_clock_notify(void) 157462306a36Sopenharmony_ci{ 157562306a36Sopenharmony_ci int cpu; 157662306a36Sopenharmony_ci 157762306a36Sopenharmony_ci for_each_possible_cpu(cpu) 157862306a36Sopenharmony_ci set_bit(0, &per_cpu(tick_cpu_sched, cpu).check_clocks); 157962306a36Sopenharmony_ci} 158062306a36Sopenharmony_ci 158162306a36Sopenharmony_ci/* 158262306a36Sopenharmony_ci * Async notification about clock event changes 158362306a36Sopenharmony_ci */ 158462306a36Sopenharmony_civoid tick_oneshot_notify(void) 158562306a36Sopenharmony_ci{ 158662306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 158762306a36Sopenharmony_ci 158862306a36Sopenharmony_ci set_bit(0, &ts->check_clocks); 158962306a36Sopenharmony_ci} 159062306a36Sopenharmony_ci 159162306a36Sopenharmony_ci/* 159262306a36Sopenharmony_ci * Check, if a change happened, which makes oneshot possible. 159362306a36Sopenharmony_ci * 159462306a36Sopenharmony_ci * Called cyclic from the hrtimer softirq (driven by the timer 159562306a36Sopenharmony_ci * softirq) allow_nohz signals, that we can switch into low-res nohz 159662306a36Sopenharmony_ci * mode, because high resolution timers are disabled (either compile 159762306a36Sopenharmony_ci * or runtime). Called with interrupts disabled. 159862306a36Sopenharmony_ci */ 159962306a36Sopenharmony_ciint tick_check_oneshot_change(int allow_nohz) 160062306a36Sopenharmony_ci{ 160162306a36Sopenharmony_ci struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); 160262306a36Sopenharmony_ci 160362306a36Sopenharmony_ci if (!test_and_clear_bit(0, &ts->check_clocks)) 160462306a36Sopenharmony_ci return 0; 160562306a36Sopenharmony_ci 160662306a36Sopenharmony_ci if (ts->nohz_mode != NOHZ_MODE_INACTIVE) 160762306a36Sopenharmony_ci return 0; 160862306a36Sopenharmony_ci 160962306a36Sopenharmony_ci if (!timekeeping_valid_for_hres() || !tick_is_oneshot_available()) 161062306a36Sopenharmony_ci return 0; 161162306a36Sopenharmony_ci 161262306a36Sopenharmony_ci if (!allow_nohz) 161362306a36Sopenharmony_ci return 1; 161462306a36Sopenharmony_ci 161562306a36Sopenharmony_ci tick_nohz_switch_to_nohz(); 161662306a36Sopenharmony_ci return 0; 161762306a36Sopenharmony_ci} 1618