|  | /* SPDX-License-Identifier: GPL-2.0 */ | 
|  | /* | 
|  | * Tick related global functions | 
|  | */ | 
|  | #ifndef _LINUX_TICK_H | 
|  | #define _LINUX_TICK_H | 
|  |  | 
|  | #include <linux/clockchips.h> | 
|  | #include <linux/irqflags.h> | 
|  | #include <linux/percpu.h> | 
|  | #include <linux/context_tracking_state.h> | 
|  | #include <linux/cpumask.h> | 
|  | #include <linux/sched.h> | 
|  | #include <linux/rcupdate.h> | 
|  |  | 
|  | #ifdef CONFIG_GENERIC_CLOCKEVENTS | 
|  | extern void __init tick_init(void); | 
|  | /* Should be core only, but ARM BL switcher requires it */ | 
|  | extern void tick_suspend_local(void); | 
|  | /* Should be core only, but XEN resume magic and ARM BL switcher require it */ | 
|  | extern void tick_resume_local(void); | 
|  | extern void tick_handover_do_timer(void); | 
|  | extern void tick_cleanup_dead_cpu(int cpu); | 
|  | #else /* CONFIG_GENERIC_CLOCKEVENTS */ | 
|  | static inline void tick_init(void) { } | 
|  | static inline void tick_suspend_local(void) { } | 
|  | static inline void tick_resume_local(void) { } | 
|  | static inline void tick_handover_do_timer(void) { } | 
|  | static inline void tick_cleanup_dead_cpu(int cpu) { } | 
|  | #endif /* !CONFIG_GENERIC_CLOCKEVENTS */ | 
|  |  | 
|  | #if defined(CONFIG_GENERIC_CLOCKEVENTS) && defined(CONFIG_SUSPEND) | 
|  | extern void tick_freeze(void); | 
|  | extern void tick_unfreeze(void); | 
|  | #else | 
|  | static inline void tick_freeze(void) { } | 
|  | static inline void tick_unfreeze(void) { } | 
|  | #endif | 
|  |  | 
|  | #ifdef CONFIG_TICK_ONESHOT | 
|  | extern void tick_irq_enter(void); | 
|  | #  ifndef arch_needs_cpu | 
|  | #   define arch_needs_cpu() (0) | 
|  | #  endif | 
|  | # else | 
|  | static inline void tick_irq_enter(void) { } | 
|  | #endif | 
|  |  | 
|  | #if defined(CONFIG_GENERIC_CLOCKEVENTS_BROADCAST) && defined(CONFIG_TICK_ONESHOT) | 
|  | extern void hotplug_cpu__broadcast_tick_pull(int dead_cpu); | 
|  | #else | 
|  | static inline void hotplug_cpu__broadcast_tick_pull(int dead_cpu) { } | 
|  | #endif | 
|  |  | 
|  | enum tick_broadcast_mode { | 
|  | TICK_BROADCAST_OFF, | 
|  | TICK_BROADCAST_ON, | 
|  | TICK_BROADCAST_FORCE, | 
|  | }; | 
|  |  | 
|  | enum tick_broadcast_state { | 
|  | TICK_BROADCAST_EXIT, | 
|  | TICK_BROADCAST_ENTER, | 
|  | }; | 
|  |  | 
|  | #ifdef CONFIG_GENERIC_CLOCKEVENTS_BROADCAST | 
|  | extern void tick_broadcast_control(enum tick_broadcast_mode mode); | 
|  | #else | 
|  | static inline void tick_broadcast_control(enum tick_broadcast_mode mode) { } | 
|  | #endif /* BROADCAST */ | 
|  |  | 
|  | #if defined(CONFIG_GENERIC_CLOCKEVENTS_BROADCAST) && defined(CONFIG_HOTPLUG_CPU) | 
|  | extern void tick_offline_cpu(unsigned int cpu); | 
|  | #else | 
|  | static inline void tick_offline_cpu(unsigned int cpu) { } | 
|  | #endif | 
|  |  | 
|  | #ifdef CONFIG_GENERIC_CLOCKEVENTS | 
|  | extern int tick_broadcast_oneshot_control(enum tick_broadcast_state state); | 
|  | #else | 
|  | static inline int tick_broadcast_oneshot_control(enum tick_broadcast_state state) | 
|  | { | 
|  | return 0; | 
|  | } | 
|  | #endif | 
|  |  | 
|  | static inline void tick_broadcast_enable(void) | 
|  | { | 
|  | tick_broadcast_control(TICK_BROADCAST_ON); | 
|  | } | 
|  | static inline void tick_broadcast_disable(void) | 
|  | { | 
|  | tick_broadcast_control(TICK_BROADCAST_OFF); | 
|  | } | 
|  | static inline void tick_broadcast_force(void) | 
|  | { | 
|  | tick_broadcast_control(TICK_BROADCAST_FORCE); | 
|  | } | 
|  | static inline int tick_broadcast_enter(void) | 
|  | { | 
|  | return tick_broadcast_oneshot_control(TICK_BROADCAST_ENTER); | 
|  | } | 
|  | static inline void tick_broadcast_exit(void) | 
|  | { | 
|  | tick_broadcast_oneshot_control(TICK_BROADCAST_EXIT); | 
|  | } | 
|  |  | 
|  | enum tick_dep_bits { | 
|  | TICK_DEP_BIT_POSIX_TIMER	= 0, | 
|  | TICK_DEP_BIT_PERF_EVENTS	= 1, | 
|  | TICK_DEP_BIT_SCHED		= 2, | 
|  | TICK_DEP_BIT_CLOCK_UNSTABLE	= 3, | 
|  | TICK_DEP_BIT_RCU		= 4, | 
|  | TICK_DEP_BIT_RCU_EXP		= 5 | 
|  | }; | 
|  | #define TICK_DEP_BIT_MAX TICK_DEP_BIT_RCU_EXP | 
|  |  | 
|  | #define TICK_DEP_MASK_NONE		0 | 
|  | #define TICK_DEP_MASK_POSIX_TIMER	(1 << TICK_DEP_BIT_POSIX_TIMER) | 
|  | #define TICK_DEP_MASK_PERF_EVENTS	(1 << TICK_DEP_BIT_PERF_EVENTS) | 
|  | #define TICK_DEP_MASK_SCHED		(1 << TICK_DEP_BIT_SCHED) | 
|  | #define TICK_DEP_MASK_CLOCK_UNSTABLE	(1 << TICK_DEP_BIT_CLOCK_UNSTABLE) | 
|  | #define TICK_DEP_MASK_RCU		(1 << TICK_DEP_BIT_RCU) | 
|  | #define TICK_DEP_MASK_RCU_EXP		(1 << TICK_DEP_BIT_RCU_EXP) | 
|  |  | 
|  | #ifdef CONFIG_NO_HZ_COMMON | 
|  | extern bool tick_nohz_enabled; | 
|  | extern bool tick_nohz_tick_stopped(void); | 
|  | extern bool tick_nohz_tick_stopped_cpu(int cpu); | 
|  | extern void tick_nohz_idle_stop_tick(void); | 
|  | extern void tick_nohz_idle_retain_tick(void); | 
|  | extern void tick_nohz_idle_restart_tick(void); | 
|  | extern void tick_nohz_idle_enter(void); | 
|  | extern void tick_nohz_idle_exit(void); | 
|  | extern void tick_nohz_irq_exit(void); | 
|  | extern bool tick_nohz_idle_got_tick(void); | 
|  | extern ktime_t tick_nohz_get_next_hrtimer(void); | 
|  | extern ktime_t tick_nohz_get_sleep_length(ktime_t *delta_next); | 
|  | extern unsigned long tick_nohz_get_idle_calls(void); | 
|  | extern unsigned long tick_nohz_get_idle_calls_cpu(int cpu); | 
|  | extern u64 get_cpu_idle_time_us(int cpu, u64 *last_update_time); | 
|  | extern u64 get_cpu_iowait_time_us(int cpu, u64 *last_update_time); | 
|  |  | 
|  | static inline void tick_nohz_idle_stop_tick_protected(void) | 
|  | { | 
|  | local_irq_disable(); | 
|  | tick_nohz_idle_stop_tick(); | 
|  | local_irq_enable(); | 
|  | } | 
|  |  | 
|  | #else /* !CONFIG_NO_HZ_COMMON */ | 
|  | #define tick_nohz_enabled (0) | 
|  | static inline int tick_nohz_tick_stopped(void) { return 0; } | 
|  | static inline int tick_nohz_tick_stopped_cpu(int cpu) { return 0; } | 
|  | static inline void tick_nohz_idle_stop_tick(void) { } | 
|  | static inline void tick_nohz_idle_retain_tick(void) { } | 
|  | static inline void tick_nohz_idle_restart_tick(void) { } | 
|  | static inline void tick_nohz_idle_enter(void) { } | 
|  | static inline void tick_nohz_idle_exit(void) { } | 
|  | static inline bool tick_nohz_idle_got_tick(void) { return false; } | 
|  | static inline ktime_t tick_nohz_get_next_hrtimer(void) | 
|  | { | 
|  | /* Next wake up is the tick period, assume it starts now */ | 
|  | return ktime_add(ktime_get(), TICK_NSEC); | 
|  | } | 
|  | static inline ktime_t tick_nohz_get_sleep_length(ktime_t *delta_next) | 
|  | { | 
|  | *delta_next = TICK_NSEC; | 
|  | return *delta_next; | 
|  | } | 
|  | static inline u64 get_cpu_idle_time_us(int cpu, u64 *unused) { return -1; } | 
|  | static inline u64 get_cpu_iowait_time_us(int cpu, u64 *unused) { return -1; } | 
|  |  | 
|  | static inline void tick_nohz_idle_stop_tick_protected(void) { } | 
|  | #endif /* !CONFIG_NO_HZ_COMMON */ | 
|  |  | 
|  | #ifdef CONFIG_NO_HZ_FULL | 
|  | extern bool tick_nohz_full_running; | 
|  | extern cpumask_var_t tick_nohz_full_mask; | 
|  |  | 
|  | static inline bool tick_nohz_full_enabled(void) | 
|  | { | 
|  | if (!context_tracking_enabled()) | 
|  | return false; | 
|  |  | 
|  | return tick_nohz_full_running; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Check if a CPU is part of the nohz_full subset. Arrange for evaluating | 
|  | * the cpu expression (typically smp_processor_id()) _after_ the static | 
|  | * key. | 
|  | */ | 
|  | #define tick_nohz_full_cpu(_cpu) ({					\ | 
|  | bool __ret = false;						\ | 
|  | if (tick_nohz_full_enabled())					\ | 
|  | __ret = cpumask_test_cpu((_cpu), tick_nohz_full_mask);	\ | 
|  | __ret;								\ | 
|  | }) | 
|  |  | 
|  | static inline void tick_nohz_full_add_cpus_to(struct cpumask *mask) | 
|  | { | 
|  | if (tick_nohz_full_enabled()) | 
|  | cpumask_or(mask, mask, tick_nohz_full_mask); | 
|  | } | 
|  |  | 
|  | extern void tick_nohz_dep_set(enum tick_dep_bits bit); | 
|  | extern void tick_nohz_dep_clear(enum tick_dep_bits bit); | 
|  | extern void tick_nohz_dep_set_cpu(int cpu, enum tick_dep_bits bit); | 
|  | extern void tick_nohz_dep_clear_cpu(int cpu, enum tick_dep_bits bit); | 
|  | extern void tick_nohz_dep_set_task(struct task_struct *tsk, | 
|  | enum tick_dep_bits bit); | 
|  | extern void tick_nohz_dep_clear_task(struct task_struct *tsk, | 
|  | enum tick_dep_bits bit); | 
|  | extern void tick_nohz_dep_set_signal(struct task_struct *tsk, | 
|  | enum tick_dep_bits bit); | 
|  | extern void tick_nohz_dep_clear_signal(struct signal_struct *signal, | 
|  | enum tick_dep_bits bit); | 
|  |  | 
|  | /* | 
|  | * The below are tick_nohz_[set,clear]_dep() wrappers that optimize off-cases | 
|  | * on top of static keys. | 
|  | */ | 
|  | static inline void tick_dep_set(enum tick_dep_bits bit) | 
|  | { | 
|  | if (tick_nohz_full_enabled()) | 
|  | tick_nohz_dep_set(bit); | 
|  | } | 
|  |  | 
|  | static inline void tick_dep_clear(enum tick_dep_bits bit) | 
|  | { | 
|  | if (tick_nohz_full_enabled()) | 
|  | tick_nohz_dep_clear(bit); | 
|  | } | 
|  |  | 
|  | static inline void tick_dep_set_cpu(int cpu, enum tick_dep_bits bit) | 
|  | { | 
|  | if (tick_nohz_full_cpu(cpu)) | 
|  | tick_nohz_dep_set_cpu(cpu, bit); | 
|  | } | 
|  |  | 
|  | static inline void tick_dep_clear_cpu(int cpu, enum tick_dep_bits bit) | 
|  | { | 
|  | if (tick_nohz_full_cpu(cpu)) | 
|  | tick_nohz_dep_clear_cpu(cpu, bit); | 
|  | } | 
|  |  | 
|  | static inline void tick_dep_set_task(struct task_struct *tsk, | 
|  | enum tick_dep_bits bit) | 
|  | { | 
|  | if (tick_nohz_full_enabled()) | 
|  | tick_nohz_dep_set_task(tsk, bit); | 
|  | } | 
|  | static inline void tick_dep_clear_task(struct task_struct *tsk, | 
|  | enum tick_dep_bits bit) | 
|  | { | 
|  | if (tick_nohz_full_enabled()) | 
|  | tick_nohz_dep_clear_task(tsk, bit); | 
|  | } | 
|  | static inline void tick_dep_set_signal(struct task_struct *tsk, | 
|  | enum tick_dep_bits bit) | 
|  | { | 
|  | if (tick_nohz_full_enabled()) | 
|  | tick_nohz_dep_set_signal(tsk, bit); | 
|  | } | 
|  | static inline void tick_dep_clear_signal(struct signal_struct *signal, | 
|  | enum tick_dep_bits bit) | 
|  | { | 
|  | if (tick_nohz_full_enabled()) | 
|  | tick_nohz_dep_clear_signal(signal, bit); | 
|  | } | 
|  |  | 
|  | extern void tick_nohz_full_kick_cpu(int cpu); | 
|  | extern void __tick_nohz_task_switch(void); | 
|  | extern void __init tick_nohz_full_setup(cpumask_var_t cpumask); | 
|  | #else | 
|  | static inline bool tick_nohz_full_enabled(void) { return false; } | 
|  | static inline bool tick_nohz_full_cpu(int cpu) { return false; } | 
|  | static inline void tick_nohz_full_add_cpus_to(struct cpumask *mask) { } | 
|  |  | 
|  | static inline void tick_nohz_dep_set_cpu(int cpu, enum tick_dep_bits bit) { } | 
|  | static inline void tick_nohz_dep_clear_cpu(int cpu, enum tick_dep_bits bit) { } | 
|  |  | 
|  | static inline void tick_dep_set(enum tick_dep_bits bit) { } | 
|  | static inline void tick_dep_clear(enum tick_dep_bits bit) { } | 
|  | static inline void tick_dep_set_cpu(int cpu, enum tick_dep_bits bit) { } | 
|  | static inline void tick_dep_clear_cpu(int cpu, enum tick_dep_bits bit) { } | 
|  | static inline void tick_dep_set_task(struct task_struct *tsk, | 
|  | enum tick_dep_bits bit) { } | 
|  | static inline void tick_dep_clear_task(struct task_struct *tsk, | 
|  | enum tick_dep_bits bit) { } | 
|  | static inline void tick_dep_set_signal(struct task_struct *tsk, | 
|  | enum tick_dep_bits bit) { } | 
|  | static inline void tick_dep_clear_signal(struct signal_struct *signal, | 
|  | enum tick_dep_bits bit) { } | 
|  |  | 
|  | static inline void tick_nohz_full_kick_cpu(int cpu) { } | 
|  | static inline void __tick_nohz_task_switch(void) { } | 
|  | static inline void tick_nohz_full_setup(cpumask_var_t cpumask) { } | 
|  | #endif | 
|  |  | 
|  | static inline void tick_nohz_task_switch(void) | 
|  | { | 
|  | if (tick_nohz_full_enabled()) | 
|  | __tick_nohz_task_switch(); | 
|  | } | 
|  |  | 
|  | static inline void tick_nohz_user_enter_prepare(void) | 
|  | { | 
|  | if (tick_nohz_full_cpu(smp_processor_id())) | 
|  | rcu_nocb_flush_deferred_wakeup(); | 
|  | } | 
|  |  | 
|  | #endif |