mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 19:45:08 +07:00
a683f390b9
The wheel clock is stale when a CPU goes into a long idle sleep. This has the side effect that timers which are queued end up in the outer wheel levels. That results in coarser granularity. To solve this, we keep track of the idle state and forward the wheel clock whenever possible. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Cc: Arjan van de Ven <arjan@infradead.org> Cc: Chris Mason <clm@fb.com> Cc: Eric Dumazet <edumazet@google.com> Cc: Frederic Weisbecker <fweisbec@gmail.com> Cc: George Spelvin <linux@sciencehorizons.net> Cc: Josh Triplett <josh@joshtriplett.org> Cc: Len Brown <lenb@kernel.org> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Rik van Riel <riel@redhat.com> Cc: rt@linutronix.de Link: http://lkml.kernel.org/r/20160704094342.512039360@linutronix.de Signed-off-by: Ingo Molnar <mingo@kernel.org>
168 lines
7.0 KiB
C
168 lines
7.0 KiB
C
/*
|
|
* tick internal variable and functions used by low/high res code
|
|
*/
|
|
#include <linux/hrtimer.h>
|
|
#include <linux/tick.h>
|
|
|
|
#include "timekeeping.h"
|
|
#include "tick-sched.h"
|
|
|
|
#ifdef CONFIG_GENERIC_CLOCKEVENTS
|
|
|
|
# define TICK_DO_TIMER_NONE -1
|
|
# define TICK_DO_TIMER_BOOT -2
|
|
|
|
DECLARE_PER_CPU(struct tick_device, tick_cpu_device);
|
|
extern ktime_t tick_next_period;
|
|
extern ktime_t tick_period;
|
|
extern int tick_do_timer_cpu __read_mostly;
|
|
|
|
extern void tick_setup_periodic(struct clock_event_device *dev, int broadcast);
|
|
extern void tick_handle_periodic(struct clock_event_device *dev);
|
|
extern void tick_check_new_device(struct clock_event_device *dev);
|
|
extern void tick_shutdown(unsigned int cpu);
|
|
extern void tick_suspend(void);
|
|
extern void tick_resume(void);
|
|
extern bool tick_check_replacement(struct clock_event_device *curdev,
|
|
struct clock_event_device *newdev);
|
|
extern void tick_install_replacement(struct clock_event_device *dev);
|
|
extern int tick_is_oneshot_available(void);
|
|
extern struct tick_device *tick_get_device(int cpu);
|
|
|
|
extern int clockevents_tick_resume(struct clock_event_device *dev);
|
|
/* Check, if the device is functional or a dummy for broadcast */
|
|
static inline int tick_device_is_functional(struct clock_event_device *dev)
|
|
{
|
|
return !(dev->features & CLOCK_EVT_FEAT_DUMMY);
|
|
}
|
|
|
|
static inline enum clock_event_state clockevent_get_state(struct clock_event_device *dev)
|
|
{
|
|
return dev->state_use_accessors;
|
|
}
|
|
|
|
static inline void clockevent_set_state(struct clock_event_device *dev,
|
|
enum clock_event_state state)
|
|
{
|
|
dev->state_use_accessors = state;
|
|
}
|
|
|
|
extern void clockevents_shutdown(struct clock_event_device *dev);
|
|
extern void clockevents_exchange_device(struct clock_event_device *old,
|
|
struct clock_event_device *new);
|
|
extern void clockevents_switch_state(struct clock_event_device *dev,
|
|
enum clock_event_state state);
|
|
extern int clockevents_program_event(struct clock_event_device *dev,
|
|
ktime_t expires, bool force);
|
|
extern void clockevents_handle_noop(struct clock_event_device *dev);
|
|
extern int __clockevents_update_freq(struct clock_event_device *dev, u32 freq);
|
|
extern ssize_t sysfs_get_uname(const char *buf, char *dst, size_t cnt);
|
|
|
|
/* Broadcasting support */
|
|
# ifdef CONFIG_GENERIC_CLOCKEVENTS_BROADCAST
|
|
extern int tick_device_uses_broadcast(struct clock_event_device *dev, int cpu);
|
|
extern void tick_install_broadcast_device(struct clock_event_device *dev);
|
|
extern int tick_is_broadcast_device(struct clock_event_device *dev);
|
|
extern void tick_shutdown_broadcast(unsigned int cpu);
|
|
extern void tick_suspend_broadcast(void);
|
|
extern void tick_resume_broadcast(void);
|
|
extern bool tick_resume_check_broadcast(void);
|
|
extern void tick_broadcast_init(void);
|
|
extern void tick_set_periodic_handler(struct clock_event_device *dev, int broadcast);
|
|
extern int tick_broadcast_update_freq(struct clock_event_device *dev, u32 freq);
|
|
extern struct tick_device *tick_get_broadcast_device(void);
|
|
extern struct cpumask *tick_get_broadcast_mask(void);
|
|
# else /* !CONFIG_GENERIC_CLOCKEVENTS_BROADCAST: */
|
|
static inline void tick_install_broadcast_device(struct clock_event_device *dev) { }
|
|
static inline int tick_is_broadcast_device(struct clock_event_device *dev) { return 0; }
|
|
static inline int tick_device_uses_broadcast(struct clock_event_device *dev, int cpu) { return 0; }
|
|
static inline void tick_do_periodic_broadcast(struct clock_event_device *d) { }
|
|
static inline void tick_shutdown_broadcast(unsigned int cpu) { }
|
|
static inline void tick_suspend_broadcast(void) { }
|
|
static inline void tick_resume_broadcast(void) { }
|
|
static inline bool tick_resume_check_broadcast(void) { return false; }
|
|
static inline void tick_broadcast_init(void) { }
|
|
static inline int tick_broadcast_update_freq(struct clock_event_device *dev, u32 freq) { return -ENODEV; }
|
|
|
|
/* Set the periodic handler in non broadcast mode */
|
|
static inline void tick_set_periodic_handler(struct clock_event_device *dev, int broadcast)
|
|
{
|
|
dev->event_handler = tick_handle_periodic;
|
|
}
|
|
# endif /* !CONFIG_GENERIC_CLOCKEVENTS_BROADCAST */
|
|
|
|
#else /* !GENERIC_CLOCKEVENTS: */
|
|
static inline void tick_suspend(void) { }
|
|
static inline void tick_resume(void) { }
|
|
#endif /* !GENERIC_CLOCKEVENTS */
|
|
|
|
/* Oneshot related functions */
|
|
#ifdef CONFIG_TICK_ONESHOT
|
|
extern void tick_setup_oneshot(struct clock_event_device *newdev,
|
|
void (*handler)(struct clock_event_device *),
|
|
ktime_t nextevt);
|
|
extern int tick_program_event(ktime_t expires, int force);
|
|
extern void tick_oneshot_notify(void);
|
|
extern int tick_switch_to_oneshot(void (*handler)(struct clock_event_device *));
|
|
extern void tick_resume_oneshot(void);
|
|
static inline bool tick_oneshot_possible(void) { return true; }
|
|
extern int tick_oneshot_mode_active(void);
|
|
extern void tick_clock_notify(void);
|
|
extern int tick_check_oneshot_change(int allow_nohz);
|
|
extern int tick_init_highres(void);
|
|
#else /* !CONFIG_TICK_ONESHOT: */
|
|
static inline
|
|
void tick_setup_oneshot(struct clock_event_device *newdev,
|
|
void (*handler)(struct clock_event_device *),
|
|
ktime_t nextevt) { BUG(); }
|
|
static inline void tick_resume_oneshot(void) { BUG(); }
|
|
static inline int tick_program_event(ktime_t expires, int force) { return 0; }
|
|
static inline void tick_oneshot_notify(void) { }
|
|
static inline bool tick_oneshot_possible(void) { return false; }
|
|
static inline int tick_oneshot_mode_active(void) { return 0; }
|
|
static inline void tick_clock_notify(void) { }
|
|
static inline int tick_check_oneshot_change(int allow_nohz) { return 0; }
|
|
#endif /* !CONFIG_TICK_ONESHOT */
|
|
|
|
/* Functions related to oneshot broadcasting */
|
|
#if defined(CONFIG_GENERIC_CLOCKEVENTS_BROADCAST) && defined(CONFIG_TICK_ONESHOT)
|
|
extern void tick_broadcast_setup_oneshot(struct clock_event_device *bc);
|
|
extern void tick_broadcast_switch_to_oneshot(void);
|
|
extern void tick_shutdown_broadcast_oneshot(unsigned int cpu);
|
|
extern int tick_broadcast_oneshot_active(void);
|
|
extern void tick_check_oneshot_broadcast_this_cpu(void);
|
|
bool tick_broadcast_oneshot_available(void);
|
|
extern struct cpumask *tick_get_broadcast_oneshot_mask(void);
|
|
#else /* !(BROADCAST && ONESHOT): */
|
|
static inline void tick_broadcast_setup_oneshot(struct clock_event_device *bc) { BUG(); }
|
|
static inline void tick_broadcast_switch_to_oneshot(void) { }
|
|
static inline void tick_shutdown_broadcast_oneshot(unsigned int cpu) { }
|
|
static inline int tick_broadcast_oneshot_active(void) { return 0; }
|
|
static inline void tick_check_oneshot_broadcast_this_cpu(void) { }
|
|
static inline bool tick_broadcast_oneshot_available(void) { return tick_oneshot_possible(); }
|
|
#endif /* !(BROADCAST && ONESHOT) */
|
|
|
|
/* NO_HZ_FULL internal */
|
|
#ifdef CONFIG_NO_HZ_FULL
|
|
extern void tick_nohz_init(void);
|
|
# else
|
|
static inline void tick_nohz_init(void) { }
|
|
#endif
|
|
|
|
#ifdef CONFIG_NO_HZ_COMMON
|
|
extern unsigned long tick_nohz_active;
|
|
#else
|
|
#define tick_nohz_active (0)
|
|
#endif
|
|
|
|
#if defined(CONFIG_SMP) && defined(CONFIG_NO_HZ_COMMON)
|
|
extern void timers_update_migration(bool update_nohz);
|
|
#else
|
|
static inline void timers_update_migration(bool update_nohz) { }
|
|
#endif
|
|
|
|
DECLARE_PER_CPU(struct hrtimer_cpu_base, hrtimer_bases);
|
|
|
|
extern u64 get_next_timer_interrupt(unsigned long basej, u64 basem);
|
|
void timer_clear_idle(void);
|