mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-24 03:07:29 +07:00
234b3840d7
This is the last instance of cpu_all_mask usage in the core framework. Replace it with cpu_possible_mask like all other instances in the clockevent drivers. This makes it possible to add a warning in the core clockevents_register_device on usage of cpu_all_mask from any clockevent drivers in the future. Signed-off-by: Sudeep Holla <sudeep.holla@arm.com> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Cc: linux-arm-kernel@lists.infradead.org Link: https://lkml.kernel.org/r/1531308264-24220-2-git-send-email-sudeep.holla@arm.com
113 lines
3.0 KiB
C
113 lines
3.0 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* linux/kernel/time/tick-broadcast-hrtimer.c
|
|
* This file emulates a local clock event device
|
|
* via a pseudo clock device.
|
|
*/
|
|
#include <linux/cpu.h>
|
|
#include <linux/err.h>
|
|
#include <linux/hrtimer.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/percpu.h>
|
|
#include <linux/profile.h>
|
|
#include <linux/clockchips.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/module.h>
|
|
|
|
#include "tick-internal.h"
|
|
|
|
static struct hrtimer bctimer;
|
|
|
|
static int bc_shutdown(struct clock_event_device *evt)
|
|
{
|
|
/*
|
|
* Note, we cannot cancel the timer here as we might
|
|
* run into the following live lock scenario:
|
|
*
|
|
* cpu 0 cpu1
|
|
* lock(broadcast_lock);
|
|
* hrtimer_interrupt()
|
|
* bc_handler()
|
|
* tick_handle_oneshot_broadcast();
|
|
* lock(broadcast_lock);
|
|
* hrtimer_cancel()
|
|
* wait_for_callback()
|
|
*/
|
|
hrtimer_try_to_cancel(&bctimer);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* This is called from the guts of the broadcast code when the cpu
|
|
* which is about to enter idle has the earliest broadcast timer event.
|
|
*/
|
|
static int bc_set_next(ktime_t expires, struct clock_event_device *bc)
|
|
{
|
|
int bc_moved;
|
|
/*
|
|
* We try to cancel the timer first. If the callback is on
|
|
* flight on some other cpu then we let it handle it. If we
|
|
* were able to cancel the timer nothing can rearm it as we
|
|
* own broadcast_lock.
|
|
*
|
|
* However we can also be called from the event handler of
|
|
* ce_broadcast_hrtimer itself when it expires. We cannot
|
|
* restart the timer because we are in the callback, but we
|
|
* can set the expiry time and let the callback return
|
|
* HRTIMER_RESTART.
|
|
*
|
|
* Since we are in the idle loop at this point and because
|
|
* hrtimer_{start/cancel} functions call into tracing,
|
|
* calls to these functions must be bound within RCU_NONIDLE.
|
|
*/
|
|
RCU_NONIDLE({
|
|
bc_moved = hrtimer_try_to_cancel(&bctimer) >= 0;
|
|
if (bc_moved)
|
|
hrtimer_start(&bctimer, expires,
|
|
HRTIMER_MODE_ABS_PINNED);});
|
|
if (bc_moved) {
|
|
/* Bind the "device" to the cpu */
|
|
bc->bound_on = smp_processor_id();
|
|
} else if (bc->bound_on == smp_processor_id()) {
|
|
hrtimer_set_expires(&bctimer, expires);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static struct clock_event_device ce_broadcast_hrtimer = {
|
|
.name = "bc_hrtimer",
|
|
.set_state_shutdown = bc_shutdown,
|
|
.set_next_ktime = bc_set_next,
|
|
.features = CLOCK_EVT_FEAT_ONESHOT |
|
|
CLOCK_EVT_FEAT_KTIME |
|
|
CLOCK_EVT_FEAT_HRTIMER,
|
|
.rating = 0,
|
|
.bound_on = -1,
|
|
.min_delta_ns = 1,
|
|
.max_delta_ns = KTIME_MAX,
|
|
.min_delta_ticks = 1,
|
|
.max_delta_ticks = ULONG_MAX,
|
|
.mult = 1,
|
|
.shift = 0,
|
|
.cpumask = cpu_possible_mask,
|
|
};
|
|
|
|
static enum hrtimer_restart bc_handler(struct hrtimer *t)
|
|
{
|
|
ce_broadcast_hrtimer.event_handler(&ce_broadcast_hrtimer);
|
|
|
|
if (clockevent_state_oneshot(&ce_broadcast_hrtimer))
|
|
if (ce_broadcast_hrtimer.next_event != KTIME_MAX)
|
|
return HRTIMER_RESTART;
|
|
|
|
return HRTIMER_NORESTART;
|
|
}
|
|
|
|
void tick_setup_hrtimer_broadcast(void)
|
|
{
|
|
hrtimer_init(&bctimer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
|
|
bctimer.function = bc_handler;
|
|
clockevents_register_device(&ce_broadcast_hrtimer);
|
|
}
|