mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 19:25:05 +07:00
699785f5d8
The circular buffers are now validated with selftests. The next interrupt index algorithm which is the hardest part to validate needs extra coverage. Add a selftest which uses the intervals stored in the arrays and insert all the values except the last one. The next event computation must return the same value as the last element which was not inserted. Signed-off-by: Daniel Lezcano <daniel.lezcano@linaro.org> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Cc: andriy.shevchenko@linux.intel.com Link: https://lkml.kernel.org/r/20190527205521.12091-9-daniel.lezcano@linaro.org
952 lines
25 KiB
C
952 lines
25 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
// Copyright (C) 2016, Linaro Ltd - Daniel Lezcano <daniel.lezcano@linaro.org>
|
|
#define pr_fmt(fmt) "irq_timings: " fmt
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/percpu.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/static_key.h>
|
|
#include <linux/init.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/idr.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/math64.h>
|
|
#include <linux/log2.h>
|
|
|
|
#include <trace/events/irq.h>
|
|
|
|
#include "internals.h"
|
|
|
|
DEFINE_STATIC_KEY_FALSE(irq_timing_enabled);
|
|
|
|
DEFINE_PER_CPU(struct irq_timings, irq_timings);
|
|
|
|
static DEFINE_IDR(irqt_stats);
|
|
|
|
void irq_timings_enable(void)
|
|
{
|
|
static_branch_enable(&irq_timing_enabled);
|
|
}
|
|
|
|
void irq_timings_disable(void)
|
|
{
|
|
static_branch_disable(&irq_timing_enabled);
|
|
}
|
|
|
|
/*
|
|
* The main goal of this algorithm is to predict the next interrupt
|
|
* occurrence on the current CPU.
|
|
*
|
|
* Currently, the interrupt timings are stored in a circular array
|
|
* buffer every time there is an interrupt, as a tuple: the interrupt
|
|
* number and the associated timestamp when the event occurred <irq,
|
|
* timestamp>.
|
|
*
|
|
* For every interrupt occurring in a short period of time, we can
|
|
* measure the elapsed time between the occurrences for the same
|
|
* interrupt and we end up with a suite of intervals. The experience
|
|
* showed the interrupts are often coming following a periodic
|
|
* pattern.
|
|
*
|
|
* The objective of the algorithm is to find out this periodic pattern
|
|
* in a fastest way and use its period to predict the next irq event.
|
|
*
|
|
* When the next interrupt event is requested, we are in the situation
|
|
* where the interrupts are disabled and the circular buffer
|
|
* containing the timings is filled with the events which happened
|
|
* after the previous next-interrupt-event request.
|
|
*
|
|
* At this point, we read the circular buffer and we fill the irq
|
|
* related statistics structure. After this step, the circular array
|
|
* containing the timings is empty because all the values are
|
|
* dispatched in their corresponding buffers.
|
|
*
|
|
* Now for each interrupt, we can predict the next event by using the
|
|
* suffix array, log interval and exponential moving average
|
|
*
|
|
* 1. Suffix array
|
|
*
|
|
* Suffix array is an array of all the suffixes of a string. It is
|
|
* widely used as a data structure for compression, text search, ...
|
|
* For instance for the word 'banana', the suffixes will be: 'banana'
|
|
* 'anana' 'nana' 'ana' 'na' 'a'
|
|
*
|
|
* Usually, the suffix array is sorted but for our purpose it is
|
|
* not necessary and won't provide any improvement in the context of
|
|
* the solved problem where we clearly define the boundaries of the
|
|
* search by a max period and min period.
|
|
*
|
|
* The suffix array will build a suite of intervals of different
|
|
* length and will look for the repetition of each suite. If the suite
|
|
* is repeating then we have the period because it is the length of
|
|
* the suite whatever its position in the buffer.
|
|
*
|
|
* 2. Log interval
|
|
*
|
|
* We saw the irq timings allow to compute the interval of the
|
|
* occurrences for a specific interrupt. We can reasonibly assume the
|
|
* longer is the interval, the higher is the error for the next event
|
|
* and we can consider storing those interval values into an array
|
|
* where each slot in the array correspond to an interval at the power
|
|
* of 2 of the index. For example, index 12 will contain values
|
|
* between 2^11 and 2^12.
|
|
*
|
|
* At the end we have an array of values where at each index defines a
|
|
* [2^index - 1, 2 ^ index] interval values allowing to store a large
|
|
* number of values inside a small array.
|
|
*
|
|
* For example, if we have the value 1123, then we store it at
|
|
* ilog2(1123) = 10 index value.
|
|
*
|
|
* Storing those value at the specific index is done by computing an
|
|
* exponential moving average for this specific slot. For instance,
|
|
* for values 1800, 1123, 1453, ... fall under the same slot (10) and
|
|
* the exponential moving average is computed every time a new value
|
|
* is stored at this slot.
|
|
*
|
|
* 3. Exponential Moving Average
|
|
*
|
|
* The EMA is largely used to track a signal for stocks or as a low
|
|
* pass filter. The magic of the formula, is it is very simple and the
|
|
* reactivity of the average can be tuned with the factors called
|
|
* alpha.
|
|
*
|
|
* The higher the alphas are, the faster the average respond to the
|
|
* signal change. In our case, if a slot in the array is a big
|
|
* interval, we can have numbers with a big difference between
|
|
* them. The impact of those differences in the average computation
|
|
* can be tuned by changing the alpha value.
|
|
*
|
|
*
|
|
* -- The algorithm --
|
|
*
|
|
* We saw the different processing above, now let's see how they are
|
|
* used together.
|
|
*
|
|
* For each interrupt:
|
|
* For each interval:
|
|
* Compute the index = ilog2(interval)
|
|
* Compute a new_ema(buffer[index], interval)
|
|
* Store the index in a circular buffer
|
|
*
|
|
* Compute the suffix array of the indexes
|
|
*
|
|
* For each suffix:
|
|
* If the suffix is reverse-found 3 times
|
|
* Return suffix
|
|
*
|
|
* Return Not found
|
|
*
|
|
* However we can not have endless suffix array to be build, it won't
|
|
* make sense and it will add an extra overhead, so we can restrict
|
|
* this to a maximum suffix length of 5 and a minimum suffix length of
|
|
* 2. The experience showed 5 is the majority of the maximum pattern
|
|
* period found for different devices.
|
|
*
|
|
* The result is a pattern finding less than 1us for an interrupt.
|
|
*
|
|
* Example based on real values:
|
|
*
|
|
* Example 1 : MMC write/read interrupt interval:
|
|
*
|
|
* 223947, 1240, 1384, 1386, 1386,
|
|
* 217416, 1236, 1384, 1386, 1387,
|
|
* 214719, 1241, 1386, 1387, 1384,
|
|
* 213696, 1234, 1384, 1386, 1388,
|
|
* 219904, 1240, 1385, 1389, 1385,
|
|
* 212240, 1240, 1386, 1386, 1386,
|
|
* 214415, 1236, 1384, 1386, 1387,
|
|
* 214276, 1234, 1384, 1388, ?
|
|
*
|
|
* For each element, apply ilog2(value)
|
|
*
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, ?
|
|
*
|
|
* Max period of 5, we take the last (max_period * 3) 15 elements as
|
|
* we can be confident if the pattern repeats itself three times it is
|
|
* a repeating pattern.
|
|
*
|
|
* 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, 8,
|
|
* 15, 8, 8, 8, ?
|
|
*
|
|
* Suffixes are:
|
|
*
|
|
* 1) 8, 15, 8, 8, 8 <- max period
|
|
* 2) 8, 15, 8, 8
|
|
* 3) 8, 15, 8
|
|
* 4) 8, 15 <- min period
|
|
*
|
|
* From there we search the repeating pattern for each suffix.
|
|
*
|
|
* buffer: 8, 15, 8, 8, 8, 8, 15, 8, 8, 8, 8, 15, 8, 8, 8
|
|
* | | | | | | | | | | | | | | |
|
|
* 8, 15, 8, 8, 8 | | | | | | | | | |
|
|
* 8, 15, 8, 8, 8 | | | | |
|
|
* 8, 15, 8, 8, 8
|
|
*
|
|
* When moving the suffix, we found exactly 3 matches.
|
|
*
|
|
* The first suffix with period 5 is repeating.
|
|
*
|
|
* The next event is (3 * max_period) % suffix_period
|
|
*
|
|
* In this example, the result 0, so the next event is suffix[0] => 8
|
|
*
|
|
* However, 8 is the index in the array of exponential moving average
|
|
* which was calculated on the fly when storing the values, so the
|
|
* interval is ema[8] = 1366
|
|
*
|
|
*
|
|
* Example 2:
|
|
*
|
|
* 4, 3, 5, 100,
|
|
* 3, 3, 5, 117,
|
|
* 4, 4, 5, 112,
|
|
* 4, 3, 4, 110,
|
|
* 3, 5, 3, 117,
|
|
* 4, 4, 5, 112,
|
|
* 4, 3, 4, 110,
|
|
* 3, 4, 5, 112,
|
|
* 4, 3, 4, 110
|
|
*
|
|
* ilog2
|
|
*
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4
|
|
*
|
|
* Max period 5:
|
|
* 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4,
|
|
* 0, 0, 0, 4
|
|
*
|
|
* Suffixes:
|
|
*
|
|
* 1) 0, 0, 4, 0, 0
|
|
* 2) 0, 0, 4, 0
|
|
* 3) 0, 0, 4
|
|
* 4) 0, 0
|
|
*
|
|
* buffer: 0, 0, 4, 0, 0, 0, 4, 0, 0, 0, 4, 0, 0, 0, 4
|
|
* | | | | | | X
|
|
* 0, 0, 4, 0, 0, | X
|
|
* 0, 0
|
|
*
|
|
* buffer: 0, 0, 4, 0, 0, 0, 4, 0, 0, 0, 4, 0, 0, 0, 4
|
|
* | | | | | | | | | | | | | | |
|
|
* 0, 0, 4, 0, | | | | | | | | | | |
|
|
* 0, 0, 4, 0, | | | | | | |
|
|
* 0, 0, 4, 0, | | |
|
|
* 0 0 4
|
|
*
|
|
* Pattern is found 3 times, the remaining is 1 which results from
|
|
* (max_period * 3) % suffix_period. This value is the index in the
|
|
* suffix arrays. The suffix array for a period 4 has the value 4
|
|
* at index 1.
|
|
*/
|
|
#define EMA_ALPHA_VAL 64
|
|
#define EMA_ALPHA_SHIFT 7
|
|
|
|
#define PREDICTION_PERIOD_MIN 3
|
|
#define PREDICTION_PERIOD_MAX 5
|
|
#define PREDICTION_FACTOR 4
|
|
#define PREDICTION_MAX 10 /* 2 ^ PREDICTION_MAX useconds */
|
|
#define PREDICTION_BUFFER_SIZE 16 /* slots for EMAs, hardly more than 16 */
|
|
|
|
/*
|
|
* Number of elements in the circular buffer: If it happens it was
|
|
* flushed before, then the number of elements could be smaller than
|
|
* IRQ_TIMINGS_SIZE, so the count is used, otherwise the array size is
|
|
* used as we wrapped. The index begins from zero when we did not
|
|
* wrap. That could be done in a nicer way with the proper circular
|
|
* array structure type but with the cost of extra computation in the
|
|
* interrupt handler hot path. We choose efficiency.
|
|
*/
|
|
#define for_each_irqts(i, irqts) \
|
|
for (i = irqts->count < IRQ_TIMINGS_SIZE ? \
|
|
0 : irqts->count & IRQ_TIMINGS_MASK, \
|
|
irqts->count = min(IRQ_TIMINGS_SIZE, \
|
|
irqts->count); \
|
|
irqts->count > 0; irqts->count--, \
|
|
i = (i + 1) & IRQ_TIMINGS_MASK)
|
|
|
|
struct irqt_stat {
|
|
u64 last_ts;
|
|
u64 ema_time[PREDICTION_BUFFER_SIZE];
|
|
int timings[IRQ_TIMINGS_SIZE];
|
|
int circ_timings[IRQ_TIMINGS_SIZE];
|
|
int count;
|
|
};
|
|
|
|
/*
|
|
* Exponential moving average computation
|
|
*/
|
|
static u64 irq_timings_ema_new(u64 value, u64 ema_old)
|
|
{
|
|
s64 diff;
|
|
|
|
if (unlikely(!ema_old))
|
|
return value;
|
|
|
|
diff = (value - ema_old) * EMA_ALPHA_VAL;
|
|
/*
|
|
* We can use a s64 type variable to be added with the u64
|
|
* ema_old variable as this one will never have its topmost
|
|
* bit set, it will be always smaller than 2^63 nanosec
|
|
* interrupt interval (292 years).
|
|
*/
|
|
return ema_old + (diff >> EMA_ALPHA_SHIFT);
|
|
}
|
|
|
|
static int irq_timings_next_event_index(int *buffer, size_t len, int period_max)
|
|
{
|
|
int period;
|
|
|
|
/*
|
|
* Move the beginning pointer to the end minus the max period x 3.
|
|
* We are at the point we can begin searching the pattern
|
|
*/
|
|
buffer = &buffer[len - (period_max * 3)];
|
|
|
|
/* Adjust the length to the maximum allowed period x 3 */
|
|
len = period_max * 3;
|
|
|
|
/*
|
|
* The buffer contains the suite of intervals, in a ilog2
|
|
* basis, we are looking for a repetition. We point the
|
|
* beginning of the search three times the length of the
|
|
* period beginning at the end of the buffer. We do that for
|
|
* each suffix.
|
|
*/
|
|
for (period = period_max; period >= PREDICTION_PERIOD_MIN; period--) {
|
|
|
|
/*
|
|
* The first comparison always succeed because the
|
|
* suffix is deduced from the first n-period bytes of
|
|
* the buffer and we compare the initial suffix with
|
|
* itself, so we can skip the first iteration.
|
|
*/
|
|
int idx = period;
|
|
size_t size = period;
|
|
|
|
/*
|
|
* We look if the suite with period 'i' repeat
|
|
* itself. If it is truncated at the end, as it
|
|
* repeats we can use the period to find out the next
|
|
* element with the modulo.
|
|
*/
|
|
while (!memcmp(buffer, &buffer[idx], size * sizeof(int))) {
|
|
|
|
/*
|
|
* Move the index in a period basis
|
|
*/
|
|
idx += size;
|
|
|
|
/*
|
|
* If this condition is reached, all previous
|
|
* memcmp were successful, so the period is
|
|
* found.
|
|
*/
|
|
if (idx == len)
|
|
return buffer[len % period];
|
|
|
|
/*
|
|
* If the remaining elements to compare are
|
|
* smaller than the period, readjust the size
|
|
* of the comparison for the last iteration.
|
|
*/
|
|
if (len - idx < period)
|
|
size = len - idx;
|
|
}
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
static u64 __irq_timings_next_event(struct irqt_stat *irqs, int irq, u64 now)
|
|
{
|
|
int index, i, period_max, count, start, min = INT_MAX;
|
|
|
|
if ((now - irqs->last_ts) >= NSEC_PER_SEC) {
|
|
irqs->count = irqs->last_ts = 0;
|
|
return U64_MAX;
|
|
}
|
|
|
|
/*
|
|
* As we want to find three times the repetition, we need a
|
|
* number of intervals greater or equal to three times the
|
|
* maximum period, otherwise we truncate the max period.
|
|
*/
|
|
period_max = irqs->count > (3 * PREDICTION_PERIOD_MAX) ?
|
|
PREDICTION_PERIOD_MAX : irqs->count / 3;
|
|
|
|
/*
|
|
* If we don't have enough irq timings for this prediction,
|
|
* just bail out.
|
|
*/
|
|
if (period_max <= PREDICTION_PERIOD_MIN)
|
|
return U64_MAX;
|
|
|
|
/*
|
|
* 'count' will depends if the circular buffer wrapped or not
|
|
*/
|
|
count = irqs->count < IRQ_TIMINGS_SIZE ?
|
|
irqs->count : IRQ_TIMINGS_SIZE;
|
|
|
|
start = irqs->count < IRQ_TIMINGS_SIZE ?
|
|
0 : (irqs->count & IRQ_TIMINGS_MASK);
|
|
|
|
/*
|
|
* Copy the content of the circular buffer into another buffer
|
|
* in order to linearize the buffer instead of dealing with
|
|
* wrapping indexes and shifted array which will be prone to
|
|
* error and extremelly difficult to debug.
|
|
*/
|
|
for (i = 0; i < count; i++) {
|
|
int index = (start + i) & IRQ_TIMINGS_MASK;
|
|
|
|
irqs->timings[i] = irqs->circ_timings[index];
|
|
min = min_t(int, irqs->timings[i], min);
|
|
}
|
|
|
|
index = irq_timings_next_event_index(irqs->timings, count, period_max);
|
|
if (index < 0)
|
|
return irqs->last_ts + irqs->ema_time[min];
|
|
|
|
return irqs->last_ts + irqs->ema_time[index];
|
|
}
|
|
|
|
static __always_inline int irq_timings_interval_index(u64 interval)
|
|
{
|
|
/*
|
|
* The PREDICTION_FACTOR increase the interval size for the
|
|
* array of exponential average.
|
|
*/
|
|
u64 interval_us = (interval >> 10) / PREDICTION_FACTOR;
|
|
|
|
return likely(interval_us) ? ilog2(interval_us) : 0;
|
|
}
|
|
|
|
static __always_inline void __irq_timings_store(int irq, struct irqt_stat *irqs,
|
|
u64 interval)
|
|
{
|
|
int index;
|
|
|
|
/*
|
|
* Get the index in the ema table for this interrupt.
|
|
*/
|
|
index = irq_timings_interval_index(interval);
|
|
|
|
/*
|
|
* Store the index as an element of the pattern in another
|
|
* circular array.
|
|
*/
|
|
irqs->circ_timings[irqs->count & IRQ_TIMINGS_MASK] = index;
|
|
|
|
irqs->ema_time[index] = irq_timings_ema_new(interval,
|
|
irqs->ema_time[index]);
|
|
|
|
irqs->count++;
|
|
}
|
|
|
|
static inline void irq_timings_store(int irq, struct irqt_stat *irqs, u64 ts)
|
|
{
|
|
u64 old_ts = irqs->last_ts;
|
|
u64 interval;
|
|
|
|
/*
|
|
* The timestamps are absolute time values, we need to compute
|
|
* the timing interval between two interrupts.
|
|
*/
|
|
irqs->last_ts = ts;
|
|
|
|
/*
|
|
* The interval type is u64 in order to deal with the same
|
|
* type in our computation, that prevent mindfuck issues with
|
|
* overflow, sign and division.
|
|
*/
|
|
interval = ts - old_ts;
|
|
|
|
/*
|
|
* The interrupt triggered more than one second apart, that
|
|
* ends the sequence as predictible for our purpose. In this
|
|
* case, assume we have the beginning of a sequence and the
|
|
* timestamp is the first value. As it is impossible to
|
|
* predict anything at this point, return.
|
|
*
|
|
* Note the first timestamp of the sequence will always fall
|
|
* in this test because the old_ts is zero. That is what we
|
|
* want as we need another timestamp to compute an interval.
|
|
*/
|
|
if (interval >= NSEC_PER_SEC) {
|
|
irqs->count = 0;
|
|
return;
|
|
}
|
|
|
|
__irq_timings_store(irq, irqs, interval);
|
|
}
|
|
|
|
/**
|
|
* irq_timings_next_event - Return when the next event is supposed to arrive
|
|
*
|
|
* During the last busy cycle, the number of interrupts is incremented
|
|
* and stored in the irq_timings structure. This information is
|
|
* necessary to:
|
|
*
|
|
* - know if the index in the table wrapped up:
|
|
*
|
|
* If more than the array size interrupts happened during the
|
|
* last busy/idle cycle, the index wrapped up and we have to
|
|
* begin with the next element in the array which is the last one
|
|
* in the sequence, otherwise it is a the index 0.
|
|
*
|
|
* - have an indication of the interrupts activity on this CPU
|
|
* (eg. irq/sec)
|
|
*
|
|
* The values are 'consumed' after inserting in the statistical model,
|
|
* thus the count is reinitialized.
|
|
*
|
|
* The array of values **must** be browsed in the time direction, the
|
|
* timestamp must increase between an element and the next one.
|
|
*
|
|
* Returns a nanosec time based estimation of the earliest interrupt,
|
|
* U64_MAX otherwise.
|
|
*/
|
|
u64 irq_timings_next_event(u64 now)
|
|
{
|
|
struct irq_timings *irqts = this_cpu_ptr(&irq_timings);
|
|
struct irqt_stat *irqs;
|
|
struct irqt_stat __percpu *s;
|
|
u64 ts, next_evt = U64_MAX;
|
|
int i, irq = 0;
|
|
|
|
/*
|
|
* This function must be called with the local irq disabled in
|
|
* order to prevent the timings circular buffer to be updated
|
|
* while we are reading it.
|
|
*/
|
|
lockdep_assert_irqs_disabled();
|
|
|
|
if (!irqts->count)
|
|
return next_evt;
|
|
|
|
/*
|
|
* Number of elements in the circular buffer: If it happens it
|
|
* was flushed before, then the number of elements could be
|
|
* smaller than IRQ_TIMINGS_SIZE, so the count is used,
|
|
* otherwise the array size is used as we wrapped. The index
|
|
* begins from zero when we did not wrap. That could be done
|
|
* in a nicer way with the proper circular array structure
|
|
* type but with the cost of extra computation in the
|
|
* interrupt handler hot path. We choose efficiency.
|
|
*
|
|
* Inject measured irq/timestamp to the pattern prediction
|
|
* model while decrementing the counter because we consume the
|
|
* data from our circular buffer.
|
|
*/
|
|
for_each_irqts(i, irqts) {
|
|
irq = irq_timing_decode(irqts->values[i], &ts);
|
|
s = idr_find(&irqt_stats, irq);
|
|
if (s)
|
|
irq_timings_store(irq, this_cpu_ptr(s), ts);
|
|
}
|
|
|
|
/*
|
|
* Look in the list of interrupts' statistics, the earliest
|
|
* next event.
|
|
*/
|
|
idr_for_each_entry(&irqt_stats, s, i) {
|
|
|
|
irqs = this_cpu_ptr(s);
|
|
|
|
ts = __irq_timings_next_event(irqs, i, now);
|
|
if (ts <= now)
|
|
return now;
|
|
|
|
if (ts < next_evt)
|
|
next_evt = ts;
|
|
}
|
|
|
|
return next_evt;
|
|
}
|
|
|
|
void irq_timings_free(int irq)
|
|
{
|
|
struct irqt_stat __percpu *s;
|
|
|
|
s = idr_find(&irqt_stats, irq);
|
|
if (s) {
|
|
free_percpu(s);
|
|
idr_remove(&irqt_stats, irq);
|
|
}
|
|
}
|
|
|
|
int irq_timings_alloc(int irq)
|
|
{
|
|
struct irqt_stat __percpu *s;
|
|
int id;
|
|
|
|
/*
|
|
* Some platforms can have the same private interrupt per cpu,
|
|
* so this function may be be called several times with the
|
|
* same interrupt number. Just bail out in case the per cpu
|
|
* stat structure is already allocated.
|
|
*/
|
|
s = idr_find(&irqt_stats, irq);
|
|
if (s)
|
|
return 0;
|
|
|
|
s = alloc_percpu(*s);
|
|
if (!s)
|
|
return -ENOMEM;
|
|
|
|
idr_preload(GFP_KERNEL);
|
|
id = idr_alloc(&irqt_stats, s, irq, irq + 1, GFP_NOWAIT);
|
|
idr_preload_end();
|
|
|
|
if (id < 0) {
|
|
free_percpu(s);
|
|
return id;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_TEST_IRQ_TIMINGS
|
|
struct timings_intervals {
|
|
u64 *intervals;
|
|
size_t count;
|
|
};
|
|
|
|
/*
|
|
* Intervals are given in nanosecond base
|
|
*/
|
|
static u64 intervals0[] __initdata = {
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000, 500000,
|
|
10000, 50000, 200000,
|
|
};
|
|
|
|
static u64 intervals1[] __initdata = {
|
|
223947000, 1240000, 1384000, 1386000, 1386000,
|
|
217416000, 1236000, 1384000, 1386000, 1387000,
|
|
214719000, 1241000, 1386000, 1387000, 1384000,
|
|
213696000, 1234000, 1384000, 1386000, 1388000,
|
|
219904000, 1240000, 1385000, 1389000, 1385000,
|
|
212240000, 1240000, 1386000, 1386000, 1386000,
|
|
214415000, 1236000, 1384000, 1386000, 1387000,
|
|
214276000, 1234000,
|
|
};
|
|
|
|
static u64 intervals2[] __initdata = {
|
|
4000, 3000, 5000, 100000,
|
|
3000, 3000, 5000, 117000,
|
|
4000, 4000, 5000, 112000,
|
|
4000, 3000, 4000, 110000,
|
|
3000, 5000, 3000, 117000,
|
|
4000, 4000, 5000, 112000,
|
|
4000, 3000, 4000, 110000,
|
|
3000, 4000, 5000, 112000,
|
|
4000,
|
|
};
|
|
|
|
static u64 intervals3[] __initdata = {
|
|
1385000, 212240000, 1240000,
|
|
1386000, 214415000, 1236000,
|
|
1384000, 214276000, 1234000,
|
|
1386000, 214415000, 1236000,
|
|
1385000, 212240000, 1240000,
|
|
1386000, 214415000, 1236000,
|
|
1384000, 214276000, 1234000,
|
|
1386000, 214415000, 1236000,
|
|
1385000, 212240000, 1240000,
|
|
};
|
|
|
|
static u64 intervals4[] __initdata = {
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000, 50000, 10000, 50000,
|
|
10000,
|
|
};
|
|
|
|
static struct timings_intervals tis[] __initdata = {
|
|
{ intervals0, ARRAY_SIZE(intervals0) },
|
|
{ intervals1, ARRAY_SIZE(intervals1) },
|
|
{ intervals2, ARRAY_SIZE(intervals2) },
|
|
{ intervals3, ARRAY_SIZE(intervals3) },
|
|
{ intervals4, ARRAY_SIZE(intervals4) },
|
|
};
|
|
|
|
static int __init irq_timings_test_next_index(struct timings_intervals *ti)
|
|
{
|
|
int _buffer[IRQ_TIMINGS_SIZE];
|
|
int buffer[IRQ_TIMINGS_SIZE];
|
|
int index, start, i, count, period_max;
|
|
|
|
count = ti->count - 1;
|
|
|
|
period_max = count > (3 * PREDICTION_PERIOD_MAX) ?
|
|
PREDICTION_PERIOD_MAX : count / 3;
|
|
|
|
/*
|
|
* Inject all values except the last one which will be used
|
|
* to compare with the next index result.
|
|
*/
|
|
pr_debug("index suite: ");
|
|
|
|
for (i = 0; i < count; i++) {
|
|
index = irq_timings_interval_index(ti->intervals[i]);
|
|
_buffer[i & IRQ_TIMINGS_MASK] = index;
|
|
pr_cont("%d ", index);
|
|
}
|
|
|
|
start = count < IRQ_TIMINGS_SIZE ? 0 :
|
|
count & IRQ_TIMINGS_MASK;
|
|
|
|
count = min_t(int, count, IRQ_TIMINGS_SIZE);
|
|
|
|
for (i = 0; i < count; i++) {
|
|
int index = (start + i) & IRQ_TIMINGS_MASK;
|
|
buffer[i] = _buffer[index];
|
|
}
|
|
|
|
index = irq_timings_next_event_index(buffer, count, period_max);
|
|
i = irq_timings_interval_index(ti->intervals[ti->count - 1]);
|
|
|
|
if (index != i) {
|
|
pr_err("Expected (%d) and computed (%d) next indexes differ\n",
|
|
i, index);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init irq_timings_next_index_selftest(void)
|
|
{
|
|
int i, ret;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(tis); i++) {
|
|
|
|
pr_info("---> Injecting intervals number #%d (count=%zd)\n",
|
|
i, tis[i].count);
|
|
|
|
ret = irq_timings_test_next_index(&tis[i]);
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __init irq_timings_test_irqs(struct timings_intervals *ti)
|
|
{
|
|
struct irqt_stat __percpu *s;
|
|
struct irqt_stat *irqs;
|
|
int i, index, ret, irq = 0xACE5;
|
|
|
|
ret = irq_timings_alloc(irq);
|
|
if (ret) {
|
|
pr_err("Failed to allocate irq timings\n");
|
|
return ret;
|
|
}
|
|
|
|
s = idr_find(&irqt_stats, irq);
|
|
if (!s) {
|
|
ret = -EIDRM;
|
|
goto out;
|
|
}
|
|
|
|
irqs = this_cpu_ptr(s);
|
|
|
|
for (i = 0; i < ti->count; i++) {
|
|
|
|
index = irq_timings_interval_index(ti->intervals[i]);
|
|
pr_debug("%d: interval=%llu ema_index=%d\n",
|
|
i, ti->intervals[i], index);
|
|
|
|
__irq_timings_store(irq, irqs, ti->intervals[i]);
|
|
if (irqs->circ_timings[i & IRQ_TIMINGS_MASK] != index) {
|
|
pr_err("Failed to store in the circular buffer\n");
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if (irqs->count != ti->count) {
|
|
pr_err("Count differs\n");
|
|
goto out;
|
|
}
|
|
|
|
ret = 0;
|
|
out:
|
|
irq_timings_free(irq);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __init irq_timings_irqs_selftest(void)
|
|
{
|
|
int i, ret;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(tis); i++) {
|
|
pr_info("---> Injecting intervals number #%d (count=%zd)\n",
|
|
i, tis[i].count);
|
|
ret = irq_timings_test_irqs(&tis[i]);
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __init irq_timings_test_irqts(struct irq_timings *irqts,
|
|
unsigned count)
|
|
{
|
|
int start = count >= IRQ_TIMINGS_SIZE ? count - IRQ_TIMINGS_SIZE : 0;
|
|
int i, irq, oirq = 0xBEEF;
|
|
u64 ots = 0xDEAD, ts;
|
|
|
|
/*
|
|
* Fill the circular buffer by using the dedicated function.
|
|
*/
|
|
for (i = 0; i < count; i++) {
|
|
pr_debug("%d: index=%d, ts=%llX irq=%X\n",
|
|
i, i & IRQ_TIMINGS_MASK, ots + i, oirq + i);
|
|
|
|
irq_timings_push(ots + i, oirq + i);
|
|
}
|
|
|
|
/*
|
|
* Compute the first elements values after the index wrapped
|
|
* up or not.
|
|
*/
|
|
ots += start;
|
|
oirq += start;
|
|
|
|
/*
|
|
* Test the circular buffer count is correct.
|
|
*/
|
|
pr_debug("---> Checking timings array count (%d) is right\n", count);
|
|
if (WARN_ON(irqts->count != count))
|
|
return -EINVAL;
|
|
|
|
/*
|
|
* Test the macro allowing to browse all the irqts.
|
|
*/
|
|
pr_debug("---> Checking the for_each_irqts() macro\n");
|
|
for_each_irqts(i, irqts) {
|
|
|
|
irq = irq_timing_decode(irqts->values[i], &ts);
|
|
|
|
pr_debug("index=%d, ts=%llX / %llX, irq=%X / %X\n",
|
|
i, ts, ots, irq, oirq);
|
|
|
|
if (WARN_ON(ts != ots || irq != oirq))
|
|
return -EINVAL;
|
|
|
|
ots++; oirq++;
|
|
}
|
|
|
|
/*
|
|
* The circular buffer should have be flushed when browsed
|
|
* with for_each_irqts
|
|
*/
|
|
pr_debug("---> Checking timings array is empty after browsing it\n");
|
|
if (WARN_ON(irqts->count))
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init irq_timings_irqts_selftest(void)
|
|
{
|
|
struct irq_timings *irqts = this_cpu_ptr(&irq_timings);
|
|
int i, ret;
|
|
|
|
/*
|
|
* Test the circular buffer with different number of
|
|
* elements. The purpose is to test at the limits (empty, half
|
|
* full, full, wrapped with the cursor at the boundaries,
|
|
* wrapped several times, etc ...
|
|
*/
|
|
int count[] = { 0,
|
|
IRQ_TIMINGS_SIZE >> 1,
|
|
IRQ_TIMINGS_SIZE,
|
|
IRQ_TIMINGS_SIZE + (IRQ_TIMINGS_SIZE >> 1),
|
|
2 * IRQ_TIMINGS_SIZE,
|
|
(2 * IRQ_TIMINGS_SIZE) + 3,
|
|
};
|
|
|
|
for (i = 0; i < ARRAY_SIZE(count); i++) {
|
|
|
|
pr_info("---> Checking the timings with %d/%d values\n",
|
|
count[i], IRQ_TIMINGS_SIZE);
|
|
|
|
ret = irq_timings_test_irqts(irqts, count[i]);
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __init irq_timings_selftest(void)
|
|
{
|
|
int ret;
|
|
|
|
pr_info("------------------- selftest start -----------------\n");
|
|
|
|
/*
|
|
* At this point, we don't except any subsystem to use the irq
|
|
* timings but us, so it should not be enabled.
|
|
*/
|
|
if (static_branch_unlikely(&irq_timing_enabled)) {
|
|
pr_warn("irq timings already initialized, skipping selftest\n");
|
|
return 0;
|
|
}
|
|
|
|
ret = irq_timings_irqts_selftest();
|
|
if (ret)
|
|
goto out;
|
|
|
|
ret = irq_timings_irqs_selftest();
|
|
if (ret)
|
|
goto out;
|
|
|
|
ret = irq_timings_next_index_selftest();
|
|
out:
|
|
pr_info("---------- selftest end with %s -----------\n",
|
|
ret ? "failure" : "success");
|
|
|
|
return ret;
|
|
}
|
|
early_initcall(irq_timings_selftest);
|
|
#endif
|