mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-30 10:36:39 +07:00
cee4393989
Commit e31d28b6ab
("trace: Eliminate cond_resched_rcu_qs() in favor
of cond_resched()") substituted cond_resched() for the earlier call
to cond_resched_rcu_qs(). However, the new-age cond_resched() does
not do anything to help RCU-tasks grace periods because (1) RCU-tasks
is only enabled when CONFIG_PREEMPT=y and (2) cond_resched() is a
complete no-op when preemption is enabled. This situation results
in hangs when running the trace benchmarks.
A number of potential fixes were discussed on LKML
(https://lkml.kernel.org/r/20180224151240.0d63a059@vmware.local.home),
including making cond_resched() not be a no-op; making cond_resched()
not be a no-op, but only when running tracing benchmarks; reverting
the aforementioned commit (which works because cond_resched_rcu_qs()
does provide an RCU-tasks quiescent state; and adding a call to the
scheduler/RCU rcu_note_voluntary_context_switch() function. All were
deemed unsatisfactory, either due to added cond_resched() overhead or
due to magic functions inviting cargo culting.
This commit renames cond_resched_rcu_qs() to cond_resched_tasks_rcu_qs(),
which provides a clear hint as to what this function is doing and
why and where it should be used, and then replaces the call to
cond_resched() with cond_resched_tasks_rcu_qs() in the trace benchmark's
benchmark_event_kthread() function.
Reported-by: Steven Rostedt <rostedt@goodmis.org>
Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Tested-by: Nicholas Piggin <npiggin@gmail.com>
230 lines
5.2 KiB
C
230 lines
5.2 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <linux/delay.h>
|
|
#include <linux/module.h>
|
|
#include <linux/kthread.h>
|
|
#include <linux/trace_clock.h>
|
|
|
|
#define CREATE_TRACE_POINTS
|
|
#include "trace_benchmark.h"
|
|
|
|
static struct task_struct *bm_event_thread;
|
|
|
|
static char bm_str[BENCHMARK_EVENT_STRLEN] = "START";
|
|
|
|
static u64 bm_total;
|
|
static u64 bm_totalsq;
|
|
static u64 bm_last;
|
|
static u64 bm_max;
|
|
static u64 bm_min;
|
|
static u64 bm_first;
|
|
static u64 bm_cnt;
|
|
static u64 bm_stddev;
|
|
static unsigned int bm_avg;
|
|
static unsigned int bm_std;
|
|
|
|
static bool ok_to_run;
|
|
|
|
/*
|
|
* This gets called in a loop recording the time it took to write
|
|
* the tracepoint. What it writes is the time statistics of the last
|
|
* tracepoint write. As there is nothing to write the first time
|
|
* it simply writes "START". As the first write is cold cache and
|
|
* the rest is hot, we save off that time in bm_first and it is
|
|
* reported as "first", which is shown in the second write to the
|
|
* tracepoint. The "first" field is writen within the statics from
|
|
* then on but never changes.
|
|
*/
|
|
static void trace_do_benchmark(void)
|
|
{
|
|
u64 start;
|
|
u64 stop;
|
|
u64 delta;
|
|
u64 stddev;
|
|
u64 seed;
|
|
u64 last_seed;
|
|
unsigned int avg;
|
|
unsigned int std = 0;
|
|
|
|
/* Only run if the tracepoint is actually active */
|
|
if (!trace_benchmark_event_enabled() || !tracing_is_on())
|
|
return;
|
|
|
|
local_irq_disable();
|
|
start = trace_clock_local();
|
|
trace_benchmark_event(bm_str);
|
|
stop = trace_clock_local();
|
|
local_irq_enable();
|
|
|
|
bm_cnt++;
|
|
|
|
delta = stop - start;
|
|
|
|
/*
|
|
* The first read is cold cached, keep it separate from the
|
|
* other calculations.
|
|
*/
|
|
if (bm_cnt == 1) {
|
|
bm_first = delta;
|
|
scnprintf(bm_str, BENCHMARK_EVENT_STRLEN,
|
|
"first=%llu [COLD CACHED]", bm_first);
|
|
return;
|
|
}
|
|
|
|
bm_last = delta;
|
|
|
|
if (delta > bm_max)
|
|
bm_max = delta;
|
|
if (!bm_min || delta < bm_min)
|
|
bm_min = delta;
|
|
|
|
/*
|
|
* When bm_cnt is greater than UINT_MAX, it breaks the statistics
|
|
* accounting. Freeze the statistics when that happens.
|
|
* We should have enough data for the avg and stddev anyway.
|
|
*/
|
|
if (bm_cnt > UINT_MAX) {
|
|
scnprintf(bm_str, BENCHMARK_EVENT_STRLEN,
|
|
"last=%llu first=%llu max=%llu min=%llu ** avg=%u std=%d std^2=%lld",
|
|
bm_last, bm_first, bm_max, bm_min, bm_avg, bm_std, bm_stddev);
|
|
return;
|
|
}
|
|
|
|
bm_total += delta;
|
|
bm_totalsq += delta * delta;
|
|
|
|
|
|
if (bm_cnt > 1) {
|
|
/*
|
|
* Apply Welford's method to calculate standard deviation:
|
|
* s^2 = 1 / (n * (n-1)) * (n * \Sum (x_i)^2 - (\Sum x_i)^2)
|
|
*/
|
|
stddev = (u64)bm_cnt * bm_totalsq - bm_total * bm_total;
|
|
do_div(stddev, (u32)bm_cnt);
|
|
do_div(stddev, (u32)bm_cnt - 1);
|
|
} else
|
|
stddev = 0;
|
|
|
|
delta = bm_total;
|
|
do_div(delta, bm_cnt);
|
|
avg = delta;
|
|
|
|
if (stddev > 0) {
|
|
int i = 0;
|
|
/*
|
|
* stddev is the square of standard deviation but
|
|
* we want the actualy number. Use the average
|
|
* as our seed to find the std.
|
|
*
|
|
* The next try is:
|
|
* x = (x + N/x) / 2
|
|
*
|
|
* Where N is the squared number to find the square
|
|
* root of.
|
|
*/
|
|
seed = avg;
|
|
do {
|
|
last_seed = seed;
|
|
seed = stddev;
|
|
if (!last_seed)
|
|
break;
|
|
do_div(seed, last_seed);
|
|
seed += last_seed;
|
|
do_div(seed, 2);
|
|
} while (i++ < 10 && last_seed != seed);
|
|
|
|
std = seed;
|
|
}
|
|
|
|
scnprintf(bm_str, BENCHMARK_EVENT_STRLEN,
|
|
"last=%llu first=%llu max=%llu min=%llu avg=%u std=%d std^2=%lld",
|
|
bm_last, bm_first, bm_max, bm_min, avg, std, stddev);
|
|
|
|
bm_std = std;
|
|
bm_avg = avg;
|
|
bm_stddev = stddev;
|
|
}
|
|
|
|
static int benchmark_event_kthread(void *arg)
|
|
{
|
|
/* sleep a bit to make sure the tracepoint gets activated */
|
|
msleep(100);
|
|
|
|
while (!kthread_should_stop()) {
|
|
|
|
trace_do_benchmark();
|
|
|
|
/*
|
|
* We don't go to sleep, but let others run as well.
|
|
* This is bascially a "yield()" to let any task that
|
|
* wants to run, schedule in, but if the CPU is idle,
|
|
* we'll keep burning cycles.
|
|
*
|
|
* Note the tasks_rcu_qs() version of cond_resched() will
|
|
* notify synchronize_rcu_tasks() that this thread has
|
|
* passed a quiescent state for rcu_tasks. Otherwise
|
|
* this thread will never voluntarily schedule which would
|
|
* block synchronize_rcu_tasks() indefinitely.
|
|
*/
|
|
cond_resched_tasks_rcu_qs();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* When the benchmark tracepoint is enabled, it calls this
|
|
* function and the thread that calls the tracepoint is created.
|
|
*/
|
|
int trace_benchmark_reg(void)
|
|
{
|
|
if (!ok_to_run) {
|
|
pr_warning("trace benchmark cannot be started via kernel command line\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
bm_event_thread = kthread_run(benchmark_event_kthread,
|
|
NULL, "event_benchmark");
|
|
if (IS_ERR(bm_event_thread)) {
|
|
pr_warning("trace benchmark failed to create kernel thread\n");
|
|
return PTR_ERR(bm_event_thread);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* When the benchmark tracepoint is disabled, it calls this
|
|
* function and the thread that calls the tracepoint is deleted
|
|
* and all the numbers are reset.
|
|
*/
|
|
void trace_benchmark_unreg(void)
|
|
{
|
|
if (!bm_event_thread)
|
|
return;
|
|
|
|
kthread_stop(bm_event_thread);
|
|
bm_event_thread = NULL;
|
|
|
|
strcpy(bm_str, "START");
|
|
bm_total = 0;
|
|
bm_totalsq = 0;
|
|
bm_last = 0;
|
|
bm_max = 0;
|
|
bm_min = 0;
|
|
bm_cnt = 0;
|
|
/* These don't need to be reset but reset them anyway */
|
|
bm_first = 0;
|
|
bm_std = 0;
|
|
bm_avg = 0;
|
|
bm_stddev = 0;
|
|
}
|
|
|
|
static __init int ok_to_run_trace_benchmark(void)
|
|
{
|
|
ok_to_run = true;
|
|
|
|
return 0;
|
|
}
|
|
|
|
early_initcall(ok_to_run_trace_benchmark);
|