mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-11 19:26:43 +07:00
60ba770227
Now that the pluging tracers use macros to create the structures and automate the exporting of their formats to the format files, they also automatically get a filter file. This patch adds the code to implement the filter logic in the trace recordings. Signed-off-by: Steven Rostedt <rostedt@goodmis.org>
186 lines
4.2 KiB
C
186 lines
4.2 KiB
C
/*
|
|
* ring buffer based initcalls tracer
|
|
*
|
|
* Copyright (C) 2008 Frederic Weisbecker <fweisbec@gmail.com>
|
|
*
|
|
*/
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/debugfs.h>
|
|
#include <linux/ftrace.h>
|
|
#include <linux/kallsyms.h>
|
|
#include <linux/time.h>
|
|
|
|
#include "trace.h"
|
|
#include "trace_output.h"
|
|
|
|
static struct trace_array *boot_trace;
|
|
static bool pre_initcalls_finished;
|
|
|
|
/* Tells the boot tracer that the pre_smp_initcalls are finished.
|
|
* So we are ready .
|
|
* It doesn't enable sched events tracing however.
|
|
* You have to call enable_boot_trace to do so.
|
|
*/
|
|
void start_boot_trace(void)
|
|
{
|
|
pre_initcalls_finished = true;
|
|
}
|
|
|
|
void enable_boot_trace(void)
|
|
{
|
|
if (boot_trace && pre_initcalls_finished)
|
|
tracing_start_sched_switch_record();
|
|
}
|
|
|
|
void disable_boot_trace(void)
|
|
{
|
|
if (boot_trace && pre_initcalls_finished)
|
|
tracing_stop_sched_switch_record();
|
|
}
|
|
|
|
static int boot_trace_init(struct trace_array *tr)
|
|
{
|
|
boot_trace = tr;
|
|
|
|
if (!tr)
|
|
return 0;
|
|
|
|
tracing_reset_online_cpus(tr);
|
|
|
|
tracing_sched_switch_assign_trace(tr);
|
|
return 0;
|
|
}
|
|
|
|
static enum print_line_t
|
|
initcall_call_print_line(struct trace_iterator *iter)
|
|
{
|
|
struct trace_entry *entry = iter->ent;
|
|
struct trace_seq *s = &iter->seq;
|
|
struct trace_boot_call *field;
|
|
struct boot_trace_call *call;
|
|
u64 ts;
|
|
unsigned long nsec_rem;
|
|
int ret;
|
|
|
|
trace_assign_type(field, entry);
|
|
call = &field->boot_call;
|
|
ts = iter->ts;
|
|
nsec_rem = do_div(ts, NSEC_PER_SEC);
|
|
|
|
ret = trace_seq_printf(s, "[%5ld.%09ld] calling %s @ %i\n",
|
|
(unsigned long)ts, nsec_rem, call->func, call->caller);
|
|
|
|
if (!ret)
|
|
return TRACE_TYPE_PARTIAL_LINE;
|
|
else
|
|
return TRACE_TYPE_HANDLED;
|
|
}
|
|
|
|
static enum print_line_t
|
|
initcall_ret_print_line(struct trace_iterator *iter)
|
|
{
|
|
struct trace_entry *entry = iter->ent;
|
|
struct trace_seq *s = &iter->seq;
|
|
struct trace_boot_ret *field;
|
|
struct boot_trace_ret *init_ret;
|
|
u64 ts;
|
|
unsigned long nsec_rem;
|
|
int ret;
|
|
|
|
trace_assign_type(field, entry);
|
|
init_ret = &field->boot_ret;
|
|
ts = iter->ts;
|
|
nsec_rem = do_div(ts, NSEC_PER_SEC);
|
|
|
|
ret = trace_seq_printf(s, "[%5ld.%09ld] initcall %s "
|
|
"returned %d after %llu msecs\n",
|
|
(unsigned long) ts,
|
|
nsec_rem,
|
|
init_ret->func, init_ret->result, init_ret->duration);
|
|
|
|
if (!ret)
|
|
return TRACE_TYPE_PARTIAL_LINE;
|
|
else
|
|
return TRACE_TYPE_HANDLED;
|
|
}
|
|
|
|
static enum print_line_t initcall_print_line(struct trace_iterator *iter)
|
|
{
|
|
struct trace_entry *entry = iter->ent;
|
|
|
|
switch (entry->type) {
|
|
case TRACE_BOOT_CALL:
|
|
return initcall_call_print_line(iter);
|
|
case TRACE_BOOT_RET:
|
|
return initcall_ret_print_line(iter);
|
|
default:
|
|
return TRACE_TYPE_UNHANDLED;
|
|
}
|
|
}
|
|
|
|
struct tracer boot_tracer __read_mostly =
|
|
{
|
|
.name = "initcall",
|
|
.init = boot_trace_init,
|
|
.reset = tracing_reset_online_cpus,
|
|
.print_line = initcall_print_line,
|
|
};
|
|
|
|
void trace_boot_call(struct boot_trace_call *bt, initcall_t fn)
|
|
{
|
|
struct ftrace_event_call *call = &event_boot_call;
|
|
struct ring_buffer_event *event;
|
|
struct ring_buffer *buffer;
|
|
struct trace_boot_call *entry;
|
|
struct trace_array *tr = boot_trace;
|
|
|
|
if (!tr || !pre_initcalls_finished)
|
|
return;
|
|
|
|
/* Get its name now since this function could
|
|
* disappear because it is in the .init section.
|
|
*/
|
|
sprint_symbol(bt->func, (unsigned long)fn);
|
|
preempt_disable();
|
|
|
|
buffer = tr->buffer;
|
|
event = trace_buffer_lock_reserve(buffer, TRACE_BOOT_CALL,
|
|
sizeof(*entry), 0, 0);
|
|
if (!event)
|
|
goto out;
|
|
entry = ring_buffer_event_data(event);
|
|
entry->boot_call = *bt;
|
|
if (!filter_check_discard(call, entry, buffer, event))
|
|
trace_buffer_unlock_commit(buffer, event, 0, 0);
|
|
out:
|
|
preempt_enable();
|
|
}
|
|
|
|
void trace_boot_ret(struct boot_trace_ret *bt, initcall_t fn)
|
|
{
|
|
struct ftrace_event_call *call = &event_boot_ret;
|
|
struct ring_buffer_event *event;
|
|
struct ring_buffer *buffer;
|
|
struct trace_boot_ret *entry;
|
|
struct trace_array *tr = boot_trace;
|
|
|
|
if (!tr || !pre_initcalls_finished)
|
|
return;
|
|
|
|
sprint_symbol(bt->func, (unsigned long)fn);
|
|
preempt_disable();
|
|
|
|
buffer = tr->buffer;
|
|
event = trace_buffer_lock_reserve(buffer, TRACE_BOOT_RET,
|
|
sizeof(*entry), 0, 0);
|
|
if (!event)
|
|
goto out;
|
|
entry = ring_buffer_event_data(event);
|
|
entry->boot_ret = *bt;
|
|
if (!filter_check_discard(call, entry, buffer, event))
|
|
trace_buffer_unlock_commit(buffer, event, 0, 0);
|
|
out:
|
|
preempt_enable();
|
|
}
|