mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-26 00:25:33 +07:00
e9666d10a5
Currently, CONFIG_JUMP_LABEL just means "I _want_ to use jump label". The jump label is controlled by HAVE_JUMP_LABEL, which is defined like this: #if defined(CC_HAVE_ASM_GOTO) && defined(CONFIG_JUMP_LABEL) # define HAVE_JUMP_LABEL #endif We can improve this by testing 'asm goto' support in Kconfig, then make JUMP_LABEL depend on CC_HAS_ASM_GOTO. Ugly #ifdef HAVE_JUMP_LABEL will go away, and CONFIG_JUMP_LABEL will match to the real kernel capability. Signed-off-by: Masahiro Yamada <yamada.masahiro@socionext.com> Acked-by: Michael Ellerman <mpe@ellerman.id.au> (powerpc) Tested-by: Sedat Dilek <sedat.dilek@gmail.com>
89 lines
1.6 KiB
C
89 lines
1.6 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <linux/percpu.h>
|
|
#include <linux/jump_label.h>
|
|
#include <asm/trace.h>
|
|
#include <asm/asm-prototypes.h>
|
|
|
|
#ifdef CONFIG_JUMP_LABEL
|
|
struct static_key opal_tracepoint_key = STATIC_KEY_INIT;
|
|
|
|
int opal_tracepoint_regfunc(void)
|
|
{
|
|
static_key_slow_inc(&opal_tracepoint_key);
|
|
return 0;
|
|
}
|
|
|
|
void opal_tracepoint_unregfunc(void)
|
|
{
|
|
static_key_slow_dec(&opal_tracepoint_key);
|
|
}
|
|
#else
|
|
/*
|
|
* We optimise OPAL calls by placing opal_tracepoint_refcount
|
|
* directly in the TOC so we can check if the opal tracepoints are
|
|
* enabled via a single load.
|
|
*/
|
|
|
|
/* NB: reg/unreg are called while guarded with the tracepoints_mutex */
|
|
extern long opal_tracepoint_refcount;
|
|
|
|
int opal_tracepoint_regfunc(void)
|
|
{
|
|
opal_tracepoint_refcount++;
|
|
return 0;
|
|
}
|
|
|
|
void opal_tracepoint_unregfunc(void)
|
|
{
|
|
opal_tracepoint_refcount--;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* Since the tracing code might execute OPAL calls we need to guard against
|
|
* recursion.
|
|
*/
|
|
static DEFINE_PER_CPU(unsigned int, opal_trace_depth);
|
|
|
|
void __trace_opal_entry(unsigned long opcode, unsigned long *args)
|
|
{
|
|
unsigned long flags;
|
|
unsigned int *depth;
|
|
|
|
local_irq_save(flags);
|
|
|
|
depth = this_cpu_ptr(&opal_trace_depth);
|
|
|
|
if (*depth)
|
|
goto out;
|
|
|
|
(*depth)++;
|
|
preempt_disable();
|
|
trace_opal_entry(opcode, args);
|
|
(*depth)--;
|
|
|
|
out:
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
void __trace_opal_exit(long opcode, unsigned long retval)
|
|
{
|
|
unsigned long flags;
|
|
unsigned int *depth;
|
|
|
|
local_irq_save(flags);
|
|
|
|
depth = this_cpu_ptr(&opal_trace_depth);
|
|
|
|
if (*depth)
|
|
goto out;
|
|
|
|
(*depth)++;
|
|
trace_opal_exit(opcode, retval);
|
|
preempt_enable();
|
|
(*depth)--;
|
|
|
|
out:
|
|
local_irq_restore(flags);
|
|
}
|