mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-25 17:25:07 +07:00
33a58980ff
The generic VDSO implementation uses the Y2038 safe clock_gettime64() and
clock_getres_time64() syscalls as fallback for 32bit VDSO. This breaks
seccomp setups because these syscalls might be not (yet) allowed.
Implement the 32bit variants which use the legacy syscalls and select the
variant in the core library.
The 64bit time variants are not removed because they are required for the
time64 based vdso accessors.
Fixes: 00b26474c2
("lib/vdso: Provide generic VDSO implementation")
Reported-by: Sean Christopherson <sean.j.christopherson@intel.com>
Reported-by: Paul Bolle <pebolle@tiscali.nl>
Suggested-by: Andy Lutomirski <luto@kernel.org>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Tested-by: Vincenzo Frascino <vincenzo.frascino@arm.com>
Reviewed-by: Vincenzo Frascino <vincenzo.frascino@arm.com>
Link: https://lkml.kernel.org/r/20190728131648.971361611@linutronix.de
167 lines
4.0 KiB
C
167 lines
4.0 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Copyright (C) 2018 ARM Limited
|
|
*/
|
|
#ifndef __ASM_VDSO_GETTIMEOFDAY_H
|
|
#define __ASM_VDSO_GETTIMEOFDAY_H
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
#include <asm/unistd.h>
|
|
#include <uapi/linux/time.h>
|
|
|
|
#include <asm/vdso/compat_barrier.h>
|
|
|
|
#define __VDSO_USE_SYSCALL ULLONG_MAX
|
|
|
|
#define VDSO_HAS_CLOCK_GETRES 1
|
|
|
|
#define VDSO_HAS_32BIT_FALLBACK 1
|
|
|
|
static __always_inline
|
|
int gettimeofday_fallback(struct __kernel_old_timeval *_tv,
|
|
struct timezone *_tz)
|
|
{
|
|
register struct timezone *tz asm("r1") = _tz;
|
|
register struct __kernel_old_timeval *tv asm("r0") = _tv;
|
|
register long ret asm ("r0");
|
|
register long nr asm("r7") = __NR_compat_gettimeofday;
|
|
|
|
asm volatile(
|
|
" swi #0\n"
|
|
: "=r" (ret)
|
|
: "r" (tv), "r" (tz), "r" (nr)
|
|
: "memory");
|
|
|
|
return ret;
|
|
}
|
|
|
|
static __always_inline
|
|
long clock_gettime_fallback(clockid_t _clkid, struct __kernel_timespec *_ts)
|
|
{
|
|
register struct __kernel_timespec *ts asm("r1") = _ts;
|
|
register clockid_t clkid asm("r0") = _clkid;
|
|
register long ret asm ("r0");
|
|
register long nr asm("r7") = __NR_compat_clock_gettime64;
|
|
|
|
asm volatile(
|
|
" swi #0\n"
|
|
: "=r" (ret)
|
|
: "r" (clkid), "r" (ts), "r" (nr)
|
|
: "memory");
|
|
|
|
return ret;
|
|
}
|
|
|
|
static __always_inline
|
|
long clock_gettime32_fallback(clockid_t _clkid, struct old_timespec32 *_ts)
|
|
{
|
|
register struct old_timespec32 *ts asm("r1") = _ts;
|
|
register clockid_t clkid asm("r0") = _clkid;
|
|
register long ret asm ("r0");
|
|
register long nr asm("r7") = __NR_compat_clock_gettime;
|
|
|
|
asm volatile(
|
|
" swi #0\n"
|
|
: "=r" (ret)
|
|
: "r" (clkid), "r" (ts), "r" (nr)
|
|
: "memory");
|
|
|
|
return ret;
|
|
}
|
|
|
|
static __always_inline
|
|
int clock_getres_fallback(clockid_t _clkid, struct __kernel_timespec *_ts)
|
|
{
|
|
register struct __kernel_timespec *ts asm("r1") = _ts;
|
|
register clockid_t clkid asm("r0") = _clkid;
|
|
register long ret asm ("r0");
|
|
register long nr asm("r7") = __NR_compat_clock_getres_time64;
|
|
|
|
/* The checks below are required for ABI consistency with arm */
|
|
if ((_clkid >= MAX_CLOCKS) && (_ts == NULL))
|
|
return -EINVAL;
|
|
|
|
asm volatile(
|
|
" swi #0\n"
|
|
: "=r" (ret)
|
|
: "r" (clkid), "r" (ts), "r" (nr)
|
|
: "memory");
|
|
|
|
return ret;
|
|
}
|
|
|
|
static __always_inline
|
|
int clock_getres32_fallback(clockid_t _clkid, struct old_timespec32 *_ts)
|
|
{
|
|
register struct old_timespec32 *ts asm("r1") = _ts;
|
|
register clockid_t clkid asm("r0") = _clkid;
|
|
register long ret asm ("r0");
|
|
register long nr asm("r7") = __NR_compat_clock_getres;
|
|
|
|
/* The checks below are required for ABI consistency with arm */
|
|
if ((_clkid >= MAX_CLOCKS) && (_ts == NULL))
|
|
return -EINVAL;
|
|
|
|
asm volatile(
|
|
" swi #0\n"
|
|
: "=r" (ret)
|
|
: "r" (clkid), "r" (ts), "r" (nr)
|
|
: "memory");
|
|
|
|
return ret;
|
|
}
|
|
|
|
static __always_inline u64 __arch_get_hw_counter(s32 clock_mode)
|
|
{
|
|
u64 res;
|
|
|
|
/*
|
|
* clock_mode == 0 implies that vDSO are enabled otherwise
|
|
* fallback on syscall.
|
|
*/
|
|
if (clock_mode)
|
|
return __VDSO_USE_SYSCALL;
|
|
|
|
/*
|
|
* This isb() is required to prevent that the counter value
|
|
* is speculated.
|
|
*/
|
|
isb();
|
|
asm volatile("mrrc p15, 1, %Q0, %R0, c14" : "=r" (res));
|
|
/*
|
|
* This isb() is required to prevent that the seq lock is
|
|
* speculated.
|
|
*/
|
|
isb();
|
|
|
|
return res;
|
|
}
|
|
|
|
static __always_inline const struct vdso_data *__arch_get_vdso_data(void)
|
|
{
|
|
const struct vdso_data *ret;
|
|
|
|
/*
|
|
* This simply puts &_vdso_data into ret. The reason why we don't use
|
|
* `ret = _vdso_data` is that the compiler tends to optimise this in a
|
|
* very suboptimal way: instead of keeping &_vdso_data in a register,
|
|
* it goes through a relocation almost every time _vdso_data must be
|
|
* accessed (even in subfunctions). This is both time and space
|
|
* consuming: each relocation uses a word in the code section, and it
|
|
* has to be loaded at runtime.
|
|
*
|
|
* This trick hides the assignment from the compiler. Since it cannot
|
|
* track where the pointer comes from, it will only use one relocation
|
|
* where __arch_get_vdso_data() is called, and then keep the result in
|
|
* a register.
|
|
*/
|
|
asm volatile("mov %0, %1" : "=r"(ret) : "r"(_vdso_data));
|
|
|
|
return ret;
|
|
}
|
|
|
|
#endif /* !__ASSEMBLY__ */
|
|
|
|
#endif /* __ASM_VDSO_GETTIMEOFDAY_H */
|