mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-16 15:56:47 +07:00
a575813bfe
Reported by syzkaller: BUG: unable to handle kernel paging request at ffffffffc07f6a2e IP: report_bug+0x94/0x120 PGD 348e12067 P4D 348e12067 PUD 348e14067 PMD 3cbd84067 PTE 80000003f7e87161 Oops: 0003 [#1] SMP CPU: 2 PID: 7091 Comm: kvm_load_guest_ Tainted: G OE 4.11.0+ #8 task: ffff92fdfb525400 task.stack: ffffbda6c3d04000 RIP: 0010:report_bug+0x94/0x120 RSP: 0018:ffffbda6c3d07b20 EFLAGS: 00010202 do_trap+0x156/0x170 do_error_trap+0xa3/0x170 ? kvm_load_guest_fpu.part.175+0x12a/0x170 [kvm] ? mark_held_locks+0x79/0xa0 ? retint_kernel+0x10/0x10 ? trace_hardirqs_off_thunk+0x1a/0x1c do_invalid_op+0x20/0x30 invalid_op+0x1e/0x30 RIP: 0010:kvm_load_guest_fpu.part.175+0x12a/0x170 [kvm] ? kvm_load_guest_fpu.part.175+0x1c/0x170 [kvm] kvm_arch_vcpu_ioctl_run+0xed6/0x1b70 [kvm] kvm_vcpu_ioctl+0x384/0x780 [kvm] ? kvm_vcpu_ioctl+0x384/0x780 [kvm] ? sched_clock+0x13/0x20 ? __do_page_fault+0x2a0/0x550 do_vfs_ioctl+0xa4/0x700 ? up_read+0x1f/0x40 ? __do_page_fault+0x2a0/0x550 SyS_ioctl+0x79/0x90 entry_SYSCALL_64_fastpath+0x23/0xc2 SDM mentioned that "The MXCSR has several reserved bits, and attempting to write a 1 to any of these bits will cause a general-protection exception(#GP) to be generated". The syzkaller forks' testcase overrides xsave area w/ random values and steps on the reserved bits of MXCSR register. The damaged MXCSR register values of guest will be restored to SSEx MXCSR register before vmentry. This patch fixes it by catching userspace override MXCSR register reserved bits w/ random values and bails out immediately. Reported-by: Andrey Konovalov <andreyknvl@google.com> Reviewed-by: Paolo Bonzini <pbonzini@redhat.com> Cc: Paolo Bonzini <pbonzini@redhat.com> Cc: Radim Krčmář <rkrcmar@redhat.com> Cc: stable@vger.kernel.org Signed-off-by: Wanpeng Li <wanpeng.li@hotmail.com> Signed-off-by: Radim Krčmář <rkrcmar@redhat.com>
293 lines
7.2 KiB
C
293 lines
7.2 KiB
C
/*
|
|
* x86 FPU boot time init code:
|
|
*/
|
|
#include <asm/fpu/internal.h>
|
|
#include <asm/tlbflush.h>
|
|
#include <asm/setup.h>
|
|
#include <asm/cmdline.h>
|
|
|
|
#include <linux/sched.h>
|
|
#include <linux/sched/task.h>
|
|
#include <linux/init.h>
|
|
|
|
/*
|
|
* Initialize the registers found in all CPUs, CR0 and CR4:
|
|
*/
|
|
static void fpu__init_cpu_generic(void)
|
|
{
|
|
unsigned long cr0;
|
|
unsigned long cr4_mask = 0;
|
|
|
|
if (boot_cpu_has(X86_FEATURE_FXSR))
|
|
cr4_mask |= X86_CR4_OSFXSR;
|
|
if (boot_cpu_has(X86_FEATURE_XMM))
|
|
cr4_mask |= X86_CR4_OSXMMEXCPT;
|
|
if (cr4_mask)
|
|
cr4_set_bits(cr4_mask);
|
|
|
|
cr0 = read_cr0();
|
|
cr0 &= ~(X86_CR0_TS|X86_CR0_EM); /* clear TS and EM */
|
|
if (!boot_cpu_has(X86_FEATURE_FPU))
|
|
cr0 |= X86_CR0_EM;
|
|
write_cr0(cr0);
|
|
|
|
/* Flush out any pending x87 state: */
|
|
#ifdef CONFIG_MATH_EMULATION
|
|
if (!boot_cpu_has(X86_FEATURE_FPU))
|
|
fpstate_init_soft(¤t->thread.fpu.state.soft);
|
|
else
|
|
#endif
|
|
asm volatile ("fninit");
|
|
}
|
|
|
|
/*
|
|
* Enable all supported FPU features. Called when a CPU is brought online:
|
|
*/
|
|
void fpu__init_cpu(void)
|
|
{
|
|
fpu__init_cpu_generic();
|
|
fpu__init_cpu_xstate();
|
|
}
|
|
|
|
static bool fpu__probe_without_cpuid(void)
|
|
{
|
|
unsigned long cr0;
|
|
u16 fsw, fcw;
|
|
|
|
fsw = fcw = 0xffff;
|
|
|
|
cr0 = read_cr0();
|
|
cr0 &= ~(X86_CR0_TS | X86_CR0_EM);
|
|
write_cr0(cr0);
|
|
|
|
asm volatile("fninit ; fnstsw %0 ; fnstcw %1" : "+m" (fsw), "+m" (fcw));
|
|
|
|
pr_info("x86/fpu: Probing for FPU: FSW=0x%04hx FCW=0x%04hx\n", fsw, fcw);
|
|
|
|
return fsw == 0 && (fcw & 0x103f) == 0x003f;
|
|
}
|
|
|
|
static void fpu__init_system_early_generic(struct cpuinfo_x86 *c)
|
|
{
|
|
if (!boot_cpu_has(X86_FEATURE_CPUID) &&
|
|
!test_bit(X86_FEATURE_FPU, (unsigned long *)cpu_caps_cleared)) {
|
|
if (fpu__probe_without_cpuid())
|
|
setup_force_cpu_cap(X86_FEATURE_FPU);
|
|
else
|
|
setup_clear_cpu_cap(X86_FEATURE_FPU);
|
|
}
|
|
|
|
#ifndef CONFIG_MATH_EMULATION
|
|
if (!test_cpu_cap(&boot_cpu_data, X86_FEATURE_FPU)) {
|
|
pr_emerg("x86/fpu: Giving up, no FPU found and no math emulation present\n");
|
|
for (;;)
|
|
asm volatile("hlt");
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* Boot time FPU feature detection code:
|
|
*/
|
|
unsigned int mxcsr_feature_mask __read_mostly = 0xffffffffu;
|
|
EXPORT_SYMBOL_GPL(mxcsr_feature_mask);
|
|
|
|
static void __init fpu__init_system_mxcsr(void)
|
|
{
|
|
unsigned int mask = 0;
|
|
|
|
if (boot_cpu_has(X86_FEATURE_FXSR)) {
|
|
/* Static because GCC does not get 16-byte stack alignment right: */
|
|
static struct fxregs_state fxregs __initdata;
|
|
|
|
asm volatile("fxsave %0" : "+m" (fxregs));
|
|
|
|
mask = fxregs.mxcsr_mask;
|
|
|
|
/*
|
|
* If zero then use the default features mask,
|
|
* which has all features set, except the
|
|
* denormals-are-zero feature bit:
|
|
*/
|
|
if (mask == 0)
|
|
mask = 0x0000ffbf;
|
|
}
|
|
mxcsr_feature_mask &= mask;
|
|
}
|
|
|
|
/*
|
|
* Once per bootup FPU initialization sequences that will run on most x86 CPUs:
|
|
*/
|
|
static void __init fpu__init_system_generic(void)
|
|
{
|
|
/*
|
|
* Set up the legacy init FPU context. (xstate init might overwrite this
|
|
* with a more modern format, if the CPU supports it.)
|
|
*/
|
|
fpstate_init(&init_fpstate);
|
|
|
|
fpu__init_system_mxcsr();
|
|
}
|
|
|
|
/*
|
|
* Size of the FPU context state. All tasks in the system use the
|
|
* same context size, regardless of what portion they use.
|
|
* This is inherent to the XSAVE architecture which puts all state
|
|
* components into a single, continuous memory block:
|
|
*/
|
|
unsigned int fpu_kernel_xstate_size;
|
|
EXPORT_SYMBOL_GPL(fpu_kernel_xstate_size);
|
|
|
|
/* Get alignment of the TYPE. */
|
|
#define TYPE_ALIGN(TYPE) offsetof(struct { char x; TYPE test; }, test)
|
|
|
|
/*
|
|
* Enforce that 'MEMBER' is the last field of 'TYPE'.
|
|
*
|
|
* Align the computed size with alignment of the TYPE,
|
|
* because that's how C aligns structs.
|
|
*/
|
|
#define CHECK_MEMBER_AT_END_OF(TYPE, MEMBER) \
|
|
BUILD_BUG_ON(sizeof(TYPE) != ALIGN(offsetofend(TYPE, MEMBER), \
|
|
TYPE_ALIGN(TYPE)))
|
|
|
|
/*
|
|
* We append the 'struct fpu' to the task_struct:
|
|
*/
|
|
static void __init fpu__init_task_struct_size(void)
|
|
{
|
|
int task_size = sizeof(struct task_struct);
|
|
|
|
/*
|
|
* Subtract off the static size of the register state.
|
|
* It potentially has a bunch of padding.
|
|
*/
|
|
task_size -= sizeof(((struct task_struct *)0)->thread.fpu.state);
|
|
|
|
/*
|
|
* Add back the dynamically-calculated register state
|
|
* size.
|
|
*/
|
|
task_size += fpu_kernel_xstate_size;
|
|
|
|
/*
|
|
* We dynamically size 'struct fpu', so we require that
|
|
* it be at the end of 'thread_struct' and that
|
|
* 'thread_struct' be at the end of 'task_struct'. If
|
|
* you hit a compile error here, check the structure to
|
|
* see if something got added to the end.
|
|
*/
|
|
CHECK_MEMBER_AT_END_OF(struct fpu, state);
|
|
CHECK_MEMBER_AT_END_OF(struct thread_struct, fpu);
|
|
CHECK_MEMBER_AT_END_OF(struct task_struct, thread);
|
|
|
|
arch_task_struct_size = task_size;
|
|
}
|
|
|
|
/*
|
|
* Set up the user and kernel xstate sizes based on the legacy FPU context size.
|
|
*
|
|
* We set this up first, and later it will be overwritten by
|
|
* fpu__init_system_xstate() if the CPU knows about xstates.
|
|
*/
|
|
static void __init fpu__init_system_xstate_size_legacy(void)
|
|
{
|
|
static int on_boot_cpu __initdata = 1;
|
|
|
|
WARN_ON_FPU(!on_boot_cpu);
|
|
on_boot_cpu = 0;
|
|
|
|
/*
|
|
* Note that xstate sizes might be overwritten later during
|
|
* fpu__init_system_xstate().
|
|
*/
|
|
|
|
if (!boot_cpu_has(X86_FEATURE_FPU)) {
|
|
/*
|
|
* Disable xsave as we do not support it if i387
|
|
* emulation is enabled.
|
|
*/
|
|
setup_clear_cpu_cap(X86_FEATURE_XSAVE);
|
|
setup_clear_cpu_cap(X86_FEATURE_XSAVEOPT);
|
|
fpu_kernel_xstate_size = sizeof(struct swregs_state);
|
|
} else {
|
|
if (boot_cpu_has(X86_FEATURE_FXSR))
|
|
fpu_kernel_xstate_size =
|
|
sizeof(struct fxregs_state);
|
|
else
|
|
fpu_kernel_xstate_size =
|
|
sizeof(struct fregs_state);
|
|
}
|
|
|
|
fpu_user_xstate_size = fpu_kernel_xstate_size;
|
|
}
|
|
|
|
/*
|
|
* Find supported xfeatures based on cpu features and command-line input.
|
|
* This must be called after fpu__init_parse_early_param() is called and
|
|
* xfeatures_mask is enumerated.
|
|
*/
|
|
u64 __init fpu__get_supported_xfeatures_mask(void)
|
|
{
|
|
return XCNTXT_MASK;
|
|
}
|
|
|
|
/* Legacy code to initialize eager fpu mode. */
|
|
static void __init fpu__init_system_ctx_switch(void)
|
|
{
|
|
static bool on_boot_cpu __initdata = 1;
|
|
|
|
WARN_ON_FPU(!on_boot_cpu);
|
|
on_boot_cpu = 0;
|
|
|
|
WARN_ON_FPU(current->thread.fpu.fpstate_active);
|
|
}
|
|
|
|
/*
|
|
* We parse fpu parameters early because fpu__init_system() is executed
|
|
* before parse_early_param().
|
|
*/
|
|
static void __init fpu__init_parse_early_param(void)
|
|
{
|
|
if (cmdline_find_option_bool(boot_command_line, "no387"))
|
|
setup_clear_cpu_cap(X86_FEATURE_FPU);
|
|
|
|
if (cmdline_find_option_bool(boot_command_line, "nofxsr")) {
|
|
setup_clear_cpu_cap(X86_FEATURE_FXSR);
|
|
setup_clear_cpu_cap(X86_FEATURE_FXSR_OPT);
|
|
setup_clear_cpu_cap(X86_FEATURE_XMM);
|
|
}
|
|
|
|
if (cmdline_find_option_bool(boot_command_line, "noxsave"))
|
|
fpu__xstate_clear_all_cpu_caps();
|
|
|
|
if (cmdline_find_option_bool(boot_command_line, "noxsaveopt"))
|
|
setup_clear_cpu_cap(X86_FEATURE_XSAVEOPT);
|
|
|
|
if (cmdline_find_option_bool(boot_command_line, "noxsaves"))
|
|
setup_clear_cpu_cap(X86_FEATURE_XSAVES);
|
|
}
|
|
|
|
/*
|
|
* Called on the boot CPU once per system bootup, to set up the initial
|
|
* FPU state that is later cloned into all processes:
|
|
*/
|
|
void __init fpu__init_system(struct cpuinfo_x86 *c)
|
|
{
|
|
fpu__init_parse_early_param();
|
|
fpu__init_system_early_generic(c);
|
|
|
|
/*
|
|
* The FPU has to be operational for some of the
|
|
* later FPU init activities:
|
|
*/
|
|
fpu__init_cpu();
|
|
|
|
fpu__init_system_generic();
|
|
fpu__init_system_xstate_size_legacy();
|
|
fpu__init_system_xstate();
|
|
fpu__init_task_struct_size();
|
|
|
|
fpu__init_system_ctx_switch();
|
|
}
|