mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-16 05:36:42 +07:00
fd89a13792
This patch fixes machine crashes which occur when heavily exercising the CPU hotplug codepaths on a 32-bit kernel. These crashes are caused by AMD Erratum 383 and result in a fatal machine check exception. Here's the scenario: 1. On 32-bit, the swapper_pg_dir page table is used as the initial page table for booting a secondary CPU. 2. To make this work, swapper_pg_dir needs a direct mapping of physical memory in it (the low mappings). By adding those low, large page (2M) mappings (PAE kernel), we create the necessary conditions for Erratum 383 to occur. 3. Other CPUs which do not participate in the off- and onlining game may use swapper_pg_dir while the low mappings are present (when leave_mm is called). For all steps below, the CPU referred to is a CPU that is using swapper_pg_dir, and not the CPU which is being onlined. 4. The presence of the low mappings in swapper_pg_dir can result in TLB entries for addresses below __PAGE_OFFSET to be established speculatively. These TLB entries are marked global and large. 5. When the CPU with such TLB entry switches to another page table, this TLB entry remains because it is global. 6. The process then generates an access to an address covered by the above TLB entry but there is a permission mismatch - the TLB entry covers a large global page not accessible to userspace. 7. Due to this permission mismatch a new 4kb, user TLB entry gets established. Further, Erratum 383 provides for a small window of time where both TLB entries are present. This results in an uncorrectable machine check exception signalling a TLB multimatch which panics the machine. There are two ways to fix this issue: 1. Always do a global TLB flush when a new cr3 is loaded and the old page table was swapper_pg_dir. I consider this a hack hard to understand and with performance implications 2. Do not use swapper_pg_dir to boot secondary CPUs like 64-bit does. This patch implements solution 2. It introduces a trampoline_pg_dir which has the same layout as swapper_pg_dir with low_mappings. This page table is used as the initial page table of the booting CPU. Later in the bringup process, it switches to swapper_pg_dir and does a global TLB flush. This fixes the crashes in our test cases. -v2: switch to swapper_pg_dir right after entering start_secondary() so that we are able to access percpu data which might not be mapped in the trampoline page table. Signed-off-by: Joerg Roedel <joerg.roedel@amd.com> LKML-Reference: <20100816123833.GB28147@aftab> Signed-off-by: Borislav Petkov <borislav.petkov@amd.com> Signed-off-by: H. Peter Anvin <hpa@zytor.com>
58 lines
1.4 KiB
C
58 lines
1.4 KiB
C
#include <linux/io.h>
|
|
|
|
#include <asm/trampoline.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/e820.h>
|
|
|
|
#if defined(CONFIG_X86_64) && defined(CONFIG_ACPI_SLEEP)
|
|
#define __trampinit
|
|
#define __trampinitdata
|
|
#else
|
|
#define __trampinit __cpuinit
|
|
#define __trampinitdata __cpuinitdata
|
|
#endif
|
|
|
|
/* ready for x86_64 and x86 */
|
|
unsigned char *__trampinitdata trampoline_base;
|
|
|
|
void __init reserve_trampoline_memory(void)
|
|
{
|
|
unsigned long mem;
|
|
|
|
/* Has to be in very low memory so we can execute real-mode AP code. */
|
|
mem = find_e820_area(0, 1<<20, TRAMPOLINE_SIZE, PAGE_SIZE);
|
|
if (mem == -1L)
|
|
panic("Cannot allocate trampoline\n");
|
|
|
|
trampoline_base = __va(mem);
|
|
reserve_early(mem, mem + TRAMPOLINE_SIZE, "TRAMPOLINE");
|
|
}
|
|
|
|
/*
|
|
* Currently trivial. Write the real->protected mode
|
|
* bootstrap into the page concerned. The caller
|
|
* has made sure it's suitably aligned.
|
|
*/
|
|
unsigned long __trampinit setup_trampoline(void)
|
|
{
|
|
memcpy(trampoline_base, trampoline_data, TRAMPOLINE_SIZE);
|
|
return virt_to_phys(trampoline_base);
|
|
}
|
|
|
|
void __init setup_trampoline_page_table(void)
|
|
{
|
|
#ifdef CONFIG_X86_32
|
|
/* Copy kernel address range */
|
|
clone_pgd_range(trampoline_pg_dir + KERNEL_PGD_BOUNDARY,
|
|
swapper_pg_dir + KERNEL_PGD_BOUNDARY,
|
|
min_t(unsigned long, KERNEL_PGD_PTRS,
|
|
KERNEL_PGD_BOUNDARY));
|
|
|
|
/* Initialize low mappings */
|
|
clone_pgd_range(trampoline_pg_dir,
|
|
swapper_pg_dir + KERNEL_PGD_BOUNDARY,
|
|
min_t(unsigned long, KERNEL_PGD_PTRS,
|
|
KERNEL_PGD_BOUNDARY));
|
|
#endif
|
|
}
|