mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-16 04:37:07 +07:00
58c8d17f2e
Signed-off-by: Nicholas Piggin <npiggin@gmail.com> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
653 lines
19 KiB
C
653 lines
19 KiB
C
/*
|
|
* Machine check exception handling CPU-side for power7 and power8
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
|
|
*
|
|
* Copyright 2013 IBM Corporation
|
|
* Author: Mahesh Salgaonkar <mahesh@linux.vnet.ibm.com>
|
|
*/
|
|
|
|
#undef DEBUG
|
|
#define pr_fmt(fmt) "mce_power: " fmt
|
|
|
|
#include <linux/types.h>
|
|
#include <linux/ptrace.h>
|
|
#include <asm/mmu.h>
|
|
#include <asm/mce.h>
|
|
#include <asm/machdep.h>
|
|
|
|
static void flush_tlb_206(unsigned int num_sets, unsigned int action)
|
|
{
|
|
unsigned long rb;
|
|
unsigned int i;
|
|
|
|
switch (action) {
|
|
case TLB_INVAL_SCOPE_GLOBAL:
|
|
rb = TLBIEL_INVAL_SET;
|
|
break;
|
|
case TLB_INVAL_SCOPE_LPID:
|
|
rb = TLBIEL_INVAL_SET_LPID;
|
|
break;
|
|
default:
|
|
BUG();
|
|
break;
|
|
}
|
|
|
|
asm volatile("ptesync" : : : "memory");
|
|
for (i = 0; i < num_sets; i++) {
|
|
asm volatile("tlbiel %0" : : "r" (rb));
|
|
rb += 1 << TLBIEL_INVAL_SET_SHIFT;
|
|
}
|
|
asm volatile("ptesync" : : : "memory");
|
|
}
|
|
|
|
/*
|
|
* Generic routines to flush TLB on POWER processors. These routines
|
|
* are used as flush_tlb hook in the cpu_spec.
|
|
*
|
|
* action => TLB_INVAL_SCOPE_GLOBAL: Invalidate all TLBs.
|
|
* TLB_INVAL_SCOPE_LPID: Invalidate TLB for current LPID.
|
|
*/
|
|
void __flush_tlb_power7(unsigned int action)
|
|
{
|
|
flush_tlb_206(POWER7_TLB_SETS, action);
|
|
}
|
|
|
|
void __flush_tlb_power8(unsigned int action)
|
|
{
|
|
flush_tlb_206(POWER8_TLB_SETS, action);
|
|
}
|
|
|
|
void __flush_tlb_power9(unsigned int action)
|
|
{
|
|
if (radix_enabled())
|
|
flush_tlb_206(POWER9_TLB_SETS_RADIX, action);
|
|
|
|
flush_tlb_206(POWER9_TLB_SETS_HASH, action);
|
|
}
|
|
|
|
|
|
/* flush SLBs and reload */
|
|
#ifdef CONFIG_PPC_STD_MMU_64
|
|
static void flush_and_reload_slb(void)
|
|
{
|
|
struct slb_shadow *slb;
|
|
unsigned long i, n;
|
|
|
|
/* Invalidate all SLBs */
|
|
asm volatile("slbmte %0,%0; slbia" : : "r" (0));
|
|
|
|
#ifdef CONFIG_KVM_BOOK3S_HANDLER
|
|
/*
|
|
* If machine check is hit when in guest or in transition, we will
|
|
* only flush the SLBs and continue.
|
|
*/
|
|
if (get_paca()->kvm_hstate.in_guest)
|
|
return;
|
|
#endif
|
|
|
|
/* For host kernel, reload the SLBs from shadow SLB buffer. */
|
|
slb = get_slb_shadow();
|
|
if (!slb)
|
|
return;
|
|
|
|
n = min_t(u32, be32_to_cpu(slb->persistent), SLB_MIN_SIZE);
|
|
|
|
/* Load up the SLB entries from shadow SLB */
|
|
for (i = 0; i < n; i++) {
|
|
unsigned long rb = be64_to_cpu(slb->save_area[i].esid);
|
|
unsigned long rs = be64_to_cpu(slb->save_area[i].vsid);
|
|
|
|
rb = (rb & ~0xFFFul) | i;
|
|
asm volatile("slbmte %0,%1" : : "r" (rs), "r" (rb));
|
|
}
|
|
}
|
|
#endif
|
|
|
|
static void flush_erat(void)
|
|
{
|
|
asm volatile(PPC_INVALIDATE_ERAT : : :"memory");
|
|
}
|
|
|
|
#define MCE_FLUSH_SLB 1
|
|
#define MCE_FLUSH_TLB 2
|
|
#define MCE_FLUSH_ERAT 3
|
|
|
|
static int mce_flush(int what)
|
|
{
|
|
#ifdef CONFIG_PPC_STD_MMU_64
|
|
if (what == MCE_FLUSH_SLB) {
|
|
flush_and_reload_slb();
|
|
return 1;
|
|
}
|
|
#endif
|
|
if (what == MCE_FLUSH_ERAT) {
|
|
flush_erat();
|
|
return 1;
|
|
}
|
|
if (what == MCE_FLUSH_TLB) {
|
|
if (cur_cpu_spec && cur_cpu_spec->flush_tlb) {
|
|
cur_cpu_spec->flush_tlb(TLB_INVAL_SCOPE_GLOBAL);
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mce_handle_flush_derrors(uint64_t dsisr, uint64_t slb, uint64_t tlb, uint64_t erat)
|
|
{
|
|
if ((dsisr & slb) && mce_flush(MCE_FLUSH_SLB))
|
|
dsisr &= ~slb;
|
|
if ((dsisr & erat) && mce_flush(MCE_FLUSH_ERAT))
|
|
dsisr &= ~erat;
|
|
if ((dsisr & tlb) && mce_flush(MCE_FLUSH_TLB))
|
|
dsisr &= ~tlb;
|
|
/* Any other errors we don't understand? */
|
|
if (dsisr)
|
|
return 0;
|
|
return 1;
|
|
}
|
|
|
|
|
|
/*
|
|
* Machine Check bits on power7 and power8
|
|
*/
|
|
#define P7_SRR1_MC_LOADSTORE(srr1) ((srr1) & PPC_BIT(42)) /* P8 too */
|
|
|
|
/* SRR1 bits for machine check (On Power7 and Power8) */
|
|
#define P7_SRR1_MC_IFETCH(srr1) ((srr1) & PPC_BITMASK(43, 45)) /* P8 too */
|
|
|
|
#define P7_SRR1_MC_IFETCH_UE (0x1 << PPC_BITLSHIFT(45)) /* P8 too */
|
|
#define P7_SRR1_MC_IFETCH_SLB_PARITY (0x2 << PPC_BITLSHIFT(45)) /* P8 too */
|
|
#define P7_SRR1_MC_IFETCH_SLB_MULTIHIT (0x3 << PPC_BITLSHIFT(45)) /* P8 too */
|
|
#define P7_SRR1_MC_IFETCH_SLB_BOTH (0x4 << PPC_BITLSHIFT(45))
|
|
#define P7_SRR1_MC_IFETCH_TLB_MULTIHIT (0x5 << PPC_BITLSHIFT(45)) /* P8 too */
|
|
#define P7_SRR1_MC_IFETCH_UE_TLB_RELOAD (0x6 << PPC_BITLSHIFT(45)) /* P8 too */
|
|
#define P7_SRR1_MC_IFETCH_UE_IFU_INTERNAL (0x7 << PPC_BITLSHIFT(45))
|
|
|
|
/* SRR1 bits for machine check (On Power8) */
|
|
#define P8_SRR1_MC_IFETCH_ERAT_MULTIHIT (0x4 << PPC_BITLSHIFT(45))
|
|
|
|
/* DSISR bits for machine check (On Power7 and Power8) */
|
|
#define P7_DSISR_MC_UE (PPC_BIT(48)) /* P8 too */
|
|
#define P7_DSISR_MC_UE_TABLEWALK (PPC_BIT(49)) /* P8 too */
|
|
#define P7_DSISR_MC_ERAT_MULTIHIT (PPC_BIT(52)) /* P8 too */
|
|
#define P7_DSISR_MC_TLB_MULTIHIT_MFTLB (PPC_BIT(53)) /* P8 too */
|
|
#define P7_DSISR_MC_SLB_PARITY_MFSLB (PPC_BIT(55)) /* P8 too */
|
|
#define P7_DSISR_MC_SLB_MULTIHIT (PPC_BIT(56)) /* P8 too */
|
|
#define P7_DSISR_MC_SLB_MULTIHIT_PARITY (PPC_BIT(57)) /* P8 too */
|
|
|
|
/*
|
|
* DSISR bits for machine check (Power8) in addition to above.
|
|
* Secondary DERAT Multihit
|
|
*/
|
|
#define P8_DSISR_MC_ERAT_MULTIHIT_SEC (PPC_BIT(54))
|
|
|
|
/* SLB error bits */
|
|
#define P7_DSISR_MC_SLB_ERRORS (P7_DSISR_MC_ERAT_MULTIHIT | \
|
|
P7_DSISR_MC_SLB_PARITY_MFSLB | \
|
|
P7_DSISR_MC_SLB_MULTIHIT | \
|
|
P7_DSISR_MC_SLB_MULTIHIT_PARITY)
|
|
|
|
#define P8_DSISR_MC_SLB_ERRORS (P7_DSISR_MC_SLB_ERRORS | \
|
|
P8_DSISR_MC_ERAT_MULTIHIT_SEC)
|
|
|
|
/*
|
|
* Machine Check bits on power9
|
|
*/
|
|
#define P9_SRR1_MC_LOADSTORE(srr1) (((srr1) >> PPC_BITLSHIFT(42)) & 1)
|
|
|
|
#define P9_SRR1_MC_IFETCH(srr1) ( \
|
|
PPC_BITEXTRACT(srr1, 45, 0) | \
|
|
PPC_BITEXTRACT(srr1, 44, 1) | \
|
|
PPC_BITEXTRACT(srr1, 43, 2) | \
|
|
PPC_BITEXTRACT(srr1, 36, 3) )
|
|
|
|
/* 0 is reserved */
|
|
#define P9_SRR1_MC_IFETCH_UE 1
|
|
#define P9_SRR1_MC_IFETCH_SLB_PARITY 2
|
|
#define P9_SRR1_MC_IFETCH_SLB_MULTIHIT 3
|
|
#define P9_SRR1_MC_IFETCH_ERAT_MULTIHIT 4
|
|
#define P9_SRR1_MC_IFETCH_TLB_MULTIHIT 5
|
|
#define P9_SRR1_MC_IFETCH_UE_TLB_RELOAD 6
|
|
/* 7 is reserved */
|
|
#define P9_SRR1_MC_IFETCH_LINK_TIMEOUT 8
|
|
#define P9_SRR1_MC_IFETCH_LINK_TABLEWALK_TIMEOUT 9
|
|
/* 10 ? */
|
|
#define P9_SRR1_MC_IFETCH_RA 11
|
|
#define P9_SRR1_MC_IFETCH_RA_TABLEWALK 12
|
|
#define P9_SRR1_MC_IFETCH_RA_ASYNC_STORE 13
|
|
#define P9_SRR1_MC_IFETCH_LINK_ASYNC_STORE_TIMEOUT 14
|
|
#define P9_SRR1_MC_IFETCH_RA_TABLEWALK_FOREIGN 15
|
|
|
|
/* DSISR bits for machine check (On Power9) */
|
|
#define P9_DSISR_MC_UE (PPC_BIT(48))
|
|
#define P9_DSISR_MC_UE_TABLEWALK (PPC_BIT(49))
|
|
#define P9_DSISR_MC_LINK_LOAD_TIMEOUT (PPC_BIT(50))
|
|
#define P9_DSISR_MC_LINK_TABLEWALK_TIMEOUT (PPC_BIT(51))
|
|
#define P9_DSISR_MC_ERAT_MULTIHIT (PPC_BIT(52))
|
|
#define P9_DSISR_MC_TLB_MULTIHIT_MFTLB (PPC_BIT(53))
|
|
#define P9_DSISR_MC_USER_TLBIE (PPC_BIT(54))
|
|
#define P9_DSISR_MC_SLB_PARITY_MFSLB (PPC_BIT(55))
|
|
#define P9_DSISR_MC_SLB_MULTIHIT_MFSLB (PPC_BIT(56))
|
|
#define P9_DSISR_MC_RA_LOAD (PPC_BIT(57))
|
|
#define P9_DSISR_MC_RA_TABLEWALK (PPC_BIT(58))
|
|
#define P9_DSISR_MC_RA_TABLEWALK_FOREIGN (PPC_BIT(59))
|
|
#define P9_DSISR_MC_RA_FOREIGN (PPC_BIT(60))
|
|
|
|
/* SLB error bits */
|
|
#define P9_DSISR_MC_SLB_ERRORS (P9_DSISR_MC_ERAT_MULTIHIT | \
|
|
P9_DSISR_MC_SLB_PARITY_MFSLB | \
|
|
P9_DSISR_MC_SLB_MULTIHIT_MFSLB)
|
|
|
|
static long mce_handle_derror_p7(uint64_t dsisr)
|
|
{
|
|
return mce_handle_flush_derrors(dsisr,
|
|
P7_DSISR_MC_SLB_ERRORS,
|
|
P7_DSISR_MC_TLB_MULTIHIT_MFTLB,
|
|
0);
|
|
}
|
|
|
|
static long mce_handle_ierror_p7(uint64_t srr1)
|
|
{
|
|
switch (P7_SRR1_MC_IFETCH(srr1)) {
|
|
case P7_SRR1_MC_IFETCH_SLB_PARITY:
|
|
case P7_SRR1_MC_IFETCH_SLB_MULTIHIT:
|
|
case P7_SRR1_MC_IFETCH_SLB_BOTH:
|
|
return mce_flush(MCE_FLUSH_SLB);
|
|
|
|
case P7_SRR1_MC_IFETCH_TLB_MULTIHIT:
|
|
return mce_flush(MCE_FLUSH_TLB);
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static void mce_get_common_ierror(struct mce_error_info *mce_err, uint64_t srr1)
|
|
{
|
|
switch (P7_SRR1_MC_IFETCH(srr1)) {
|
|
case P7_SRR1_MC_IFETCH_SLB_PARITY:
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_PARITY;
|
|
break;
|
|
case P7_SRR1_MC_IFETCH_SLB_MULTIHIT:
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_MULTIHIT;
|
|
break;
|
|
case P7_SRR1_MC_IFETCH_TLB_MULTIHIT:
|
|
mce_err->error_type = MCE_ERROR_TYPE_TLB;
|
|
mce_err->u.tlb_error_type = MCE_TLB_ERROR_MULTIHIT;
|
|
break;
|
|
case P7_SRR1_MC_IFETCH_UE:
|
|
case P7_SRR1_MC_IFETCH_UE_IFU_INTERNAL:
|
|
mce_err->error_type = MCE_ERROR_TYPE_UE;
|
|
mce_err->u.ue_error_type = MCE_UE_ERROR_IFETCH;
|
|
break;
|
|
case P7_SRR1_MC_IFETCH_UE_TLB_RELOAD:
|
|
mce_err->error_type = MCE_ERROR_TYPE_UE;
|
|
mce_err->u.ue_error_type =
|
|
MCE_UE_ERROR_PAGE_TABLE_WALK_IFETCH;
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void mce_get_ierror_p7(struct mce_error_info *mce_err, uint64_t srr1)
|
|
{
|
|
mce_get_common_ierror(mce_err, srr1);
|
|
if (P7_SRR1_MC_IFETCH(srr1) == P7_SRR1_MC_IFETCH_SLB_BOTH) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_INDETERMINATE;
|
|
}
|
|
}
|
|
|
|
static void mce_get_derror_p7(struct mce_error_info *mce_err, uint64_t dsisr)
|
|
{
|
|
if (dsisr & P7_DSISR_MC_UE) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_UE;
|
|
mce_err->u.ue_error_type = MCE_UE_ERROR_LOAD_STORE;
|
|
} else if (dsisr & P7_DSISR_MC_UE_TABLEWALK) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_UE;
|
|
mce_err->u.ue_error_type =
|
|
MCE_UE_ERROR_PAGE_TABLE_WALK_LOAD_STORE;
|
|
} else if (dsisr & P7_DSISR_MC_ERAT_MULTIHIT) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_ERAT;
|
|
mce_err->u.erat_error_type = MCE_ERAT_ERROR_MULTIHIT;
|
|
} else if (dsisr & P7_DSISR_MC_SLB_MULTIHIT) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_MULTIHIT;
|
|
} else if (dsisr & P7_DSISR_MC_SLB_PARITY_MFSLB) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_PARITY;
|
|
} else if (dsisr & P7_DSISR_MC_TLB_MULTIHIT_MFTLB) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_TLB;
|
|
mce_err->u.tlb_error_type = MCE_TLB_ERROR_MULTIHIT;
|
|
} else if (dsisr & P7_DSISR_MC_SLB_MULTIHIT_PARITY) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_INDETERMINATE;
|
|
}
|
|
}
|
|
|
|
static long mce_handle_ue_error(struct pt_regs *regs)
|
|
{
|
|
long handled = 0;
|
|
|
|
/*
|
|
* On specific SCOM read via MMIO we may get a machine check
|
|
* exception with SRR0 pointing inside opal. If that is the
|
|
* case OPAL may have recovery address to re-read SCOM data in
|
|
* different way and hence we can recover from this MC.
|
|
*/
|
|
|
|
if (ppc_md.mce_check_early_recovery) {
|
|
if (ppc_md.mce_check_early_recovery(regs))
|
|
handled = 1;
|
|
}
|
|
return handled;
|
|
}
|
|
|
|
long __machine_check_early_realmode_p7(struct pt_regs *regs)
|
|
{
|
|
uint64_t srr1, nip, addr;
|
|
long handled = 1;
|
|
struct mce_error_info mce_error_info = { 0 };
|
|
|
|
mce_error_info.severity = MCE_SEV_ERROR_SYNC;
|
|
mce_error_info.initiator = MCE_INITIATOR_CPU;
|
|
|
|
srr1 = regs->msr;
|
|
nip = regs->nip;
|
|
|
|
/*
|
|
* Handle memory errors depending whether this was a load/store or
|
|
* ifetch exception. Also, populate the mce error_type and
|
|
* type-specific error_type from either SRR1 or DSISR, depending
|
|
* whether this was a load/store or ifetch exception
|
|
*/
|
|
if (P7_SRR1_MC_LOADSTORE(srr1)) {
|
|
handled = mce_handle_derror_p7(regs->dsisr);
|
|
mce_get_derror_p7(&mce_error_info, regs->dsisr);
|
|
addr = regs->dar;
|
|
} else {
|
|
handled = mce_handle_ierror_p7(srr1);
|
|
mce_get_ierror_p7(&mce_error_info, srr1);
|
|
addr = regs->nip;
|
|
}
|
|
|
|
/* Handle UE error. */
|
|
if (mce_error_info.error_type == MCE_ERROR_TYPE_UE)
|
|
handled = mce_handle_ue_error(regs);
|
|
|
|
save_mce_event(regs, handled, &mce_error_info, nip, addr);
|
|
return handled;
|
|
}
|
|
|
|
static void mce_get_ierror_p8(struct mce_error_info *mce_err, uint64_t srr1)
|
|
{
|
|
mce_get_common_ierror(mce_err, srr1);
|
|
if (P7_SRR1_MC_IFETCH(srr1) == P8_SRR1_MC_IFETCH_ERAT_MULTIHIT) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_ERAT;
|
|
mce_err->u.erat_error_type = MCE_ERAT_ERROR_MULTIHIT;
|
|
}
|
|
}
|
|
|
|
static void mce_get_derror_p8(struct mce_error_info *mce_err, uint64_t dsisr)
|
|
{
|
|
mce_get_derror_p7(mce_err, dsisr);
|
|
if (dsisr & P8_DSISR_MC_ERAT_MULTIHIT_SEC) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_ERAT;
|
|
mce_err->u.erat_error_type = MCE_ERAT_ERROR_MULTIHIT;
|
|
}
|
|
}
|
|
|
|
static long mce_handle_ierror_p8(uint64_t srr1)
|
|
{
|
|
switch (P7_SRR1_MC_IFETCH(srr1)) {
|
|
case P7_SRR1_MC_IFETCH_SLB_PARITY:
|
|
case P7_SRR1_MC_IFETCH_SLB_MULTIHIT:
|
|
case P8_SRR1_MC_IFETCH_ERAT_MULTIHIT:
|
|
return mce_flush(MCE_FLUSH_SLB);
|
|
|
|
case P7_SRR1_MC_IFETCH_TLB_MULTIHIT:
|
|
return mce_flush(MCE_FLUSH_TLB);
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static long mce_handle_derror_p8(uint64_t dsisr)
|
|
{
|
|
return mce_handle_flush_derrors(dsisr,
|
|
P8_DSISR_MC_SLB_ERRORS,
|
|
P7_DSISR_MC_TLB_MULTIHIT_MFTLB,
|
|
0);
|
|
}
|
|
|
|
long __machine_check_early_realmode_p8(struct pt_regs *regs)
|
|
{
|
|
uint64_t srr1, nip, addr;
|
|
long handled = 1;
|
|
struct mce_error_info mce_error_info = { 0 };
|
|
|
|
mce_error_info.severity = MCE_SEV_ERROR_SYNC;
|
|
mce_error_info.initiator = MCE_INITIATOR_CPU;
|
|
|
|
srr1 = regs->msr;
|
|
nip = regs->nip;
|
|
|
|
if (P7_SRR1_MC_LOADSTORE(srr1)) {
|
|
handled = mce_handle_derror_p8(regs->dsisr);
|
|
mce_get_derror_p8(&mce_error_info, regs->dsisr);
|
|
addr = regs->dar;
|
|
} else {
|
|
handled = mce_handle_ierror_p8(srr1);
|
|
mce_get_ierror_p8(&mce_error_info, srr1);
|
|
addr = regs->nip;
|
|
}
|
|
|
|
/* Handle UE error. */
|
|
if (mce_error_info.error_type == MCE_ERROR_TYPE_UE)
|
|
handled = mce_handle_ue_error(regs);
|
|
|
|
save_mce_event(regs, handled, &mce_error_info, nip, addr);
|
|
return handled;
|
|
}
|
|
|
|
static int mce_handle_derror_p9(struct pt_regs *regs)
|
|
{
|
|
uint64_t dsisr = regs->dsisr;
|
|
|
|
return mce_handle_flush_derrors(dsisr,
|
|
P9_DSISR_MC_SLB_PARITY_MFSLB |
|
|
P9_DSISR_MC_SLB_MULTIHIT_MFSLB,
|
|
|
|
P9_DSISR_MC_TLB_MULTIHIT_MFTLB,
|
|
|
|
P9_DSISR_MC_ERAT_MULTIHIT);
|
|
}
|
|
|
|
static int mce_handle_ierror_p9(struct pt_regs *regs)
|
|
{
|
|
uint64_t srr1 = regs->msr;
|
|
|
|
switch (P9_SRR1_MC_IFETCH(srr1)) {
|
|
case P9_SRR1_MC_IFETCH_SLB_PARITY:
|
|
case P9_SRR1_MC_IFETCH_SLB_MULTIHIT:
|
|
return mce_flush(MCE_FLUSH_SLB);
|
|
case P9_SRR1_MC_IFETCH_TLB_MULTIHIT:
|
|
return mce_flush(MCE_FLUSH_TLB);
|
|
case P9_SRR1_MC_IFETCH_ERAT_MULTIHIT:
|
|
return mce_flush(MCE_FLUSH_ERAT);
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static void mce_get_derror_p9(struct pt_regs *regs,
|
|
struct mce_error_info *mce_err, uint64_t *addr)
|
|
{
|
|
uint64_t dsisr = regs->dsisr;
|
|
|
|
mce_err->severity = MCE_SEV_ERROR_SYNC;
|
|
mce_err->initiator = MCE_INITIATOR_CPU;
|
|
|
|
if (dsisr & P9_DSISR_MC_USER_TLBIE)
|
|
*addr = regs->nip;
|
|
else
|
|
*addr = regs->dar;
|
|
|
|
if (dsisr & P9_DSISR_MC_UE) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_UE;
|
|
mce_err->u.ue_error_type = MCE_UE_ERROR_LOAD_STORE;
|
|
} else if (dsisr & P9_DSISR_MC_UE_TABLEWALK) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_UE;
|
|
mce_err->u.ue_error_type = MCE_UE_ERROR_PAGE_TABLE_WALK_LOAD_STORE;
|
|
} else if (dsisr & P9_DSISR_MC_LINK_LOAD_TIMEOUT) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_LINK;
|
|
mce_err->u.link_error_type = MCE_LINK_ERROR_LOAD_TIMEOUT;
|
|
} else if (dsisr & P9_DSISR_MC_LINK_TABLEWALK_TIMEOUT) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_LINK;
|
|
mce_err->u.link_error_type = MCE_LINK_ERROR_PAGE_TABLE_WALK_LOAD_STORE_TIMEOUT;
|
|
} else if (dsisr & P9_DSISR_MC_ERAT_MULTIHIT) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_ERAT;
|
|
mce_err->u.erat_error_type = MCE_ERAT_ERROR_MULTIHIT;
|
|
} else if (dsisr & P9_DSISR_MC_TLB_MULTIHIT_MFTLB) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_TLB;
|
|
mce_err->u.tlb_error_type = MCE_TLB_ERROR_MULTIHIT;
|
|
} else if (dsisr & P9_DSISR_MC_USER_TLBIE) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_USER;
|
|
mce_err->u.user_error_type = MCE_USER_ERROR_TLBIE;
|
|
} else if (dsisr & P9_DSISR_MC_SLB_PARITY_MFSLB) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_PARITY;
|
|
} else if (dsisr & P9_DSISR_MC_SLB_MULTIHIT_MFSLB) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_MULTIHIT;
|
|
} else if (dsisr & P9_DSISR_MC_RA_LOAD) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_RA;
|
|
mce_err->u.ra_error_type = MCE_RA_ERROR_LOAD;
|
|
} else if (dsisr & P9_DSISR_MC_RA_TABLEWALK) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_RA;
|
|
mce_err->u.ra_error_type = MCE_RA_ERROR_PAGE_TABLE_WALK_LOAD_STORE;
|
|
} else if (dsisr & P9_DSISR_MC_RA_TABLEWALK_FOREIGN) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_RA;
|
|
mce_err->u.ra_error_type = MCE_RA_ERROR_PAGE_TABLE_WALK_LOAD_STORE_FOREIGN;
|
|
} else if (dsisr & P9_DSISR_MC_RA_FOREIGN) {
|
|
mce_err->error_type = MCE_ERROR_TYPE_RA;
|
|
mce_err->u.ra_error_type = MCE_RA_ERROR_LOAD_STORE_FOREIGN;
|
|
}
|
|
}
|
|
|
|
static void mce_get_ierror_p9(struct pt_regs *regs,
|
|
struct mce_error_info *mce_err, uint64_t *addr)
|
|
{
|
|
uint64_t srr1 = regs->msr;
|
|
|
|
switch (P9_SRR1_MC_IFETCH(srr1)) {
|
|
case P9_SRR1_MC_IFETCH_RA_ASYNC_STORE:
|
|
case P9_SRR1_MC_IFETCH_LINK_ASYNC_STORE_TIMEOUT:
|
|
mce_err->severity = MCE_SEV_FATAL;
|
|
break;
|
|
default:
|
|
mce_err->severity = MCE_SEV_ERROR_SYNC;
|
|
break;
|
|
}
|
|
|
|
mce_err->initiator = MCE_INITIATOR_CPU;
|
|
|
|
*addr = regs->nip;
|
|
|
|
switch (P9_SRR1_MC_IFETCH(srr1)) {
|
|
case P9_SRR1_MC_IFETCH_UE:
|
|
mce_err->error_type = MCE_ERROR_TYPE_UE;
|
|
mce_err->u.ue_error_type = MCE_UE_ERROR_IFETCH;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_SLB_PARITY:
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_PARITY;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_SLB_MULTIHIT:
|
|
mce_err->error_type = MCE_ERROR_TYPE_SLB;
|
|
mce_err->u.slb_error_type = MCE_SLB_ERROR_MULTIHIT;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_ERAT_MULTIHIT:
|
|
mce_err->error_type = MCE_ERROR_TYPE_ERAT;
|
|
mce_err->u.erat_error_type = MCE_ERAT_ERROR_MULTIHIT;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_TLB_MULTIHIT:
|
|
mce_err->error_type = MCE_ERROR_TYPE_TLB;
|
|
mce_err->u.tlb_error_type = MCE_TLB_ERROR_MULTIHIT;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_UE_TLB_RELOAD:
|
|
mce_err->error_type = MCE_ERROR_TYPE_UE;
|
|
mce_err->u.ue_error_type = MCE_UE_ERROR_PAGE_TABLE_WALK_IFETCH;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_LINK_TIMEOUT:
|
|
mce_err->error_type = MCE_ERROR_TYPE_LINK;
|
|
mce_err->u.link_error_type = MCE_LINK_ERROR_IFETCH_TIMEOUT;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_LINK_TABLEWALK_TIMEOUT:
|
|
mce_err->error_type = MCE_ERROR_TYPE_LINK;
|
|
mce_err->u.link_error_type = MCE_LINK_ERROR_PAGE_TABLE_WALK_IFETCH_TIMEOUT;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_RA:
|
|
mce_err->error_type = MCE_ERROR_TYPE_RA;
|
|
mce_err->u.ra_error_type = MCE_RA_ERROR_IFETCH;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_RA_TABLEWALK:
|
|
mce_err->error_type = MCE_ERROR_TYPE_RA;
|
|
mce_err->u.ra_error_type = MCE_RA_ERROR_PAGE_TABLE_WALK_IFETCH;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_RA_ASYNC_STORE:
|
|
mce_err->error_type = MCE_ERROR_TYPE_RA;
|
|
mce_err->u.ra_error_type = MCE_RA_ERROR_STORE;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_LINK_ASYNC_STORE_TIMEOUT:
|
|
mce_err->error_type = MCE_ERROR_TYPE_LINK;
|
|
mce_err->u.link_error_type = MCE_LINK_ERROR_STORE_TIMEOUT;
|
|
break;
|
|
case P9_SRR1_MC_IFETCH_RA_TABLEWALK_FOREIGN:
|
|
mce_err->error_type = MCE_ERROR_TYPE_RA;
|
|
mce_err->u.ra_error_type = MCE_RA_ERROR_PAGE_TABLE_WALK_IFETCH_FOREIGN;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
long __machine_check_early_realmode_p9(struct pt_regs *regs)
|
|
{
|
|
uint64_t nip, addr;
|
|
long handled;
|
|
struct mce_error_info mce_error_info = { 0 };
|
|
|
|
nip = regs->nip;
|
|
|
|
if (P9_SRR1_MC_LOADSTORE(regs->msr)) {
|
|
handled = mce_handle_derror_p9(regs);
|
|
mce_get_derror_p9(regs, &mce_error_info, &addr);
|
|
} else {
|
|
handled = mce_handle_ierror_p9(regs);
|
|
mce_get_ierror_p9(regs, &mce_error_info, &addr);
|
|
}
|
|
|
|
/* Handle UE error. */
|
|
if (mce_error_info.error_type == MCE_ERROR_TYPE_UE)
|
|
handled = mce_handle_ue_error(regs);
|
|
|
|
save_mce_event(regs, handled, &mce_error_info, nip, addr);
|
|
return handled;
|
|
}
|