2019-06-03 12:44:50 +07:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2014-11-14 22:54:09 +07:00
|
|
|
/*
|
|
|
|
* Contains CPU specific errata definitions
|
|
|
|
*
|
|
|
|
* Copyright (C) 2014 ARM Ltd.
|
|
|
|
*/
|
|
|
|
|
2018-06-05 18:50:07 +07:00
|
|
|
#include <linux/arm-smccc.h>
|
2014-11-14 22:54:09 +07:00
|
|
|
#include <linux/types.h>
|
2019-04-13 03:39:32 +07:00
|
|
|
#include <linux/cpu.h>
|
2014-11-14 22:54:09 +07:00
|
|
|
#include <asm/cpu.h>
|
|
|
|
#include <asm/cputype.h>
|
|
|
|
#include <asm/cpufeature.h>
|
2020-02-19 02:58:39 +07:00
|
|
|
#include <asm/kvm_asm.h>
|
2019-04-09 22:26:21 +07:00
|
|
|
#include <asm/smp_plat.h>
|
2014-11-14 22:54:09 +07:00
|
|
|
|
2014-11-14 22:54:10 +07:00
|
|
|
static bool __maybe_unused
|
2016-04-22 18:25:31 +07:00
|
|
|
is_affected_midr_range(const struct arm64_cpu_capabilities *entry, int scope)
|
2014-11-14 22:54:10 +07:00
|
|
|
{
|
2018-03-07 00:15:34 +07:00
|
|
|
const struct arm64_midr_revidr *fix;
|
|
|
|
u32 midr = read_cpuid_id(), revidr;
|
|
|
|
|
2016-04-22 18:25:31 +07:00
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
2018-03-26 21:12:44 +07:00
|
|
|
if (!is_midr_in_range(midr, &entry->midr_range))
|
2018-03-07 00:15:34 +07:00
|
|
|
return false;
|
|
|
|
|
|
|
|
midr &= MIDR_REVISION_MASK | MIDR_VARIANT_MASK;
|
|
|
|
revidr = read_cpuid(REVIDR_EL1);
|
|
|
|
for (fix = entry->fixed_revs; fix && fix->revidr_mask; fix++)
|
|
|
|
if (midr == fix->midr_rv && (revidr & fix->revidr_mask))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return true;
|
2014-11-14 22:54:10 +07:00
|
|
|
}
|
|
|
|
|
2018-03-26 21:12:45 +07:00
|
|
|
static bool __maybe_unused
|
|
|
|
is_affected_midr_range_list(const struct arm64_cpu_capabilities *entry,
|
|
|
|
int scope)
|
2014-11-14 22:54:10 +07:00
|
|
|
{
|
2016-04-22 18:25:31 +07:00
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
2018-03-26 21:12:45 +07:00
|
|
|
return is_midr_in_range_list(read_cpuid_id(), entry->midr_range_list);
|
2014-11-14 22:54:10 +07:00
|
|
|
}
|
|
|
|
|
2017-12-14 05:19:37 +07:00
|
|
|
static bool __maybe_unused
|
|
|
|
is_kryo_midr(const struct arm64_cpu_capabilities *entry, int scope)
|
|
|
|
{
|
|
|
|
u32 model;
|
|
|
|
|
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
|
|
|
|
|
|
model = read_cpuid_id();
|
|
|
|
model &= MIDR_IMPLEMENTOR_MASK | (0xf00 << MIDR_PARTNUM_SHIFT) |
|
|
|
|
MIDR_ARCHITECTURE_MASK;
|
|
|
|
|
2018-03-26 21:12:44 +07:00
|
|
|
return model == entry->midr_range.model;
|
2017-12-14 05:19:37 +07:00
|
|
|
}
|
|
|
|
|
2016-09-09 20:07:16 +07:00
|
|
|
static bool
|
2018-07-05 05:07:46 +07:00
|
|
|
has_mismatched_cache_type(const struct arm64_cpu_capabilities *entry,
|
|
|
|
int scope)
|
2016-09-09 20:07:16 +07:00
|
|
|
{
|
2018-10-09 20:47:06 +07:00
|
|
|
u64 mask = arm64_ftr_reg_ctrel0.strict_mask;
|
|
|
|
u64 sys = arm64_ftr_reg_ctrel0.sys_val & mask;
|
|
|
|
u64 ctr_raw, ctr_real;
|
2018-07-05 05:07:46 +07:00
|
|
|
|
2016-09-09 20:07:16 +07:00
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
2018-10-09 20:47:06 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* We want to make sure that all the CPUs in the system expose
|
|
|
|
* a consistent CTR_EL0 to make sure that applications behaves
|
|
|
|
* correctly with migration.
|
|
|
|
*
|
|
|
|
* If a CPU has CTR_EL0.IDC but does not advertise it via CTR_EL0 :
|
|
|
|
*
|
|
|
|
* 1) It is safe if the system doesn't support IDC, as CPU anyway
|
|
|
|
* reports IDC = 0, consistent with the rest.
|
|
|
|
*
|
|
|
|
* 2) If the system has IDC, it is still safe as we trap CTR_EL0
|
|
|
|
* access on this CPU via the ARM64_HAS_CACHE_IDC capability.
|
|
|
|
*
|
|
|
|
* So, we need to make sure either the raw CTR_EL0 or the effective
|
|
|
|
* CTR_EL0 matches the system's copy to allow a secondary CPU to boot.
|
|
|
|
*/
|
|
|
|
ctr_raw = read_cpuid_cachetype() & mask;
|
|
|
|
ctr_real = read_cpuid_effective_cachetype() & mask;
|
|
|
|
|
|
|
|
return (ctr_real != sys) && (ctr_raw != sys);
|
2016-09-09 20:07:16 +07:00
|
|
|
}
|
|
|
|
|
2018-03-26 21:12:28 +07:00
|
|
|
static void
|
2019-10-18 00:42:58 +07:00
|
|
|
cpu_enable_trap_ctr_access(const struct arm64_cpu_capabilities *cap)
|
2016-09-09 20:07:16 +07:00
|
|
|
{
|
2018-10-09 20:47:07 +07:00
|
|
|
u64 mask = arm64_ftr_reg_ctrel0.strict_mask;
|
2019-10-18 00:42:58 +07:00
|
|
|
bool enable_uct_trap = false;
|
2018-10-09 20:47:07 +07:00
|
|
|
|
|
|
|
/* Trap CTR_EL0 access on this CPU, only if it has a mismatch */
|
|
|
|
if ((read_cpuid_cachetype() & mask) !=
|
|
|
|
(arm64_ftr_reg_ctrel0.sys_val & mask))
|
2019-10-18 00:42:58 +07:00
|
|
|
enable_uct_trap = true;
|
|
|
|
|
|
|
|
/* ... or if the system is affected by an erratum */
|
|
|
|
if (cap->capability == ARM64_WORKAROUND_1542419)
|
|
|
|
enable_uct_trap = true;
|
|
|
|
|
|
|
|
if (enable_uct_trap)
|
2018-10-09 20:47:07 +07:00
|
|
|
sysreg_clear_set(sctlr_el1, SCTLR_EL1_UCT, 0);
|
2016-09-09 20:07:16 +07:00
|
|
|
}
|
|
|
|
|
2018-03-13 19:40:39 +07:00
|
|
|
atomic_t arm64_el2_vector_last_slot = ATOMIC_INIT(-1);
|
|
|
|
|
2018-01-03 18:17:58 +07:00
|
|
|
#include <asm/mmu_context.h>
|
|
|
|
#include <asm/cacheflush.h>
|
|
|
|
|
|
|
|
DEFINE_PER_CPU_READ_MOSTLY(struct bp_hardening_data, bp_hardening_data);
|
|
|
|
|
2018-04-10 17:36:45 +07:00
|
|
|
#ifdef CONFIG_KVM_INDIRECT_VECTORS
|
2018-01-03 18:17:58 +07:00
|
|
|
static void __copy_hyp_vect_bpi(int slot, const char *hyp_vecs_start,
|
|
|
|
const char *hyp_vecs_end)
|
|
|
|
{
|
2020-02-19 02:58:38 +07:00
|
|
|
void *dst = lm_alias(__bp_harden_hyp_vecs + slot * SZ_2K);
|
2018-01-03 18:17:58 +07:00
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < SZ_2K; i += 0x80)
|
|
|
|
memcpy(dst + i, hyp_vecs_start, hyp_vecs_end - hyp_vecs_start);
|
|
|
|
|
2018-06-11 20:22:09 +07:00
|
|
|
__flush_icache_range((uintptr_t)dst, (uintptr_t)dst + SZ_2K);
|
2018-01-03 18:17:58 +07:00
|
|
|
}
|
|
|
|
|
2019-04-16 04:21:23 +07:00
|
|
|
static void install_bp_hardening_cb(bp_hardening_cb_t fn,
|
|
|
|
const char *hyp_vecs_start,
|
|
|
|
const char *hyp_vecs_end)
|
2018-01-03 18:17:58 +07:00
|
|
|
{
|
2018-11-27 22:35:21 +07:00
|
|
|
static DEFINE_RAW_SPINLOCK(bp_lock);
|
2018-01-03 18:17:58 +07:00
|
|
|
int cpu, slot = -1;
|
|
|
|
|
2018-09-22 03:49:19 +07:00
|
|
|
/*
|
2019-09-23 16:12:29 +07:00
|
|
|
* detect_harden_bp_fw() passes NULL for the hyp_vecs start/end if
|
|
|
|
* we're a guest. Skip the hyp-vectors work.
|
2018-09-22 03:49:19 +07:00
|
|
|
*/
|
|
|
|
if (!hyp_vecs_start) {
|
|
|
|
__this_cpu_write(bp_hardening_data.fn, fn);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2018-11-27 22:35:21 +07:00
|
|
|
raw_spin_lock(&bp_lock);
|
2018-01-03 18:17:58 +07:00
|
|
|
for_each_possible_cpu(cpu) {
|
|
|
|
if (per_cpu(bp_hardening_data.fn, cpu) == fn) {
|
|
|
|
slot = per_cpu(bp_hardening_data.hyp_vectors_slot, cpu);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (slot == -1) {
|
2018-03-13 19:40:39 +07:00
|
|
|
slot = atomic_inc_return(&arm64_el2_vector_last_slot);
|
|
|
|
BUG_ON(slot >= BP_HARDEN_EL2_SLOTS);
|
2018-01-03 18:17:58 +07:00
|
|
|
__copy_hyp_vect_bpi(slot, hyp_vecs_start, hyp_vecs_end);
|
|
|
|
}
|
|
|
|
|
|
|
|
__this_cpu_write(bp_hardening_data.hyp_vectors_slot, slot);
|
|
|
|
__this_cpu_write(bp_hardening_data.fn, fn);
|
2018-11-27 22:35:21 +07:00
|
|
|
raw_spin_unlock(&bp_lock);
|
2018-01-03 18:17:58 +07:00
|
|
|
}
|
|
|
|
#else
|
2019-04-16 04:21:23 +07:00
|
|
|
static void install_bp_hardening_cb(bp_hardening_cb_t fn,
|
2018-01-03 18:17:58 +07:00
|
|
|
const char *hyp_vecs_start,
|
|
|
|
const char *hyp_vecs_end)
|
|
|
|
{
|
|
|
|
__this_cpu_write(bp_hardening_data.fn, fn);
|
|
|
|
}
|
2018-04-10 17:36:45 +07:00
|
|
|
#endif /* CONFIG_KVM_INDIRECT_VECTORS */
|
2018-01-03 18:17:58 +07:00
|
|
|
|
2018-02-07 00:56:20 +07:00
|
|
|
#include <linux/arm-smccc.h>
|
2018-01-03 19:46:21 +07:00
|
|
|
|
2020-03-11 06:25:44 +07:00
|
|
|
static void __maybe_unused call_smc_arch_workaround_1(void)
|
2018-02-07 00:56:20 +07:00
|
|
|
{
|
|
|
|
arm_smccc_1_1_smc(ARM_SMCCC_ARCH_WORKAROUND_1, NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void call_hvc_arch_workaround_1(void)
|
|
|
|
{
|
|
|
|
arm_smccc_1_1_hvc(ARM_SMCCC_ARCH_WORKAROUND_1, NULL);
|
|
|
|
}
|
|
|
|
|
2018-04-10 17:36:42 +07:00
|
|
|
static void qcom_link_stack_sanitization(void)
|
|
|
|
{
|
|
|
|
u64 tmp;
|
|
|
|
|
|
|
|
asm volatile("mov %0, x30 \n"
|
|
|
|
".rept 16 \n"
|
|
|
|
"bl . + 4 \n"
|
|
|
|
".endr \n"
|
|
|
|
"mov x30, %0 \n"
|
|
|
|
: "=&r" (tmp));
|
|
|
|
}
|
|
|
|
|
2019-04-16 04:21:20 +07:00
|
|
|
static bool __nospectre_v2;
|
|
|
|
static int __init parse_nospectre_v2(char *str)
|
|
|
|
{
|
|
|
|
__nospectre_v2 = true;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_param("nospectre_v2", parse_nospectre_v2);
|
|
|
|
|
2019-04-16 04:21:23 +07:00
|
|
|
/*
|
|
|
|
* -1: No workaround
|
|
|
|
* 0: No workaround required
|
|
|
|
* 1: Workaround installed
|
|
|
|
*/
|
|
|
|
static int detect_harden_bp_fw(void)
|
2018-02-07 00:56:20 +07:00
|
|
|
{
|
|
|
|
bp_hardening_cb_t cb;
|
|
|
|
void *smccc_start, *smccc_end;
|
|
|
|
struct arm_smccc_res res;
|
2018-04-10 17:36:42 +07:00
|
|
|
u32 midr = read_cpuid_id();
|
2018-02-07 00:56:20 +07:00
|
|
|
|
2019-10-21 22:28:22 +07:00
|
|
|
arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID,
|
|
|
|
ARM_SMCCC_ARCH_WORKAROUND_1, &res);
|
|
|
|
|
|
|
|
switch ((int)res.a0) {
|
|
|
|
case 1:
|
|
|
|
/* Firmware says we're just fine */
|
|
|
|
return 0;
|
|
|
|
case 0:
|
|
|
|
break;
|
|
|
|
default:
|
2019-04-16 04:21:23 +07:00
|
|
|
return -1;
|
2019-10-21 22:28:22 +07:00
|
|
|
}
|
2018-02-07 00:56:20 +07:00
|
|
|
|
2019-08-09 20:22:41 +07:00
|
|
|
switch (arm_smccc_1_1_get_conduit()) {
|
|
|
|
case SMCCC_CONDUIT_HVC:
|
2019-10-21 22:28:22 +07:00
|
|
|
cb = call_hvc_arch_workaround_1;
|
|
|
|
/* This is a guest, no need to patch KVM vectors */
|
|
|
|
smccc_start = NULL;
|
|
|
|
smccc_end = NULL;
|
2018-02-07 00:56:20 +07:00
|
|
|
break;
|
|
|
|
|
2020-02-19 02:58:39 +07:00
|
|
|
#if IS_ENABLED(CONFIG_KVM_ARM_HOST)
|
2019-08-09 20:22:41 +07:00
|
|
|
case SMCCC_CONDUIT_SMC:
|
2019-10-21 22:28:22 +07:00
|
|
|
cb = call_smc_arch_workaround_1;
|
2020-02-19 02:58:39 +07:00
|
|
|
smccc_start = __smccc_workaround_1_smc;
|
|
|
|
smccc_end = __smccc_workaround_1_smc +
|
|
|
|
__SMCCC_WORKAROUND_1_SMC_SZ;
|
2018-02-07 00:56:20 +07:00
|
|
|
break;
|
2020-02-19 02:58:39 +07:00
|
|
|
#endif
|
2018-02-07 00:56:20 +07:00
|
|
|
|
|
|
|
default:
|
2019-04-16 04:21:23 +07:00
|
|
|
return -1;
|
2018-02-07 00:56:20 +07:00
|
|
|
}
|
|
|
|
|
2018-04-10 17:36:42 +07:00
|
|
|
if (((midr & MIDR_CPU_MODEL_MASK) == MIDR_QCOM_FALKOR) ||
|
|
|
|
((midr & MIDR_CPU_MODEL_MASK) == MIDR_QCOM_FALKOR_V1))
|
|
|
|
cb = qcom_link_stack_sanitization;
|
|
|
|
|
2019-04-16 04:21:25 +07:00
|
|
|
if (IS_ENABLED(CONFIG_HARDEN_BRANCH_PREDICTOR))
|
|
|
|
install_bp_hardening_cb(cb, smccc_start, smccc_end);
|
2018-02-07 00:56:20 +07:00
|
|
|
|
2019-04-16 04:21:23 +07:00
|
|
|
return 1;
|
2018-01-03 19:46:21 +07:00
|
|
|
}
|
2018-01-03 18:17:58 +07:00
|
|
|
|
2018-05-29 19:11:07 +07:00
|
|
|
DEFINE_PER_CPU_READ_MOSTLY(u64, arm64_ssbd_callback_required);
|
|
|
|
|
2018-05-29 19:11:09 +07:00
|
|
|
int ssbd_state __read_mostly = ARM64_SSBD_KERNEL;
|
2019-04-16 04:21:28 +07:00
|
|
|
static bool __ssb_safe = true;
|
2018-05-29 19:11:09 +07:00
|
|
|
|
|
|
|
static const struct ssbd_options {
|
|
|
|
const char *str;
|
|
|
|
int state;
|
|
|
|
} ssbd_options[] = {
|
|
|
|
{ "force-on", ARM64_SSBD_FORCE_ENABLE, },
|
|
|
|
{ "force-off", ARM64_SSBD_FORCE_DISABLE, },
|
|
|
|
{ "kernel", ARM64_SSBD_KERNEL, },
|
|
|
|
};
|
|
|
|
|
|
|
|
static int __init ssbd_cfg(char *buf)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (!buf || !buf[0])
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ssbd_options); i++) {
|
|
|
|
int len = strlen(ssbd_options[i].str);
|
|
|
|
|
|
|
|
if (strncmp(buf, ssbd_options[i].str, len))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
ssbd_state = ssbd_options[i].state;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
early_param("ssbd", ssbd_cfg);
|
|
|
|
|
2018-05-29 19:11:06 +07:00
|
|
|
void __init arm64_update_smccc_conduit(struct alt_instr *alt,
|
|
|
|
__le32 *origptr, __le32 *updptr,
|
|
|
|
int nr_inst)
|
|
|
|
{
|
|
|
|
u32 insn;
|
|
|
|
|
|
|
|
BUG_ON(nr_inst != 1);
|
|
|
|
|
2019-08-09 20:22:41 +07:00
|
|
|
switch (arm_smccc_1_1_get_conduit()) {
|
|
|
|
case SMCCC_CONDUIT_HVC:
|
2018-05-29 19:11:06 +07:00
|
|
|
insn = aarch64_insn_get_hvc_value();
|
|
|
|
break;
|
2019-08-09 20:22:41 +07:00
|
|
|
case SMCCC_CONDUIT_SMC:
|
2018-05-29 19:11:06 +07:00
|
|
|
insn = aarch64_insn_get_smc_value();
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
*updptr = cpu_to_le32(insn);
|
|
|
|
}
|
2018-05-29 19:11:08 +07:00
|
|
|
|
2018-05-29 19:11:11 +07:00
|
|
|
void __init arm64_enable_wa2_handling(struct alt_instr *alt,
|
|
|
|
__le32 *origptr, __le32 *updptr,
|
|
|
|
int nr_inst)
|
|
|
|
{
|
|
|
|
BUG_ON(nr_inst != 1);
|
|
|
|
/*
|
|
|
|
* Only allow mitigation on EL1 entry/exit and guest
|
|
|
|
* ARCH_WORKAROUND_2 handling if the SSBD state allows it to
|
|
|
|
* be flipped.
|
|
|
|
*/
|
|
|
|
if (arm64_get_ssbd_state() == ARM64_SSBD_KERNEL)
|
|
|
|
*updptr = cpu_to_le32(aarch64_insn_gen_nop());
|
|
|
|
}
|
|
|
|
|
2018-05-29 19:11:12 +07:00
|
|
|
void arm64_set_ssbd_mitigation(bool state)
|
2018-05-29 19:11:08 +07:00
|
|
|
{
|
2019-10-21 22:28:22 +07:00
|
|
|
int conduit;
|
|
|
|
|
2019-04-16 04:21:27 +07:00
|
|
|
if (!IS_ENABLED(CONFIG_ARM64_SSBD)) {
|
|
|
|
pr_info_once("SSBD disabled by kernel configuration\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2018-08-07 19:47:06 +07:00
|
|
|
if (this_cpu_has_cap(ARM64_SSBS)) {
|
|
|
|
if (state)
|
|
|
|
asm volatile(SET_PSTATE_SSBS(0));
|
|
|
|
else
|
|
|
|
asm volatile(SET_PSTATE_SSBS(1));
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2019-10-21 22:28:22 +07:00
|
|
|
conduit = arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_WORKAROUND_2, state,
|
|
|
|
NULL);
|
2018-05-29 19:11:08 +07:00
|
|
|
|
2019-10-21 22:28:22 +07:00
|
|
|
WARN_ON_ONCE(conduit == SMCCC_CONDUIT_NONE);
|
2018-05-29 19:11:08 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static bool has_ssbd_mitigation(const struct arm64_cpu_capabilities *entry,
|
|
|
|
int scope)
|
|
|
|
{
|
|
|
|
struct arm_smccc_res res;
|
2018-05-29 19:11:09 +07:00
|
|
|
bool required = true;
|
|
|
|
s32 val;
|
2019-04-16 04:21:28 +07:00
|
|
|
bool this_cpu_safe = false;
|
2019-10-21 22:28:22 +07:00
|
|
|
int conduit;
|
2018-05-29 19:11:08 +07:00
|
|
|
|
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
|
|
|
2019-04-13 03:39:32 +07:00
|
|
|
if (cpu_mitigations_off())
|
|
|
|
ssbd_state = ARM64_SSBD_FORCE_DISABLE;
|
|
|
|
|
2019-04-30 22:58:56 +07:00
|
|
|
/* delay setting __ssb_safe until we get a firmware response */
|
|
|
|
if (is_midr_in_range_list(read_cpuid_id(), entry->midr_range_list))
|
|
|
|
this_cpu_safe = true;
|
|
|
|
|
2018-08-07 19:47:06 +07:00
|
|
|
if (this_cpu_has_cap(ARM64_SSBS)) {
|
2019-04-30 22:58:56 +07:00
|
|
|
if (!this_cpu_safe)
|
|
|
|
__ssb_safe = false;
|
2018-08-07 19:47:06 +07:00
|
|
|
required = false;
|
|
|
|
goto out_printmsg;
|
|
|
|
}
|
|
|
|
|
2019-10-21 22:28:22 +07:00
|
|
|
conduit = arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID,
|
|
|
|
ARM_SMCCC_ARCH_WORKAROUND_2, &res);
|
2018-05-29 19:11:08 +07:00
|
|
|
|
2019-10-21 22:28:22 +07:00
|
|
|
if (conduit == SMCCC_CONDUIT_NONE) {
|
2018-05-29 19:11:09 +07:00
|
|
|
ssbd_state = ARM64_SSBD_UNKNOWN;
|
2019-04-16 04:21:28 +07:00
|
|
|
if (!this_cpu_safe)
|
|
|
|
__ssb_safe = false;
|
2018-05-29 19:11:09 +07:00
|
|
|
return false;
|
2018-05-29 19:11:08 +07:00
|
|
|
}
|
|
|
|
|
2018-05-29 19:11:09 +07:00
|
|
|
val = (s32)res.a0;
|
|
|
|
|
|
|
|
switch (val) {
|
|
|
|
case SMCCC_RET_NOT_SUPPORTED:
|
|
|
|
ssbd_state = ARM64_SSBD_UNKNOWN;
|
2019-04-16 04:21:28 +07:00
|
|
|
if (!this_cpu_safe)
|
|
|
|
__ssb_safe = false;
|
2018-05-29 19:11:09 +07:00
|
|
|
return false;
|
|
|
|
|
2019-04-16 04:21:28 +07:00
|
|
|
/* machines with mixed mitigation requirements must not return this */
|
2018-05-29 19:11:09 +07:00
|
|
|
case SMCCC_RET_NOT_REQUIRED:
|
|
|
|
pr_info_once("%s mitigation not required\n", entry->desc);
|
|
|
|
ssbd_state = ARM64_SSBD_MITIGATED;
|
|
|
|
return false;
|
|
|
|
|
|
|
|
case SMCCC_RET_SUCCESS:
|
2019-04-16 04:21:28 +07:00
|
|
|
__ssb_safe = false;
|
2018-05-29 19:11:09 +07:00
|
|
|
required = true;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 1: /* Mitigation not required on this CPU */
|
|
|
|
required = false;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
WARN_ON(1);
|
2019-04-16 04:21:28 +07:00
|
|
|
if (!this_cpu_safe)
|
|
|
|
__ssb_safe = false;
|
2018-05-29 19:11:09 +07:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (ssbd_state) {
|
|
|
|
case ARM64_SSBD_FORCE_DISABLE:
|
|
|
|
arm64_set_ssbd_mitigation(false);
|
|
|
|
required = false;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case ARM64_SSBD_KERNEL:
|
|
|
|
if (required) {
|
|
|
|
__this_cpu_write(arm64_ssbd_callback_required, 1);
|
|
|
|
arm64_set_ssbd_mitigation(true);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case ARM64_SSBD_FORCE_ENABLE:
|
2018-05-29 19:11:08 +07:00
|
|
|
arm64_set_ssbd_mitigation(true);
|
2018-05-29 19:11:09 +07:00
|
|
|
required = true;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
WARN_ON(1);
|
|
|
|
break;
|
2018-05-29 19:11:08 +07:00
|
|
|
}
|
|
|
|
|
2018-08-07 19:47:06 +07:00
|
|
|
out_printmsg:
|
|
|
|
switch (ssbd_state) {
|
|
|
|
case ARM64_SSBD_FORCE_DISABLE:
|
|
|
|
pr_info_once("%s disabled from command-line\n", entry->desc);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case ARM64_SSBD_FORCE_ENABLE:
|
|
|
|
pr_info_once("%s forced from command-line\n", entry->desc);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2018-05-29 19:11:09 +07:00
|
|
|
return required;
|
2018-05-29 19:11:08 +07:00
|
|
|
}
|
2018-05-29 19:11:06 +07:00
|
|
|
|
2019-04-16 04:21:28 +07:00
|
|
|
/* known invulnerable cores */
|
|
|
|
static const struct midr_range arm64_ssb_cpus[] = {
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A35),
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A53),
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A55),
|
2019-11-01 04:47:24 +07:00
|
|
|
MIDR_ALL_VERSIONS(MIDR_BRAHMA_B53),
|
2019-04-16 04:21:28 +07:00
|
|
|
{},
|
|
|
|
};
|
|
|
|
|
2019-04-29 19:03:57 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1463225
|
|
|
|
DEFINE_PER_CPU(int, __in_cortex_a76_erratum_1463225_wa);
|
|
|
|
|
|
|
|
static bool
|
|
|
|
has_cortex_a76_erratum_1463225(const struct arm64_cpu_capabilities *entry,
|
|
|
|
int scope)
|
|
|
|
{
|
|
|
|
u32 midr = read_cpuid_id();
|
|
|
|
/* Cortex-A76 r0p0 - r3p1 */
|
|
|
|
struct midr_range range = MIDR_RANGE(MIDR_CORTEX_A76, 0, 0, 3, 1);
|
|
|
|
|
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
|
|
return is_midr_in_range(midr, &range) && is_kernel_in_hyp_mode();
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2018-08-07 19:53:41 +07:00
|
|
|
static void __maybe_unused
|
|
|
|
cpu_enable_cache_maint_trap(const struct arm64_cpu_capabilities *__unused)
|
|
|
|
{
|
|
|
|
sysreg_clear_set(sctlr_el1, SCTLR_EL1_UCI, 0);
|
|
|
|
}
|
|
|
|
|
2018-03-26 21:12:43 +07:00
|
|
|
#define CAP_MIDR_RANGE(model, v_min, r_min, v_max, r_max) \
|
|
|
|
.matches = is_affected_midr_range, \
|
2018-03-26 21:12:44 +07:00
|
|
|
.midr_range = MIDR_RANGE(model, v_min, r_min, v_max, r_max)
|
2018-03-26 21:12:43 +07:00
|
|
|
|
|
|
|
#define CAP_MIDR_ALL_VERSIONS(model) \
|
|
|
|
.matches = is_affected_midr_range, \
|
2018-03-26 21:12:44 +07:00
|
|
|
.midr_range = MIDR_ALL_VERSIONS(model)
|
2017-02-01 21:38:46 +07:00
|
|
|
|
2018-03-07 00:15:34 +07:00
|
|
|
#define MIDR_FIXED(rev, revidr_mask) \
|
|
|
|
.fixed_revs = (struct arm64_midr_revidr[]){{ (rev), (revidr_mask) }, {}}
|
|
|
|
|
2018-03-26 21:12:43 +07:00
|
|
|
#define ERRATA_MIDR_RANGE(model, v_min, r_min, v_max, r_max) \
|
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM, \
|
|
|
|
CAP_MIDR_RANGE(model, v_min, r_min, v_max, r_max)
|
|
|
|
|
2018-03-26 21:12:45 +07:00
|
|
|
#define CAP_MIDR_RANGE_LIST(list) \
|
|
|
|
.matches = is_affected_midr_range_list, \
|
|
|
|
.midr_range_list = list
|
|
|
|
|
2018-03-26 21:12:43 +07:00
|
|
|
/* Errata affecting a range of revisions of given model variant */
|
|
|
|
#define ERRATA_MIDR_REV_RANGE(m, var, r_min, r_max) \
|
|
|
|
ERRATA_MIDR_RANGE(m, var, r_min, var, r_max)
|
|
|
|
|
|
|
|
/* Errata affecting a single variant/revision of a model */
|
|
|
|
#define ERRATA_MIDR_REV(model, var, rev) \
|
|
|
|
ERRATA_MIDR_RANGE(model, var, rev, var, rev)
|
|
|
|
|
|
|
|
/* Errata affecting all variants/revisions of a given a model */
|
|
|
|
#define ERRATA_MIDR_ALL_VERSIONS(model) \
|
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM, \
|
|
|
|
CAP_MIDR_ALL_VERSIONS(model)
|
|
|
|
|
2018-03-26 21:12:45 +07:00
|
|
|
/* Errata affecting a list of midr ranges, with same work around */
|
|
|
|
#define ERRATA_MIDR_RANGE_LIST(midr_list) \
|
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM, \
|
|
|
|
CAP_MIDR_RANGE_LIST(midr_list)
|
|
|
|
|
2019-04-16 04:21:26 +07:00
|
|
|
/* Track overall mitigation state. We are only mitigated if all cores are ok */
|
|
|
|
static bool __hardenbp_enab = true;
|
|
|
|
static bool __spectrev2_safe = true;
|
|
|
|
|
2019-05-03 21:27:48 +07:00
|
|
|
int get_spectre_v2_workaround_state(void)
|
|
|
|
{
|
|
|
|
if (__spectrev2_safe)
|
|
|
|
return ARM64_BP_HARDEN_NOT_REQUIRED;
|
|
|
|
|
|
|
|
if (!__hardenbp_enab)
|
|
|
|
return ARM64_BP_HARDEN_UNKNOWN;
|
|
|
|
|
|
|
|
return ARM64_BP_HARDEN_WA_NEEDED;
|
|
|
|
}
|
|
|
|
|
2018-03-26 21:12:45 +07:00
|
|
|
/*
|
2019-04-16 04:21:23 +07:00
|
|
|
* List of CPUs that do not need any Spectre-v2 mitigation at all.
|
2018-03-26 21:12:45 +07:00
|
|
|
*/
|
2019-04-16 04:21:23 +07:00
|
|
|
static const struct midr_range spectre_v2_safe_list[] = {
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A35),
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A53),
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A55),
|
2019-11-01 04:47:24 +07:00
|
|
|
MIDR_ALL_VERSIONS(MIDR_BRAHMA_B53),
|
2019-12-20 16:17:10 +07:00
|
|
|
MIDR_ALL_VERSIONS(MIDR_HISI_TSV110),
|
2020-01-17 03:03:53 +07:00
|
|
|
MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_3XX_SILVER),
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_QCOM_KRYO_4XX_SILVER),
|
2019-04-16 04:21:23 +07:00
|
|
|
{ /* sentinel */ }
|
2018-03-26 21:12:45 +07:00
|
|
|
};
|
|
|
|
|
2019-04-16 04:21:26 +07:00
|
|
|
/*
|
|
|
|
* Track overall bp hardening for all heterogeneous cores in the machine.
|
|
|
|
* We are only considered "safe" if all booted cores are known safe.
|
|
|
|
*/
|
2019-04-16 04:21:23 +07:00
|
|
|
static bool __maybe_unused
|
|
|
|
check_branch_predictor(const struct arm64_cpu_capabilities *entry, int scope)
|
|
|
|
{
|
|
|
|
int need_wa;
|
|
|
|
|
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
|
|
|
|
|
|
/* If the CPU has CSV2 set, we're safe */
|
|
|
|
if (cpuid_feature_extract_unsigned_field(read_cpuid(ID_AA64PFR0_EL1),
|
|
|
|
ID_AA64PFR0_CSV2_SHIFT))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
/* Alternatively, we have a list of unaffected CPUs */
|
|
|
|
if (is_midr_in_range_list(read_cpuid_id(), spectre_v2_safe_list))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
/* Fallback to firmware detection */
|
|
|
|
need_wa = detect_harden_bp_fw();
|
|
|
|
if (!need_wa)
|
|
|
|
return false;
|
|
|
|
|
2019-04-16 04:21:26 +07:00
|
|
|
__spectrev2_safe = false;
|
|
|
|
|
2019-04-16 04:21:25 +07:00
|
|
|
if (!IS_ENABLED(CONFIG_HARDEN_BRANCH_PREDICTOR)) {
|
|
|
|
pr_warn_once("spectrev2 mitigation disabled by kernel configuration\n");
|
|
|
|
__hardenbp_enab = false;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2019-04-16 04:21:23 +07:00
|
|
|
/* forced off */
|
2019-04-13 03:39:32 +07:00
|
|
|
if (__nospectre_v2 || cpu_mitigations_off()) {
|
2019-04-16 04:21:23 +07:00
|
|
|
pr_info_once("spectrev2 mitigation disabled by command line option\n");
|
2019-04-16 04:21:26 +07:00
|
|
|
__hardenbp_enab = false;
|
2019-04-16 04:21:23 +07:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2019-04-16 04:21:26 +07:00
|
|
|
if (need_wa < 0) {
|
2019-04-16 04:21:23 +07:00
|
|
|
pr_warn_once("ARM_SMCCC_ARCH_WORKAROUND_1 missing from firmware\n");
|
2019-04-16 04:21:26 +07:00
|
|
|
__hardenbp_enab = false;
|
|
|
|
}
|
2019-04-16 04:21:23 +07:00
|
|
|
|
|
|
|
return (need_wa > 0);
|
|
|
|
}
|
2017-02-01 21:38:46 +07:00
|
|
|
|
2019-04-09 22:26:21 +07:00
|
|
|
static const __maybe_unused struct midr_range tx2_family_cpus[] = {
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_BRCM_VULCAN),
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_CAVIUM_THUNDERX2),
|
|
|
|
{},
|
|
|
|
};
|
|
|
|
|
|
|
|
static bool __maybe_unused
|
|
|
|
needs_tx2_tvm_workaround(const struct arm64_cpu_capabilities *entry,
|
|
|
|
int scope)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (!is_affected_midr_range_list(entry, scope) ||
|
|
|
|
!is_hyp_mode_available())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
if (MPIDR_AFFINITY_LEVEL(cpu_logical_map(i), 0) != 0)
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2019-10-18 00:42:58 +07:00
|
|
|
static bool __maybe_unused
|
|
|
|
has_neoverse_n1_erratum_1542419(const struct arm64_cpu_capabilities *entry,
|
|
|
|
int scope)
|
|
|
|
{
|
|
|
|
u32 midr = read_cpuid_id();
|
|
|
|
bool has_dic = read_cpuid_cachetype() & BIT(CTR_DIC_SHIFT);
|
|
|
|
const struct midr_range range = MIDR_ALL_VERSIONS(MIDR_NEOVERSE_N1);
|
|
|
|
|
|
|
|
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
|
|
|
|
return is_midr_in_range(midr, &range) && has_dic;
|
|
|
|
}
|
2018-04-10 17:36:43 +07:00
|
|
|
|
2018-11-24 00:25:52 +07:00
|
|
|
#if defined(CONFIG_HARDEN_EL2_VECTORS) || defined(CONFIG_ARM64_ERRATUM_1319367)
|
2018-04-10 17:36:43 +07:00
|
|
|
|
2018-11-24 00:25:52 +07:00
|
|
|
static const struct midr_range ca57_a72[] = {
|
2018-04-10 17:36:43 +07:00
|
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A57),
|
|
|
|
MIDR_ALL_VERSIONS(MIDR_CORTEX_A72),
|
|
|
|
{},
|
|
|
|
};
|
|
|
|
|
2018-03-28 18:46:07 +07:00
|
|
|
#endif
|
|
|
|
|
2018-11-19 18:27:28 +07:00
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_REPEAT_TLBI
|
2019-10-30 06:27:38 +07:00
|
|
|
static const struct arm64_cpu_capabilities arm64_repeat_tlbi_list[] = {
|
2018-11-19 18:27:28 +07:00
|
|
|
#ifdef CONFIG_QCOM_FALKOR_ERRATUM_1009
|
2019-10-30 06:27:38 +07:00
|
|
|
{
|
|
|
|
ERRATA_MIDR_REV(MIDR_QCOM_FALKOR_V1, 0, 0)
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.midr_range.model = MIDR_QCOM_KRYO,
|
|
|
|
.matches = is_kryo_midr,
|
|
|
|
},
|
2018-11-19 18:27:28 +07:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1286807
|
2019-10-30 06:27:38 +07:00
|
|
|
{
|
|
|
|
ERRATA_MIDR_RANGE(MIDR_CORTEX_A76, 0, 0, 3, 0),
|
|
|
|
},
|
2018-11-19 18:27:28 +07:00
|
|
|
#endif
|
|
|
|
{},
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
2018-12-01 00:18:01 +07:00
|
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_27456
|
2019-01-08 23:19:01 +07:00
|
|
|
const struct midr_range cavium_erratum_27456_cpus[] = {
|
2018-12-01 00:18:01 +07:00
|
|
|
/* Cavium ThunderX, T88 pass 1.x - 2.1 */
|
|
|
|
MIDR_RANGE(MIDR_THUNDERX, 0, 0, 1, 1),
|
|
|
|
/* Cavium ThunderX, T81 pass 1.0 */
|
|
|
|
MIDR_REV(MIDR_THUNDERX_81XX, 0, 0),
|
|
|
|
{},
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_30115
|
|
|
|
static const struct midr_range cavium_erratum_30115_cpus[] = {
|
|
|
|
/* Cavium ThunderX, T88 pass 1.x - 2.2 */
|
|
|
|
MIDR_RANGE(MIDR_THUNDERX, 0, 0, 1, 2),
|
|
|
|
/* Cavium ThunderX, T81 pass 1.0 - 1.2 */
|
|
|
|
MIDR_REV_RANGE(MIDR_THUNDERX_81XX, 0, 0, 2),
|
|
|
|
/* Cavium ThunderX, T83 pass 1.0 */
|
|
|
|
MIDR_REV(MIDR_THUNDERX_83XX, 0, 0),
|
|
|
|
{},
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
2018-12-01 00:18:02 +07:00
|
|
|
#ifdef CONFIG_QCOM_FALKOR_ERRATUM_1003
|
|
|
|
static const struct arm64_cpu_capabilities qcom_erratum_1003_list[] = {
|
|
|
|
{
|
|
|
|
ERRATA_MIDR_REV(MIDR_QCOM_FALKOR_V1, 0, 0),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.midr_range.model = MIDR_QCOM_KRYO,
|
|
|
|
.matches = is_kryo_midr,
|
|
|
|
},
|
|
|
|
{},
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
2018-12-01 00:18:00 +07:00
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_CLEAN_CACHE
|
|
|
|
static const struct midr_range workaround_clean_cache[] = {
|
2014-11-14 22:54:12 +07:00
|
|
|
#if defined(CONFIG_ARM64_ERRATUM_826319) || \
|
|
|
|
defined(CONFIG_ARM64_ERRATUM_827319) || \
|
|
|
|
defined(CONFIG_ARM64_ERRATUM_824069)
|
2018-12-01 00:18:00 +07:00
|
|
|
/* Cortex-A53 r0p[012]: ARM errata 826319, 827319, 824069 */
|
|
|
|
MIDR_REV_RANGE(MIDR_CORTEX_A53, 0, 0, 2),
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_819472
|
|
|
|
/* Cortex-A53 r0p[01] : ARM errata 819472 */
|
|
|
|
MIDR_REV_RANGE(MIDR_CORTEX_A53, 0, 0, 1),
|
2014-11-14 22:54:12 +07:00
|
|
|
#endif
|
2018-12-01 00:18:00 +07:00
|
|
|
{},
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
2019-05-23 17:24:50 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1418040
|
|
|
|
/*
|
|
|
|
* - 1188873 affects r0p0 to r2p0
|
|
|
|
* - 1418040 affects r0p0 to r3p1
|
|
|
|
*/
|
|
|
|
static const struct midr_range erratum_1418040_list[] = {
|
|
|
|
/* Cortex-A76 r0p0 to r3p1 */
|
|
|
|
MIDR_RANGE(MIDR_CORTEX_A76, 0, 0, 3, 1),
|
|
|
|
/* Neoverse-N1 r0p0 to r3p1 */
|
|
|
|
MIDR_RANGE(MIDR_NEOVERSE_N1, 0, 0, 3, 1),
|
2019-04-15 19:03:54 +07:00
|
|
|
{},
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
2019-11-01 04:47:23 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_845719
|
|
|
|
static const struct midr_range erratum_845719_list[] = {
|
|
|
|
/* Cortex-A53 r0p[01234] */
|
|
|
|
MIDR_REV_RANGE(MIDR_CORTEX_A53, 0, 0, 4),
|
|
|
|
/* Brahma-B53 r0p[0] */
|
|
|
|
MIDR_REV(MIDR_BRAHMA_B53, 0, 0),
|
|
|
|
{},
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
2019-11-01 04:47:25 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_843419
|
|
|
|
static const struct arm64_cpu_capabilities erratum_843419_list[] = {
|
|
|
|
{
|
|
|
|
/* Cortex-A53 r0p[01234] */
|
|
|
|
.matches = is_affected_midr_range,
|
|
|
|
ERRATA_MIDR_REV_RANGE(MIDR_CORTEX_A53, 0, 0, 4),
|
|
|
|
MIDR_FIXED(0x4, BIT(8)),
|
|
|
|
},
|
|
|
|
{
|
|
|
|
/* Brahma-B53 r0p[0] */
|
|
|
|
.matches = is_affected_midr_range,
|
|
|
|
ERRATA_MIDR_REV(MIDR_BRAHMA_B53, 0, 0),
|
|
|
|
},
|
|
|
|
{},
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
2019-12-16 18:56:29 +07:00
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_SPECULATIVE_AT_VHE
|
|
|
|
static const struct midr_range erratum_speculative_at_vhe_list[] = {
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1165522
|
|
|
|
/* Cortex A76 r0p0 to r2p0 */
|
|
|
|
MIDR_RANGE(MIDR_CORTEX_A76, 0, 0, 2, 0),
|
2019-12-16 18:56:31 +07:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1530923
|
|
|
|
/* Cortex A55 r0p0 to r2p0 */
|
|
|
|
MIDR_RANGE(MIDR_CORTEX_A55, 0, 0, 2, 0),
|
2019-12-16 18:56:29 +07:00
|
|
|
#endif
|
|
|
|
{},
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
2018-12-01 00:18:00 +07:00
|
|
|
const struct arm64_cpu_capabilities arm64_errata[] = {
|
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_CLEAN_CACHE
|
2014-11-14 22:54:12 +07:00
|
|
|
{
|
2018-12-01 00:18:00 +07:00
|
|
|
.desc = "ARM errata 826319, 827319, 824069, 819472",
|
2014-11-14 22:54:12 +07:00
|
|
|
.capability = ARM64_WORKAROUND_CLEAN_CACHE,
|
2018-12-01 00:18:00 +07:00
|
|
|
ERRATA_MIDR_RANGE_LIST(workaround_clean_cache),
|
2018-03-26 21:12:28 +07:00
|
|
|
.cpu_enable = cpu_enable_cache_maint_trap,
|
2014-11-14 22:54:12 +07:00
|
|
|
},
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_832075
|
2014-11-14 22:54:10 +07:00
|
|
|
{
|
2014-11-14 22:54:11 +07:00
|
|
|
/* Cortex-A57 r0p0 - r1p2 */
|
|
|
|
.desc = "ARM erratum 832075",
|
|
|
|
.capability = ARM64_WORKAROUND_DEVICE_LOAD_ACQUIRE,
|
2018-03-26 21:12:43 +07:00
|
|
|
ERRATA_MIDR_RANGE(MIDR_CORTEX_A57,
|
|
|
|
0, 0,
|
|
|
|
1, 2),
|
2014-11-14 22:54:11 +07:00
|
|
|
},
|
2015-03-24 02:07:02 +07:00
|
|
|
#endif
|
2015-11-16 17:28:18 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_834220
|
|
|
|
{
|
|
|
|
/* Cortex-A57 r0p0 - r1p2 */
|
|
|
|
.desc = "ARM erratum 834220",
|
|
|
|
.capability = ARM64_WORKAROUND_834220,
|
2018-03-26 21:12:43 +07:00
|
|
|
ERRATA_MIDR_RANGE(MIDR_CORTEX_A57,
|
|
|
|
0, 0,
|
|
|
|
1, 2),
|
2015-11-16 17:28:18 +07:00
|
|
|
},
|
|
|
|
#endif
|
2018-03-07 00:15:35 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_843419
|
|
|
|
{
|
|
|
|
.desc = "ARM erratum 843419",
|
|
|
|
.capability = ARM64_WORKAROUND_843419,
|
2019-11-01 04:47:25 +07:00
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
|
|
.matches = cpucap_multi_entry_cap_matches,
|
|
|
|
.match_list = erratum_843419_list,
|
2015-11-16 17:28:18 +07:00
|
|
|
},
|
|
|
|
#endif
|
2015-03-24 02:07:02 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_845719
|
|
|
|
{
|
|
|
|
.desc = "ARM erratum 845719",
|
|
|
|
.capability = ARM64_WORKAROUND_845719,
|
2019-11-01 04:47:23 +07:00
|
|
|
ERRATA_MIDR_RANGE_LIST(erratum_845719_list),
|
2015-03-24 02:07:02 +07:00
|
|
|
},
|
2015-09-22 03:58:35 +07:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_23154
|
|
|
|
{
|
|
|
|
/* Cavium ThunderX, pass 1.x */
|
|
|
|
.desc = "Cavium erratum 23154",
|
|
|
|
.capability = ARM64_WORKAROUND_CAVIUM_23154,
|
2018-03-26 21:12:43 +07:00
|
|
|
ERRATA_MIDR_REV_RANGE(MIDR_THUNDERX, 0, 0, 1),
|
2015-09-22 03:58:35 +07:00
|
|
|
},
|
2016-02-25 08:44:57 +07:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_27456
|
|
|
|
{
|
2016-07-07 11:48:17 +07:00
|
|
|
.desc = "Cavium erratum 27456",
|
|
|
|
.capability = ARM64_WORKAROUND_CAVIUM_27456,
|
2018-12-01 00:18:01 +07:00
|
|
|
ERRATA_MIDR_RANGE_LIST(cavium_erratum_27456_cpus),
|
2016-07-07 11:48:17 +07:00
|
|
|
},
|
2017-06-09 18:49:48 +07:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_CAVIUM_ERRATUM_30115
|
|
|
|
{
|
|
|
|
.desc = "Cavium erratum 30115",
|
|
|
|
.capability = ARM64_WORKAROUND_CAVIUM_30115,
|
2018-12-01 00:18:01 +07:00
|
|
|
ERRATA_MIDR_RANGE_LIST(cavium_erratum_30115_cpus),
|
2017-06-09 18:49:48 +07:00
|
|
|
},
|
2014-11-14 22:54:12 +07:00
|
|
|
#endif
|
2016-09-09 20:07:16 +07:00
|
|
|
{
|
2018-09-19 17:41:21 +07:00
|
|
|
.desc = "Mismatched cache type (CTR_EL0)",
|
2018-07-05 05:07:46 +07:00
|
|
|
.capability = ARM64_MISMATCHED_CACHE_TYPE,
|
|
|
|
.matches = has_mismatched_cache_type,
|
2018-03-26 21:12:32 +07:00
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
2018-03-26 21:12:28 +07:00
|
|
|
.cpu_enable = cpu_enable_trap_ctr_access,
|
2016-09-09 20:07:16 +07:00
|
|
|
},
|
2017-02-09 03:08:37 +07:00
|
|
|
#ifdef CONFIG_QCOM_FALKOR_ERRATUM_1003
|
|
|
|
{
|
2018-12-01 00:18:02 +07:00
|
|
|
.desc = "Qualcomm Technologies Falkor/Kryo erratum 1003",
|
2017-12-14 05:19:37 +07:00
|
|
|
.capability = ARM64_WORKAROUND_QCOM_FALKOR_E1003,
|
2019-10-30 00:15:39 +07:00
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
2018-12-12 22:53:54 +07:00
|
|
|
.matches = cpucap_multi_entry_cap_matches,
|
2018-12-01 00:18:02 +07:00
|
|
|
.match_list = qcom_erratum_1003_list,
|
2017-12-14 05:19:37 +07:00
|
|
|
},
|
2017-02-09 03:08:37 +07:00
|
|
|
#endif
|
2018-11-19 18:27:28 +07:00
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_REPEAT_TLBI
|
2017-02-01 00:50:19 +07:00
|
|
|
{
|
2018-11-19 18:27:28 +07:00
|
|
|
.desc = "Qualcomm erratum 1009, ARM erratum 1286807",
|
2017-02-01 00:50:19 +07:00
|
|
|
.capability = ARM64_WORKAROUND_REPEAT_TLBI,
|
2019-10-30 06:27:38 +07:00
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
|
|
.matches = cpucap_multi_entry_cap_matches,
|
|
|
|
.match_list = arm64_repeat_tlbi_list,
|
2017-02-01 00:50:19 +07:00
|
|
|
},
|
2017-03-21 00:18:06 +07:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_858921
|
|
|
|
{
|
|
|
|
/* Cortex-A73 all versions */
|
|
|
|
.desc = "ARM erratum 858921",
|
|
|
|
.capability = ARM64_WORKAROUND_858921,
|
2018-03-26 21:12:43 +07:00
|
|
|
ERRATA_MIDR_ALL_VERSIONS(MIDR_CORTEX_A73),
|
2017-03-21 00:18:06 +07:00
|
|
|
},
|
2018-01-03 19:46:21 +07:00
|
|
|
#endif
|
|
|
|
{
|
|
|
|
.capability = ARM64_HARDEN_BRANCH_PREDICTOR,
|
2019-04-16 04:21:23 +07:00
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
|
|
.matches = check_branch_predictor,
|
2018-01-19 19:22:47 +07:00
|
|
|
},
|
2018-02-15 18:49:20 +07:00
|
|
|
#ifdef CONFIG_HARDEN_EL2_VECTORS
|
|
|
|
{
|
2018-04-10 17:36:43 +07:00
|
|
|
.desc = "EL2 vector hardening",
|
2018-02-15 18:49:20 +07:00
|
|
|
.capability = ARM64_HARDEN_EL2_VECTORS,
|
2018-11-24 00:25:52 +07:00
|
|
|
ERRATA_MIDR_RANGE_LIST(ca57_a72),
|
2018-02-15 18:49:20 +07:00
|
|
|
},
|
2018-05-29 19:11:08 +07:00
|
|
|
#endif
|
|
|
|
{
|
|
|
|
.desc = "Speculative Store Bypass Disable",
|
|
|
|
.capability = ARM64_SSBD,
|
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
|
|
.matches = has_ssbd_mitigation,
|
2019-04-16 04:21:28 +07:00
|
|
|
.midr_range_list = arm64_ssb_cpus,
|
2018-05-29 19:11:08 +07:00
|
|
|
},
|
2019-05-23 17:24:50 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1418040
|
2018-09-27 23:15:34 +07:00
|
|
|
{
|
2019-05-23 17:24:50 +07:00
|
|
|
.desc = "ARM erratum 1418040",
|
|
|
|
.capability = ARM64_WORKAROUND_1418040,
|
|
|
|
ERRATA_MIDR_RANGE_LIST(erratum_1418040_list),
|
2018-09-27 23:15:34 +07:00
|
|
|
},
|
2018-12-07 00:31:23 +07:00
|
|
|
#endif
|
2019-12-16 18:56:29 +07:00
|
|
|
#ifdef CONFIG_ARM64_WORKAROUND_SPECULATIVE_AT_VHE
|
2018-12-07 00:31:23 +07:00
|
|
|
{
|
2019-12-16 18:56:31 +07:00
|
|
|
.desc = "ARM errata 1165522, 1530923",
|
2019-12-16 18:56:29 +07:00
|
|
|
.capability = ARM64_WORKAROUND_SPECULATIVE_AT_VHE,
|
|
|
|
ERRATA_MIDR_RANGE_LIST(erratum_speculative_at_vhe_list),
|
2018-12-07 00:31:23 +07:00
|
|
|
},
|
2019-04-29 19:03:57 +07:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1463225
|
|
|
|
{
|
|
|
|
.desc = "ARM erratum 1463225",
|
|
|
|
.capability = ARM64_WORKAROUND_1463225,
|
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
|
|
.matches = has_cortex_a76_erratum_1463225,
|
|
|
|
},
|
2019-04-09 22:26:21 +07:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_CAVIUM_TX2_ERRATUM_219
|
|
|
|
{
|
|
|
|
.desc = "Cavium ThunderX2 erratum 219 (KVM guest sysreg trapping)",
|
|
|
|
.capability = ARM64_WORKAROUND_CAVIUM_TX2_219_TVM,
|
|
|
|
ERRATA_MIDR_RANGE_LIST(tx2_family_cpus),
|
|
|
|
.matches = needs_tx2_tvm_workaround,
|
|
|
|
},
|
2019-04-09 22:22:24 +07:00
|
|
|
{
|
|
|
|
.desc = "Cavium ThunderX2 erratum 219 (PRFM removal)",
|
|
|
|
.capability = ARM64_WORKAROUND_CAVIUM_TX2_219_PRFM,
|
|
|
|
ERRATA_MIDR_RANGE_LIST(tx2_family_cpus),
|
|
|
|
},
|
2019-10-28 23:12:40 +07:00
|
|
|
#endif
|
2019-10-18 00:42:58 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1542419
|
|
|
|
{
|
|
|
|
/* we depend on the firmware portion for correctness */
|
|
|
|
.desc = "ARM erratum 1542419 (kernel portion)",
|
|
|
|
.capability = ARM64_WORKAROUND_1542419,
|
|
|
|
.type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM,
|
|
|
|
.matches = has_neoverse_n1_erratum_1542419,
|
|
|
|
.cpu_enable = cpu_enable_trap_ctr_access,
|
|
|
|
},
|
Merge branch 'kvm-arm64/erratum-1319367' of git://git.kernel.org/pub/scm/linux/kernel/git/maz/arm-platforms into for-next/core
Similarly to erratum 1165522 that affects Cortex-A76, A57 and A72
respectively suffer from errata 1319537 and 1319367, potentially
resulting in TLB corruption if the CPU speculates an AT instruction
while switching guests.
The fix is slightly more involved since we don't have VHE to help us
here, but the idea is the same: when switching a guest in, we must
prevent any speculated AT from being able to parse the page tables
until S2 is up and running. Only at this stage can we allow AT to take
place.
For this, we always restore the guest sysregs first, except for its
SCTLR and TCR registers, which must be set with SCTLR.M=1 and
TCR.EPD{0,1} = {1, 1}, effectively disabling the PTW and TLB
allocation. Once S2 is setup, we restore the guest's SCTLR and
TCR. Similar things must be done on TLB invalidation...
* 'kvm-arm64/erratum-1319367' of git://git.kernel.org/pub/scm/linux/kernel/git/maz/arm-platforms:
arm64: Enable and document ARM errata 1319367 and 1319537
arm64: KVM: Prevent speculative S1 PTW when restoring vcpu context
arm64: KVM: Disable EL1 PTW when invalidating S2 TLBs
arm64: KVM: Reorder system register restoration and stage-2 activation
arm64: Add ARM64_WORKAROUND_1319367 for all A57 and A72 versions
2019-10-28 23:22:49 +07:00
|
|
|
#endif
|
2018-11-24 00:25:52 +07:00
|
|
|
#ifdef CONFIG_ARM64_ERRATUM_1319367
|
|
|
|
{
|
|
|
|
.desc = "ARM erratum 1319367",
|
2019-12-16 18:56:30 +07:00
|
|
|
.capability = ARM64_WORKAROUND_SPECULATIVE_AT_NVHE,
|
2018-11-24 00:25:52 +07:00
|
|
|
ERRATA_MIDR_RANGE_LIST(ca57_a72),
|
|
|
|
},
|
2017-02-01 00:50:19 +07:00
|
|
|
#endif
|
2014-11-14 22:54:11 +07:00
|
|
|
{
|
2014-11-14 22:54:10 +07:00
|
|
|
}
|
2014-11-14 22:54:09 +07:00
|
|
|
};
|
2019-04-16 04:21:21 +07:00
|
|
|
|
|
|
|
ssize_t cpu_show_spectre_v1(struct device *dev, struct device_attribute *attr,
|
|
|
|
char *buf)
|
|
|
|
{
|
|
|
|
return sprintf(buf, "Mitigation: __user pointer sanitization\n");
|
|
|
|
}
|
2019-04-16 04:21:26 +07:00
|
|
|
|
|
|
|
ssize_t cpu_show_spectre_v2(struct device *dev, struct device_attribute *attr,
|
|
|
|
char *buf)
|
|
|
|
{
|
2019-05-03 21:27:48 +07:00
|
|
|
switch (get_spectre_v2_workaround_state()) {
|
|
|
|
case ARM64_BP_HARDEN_NOT_REQUIRED:
|
2019-04-16 04:21:26 +07:00
|
|
|
return sprintf(buf, "Not affected\n");
|
2019-05-03 21:27:48 +07:00
|
|
|
case ARM64_BP_HARDEN_WA_NEEDED:
|
2019-04-16 04:21:26 +07:00
|
|
|
return sprintf(buf, "Mitigation: Branch predictor hardening\n");
|
2019-05-03 21:27:48 +07:00
|
|
|
case ARM64_BP_HARDEN_UNKNOWN:
|
|
|
|
default:
|
|
|
|
return sprintf(buf, "Vulnerable\n");
|
|
|
|
}
|
2019-04-16 04:21:26 +07:00
|
|
|
}
|
2019-04-16 04:21:28 +07:00
|
|
|
|
|
|
|
ssize_t cpu_show_spec_store_bypass(struct device *dev,
|
|
|
|
struct device_attribute *attr, char *buf)
|
|
|
|
{
|
|
|
|
if (__ssb_safe)
|
|
|
|
return sprintf(buf, "Not affected\n");
|
|
|
|
|
|
|
|
switch (ssbd_state) {
|
|
|
|
case ARM64_SSBD_KERNEL:
|
|
|
|
case ARM64_SSBD_FORCE_ENABLE:
|
|
|
|
if (IS_ENABLED(CONFIG_ARM64_SSBD))
|
|
|
|
return sprintf(buf,
|
|
|
|
"Mitigation: Speculative Store Bypass disabled via prctl\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
return sprintf(buf, "Vulnerable\n");
|
|
|
|
}
|