mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-12 00:56:41 +07:00
1855a2c4ce
When we bring up a new CPU via INIT/startup IPI messages, the CPU that's coming up sends a xTPR message to the chipset. Intel chipsets (at least) don't provide any architectural guarantee on what the chipset will do with this message. For example, the E850x chipsets uses this xTPR message to interpret the interrupt operating mode of the platform. When the CPU coming online sends this message, it always indicates that it is in logical flat mode. For the CPU hotplug case, the platform may already be functioning in cluster APIC mode at this time, the chipset can get confused and mishandle I/O device and IPI interrupt routing. The situation eventually gets corrected when the new CPU sends another xTPR update when we switch it to cluster mode, but there's a window during which the chipset may be in an inconsistent state. This patch avoids this problem by using the flat physical interrupt delivery mode instead of cluster mode for bigsmp (>8 cpu) support. Signed-off-by: Ashok Raj <ashok.raj@intel.com> Signed-off-by: Venkatesh Pallipadi <venkatesh.pallipadi@intel.com> Cc: Andi Kleen <ak@muc.de> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
159 lines
3.4 KiB
C
159 lines
3.4 KiB
C
#ifndef __ASM_MACH_APIC_H
|
|
#define __ASM_MACH_APIC_H
|
|
|
|
|
|
extern u8 bios_cpu_apicid[];
|
|
|
|
#define xapic_phys_to_log_apicid(cpu) (bios_cpu_apicid[cpu])
|
|
#define esr_disable (1)
|
|
|
|
static inline int apic_id_registered(void)
|
|
{
|
|
return (1);
|
|
}
|
|
|
|
/* Round robin the irqs amoung the online cpus */
|
|
static inline cpumask_t target_cpus(void)
|
|
{
|
|
static unsigned long cpu = NR_CPUS;
|
|
do {
|
|
if (cpu >= NR_CPUS)
|
|
cpu = first_cpu(cpu_online_map);
|
|
else
|
|
cpu = next_cpu(cpu, cpu_online_map);
|
|
} while (cpu >= NR_CPUS);
|
|
return cpumask_of_cpu(cpu);
|
|
}
|
|
|
|
#undef APIC_DEST_LOGICAL
|
|
#define APIC_DEST_LOGICAL 0
|
|
#define TARGET_CPUS (target_cpus())
|
|
#define APIC_DFR_VALUE (APIC_DFR_FLAT)
|
|
#define INT_DELIVERY_MODE (dest_Fixed)
|
|
#define INT_DEST_MODE (0) /* phys delivery to target proc */
|
|
#define NO_BALANCE_IRQ (0)
|
|
#define WAKE_SECONDARY_VIA_INIT
|
|
|
|
|
|
static inline unsigned long check_apicid_used(physid_mask_t bitmap, int apicid)
|
|
{
|
|
return (0);
|
|
}
|
|
|
|
static inline unsigned long check_apicid_present(int bit)
|
|
{
|
|
return (1);
|
|
}
|
|
|
|
static inline unsigned long calculate_ldr(int cpu)
|
|
{
|
|
unsigned long val, id;
|
|
val = apic_read(APIC_LDR) & ~APIC_LDR_MASK;
|
|
id = xapic_phys_to_log_apicid(cpu);
|
|
val |= SET_APIC_LOGICAL_ID(id);
|
|
return val;
|
|
}
|
|
|
|
/*
|
|
* Set up the logical destination ID.
|
|
*
|
|
* Intel recommends to set DFR, LDR and TPR before enabling
|
|
* an APIC. See e.g. "AP-388 82489DX User's Manual" (Intel
|
|
* document number 292116). So here it goes...
|
|
*/
|
|
static inline void init_apic_ldr(void)
|
|
{
|
|
unsigned long val;
|
|
int cpu = smp_processor_id();
|
|
|
|
apic_write_around(APIC_DFR, APIC_DFR_VALUE);
|
|
val = calculate_ldr(cpu);
|
|
apic_write_around(APIC_LDR, val);
|
|
}
|
|
|
|
static inline void clustered_apic_check(void)
|
|
{
|
|
printk("Enabling APIC mode: %s. Using %d I/O APICs\n",
|
|
"Physflat", nr_ioapics);
|
|
}
|
|
|
|
static inline int multi_timer_check(int apic, int irq)
|
|
{
|
|
return (0);
|
|
}
|
|
|
|
static inline int apicid_to_node(int logical_apicid)
|
|
{
|
|
return (0);
|
|
}
|
|
|
|
static inline int cpu_present_to_apicid(int mps_cpu)
|
|
{
|
|
if (mps_cpu < NR_CPUS)
|
|
return (int) bios_cpu_apicid[mps_cpu];
|
|
|
|
return BAD_APICID;
|
|
}
|
|
|
|
static inline physid_mask_t apicid_to_cpu_present(int phys_apicid)
|
|
{
|
|
return physid_mask_of_physid(phys_apicid);
|
|
}
|
|
|
|
extern u8 cpu_2_logical_apicid[];
|
|
/* Mapping from cpu number to logical apicid */
|
|
static inline int cpu_to_logical_apicid(int cpu)
|
|
{
|
|
if (cpu >= NR_CPUS)
|
|
return BAD_APICID;
|
|
return cpu_physical_id(cpu);
|
|
}
|
|
|
|
static inline int mpc_apic_id(struct mpc_config_processor *m,
|
|
struct mpc_config_translation *translation_record)
|
|
{
|
|
printk("Processor #%d %ld:%ld APIC version %d\n",
|
|
m->mpc_apicid,
|
|
(m->mpc_cpufeature & CPU_FAMILY_MASK) >> 8,
|
|
(m->mpc_cpufeature & CPU_MODEL_MASK) >> 4,
|
|
m->mpc_apicver);
|
|
return m->mpc_apicid;
|
|
}
|
|
|
|
static inline physid_mask_t ioapic_phys_id_map(physid_mask_t phys_map)
|
|
{
|
|
/* For clustered we don't have a good way to do this yet - hack */
|
|
return physids_promote(0xFFL);
|
|
}
|
|
|
|
static inline void setup_portio_remap(void)
|
|
{
|
|
}
|
|
|
|
static inline void enable_apic_mode(void)
|
|
{
|
|
}
|
|
|
|
static inline int check_phys_apicid_present(int boot_cpu_physical_apicid)
|
|
{
|
|
return (1);
|
|
}
|
|
|
|
/* As we are using single CPU as destination, pick only one CPU here */
|
|
static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask)
|
|
{
|
|
int cpu;
|
|
int apicid;
|
|
|
|
cpu = first_cpu(cpumask);
|
|
apicid = cpu_to_logical_apicid(cpu);
|
|
return apicid;
|
|
}
|
|
|
|
static inline u32 phys_pkg_id(u32 cpuid_apic, int index_msb)
|
|
{
|
|
return cpuid_apic >> index_msb;
|
|
}
|
|
|
|
#endif /* __ASM_MACH_APIC_H */
|