linux_dsm_epyc7002/include/asm-x86/smp_64.h
Glauber de Oliveira Costa 7e1efc0cde x86: unify extern masks declaration
take them off smp_{32,64}.h and move to smp.h

Signed-off-by: Glauber Costa <gcosta@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-04-17 17:41:00 +02:00

75 lines
1.7 KiB
C

#ifndef __ASM_SMP_H
#define __ASM_SMP_H
#include <linux/cpumask.h>
#include <linux/init.h>
/*
* We need the APIC definitions automatically as part of 'smp.h'
*/
#include <asm/apic.h>
#include <asm/io_apic.h>
#include <asm/mpspec.h>
#include <asm/pda.h>
#include <asm/thread_info.h>
extern cpumask_t cpu_initialized;
extern cpumask_t cpu_callin_map;
extern int smp_call_function_mask(cpumask_t mask, void (*func)(void *),
void *info, int wait);
static inline int cpu_present_to_apicid(int mps_cpu)
{
if (cpu_present(mps_cpu))
return (int)per_cpu(x86_bios_cpu_apicid, mps_cpu);
else
return BAD_APICID;
}
#ifdef CONFIG_SMP
#define raw_smp_processor_id() read_pda(cpunumber)
#define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu)
#define stack_smp_processor_id() \
({ \
struct thread_info *ti; \
__asm__("andq %%rsp,%0; ":"=r" (ti) : "0" (CURRENT_MASK)); \
ti->cpu; \
})
/*
* On x86 all CPUs are mapped 1:1 to the APIC space. This simplifies
* scheduling and IPI sending and compresses data structures.
*/
static inline int num_booting_cpus(void)
{
return cpus_weight(cpu_callout_map);
}
#else /* CONFIG_SMP */
extern unsigned int boot_cpu_id;
#define cpu_physical_id(cpu) boot_cpu_id
#define stack_smp_processor_id() 0
#endif /* !CONFIG_SMP */
#define safe_smp_processor_id() smp_processor_id()
static __inline int logical_smp_processor_id(void)
{
/* we don't want to mark this access volatile - bad code generation */
return GET_APIC_LOGICAL_ID(*(u32 *)(APIC_BASE + APIC_LDR));
}
static inline int hard_smp_processor_id(void)
{
/* we don't want to mark this access volatile - bad code generation */
return GET_APIC_ID(*(u32 *)(APIC_BASE + APIC_ID));
}
#endif