mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2025-02-11 15:25:07 +07:00
![Paul Gortmaker](/assets/img/avatar_default.png)
The __cpuinit type of throwaway sections might have made sense
some time ago when RAM was more constrained, but now the savings
do not offset the cost and complications. For example, the fix in
commit 5e427ec2d0
("x86: Fix bit corruption at CPU resume time")
is a good example of the nasty type of bugs that can be created
with improper use of the various __init prefixes.
After a discussion on LKML[1] it was decided that cpuinit should go
the way of devinit and be phased out. Once all the users are gone,
we can then finally remove the macros themselves from linux/init.h.
Note that some harmless section mismatch warnings may result, since
notify_cpu_starting() and cpu_up() are arch independent (kernel/cpu.c)
are flagged as __cpuinit -- so if we remove the __cpuinit from
arch specific callers, we will also get section mismatch warnings.
As an intermediate step, we intend to turn the linux/init.h cpuinit
content into no-ops as early as possible, since that will get rid
of these warnings. In any case, they are temporary and harmless.
This removes all the arch/cris uses of the __cpuinit macros from
all C files. Currently cris does not have any __CPUINIT used in
assembly files.
[1] https://lkml.org/lkml/2013/5/20/589
Cc: Mikael Starvik <starvik@axis.com>
Cc: Jesper Nilsson <jesper.nilsson@axis.com>
Cc: linux-cris-kernel@axis.com
Acked-by: Jesper Nilsson <jesper.nilsson@axis.com>
Signed-off-by: Paul Gortmaker <paul.gortmaker@windriver.com>
359 lines
8.0 KiB
C
359 lines
8.0 KiB
C
#include <linux/types.h>
|
|
#include <asm/delay.h>
|
|
#include <irq.h>
|
|
#include <hwregs/intr_vect.h>
|
|
#include <hwregs/intr_vect_defs.h>
|
|
#include <asm/tlbflush.h>
|
|
#include <asm/mmu_context.h>
|
|
#include <hwregs/asm/mmu_defs_asm.h>
|
|
#include <hwregs/supp_reg.h>
|
|
#include <linux/atomic.h>
|
|
|
|
#include <linux/err.h>
|
|
#include <linux/init.h>
|
|
#include <linux/timex.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/cpumask.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/module.h>
|
|
|
|
#define IPI_SCHEDULE 1
|
|
#define IPI_CALL 2
|
|
#define IPI_FLUSH_TLB 4
|
|
#define IPI_BOOT 8
|
|
|
|
#define FLUSH_ALL (void*)0xffffffff
|
|
|
|
/* Vector of locks used for various atomic operations */
|
|
spinlock_t cris_atomic_locks[] = {
|
|
[0 ... LOCK_COUNT - 1] = __SPIN_LOCK_UNLOCKED(cris_atomic_locks)
|
|
};
|
|
|
|
/* CPU masks */
|
|
cpumask_t phys_cpu_present_map = CPU_MASK_NONE;
|
|
EXPORT_SYMBOL(phys_cpu_present_map);
|
|
|
|
/* Variables used during SMP boot */
|
|
volatile int cpu_now_booting = 0;
|
|
volatile struct thread_info *smp_init_current_idle_thread;
|
|
|
|
/* Variables used during IPI */
|
|
static DEFINE_SPINLOCK(call_lock);
|
|
static DEFINE_SPINLOCK(tlbstate_lock);
|
|
|
|
struct call_data_struct {
|
|
void (*func) (void *info);
|
|
void *info;
|
|
int wait;
|
|
};
|
|
|
|
static struct call_data_struct * call_data;
|
|
|
|
static struct mm_struct* flush_mm;
|
|
static struct vm_area_struct* flush_vma;
|
|
static unsigned long flush_addr;
|
|
|
|
/* Mode registers */
|
|
static unsigned long irq_regs[NR_CPUS] = {
|
|
regi_irq,
|
|
regi_irq2
|
|
};
|
|
|
|
static irqreturn_t crisv32_ipi_interrupt(int irq, void *dev_id);
|
|
static int send_ipi(int vector, int wait, cpumask_t cpu_mask);
|
|
static struct irqaction irq_ipi = {
|
|
.handler = crisv32_ipi_interrupt,
|
|
.flags = IRQF_DISABLED,
|
|
.name = "ipi",
|
|
};
|
|
|
|
extern void cris_mmu_init(void);
|
|
extern void cris_timer_init(void);
|
|
|
|
/* SMP initialization */
|
|
void __init smp_prepare_cpus(unsigned int max_cpus)
|
|
{
|
|
int i;
|
|
|
|
/* From now on we can expect IPIs so set them up */
|
|
setup_irq(IPI_INTR_VECT, &irq_ipi);
|
|
|
|
/* Mark all possible CPUs as present */
|
|
for (i = 0; i < max_cpus; i++)
|
|
cpumask_set_cpu(i, &phys_cpu_present_map);
|
|
}
|
|
|
|
void smp_prepare_boot_cpu(void)
|
|
{
|
|
/* PGD pointer has moved after per_cpu initialization so
|
|
* update the MMU.
|
|
*/
|
|
pgd_t **pgd;
|
|
pgd = (pgd_t**)&per_cpu(current_pgd, smp_processor_id());
|
|
|
|
SUPP_BANK_SEL(1);
|
|
SUPP_REG_WR(RW_MM_TLB_PGD, pgd);
|
|
SUPP_BANK_SEL(2);
|
|
SUPP_REG_WR(RW_MM_TLB_PGD, pgd);
|
|
|
|
set_cpu_online(0, true);
|
|
cpumask_set_cpu(0, &phys_cpu_present_map);
|
|
set_cpu_possible(0, true);
|
|
}
|
|
|
|
void __init smp_cpus_done(unsigned int max_cpus)
|
|
{
|
|
}
|
|
|
|
/* Bring one cpu online.*/
|
|
static int __init
|
|
smp_boot_one_cpu(int cpuid, struct task_struct idle)
|
|
{
|
|
unsigned timeout;
|
|
cpumask_t cpu_mask;
|
|
|
|
cpumask_clear(&cpu_mask);
|
|
task_thread_info(idle)->cpu = cpuid;
|
|
|
|
/* Information to the CPU that is about to boot */
|
|
smp_init_current_idle_thread = task_thread_info(idle);
|
|
cpu_now_booting = cpuid;
|
|
|
|
/* Kick it */
|
|
set_cpu_online(cpuid, true);
|
|
cpumask_set_cpu(cpuid, &cpu_mask);
|
|
send_ipi(IPI_BOOT, 0, cpu_mask);
|
|
set_cpu_online(cpuid, false);
|
|
|
|
/* Wait for CPU to come online */
|
|
for (timeout = 0; timeout < 10000; timeout++) {
|
|
if(cpu_online(cpuid)) {
|
|
cpu_now_booting = 0;
|
|
smp_init_current_idle_thread = NULL;
|
|
return 0; /* CPU online */
|
|
}
|
|
udelay(100);
|
|
barrier();
|
|
}
|
|
|
|
printk(KERN_CRIT "SMP: CPU:%d is stuck.\n", cpuid);
|
|
return -1;
|
|
}
|
|
|
|
/* Secondary CPUs starts using C here. Here we need to setup CPU
|
|
* specific stuff such as the local timer and the MMU. */
|
|
void __init smp_callin(void)
|
|
{
|
|
int cpu = cpu_now_booting;
|
|
reg_intr_vect_rw_mask vect_mask = {0};
|
|
|
|
/* Initialise the idle task for this CPU */
|
|
atomic_inc(&init_mm.mm_count);
|
|
current->active_mm = &init_mm;
|
|
|
|
/* Set up MMU */
|
|
cris_mmu_init();
|
|
__flush_tlb_all();
|
|
|
|
/* Setup local timer. */
|
|
cris_timer_init();
|
|
|
|
/* Enable IRQ and idle */
|
|
REG_WR(intr_vect, irq_regs[cpu], rw_mask, vect_mask);
|
|
crisv32_unmask_irq(IPI_INTR_VECT);
|
|
crisv32_unmask_irq(TIMER0_INTR_VECT);
|
|
preempt_disable();
|
|
notify_cpu_starting(cpu);
|
|
local_irq_enable();
|
|
|
|
set_cpu_online(cpu, true);
|
|
cpu_startup_entry(CPUHP_ONLINE);
|
|
}
|
|
|
|
/* Stop execution on this CPU.*/
|
|
void stop_this_cpu(void* dummy)
|
|
{
|
|
local_irq_disable();
|
|
asm volatile("halt");
|
|
}
|
|
|
|
/* Other calls */
|
|
void smp_send_stop(void)
|
|
{
|
|
smp_call_function(stop_this_cpu, NULL, 0);
|
|
}
|
|
|
|
int setup_profiling_timer(unsigned int multiplier)
|
|
{
|
|
return -EINVAL;
|
|
}
|
|
|
|
|
|
/* cache_decay_ticks is used by the scheduler to decide if a process
|
|
* is "hot" on one CPU. A higher value means a higher penalty to move
|
|
* a process to another CPU. Our cache is rather small so we report
|
|
* 1 tick.
|
|
*/
|
|
unsigned long cache_decay_ticks = 1;
|
|
|
|
int __cpu_up(unsigned int cpu, struct task_struct *tidle)
|
|
{
|
|
smp_boot_one_cpu(cpu, tidle);
|
|
return cpu_online(cpu) ? 0 : -ENOSYS;
|
|
}
|
|
|
|
void smp_send_reschedule(int cpu)
|
|
{
|
|
cpumask_t cpu_mask;
|
|
cpumask_clear(&cpu_mask);
|
|
cpumask_set_cpu(cpu, &cpu_mask);
|
|
send_ipi(IPI_SCHEDULE, 0, cpu_mask);
|
|
}
|
|
|
|
/* TLB flushing
|
|
*
|
|
* Flush needs to be done on the local CPU and on any other CPU that
|
|
* may have the same mapping. The mm->cpu_vm_mask is used to keep track
|
|
* of which CPUs that a specific process has been executed on.
|
|
*/
|
|
void flush_tlb_common(struct mm_struct* mm, struct vm_area_struct* vma, unsigned long addr)
|
|
{
|
|
unsigned long flags;
|
|
cpumask_t cpu_mask;
|
|
|
|
spin_lock_irqsave(&tlbstate_lock, flags);
|
|
cpu_mask = (mm == FLUSH_ALL ? cpu_all_mask : *mm_cpumask(mm));
|
|
cpumask_clear_cpu(smp_processor_id(), &cpu_mask);
|
|
flush_mm = mm;
|
|
flush_vma = vma;
|
|
flush_addr = addr;
|
|
send_ipi(IPI_FLUSH_TLB, 1, cpu_mask);
|
|
spin_unlock_irqrestore(&tlbstate_lock, flags);
|
|
}
|
|
|
|
void flush_tlb_all(void)
|
|
{
|
|
__flush_tlb_all();
|
|
flush_tlb_common(FLUSH_ALL, FLUSH_ALL, 0);
|
|
}
|
|
|
|
void flush_tlb_mm(struct mm_struct *mm)
|
|
{
|
|
__flush_tlb_mm(mm);
|
|
flush_tlb_common(mm, FLUSH_ALL, 0);
|
|
/* No more mappings in other CPUs */
|
|
cpumask_clear(mm_cpumask(mm));
|
|
cpumask_set_cpu(smp_processor_id(), mm_cpumask(mm));
|
|
}
|
|
|
|
void flush_tlb_page(struct vm_area_struct *vma,
|
|
unsigned long addr)
|
|
{
|
|
__flush_tlb_page(vma, addr);
|
|
flush_tlb_common(vma->vm_mm, vma, addr);
|
|
}
|
|
|
|
/* Inter processor interrupts
|
|
*
|
|
* The IPIs are used for:
|
|
* * Force a schedule on a CPU
|
|
* * FLush TLB on other CPUs
|
|
* * Call a function on other CPUs
|
|
*/
|
|
|
|
int send_ipi(int vector, int wait, cpumask_t cpu_mask)
|
|
{
|
|
int i = 0;
|
|
reg_intr_vect_rw_ipi ipi = REG_RD(intr_vect, irq_regs[i], rw_ipi);
|
|
int ret = 0;
|
|
|
|
/* Calculate CPUs to send to. */
|
|
cpumask_and(&cpu_mask, &cpu_mask, cpu_online_mask);
|
|
|
|
/* Send the IPI. */
|
|
for_each_cpu(i, &cpu_mask)
|
|
{
|
|
ipi.vector |= vector;
|
|
REG_WR(intr_vect, irq_regs[i], rw_ipi, ipi);
|
|
}
|
|
|
|
/* Wait for IPI to finish on other CPUS */
|
|
if (wait) {
|
|
for_each_cpu(i, &cpu_mask) {
|
|
int j;
|
|
for (j = 0 ; j < 1000; j++) {
|
|
ipi = REG_RD(intr_vect, irq_regs[i], rw_ipi);
|
|
if (!ipi.vector)
|
|
break;
|
|
udelay(100);
|
|
}
|
|
|
|
/* Timeout? */
|
|
if (ipi.vector) {
|
|
printk("SMP call timeout from %d to %d\n", smp_processor_id(), i);
|
|
ret = -ETIMEDOUT;
|
|
dump_stack();
|
|
}
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* You must not call this function with disabled interrupts or from a
|
|
* hardware interrupt handler or from a bottom half handler.
|
|
*/
|
|
int smp_call_function(void (*func)(void *info), void *info, int wait)
|
|
{
|
|
cpumask_t cpu_mask;
|
|
struct call_data_struct data;
|
|
int ret;
|
|
|
|
cpumask_setall(&cpu_mask);
|
|
cpumask_clear_cpu(smp_processor_id(), &cpu_mask);
|
|
|
|
WARN_ON(irqs_disabled());
|
|
|
|
data.func = func;
|
|
data.info = info;
|
|
data.wait = wait;
|
|
|
|
spin_lock(&call_lock);
|
|
call_data = &data;
|
|
ret = send_ipi(IPI_CALL, wait, cpu_mask);
|
|
spin_unlock(&call_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
irqreturn_t crisv32_ipi_interrupt(int irq, void *dev_id)
|
|
{
|
|
void (*func) (void *info) = call_data->func;
|
|
void *info = call_data->info;
|
|
reg_intr_vect_rw_ipi ipi;
|
|
|
|
ipi = REG_RD(intr_vect, irq_regs[smp_processor_id()], rw_ipi);
|
|
|
|
if (ipi.vector & IPI_SCHEDULE) {
|
|
scheduler_ipi();
|
|
}
|
|
if (ipi.vector & IPI_CALL) {
|
|
func(info);
|
|
}
|
|
if (ipi.vector & IPI_FLUSH_TLB) {
|
|
if (flush_mm == FLUSH_ALL)
|
|
__flush_tlb_all();
|
|
else if (flush_vma == FLUSH_ALL)
|
|
__flush_tlb_mm(flush_mm);
|
|
else
|
|
__flush_tlb_page(flush_vma, flush_addr);
|
|
}
|
|
|
|
ipi.vector = 0;
|
|
REG_WR(intr_vect, irq_regs[smp_processor_id()], rw_ipi, ipi);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|