2011-02-14 14:33:10 +07:00
|
|
|
/* linux/arch/arm/mach-exynos4/platsmp.c
|
2010-07-26 19:08:52 +07:00
|
|
|
*
|
2011-02-14 14:33:10 +07:00
|
|
|
* Copyright (c) 2010-2011 Samsung Electronics Co., Ltd.
|
|
|
|
* http://www.samsung.com
|
2010-07-26 19:08:52 +07:00
|
|
|
*
|
|
|
|
* Cloned from linux/arch/arm/mach-vexpress/platsmp.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 2002 ARM Ltd.
|
|
|
|
* All Rights Reserved
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/errno.h>
|
|
|
|
#include <linux/delay.h>
|
|
|
|
#include <linux/device.h>
|
|
|
|
#include <linux/jiffies.h>
|
|
|
|
#include <linux/smp.h>
|
|
|
|
#include <linux/io.h>
|
|
|
|
|
|
|
|
#include <asm/cacheflush.h>
|
2011-04-03 19:01:30 +07:00
|
|
|
#include <asm/hardware/gic.h>
|
2012-01-20 18:01:12 +07:00
|
|
|
#include <asm/smp_plat.h>
|
2010-07-26 19:08:52 +07:00
|
|
|
#include <asm/smp_scu.h>
|
|
|
|
|
|
|
|
#include <mach/hardware.h>
|
|
|
|
#include <mach/regs-clock.h>
|
2011-07-16 11:39:09 +07:00
|
|
|
#include <mach/regs-pmu.h>
|
2010-07-26 19:08:52 +07:00
|
|
|
|
2011-08-20 11:41:21 +07:00
|
|
|
#include <plat/cpu.h>
|
|
|
|
|
2011-02-14 14:33:10 +07:00
|
|
|
extern void exynos4_secondary_startup(void);
|
2010-07-26 19:08:52 +07:00
|
|
|
|
2011-08-20 11:41:21 +07:00
|
|
|
#define CPU1_BOOT_REG (samsung_rev() == EXYNOS4210_REV_1_1 ? \
|
|
|
|
S5P_INFORM5 : S5P_VA_SYSRAM)
|
2011-07-16 11:39:09 +07:00
|
|
|
|
2010-07-26 19:08:52 +07:00
|
|
|
/*
|
|
|
|
* control for which core is the next to come out of the secondary
|
|
|
|
* boot "holding pen"
|
|
|
|
*/
|
|
|
|
|
|
|
|
volatile int __cpuinitdata pen_release = -1;
|
|
|
|
|
ARM: Fix subtle race in CPU pen_release hotplug code
There is a subtle race in the CPU hotplug code, where a CPU which has
been offlined can online itself before being requested, which results
in things going astray on the next online/offline cycle.
What happens in the normal online/offline/online cycle is:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads -1
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
However, as the write of -1 of pen_release is not fully flushed back to
memory, and the checking of pen_release is done with caches disabled,
this allows CPU3 the opportunity to read the old value of pen_release:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads 3
starts boot
pen_release = -1
requests boot of CPU3
pen_release = 3
flush cache line
Fix this by grouping the write of pen_release along with its cache line
flushing code to ensure that any update to pen_release is always pushed
out to physical memory.
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
2010-12-18 17:53:12 +07:00
|
|
|
/*
|
|
|
|
* Write pen_release in a way that is guaranteed to be visible to all
|
|
|
|
* observers, irrespective of whether they're taking part in coherency
|
|
|
|
* or not. This is necessary for the hotplug code to work reliably.
|
|
|
|
*/
|
|
|
|
static void write_pen_release(int val)
|
|
|
|
{
|
|
|
|
pen_release = val;
|
|
|
|
smp_wmb();
|
|
|
|
__cpuc_flush_dcache_area((void *)&pen_release, sizeof(pen_release));
|
|
|
|
outer_clean_range(__pa(&pen_release), __pa(&pen_release + 1));
|
|
|
|
}
|
|
|
|
|
2010-07-26 19:08:52 +07:00
|
|
|
static void __iomem *scu_base_addr(void)
|
|
|
|
{
|
|
|
|
return (void __iomem *)(S5P_VA_SCU);
|
|
|
|
}
|
|
|
|
|
|
|
|
static DEFINE_SPINLOCK(boot_lock);
|
|
|
|
|
|
|
|
void __cpuinit platform_secondary_init(unsigned int cpu)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* if any interrupts are already enabled for the primary
|
|
|
|
* core (e.g. timer irq), then they will not have been enabled
|
|
|
|
* for us: do so
|
|
|
|
*/
|
2011-11-12 23:09:49 +07:00
|
|
|
gic_secondary_init(0);
|
2010-07-26 19:08:52 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* let the primary processor know we're out of the
|
|
|
|
* pen, then head off into the C entry point
|
|
|
|
*/
|
ARM: Fix subtle race in CPU pen_release hotplug code
There is a subtle race in the CPU hotplug code, where a CPU which has
been offlined can online itself before being requested, which results
in things going astray on the next online/offline cycle.
What happens in the normal online/offline/online cycle is:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads -1
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
However, as the write of -1 of pen_release is not fully flushed back to
memory, and the checking of pen_release is done with caches disabled,
this allows CPU3 the opportunity to read the old value of pen_release:
CPU0 CPU3
requests boot of CPU3
pen_release = 3
flush cache line
checks pen_release, reads 3
starts boot
pen_release = -1
... requests CPU3 offline ...
... dies ...
checks pen_release, reads 3
starts boot
pen_release = -1
requests boot of CPU3
pen_release = 3
flush cache line
Fix this by grouping the write of pen_release along with its cache line
flushing code to ensure that any update to pen_release is always pushed
out to physical memory.
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
2010-12-18 17:53:12 +07:00
|
|
|
write_pen_release(-1);
|
2010-07-26 19:08:52 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Synchronise with the boot thread.
|
|
|
|
*/
|
|
|
|
spin_lock(&boot_lock);
|
|
|
|
spin_unlock(&boot_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
int __cpuinit boot_secondary(unsigned int cpu, struct task_struct *idle)
|
|
|
|
{
|
|
|
|
unsigned long timeout;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Set synchronisation state between this boot processor
|
|
|
|
* and the secondary one
|
|
|
|
*/
|
|
|
|
spin_lock(&boot_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The secondary processor is waiting to be released from
|
|
|
|
* the holding pen - release it, then wait for it to flag
|
|
|
|
* that it has been released by resetting pen_release.
|
|
|
|
*
|
|
|
|
* Note that "pen_release" is the hardware CPU ID, whereas
|
|
|
|
* "cpu" is Linux's internal ID.
|
|
|
|
*/
|
2011-08-09 17:29:19 +07:00
|
|
|
write_pen_release(cpu_logical_map(cpu));
|
2010-07-26 19:08:52 +07:00
|
|
|
|
2011-07-16 11:39:09 +07:00
|
|
|
if (!(__raw_readl(S5P_ARM_CORE1_STATUS) & S5P_CORE_LOCAL_PWR_EN)) {
|
|
|
|
__raw_writel(S5P_CORE_LOCAL_PWR_EN,
|
|
|
|
S5P_ARM_CORE1_CONFIGURATION);
|
|
|
|
|
|
|
|
timeout = 10;
|
|
|
|
|
|
|
|
/* wait max 10 ms until cpu1 is on */
|
|
|
|
while ((__raw_readl(S5P_ARM_CORE1_STATUS)
|
|
|
|
& S5P_CORE_LOCAL_PWR_EN) != S5P_CORE_LOCAL_PWR_EN) {
|
|
|
|
if (timeout-- == 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
mdelay(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (timeout == 0) {
|
|
|
|
printk(KERN_ERR "cpu1 power enable failed");
|
|
|
|
spin_unlock(&boot_lock);
|
|
|
|
return -ETIMEDOUT;
|
|
|
|
}
|
|
|
|
}
|
2010-07-26 19:08:52 +07:00
|
|
|
/*
|
|
|
|
* Send the secondary CPU a soft interrupt, thereby causing
|
|
|
|
* the boot monitor to read the system wide flags register,
|
|
|
|
* and branch to the address found there.
|
|
|
|
*/
|
|
|
|
|
|
|
|
timeout = jiffies + (1 * HZ);
|
|
|
|
while (time_before(jiffies, timeout)) {
|
|
|
|
smp_rmb();
|
2011-07-16 11:39:09 +07:00
|
|
|
|
2012-01-10 04:39:19 +07:00
|
|
|
__raw_writel(virt_to_phys(exynos4_secondary_startup),
|
2011-07-16 11:39:09 +07:00
|
|
|
CPU1_BOOT_REG);
|
|
|
|
gic_raise_softirq(cpumask_of(cpu), 1);
|
|
|
|
|
2010-07-26 19:08:52 +07:00
|
|
|
if (pen_release == -1)
|
|
|
|
break;
|
|
|
|
|
|
|
|
udelay(10);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* now the secondary core is starting up let it run its
|
|
|
|
* calibrations, then wait for it to finish
|
|
|
|
*/
|
|
|
|
spin_unlock(&boot_lock);
|
|
|
|
|
|
|
|
return pen_release != -1 ? -ENOSYS : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Initialise the CPU possible map early - this describes the CPUs
|
|
|
|
* which may be present or become present in the system.
|
|
|
|
*/
|
|
|
|
|
|
|
|
void __init smp_init_cpus(void)
|
|
|
|
{
|
|
|
|
void __iomem *scu_base = scu_base_addr();
|
|
|
|
unsigned int i, ncores;
|
|
|
|
|
|
|
|
ncores = scu_base ? scu_get_core_count(scu_base) : 1;
|
|
|
|
|
|
|
|
/* sanity check */
|
2011-10-21 04:04:18 +07:00
|
|
|
if (ncores > nr_cpu_ids) {
|
|
|
|
pr_warn("SMP: %u cores greater than maximum (%u), clipping\n",
|
|
|
|
ncores, nr_cpu_ids);
|
|
|
|
ncores = nr_cpu_ids;
|
2010-07-26 19:08:52 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < ncores; i++)
|
|
|
|
set_cpu_possible(i, true);
|
2011-04-03 19:01:30 +07:00
|
|
|
|
|
|
|
set_smp_cross_call(gic_raise_softirq);
|
2010-07-26 19:08:52 +07:00
|
|
|
}
|
|
|
|
|
2010-12-03 18:09:48 +07:00
|
|
|
void __init platform_smp_prepare_cpus(unsigned int max_cpus)
|
2010-07-26 19:08:52 +07:00
|
|
|
{
|
|
|
|
|
2010-12-03 18:09:48 +07:00
|
|
|
scu_enable(scu_base_addr());
|
|
|
|
|
2010-07-26 19:08:52 +07:00
|
|
|
/*
|
2010-12-03 18:09:48 +07:00
|
|
|
* Write the address of secondary startup into the
|
|
|
|
* system-wide flags register. The boot monitor waits
|
|
|
|
* until it receives a soft interrupt, and then the
|
|
|
|
* secondary CPU branches to this address.
|
2010-07-26 19:08:52 +07:00
|
|
|
*/
|
2012-01-10 04:39:19 +07:00
|
|
|
__raw_writel(virt_to_phys(exynos4_secondary_startup),
|
2011-08-20 11:41:21 +07:00
|
|
|
CPU1_BOOT_REG);
|
2010-07-26 19:08:52 +07:00
|
|
|
}
|