mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-25 05:04:09 +08:00
02b4e2756e
All ARMv5 and older CPUs invalidate their caches in the early assembly setup function, prior to enabling the MMU. This is because the L1 cache should not contain any data relevant to the execution of the kernel at this point; all data should have been flushed out to memory. This requirement should also be true for ARMv6 and ARMv7 CPUs - indeed, these typically do not search their caches when caching is disabled (as it needs to be when the MMU is disabled) so this change should be safe. ARMv7 allows there to be CPUs which search their caches while caching is disabled, and it's permitted that the cache is uninitialised at boot; for these, the architecture reference manual requires that an implementation specific code sequence is used immediately after reset to ensure that the cache is placed into a sane state. Such functionality is definitely outside the remit of the Linux kernel, and must be done by the SoC's firmware before _any_ CPU gets to the Linux kernel. Changing the data cache clean+invalidate to a mere invalidate allows us to get rid of a lot of platform specific hacks around this issue for their secondary CPU bringup paths - some of which were buggy. Reviewed-by: Florian Fainelli <f.fainelli@gmail.com> Tested-by: Florian Fainelli <f.fainelli@gmail.com> Tested-by: Heiko Stuebner <heiko@sntech.de> Tested-by: Dinh Nguyen <dinguyen@opensource.altera.com> Acked-by: Sebastian Hesselbarth <sebastian.hesselbarth@gmail.com> Tested-by: Sebastian Hesselbarth <sebastian.hesselbarth@gmail.com> Acked-by: Shawn Guo <shawn.guo@linaro.org> Tested-by: Thierry Reding <treding@nvidia.com> Acked-by: Thierry Reding <treding@nvidia.com> Tested-by: Geert Uytterhoeven <geert+renesas@glider.be> Tested-by: Michal Simek <michal.simek@xilinx.com> Tested-by: Wei Xu <xuwei5@hisilicon.com> Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
239 lines
5.3 KiB
C
239 lines
5.3 KiB
C
/*
|
|
* SMP support for SoCs with APMU
|
|
*
|
|
* Copyright (C) 2014 Renesas Electronics Corporation
|
|
* Copyright (C) 2013 Magnus Damm
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*/
|
|
#include <linux/cpu_pm.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/init.h>
|
|
#include <linux/io.h>
|
|
#include <linux/ioport.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/suspend.h>
|
|
#include <linux/threads.h>
|
|
#include <asm/cacheflush.h>
|
|
#include <asm/cp15.h>
|
|
#include <asm/proc-fns.h>
|
|
#include <asm/smp_plat.h>
|
|
#include <asm/suspend.h>
|
|
#include "common.h"
|
|
#include "platsmp-apmu.h"
|
|
|
|
static struct {
|
|
void __iomem *iomem;
|
|
int bit;
|
|
} apmu_cpus[NR_CPUS];
|
|
|
|
#define WUPCR_OFFS 0x10
|
|
#define PSTR_OFFS 0x40
|
|
#define CPUNCR_OFFS(n) (0x100 + (0x10 * (n)))
|
|
|
|
static int __maybe_unused apmu_power_on(void __iomem *p, int bit)
|
|
{
|
|
/* request power on */
|
|
writel_relaxed(BIT(bit), p + WUPCR_OFFS);
|
|
|
|
/* wait for APMU to finish */
|
|
while (readl_relaxed(p + WUPCR_OFFS) != 0)
|
|
;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int apmu_power_off(void __iomem *p, int bit)
|
|
{
|
|
/* request Core Standby for next WFI */
|
|
writel_relaxed(3, p + CPUNCR_OFFS(bit));
|
|
return 0;
|
|
}
|
|
|
|
static int __maybe_unused apmu_power_off_poll(void __iomem *p, int bit)
|
|
{
|
|
int k;
|
|
|
|
for (k = 0; k < 1000; k++) {
|
|
if (((readl_relaxed(p + PSTR_OFFS) >> (bit * 4)) & 0x03) == 3)
|
|
return 1;
|
|
|
|
mdelay(1);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int apmu_wrap(int cpu, int (*fn)(void __iomem *p, int cpu))
|
|
{
|
|
void __iomem *p = apmu_cpus[cpu].iomem;
|
|
|
|
return p ? fn(p, apmu_cpus[cpu].bit) : -EINVAL;
|
|
}
|
|
|
|
static void apmu_init_cpu(struct resource *res, int cpu, int bit)
|
|
{
|
|
if ((cpu >= ARRAY_SIZE(apmu_cpus)) || apmu_cpus[cpu].iomem)
|
|
return;
|
|
|
|
apmu_cpus[cpu].iomem = ioremap_nocache(res->start, resource_size(res));
|
|
apmu_cpus[cpu].bit = bit;
|
|
|
|
pr_debug("apmu ioremap %d %d %pr\n", cpu, bit, res);
|
|
}
|
|
|
|
static void apmu_parse_cfg(void (*fn)(struct resource *res, int cpu, int bit),
|
|
struct rcar_apmu_config *apmu_config, int num)
|
|
{
|
|
u32 id;
|
|
int k;
|
|
int bit, index;
|
|
bool is_allowed;
|
|
|
|
for (k = 0; k < num; k++) {
|
|
/* only enable the cluster that includes the boot CPU */
|
|
is_allowed = false;
|
|
for (bit = 0; bit < ARRAY_SIZE(apmu_config[k].cpus); bit++) {
|
|
id = apmu_config[k].cpus[bit];
|
|
if (id >= 0) {
|
|
if (id == cpu_logical_map(0))
|
|
is_allowed = true;
|
|
}
|
|
}
|
|
if (!is_allowed)
|
|
continue;
|
|
|
|
for (bit = 0; bit < ARRAY_SIZE(apmu_config[k].cpus); bit++) {
|
|
id = apmu_config[k].cpus[bit];
|
|
if (id >= 0) {
|
|
index = get_logical_index(id);
|
|
if (index >= 0)
|
|
fn(&apmu_config[k].iomem, index, bit);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void __init shmobile_smp_apmu_prepare_cpus(unsigned int max_cpus,
|
|
struct rcar_apmu_config *apmu_config,
|
|
int num)
|
|
{
|
|
/* install boot code shared by all CPUs */
|
|
shmobile_boot_fn = virt_to_phys(shmobile_smp_boot);
|
|
shmobile_boot_arg = MPIDR_HWID_BITMASK;
|
|
|
|
/* perform per-cpu setup */
|
|
apmu_parse_cfg(apmu_init_cpu, apmu_config, num);
|
|
}
|
|
|
|
#ifdef CONFIG_SMP
|
|
int shmobile_smp_apmu_boot_secondary(unsigned int cpu, struct task_struct *idle)
|
|
{
|
|
/* For this particular CPU register boot vector */
|
|
shmobile_smp_hook(cpu, virt_to_phys(secondary_startup), 0);
|
|
|
|
return apmu_wrap(cpu, apmu_power_on);
|
|
}
|
|
#endif
|
|
|
|
#if defined(CONFIG_HOTPLUG_CPU) || defined(CONFIG_SUSPEND)
|
|
/* nicked from arch/arm/mach-exynos/hotplug.c */
|
|
static inline void cpu_enter_lowpower_a15(void)
|
|
{
|
|
unsigned int v;
|
|
|
|
asm volatile(
|
|
" mrc p15, 0, %0, c1, c0, 0\n"
|
|
" bic %0, %0, %1\n"
|
|
" mcr p15, 0, %0, c1, c0, 0\n"
|
|
: "=&r" (v)
|
|
: "Ir" (CR_C)
|
|
: "cc");
|
|
|
|
flush_cache_louis();
|
|
|
|
asm volatile(
|
|
/*
|
|
* Turn off coherency
|
|
*/
|
|
" mrc p15, 0, %0, c1, c0, 1\n"
|
|
" bic %0, %0, %1\n"
|
|
" mcr p15, 0, %0, c1, c0, 1\n"
|
|
: "=&r" (v)
|
|
: "Ir" (0x40)
|
|
: "cc");
|
|
|
|
isb();
|
|
dsb();
|
|
}
|
|
|
|
void shmobile_smp_apmu_cpu_shutdown(unsigned int cpu)
|
|
{
|
|
|
|
/* Select next sleep mode using the APMU */
|
|
apmu_wrap(cpu, apmu_power_off);
|
|
|
|
/* Do ARM specific CPU shutdown */
|
|
cpu_enter_lowpower_a15();
|
|
}
|
|
|
|
static inline void cpu_leave_lowpower(void)
|
|
{
|
|
unsigned int v;
|
|
|
|
asm volatile("mrc p15, 0, %0, c1, c0, 0\n"
|
|
" orr %0, %0, %1\n"
|
|
" mcr p15, 0, %0, c1, c0, 0\n"
|
|
" mrc p15, 0, %0, c1, c0, 1\n"
|
|
" orr %0, %0, %2\n"
|
|
" mcr p15, 0, %0, c1, c0, 1\n"
|
|
: "=&r" (v)
|
|
: "Ir" (CR_C), "Ir" (0x40)
|
|
: "cc");
|
|
}
|
|
#endif
|
|
|
|
#if defined(CONFIG_HOTPLUG_CPU)
|
|
void shmobile_smp_apmu_cpu_die(unsigned int cpu)
|
|
{
|
|
/* For this particular CPU deregister boot vector */
|
|
shmobile_smp_hook(cpu, 0, 0);
|
|
|
|
/* Shutdown CPU core */
|
|
shmobile_smp_apmu_cpu_shutdown(cpu);
|
|
|
|
/* jump to shared mach-shmobile sleep / reset code */
|
|
shmobile_smp_sleep();
|
|
}
|
|
|
|
int shmobile_smp_apmu_cpu_kill(unsigned int cpu)
|
|
{
|
|
return apmu_wrap(cpu, apmu_power_off_poll);
|
|
}
|
|
#endif
|
|
|
|
#if defined(CONFIG_SUSPEND)
|
|
static int shmobile_smp_apmu_do_suspend(unsigned long cpu)
|
|
{
|
|
shmobile_smp_hook(cpu, virt_to_phys(cpu_resume), 0);
|
|
shmobile_smp_apmu_cpu_shutdown(cpu);
|
|
cpu_do_idle(); /* WFI selects Core Standby */
|
|
return 1;
|
|
}
|
|
|
|
static int shmobile_smp_apmu_enter_suspend(suspend_state_t state)
|
|
{
|
|
cpu_suspend(smp_processor_id(), shmobile_smp_apmu_do_suspend);
|
|
cpu_leave_lowpower();
|
|
return 0;
|
|
}
|
|
|
|
void __init shmobile_smp_apmu_suspend_init(void)
|
|
{
|
|
shmobile_suspend_ops.enter = shmobile_smp_apmu_enter_suspend;
|
|
}
|
|
#endif
|