mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-29 15:43:59 +08:00
edfaf05c2f
All OMAP IP blocks expect LE data, but CPU may operate in BE mode. Need to use endian neutral functions to read/write h/w registers. I.e instead of __raw_read[lw] and __raw_write[lw] functions code need to use read[lw]_relaxed and write[lw]_relaxed functions. If the first simply reads/writes register, the second will byteswap it if host operates in BE mode. Changes are trivial sed like replacement of __raw_xxx functions with xxx_relaxed variant. Signed-off-by: Victor Kamensky <victor.kamensky@linaro.org> Signed-off-by: Taras Kondratiuk <taras.kondratiuk@linaro.org> Signed-off-by: Tony Lindgren <tony@atomide.com>
309 lines
8.4 KiB
C
309 lines
8.4 KiB
C
/*
|
|
*
|
|
* OMAP SRAM detection and management
|
|
*
|
|
* Copyright (C) 2005 Nokia Corporation
|
|
* Written by Tony Lindgren <tony@atomide.com>
|
|
*
|
|
* Copyright (C) 2009-2012 Texas Instruments
|
|
* Added OMAP4/5 support - Santosh Shilimkar <santosh.shilimkar@ti.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/init.h>
|
|
#include <linux/io.h>
|
|
|
|
#include <asm/fncpy.h>
|
|
#include <asm/tlb.h>
|
|
#include <asm/cacheflush.h>
|
|
|
|
#include <asm/mach/map.h>
|
|
|
|
#include "soc.h"
|
|
#include "iomap.h"
|
|
#include "prm2xxx_3xxx.h"
|
|
#include "sdrc.h"
|
|
#include "sram.h"
|
|
|
|
#define OMAP2_SRAM_PUB_PA (OMAP2_SRAM_PA + 0xf800)
|
|
#define OMAP3_SRAM_PUB_PA (OMAP3_SRAM_PA + 0x8000)
|
|
#ifdef CONFIG_OMAP4_ERRATA_I688
|
|
#define OMAP4_SRAM_PUB_PA OMAP4_SRAM_PA
|
|
#else
|
|
#define OMAP4_SRAM_PUB_PA (OMAP4_SRAM_PA + 0x4000)
|
|
#endif
|
|
#define OMAP5_SRAM_PA 0x40300000
|
|
|
|
#define SRAM_BOOTLOADER_SZ 0x00
|
|
|
|
#define OMAP24XX_VA_REQINFOPERM0 OMAP2_L3_IO_ADDRESS(0x68005048)
|
|
#define OMAP24XX_VA_READPERM0 OMAP2_L3_IO_ADDRESS(0x68005050)
|
|
#define OMAP24XX_VA_WRITEPERM0 OMAP2_L3_IO_ADDRESS(0x68005058)
|
|
|
|
#define OMAP34XX_VA_REQINFOPERM0 OMAP2_L3_IO_ADDRESS(0x68012848)
|
|
#define OMAP34XX_VA_READPERM0 OMAP2_L3_IO_ADDRESS(0x68012850)
|
|
#define OMAP34XX_VA_WRITEPERM0 OMAP2_L3_IO_ADDRESS(0x68012858)
|
|
#define OMAP34XX_VA_ADDR_MATCH2 OMAP2_L3_IO_ADDRESS(0x68012880)
|
|
#define OMAP34XX_VA_SMS_RG_ATT0 OMAP2_L3_IO_ADDRESS(0x6C000048)
|
|
|
|
#define GP_DEVICE 0x300
|
|
|
|
#define ROUND_DOWN(value,boundary) ((value) & (~((boundary)-1)))
|
|
|
|
static unsigned long omap_sram_start;
|
|
static unsigned long omap_sram_skip;
|
|
static unsigned long omap_sram_size;
|
|
|
|
/*
|
|
* Depending on the target RAMFS firewall setup, the public usable amount of
|
|
* SRAM varies. The default accessible size for all device types is 2k. A GP
|
|
* device allows ARM11 but not other initiators for full size. This
|
|
* functionality seems ok until some nice security API happens.
|
|
*/
|
|
static int is_sram_locked(void)
|
|
{
|
|
if (OMAP2_DEVICE_TYPE_GP == omap_type()) {
|
|
/* RAMFW: R/W access to all initiators for all qualifier sets */
|
|
if (cpu_is_omap242x()) {
|
|
writel_relaxed(0xFF, OMAP24XX_VA_REQINFOPERM0); /* all q-vects */
|
|
writel_relaxed(0xCFDE, OMAP24XX_VA_READPERM0); /* all i-read */
|
|
writel_relaxed(0xCFDE, OMAP24XX_VA_WRITEPERM0); /* all i-write */
|
|
}
|
|
if (cpu_is_omap34xx()) {
|
|
writel_relaxed(0xFFFF, OMAP34XX_VA_REQINFOPERM0); /* all q-vects */
|
|
writel_relaxed(0xFFFF, OMAP34XX_VA_READPERM0); /* all i-read */
|
|
writel_relaxed(0xFFFF, OMAP34XX_VA_WRITEPERM0); /* all i-write */
|
|
writel_relaxed(0x0, OMAP34XX_VA_ADDR_MATCH2);
|
|
writel_relaxed(0xFFFFFFFF, OMAP34XX_VA_SMS_RG_ATT0);
|
|
}
|
|
return 0;
|
|
} else
|
|
return 1; /* assume locked with no PPA or security driver */
|
|
}
|
|
|
|
/*
|
|
* The amount of SRAM depends on the core type.
|
|
* Note that we cannot try to test for SRAM here because writes
|
|
* to secure SRAM will hang the system. Also the SRAM is not
|
|
* yet mapped at this point.
|
|
*/
|
|
static void __init omap_detect_sram(void)
|
|
{
|
|
omap_sram_skip = SRAM_BOOTLOADER_SZ;
|
|
if (is_sram_locked()) {
|
|
if (cpu_is_omap34xx()) {
|
|
omap_sram_start = OMAP3_SRAM_PUB_PA;
|
|
if ((omap_type() == OMAP2_DEVICE_TYPE_EMU) ||
|
|
(omap_type() == OMAP2_DEVICE_TYPE_SEC)) {
|
|
omap_sram_size = 0x7000; /* 28K */
|
|
omap_sram_skip += SZ_16K;
|
|
} else {
|
|
omap_sram_size = 0x8000; /* 32K */
|
|
}
|
|
} else if (cpu_is_omap44xx()) {
|
|
omap_sram_start = OMAP4_SRAM_PUB_PA;
|
|
omap_sram_size = 0xa000; /* 40K */
|
|
} else if (soc_is_omap54xx()) {
|
|
omap_sram_start = OMAP5_SRAM_PA;
|
|
omap_sram_size = SZ_128K; /* 128KB */
|
|
} else {
|
|
omap_sram_start = OMAP2_SRAM_PUB_PA;
|
|
omap_sram_size = 0x800; /* 2K */
|
|
}
|
|
} else {
|
|
if (soc_is_am33xx()) {
|
|
omap_sram_start = AM33XX_SRAM_PA;
|
|
omap_sram_size = 0x10000; /* 64K */
|
|
} else if (soc_is_am43xx()) {
|
|
omap_sram_start = AM33XX_SRAM_PA;
|
|
omap_sram_size = SZ_256K;
|
|
} else if (cpu_is_omap34xx()) {
|
|
omap_sram_start = OMAP3_SRAM_PA;
|
|
omap_sram_size = 0x10000; /* 64K */
|
|
} else if (cpu_is_omap44xx()) {
|
|
omap_sram_start = OMAP4_SRAM_PA;
|
|
omap_sram_size = 0xe000; /* 56K */
|
|
} else if (soc_is_omap54xx()) {
|
|
omap_sram_start = OMAP5_SRAM_PA;
|
|
omap_sram_size = SZ_128K; /* 128KB */
|
|
} else {
|
|
omap_sram_start = OMAP2_SRAM_PA;
|
|
if (cpu_is_omap242x())
|
|
omap_sram_size = 0xa0000; /* 640K */
|
|
else if (cpu_is_omap243x())
|
|
omap_sram_size = 0x10000; /* 64K */
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Note that we cannot use ioremap for SRAM, as clock init needs SRAM early.
|
|
*/
|
|
static void __init omap2_map_sram(void)
|
|
{
|
|
int cached = 1;
|
|
|
|
#ifdef CONFIG_OMAP4_ERRATA_I688
|
|
if (cpu_is_omap44xx()) {
|
|
omap_sram_start += PAGE_SIZE;
|
|
omap_sram_size -= SZ_16K;
|
|
}
|
|
#endif
|
|
if (cpu_is_omap34xx()) {
|
|
/*
|
|
* SRAM must be marked as non-cached on OMAP3 since the
|
|
* CORE DPLL M2 divider change code (in SRAM) runs with the
|
|
* SDRAM controller disabled, and if it is marked cached,
|
|
* the ARM may attempt to write cache lines back to SDRAM
|
|
* which will cause the system to hang.
|
|
*/
|
|
cached = 0;
|
|
}
|
|
|
|
omap_map_sram(omap_sram_start, omap_sram_size,
|
|
omap_sram_skip, cached);
|
|
}
|
|
|
|
static void (*_omap2_sram_ddr_init)(u32 *slow_dll_ctrl, u32 fast_dll_ctrl,
|
|
u32 base_cs, u32 force_unlock);
|
|
|
|
void omap2_sram_ddr_init(u32 *slow_dll_ctrl, u32 fast_dll_ctrl,
|
|
u32 base_cs, u32 force_unlock)
|
|
{
|
|
BUG_ON(!_omap2_sram_ddr_init);
|
|
_omap2_sram_ddr_init(slow_dll_ctrl, fast_dll_ctrl,
|
|
base_cs, force_unlock);
|
|
}
|
|
|
|
static void (*_omap2_sram_reprogram_sdrc)(u32 perf_level, u32 dll_val,
|
|
u32 mem_type);
|
|
|
|
void omap2_sram_reprogram_sdrc(u32 perf_level, u32 dll_val, u32 mem_type)
|
|
{
|
|
BUG_ON(!_omap2_sram_reprogram_sdrc);
|
|
_omap2_sram_reprogram_sdrc(perf_level, dll_val, mem_type);
|
|
}
|
|
|
|
static u32 (*_omap2_set_prcm)(u32 dpll_ctrl_val, u32 sdrc_rfr_val, int bypass);
|
|
|
|
u32 omap2_set_prcm(u32 dpll_ctrl_val, u32 sdrc_rfr_val, int bypass)
|
|
{
|
|
BUG_ON(!_omap2_set_prcm);
|
|
return _omap2_set_prcm(dpll_ctrl_val, sdrc_rfr_val, bypass);
|
|
}
|
|
|
|
#ifdef CONFIG_SOC_OMAP2420
|
|
static int __init omap242x_sram_init(void)
|
|
{
|
|
_omap2_sram_ddr_init = omap_sram_push(omap242x_sram_ddr_init,
|
|
omap242x_sram_ddr_init_sz);
|
|
|
|
_omap2_sram_reprogram_sdrc = omap_sram_push(omap242x_sram_reprogram_sdrc,
|
|
omap242x_sram_reprogram_sdrc_sz);
|
|
|
|
_omap2_set_prcm = omap_sram_push(omap242x_sram_set_prcm,
|
|
omap242x_sram_set_prcm_sz);
|
|
|
|
return 0;
|
|
}
|
|
#else
|
|
static inline int omap242x_sram_init(void)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
#ifdef CONFIG_SOC_OMAP2430
|
|
static int __init omap243x_sram_init(void)
|
|
{
|
|
_omap2_sram_ddr_init = omap_sram_push(omap243x_sram_ddr_init,
|
|
omap243x_sram_ddr_init_sz);
|
|
|
|
_omap2_sram_reprogram_sdrc = omap_sram_push(omap243x_sram_reprogram_sdrc,
|
|
omap243x_sram_reprogram_sdrc_sz);
|
|
|
|
_omap2_set_prcm = omap_sram_push(omap243x_sram_set_prcm,
|
|
omap243x_sram_set_prcm_sz);
|
|
|
|
return 0;
|
|
}
|
|
#else
|
|
static inline int omap243x_sram_init(void)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARCH_OMAP3
|
|
|
|
static u32 (*_omap3_sram_configure_core_dpll)(
|
|
u32 m2, u32 unlock_dll, u32 f, u32 inc,
|
|
u32 sdrc_rfr_ctrl_0, u32 sdrc_actim_ctrl_a_0,
|
|
u32 sdrc_actim_ctrl_b_0, u32 sdrc_mr_0,
|
|
u32 sdrc_rfr_ctrl_1, u32 sdrc_actim_ctrl_a_1,
|
|
u32 sdrc_actim_ctrl_b_1, u32 sdrc_mr_1);
|
|
|
|
u32 omap3_configure_core_dpll(u32 m2, u32 unlock_dll, u32 f, u32 inc,
|
|
u32 sdrc_rfr_ctrl_0, u32 sdrc_actim_ctrl_a_0,
|
|
u32 sdrc_actim_ctrl_b_0, u32 sdrc_mr_0,
|
|
u32 sdrc_rfr_ctrl_1, u32 sdrc_actim_ctrl_a_1,
|
|
u32 sdrc_actim_ctrl_b_1, u32 sdrc_mr_1)
|
|
{
|
|
BUG_ON(!_omap3_sram_configure_core_dpll);
|
|
return _omap3_sram_configure_core_dpll(
|
|
m2, unlock_dll, f, inc,
|
|
sdrc_rfr_ctrl_0, sdrc_actim_ctrl_a_0,
|
|
sdrc_actim_ctrl_b_0, sdrc_mr_0,
|
|
sdrc_rfr_ctrl_1, sdrc_actim_ctrl_a_1,
|
|
sdrc_actim_ctrl_b_1, sdrc_mr_1);
|
|
}
|
|
|
|
void omap3_sram_restore_context(void)
|
|
{
|
|
omap_sram_reset();
|
|
|
|
_omap3_sram_configure_core_dpll =
|
|
omap_sram_push(omap3_sram_configure_core_dpll,
|
|
omap3_sram_configure_core_dpll_sz);
|
|
omap_push_sram_idle();
|
|
}
|
|
|
|
static inline int omap34xx_sram_init(void)
|
|
{
|
|
omap3_sram_restore_context();
|
|
return 0;
|
|
}
|
|
#else
|
|
static inline int omap34xx_sram_init(void)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif /* CONFIG_ARCH_OMAP3 */
|
|
|
|
static inline int am33xx_sram_init(void)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
int __init omap_sram_init(void)
|
|
{
|
|
omap_detect_sram();
|
|
omap2_map_sram();
|
|
|
|
if (cpu_is_omap242x())
|
|
omap242x_sram_init();
|
|
else if (cpu_is_omap2430())
|
|
omap243x_sram_init();
|
|
else if (soc_is_am33xx())
|
|
am33xx_sram_init();
|
|
else if (cpu_is_omap34xx())
|
|
omap34xx_sram_init();
|
|
|
|
return 0;
|
|
}
|