mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-15 00:04:15 +08:00
b4d20eff64
Instead of defining all cache flush operations with an assembly macro in proc-macros.S, provide an explicit struct cpu_cache_fns for each CPU cache type in mm/cache.c. As a side effect from rewriting the vtables in C, we can avoid the aliasing for the "louis" cache callback, instead we can just assign the NN_flush_kern_cache_all() function to the louis callback in the C vtable. As the louis cache callback is called explicitly (not through the vtable) if we only have one type of cache support compiled in, we need an ifdef quirk for this in the !MULTI_CACHE case. Feroceon and XScale have some dma mapping quirk, in this case we can just define two structs and assign all but one callback to the main implementation; since each of them invoked define_cache_functions twice they require MULTI_CACHE by definition so the compiled-in shortcut is not used on these variants. Tested-by: Kees Cook <keescook@chromium.org> Reviewed-by: Sami Tolvanen <samitolvanen@google.com> Signed-off-by: Linus Walleij <linus.walleij@linaro.org> Signed-off-by: Russell King (Oracle) <rmk+kernel@armlinux.org.uk>
664 lines
27 KiB
C
664 lines
27 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* This file defines C prototypes for the low-level cache assembly functions
|
|
* and populates a vtable for each selected ARM CPU cache type.
|
|
*/
|
|
|
|
#include <linux/types.h>
|
|
#include <asm/cacheflush.h>
|
|
|
|
#ifdef CONFIG_CPU_CACHE_V4
|
|
void v4_flush_icache_all(void);
|
|
void v4_flush_kern_cache_all(void);
|
|
void v4_flush_user_cache_all(void);
|
|
void v4_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void v4_coherent_kern_range(unsigned long, unsigned long);
|
|
int v4_coherent_user_range(unsigned long, unsigned long);
|
|
void v4_flush_kern_dcache_area(void *, size_t);
|
|
void v4_dma_map_area(const void *, size_t, int);
|
|
void v4_dma_unmap_area(const void *, size_t, int);
|
|
void v4_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns v4_cache_fns __initconst = {
|
|
.flush_icache_all = v4_flush_icache_all,
|
|
.flush_kern_all = v4_flush_kern_cache_all,
|
|
.flush_kern_louis = v4_flush_kern_cache_all,
|
|
.flush_user_all = v4_flush_user_cache_all,
|
|
.flush_user_range = v4_flush_user_cache_range,
|
|
.coherent_kern_range = v4_coherent_kern_range,
|
|
.coherent_user_range = v4_coherent_user_range,
|
|
.flush_kern_dcache_area = v4_flush_kern_dcache_area,
|
|
.dma_map_area = v4_dma_map_area,
|
|
.dma_unmap_area = v4_dma_unmap_area,
|
|
.dma_flush_range = v4_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
/* V4 write-back cache "V4WB" */
|
|
#ifdef CONFIG_CPU_CACHE_V4WB
|
|
void v4wb_flush_icache_all(void);
|
|
void v4wb_flush_kern_cache_all(void);
|
|
void v4wb_flush_user_cache_all(void);
|
|
void v4wb_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void v4wb_coherent_kern_range(unsigned long, unsigned long);
|
|
int v4wb_coherent_user_range(unsigned long, unsigned long);
|
|
void v4wb_flush_kern_dcache_area(void *, size_t);
|
|
void v4wb_dma_map_area(const void *, size_t, int);
|
|
void v4wb_dma_unmap_area(const void *, size_t, int);
|
|
void v4wb_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns v4wb_cache_fns __initconst = {
|
|
.flush_icache_all = v4wb_flush_icache_all,
|
|
.flush_kern_all = v4wb_flush_kern_cache_all,
|
|
.flush_kern_louis = v4wb_flush_kern_cache_all,
|
|
.flush_user_all = v4wb_flush_user_cache_all,
|
|
.flush_user_range = v4wb_flush_user_cache_range,
|
|
.coherent_kern_range = v4wb_coherent_kern_range,
|
|
.coherent_user_range = v4wb_coherent_user_range,
|
|
.flush_kern_dcache_area = v4wb_flush_kern_dcache_area,
|
|
.dma_map_area = v4wb_dma_map_area,
|
|
.dma_unmap_area = v4wb_dma_unmap_area,
|
|
.dma_flush_range = v4wb_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
/* V4 write-through cache "V4WT" */
|
|
#ifdef CONFIG_CPU_CACHE_V4WT
|
|
void v4wt_flush_icache_all(void);
|
|
void v4wt_flush_kern_cache_all(void);
|
|
void v4wt_flush_user_cache_all(void);
|
|
void v4wt_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void v4wt_coherent_kern_range(unsigned long, unsigned long);
|
|
int v4wt_coherent_user_range(unsigned long, unsigned long);
|
|
void v4wt_flush_kern_dcache_area(void *, size_t);
|
|
void v4wt_dma_map_area(const void *, size_t, int);
|
|
void v4wt_dma_unmap_area(const void *, size_t, int);
|
|
void v4wt_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns v4wt_cache_fns __initconst = {
|
|
.flush_icache_all = v4wt_flush_icache_all,
|
|
.flush_kern_all = v4wt_flush_kern_cache_all,
|
|
.flush_kern_louis = v4wt_flush_kern_cache_all,
|
|
.flush_user_all = v4wt_flush_user_cache_all,
|
|
.flush_user_range = v4wt_flush_user_cache_range,
|
|
.coherent_kern_range = v4wt_coherent_kern_range,
|
|
.coherent_user_range = v4wt_coherent_user_range,
|
|
.flush_kern_dcache_area = v4wt_flush_kern_dcache_area,
|
|
.dma_map_area = v4wt_dma_map_area,
|
|
.dma_unmap_area = v4wt_dma_unmap_area,
|
|
.dma_flush_range = v4wt_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
/* Faraday FA526 cache */
|
|
#ifdef CONFIG_CPU_CACHE_FA
|
|
void fa_flush_icache_all(void);
|
|
void fa_flush_kern_cache_all(void);
|
|
void fa_flush_user_cache_all(void);
|
|
void fa_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void fa_coherent_kern_range(unsigned long, unsigned long);
|
|
int fa_coherent_user_range(unsigned long, unsigned long);
|
|
void fa_flush_kern_dcache_area(void *, size_t);
|
|
void fa_dma_map_area(const void *, size_t, int);
|
|
void fa_dma_unmap_area(const void *, size_t, int);
|
|
void fa_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns fa_cache_fns __initconst = {
|
|
.flush_icache_all = fa_flush_icache_all,
|
|
.flush_kern_all = fa_flush_kern_cache_all,
|
|
.flush_kern_louis = fa_flush_kern_cache_all,
|
|
.flush_user_all = fa_flush_user_cache_all,
|
|
.flush_user_range = fa_flush_user_cache_range,
|
|
.coherent_kern_range = fa_coherent_kern_range,
|
|
.coherent_user_range = fa_coherent_user_range,
|
|
.flush_kern_dcache_area = fa_flush_kern_dcache_area,
|
|
.dma_map_area = fa_dma_map_area,
|
|
.dma_unmap_area = fa_dma_unmap_area,
|
|
.dma_flush_range = fa_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_CACHE_V6
|
|
void v6_flush_icache_all(void);
|
|
void v6_flush_kern_cache_all(void);
|
|
void v6_flush_user_cache_all(void);
|
|
void v6_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void v6_coherent_kern_range(unsigned long, unsigned long);
|
|
int v6_coherent_user_range(unsigned long, unsigned long);
|
|
void v6_flush_kern_dcache_area(void *, size_t);
|
|
void v6_dma_map_area(const void *, size_t, int);
|
|
void v6_dma_unmap_area(const void *, size_t, int);
|
|
void v6_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns v6_cache_fns __initconst = {
|
|
.flush_icache_all = v6_flush_icache_all,
|
|
.flush_kern_all = v6_flush_kern_cache_all,
|
|
.flush_kern_louis = v6_flush_kern_cache_all,
|
|
.flush_user_all = v6_flush_user_cache_all,
|
|
.flush_user_range = v6_flush_user_cache_range,
|
|
.coherent_kern_range = v6_coherent_kern_range,
|
|
.coherent_user_range = v6_coherent_user_range,
|
|
.flush_kern_dcache_area = v6_flush_kern_dcache_area,
|
|
.dma_map_area = v6_dma_map_area,
|
|
.dma_unmap_area = v6_dma_unmap_area,
|
|
.dma_flush_range = v6_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_CACHE_V7
|
|
void v7_flush_icache_all(void);
|
|
void v7_flush_kern_cache_all(void);
|
|
void v7_flush_kern_cache_louis(void);
|
|
void v7_flush_user_cache_all(void);
|
|
void v7_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void v7_coherent_kern_range(unsigned long, unsigned long);
|
|
int v7_coherent_user_range(unsigned long, unsigned long);
|
|
void v7_flush_kern_dcache_area(void *, size_t);
|
|
void v7_dma_map_area(const void *, size_t, int);
|
|
void v7_dma_unmap_area(const void *, size_t, int);
|
|
void v7_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns v7_cache_fns __initconst = {
|
|
.flush_icache_all = v7_flush_icache_all,
|
|
.flush_kern_all = v7_flush_kern_cache_all,
|
|
.flush_kern_louis = v7_flush_kern_cache_louis,
|
|
.flush_user_all = v7_flush_user_cache_all,
|
|
.flush_user_range = v7_flush_user_cache_range,
|
|
.coherent_kern_range = v7_coherent_kern_range,
|
|
.coherent_user_range = v7_coherent_user_range,
|
|
.flush_kern_dcache_area = v7_flush_kern_dcache_area,
|
|
.dma_map_area = v7_dma_map_area,
|
|
.dma_unmap_area = v7_dma_unmap_area,
|
|
.dma_flush_range = v7_dma_flush_range,
|
|
};
|
|
|
|
/* Special quirky cache flush function for Broadcom B15 v7 caches */
|
|
void b15_flush_kern_cache_all(void);
|
|
|
|
struct cpu_cache_fns b15_cache_fns __initconst = {
|
|
.flush_icache_all = v7_flush_icache_all,
|
|
#ifdef CONFIG_CACHE_B15_RAC
|
|
.flush_kern_all = b15_flush_kern_cache_all,
|
|
#else
|
|
.flush_kern_all = v7_flush_kern_cache_all,
|
|
#endif
|
|
.flush_kern_louis = v7_flush_kern_cache_louis,
|
|
.flush_user_all = v7_flush_user_cache_all,
|
|
.flush_user_range = v7_flush_user_cache_range,
|
|
.coherent_kern_range = v7_coherent_kern_range,
|
|
.coherent_user_range = v7_coherent_user_range,
|
|
.flush_kern_dcache_area = v7_flush_kern_dcache_area,
|
|
.dma_map_area = v7_dma_map_area,
|
|
.dma_unmap_area = v7_dma_unmap_area,
|
|
.dma_flush_range = v7_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
/* The NOP cache is just a set of dummy stubs that by definition does nothing */
|
|
#ifdef CONFIG_CPU_CACHE_NOP
|
|
void nop_flush_icache_all(void);
|
|
void nop_flush_kern_cache_all(void);
|
|
void nop_flush_user_cache_all(void);
|
|
void nop_flush_user_cache_range(unsigned long start, unsigned long end, unsigned int flags);
|
|
void nop_coherent_kern_range(unsigned long start, unsigned long end);
|
|
int nop_coherent_user_range(unsigned long, unsigned long);
|
|
void nop_flush_kern_dcache_area(void *kaddr, size_t size);
|
|
void nop_dma_map_area(const void *start, size_t size, int flags);
|
|
void nop_dma_unmap_area(const void *start, size_t size, int flags);
|
|
void nop_dma_flush_range(const void *start, const void *end);
|
|
|
|
struct cpu_cache_fns nop_cache_fns __initconst = {
|
|
.flush_icache_all = nop_flush_icache_all,
|
|
.flush_kern_all = nop_flush_kern_cache_all,
|
|
.flush_kern_louis = nop_flush_kern_cache_all,
|
|
.flush_user_all = nop_flush_user_cache_all,
|
|
.flush_user_range = nop_flush_user_cache_range,
|
|
.coherent_kern_range = nop_coherent_kern_range,
|
|
.coherent_user_range = nop_coherent_user_range,
|
|
.flush_kern_dcache_area = nop_flush_kern_dcache_area,
|
|
.dma_map_area = nop_dma_map_area,
|
|
.dma_unmap_area = nop_dma_unmap_area,
|
|
.dma_flush_range = nop_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_CACHE_V7M
|
|
void v7m_flush_icache_all(void);
|
|
void v7m_flush_kern_cache_all(void);
|
|
void v7m_flush_user_cache_all(void);
|
|
void v7m_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void v7m_coherent_kern_range(unsigned long, unsigned long);
|
|
int v7m_coherent_user_range(unsigned long, unsigned long);
|
|
void v7m_flush_kern_dcache_area(void *, size_t);
|
|
void v7m_dma_map_area(const void *, size_t, int);
|
|
void v7m_dma_unmap_area(const void *, size_t, int);
|
|
void v7m_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns v7m_cache_fns __initconst = {
|
|
.flush_icache_all = v7m_flush_icache_all,
|
|
.flush_kern_all = v7m_flush_kern_cache_all,
|
|
.flush_kern_louis = v7m_flush_kern_cache_all,
|
|
.flush_user_all = v7m_flush_user_cache_all,
|
|
.flush_user_range = v7m_flush_user_cache_range,
|
|
.coherent_kern_range = v7m_coherent_kern_range,
|
|
.coherent_user_range = v7m_coherent_user_range,
|
|
.flush_kern_dcache_area = v7m_flush_kern_dcache_area,
|
|
.dma_map_area = v7m_dma_map_area,
|
|
.dma_unmap_area = v7m_dma_unmap_area,
|
|
.dma_flush_range = v7m_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_ARM1020
|
|
void arm1020_flush_icache_all(void);
|
|
void arm1020_flush_kern_cache_all(void);
|
|
void arm1020_flush_user_cache_all(void);
|
|
void arm1020_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm1020_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm1020_coherent_user_range(unsigned long, unsigned long);
|
|
void arm1020_flush_kern_dcache_area(void *, size_t);
|
|
void arm1020_dma_map_area(const void *, size_t, int);
|
|
void arm1020_dma_unmap_area(const void *, size_t, int);
|
|
void arm1020_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm1020_cache_fns __initconst = {
|
|
.flush_icache_all = arm1020_flush_icache_all,
|
|
.flush_kern_all = arm1020_flush_kern_cache_all,
|
|
.flush_kern_louis = arm1020_flush_kern_cache_all,
|
|
.flush_user_all = arm1020_flush_user_cache_all,
|
|
.flush_user_range = arm1020_flush_user_cache_range,
|
|
.coherent_kern_range = arm1020_coherent_kern_range,
|
|
.coherent_user_range = arm1020_coherent_user_range,
|
|
.flush_kern_dcache_area = arm1020_flush_kern_dcache_area,
|
|
.dma_map_area = arm1020_dma_map_area,
|
|
.dma_unmap_area = arm1020_dma_unmap_area,
|
|
.dma_flush_range = arm1020_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_ARM1020E
|
|
void arm1020e_flush_icache_all(void);
|
|
void arm1020e_flush_kern_cache_all(void);
|
|
void arm1020e_flush_user_cache_all(void);
|
|
void arm1020e_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm1020e_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm1020e_coherent_user_range(unsigned long, unsigned long);
|
|
void arm1020e_flush_kern_dcache_area(void *, size_t);
|
|
void arm1020e_dma_map_area(const void *, size_t, int);
|
|
void arm1020e_dma_unmap_area(const void *, size_t, int);
|
|
void arm1020e_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm1020e_cache_fns __initconst = {
|
|
.flush_icache_all = arm1020e_flush_icache_all,
|
|
.flush_kern_all = arm1020e_flush_kern_cache_all,
|
|
.flush_kern_louis = arm1020e_flush_kern_cache_all,
|
|
.flush_user_all = arm1020e_flush_user_cache_all,
|
|
.flush_user_range = arm1020e_flush_user_cache_range,
|
|
.coherent_kern_range = arm1020e_coherent_kern_range,
|
|
.coherent_user_range = arm1020e_coherent_user_range,
|
|
.flush_kern_dcache_area = arm1020e_flush_kern_dcache_area,
|
|
.dma_map_area = arm1020e_dma_map_area,
|
|
.dma_unmap_area = arm1020e_dma_unmap_area,
|
|
.dma_flush_range = arm1020e_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_ARM1022
|
|
void arm1022_flush_icache_all(void);
|
|
void arm1022_flush_kern_cache_all(void);
|
|
void arm1022_flush_user_cache_all(void);
|
|
void arm1022_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm1022_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm1022_coherent_user_range(unsigned long, unsigned long);
|
|
void arm1022_flush_kern_dcache_area(void *, size_t);
|
|
void arm1022_dma_map_area(const void *, size_t, int);
|
|
void arm1022_dma_unmap_area(const void *, size_t, int);
|
|
void arm1022_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm1022_cache_fns __initconst = {
|
|
.flush_icache_all = arm1022_flush_icache_all,
|
|
.flush_kern_all = arm1022_flush_kern_cache_all,
|
|
.flush_kern_louis = arm1022_flush_kern_cache_all,
|
|
.flush_user_all = arm1022_flush_user_cache_all,
|
|
.flush_user_range = arm1022_flush_user_cache_range,
|
|
.coherent_kern_range = arm1022_coherent_kern_range,
|
|
.coherent_user_range = arm1022_coherent_user_range,
|
|
.flush_kern_dcache_area = arm1022_flush_kern_dcache_area,
|
|
.dma_map_area = arm1022_dma_map_area,
|
|
.dma_unmap_area = arm1022_dma_unmap_area,
|
|
.dma_flush_range = arm1022_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_ARM1026
|
|
void arm1026_flush_icache_all(void);
|
|
void arm1026_flush_kern_cache_all(void);
|
|
void arm1026_flush_user_cache_all(void);
|
|
void arm1026_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm1026_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm1026_coherent_user_range(unsigned long, unsigned long);
|
|
void arm1026_flush_kern_dcache_area(void *, size_t);
|
|
void arm1026_dma_map_area(const void *, size_t, int);
|
|
void arm1026_dma_unmap_area(const void *, size_t, int);
|
|
void arm1026_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm1026_cache_fns __initconst = {
|
|
.flush_icache_all = arm1026_flush_icache_all,
|
|
.flush_kern_all = arm1026_flush_kern_cache_all,
|
|
.flush_kern_louis = arm1026_flush_kern_cache_all,
|
|
.flush_user_all = arm1026_flush_user_cache_all,
|
|
.flush_user_range = arm1026_flush_user_cache_range,
|
|
.coherent_kern_range = arm1026_coherent_kern_range,
|
|
.coherent_user_range = arm1026_coherent_user_range,
|
|
.flush_kern_dcache_area = arm1026_flush_kern_dcache_area,
|
|
.dma_map_area = arm1026_dma_map_area,
|
|
.dma_unmap_area = arm1026_dma_unmap_area,
|
|
.dma_flush_range = arm1026_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#if defined(CONFIG_CPU_ARM920T) && !defined(CONFIG_CPU_DCACHE_WRITETHROUGH)
|
|
void arm920_flush_icache_all(void);
|
|
void arm920_flush_kern_cache_all(void);
|
|
void arm920_flush_user_cache_all(void);
|
|
void arm920_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm920_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm920_coherent_user_range(unsigned long, unsigned long);
|
|
void arm920_flush_kern_dcache_area(void *, size_t);
|
|
void arm920_dma_map_area(const void *, size_t, int);
|
|
void arm920_dma_unmap_area(const void *, size_t, int);
|
|
void arm920_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm920_cache_fns __initconst = {
|
|
.flush_icache_all = arm920_flush_icache_all,
|
|
.flush_kern_all = arm920_flush_kern_cache_all,
|
|
.flush_kern_louis = arm920_flush_kern_cache_all,
|
|
.flush_user_all = arm920_flush_user_cache_all,
|
|
.flush_user_range = arm920_flush_user_cache_range,
|
|
.coherent_kern_range = arm920_coherent_kern_range,
|
|
.coherent_user_range = arm920_coherent_user_range,
|
|
.flush_kern_dcache_area = arm920_flush_kern_dcache_area,
|
|
.dma_map_area = arm920_dma_map_area,
|
|
.dma_unmap_area = arm920_dma_unmap_area,
|
|
.dma_flush_range = arm920_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#if defined(CONFIG_CPU_ARM922T) && !defined(CONFIG_CPU_DCACHE_WRITETHROUGH)
|
|
void arm922_flush_icache_all(void);
|
|
void arm922_flush_kern_cache_all(void);
|
|
void arm922_flush_user_cache_all(void);
|
|
void arm922_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm922_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm922_coherent_user_range(unsigned long, unsigned long);
|
|
void arm922_flush_kern_dcache_area(void *, size_t);
|
|
void arm922_dma_map_area(const void *, size_t, int);
|
|
void arm922_dma_unmap_area(const void *, size_t, int);
|
|
void arm922_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm922_cache_fns __initconst = {
|
|
.flush_icache_all = arm922_flush_icache_all,
|
|
.flush_kern_all = arm922_flush_kern_cache_all,
|
|
.flush_kern_louis = arm922_flush_kern_cache_all,
|
|
.flush_user_all = arm922_flush_user_cache_all,
|
|
.flush_user_range = arm922_flush_user_cache_range,
|
|
.coherent_kern_range = arm922_coherent_kern_range,
|
|
.coherent_user_range = arm922_coherent_user_range,
|
|
.flush_kern_dcache_area = arm922_flush_kern_dcache_area,
|
|
.dma_map_area = arm922_dma_map_area,
|
|
.dma_unmap_area = arm922_dma_unmap_area,
|
|
.dma_flush_range = arm922_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_ARM925T
|
|
void arm925_flush_icache_all(void);
|
|
void arm925_flush_kern_cache_all(void);
|
|
void arm925_flush_user_cache_all(void);
|
|
void arm925_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm925_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm925_coherent_user_range(unsigned long, unsigned long);
|
|
void arm925_flush_kern_dcache_area(void *, size_t);
|
|
void arm925_dma_map_area(const void *, size_t, int);
|
|
void arm925_dma_unmap_area(const void *, size_t, int);
|
|
void arm925_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm925_cache_fns __initconst = {
|
|
.flush_icache_all = arm925_flush_icache_all,
|
|
.flush_kern_all = arm925_flush_kern_cache_all,
|
|
.flush_kern_louis = arm925_flush_kern_cache_all,
|
|
.flush_user_all = arm925_flush_user_cache_all,
|
|
.flush_user_range = arm925_flush_user_cache_range,
|
|
.coherent_kern_range = arm925_coherent_kern_range,
|
|
.coherent_user_range = arm925_coherent_user_range,
|
|
.flush_kern_dcache_area = arm925_flush_kern_dcache_area,
|
|
.dma_map_area = arm925_dma_map_area,
|
|
.dma_unmap_area = arm925_dma_unmap_area,
|
|
.dma_flush_range = arm925_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_ARM926T
|
|
void arm926_flush_icache_all(void);
|
|
void arm926_flush_kern_cache_all(void);
|
|
void arm926_flush_user_cache_all(void);
|
|
void arm926_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm926_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm926_coherent_user_range(unsigned long, unsigned long);
|
|
void arm926_flush_kern_dcache_area(void *, size_t);
|
|
void arm926_dma_map_area(const void *, size_t, int);
|
|
void arm926_dma_unmap_area(const void *, size_t, int);
|
|
void arm926_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm926_cache_fns __initconst = {
|
|
.flush_icache_all = arm926_flush_icache_all,
|
|
.flush_kern_all = arm926_flush_kern_cache_all,
|
|
.flush_kern_louis = arm926_flush_kern_cache_all,
|
|
.flush_user_all = arm926_flush_user_cache_all,
|
|
.flush_user_range = arm926_flush_user_cache_range,
|
|
.coherent_kern_range = arm926_coherent_kern_range,
|
|
.coherent_user_range = arm926_coherent_user_range,
|
|
.flush_kern_dcache_area = arm926_flush_kern_dcache_area,
|
|
.dma_map_area = arm926_dma_map_area,
|
|
.dma_unmap_area = arm926_dma_unmap_area,
|
|
.dma_flush_range = arm926_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_ARM940T
|
|
void arm940_flush_icache_all(void);
|
|
void arm940_flush_kern_cache_all(void);
|
|
void arm940_flush_user_cache_all(void);
|
|
void arm940_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm940_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm940_coherent_user_range(unsigned long, unsigned long);
|
|
void arm940_flush_kern_dcache_area(void *, size_t);
|
|
void arm940_dma_map_area(const void *, size_t, int);
|
|
void arm940_dma_unmap_area(const void *, size_t, int);
|
|
void arm940_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm940_cache_fns __initconst = {
|
|
.flush_icache_all = arm940_flush_icache_all,
|
|
.flush_kern_all = arm940_flush_kern_cache_all,
|
|
.flush_kern_louis = arm940_flush_kern_cache_all,
|
|
.flush_user_all = arm940_flush_user_cache_all,
|
|
.flush_user_range = arm940_flush_user_cache_range,
|
|
.coherent_kern_range = arm940_coherent_kern_range,
|
|
.coherent_user_range = arm940_coherent_user_range,
|
|
.flush_kern_dcache_area = arm940_flush_kern_dcache_area,
|
|
.dma_map_area = arm940_dma_map_area,
|
|
.dma_unmap_area = arm940_dma_unmap_area,
|
|
.dma_flush_range = arm940_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_ARM946E
|
|
void arm946_flush_icache_all(void);
|
|
void arm946_flush_kern_cache_all(void);
|
|
void arm946_flush_user_cache_all(void);
|
|
void arm946_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void arm946_coherent_kern_range(unsigned long, unsigned long);
|
|
int arm946_coherent_user_range(unsigned long, unsigned long);
|
|
void arm946_flush_kern_dcache_area(void *, size_t);
|
|
void arm946_dma_map_area(const void *, size_t, int);
|
|
void arm946_dma_unmap_area(const void *, size_t, int);
|
|
void arm946_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns arm946_cache_fns __initconst = {
|
|
.flush_icache_all = arm946_flush_icache_all,
|
|
.flush_kern_all = arm946_flush_kern_cache_all,
|
|
.flush_kern_louis = arm946_flush_kern_cache_all,
|
|
.flush_user_all = arm946_flush_user_cache_all,
|
|
.flush_user_range = arm946_flush_user_cache_range,
|
|
.coherent_kern_range = arm946_coherent_kern_range,
|
|
.coherent_user_range = arm946_coherent_user_range,
|
|
.flush_kern_dcache_area = arm946_flush_kern_dcache_area,
|
|
.dma_map_area = arm946_dma_map_area,
|
|
.dma_unmap_area = arm946_dma_unmap_area,
|
|
.dma_flush_range = arm946_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_XSCALE
|
|
void xscale_flush_icache_all(void);
|
|
void xscale_flush_kern_cache_all(void);
|
|
void xscale_flush_user_cache_all(void);
|
|
void xscale_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void xscale_coherent_kern_range(unsigned long, unsigned long);
|
|
int xscale_coherent_user_range(unsigned long, unsigned long);
|
|
void xscale_flush_kern_dcache_area(void *, size_t);
|
|
void xscale_dma_map_area(const void *, size_t, int);
|
|
void xscale_dma_unmap_area(const void *, size_t, int);
|
|
void xscale_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns xscale_cache_fns __initconst = {
|
|
.flush_icache_all = xscale_flush_icache_all,
|
|
.flush_kern_all = xscale_flush_kern_cache_all,
|
|
.flush_kern_louis = xscale_flush_kern_cache_all,
|
|
.flush_user_all = xscale_flush_user_cache_all,
|
|
.flush_user_range = xscale_flush_user_cache_range,
|
|
.coherent_kern_range = xscale_coherent_kern_range,
|
|
.coherent_user_range = xscale_coherent_user_range,
|
|
.flush_kern_dcache_area = xscale_flush_kern_dcache_area,
|
|
.dma_map_area = xscale_dma_map_area,
|
|
.dma_unmap_area = xscale_dma_unmap_area,
|
|
.dma_flush_range = xscale_dma_flush_range,
|
|
};
|
|
|
|
/* The 80200 A0 and A1 need a special quirk for dma_map_area() */
|
|
void xscale_80200_A0_A1_dma_map_area(const void *, size_t, int);
|
|
|
|
struct cpu_cache_fns xscale_80200_A0_A1_cache_fns __initconst = {
|
|
.flush_icache_all = xscale_flush_icache_all,
|
|
.flush_kern_all = xscale_flush_kern_cache_all,
|
|
.flush_kern_louis = xscale_flush_kern_cache_all,
|
|
.flush_user_all = xscale_flush_user_cache_all,
|
|
.flush_user_range = xscale_flush_user_cache_range,
|
|
.coherent_kern_range = xscale_coherent_kern_range,
|
|
.coherent_user_range = xscale_coherent_user_range,
|
|
.flush_kern_dcache_area = xscale_flush_kern_dcache_area,
|
|
.dma_map_area = xscale_80200_A0_A1_dma_map_area,
|
|
.dma_unmap_area = xscale_dma_unmap_area,
|
|
.dma_flush_range = xscale_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_XSC3
|
|
void xsc3_flush_icache_all(void);
|
|
void xsc3_flush_kern_cache_all(void);
|
|
void xsc3_flush_user_cache_all(void);
|
|
void xsc3_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void xsc3_coherent_kern_range(unsigned long, unsigned long);
|
|
int xsc3_coherent_user_range(unsigned long, unsigned long);
|
|
void xsc3_flush_kern_dcache_area(void *, size_t);
|
|
void xsc3_dma_map_area(const void *, size_t, int);
|
|
void xsc3_dma_unmap_area(const void *, size_t, int);
|
|
void xsc3_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns xsc3_cache_fns __initconst = {
|
|
.flush_icache_all = xsc3_flush_icache_all,
|
|
.flush_kern_all = xsc3_flush_kern_cache_all,
|
|
.flush_kern_louis = xsc3_flush_kern_cache_all,
|
|
.flush_user_all = xsc3_flush_user_cache_all,
|
|
.flush_user_range = xsc3_flush_user_cache_range,
|
|
.coherent_kern_range = xsc3_coherent_kern_range,
|
|
.coherent_user_range = xsc3_coherent_user_range,
|
|
.flush_kern_dcache_area = xsc3_flush_kern_dcache_area,
|
|
.dma_map_area = xsc3_dma_map_area,
|
|
.dma_unmap_area = xsc3_dma_unmap_area,
|
|
.dma_flush_range = xsc3_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_MOHAWK
|
|
void mohawk_flush_icache_all(void);
|
|
void mohawk_flush_kern_cache_all(void);
|
|
void mohawk_flush_user_cache_all(void);
|
|
void mohawk_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void mohawk_coherent_kern_range(unsigned long, unsigned long);
|
|
int mohawk_coherent_user_range(unsigned long, unsigned long);
|
|
void mohawk_flush_kern_dcache_area(void *, size_t);
|
|
void mohawk_dma_map_area(const void *, size_t, int);
|
|
void mohawk_dma_unmap_area(const void *, size_t, int);
|
|
void mohawk_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns mohawk_cache_fns __initconst = {
|
|
.flush_icache_all = mohawk_flush_icache_all,
|
|
.flush_kern_all = mohawk_flush_kern_cache_all,
|
|
.flush_kern_louis = mohawk_flush_kern_cache_all,
|
|
.flush_user_all = mohawk_flush_user_cache_all,
|
|
.flush_user_range = mohawk_flush_user_cache_range,
|
|
.coherent_kern_range = mohawk_coherent_kern_range,
|
|
.coherent_user_range = mohawk_coherent_user_range,
|
|
.flush_kern_dcache_area = mohawk_flush_kern_dcache_area,
|
|
.dma_map_area = mohawk_dma_map_area,
|
|
.dma_unmap_area = mohawk_dma_unmap_area,
|
|
.dma_flush_range = mohawk_dma_flush_range,
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_FEROCEON
|
|
void feroceon_flush_icache_all(void);
|
|
void feroceon_flush_kern_cache_all(void);
|
|
void feroceon_flush_user_cache_all(void);
|
|
void feroceon_flush_user_cache_range(unsigned long, unsigned long, unsigned int);
|
|
void feroceon_coherent_kern_range(unsigned long, unsigned long);
|
|
int feroceon_coherent_user_range(unsigned long, unsigned long);
|
|
void feroceon_flush_kern_dcache_area(void *, size_t);
|
|
void feroceon_dma_map_area(const void *, size_t, int);
|
|
void feroceon_dma_unmap_area(const void *, size_t, int);
|
|
void feroceon_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns feroceon_cache_fns __initconst = {
|
|
.flush_icache_all = feroceon_flush_icache_all,
|
|
.flush_kern_all = feroceon_flush_kern_cache_all,
|
|
.flush_kern_louis = feroceon_flush_kern_cache_all,
|
|
.flush_user_all = feroceon_flush_user_cache_all,
|
|
.flush_user_range = feroceon_flush_user_cache_range,
|
|
.coherent_kern_range = feroceon_coherent_kern_range,
|
|
.coherent_user_range = feroceon_coherent_user_range,
|
|
.flush_kern_dcache_area = feroceon_flush_kern_dcache_area,
|
|
.dma_map_area = feroceon_dma_map_area,
|
|
.dma_unmap_area = feroceon_dma_unmap_area,
|
|
.dma_flush_range = feroceon_dma_flush_range,
|
|
};
|
|
|
|
void feroceon_range_flush_kern_dcache_area(void *, size_t);
|
|
void feroceon_range_dma_map_area(const void *, size_t, int);
|
|
void feroceon_range_dma_flush_range(const void *, const void *);
|
|
|
|
struct cpu_cache_fns feroceon_range_cache_fns __initconst = {
|
|
.flush_icache_all = feroceon_flush_icache_all,
|
|
.flush_kern_all = feroceon_flush_kern_cache_all,
|
|
.flush_kern_louis = feroceon_flush_kern_cache_all,
|
|
.flush_user_all = feroceon_flush_user_cache_all,
|
|
.flush_user_range = feroceon_flush_user_cache_range,
|
|
.coherent_kern_range = feroceon_coherent_kern_range,
|
|
.coherent_user_range = feroceon_coherent_user_range,
|
|
.flush_kern_dcache_area = feroceon_range_flush_kern_dcache_area,
|
|
.dma_map_area = feroceon_range_dma_map_area,
|
|
.dma_unmap_area = feroceon_dma_unmap_area,
|
|
.dma_flush_range = feroceon_range_dma_flush_range,
|
|
};
|
|
#endif
|