mirror of
https://github.com/qemu/qemu.git
synced 2024-12-03 16:53:53 +08:00
664a79735e
For darwin, the CTR_EL0 register is not accessible, but there are system routines that we can use. For other hosts, copy the single pointer implementation from libgcc and modify it to support the double pointer interface we require. This halves the number of cache operations required when split-rwx is enabled. Reviewed-by: Joelle van Dyne <j@getutm.app> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
147 lines
4.1 KiB
C
147 lines
4.1 KiB
C
/*
|
|
* Flush the host cpu caches.
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or later.
|
|
* See the COPYING file in the top-level directory.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/cacheflush.h"
|
|
#include "qemu/bitops.h"
|
|
|
|
|
|
#if defined(__i386__) || defined(__x86_64__) || defined(__s390__)
|
|
|
|
/* Caches are coherent and do not require flushing; symbol inline. */
|
|
|
|
#elif defined(__aarch64__)
|
|
|
|
#ifdef CONFIG_DARWIN
|
|
/* Apple does not expose CTR_EL0, so we must use system interfaces. */
|
|
extern void sys_icache_invalidate(void *start, size_t len);
|
|
extern void sys_dcache_flush(void *start, size_t len);
|
|
void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
|
|
{
|
|
sys_dcache_flush((void *)rw, len);
|
|
sys_icache_invalidate((void *)rx, len);
|
|
}
|
|
#else
|
|
|
|
/*
|
|
* TODO: unify this with cacheinfo.c.
|
|
* We want to save the whole contents of CTR_EL0, so that we
|
|
* have more than the linesize, but also IDC and DIC.
|
|
*/
|
|
static unsigned int save_ctr_el0;
|
|
static void __attribute__((constructor)) init_ctr_el0(void)
|
|
{
|
|
asm volatile("mrs\t%0, ctr_el0" : "=r"(save_ctr_el0));
|
|
}
|
|
|
|
/*
|
|
* This is a copy of gcc's __aarch64_sync_cache_range, modified
|
|
* to fit this three-operand interface.
|
|
*/
|
|
void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
|
|
{
|
|
const unsigned CTR_IDC = 1u << 28;
|
|
const unsigned CTR_DIC = 1u << 29;
|
|
const unsigned int ctr_el0 = save_ctr_el0;
|
|
const uintptr_t icache_lsize = 4 << extract32(ctr_el0, 0, 4);
|
|
const uintptr_t dcache_lsize = 4 << extract32(ctr_el0, 16, 4);
|
|
uintptr_t p;
|
|
|
|
/*
|
|
* If CTR_EL0.IDC is enabled, Data cache clean to the Point of Unification
|
|
* is not required for instruction to data coherence.
|
|
*/
|
|
if (!(ctr_el0 & CTR_IDC)) {
|
|
/*
|
|
* Loop over the address range, clearing one cache line at once.
|
|
* Data cache must be flushed to unification first to make sure
|
|
* the instruction cache fetches the updated data.
|
|
*/
|
|
for (p = rw & -dcache_lsize; p < rw + len; p += dcache_lsize) {
|
|
asm volatile("dc\tcvau, %0" : : "r" (p) : "memory");
|
|
}
|
|
asm volatile("dsb\tish" : : : "memory");
|
|
}
|
|
|
|
/*
|
|
* If CTR_EL0.DIC is enabled, Instruction cache cleaning to the Point
|
|
* of Unification is not required for instruction to data coherence.
|
|
*/
|
|
if (!(ctr_el0 & CTR_DIC)) {
|
|
for (p = rx & -icache_lsize; p < rx + len; p += icache_lsize) {
|
|
asm volatile("ic\tivau, %0" : : "r"(p) : "memory");
|
|
}
|
|
asm volatile ("dsb\tish" : : : "memory");
|
|
}
|
|
|
|
asm volatile("isb" : : : "memory");
|
|
}
|
|
#endif /* CONFIG_DARWIN */
|
|
|
|
#elif defined(__mips__)
|
|
|
|
#ifdef __OpenBSD__
|
|
#include <machine/sysarch.h>
|
|
#else
|
|
#include <sys/cachectl.h>
|
|
#endif
|
|
|
|
void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
|
|
{
|
|
if (rx != rw) {
|
|
cacheflush((void *)rw, len, DCACHE);
|
|
}
|
|
cacheflush((void *)rx, len, ICACHE);
|
|
}
|
|
|
|
#elif defined(__powerpc__)
|
|
|
|
void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
|
|
{
|
|
uintptr_t p, b, e;
|
|
size_t dsize = qemu_dcache_linesize;
|
|
size_t isize = qemu_icache_linesize;
|
|
|
|
b = rw & ~(dsize - 1);
|
|
e = (rw + len + dsize - 1) & ~(dsize - 1);
|
|
for (p = b; p < e; p += dsize) {
|
|
asm volatile ("dcbst 0,%0" : : "r"(p) : "memory");
|
|
}
|
|
asm volatile ("sync" : : : "memory");
|
|
|
|
b = rx & ~(isize - 1);
|
|
e = (rx + len + isize - 1) & ~(isize - 1);
|
|
for (p = b; p < e; p += isize) {
|
|
asm volatile ("icbi 0,%0" : : "r"(p) : "memory");
|
|
}
|
|
asm volatile ("sync" : : : "memory");
|
|
asm volatile ("isync" : : : "memory");
|
|
}
|
|
|
|
#elif defined(__sparc__)
|
|
|
|
void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
|
|
{
|
|
/* No additional data flush to the RW virtual address required. */
|
|
uintptr_t p, end = (rx + len + 7) & -8;
|
|
for (p = rx & -8; p < end; p += 8) {
|
|
__asm__ __volatile__("flush\t%0" : : "r" (p));
|
|
}
|
|
}
|
|
|
|
#else
|
|
|
|
void flush_idcache_range(uintptr_t rx, uintptr_t rw, size_t len)
|
|
{
|
|
if (rw != rx) {
|
|
__builtin___clear_cache((char *)rw, (char *)rw + len);
|
|
}
|
|
__builtin___clear_cache((char *)rx, (char *)rx + len);
|
|
}
|
|
|
|
#endif
|