mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-22 12:14:01 +08:00
74bf4312ff
We now use the TSB hardware assist features of the UltraSPARC MMUs. SMP is currently knowingly broken, we need to find another place to store the per-cpu base pointers. We hid them away in the TSB base register, and that obviously will not work any more :-) Another known broken case is non-8KB base page size. Also noticed that flush_tlb_all() is not referenced anywhere, only the internal __flush_tlb_all() (local cpu only) is used by the sparc64 port, so we can get rid of flush_tlb_all(). The kernel gets it's own 8KB TSB (swapper_tsb) and each address space gets it's own private 8K TSB. Later we can add code to dynamically increase the size of per-process TSB as the RSS grows. An 8KB TSB is good enough for up to about a 4MB RSS, after which the TSB starts to incur many capacity and conflict misses. We even accumulate OBP translations into the kernel TSB. Another area for refinement is large page size support. We could use a secondary address space TSB to handle those. Signed-off-by: David S. Miller <davem@davemloft.net>
55 lines
1.4 KiB
C
55 lines
1.4 KiB
C
#ifndef _SPARC64_TLBFLUSH_H
|
|
#define _SPARC64_TLBFLUSH_H
|
|
|
|
#include <linux/config.h>
|
|
#include <linux/mm.h>
|
|
#include <asm/mmu_context.h>
|
|
|
|
/* TSB flush operations. */
|
|
struct mmu_gather;
|
|
extern void flush_tsb_kernel_range(unsigned long start, unsigned long end);
|
|
extern void flush_tsb_user(struct mmu_gather *mp);
|
|
|
|
/* TLB flush operations. */
|
|
|
|
extern void flush_tlb_pending(void);
|
|
|
|
#define flush_tlb_range(vma,start,end) \
|
|
do { (void)(start); flush_tlb_pending(); } while (0)
|
|
#define flush_tlb_page(vma,addr) flush_tlb_pending()
|
|
#define flush_tlb_mm(mm) flush_tlb_pending()
|
|
|
|
/* Local cpu only. */
|
|
extern void __flush_tlb_all(void);
|
|
|
|
extern void __flush_tlb_page(unsigned long context, unsigned long page, unsigned long r);
|
|
|
|
extern void __flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
|
|
|
#ifndef CONFIG_SMP
|
|
|
|
#define flush_tlb_kernel_range(start,end) \
|
|
do { flush_tsb_kernel_range(start,end); \
|
|
__flush_tlb_kernel_range(start,end); \
|
|
} while (0)
|
|
|
|
#else /* CONFIG_SMP */
|
|
|
|
extern void smp_flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
|
|
|
#define flush_tlb_kernel_range(start, end) \
|
|
do { flush_tsb_kernel_range(start,end); \
|
|
smp_flush_tlb_kernel_range(start, end); \
|
|
} while (0)
|
|
|
|
#endif /* ! CONFIG_SMP */
|
|
|
|
static inline void flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end)
|
|
{
|
|
/* We don't use virtual page tables for TLB miss processing
|
|
* any more. Nowadays we use the TSB.
|
|
*/
|
|
}
|
|
|
|
#endif /* _SPARC64_TLBFLUSH_H */
|