mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-21 11:44:01 +08:00
078a55fc82
commit 3747069b25e419f6b51395f48127e9812abc3596 upstream.
The __cpuinit type of throwaway sections might have made sense
some time ago when RAM was more constrained, but now the savings
do not offset the cost and complications. For example, the fix in
commit 5e427ec2d0
("x86: Fix bit corruption at CPU resume time")
is a good example of the nasty type of bugs that can be created
with improper use of the various __init prefixes.
After a discussion on LKML[1] it was decided that cpuinit should go
the way of devinit and be phased out. Once all the users are gone,
we can then finally remove the macros themselves from linux/init.h.
Note that some harmless section mismatch warnings may result, since
notify_cpu_starting() and cpu_up() are arch independent (kernel/cpu.c)
and are flagged as __cpuinit -- so if we remove the __cpuinit from
the arch specific callers, we will also get section mismatch warnings.
As an intermediate step, we intend to turn the linux/init.h cpuinit
related content into no-ops as early as possible, since that will get
rid of these warnings. In any case, they are temporary and harmless.
Here, we remove all the MIPS __cpuinit from C code and __CPUINIT
from asm files. MIPS is interesting in this respect, because there
are also uasm users hiding behind their own renamed versions of the
__cpuinit macros.
[1] https://lkml.org/lkml/2013/5/20/589
[ralf@linux-mips.org: Folded in Paul's followup fix.]
Signed-off-by: Paul Gortmaker <paul.gortmaker@windriver.com>
Cc: linux-mips@linux-mips.org
Patchwork: https://patchwork.linux-mips.org/patch/5494/
Patchwork: https://patchwork.linux-mips.org/patch/5495/
Patchwork: https://patchwork.linux-mips.org/patch/5509/
Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
249 lines
5.3 KiB
C
249 lines
5.3 KiB
C
/*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* Copyright (C) 1996 David S. Miller (davem@davemloft.net)
|
|
* Copyright (C) 1997, 1998, 1999, 2000 Ralf Baechle ralf@gnu.org
|
|
* Carsten Langgaard, carstenl@mips.com
|
|
* Copyright (C) 2002 MIPS Technologies, Inc. All rights reserved.
|
|
*/
|
|
#include <linux/init.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/mm.h>
|
|
|
|
#include <asm/cpu.h>
|
|
#include <asm/bootinfo.h>
|
|
#include <asm/mmu_context.h>
|
|
#include <asm/pgtable.h>
|
|
|
|
extern void build_tlb_refill_handler(void);
|
|
|
|
#define TFP_TLB_SIZE 384
|
|
#define TFP_TLB_SET_SHIFT 7
|
|
|
|
/* CP0 hazard avoidance. */
|
|
#define BARRIER __asm__ __volatile__(".set noreorder\n\t" \
|
|
"nop; nop; nop; nop; nop; nop;\n\t" \
|
|
".set reorder\n\t")
|
|
|
|
void local_flush_tlb_all(void)
|
|
{
|
|
unsigned long flags;
|
|
unsigned long old_ctx;
|
|
int entry;
|
|
|
|
local_irq_save(flags);
|
|
/* Save old context and create impossible VPN2 value */
|
|
old_ctx = read_c0_entryhi();
|
|
write_c0_entrylo(0);
|
|
|
|
for (entry = 0; entry < TFP_TLB_SIZE; entry++) {
|
|
write_c0_tlbset(entry >> TFP_TLB_SET_SHIFT);
|
|
write_c0_vaddr(entry << PAGE_SHIFT);
|
|
write_c0_entryhi(CKSEG0 + (entry << (PAGE_SHIFT + 1)));
|
|
mtc0_tlbw_hazard();
|
|
tlb_write();
|
|
}
|
|
tlbw_use_hazard();
|
|
write_c0_entryhi(old_ctx);
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
void local_flush_tlb_mm(struct mm_struct *mm)
|
|
{
|
|
int cpu = smp_processor_id();
|
|
|
|
if (cpu_context(cpu, mm) != 0)
|
|
drop_mmu_context(mm, cpu);
|
|
}
|
|
|
|
void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
|
|
unsigned long end)
|
|
{
|
|
struct mm_struct *mm = vma->vm_mm;
|
|
int cpu = smp_processor_id();
|
|
unsigned long flags;
|
|
int oldpid, newpid, size;
|
|
|
|
if (!cpu_context(cpu, mm))
|
|
return;
|
|
|
|
size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
|
|
size = (size + 1) >> 1;
|
|
|
|
local_irq_save(flags);
|
|
|
|
if (size > TFP_TLB_SIZE / 2) {
|
|
drop_mmu_context(mm, cpu);
|
|
goto out_restore;
|
|
}
|
|
|
|
oldpid = read_c0_entryhi();
|
|
newpid = cpu_asid(cpu, mm);
|
|
|
|
write_c0_entrylo(0);
|
|
|
|
start &= PAGE_MASK;
|
|
end += (PAGE_SIZE - 1);
|
|
end &= PAGE_MASK;
|
|
while (start < end) {
|
|
signed long idx;
|
|
|
|
write_c0_vaddr(start);
|
|
write_c0_entryhi(start);
|
|
start += PAGE_SIZE;
|
|
tlb_probe();
|
|
idx = read_c0_tlbset();
|
|
if (idx < 0)
|
|
continue;
|
|
|
|
write_c0_entryhi(CKSEG0 + (idx << (PAGE_SHIFT + 1)));
|
|
tlb_write();
|
|
}
|
|
write_c0_entryhi(oldpid);
|
|
|
|
out_restore:
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
/* Usable for KV1 addresses only! */
|
|
void local_flush_tlb_kernel_range(unsigned long start, unsigned long end)
|
|
{
|
|
unsigned long size, flags;
|
|
|
|
size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
|
|
size = (size + 1) >> 1;
|
|
|
|
if (size > TFP_TLB_SIZE / 2) {
|
|
local_flush_tlb_all();
|
|
return;
|
|
}
|
|
|
|
local_irq_save(flags);
|
|
|
|
write_c0_entrylo(0);
|
|
|
|
start &= PAGE_MASK;
|
|
end += (PAGE_SIZE - 1);
|
|
end &= PAGE_MASK;
|
|
while (start < end) {
|
|
signed long idx;
|
|
|
|
write_c0_vaddr(start);
|
|
write_c0_entryhi(start);
|
|
start += PAGE_SIZE;
|
|
tlb_probe();
|
|
idx = read_c0_tlbset();
|
|
if (idx < 0)
|
|
continue;
|
|
|
|
write_c0_entryhi(CKSEG0 + (idx << (PAGE_SHIFT + 1)));
|
|
tlb_write();
|
|
}
|
|
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page)
|
|
{
|
|
int cpu = smp_processor_id();
|
|
unsigned long flags;
|
|
int oldpid, newpid;
|
|
signed long idx;
|
|
|
|
if (!cpu_context(cpu, vma->vm_mm))
|
|
return;
|
|
|
|
newpid = cpu_asid(cpu, vma->vm_mm);
|
|
page &= PAGE_MASK;
|
|
local_irq_save(flags);
|
|
oldpid = read_c0_entryhi();
|
|
write_c0_vaddr(page);
|
|
write_c0_entryhi(newpid);
|
|
tlb_probe();
|
|
idx = read_c0_tlbset();
|
|
if (idx < 0)
|
|
goto finish;
|
|
|
|
write_c0_entrylo(0);
|
|
write_c0_entryhi(CKSEG0 + (idx << (PAGE_SHIFT + 1)));
|
|
tlb_write();
|
|
|
|
finish:
|
|
write_c0_entryhi(oldpid);
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
/*
|
|
* We will need multiple versions of update_mmu_cache(), one that just
|
|
* updates the TLB with the new pte(s), and another which also checks
|
|
* for the R4k "end of page" hardware bug and does the needy.
|
|
*/
|
|
void __update_tlb(struct vm_area_struct * vma, unsigned long address, pte_t pte)
|
|
{
|
|
unsigned long flags;
|
|
pgd_t *pgdp;
|
|
pmd_t *pmdp;
|
|
pte_t *ptep;
|
|
int pid;
|
|
|
|
/*
|
|
* Handle debugger faulting in for debugee.
|
|
*/
|
|
if (current->active_mm != vma->vm_mm)
|
|
return;
|
|
|
|
pid = read_c0_entryhi() & ASID_MASK;
|
|
|
|
local_irq_save(flags);
|
|
address &= PAGE_MASK;
|
|
write_c0_vaddr(address);
|
|
write_c0_entryhi(pid);
|
|
pgdp = pgd_offset(vma->vm_mm, address);
|
|
pmdp = pmd_offset(pgdp, address);
|
|
ptep = pte_offset_map(pmdp, address);
|
|
tlb_probe();
|
|
|
|
write_c0_entrylo(pte_val(*ptep++) >> 6);
|
|
tlb_write();
|
|
|
|
write_c0_entryhi(pid);
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
static void probe_tlb(unsigned long config)
|
|
{
|
|
struct cpuinfo_mips *c = ¤t_cpu_data;
|
|
|
|
c->tlbsize = 3 * 128; /* 3 sets each 128 entries */
|
|
}
|
|
|
|
void tlb_init(void)
|
|
{
|
|
unsigned int config = read_c0_config();
|
|
unsigned long status;
|
|
|
|
probe_tlb(config);
|
|
|
|
status = read_c0_status();
|
|
status &= ~(ST0_UPS | ST0_KPS);
|
|
#ifdef CONFIG_PAGE_SIZE_4KB
|
|
status |= (TFP_PAGESIZE_4K << 32) | (TFP_PAGESIZE_4K << 36);
|
|
#elif defined(CONFIG_PAGE_SIZE_8KB)
|
|
status |= (TFP_PAGESIZE_8K << 32) | (TFP_PAGESIZE_8K << 36);
|
|
#elif defined(CONFIG_PAGE_SIZE_16KB)
|
|
status |= (TFP_PAGESIZE_16K << 32) | (TFP_PAGESIZE_16K << 36);
|
|
#elif defined(CONFIG_PAGE_SIZE_64KB)
|
|
status |= (TFP_PAGESIZE_64K << 32) | (TFP_PAGESIZE_64K << 36);
|
|
#endif
|
|
write_c0_status(status);
|
|
|
|
write_c0_wired(0);
|
|
|
|
local_flush_tlb_all();
|
|
|
|
build_tlb_refill_handler();
|
|
}
|