2019-03-27 08:41:29 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
/*
|
|
|
|
* Copyright (C) 2012 Regents of the University of California
|
|
|
|
* Copyright (C) 2017 SiFive
|
2021-02-03 17:49:07 +08:00
|
|
|
* Copyright (C) 2021 Western Digital Corporation or its affiliates.
|
2019-03-27 08:41:29 +08:00
|
|
|
*/
|
|
|
|
|
2021-02-03 17:49:07 +08:00
|
|
|
#include <linux/bitops.h>
|
|
|
|
#include <linux/cpumask.h>
|
2019-03-27 08:41:29 +08:00
|
|
|
#include <linux/mm.h>
|
2021-02-03 17:49:07 +08:00
|
|
|
#include <linux/percpu.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/spinlock.h>
|
|
|
|
#include <linux/static_key.h>
|
2019-03-27 08:41:29 +08:00
|
|
|
#include <asm/tlbflush.h>
|
|
|
|
#include <asm/cacheflush.h>
|
2019-10-18 06:21:28 +08:00
|
|
|
#include <asm/mmu_context.h>
|
2024-03-13 07:53:41 +08:00
|
|
|
#include <asm/switch_to.h>
|
2019-03-27 08:41:29 +08:00
|
|
|
|
2021-02-03 17:49:07 +08:00
|
|
|
#ifdef CONFIG_MMU
|
|
|
|
|
2021-06-06 23:20:50 +08:00
|
|
|
DEFINE_STATIC_KEY_FALSE(use_asid_allocator);
|
2021-02-03 17:49:07 +08:00
|
|
|
|
|
|
|
static unsigned long num_asids;
|
|
|
|
|
|
|
|
static atomic_long_t current_version;
|
|
|
|
|
|
|
|
static DEFINE_RAW_SPINLOCK(context_lock);
|
|
|
|
static cpumask_t context_tlb_flush_pending;
|
|
|
|
static unsigned long *context_asid_map;
|
|
|
|
|
|
|
|
static DEFINE_PER_CPU(atomic_long_t, active_context);
|
|
|
|
static DEFINE_PER_CPU(unsigned long, reserved_context);
|
|
|
|
|
|
|
|
static bool check_update_reserved_context(unsigned long cntx,
|
|
|
|
unsigned long newcntx)
|
|
|
|
{
|
|
|
|
int cpu;
|
|
|
|
bool hit = false;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Iterate over the set of reserved CONTEXT looking for a match.
|
|
|
|
* If we find one, then we can update our mm to use new CONTEXT
|
|
|
|
* (i.e. the same CONTEXT in the current_version) but we can't
|
|
|
|
* exit the loop early, since we need to ensure that all copies
|
|
|
|
* of the old CONTEXT are updated to reflect the mm. Failure to do
|
|
|
|
* so could result in us missing the reserved CONTEXT in a future
|
|
|
|
* version.
|
|
|
|
*/
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
|
|
if (per_cpu(reserved_context, cpu) == cntx) {
|
|
|
|
hit = true;
|
|
|
|
per_cpu(reserved_context, cpu) = newcntx;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return hit;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __flush_context(void)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
unsigned long cntx;
|
|
|
|
|
|
|
|
/* Must be called with context_lock held */
|
|
|
|
lockdep_assert_held(&context_lock);
|
|
|
|
|
|
|
|
/* Update the list of reserved ASIDs and the ASID bitmap. */
|
2023-05-06 17:11:41 +08:00
|
|
|
bitmap_zero(context_asid_map, num_asids);
|
2021-02-03 17:49:07 +08:00
|
|
|
|
|
|
|
/* Mark already active ASIDs as used */
|
|
|
|
for_each_possible_cpu(i) {
|
|
|
|
cntx = atomic_long_xchg_relaxed(&per_cpu(active_context, i), 0);
|
|
|
|
/*
|
|
|
|
* If this CPU has already been through a rollover, but
|
|
|
|
* hasn't run another task in the meantime, we must preserve
|
|
|
|
* its reserved CONTEXT, as this is the only trace we have of
|
|
|
|
* the process it is still running.
|
|
|
|
*/
|
|
|
|
if (cntx == 0)
|
|
|
|
cntx = per_cpu(reserved_context, i);
|
|
|
|
|
2024-03-27 12:49:50 +08:00
|
|
|
__set_bit(cntx2asid(cntx), context_asid_map);
|
2021-02-03 17:49:07 +08:00
|
|
|
per_cpu(reserved_context, i) = cntx;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Mark ASID #0 as used because it is used at boot-time */
|
|
|
|
__set_bit(0, context_asid_map);
|
|
|
|
|
|
|
|
/* Queue a TLB invalidation for each CPU on next context-switch */
|
|
|
|
cpumask_setall(&context_tlb_flush_pending);
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned long __new_context(struct mm_struct *mm)
|
|
|
|
{
|
|
|
|
static u32 cur_idx = 1;
|
|
|
|
unsigned long cntx = atomic_long_read(&mm->context.id);
|
|
|
|
unsigned long asid, ver = atomic_long_read(¤t_version);
|
|
|
|
|
|
|
|
/* Must be called with context_lock held */
|
|
|
|
lockdep_assert_held(&context_lock);
|
|
|
|
|
|
|
|
if (cntx != 0) {
|
2024-03-27 12:49:50 +08:00
|
|
|
unsigned long newcntx = ver | cntx2asid(cntx);
|
2021-02-03 17:49:07 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If our current CONTEXT was active during a rollover, we
|
|
|
|
* can continue to use it and this was just a false alarm.
|
|
|
|
*/
|
|
|
|
if (check_update_reserved_context(cntx, newcntx))
|
|
|
|
return newcntx;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We had a valid CONTEXT in a previous life, so try to
|
|
|
|
* re-use it if possible.
|
|
|
|
*/
|
2024-03-27 12:49:50 +08:00
|
|
|
if (!__test_and_set_bit(cntx2asid(cntx), context_asid_map))
|
2021-02-03 17:49:07 +08:00
|
|
|
return newcntx;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate a free ASID. If we can't find one then increment
|
|
|
|
* current_version and flush all ASIDs.
|
|
|
|
*/
|
|
|
|
asid = find_next_zero_bit(context_asid_map, num_asids, cur_idx);
|
|
|
|
if (asid != num_asids)
|
|
|
|
goto set_asid;
|
|
|
|
|
|
|
|
/* We're out of ASIDs, so increment current_version */
|
2024-03-27 12:49:51 +08:00
|
|
|
ver = atomic_long_add_return_relaxed(BIT(SATP_ASID_BITS), ¤t_version);
|
2021-02-03 17:49:07 +08:00
|
|
|
|
|
|
|
/* Flush everything */
|
|
|
|
__flush_context();
|
|
|
|
|
|
|
|
/* We have more ASIDs than CPUs, so this will always succeed */
|
|
|
|
asid = find_next_zero_bit(context_asid_map, num_asids, 1);
|
|
|
|
|
|
|
|
set_asid:
|
|
|
|
__set_bit(asid, context_asid_map);
|
|
|
|
cur_idx = asid;
|
|
|
|
return asid | ver;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void set_mm_asid(struct mm_struct *mm, unsigned int cpu)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
bool need_flush_tlb = false;
|
|
|
|
unsigned long cntx, old_active_cntx;
|
|
|
|
|
|
|
|
cntx = atomic_long_read(&mm->context.id);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If our active_context is non-zero and the context matches the
|
|
|
|
* current_version, then we update the active_context entry with a
|
|
|
|
* relaxed cmpxchg.
|
|
|
|
*
|
|
|
|
* Following is how we handle racing with a concurrent rollover:
|
|
|
|
*
|
|
|
|
* - We get a zero back from the cmpxchg and end up waiting on the
|
|
|
|
* lock. Taking the lock synchronises with the rollover and so
|
|
|
|
* we are forced to see the updated verion.
|
|
|
|
*
|
|
|
|
* - We get a valid context back from the cmpxchg then we continue
|
|
|
|
* using old ASID because __flush_context() would have marked ASID
|
|
|
|
* of active_context as used and next context switch we will
|
|
|
|
* allocate new context.
|
|
|
|
*/
|
|
|
|
old_active_cntx = atomic_long_read(&per_cpu(active_context, cpu));
|
|
|
|
if (old_active_cntx &&
|
2024-03-27 12:49:50 +08:00
|
|
|
(cntx2version(cntx) == atomic_long_read(¤t_version)) &&
|
2021-02-03 17:49:07 +08:00
|
|
|
atomic_long_cmpxchg_relaxed(&per_cpu(active_context, cpu),
|
|
|
|
old_active_cntx, cntx))
|
|
|
|
goto switch_mm_fast;
|
|
|
|
|
|
|
|
raw_spin_lock_irqsave(&context_lock, flags);
|
|
|
|
|
|
|
|
/* Check that our ASID belongs to the current_version. */
|
|
|
|
cntx = atomic_long_read(&mm->context.id);
|
2024-03-27 12:49:50 +08:00
|
|
|
if (cntx2version(cntx) != atomic_long_read(¤t_version)) {
|
2021-02-03 17:49:07 +08:00
|
|
|
cntx = __new_context(mm);
|
|
|
|
atomic_long_set(&mm->context.id, cntx);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (cpumask_test_and_clear_cpu(cpu, &context_tlb_flush_pending))
|
|
|
|
need_flush_tlb = true;
|
|
|
|
|
|
|
|
atomic_long_set(&per_cpu(active_context, cpu), cntx);
|
|
|
|
|
|
|
|
raw_spin_unlock_irqrestore(&context_lock, flags);
|
|
|
|
|
|
|
|
switch_mm_fast:
|
|
|
|
csr_write(CSR_SATP, virt_to_pfn(mm->pgd) |
|
2024-03-27 12:49:50 +08:00
|
|
|
(cntx2asid(cntx) << SATP_ASID_SHIFT) |
|
2021-12-06 18:46:51 +08:00
|
|
|
satp_mode);
|
2021-02-03 17:49:07 +08:00
|
|
|
|
|
|
|
if (need_flush_tlb)
|
|
|
|
local_flush_tlb_all();
|
|
|
|
}
|
|
|
|
|
|
|
|
static void set_mm_noasid(struct mm_struct *mm)
|
|
|
|
{
|
|
|
|
/* Switch the page table and blindly nuke entire local TLB */
|
2021-12-06 18:46:51 +08:00
|
|
|
csr_write(CSR_SATP, virt_to_pfn(mm->pgd) | satp_mode);
|
2024-03-27 12:49:53 +08:00
|
|
|
local_flush_tlb_all_asid(0);
|
2021-02-03 17:49:07 +08:00
|
|
|
}
|
|
|
|
|
riscv: asid: Fixup stale TLB entry cause application crash
After use_asid_allocator is enabled, the userspace application will
crash by stale TLB entries. Because only using cpumask_clear_cpu without
local_flush_tlb_all couldn't guarantee CPU's TLB entries were fresh.
Then set_mm_asid would cause the user space application to get a stale
value by stale TLB entry, but set_mm_noasid is okay.
Here is the symptom of the bug:
unhandled signal 11 code 0x1 (coredump)
0x0000003fd6d22524 <+4>: auipc s0,0x70
0x0000003fd6d22528 <+8>: ld s0,-148(s0) # 0x3fd6d92490
=> 0x0000003fd6d2252c <+12>: ld a5,0(s0)
(gdb) i r s0
s0 0x8082ed1cc3198b21 0x8082ed1cc3198b21
(gdb) x /2x 0x3fd6d92490
0x3fd6d92490: 0xd80ac8a8 0x0000003f
The core dump file shows that register s0 is wrong, but the value in
memory is correct. Because 'ld s0, -148(s0)' used a stale mapping entry
in TLB and got a wrong result from an incorrect physical address.
When the task ran on CPU0, which loaded/speculative-loaded the value of
address(0x3fd6d92490), then the first version of the mapping entry was
PTWed into CPU0's TLB.
When the task switched from CPU0 to CPU1 (No local_tlb_flush_all here by
asid), it happened to write a value on the address (0x3fd6d92490). It
caused do_page_fault -> wp_page_copy -> ptep_clear_flush ->
ptep_get_and_clear & flush_tlb_page.
The flush_tlb_page used mm_cpumask(mm) to determine which CPUs need TLB
flush, but CPU0 had cleared the CPU0's mm_cpumask in the previous
switch_mm. So we only flushed the CPU1 TLB and set the second version
mapping of the PTE. When the task switched from CPU1 to CPU0 again, CPU0
still used a stale TLB mapping entry which contained a wrong target
physical address. It raised a bug when the task happened to read that
value.
CPU0 CPU1
- switch 'task' in
- read addr (Fill stale mapping
entry into TLB)
- switch 'task' out (no tlb_flush)
- switch 'task' in (no tlb_flush)
- write addr cause pagefault
do_page_fault() (change to
new addr mapping)
wp_page_copy()
ptep_clear_flush()
ptep_get_and_clear()
& flush_tlb_page()
write new value into addr
- switch 'task' out (no tlb_flush)
- switch 'task' in (no tlb_flush)
- read addr again (Use stale
mapping entry in TLB)
get wrong value from old phyical
addr, BUG!
The solution is to keep all CPUs' footmarks of cpumask(mm) in switch_mm,
which could guarantee to invalidate all stale TLB entries during TLB
flush.
Fixes: 65d4b9c53017 ("RISC-V: Implement ASID allocator")
Signed-off-by: Guo Ren <guoren@linux.alibaba.com>
Signed-off-by: Guo Ren <guoren@kernel.org>
Tested-by: Lad Prabhakar <prabhakar.mahadev-lad.rj@bp.renesas.com>
Tested-by: Zong Li <zong.li@sifive.com>
Tested-by: Sergey Matyukevich <sergey.matyukevich@syntacore.com>
Cc: Anup Patel <apatel@ventanamicro.com>
Cc: Palmer Dabbelt <palmer@rivosinc.com>
Cc: stable@vger.kernel.org
Reviewed-by: Andrew Jones <ajones@ventanamicro.com>
Link: https://lore.kernel.org/r/20230226150137.1919750-3-geomatsi@gmail.com
Signed-off-by: Palmer Dabbelt <palmer@rivosinc.com>
2023-02-26 23:01:37 +08:00
|
|
|
static inline void set_mm(struct mm_struct *prev,
|
|
|
|
struct mm_struct *next, unsigned int cpu)
|
2021-02-03 17:49:07 +08:00
|
|
|
{
|
riscv: asid: Fixup stale TLB entry cause application crash
After use_asid_allocator is enabled, the userspace application will
crash by stale TLB entries. Because only using cpumask_clear_cpu without
local_flush_tlb_all couldn't guarantee CPU's TLB entries were fresh.
Then set_mm_asid would cause the user space application to get a stale
value by stale TLB entry, but set_mm_noasid is okay.
Here is the symptom of the bug:
unhandled signal 11 code 0x1 (coredump)
0x0000003fd6d22524 <+4>: auipc s0,0x70
0x0000003fd6d22528 <+8>: ld s0,-148(s0) # 0x3fd6d92490
=> 0x0000003fd6d2252c <+12>: ld a5,0(s0)
(gdb) i r s0
s0 0x8082ed1cc3198b21 0x8082ed1cc3198b21
(gdb) x /2x 0x3fd6d92490
0x3fd6d92490: 0xd80ac8a8 0x0000003f
The core dump file shows that register s0 is wrong, but the value in
memory is correct. Because 'ld s0, -148(s0)' used a stale mapping entry
in TLB and got a wrong result from an incorrect physical address.
When the task ran on CPU0, which loaded/speculative-loaded the value of
address(0x3fd6d92490), then the first version of the mapping entry was
PTWed into CPU0's TLB.
When the task switched from CPU0 to CPU1 (No local_tlb_flush_all here by
asid), it happened to write a value on the address (0x3fd6d92490). It
caused do_page_fault -> wp_page_copy -> ptep_clear_flush ->
ptep_get_and_clear & flush_tlb_page.
The flush_tlb_page used mm_cpumask(mm) to determine which CPUs need TLB
flush, but CPU0 had cleared the CPU0's mm_cpumask in the previous
switch_mm. So we only flushed the CPU1 TLB and set the second version
mapping of the PTE. When the task switched from CPU1 to CPU0 again, CPU0
still used a stale TLB mapping entry which contained a wrong target
physical address. It raised a bug when the task happened to read that
value.
CPU0 CPU1
- switch 'task' in
- read addr (Fill stale mapping
entry into TLB)
- switch 'task' out (no tlb_flush)
- switch 'task' in (no tlb_flush)
- write addr cause pagefault
do_page_fault() (change to
new addr mapping)
wp_page_copy()
ptep_clear_flush()
ptep_get_and_clear()
& flush_tlb_page()
write new value into addr
- switch 'task' out (no tlb_flush)
- switch 'task' in (no tlb_flush)
- read addr again (Use stale
mapping entry in TLB)
get wrong value from old phyical
addr, BUG!
The solution is to keep all CPUs' footmarks of cpumask(mm) in switch_mm,
which could guarantee to invalidate all stale TLB entries during TLB
flush.
Fixes: 65d4b9c53017 ("RISC-V: Implement ASID allocator")
Signed-off-by: Guo Ren <guoren@linux.alibaba.com>
Signed-off-by: Guo Ren <guoren@kernel.org>
Tested-by: Lad Prabhakar <prabhakar.mahadev-lad.rj@bp.renesas.com>
Tested-by: Zong Li <zong.li@sifive.com>
Tested-by: Sergey Matyukevich <sergey.matyukevich@syntacore.com>
Cc: Anup Patel <apatel@ventanamicro.com>
Cc: Palmer Dabbelt <palmer@rivosinc.com>
Cc: stable@vger.kernel.org
Reviewed-by: Andrew Jones <ajones@ventanamicro.com>
Link: https://lore.kernel.org/r/20230226150137.1919750-3-geomatsi@gmail.com
Signed-off-by: Palmer Dabbelt <palmer@rivosinc.com>
2023-02-26 23:01:37 +08:00
|
|
|
/*
|
|
|
|
* The mm_cpumask indicates which harts' TLBs contain the virtual
|
|
|
|
* address mapping of the mm. Compared to noasid, using asid
|
|
|
|
* can't guarantee that stale TLB entries are invalidated because
|
|
|
|
* the asid mechanism wouldn't flush TLB for every switch_mm for
|
|
|
|
* performance. So when using asid, keep all CPUs footmarks in
|
|
|
|
* cpumask() until mm reset.
|
|
|
|
*/
|
|
|
|
cpumask_set_cpu(cpu, mm_cpumask(next));
|
|
|
|
if (static_branch_unlikely(&use_asid_allocator)) {
|
|
|
|
set_mm_asid(next, cpu);
|
|
|
|
} else {
|
|
|
|
cpumask_clear_cpu(cpu, mm_cpumask(prev));
|
|
|
|
set_mm_noasid(next);
|
|
|
|
}
|
2021-02-03 17:49:07 +08:00
|
|
|
}
|
|
|
|
|
2021-05-16 20:59:42 +08:00
|
|
|
static int __init asids_init(void)
|
2021-02-03 17:49:07 +08:00
|
|
|
{
|
2024-03-27 12:49:52 +08:00
|
|
|
unsigned long asid_bits, old;
|
2021-02-03 17:49:07 +08:00
|
|
|
|
|
|
|
/* Figure-out number of ASID bits in HW */
|
|
|
|
old = csr_read(CSR_SATP);
|
|
|
|
asid_bits = old | (SATP_ASID_MASK << SATP_ASID_SHIFT);
|
|
|
|
csr_write(CSR_SATP, asid_bits);
|
|
|
|
asid_bits = (csr_read(CSR_SATP) >> SATP_ASID_SHIFT) & SATP_ASID_MASK;
|
|
|
|
asid_bits = fls_long(asid_bits);
|
|
|
|
csr_write(CSR_SATP, old);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* In the process of determining number of ASID bits (above)
|
|
|
|
* we polluted the TLB of current HART so let's do TLB flushed
|
|
|
|
* to remove unwanted TLB enteries.
|
|
|
|
*/
|
|
|
|
local_flush_tlb_all();
|
|
|
|
|
|
|
|
/* Pre-compute ASID details */
|
2021-09-09 01:30:29 +08:00
|
|
|
if (asid_bits) {
|
|
|
|
num_asids = 1 << asid_bits;
|
|
|
|
}
|
2021-02-03 17:49:07 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Use ASID allocator only if number of HW ASIDs are
|
|
|
|
* at-least twice more than CPUs
|
|
|
|
*/
|
|
|
|
if (num_asids > (2 * num_possible_cpus())) {
|
2024-03-27 12:49:51 +08:00
|
|
|
atomic_long_set(¤t_version, BIT(SATP_ASID_BITS));
|
2021-02-03 17:49:07 +08:00
|
|
|
|
2021-05-29 19:15:34 +08:00
|
|
|
context_asid_map = bitmap_zalloc(num_asids, GFP_KERNEL);
|
2021-02-03 17:49:07 +08:00
|
|
|
if (!context_asid_map)
|
|
|
|
panic("Failed to allocate bitmap for %lu ASIDs\n",
|
|
|
|
num_asids);
|
|
|
|
|
|
|
|
__set_bit(0, context_asid_map);
|
|
|
|
|
|
|
|
static_branch_enable(&use_asid_allocator);
|
|
|
|
|
|
|
|
pr_info("ASID allocator using %lu bits (%lu entries)\n",
|
|
|
|
asid_bits, num_asids);
|
|
|
|
} else {
|
2021-09-09 01:30:29 +08:00
|
|
|
pr_info("ASID allocator disabled (%lu bits)\n", asid_bits);
|
2021-02-03 17:49:07 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_initcall(asids_init);
|
|
|
|
#else
|
riscv: asid: Fixup stale TLB entry cause application crash
After use_asid_allocator is enabled, the userspace application will
crash by stale TLB entries. Because only using cpumask_clear_cpu without
local_flush_tlb_all couldn't guarantee CPU's TLB entries were fresh.
Then set_mm_asid would cause the user space application to get a stale
value by stale TLB entry, but set_mm_noasid is okay.
Here is the symptom of the bug:
unhandled signal 11 code 0x1 (coredump)
0x0000003fd6d22524 <+4>: auipc s0,0x70
0x0000003fd6d22528 <+8>: ld s0,-148(s0) # 0x3fd6d92490
=> 0x0000003fd6d2252c <+12>: ld a5,0(s0)
(gdb) i r s0
s0 0x8082ed1cc3198b21 0x8082ed1cc3198b21
(gdb) x /2x 0x3fd6d92490
0x3fd6d92490: 0xd80ac8a8 0x0000003f
The core dump file shows that register s0 is wrong, but the value in
memory is correct. Because 'ld s0, -148(s0)' used a stale mapping entry
in TLB and got a wrong result from an incorrect physical address.
When the task ran on CPU0, which loaded/speculative-loaded the value of
address(0x3fd6d92490), then the first version of the mapping entry was
PTWed into CPU0's TLB.
When the task switched from CPU0 to CPU1 (No local_tlb_flush_all here by
asid), it happened to write a value on the address (0x3fd6d92490). It
caused do_page_fault -> wp_page_copy -> ptep_clear_flush ->
ptep_get_and_clear & flush_tlb_page.
The flush_tlb_page used mm_cpumask(mm) to determine which CPUs need TLB
flush, but CPU0 had cleared the CPU0's mm_cpumask in the previous
switch_mm. So we only flushed the CPU1 TLB and set the second version
mapping of the PTE. When the task switched from CPU1 to CPU0 again, CPU0
still used a stale TLB mapping entry which contained a wrong target
physical address. It raised a bug when the task happened to read that
value.
CPU0 CPU1
- switch 'task' in
- read addr (Fill stale mapping
entry into TLB)
- switch 'task' out (no tlb_flush)
- switch 'task' in (no tlb_flush)
- write addr cause pagefault
do_page_fault() (change to
new addr mapping)
wp_page_copy()
ptep_clear_flush()
ptep_get_and_clear()
& flush_tlb_page()
write new value into addr
- switch 'task' out (no tlb_flush)
- switch 'task' in (no tlb_flush)
- read addr again (Use stale
mapping entry in TLB)
get wrong value from old phyical
addr, BUG!
The solution is to keep all CPUs' footmarks of cpumask(mm) in switch_mm,
which could guarantee to invalidate all stale TLB entries during TLB
flush.
Fixes: 65d4b9c53017 ("RISC-V: Implement ASID allocator")
Signed-off-by: Guo Ren <guoren@linux.alibaba.com>
Signed-off-by: Guo Ren <guoren@kernel.org>
Tested-by: Lad Prabhakar <prabhakar.mahadev-lad.rj@bp.renesas.com>
Tested-by: Zong Li <zong.li@sifive.com>
Tested-by: Sergey Matyukevich <sergey.matyukevich@syntacore.com>
Cc: Anup Patel <apatel@ventanamicro.com>
Cc: Palmer Dabbelt <palmer@rivosinc.com>
Cc: stable@vger.kernel.org
Reviewed-by: Andrew Jones <ajones@ventanamicro.com>
Link: https://lore.kernel.org/r/20230226150137.1919750-3-geomatsi@gmail.com
Signed-off-by: Palmer Dabbelt <palmer@rivosinc.com>
2023-02-26 23:01:37 +08:00
|
|
|
static inline void set_mm(struct mm_struct *prev,
|
|
|
|
struct mm_struct *next, unsigned int cpu)
|
2021-02-03 17:49:07 +08:00
|
|
|
{
|
|
|
|
/* Nothing to do here when there is no MMU */
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2019-03-27 08:41:29 +08:00
|
|
|
/*
|
|
|
|
* When necessary, performs a deferred icache flush for the given MM context,
|
|
|
|
* on the local CPU. RISC-V has no direct mechanism for instruction cache
|
|
|
|
* shoot downs, so instead we send an IPI that informs the remote harts they
|
|
|
|
* need to flush their local instruction caches. To avoid pathologically slow
|
|
|
|
* behavior in a common case (a bunch of single-hart processes on a many-hart
|
|
|
|
* machine, ie 'make -j') we avoid the IPIs for harts that are not currently
|
|
|
|
* executing a MM context and instead schedule a deferred local instruction
|
|
|
|
* cache flush to be performed before execution resumes on each hart. This
|
|
|
|
* actually performs that local instruction cache flush, which implicitly only
|
|
|
|
* refers to the current hart.
|
2021-05-12 01:42:31 +08:00
|
|
|
*
|
|
|
|
* The "cpu" argument must be the current local CPU number.
|
2019-03-27 08:41:29 +08:00
|
|
|
*/
|
2024-03-13 07:53:41 +08:00
|
|
|
static inline void flush_icache_deferred(struct mm_struct *mm, unsigned int cpu,
|
|
|
|
struct task_struct *task)
|
2019-03-27 08:41:29 +08:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_SMP
|
2024-03-13 07:53:41 +08:00
|
|
|
if (cpumask_test_and_clear_cpu(cpu, &mm->context.icache_stale_mask)) {
|
2019-03-27 08:41:29 +08:00
|
|
|
/*
|
|
|
|
* Ensure the remote hart's writes are visible to this hart.
|
|
|
|
* This pairs with a barrier in flush_icache_mm.
|
|
|
|
*/
|
|
|
|
smp_mb();
|
|
|
|
|
2024-03-13 07:53:41 +08:00
|
|
|
/*
|
|
|
|
* If cache will be flushed in switch_to, no need to flush here.
|
|
|
|
*/
|
|
|
|
if (!(task && switch_to_should_flush_icache(task)))
|
|
|
|
local_flush_icache_all();
|
|
|
|
}
|
2019-03-27 08:41:29 +08:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
void switch_mm(struct mm_struct *prev, struct mm_struct *next,
|
|
|
|
struct task_struct *task)
|
|
|
|
{
|
|
|
|
unsigned int cpu;
|
|
|
|
|
|
|
|
if (unlikely(prev == next))
|
|
|
|
return;
|
|
|
|
|
2024-01-31 22:49:33 +08:00
|
|
|
membarrier_arch_switch_mm(prev, next, task);
|
|
|
|
|
2019-03-27 08:41:29 +08:00
|
|
|
/*
|
|
|
|
* Mark the current MM context as inactive, and the next as
|
|
|
|
* active. This is at least used by the icache flushing
|
|
|
|
* routines in order to determine who should be flushed.
|
|
|
|
*/
|
|
|
|
cpu = smp_processor_id();
|
|
|
|
|
riscv: asid: Fixup stale TLB entry cause application crash
After use_asid_allocator is enabled, the userspace application will
crash by stale TLB entries. Because only using cpumask_clear_cpu without
local_flush_tlb_all couldn't guarantee CPU's TLB entries were fresh.
Then set_mm_asid would cause the user space application to get a stale
value by stale TLB entry, but set_mm_noasid is okay.
Here is the symptom of the bug:
unhandled signal 11 code 0x1 (coredump)
0x0000003fd6d22524 <+4>: auipc s0,0x70
0x0000003fd6d22528 <+8>: ld s0,-148(s0) # 0x3fd6d92490
=> 0x0000003fd6d2252c <+12>: ld a5,0(s0)
(gdb) i r s0
s0 0x8082ed1cc3198b21 0x8082ed1cc3198b21
(gdb) x /2x 0x3fd6d92490
0x3fd6d92490: 0xd80ac8a8 0x0000003f
The core dump file shows that register s0 is wrong, but the value in
memory is correct. Because 'ld s0, -148(s0)' used a stale mapping entry
in TLB and got a wrong result from an incorrect physical address.
When the task ran on CPU0, which loaded/speculative-loaded the value of
address(0x3fd6d92490), then the first version of the mapping entry was
PTWed into CPU0's TLB.
When the task switched from CPU0 to CPU1 (No local_tlb_flush_all here by
asid), it happened to write a value on the address (0x3fd6d92490). It
caused do_page_fault -> wp_page_copy -> ptep_clear_flush ->
ptep_get_and_clear & flush_tlb_page.
The flush_tlb_page used mm_cpumask(mm) to determine which CPUs need TLB
flush, but CPU0 had cleared the CPU0's mm_cpumask in the previous
switch_mm. So we only flushed the CPU1 TLB and set the second version
mapping of the PTE. When the task switched from CPU1 to CPU0 again, CPU0
still used a stale TLB mapping entry which contained a wrong target
physical address. It raised a bug when the task happened to read that
value.
CPU0 CPU1
- switch 'task' in
- read addr (Fill stale mapping
entry into TLB)
- switch 'task' out (no tlb_flush)
- switch 'task' in (no tlb_flush)
- write addr cause pagefault
do_page_fault() (change to
new addr mapping)
wp_page_copy()
ptep_clear_flush()
ptep_get_and_clear()
& flush_tlb_page()
write new value into addr
- switch 'task' out (no tlb_flush)
- switch 'task' in (no tlb_flush)
- read addr again (Use stale
mapping entry in TLB)
get wrong value from old phyical
addr, BUG!
The solution is to keep all CPUs' footmarks of cpumask(mm) in switch_mm,
which could guarantee to invalidate all stale TLB entries during TLB
flush.
Fixes: 65d4b9c53017 ("RISC-V: Implement ASID allocator")
Signed-off-by: Guo Ren <guoren@linux.alibaba.com>
Signed-off-by: Guo Ren <guoren@kernel.org>
Tested-by: Lad Prabhakar <prabhakar.mahadev-lad.rj@bp.renesas.com>
Tested-by: Zong Li <zong.li@sifive.com>
Tested-by: Sergey Matyukevich <sergey.matyukevich@syntacore.com>
Cc: Anup Patel <apatel@ventanamicro.com>
Cc: Palmer Dabbelt <palmer@rivosinc.com>
Cc: stable@vger.kernel.org
Reviewed-by: Andrew Jones <ajones@ventanamicro.com>
Link: https://lore.kernel.org/r/20230226150137.1919750-3-geomatsi@gmail.com
Signed-off-by: Palmer Dabbelt <palmer@rivosinc.com>
2023-02-26 23:01:37 +08:00
|
|
|
set_mm(prev, next, cpu);
|
2019-03-27 08:41:29 +08:00
|
|
|
|
2024-03-13 07:53:41 +08:00
|
|
|
flush_icache_deferred(next, cpu, task);
|
2019-03-27 08:41:29 +08:00
|
|
|
}
|