mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-24 12:44:11 +08:00
7d6be67cfd
Commit2b5067a814
("mm: mmap_lock: add tracepoints around lock acquisition") introduced TRACE_MMAP_LOCK_EVENT() macro using preempt_disable() in order to let get_mm_memcg_path() return a percpu buffer exclusively used by normal, softirq, irq and NMI contexts respectively. Commit832b507253
("mm: mmap_lock: use local locks instead of disabling preemption") replaced preempt_disable() with local_lock(&memcg_paths.lock) based on an argument that preempt_disable() has to be avoided because get_mm_memcg_path() might sleep if PREEMPT_RT=y. But syzbot started reporting inconsistent {HARDIRQ-ON-W} -> {IN-HARDIRQ-W} usage. and inconsistent {SOFTIRQ-ON-W} -> {IN-SOFTIRQ-W} usage. messages, for local_lock() does not disable IRQ. We could replace local_lock() with local_lock_irqsave() in order to suppress these messages. But this patch instead replaces percpu buffers with on-stack buffer, for the size of each buffer returned by get_memcg_path_buf() is only 256 bytes which is tolerable for allocating from current thread's kernel stack memory. Link: https://lkml.kernel.org/r/ef22d289-eadb-4ed9-863b-fbc922b33d8d@I-love.SAKURA.ne.jp Reported-by: syzbot <syzbot+40905bca570ae6784745@syzkaller.appspotmail.com> Closes: https://syzkaller.appspot.com/bug?extid=40905bca570ae6784745 Fixes:832b507253
("mm: mmap_lock: use local locks instead of disabling preemption") Signed-off-by: Tetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp> Reviewed-by: Axel Rasmussen <axelrasmussen@google.com> Cc: Nicolas Saenz Julienne <nsaenzju@redhat.com> Cc: <stable@vger.kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
112 lines
2.8 KiB
C
112 lines
2.8 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#define CREATE_TRACE_POINTS
|
|
#include <trace/events/mmap_lock.h>
|
|
|
|
#include <linux/mm.h>
|
|
#include <linux/cgroup.h>
|
|
#include <linux/memcontrol.h>
|
|
#include <linux/mmap_lock.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/percpu.h>
|
|
#include <linux/rcupdate.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/trace_events.h>
|
|
#include <linux/local_lock.h>
|
|
|
|
EXPORT_TRACEPOINT_SYMBOL(mmap_lock_start_locking);
|
|
EXPORT_TRACEPOINT_SYMBOL(mmap_lock_acquire_returned);
|
|
EXPORT_TRACEPOINT_SYMBOL(mmap_lock_released);
|
|
|
|
#ifdef CONFIG_MEMCG
|
|
|
|
static atomic_t reg_refcount;
|
|
|
|
/*
|
|
* Size of the buffer for memcg path names. Ignoring stack trace support,
|
|
* trace_events_hist.c uses MAX_FILTER_STR_VAL for this, so we also use it.
|
|
*/
|
|
#define MEMCG_PATH_BUF_SIZE MAX_FILTER_STR_VAL
|
|
|
|
int trace_mmap_lock_reg(void)
|
|
{
|
|
atomic_inc(®_refcount);
|
|
return 0;
|
|
}
|
|
|
|
void trace_mmap_lock_unreg(void)
|
|
{
|
|
atomic_dec(®_refcount);
|
|
}
|
|
|
|
#define TRACE_MMAP_LOCK_EVENT(type, mm, ...) \
|
|
do { \
|
|
char buf[MEMCG_PATH_BUF_SIZE]; \
|
|
get_mm_memcg_path(mm, buf, sizeof(buf)); \
|
|
trace_mmap_lock_##type(mm, buf, ##__VA_ARGS__); \
|
|
} while (0)
|
|
|
|
#else /* !CONFIG_MEMCG */
|
|
|
|
int trace_mmap_lock_reg(void)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
void trace_mmap_lock_unreg(void)
|
|
{
|
|
}
|
|
|
|
#define TRACE_MMAP_LOCK_EVENT(type, mm, ...) \
|
|
trace_mmap_lock_##type(mm, "", ##__VA_ARGS__)
|
|
|
|
#endif /* CONFIG_MEMCG */
|
|
|
|
#ifdef CONFIG_TRACING
|
|
#ifdef CONFIG_MEMCG
|
|
/*
|
|
* Write the given mm_struct's memcg path to a buffer. If the path cannot be
|
|
* determined or the trace event is being unregistered, empty string is written.
|
|
*/
|
|
static void get_mm_memcg_path(struct mm_struct *mm, char *buf, size_t buflen)
|
|
{
|
|
struct mem_cgroup *memcg;
|
|
|
|
buf[0] = '\0';
|
|
/* No need to get path if no trace event is registered. */
|
|
if (!atomic_read(®_refcount))
|
|
return;
|
|
memcg = get_mem_cgroup_from_mm(mm);
|
|
if (memcg == NULL)
|
|
return;
|
|
if (memcg->css.cgroup)
|
|
cgroup_path(memcg->css.cgroup, buf, buflen);
|
|
css_put(&memcg->css);
|
|
}
|
|
|
|
#endif /* CONFIG_MEMCG */
|
|
|
|
/*
|
|
* Trace calls must be in a separate file, as otherwise there's a circular
|
|
* dependency between linux/mmap_lock.h and trace/events/mmap_lock.h.
|
|
*/
|
|
|
|
void __mmap_lock_do_trace_start_locking(struct mm_struct *mm, bool write)
|
|
{
|
|
TRACE_MMAP_LOCK_EVENT(start_locking, mm, write);
|
|
}
|
|
EXPORT_SYMBOL(__mmap_lock_do_trace_start_locking);
|
|
|
|
void __mmap_lock_do_trace_acquire_returned(struct mm_struct *mm, bool write,
|
|
bool success)
|
|
{
|
|
TRACE_MMAP_LOCK_EVENT(acquire_returned, mm, write, success);
|
|
}
|
|
EXPORT_SYMBOL(__mmap_lock_do_trace_acquire_returned);
|
|
|
|
void __mmap_lock_do_trace_released(struct mm_struct *mm, bool write)
|
|
{
|
|
TRACE_MMAP_LOCK_EVENT(released, mm, write);
|
|
}
|
|
EXPORT_SYMBOL(__mmap_lock_do_trace_released);
|
|
#endif /* CONFIG_TRACING */
|