mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-16 00:34:20 +08:00
30acd0bdfb
Container admin can create new namespaces and force kernel to allocate up to several pages of memory for the namespaces and its associated structures. Net and uts namespaces have enabled accounting for such allocations. It makes sense to account for rest ones to restrict the host's memory consumption from inside the memcg-limited container. Link: https://lkml.kernel.org/r/5525bcbf-533e-da27-79b7-158686c64e13@virtuozzo.com Signed-off-by: Vasily Averin <vvs@virtuozzo.com> Acked-by: Serge Hallyn <serge@hallyn.com> Acked-by: Christian Brauner <christian.brauner@ubuntu.com> Acked-by: Kirill Tkhai <ktkhai@virtuozzo.com> Reviewed-by: Shakeel Butt <shakeelb@google.com> Cc: Alexander Viro <viro@zeniv.linux.org.uk> Cc: Alexey Dobriyan <adobriyan@gmail.com> Cc: Andrei Vagin <avagin@gmail.com> Cc: Borislav Petkov <bp@alien8.de> Cc: Borislav Petkov <bp@suse.de> Cc: Dmitry Safonov <0x7f454c46@gmail.com> Cc: "Eric W. Biederman" <ebiederm@xmission.com> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: "J. Bruce Fields" <bfields@fieldses.org> Cc: Jeff Layton <jlayton@kernel.org> Cc: Jens Axboe <axboe@kernel.dk> Cc: Jiri Slaby <jirislaby@kernel.org> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: Michal Hocko <mhocko@kernel.org> Cc: Oleg Nesterov <oleg@redhat.com> Cc: Roman Gushchin <guro@fb.com> Cc: Tejun Heo <tj@kernel.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Vladimir Davydov <vdavydov.dev@gmail.com> Cc: Yutian Yang <nglaive@gmail.com> Cc: Zefan Li <lizefan.x@bytedance.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
468 lines
10 KiB
C
468 lines
10 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Author: Andrei Vagin <avagin@openvz.org>
|
|
* Author: Dmitry Safonov <dima@arista.com>
|
|
*/
|
|
|
|
#include <linux/time_namespace.h>
|
|
#include <linux/user_namespace.h>
|
|
#include <linux/sched/signal.h>
|
|
#include <linux/sched/task.h>
|
|
#include <linux/clocksource.h>
|
|
#include <linux/seq_file.h>
|
|
#include <linux/proc_ns.h>
|
|
#include <linux/export.h>
|
|
#include <linux/time.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/cred.h>
|
|
#include <linux/err.h>
|
|
#include <linux/mm.h>
|
|
|
|
#include <vdso/datapage.h>
|
|
|
|
ktime_t do_timens_ktime_to_host(clockid_t clockid, ktime_t tim,
|
|
struct timens_offsets *ns_offsets)
|
|
{
|
|
ktime_t offset;
|
|
|
|
switch (clockid) {
|
|
case CLOCK_MONOTONIC:
|
|
offset = timespec64_to_ktime(ns_offsets->monotonic);
|
|
break;
|
|
case CLOCK_BOOTTIME:
|
|
case CLOCK_BOOTTIME_ALARM:
|
|
offset = timespec64_to_ktime(ns_offsets->boottime);
|
|
break;
|
|
default:
|
|
return tim;
|
|
}
|
|
|
|
/*
|
|
* Check that @tim value is in [offset, KTIME_MAX + offset]
|
|
* and subtract offset.
|
|
*/
|
|
if (tim < offset) {
|
|
/*
|
|
* User can specify @tim *absolute* value - if it's lesser than
|
|
* the time namespace's offset - it's already expired.
|
|
*/
|
|
tim = 0;
|
|
} else {
|
|
tim = ktime_sub(tim, offset);
|
|
if (unlikely(tim > KTIME_MAX))
|
|
tim = KTIME_MAX;
|
|
}
|
|
|
|
return tim;
|
|
}
|
|
|
|
static struct ucounts *inc_time_namespaces(struct user_namespace *ns)
|
|
{
|
|
return inc_ucount(ns, current_euid(), UCOUNT_TIME_NAMESPACES);
|
|
}
|
|
|
|
static void dec_time_namespaces(struct ucounts *ucounts)
|
|
{
|
|
dec_ucount(ucounts, UCOUNT_TIME_NAMESPACES);
|
|
}
|
|
|
|
/**
|
|
* clone_time_ns - Clone a time namespace
|
|
* @user_ns: User namespace which owns a new namespace.
|
|
* @old_ns: Namespace to clone
|
|
*
|
|
* Clone @old_ns and set the clone refcount to 1
|
|
*
|
|
* Return: The new namespace or ERR_PTR.
|
|
*/
|
|
static struct time_namespace *clone_time_ns(struct user_namespace *user_ns,
|
|
struct time_namespace *old_ns)
|
|
{
|
|
struct time_namespace *ns;
|
|
struct ucounts *ucounts;
|
|
int err;
|
|
|
|
err = -ENOSPC;
|
|
ucounts = inc_time_namespaces(user_ns);
|
|
if (!ucounts)
|
|
goto fail;
|
|
|
|
err = -ENOMEM;
|
|
ns = kmalloc(sizeof(*ns), GFP_KERNEL_ACCOUNT);
|
|
if (!ns)
|
|
goto fail_dec;
|
|
|
|
refcount_set(&ns->ns.count, 1);
|
|
|
|
ns->vvar_page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO);
|
|
if (!ns->vvar_page)
|
|
goto fail_free;
|
|
|
|
err = ns_alloc_inum(&ns->ns);
|
|
if (err)
|
|
goto fail_free_page;
|
|
|
|
ns->ucounts = ucounts;
|
|
ns->ns.ops = &timens_operations;
|
|
ns->user_ns = get_user_ns(user_ns);
|
|
ns->offsets = old_ns->offsets;
|
|
ns->frozen_offsets = false;
|
|
return ns;
|
|
|
|
fail_free_page:
|
|
__free_page(ns->vvar_page);
|
|
fail_free:
|
|
kfree(ns);
|
|
fail_dec:
|
|
dec_time_namespaces(ucounts);
|
|
fail:
|
|
return ERR_PTR(err);
|
|
}
|
|
|
|
/**
|
|
* copy_time_ns - Create timens_for_children from @old_ns
|
|
* @flags: Cloning flags
|
|
* @user_ns: User namespace which owns a new namespace.
|
|
* @old_ns: Namespace to clone
|
|
*
|
|
* If CLONE_NEWTIME specified in @flags, creates a new timens_for_children;
|
|
* adds a refcounter to @old_ns otherwise.
|
|
*
|
|
* Return: timens_for_children namespace or ERR_PTR.
|
|
*/
|
|
struct time_namespace *copy_time_ns(unsigned long flags,
|
|
struct user_namespace *user_ns, struct time_namespace *old_ns)
|
|
{
|
|
if (!(flags & CLONE_NEWTIME))
|
|
return get_time_ns(old_ns);
|
|
|
|
return clone_time_ns(user_ns, old_ns);
|
|
}
|
|
|
|
static struct timens_offset offset_from_ts(struct timespec64 off)
|
|
{
|
|
struct timens_offset ret;
|
|
|
|
ret.sec = off.tv_sec;
|
|
ret.nsec = off.tv_nsec;
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* A time namespace VVAR page has the same layout as the VVAR page which
|
|
* contains the system wide VDSO data.
|
|
*
|
|
* For a normal task the VVAR pages are installed in the normal ordering:
|
|
* VVAR
|
|
* PVCLOCK
|
|
* HVCLOCK
|
|
* TIMENS <- Not really required
|
|
*
|
|
* Now for a timens task the pages are installed in the following order:
|
|
* TIMENS
|
|
* PVCLOCK
|
|
* HVCLOCK
|
|
* VVAR
|
|
*
|
|
* The check for vdso_data->clock_mode is in the unlikely path of
|
|
* the seq begin magic. So for the non-timens case most of the time
|
|
* 'seq' is even, so the branch is not taken.
|
|
*
|
|
* If 'seq' is odd, i.e. a concurrent update is in progress, the extra check
|
|
* for vdso_data->clock_mode is a non-issue. The task is spin waiting for the
|
|
* update to finish and for 'seq' to become even anyway.
|
|
*
|
|
* Timens page has vdso_data->clock_mode set to VDSO_CLOCKMODE_TIMENS which
|
|
* enforces the time namespace handling path.
|
|
*/
|
|
static void timens_setup_vdso_data(struct vdso_data *vdata,
|
|
struct time_namespace *ns)
|
|
{
|
|
struct timens_offset *offset = vdata->offset;
|
|
struct timens_offset monotonic = offset_from_ts(ns->offsets.monotonic);
|
|
struct timens_offset boottime = offset_from_ts(ns->offsets.boottime);
|
|
|
|
vdata->seq = 1;
|
|
vdata->clock_mode = VDSO_CLOCKMODE_TIMENS;
|
|
offset[CLOCK_MONOTONIC] = monotonic;
|
|
offset[CLOCK_MONOTONIC_RAW] = monotonic;
|
|
offset[CLOCK_MONOTONIC_COARSE] = monotonic;
|
|
offset[CLOCK_BOOTTIME] = boottime;
|
|
offset[CLOCK_BOOTTIME_ALARM] = boottime;
|
|
}
|
|
|
|
/*
|
|
* Protects possibly multiple offsets writers racing each other
|
|
* and tasks entering the namespace.
|
|
*/
|
|
static DEFINE_MUTEX(offset_lock);
|
|
|
|
static void timens_set_vvar_page(struct task_struct *task,
|
|
struct time_namespace *ns)
|
|
{
|
|
struct vdso_data *vdata;
|
|
unsigned int i;
|
|
|
|
if (ns == &init_time_ns)
|
|
return;
|
|
|
|
/* Fast-path, taken by every task in namespace except the first. */
|
|
if (likely(ns->frozen_offsets))
|
|
return;
|
|
|
|
mutex_lock(&offset_lock);
|
|
/* Nothing to-do: vvar_page has been already initialized. */
|
|
if (ns->frozen_offsets)
|
|
goto out;
|
|
|
|
ns->frozen_offsets = true;
|
|
vdata = arch_get_vdso_data(page_address(ns->vvar_page));
|
|
|
|
for (i = 0; i < CS_BASES; i++)
|
|
timens_setup_vdso_data(&vdata[i], ns);
|
|
|
|
out:
|
|
mutex_unlock(&offset_lock);
|
|
}
|
|
|
|
void free_time_ns(struct time_namespace *ns)
|
|
{
|
|
dec_time_namespaces(ns->ucounts);
|
|
put_user_ns(ns->user_ns);
|
|
ns_free_inum(&ns->ns);
|
|
__free_page(ns->vvar_page);
|
|
kfree(ns);
|
|
}
|
|
|
|
static struct time_namespace *to_time_ns(struct ns_common *ns)
|
|
{
|
|
return container_of(ns, struct time_namespace, ns);
|
|
}
|
|
|
|
static struct ns_common *timens_get(struct task_struct *task)
|
|
{
|
|
struct time_namespace *ns = NULL;
|
|
struct nsproxy *nsproxy;
|
|
|
|
task_lock(task);
|
|
nsproxy = task->nsproxy;
|
|
if (nsproxy) {
|
|
ns = nsproxy->time_ns;
|
|
get_time_ns(ns);
|
|
}
|
|
task_unlock(task);
|
|
|
|
return ns ? &ns->ns : NULL;
|
|
}
|
|
|
|
static struct ns_common *timens_for_children_get(struct task_struct *task)
|
|
{
|
|
struct time_namespace *ns = NULL;
|
|
struct nsproxy *nsproxy;
|
|
|
|
task_lock(task);
|
|
nsproxy = task->nsproxy;
|
|
if (nsproxy) {
|
|
ns = nsproxy->time_ns_for_children;
|
|
get_time_ns(ns);
|
|
}
|
|
task_unlock(task);
|
|
|
|
return ns ? &ns->ns : NULL;
|
|
}
|
|
|
|
static void timens_put(struct ns_common *ns)
|
|
{
|
|
put_time_ns(to_time_ns(ns));
|
|
}
|
|
|
|
void timens_commit(struct task_struct *tsk, struct time_namespace *ns)
|
|
{
|
|
timens_set_vvar_page(tsk, ns);
|
|
vdso_join_timens(tsk, ns);
|
|
}
|
|
|
|
static int timens_install(struct nsset *nsset, struct ns_common *new)
|
|
{
|
|
struct nsproxy *nsproxy = nsset->nsproxy;
|
|
struct time_namespace *ns = to_time_ns(new);
|
|
|
|
if (!current_is_single_threaded())
|
|
return -EUSERS;
|
|
|
|
if (!ns_capable(ns->user_ns, CAP_SYS_ADMIN) ||
|
|
!ns_capable(nsset->cred->user_ns, CAP_SYS_ADMIN))
|
|
return -EPERM;
|
|
|
|
get_time_ns(ns);
|
|
put_time_ns(nsproxy->time_ns);
|
|
nsproxy->time_ns = ns;
|
|
|
|
get_time_ns(ns);
|
|
put_time_ns(nsproxy->time_ns_for_children);
|
|
nsproxy->time_ns_for_children = ns;
|
|
return 0;
|
|
}
|
|
|
|
void timens_on_fork(struct nsproxy *nsproxy, struct task_struct *tsk)
|
|
{
|
|
struct ns_common *nsc = &nsproxy->time_ns_for_children->ns;
|
|
struct time_namespace *ns = to_time_ns(nsc);
|
|
|
|
/* create_new_namespaces() already incremented the ref counter */
|
|
if (nsproxy->time_ns == nsproxy->time_ns_for_children)
|
|
return;
|
|
|
|
get_time_ns(ns);
|
|
put_time_ns(nsproxy->time_ns);
|
|
nsproxy->time_ns = ns;
|
|
|
|
timens_commit(tsk, ns);
|
|
}
|
|
|
|
static struct user_namespace *timens_owner(struct ns_common *ns)
|
|
{
|
|
return to_time_ns(ns)->user_ns;
|
|
}
|
|
|
|
static void show_offset(struct seq_file *m, int clockid, struct timespec64 *ts)
|
|
{
|
|
char *clock;
|
|
|
|
switch (clockid) {
|
|
case CLOCK_BOOTTIME:
|
|
clock = "boottime";
|
|
break;
|
|
case CLOCK_MONOTONIC:
|
|
clock = "monotonic";
|
|
break;
|
|
default:
|
|
clock = "unknown";
|
|
break;
|
|
}
|
|
seq_printf(m, "%-10s %10lld %9ld\n", clock, ts->tv_sec, ts->tv_nsec);
|
|
}
|
|
|
|
void proc_timens_show_offsets(struct task_struct *p, struct seq_file *m)
|
|
{
|
|
struct ns_common *ns;
|
|
struct time_namespace *time_ns;
|
|
|
|
ns = timens_for_children_get(p);
|
|
if (!ns)
|
|
return;
|
|
time_ns = to_time_ns(ns);
|
|
|
|
show_offset(m, CLOCK_MONOTONIC, &time_ns->offsets.monotonic);
|
|
show_offset(m, CLOCK_BOOTTIME, &time_ns->offsets.boottime);
|
|
put_time_ns(time_ns);
|
|
}
|
|
|
|
int proc_timens_set_offset(struct file *file, struct task_struct *p,
|
|
struct proc_timens_offset *offsets, int noffsets)
|
|
{
|
|
struct ns_common *ns;
|
|
struct time_namespace *time_ns;
|
|
struct timespec64 tp;
|
|
int i, err;
|
|
|
|
ns = timens_for_children_get(p);
|
|
if (!ns)
|
|
return -ESRCH;
|
|
time_ns = to_time_ns(ns);
|
|
|
|
if (!file_ns_capable(file, time_ns->user_ns, CAP_SYS_TIME)) {
|
|
put_time_ns(time_ns);
|
|
return -EPERM;
|
|
}
|
|
|
|
for (i = 0; i < noffsets; i++) {
|
|
struct proc_timens_offset *off = &offsets[i];
|
|
|
|
switch (off->clockid) {
|
|
case CLOCK_MONOTONIC:
|
|
ktime_get_ts64(&tp);
|
|
break;
|
|
case CLOCK_BOOTTIME:
|
|
ktime_get_boottime_ts64(&tp);
|
|
break;
|
|
default:
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
err = -ERANGE;
|
|
|
|
if (off->val.tv_sec > KTIME_SEC_MAX ||
|
|
off->val.tv_sec < -KTIME_SEC_MAX)
|
|
goto out;
|
|
|
|
tp = timespec64_add(tp, off->val);
|
|
/*
|
|
* KTIME_SEC_MAX is divided by 2 to be sure that KTIME_MAX is
|
|
* still unreachable.
|
|
*/
|
|
if (tp.tv_sec < 0 || tp.tv_sec > KTIME_SEC_MAX / 2)
|
|
goto out;
|
|
}
|
|
|
|
mutex_lock(&offset_lock);
|
|
if (time_ns->frozen_offsets) {
|
|
err = -EACCES;
|
|
goto out_unlock;
|
|
}
|
|
|
|
err = 0;
|
|
/* Don't report errors after this line */
|
|
for (i = 0; i < noffsets; i++) {
|
|
struct proc_timens_offset *off = &offsets[i];
|
|
struct timespec64 *offset = NULL;
|
|
|
|
switch (off->clockid) {
|
|
case CLOCK_MONOTONIC:
|
|
offset = &time_ns->offsets.monotonic;
|
|
break;
|
|
case CLOCK_BOOTTIME:
|
|
offset = &time_ns->offsets.boottime;
|
|
break;
|
|
}
|
|
|
|
*offset = off->val;
|
|
}
|
|
|
|
out_unlock:
|
|
mutex_unlock(&offset_lock);
|
|
out:
|
|
put_time_ns(time_ns);
|
|
|
|
return err;
|
|
}
|
|
|
|
const struct proc_ns_operations timens_operations = {
|
|
.name = "time",
|
|
.type = CLONE_NEWTIME,
|
|
.get = timens_get,
|
|
.put = timens_put,
|
|
.install = timens_install,
|
|
.owner = timens_owner,
|
|
};
|
|
|
|
const struct proc_ns_operations timens_for_children_operations = {
|
|
.name = "time_for_children",
|
|
.real_ns_name = "time",
|
|
.type = CLONE_NEWTIME,
|
|
.get = timens_for_children_get,
|
|
.put = timens_put,
|
|
.install = timens_install,
|
|
.owner = timens_owner,
|
|
};
|
|
|
|
struct time_namespace init_time_ns = {
|
|
.ns.count = REFCOUNT_INIT(3),
|
|
.user_ns = &init_user_ns,
|
|
.ns.inum = PROC_TIME_INIT_INO,
|
|
.ns.ops = &timens_operations,
|
|
.frozen_offsets = true,
|
|
};
|