When a new cgroup is created or a cpuset is updated, the mask supplied to it looks for its corresponding CPU translations for the restrictions to apply on. The patch also updates the display interface such that tasks within the namespace can view the corresponding virtual CPUset based on the requested CPU namespace context. Signed-off-by: Pratik R. Sampat <psampat@xxxxxxxxxxxxx> --- kernel/cgroup/cpuset.c | 57 +++++++++++++++++++++++++++++++++++++++--- 1 file changed, 54 insertions(+), 3 deletions(-) diff --git a/kernel/cgroup/cpuset.c b/kernel/cgroup/cpuset.c index adb5190c4429..eb1e950543cf 100644 --- a/kernel/cgroup/cpuset.c +++ b/kernel/cgroup/cpuset.c @@ -65,6 +65,7 @@ #include <linux/mutex.h> #include <linux/cgroup.h> #include <linux/wait.h> +#include <linux/cpu_namespace.h> DEFINE_STATIC_KEY_FALSE(cpusets_pre_enable_key); DEFINE_STATIC_KEY_FALSE(cpusets_enabled_key); @@ -1061,8 +1062,19 @@ static void update_tasks_cpumask(struct cpuset *cs) struct task_struct *task; css_task_iter_start(&cs->css, 0, &it); - while ((task = css_task_iter_next(&it))) + while ((task = css_task_iter_next(&it))) { +#ifdef CONFIG_CPU_NS + cpumask_t pcpus; + cpumask_t vcpus; + + pcpus = get_pcpus_cpuns(current->nsproxy->cpu_ns, cs->effective_cpus); + vcpus = get_vcpus_cpuns(task->nsproxy->cpu_ns, &pcpus); + cpumask_copy(&task->nsproxy->cpu_ns->v_cpuset_cpus, &vcpus); + set_cpus_allowed_ptr(task, &pcpus); +#else set_cpus_allowed_ptr(task, cs->effective_cpus); +#endif + } css_task_iter_end(&it); } @@ -2212,8 +2224,18 @@ static void cpuset_attach(struct cgroup_taskset *tset) * can_attach beforehand should guarantee that this doesn't * fail. TODO: have a better way to handle failure here */ - WARN_ON_ONCE(set_cpus_allowed_ptr(task, cpus_attach)); +#ifdef CONFIG_CPU_NS + cpumask_t pcpus; + cpumask_t vcpus; + pcpus = get_pcpus_cpuns(current->nsproxy->cpu_ns, cpus_attach); + vcpus = get_vcpus_cpuns(task->nsproxy->cpu_ns, &pcpus); + cpumask_copy(&task->nsproxy->cpu_ns->v_cpuset_cpus, &vcpus); + + WARN_ON_ONCE(set_cpus_allowed_ptr(task, &pcpus)); +#else + WARN_ON_ONCE(set_cpus_allowed_ptr(task, cpus_attach)); +#endif cpuset_change_task_nodemask(task, &cpuset_attach_nodemask_to); cpuset_update_task_spread_flag(cs, task); } @@ -2436,13 +2458,33 @@ static int cpuset_common_seq_show(struct seq_file *sf, void *v) switch (type) { case FILE_CPULIST: +#ifdef CONFIG_CPU_NS + if (current->nsproxy->cpu_ns == &init_cpu_ns) { + seq_printf(sf, "%*pbl\n", + cpumask_pr_args(cs->cpus_allowed)); + } else { + seq_printf(sf, "%*pbl\n", + cpumask_pr_args(¤t->nsproxy->cpu_ns->v_cpuset_cpus)); + } +#else seq_printf(sf, "%*pbl\n", cpumask_pr_args(cs->cpus_allowed)); +#endif break; case FILE_MEMLIST: seq_printf(sf, "%*pbl\n", nodemask_pr_args(&cs->mems_allowed)); break; case FILE_EFFECTIVE_CPULIST: +#ifdef CONFIG_CPU_NS + if (current->nsproxy->cpu_ns == &init_cpu_ns) { + seq_printf(sf, "%*pbl\n", + cpumask_pr_args(cs->effective_cpus)); + } else { + seq_printf(sf, "%*pbl\n", + cpumask_pr_args(¤t->nsproxy->cpu_ns->v_cpuset_cpus)); + } +#else seq_printf(sf, "%*pbl\n", cpumask_pr_args(cs->effective_cpus)); +#endif break; case FILE_EFFECTIVE_MEMLIST: seq_printf(sf, "%*pbl\n", nodemask_pr_args(&cs->effective_mems)); @@ -2884,9 +2926,18 @@ static void cpuset_bind(struct cgroup_subsys_state *root_css) */ static void cpuset_fork(struct task_struct *task) { +#ifdef CONFIG_CPU_NS + cpumask_t vcpus; +#endif + if (task_css_is_root(task, cpuset_cgrp_id)) return; - +#ifdef CONFIG_CPU_NS + if (task->nsproxy->cpu_ns != &init_cpu_ns) { + vcpus = get_vcpus_cpuns(task->nsproxy->cpu_ns, current->cpus_ptr); + cpumask_copy(&task->nsproxy->cpu_ns->v_cpuset_cpus, &vcpus); + } +#endif set_cpus_allowed_ptr(task, current->cpus_ptr); task->mems_allowed = current->mems_allowed; } -- 2.31.1