When there is a cpu hotplug event (CPU online or offline), the scheduling domains needed to be reconfigured and regenerated. So code is added to the hotplug functions to make them work with new reserved_cpus mask to compute the right effective_cpus for each of the affected cpusets. Signed-off-by: Waiman Long <longman@xxxxxxxxxx> --- Documentation/admin-guide/cgroup-v2.rst | 7 +++++++ kernel/cgroup/cpuset.c | 26 ++++++++++++++++++++++++-- 2 files changed, 31 insertions(+), 2 deletions(-) diff --git a/Documentation/admin-guide/cgroup-v2.rst b/Documentation/admin-guide/cgroup-v2.rst index 5ee5e77..6ef3516 100644 --- a/Documentation/admin-guide/cgroup-v2.rst +++ b/Documentation/admin-guide/cgroup-v2.rst @@ -1626,6 +1626,13 @@ Cpuset Interface Files 2) No CPU that has been distributed to child scheduling domain roots is deleted. + When all the CPUs allocated to a scheduling domain are offlined, + that scheduling domain will be temporaily gone and all the + tasks in that scheduling domain will migrate to another one that + belongs to the parent of the scheduling domain root. When any + of those offlined CPUs is onlined again, a new scheduling domain + will be re-created and the tasks will be migrated back. + Device controller ----------------- diff --git a/kernel/cgroup/cpuset.c b/kernel/cgroup/cpuset.c index b1abe3d..26ac083 100644 --- a/kernel/cgroup/cpuset.c +++ b/kernel/cgroup/cpuset.c @@ -900,7 +900,8 @@ static void update_tasks_cpumask(struct cpuset *cs) * @parent: the parent cpuset * * If the parent has reserved CPUs, include them in the list of allowable - * CPUs in computing the new effective_cpus mask. + * CPUs in computing the new effective_cpus mask. The cpu_active_mask is + * used to mask off cpus that are to be offlined. */ static void compute_effective_cpumask(struct cpumask *new_cpus, struct cpuset *cs, struct cpuset *parent) @@ -909,6 +910,7 @@ static void compute_effective_cpumask(struct cpumask *new_cpus, cpumask_or(new_cpus, parent->effective_cpus, parent->reserved_cpus); cpumask_and(new_cpus, new_cpus, cs->cpus_allowed); + cpumask_and(new_cpus, new_cpus, cpu_active_mask); } else { cpumask_and(new_cpus, cs->cpus_allowed, parent->effective_cpus); } @@ -2571,9 +2573,17 @@ static void cpuset_hotplug_update_tasks(struct cpuset *cs) goto retry; } - cpumask_and(&new_cpus, cs->cpus_allowed, parent_cs(cs)->effective_cpus); + compute_effective_cpumask(&new_cpus, cs, parent_cs(cs)); nodes_and(new_mems, cs->mems_allowed, parent_cs(cs)->effective_mems); + if (cs->nr_reserved) { + /* + * Some of the CPUs may have been distributed to child + * domain roots. So we need skip those when computing the + * real effective cpus. + */ + cpumask_andnot(&new_cpus, &new_cpus, cs->reserved_cpus); + } cpus_updated = !cpumask_equal(&new_cpus, cs->effective_cpus); mems_updated = !nodes_equal(new_mems, cs->effective_mems); @@ -2623,6 +2633,11 @@ static void cpuset_hotplug_workfn(struct work_struct *work) cpumask_copy(&new_cpus, cpu_active_mask); new_mems = node_states[N_MEMORY]; + /* + * If reserved_cpus is populated, it is likely that the check below + * will produce a false positive on cpus_updated when the cpu list + * isn't changed. It is extra work, but it is better to be safe. + */ cpus_updated = !cpumask_equal(top_cpuset.effective_cpus, &new_cpus); mems_updated = !nodes_equal(top_cpuset.effective_mems, new_mems); @@ -2631,6 +2646,13 @@ static void cpuset_hotplug_workfn(struct work_struct *work) spin_lock_irq(&callback_lock); if (!on_dfl) cpumask_copy(top_cpuset.cpus_allowed, &new_cpus); + /* + * Make sure that the reserved cpus aren't in the + * effective cpus. + */ + if (top_cpuset.nr_reserved) + cpumask_andnot(&new_cpus, &new_cpus, + top_cpuset.reserved_cpus); cpumask_copy(top_cpuset.effective_cpus, &new_cpus); spin_unlock_irq(&callback_lock); /* we don't mess with cpumasks of tasks in top_cpuset */ -- 1.8.3.1 -- To unsubscribe from this list: send the line "unsubscribe linux-doc" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html