[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20251025064844.495525-14-chenridong@huaweicloud.com>
Date: Sat, 25 Oct 2025 06:48:35 +0000
From: Chen Ridong <chenridong@...weicloud.com>
To: longman@...hat.com,
tj@...nel.org,
hannes@...xchg.org,
mkoutny@...e.com
Cc: cgroups@...r.kernel.org,
linux-kernel@...r.kernel.org,
lujialin4@...wei.com,
chenridong@...wei.com
Subject: [PATCH RFC v2 13/22] cpuset: introduce local_partition_update()
From: Chen Ridong <chenridong@...wei.com>
The local_partition_update() function replaces the command partcmd_update
previously handled within update_parent_effective_cpumask(). The update
logic follows a state-based approach:
1. Validation check: First verify if the local partition is currently valid
2. Invalidation handling: If the partition is invalid, trigger invalidation
3. State transition: If an invalid partition has no errors, transition to
valid
4. cpumasks updates: For local partition that only cpu maks changes, use
partition_update() to handle partition change.
Signed-off-by: Chen Ridong <chenridong@...wei.com>
---
kernel/cgroup/cpuset.c | 153 +++++++++++++++++++++++++++++++++++++++--
1 file changed, 148 insertions(+), 5 deletions(-)
diff --git a/kernel/cgroup/cpuset.c b/kernel/cgroup/cpuset.c
index 73a43ab58f72..49df38237c1d 100644
--- a/kernel/cgroup/cpuset.c
+++ b/kernel/cgroup/cpuset.c
@@ -1822,6 +1822,59 @@ static void remote_cpus_update(struct cpuset *cs, struct cpumask *xcpus,
remote_partition_disable(cs, tmp);
}
+static bool is_user_cpus_exclusive(struct cpuset *cs)
+{
+ struct cpuset *parent = parent_cs(cs);
+ struct cgroup_subsys_state *css;
+ struct cpuset *child;
+ bool exclusive = true;
+
+ rcu_read_lock();
+ cpuset_for_each_child(child, css, parent) {
+ if (child == cs)
+ continue;
+ if (!cpusets_are_exclusive(cs, child)) {
+ exclusive = false;
+ break;
+ }
+ }
+ rcu_read_unlock();
+ return exclusive;
+}
+
+/**
+ * validate_local_partition - Validate for local partition
+ * @cs: Target cpuset to validate
+ * @new_prs: New partition root state to validate
+ * @excpus: New exclusive effectuve CPUs mask to validate
+ * @excl_check: Flag to enable exclusive CPUs ownership validation
+ *
+ * Return: PERR_NONE if validation passes, appropriate error code otherwise
+ *
+ * Important: The caller must ensure that @cs's cpu mask is updated before
+ * invoking this function when exclusive CPU validation is required.
+ */
+static enum prs_errcode validate_local_partition(struct cpuset *cs, int new_prs,
+ struct cpumask *excpus, bool excl_check)
+{
+ struct cpuset *parent = parent_cs(cs);
+
+ /*
+ * The parent must be a partition root.
+ * The new cpumask, if present, or the current cpus_allowed must
+ * not be empty.
+ */
+ if (!is_partition_valid(parent)) {
+ return is_partition_invalid(parent)
+ ? PERR_INVPARENT : PERR_NOTPART;
+ }
+
+ if (excl_check && !is_user_cpus_exclusive(cs))
+ return PERR_NOTEXCL;
+
+ return validate_partition(cs, new_prs, excpus);
+}
+
/**
* local_partition_enable - Enable local partition for a cpuset
* @cs: Target cpuset to become a local partition root
@@ -1948,6 +2001,90 @@ static void local_partition_invalidate(struct cpuset *cs, struct tmpmasks *tmp)
}
}
+/**
+ * __local_partition_update - Update local partition configuration
+ * @cs: Target cpuset to update
+ * @xcpus: New exclusive CPU mask
+ * @excpus: New effective exclusive CPU mask
+ * @tmp: Temporary mask storage for intermediate calculations
+ * @excl_check: Flag to enable exclusivity validation
+ *
+ * Handles updates to local CPU partition configurations by validating
+ * changes, managing state transitions, and propagating updates through
+ * the cpuset hierarchy.
+ *
+ * Note on exclusivity checking: Exclusivity validation is required when
+ * transitioning from an invalid to valid partition state. However, when
+ * updating cpus_allowed or exclusive_cpus, exclusivity should have already
+ * been verified by validate_change(). In such cases, excl_check must be
+ * false since the cs cpumasks are not yet updated.
+ *
+ * Return: Partition error code (PERR_NONE indicates success)
+ */
+static int __local_partition_update(struct cpuset *cs, struct cpumask *xcpus,
+ struct cpumask *excpus, struct tmpmasks *tmp,
+ bool excl_check)
+{
+ struct cpuset *parent = parent_cs(cs);
+ int part_error = PERR_NONE; /* Partition error? */
+ int old_prs, new_prs;
+ bool cpumask_updated = false;
+
+ lockdep_assert_held(&cpuset_mutex);
+ /* For local partition only */
+ if (WARN_ON_ONCE(is_remote_partition(cs) || cs_is_member(cs)))
+ return PERR_NONE;
+
+ old_prs = cs->partition_root_state;
+ /*
+ * If new_prs < 0, it might transition to valid partition state.
+ * Use absolute value for validation checks.
+ */
+ new_prs = old_prs < 0 ? -old_prs : old_prs;
+ part_error = validate_local_partition(cs, new_prs, excpus, excl_check);
+ if (part_error) {
+ local_partition_disable(cs, part_error, tmp);
+ return part_error;
+ }
+
+ cpumask_updated = !cpumask_equal(excpus, cs->effective_xcpus);
+ /* Nothing changes, return PERR_NONE */
+ if (new_prs == old_prs && !cpumask_updated)
+ return PERR_NONE;
+
+ /*
+ * If partition was previously invalid but now passes checks,
+ * re-enable it and update related flags.
+ * Otherwise, partition state doesn't change, only cpumasks change.
+ */
+ if (is_partition_invalid(cs)) {
+ partition_enable(cs, parent, new_prs, excpus);
+ update_partition_exclusive_flag(cs, new_prs);
+ update_partition_sd_lb(cs, old_prs);
+ } else {
+ cpumask_andnot(tmp->addmask, excpus, cs->effective_xcpus);
+ cpumask_andnot(tmp->delmask, cs->effective_xcpus, excpus);
+ partition_update(cs, new_prs, xcpus, excpus, tmp);
+ }
+
+ /*
+ * Propagate changes in parent's effective_cpus down the hierarchy.
+ */
+ if (cpumask_updated) {
+ cpuset_update_tasks_cpumask(parent, tmp->addmask);
+ update_sibling_cpumasks(parent, cs, tmp);
+ }
+ return PERR_NONE;
+}
+
+static int local_partition_update(struct cpuset *cs, struct tmpmasks *tmp)
+{
+ struct cpuset *parent = parent_cs(cs);
+
+ cpumask_and(tmp->new_cpus, user_xcpus(cs), parent->effective_xcpus);
+ return __local_partition_update(cs, NULL, tmp->new_cpus, tmp, true);
+}
+
/**
* update_parent_effective_cpumask - update effective_cpus mask of parent cpuset
* @cs: The cpuset that requests change in partition root state
@@ -2410,9 +2547,16 @@ static void update_cpumasks_hier(struct cpuset *cs, struct tmpmasks *tmp,
if (!css_tryget_online(&cp->css))
continue;
rcu_read_unlock();
+ /*
+ * The tmp->new_cpus may by modified.
+ * Update effective_cpus before passing tmp to other functions.
+ */
+ spin_lock_irq(&callback_lock);
+ cpumask_copy(cp->effective_cpus, tmp->new_cpus);
+ spin_unlock_irq(&callback_lock);
if (update_parent) {
- update_parent_effective_cpumask(cp, partcmd_update, NULL, tmp);
+ local_partition_update(cp, tmp);
/*
* The cpuset partition_root_state may become
* invalid. Capture it.
@@ -2421,7 +2565,6 @@ static void update_cpumasks_hier(struct cpuset *cs, struct tmpmasks *tmp,
}
spin_lock_irq(&callback_lock);
- cpumask_copy(cp->effective_cpus, tmp->new_cpus);
cp->partition_root_state = new_prs;
if (!cpumask_empty(cp->exclusive_cpus) && (cp != cs))
compute_excpus(cp, cp->effective_xcpus);
@@ -2613,8 +2756,8 @@ static void partition_cpus_change(struct cpuset *cs, struct cpuset *trialcs,
if (trialcs->prs_err)
local_partition_invalidate(cs, tmp);
else
- update_parent_effective_cpumask(cs, partcmd_update,
- trialcs->effective_xcpus, tmp);
+ __local_partition_update(cs, trialcs->exclusive_cpus,
+ trialcs->effective_xcpus, tmp, false);
}
}
@@ -4068,7 +4211,7 @@ static void cpuset_hotplug_update_tasks(struct cpuset *cs, struct tmpmasks *tmp)
else if (is_partition_valid(parent) && is_partition_invalid(cs) &&
!cpumask_empty(user_xcpus(cs))) {
partcmd = partcmd_update;
- update_parent_effective_cpumask(cs, partcmd, NULL, tmp);
+ local_partition_update(cs, tmp);
}
if (partcmd >= 0) {
--
2.34.1
Powered by blists - more mailing lists