[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <87eea29h1a.ffs@tglx>
Date: Mon, 06 Sep 2021 11:42:25 +0200
From: Thomas Gleixner <tglx@...utronix.de>
To: Randy Dunlap <rdunlap@...radead.org>,
Dave Chinner <david@...morbit.com>
Cc: "Darrick J. Wong" <djwong@...nel.org>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Dennis Zhou <dennis@...nel.org>, Tejun Heo <tj@...nel.org>,
linux-fsdevel <linux-fsdevel@...r.kernel.org>,
linux-xfs <linux-xfs@...r.kernel.org>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
Eric Sandeen <sandeen@...deen.net>,
Christoph Hellwig <hch@....de>,
Peter Zijlstra <peterz@...radead.org>,
Ingo Molnar <mingo@...nel.org>
Subject: Re: [GIT PULL] xfs: new code for 5.15
Randy,
On Sun, Sep 05 2021 at 19:11, Randy Dunlap wrote:
> On 9/5/21 4:28 PM, Thomas Gleixner wrote:
>> + * cpuhp_setup_state() and cpuhp_setup_state_cpuslocked() install the
>> + callbacks and invoke the @startup callback (if not NULL) for all online
>> + CPUs which have currently a state greater than the newly installed
>> + state. Depending on the state section the callback is either invoked on
>> + the current CPU (PREPARE section) or on each online CPU (ONLINE
>> + section) in the context of the CPU's hotplug thread.
>> +
>> + If a callback fails for CPU N then the teardown callback for CPU
>> + 0 .. N-1 is invoked to rollback the operation. The state setup fails,
>
> CPU 0? Does one of these fail since it's not an AP?
Yes. CPU 0 is not special in any way.
The point is that the hotplug state callbacks are set up late in the
boot process or during runtime when a module is loaded or some
functionality initialized on first use.
At that time the boot CPU (0) and usually the secondary CPUs are online
already. So the driver/subsystem has two ways to bring the per CPU
functionality into operation:
1) Initialize all per CPU state manually which often involves queuing
work on each online CPU or invoking SMP function calls on the online
CPUs and if all succeeds install the callbacks. If something goes
wrong on one of the CPUs then the state has to be cleaned up on the
CPUs which had their state set up correctly already.
This of course has to be done with cpus_read_lock() held to
serialize against a concurrent CPU hotplug operation.-
2) Let the hotplug core do that work. Setup a state with the
corresponding callbacks. The core invokes the startup callback on
all online CPUs (including 0) in the correct context:
for_each_online_cpu(cpu) {
ret = invoke_callback_on/for_cpu(cpu, startup);
if (ret)
goto err;
...
Any of these callback invocations can fail even the one on the boot
CPU. In case of failure on CPU0 there is nothing to clean up, but if
the Nth CPU callback fails then the state has been established for
CPU 0 to CPU N-1 already, e.g. memory allocation, hardware setup ...
So instead of returning with a half set up functionality, the core
does the rollback on CPU 0 to CPU N-1 by invoking the teardown
callback before returning the error code.
err:
for_each_online_cpu(cpu) {
if (startup_done[cpu])
invoke_callback_on/for_cpu(cpu, teardown);
}
That means the call site does not have to mop up the half
initialized state manually.
All of that is properly serialized against CPU hotplug operations.
>> +
>> + If a callback fails for CPU N then the teardown callback for CPU
>> + 0 .. N-1 is invoked to rollback the operation, the function fails and
>
> all except the Boot CPU?
See above.
Thanks,
tglx
Powered by blists - more mailing lists