[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <Z-Qu0stgvwHF9n3q@localhost.localdomain>
Date: Wed, 26 Mar 2025 17:44:02 +0100
From: Frederic Weisbecker <frederic@...nel.org>
To: Walter Chang (張維哲) <Walter.Chang@...iatek.com>
Cc: "paulmck@...nel.org" <paulmck@...nel.org>,
"joel@...lfernandes.org" <joel@...lfernandes.org>,
"boqun.feng@...il.com" <boqun.feng@...il.com>,
"vlad.wing@...il.com" <vlad.wing@...il.com>,
"kernel-team@...a.com" <kernel-team@...a.com>,
Alex Hoh (賀振坤) <Alex.Hoh@...iatek.com>,
Cheng-Jui Wang (王正睿) <Cheng-Jui.Wang@...iatek.com>,
"usamaarif642@...il.com" <usamaarif642@...il.com>,
"anna-maria@...utronix.de" <anna-maria@...utronix.de>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"neeraj.upadhyay@....com" <neeraj.upadhyay@....com>,
"tglx@...utronix.de" <tglx@...utronix.de>,
"leitao@...ian.org" <leitao@...ian.org>,
"urezki@...il.com" <urezki@...il.com>,
Freddy Hsin (辛恒豐) <Freddy.Hsin@...iatek.com>,
"qiang.zhang1211@...il.com" <qiang.zhang1211@...il.com>,
wsd_upstream <wsd_upstream@...iatek.com>,
Xinghua Yang (杨兴华) <Xinghua.Yang@...iatek.com>,
"rcu@...r.kernel.org" <rcu@...r.kernel.org>,
Chun-Hung Wu (巫駿宏) <Chun-hung.Wu@...iatek.com>
Subject: Re: [PATCH v4] hrtimers: Force migrate away hrtimers queued after
CPUHP_AP_HRTIMERS_DYING
Hi Walter Chang,
Le Wed, Mar 26, 2025 at 05:46:38AM +0000, Walter Chang (張維哲) a écrit :
> On Tue, 2025-01-21 at 09:08 -0800, Paul E. McKenney wrote:
> > On Sat, Jan 18, 2025 at 12:24:33AM +0100, Frederic Weisbecker wrote:
> > > hrtimers are migrated away from the dying CPU to any online target
> > > at
> > > the CPUHP_AP_HRTIMERS_DYING stage in order not to delay bandwidth
> > > timers
> > > handling tasks involved in the CPU hotplug forward progress.
> > >
> > > However wake ups can still be performed by the outgoing CPU after
> > > CPUHP_AP_HRTIMERS_DYING. Those can result again in bandwidth timers
> > > being armed. Depending on several considerations (crystal ball
> > > power management based election, earliest timer already enqueued,
> > > timer
> > > migration enabled or not), the target may eventually be the current
> > > CPU even if offline. If that happens, the timer is eventually
> > > ignored.
> > >
> > > The most notable example is RCU which had to deal with each and
> > > every of
> > > those wake-ups by deferring them to an online CPU, along with
> > > related
> > > workarounds:
> > >
> > > _ e787644caf76 (rcu: Defer RCU kthreads wakeup when CPU is dying)
> > > _ 9139f93209d1 (rcu/nocb: Fix RT throttling hrtimer armed from
> > > offline CPU)
> > > _ f7345ccc62a4 (rcu/nocb: Fix rcuog wake-up from offline softirq)
> > >
> > > The problem isn't confined to RCU though as the stop machine
> > > kthread
> > > (which runs CPUHP_AP_HRTIMERS_DYING) reports its completion at the
> > > end
> > > of its work through cpu_stop_signal_done() and performs a wake up
> > > that
> > > eventually arms the deadline server timer:
> > >
> > > WARNING: CPU: 94 PID: 588 at kernel/time/hrtimer.c:1086
> > > hrtimer_start_range_ns+0x289/0x2d0
> > > CPU: 94 UID: 0 PID: 588 Comm: migration/94 Not tainted
> > > Stopper: multi_cpu_stop+0x0/0x120 <-
> > > stop_machine_cpuslocked+0x66/0xc0
> > > RIP: 0010:hrtimer_start_range_ns+0x289/0x2d0
> > > Call Trace:
> > > <TASK>
> > > ? hrtimer_start_range_ns
> > > start_dl_timer
> > > enqueue_dl_entity
> > > dl_server_start
> > > enqueue_task_fair
> > > enqueue_task
> > > ttwu_do_activate
> > > try_to_wake_up
> > > complete
> > > cpu_stopper_thread
> > > smpboot_thread_fn
> > > kthread
> > > ret_from_fork
> > > ret_from_fork_asm
> > > </TASK>
> > >
> > > Instead of providing yet another bandaid to work around the
> > > situation,
> > > fix it from hrtimers infrastructure instead: always migrate away a
> > > timer to an online target whenever it is enqueued from an offline
> > > CPU.
> > >
> > > This will also allow to revert all the above RCU disgraceful hacks.
> > >
> > > Reported-by: Vlad Poenaru <vlad.wing@...il.com>
> > > Reported-by: Usama Arif <usamaarif642@...il.com>
> > > Fixes: 5c0930ccaad5 ("hrtimers: Push pending hrtimers away from
> > > outgoing CPU earlier")
> > > Closes: 20241213203739.1519801-1-usamaarif642@...il.com
> > > Signed-off-by: Frederic Weisbecker <frederic@...nel.org>
> > > Signed-off-by: Paul E. McKenney <paulmck@...nel.org>
> >
> > This passes over-holiday testing rcutorture, so, perhaps redundantly:
> >
> > Tested-by: Paul E. McKenney <paulmck@...nel.org>
>
> Hi,
>
> I encountered the same issue even after applying this patch.
> Below are the details of the warning and call trace.
>
>
> migration/3: ------------[ cut here ]------------
> migration/3: WARNING: CPU: 3 PID: 42 at kernel/time/hrtimer.c:1125
> enqueue_hrtimer+0x7c/0xec
> migration/3: CPU: 3 UID: 0 PID: 42 Comm: migration/3 Tainted: G
> OE 6.12.18-android16-0-g59cb5a849beb-4k #1
> 0b440e43fa7b24aaa3b7e6e5d2b938948e0cacdb
> migration/3: Stopper: multi_cpu_stop+0x0/0x184 <-
> stop_machine_cpuslocked+0xc0/0x15c
It's not the first time I get such a report on an out of tree
kernel. The problem is I don't know if the tainted modules are
involved. But something is probably making an offline CPU visible within
the hierarchy on get_nohz_timer_target(). And that new warning made
that visible.
Can you try this and tell us if the warning fires?
Thanks.
diff --git a/include/linux/sched/nohz.h b/include/linux/sched/nohz.h
index 6d67e9a5af6b..f49512628269 100644
--- a/include/linux/sched/nohz.h
+++ b/include/linux/sched/nohz.h
@@ -9,6 +9,7 @@
#if defined(CONFIG_SMP) && defined(CONFIG_NO_HZ_COMMON)
extern void nohz_balance_enter_idle(int cpu);
extern int get_nohz_timer_target(void);
+extern void assert_domain_online(void);
#else
static inline void nohz_balance_enter_idle(int cpu) { }
#endif
diff --git a/kernel/cpu.c b/kernel/cpu.c
index 07455d25329c..98c8f8408403 100644
--- a/kernel/cpu.c
+++ b/kernel/cpu.c
@@ -13,6 +13,7 @@
#include <linux/sched/isolation.h>
#include <linux/sched/task.h>
#include <linux/sched/smt.h>
+#include <linux/sched/nohz.h>
#include <linux/unistd.h>
#include <linux/cpu.h>
#include <linux/oom.h>
@@ -1277,6 +1278,7 @@ static int take_cpu_down(void *_param)
if (err < 0)
return err;
+ assert_domain_online();
/*
* Must be called from CPUHP_TEARDOWN_CPU, which means, as we are going
* down, that the current state is CPUHP_TEARDOWN_CPU - 1.
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 175a5a7ac107..88157b1645cc 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -1163,6 +1163,20 @@ void resched_cpu(int cpu)
#ifdef CONFIG_SMP
#ifdef CONFIG_NO_HZ_COMMON
+void assert_domain_online(void)
+{
+ int cpu = smp_processor_id();
+ int i;
+ struct sched_domain *sd;
+
+ guard(rcu)();
+
+ for_each_domain(cpu, sd) {
+ for_each_cpu(i, sched_domain_span(sd)) {
+ WARN_ON_ONCE(cpu_is_offline(i));
+ }
+ }
+}
/*
* In the semi idle case, use the nearest busy CPU for migrating timers
* from an idle CPU. This is good for power-savings.
Powered by blists - more mailing lists