lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Message-Id: <20130114154914.6d69eb27.akpm@linux-foundation.org> Date: Mon, 14 Jan 2013 15:49:14 -0800 From: Andrew Morton <akpm@...ux-foundation.org> To: Colin Cross <ccross@...roid.com> Cc: linux-kernel@...r.kernel.org, Don Zickus <dzickus@...hat.com>, Ingo Molnar <mingo@...nel.org>, Thomas Gleixner <tglx@...utronix.de>, liu chuansheng <chuansheng.liu@...el.com>, linux-arm-kernel@...ts.infradead.org, Russell King - ARM Linux <linux@....linux.org.uk>, Frederic Weisbecker <fweisbec@...il.com>, Tony Lindgren <tony@...mide.com> Subject: Re: [PATCH v2] hardlockup: detect hard lockups without NMIs using secondary cpus On Fri, 11 Jan 2013 13:51:48 -0800 Colin Cross <ccross@...roid.com> wrote: > Emulate NMIs on systems where they are not available by using timer > interrupts on other cpus. Each cpu will use its softlockup hrtimer > to check that the next cpu is processing hrtimer interrupts by > verifying that a counter is increasing. Seems sensible. > This patch is useful on systems where the hardlockup detector is not > available due to a lack of NMIs, for example most ARM SoCs. > Without this patch any cpu stuck with interrupts disabled can > cause a hardware watchdog reset with no debugging information, > but with this patch the kernel can detect the lockup and panic, > which can result in useful debugging info. But we don't get the target cpu's stack, yes? That's a pretty big loss. > > ... > > +#ifdef CONFIG_HARDLOCKUP_DETECTOR_OTHER_CPU > +static unsigned int watchdog_next_cpu(unsigned int cpu) > +{ > + cpumask_t cpus = watchdog_cpus; cpumask_t can be tremendously huge and putting one on the stack is risky. Can we use watchdog_cpus directly here? Perhaps with a lock? or take a copy into a static local, with a lock? > + unsigned int next_cpu; > + > + next_cpu = cpumask_next(cpu, &cpus); > + if (next_cpu >= nr_cpu_ids) > + next_cpu = cpumask_first(&cpus); > + > + if (next_cpu == cpu) > + return nr_cpu_ids; > + > + return next_cpu; > +} > + > +static int is_hardlockup_other_cpu(unsigned int cpu) > +{ > + unsigned long hrint = per_cpu(hrtimer_interrupts, cpu); > + > + if (per_cpu(hrtimer_interrupts_saved, cpu) == hrint) > + return 1; > + > + per_cpu(hrtimer_interrupts_saved, cpu) = hrint; > + return 0; > +} This could return a bool type. > +static void watchdog_check_hardlockup_other_cpu(void) > +{ > + unsigned int next_cpu; > + > + /* > + * Test for hardlockups every 3 samples. The sample period is > + * watchdog_thresh * 2 / 5, so 3 samples gets us back to slightly over > + * watchdog_thresh (over by 20%). > + */ > + if (__this_cpu_read(hrtimer_interrupts) % 3 != 0) > + return; The hardwired interval Seems Wrong. watchdog_thresh is tunable at runtime. The comment could do with some fleshing out. *why* do we want to test at an interval "slightly over watchdog_thresh"? What's going on here? > + /* check for a hardlockup on the next cpu */ > + next_cpu = watchdog_next_cpu(smp_processor_id()); > + if (next_cpu >= nr_cpu_ids) > + return; > + > + smp_rmb(); Mystery barrier (always) needs an explanatory comment, please. > + if (per_cpu(watchdog_nmi_touch, next_cpu) == true) { > + per_cpu(watchdog_nmi_touch, next_cpu) = false; > + return; > + } I wonder if a well-timed CPU plug/unplug could result in two CPUs simultaneously checking one other CPU's state. > + if (is_hardlockup_other_cpu(next_cpu)) { > + /* only warn once */ > + if (per_cpu(hard_watchdog_warn, next_cpu) == true) > + return; > + > + if (hardlockup_panic) > + panic("Watchdog detected hard LOCKUP on cpu %u", next_cpu); > + else > + WARN(1, "Watchdog detected hard LOCKUP on cpu %u", next_cpu); I suggest we use messages here which make it clear to people who read kernel output that this was triggered by hrtimers, not by NMI. Most importantly because people will need to know that the CPU which locked up is *not this CPU* and that any backtrace from the reporting CPU is misleading. Also, there was never any sense in making the LOCKUP all-caps ;) > + per_cpu(hard_watchdog_warn, next_cpu) = true; > + } else { > + per_cpu(hard_watchdog_warn, next_cpu) = false; > + } > +} > +#else > +static inline void watchdog_check_hardlockup_other_cpu(void) { return; } > +#endif > + > > ... > > --- a/lib/Kconfig.debug > +++ b/lib/Kconfig.debug > @@ -191,15 +191,27 @@ config LOCKUP_DETECTOR > The overhead should be minimal. A periodic hrtimer runs to > generate interrupts and kick the watchdog task every 4 seconds. > An NMI is generated every 10 seconds or so to check for hardlockups. > + If NMIs are not available on the platform, every 12 seconds the hm. Is the old "4 seconds" still true/accurate/complete? > + hrtimer interrupt on one cpu will be used to check for hardlockups > + on the next cpu. > > The frequency of hrtimer and NMI events and the soft and hard lockup > thresholds can be controlled through the sysctl watchdog_thresh. > > -config HARDLOCKUP_DETECTOR > +config HARDLOCKUP_DETECTOR_NMI > def_bool y > depends on LOCKUP_DETECTOR && !HAVE_NMI_WATCHDOG > depends on PERF_EVENTS && HAVE_PERF_EVENTS_NMI Confused. I'd have expected this to depend on HAVE_NMI_WATCHDOG, rather than -no-that. What does "HAVE_NMI_WATCHDOG" actually mean and what's happening here? > > ... > -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@...r.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists