[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <alpine.LFD.2.11.1405091640090.980@knanqh.ubzr>
Date: Fri, 9 May 2014 17:05:43 -0400 (EDT)
From: Nicolas Pitre <nicolas.pitre@...aro.org>
To: Russell King - ARM Linux <linux@....linux.org.uk>
cc: Doug Anderson <dianders@...omium.org>,
Viresh Kumar <viresh.kumar@...aro.org>,
"Rafael J. Wysocki" <rjw@...ysocki.net>,
Will Deacon <will.deacon@....com>,
John Stultz <john.stultz@...aro.org>,
David Riley <davidriley@...omium.org>,
"olof@...om.net" <olof@...om.net>,
Sonny Rao <sonnyrao@...omium.org>,
Richard Zhao <richard.zhao@...aro.org>,
Santosh Shilimkar <santosh.shilimkar@...com>,
Shawn Guo <shawn.guo@...aro.org>,
Stephen Boyd <sboyd@...eaurora.org>,
Marc Zyngier <marc.zyngier@....com>,
Stephen Warren <swarren@...dia.com>,
Paul Gortmaker <paul.gortmaker@...driver.com>,
"linux-arm-kernel@...ts.infradead.org"
<linux-arm-kernel@...ts.infradead.org>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] ARM: Don't ever downscale loops_per_jiffy in SMP
systems#
On Fri, 9 May 2014, Russell King - ARM Linux wrote:
> I'd much prefer just printing a warning at kernel boot time to report
> that the kernel is running with features which would make udelay() less
> than accurate.
What if there is simply no timer to rely upon, as in those cases where
interrupts are needed for time keeping to make progress? We should do
better than simply saying "sorry your kernel should irradicate every
udelay() usage to be reliable".
And I mean "reliable" which is not exactly the same as "accurate".
Reliable means "never *significantly* shorter".
> Remember, it should be usable for _short_ delays on slow machines as
> well as other stuff, and if we're going to start throwing stuff like
> the above at it, it's going to become very inefficient.
You said that udelay can be much longer than expected due to various
reasons.
You also said that the IRQ handler overhead during udelay calibration
makes actual delays slightli shorter than expected.
I'm suggesting the addition of a slight overhead that is much smaller
than the IRQ handler here. That shouldn't impact things masurably.
I'd certainly like Doug to run his udelay timing test with the following
patch to see if it solves the problem.
diff --git a/arch/arm/include/asm/delay.h b/arch/arm/include/asm/delay.h
index dff714d886..74fb571a55 100644
--- a/arch/arm/include/asm/delay.h
+++ b/arch/arm/include/asm/delay.h
@@ -57,11 +57,6 @@ extern void __bad_udelay(void);
__const_udelay((n) * UDELAY_MULT)) : \
__udelay(n))
-/* Loop-based definitions for assembly code. */
-extern void __loop_delay(unsigned long loops);
-extern void __loop_udelay(unsigned long usecs);
-extern void __loop_const_udelay(unsigned long);
-
/* Delay-loop timer registration. */
#define ARCH_HAS_READ_CURRENT_TIMER
extern void register_current_timer_delay(const struct delay_timer *timer);
diff --git a/arch/arm/lib/delay.c b/arch/arm/lib/delay.c
index 5306de3501..9150d31c2d 100644
--- a/arch/arm/lib/delay.c
+++ b/arch/arm/lib/delay.c
@@ -25,6 +25,11 @@
#include <linux/module.h>
#include <linux/timex.h>
+/* Loop-based definitions for assembly code. */
+extern void __loop_delay(unsigned long loops);
+extern void __loop_udelay(unsigned long usecs);
+extern void __loop_const_udelay(unsigned long);
+
/*
* Default to the loop-based delay implementation.
*/
@@ -34,6 +39,85 @@ struct arm_delay_ops arm_delay_ops = {
.udelay = __loop_udelay,
};
+#if defined(CONFIG_CPU_FREQ) && (defined(CONFIG_SMP) || defined(CONFIG_PREEMPT))
+
+#include <linux/cpufreq.h>
+
+/*
+ * Another CPU/thread might increase the CPU clock in the middle of
+ * the loop based delay routine and the newly scaled LPJ value won't be
+ * accounted for, resulting in a possibly significantly shorter delay than
+ * expected. Let's make sure this occurrence is trapped and compensated.
+ */
+
+static int __loop_seq;
+static unsigned int __loop_security_factor;
+
+#define __safe_loop_(type) \
+static void __safe_loop_##type(unsigned long val) \
+{ \
+ int seq_count = __loop_seq; \
+ __loop_##type(val); \
+ if (seq_count != __loop_seq) \
+ __loop_##type(val * __loop_security_factor); \
+}
+
+__safe_loop_(delay)
+__safe_loop_(const_udelay)
+__safe_loop_(udelay)
+
+static int cpufreq_callback(struct notifier_block *nb,
+ unsigned long val, void *data)
+{
+ struct cpufreq_freqs *freq = data;
+ unsigned int f;
+
+ if ((freq->flags & CPUFREQ_CONST_LOOPS) ||
+ freq->old >= freq->new)
+ return NOTIFY_OK;
+
+ switch (val) {
+ case CPUFREQ_PRECHANGE:
+ /* Remember the largest security factor ever needed */
+ f = DIV_ROUND_UP(freq->new, freq->old) - 1;
+ if (__loop_security_factor < f)
+ __loop_security_factor = f;
+ /* fallthrough */
+ case CPUFREQ_POSTCHANGE:
+ __loop_seq++;
+ }
+ return NOTIFY_OK;
+}
+
+static struct notifier_block cpufreq_notifier = {
+ .notifier_call = cpufreq_callback,
+};
+
+static int __init init_safe_loop_delays(void)
+{
+ int err;
+
+ /*
+ * Bail out if the default loop based implementation has
+ * already been replaced by something better.
+ */
+ if (arm_delay_ops.udelay != __loop_udelay)
+ return 0;
+
+ __loop_security_factor = 1;
+ err = cpufreq_register_notifier(&cpufreq_notifier,
+ CPUFREQ_TRANSITION_NOTIFIER);
+ if (!err) {
+ arm_delay_ops.delay = __safe_loop_delay;
+ arm_delay_ops.const_udelay = __safe_loop_const_udelay;
+ arm_delay_ops.udelay = __safe_loop_udelay;
+ }
+ return err;
+}
+core_initcall(init_safe_loop_delays);
+
+#endif
+
static const struct delay_timer *delay_timer;
static bool delay_calibrated;
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists