lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Fri, 06 May 2011 22:04:27 +0200
From:	Eric Dumazet <eric.dumazet@...il.com>
To:	Andi Kleen <andi@...stfloor.org>
Cc:	Thomas Gleixner <tglx@...utronix.de>,
	john stultz <johnstul@...ibm.com>,
	lkml <linux-kernel@...r.kernel.org>,
	Paul Mackerras <paulus@...ba.org>,
	"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>,
	Anton Blanchard <anton@...ba.org>, Ingo Molnar <mingo@...e.hu>
Subject: Re: [RFC] time: xtime_lock is held too long

Le vendredi 06 mai 2011 à 21:26 +0200, Eric Dumazet a écrit :
> Le vendredi 06 mai 2011 à 19:50 +0200, Andi Kleen a écrit :
> > On Fri, May 06, 2011 at 07:42:47PM +0200, Eric Dumazet wrote:
> > > Le vendredi 06 mai 2011 à 18:59 +0200, Andi Kleen a écrit :
> > > 
> > > > If you have a better way to make it faster please share it.
> > > 
> > > Ideally we could use RCU :)
> > 
> > Hmm, I didn't think my case had a lot of loops in the seqlock -- just
> > expensive cache misses -- but I should double check.
> > 
> > For the lots of loop case we probably need to understand first why you
> > iterate that often.
> 
> Yep, I'll try to investigate on this
> 

So apparently some calls to tick_do_update_jiffies64() are pretty
expensive :

[  369.334399] maxhold=1191627
[  369.334450] ------------[ cut here ]------------
[  369.334505] WARNING: at include/linux/seqlock.h:81 tick_do_update_jiffies64+0x169/0x1c0()
[  369.334579] Hardware name: ProLiant BL460c G1
[  369.334631] Modules linked in: xt_hashlimit ipmi_devintf af_packet ipmi_si ipmi_msghandler hpilo tg3 bonding
[  369.334989] Pid: 0, comm: swapper Tainted: G        W   2.6.39-rc6-00097-g6ac1576-dirty #547
[  369.335000] Call Trace:
[  369.335000]  [<c05b395f>] ? printk+0x18/0x21
[  369.335000]  [<c023b90d>] warn_slowpath_common+0x6d/0xa0
[  369.335000]  [<c02661f9>] ? tick_do_update_jiffies64+0x169/0x1c0
[  369.335000]  [<c02661f9>] ? tick_do_update_jiffies64+0x169/0x1c0
[  369.335000]  [<c023b95d>] warn_slowpath_null+0x1d/0x20
[  369.335000]  [<c02661f9>] tick_do_update_jiffies64+0x169/0x1c0
[  369.335000]  [<c0266c18>] tick_check_idle+0x78/0xa0
[  369.335000]  [<c0241734>] irq_enter+0x54/0x60
[  369.335000]  [<c021ac3c>] smp_apic_timer_interrupt+0x2c/0x90
[  369.335000]  [<c025b452>] ? hrtimer_start+0x22/0x30
[  369.335000]  [<c05b6d2a>] apic_timer_interrupt+0x2a/0x30
[  369.335000]  [<c02098e4>] ? mwait_idle+0x84/0x180
[  369.335000]  [<c02015f6>] cpu_idle+0x46/0x80
[  369.335000]  [<c059f41d>] rest_init+0x5d/0x70
[  369.335000]  [<c07c8710>] start_kernel+0x2d6/0x2dc
[  369.335000]  [<c07c81c2>] ? loglevel+0x1a/0x1a
[  369.335000]  [<c07c80bd>] i386_start_kernel+0xbd/0xc5
[  369.335000] ---[ end trace 4eaa2a86a8e2da80 ]---


I used following hack/patch :

diff --git a/include/linux/seqlock.h b/include/linux/seqlock.h
index e98cd2e..1e3ba24 100644
--- a/include/linux/seqlock.h
+++ b/include/linux/seqlock.h
@@ -32,6 +32,8 @@
 typedef struct {
 	unsigned sequence;
 	spinlock_t lock;
+	unsigned long long maxhold;
+	unsigned long long t0;
 } seqlock_t;
 
 /*
@@ -48,6 +50,7 @@ typedef struct {
 	do {						\
 		(x)->sequence = 0;			\
 		spin_lock_init(&(x)->lock);		\
+		(x)->maxhold = 0;			\
 	} while (0)
 
 #define DEFINE_SEQLOCK(x) \
@@ -61,13 +64,22 @@ static inline void write_seqlock(seqlock_t *sl)
 {
 	spin_lock(&sl->lock);
 	++sl->sequence;
+	sl->t0 = __native_read_tsc();
 	smp_wmb();
 }
 
 static inline void write_sequnlock(seqlock_t *sl)
 {
+	unsigned long long t1;
 	smp_wmb();
 	sl->sequence++;
+	t1 = __native_read_tsc();
+	t1 -= sl->t0;
+	if (unlikely(t1 > sl->maxhold)) {
+		sl->maxhold = t1;
+		pr_err("maxhold=%llu\n", t1);
+		WARN_ON(1);
+	}
 	spin_unlock(&sl->lock);
 }
 
@@ -77,13 +89,14 @@ static inline int write_tryseqlock(seqlock_t *sl)
 
 	if (ret) {
 		++sl->sequence;
+		sl->t0 = __native_read_tsc();
 		smp_wmb();
 	}
 	return ret;
 }
 
 /* Start of read calculation -- fetch last complete writer token */
-static __always_inline unsigned read_seqbegin(const seqlock_t *sl)
+static __always_inline unsigned __read_seqbegin(const seqlock_t *sl, unsigned long *loop)
 {
 	unsigned ret;
 
@@ -91,6 +104,8 @@ repeat:
 	ret = sl->sequence;
 	smp_rmb();
 	if (unlikely(ret & 1)) {
+		if (loop)
+			(*loop)++;
 		cpu_relax();
 		goto repeat;
 	}
@@ -98,6 +113,11 @@ repeat:
 	return ret;
 }
 
+static __always_inline unsigned read_seqbegin(const seqlock_t *sl)
+{
+	return __read_seqbegin(sl, NULL);
+}
+
 /*
  * Test if reader processed invalid data.
  *
diff --git a/kernel/time/timekeeping.c b/kernel/time/timekeeping.c
index 8ad5d57..db84027 100644
--- a/kernel/time/timekeeping.c
+++ b/kernel/time/timekeeping.c
@@ -237,20 +237,28 @@ void getnstimeofday(struct timespec *ts)
 
 EXPORT_SYMBOL(getnstimeofday);
 
+static DEFINE_PER_CPU(unsigned long, loopmax);
 ktime_t ktime_get(void)
 {
 	unsigned int seq;
 	s64 secs, nsecs;
+	unsigned long flags, loop = 0;
 
 	WARN_ON(timekeeping_suspended);
 
+	local_irq_save(flags);
 	do {
-		seq = read_seqbegin(&xtime_lock);
+		seq = __read_seqbegin(&xtime_lock, &loop);
 		secs = xtime.tv_sec + wall_to_monotonic.tv_sec;
 		nsecs = xtime.tv_nsec + wall_to_monotonic.tv_nsec;
 		nsecs += timekeeping_get_ns();
 
 	} while (read_seqretry(&xtime_lock, seq));
+	if (unlikely(loop > __this_cpu_read(loopmax))) {
+		this_cpu_write(loopmax, loop);
+		pr_err("cpu%d ktime_get() looped %lu times\n", smp_processor_id(), loop);
+	}
+	local_irq_restore(flags);
 	/*
 	 * Use ktime_set/ktime_add_ns to create a proper ktime on
 	 * 32-bit architectures without CONFIG_KTIME_SCALAR.


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ