lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20120629172320.GA2416@linux.vnet.ibm.com>
Date:	Fri, 29 Jun 2012 10:23:20 -0700
From:	"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
To:	Sasha Levin <levinsasha928@...il.com>
Cc:	Dave Jones <davej@...hat.com>,
	"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
	Peter Zijlstra <peterz@...radead.org>
Subject: Re: rcu: BUG: spinlock recursion on CPU#3, trinity-child19/5970

On Fri, Jun 29, 2012 at 12:09:44PM +0200, Sasha Levin wrote:
> Hi Paul,
> 
> I think I've stumbled on another bug that will increase your paranoia levels even further.
> 
> I got the following lockup when fuzzing with trinity inside a KVM tools guest, using latest linux-next.
> 
> It appears that it was caused by a03d6178 ("rcu: Move RCU grace-period cleanup into kthread"). This issue doesn't reproduce easily though, it took some fuzzing before hitting it.

Hmmm...  If the preemption at that point in __rcu_read_unlock() is
required to make this happen, then it would be pretty hard to hit.
I suspect that you can make it reproduce more quickly by putting
a udelay(10) or similar right after the assignment of INT_MIN to
t->rcu_read_lock_nesting in __rcu_read_unlock() in kernel/rcupdate.c.
Can this be reproduced while running with lockdep enabled?

Please see below for an untested patch that gets RCU out of this loop,
but it is quite possible that something else is involved here, so it would
be very good to get a lockdep run, if possible.  My concern stems from the
fact that interrupts were enabled during the call to __rcu_read_unlock()
-- otherwise it would not have been preempted -- so the runqueue locks
were presumably not held on entry to __rcu_read_unlock().

							Thanx, Paul

------------------------------------------------------------------------

rcu: Avoid grace-period kthread wakeups from unsafe environments

A soft lockup involving the run-queue locks involved the wakeup of
the grace-period kthread.  This commit therefore defers the wakeup
to softirq context.

Reported-by: Sasha Levin <levinsasha928@...il.com>
Signed-off-by: Paul E. McKenney <paulmck@...ux.vnet.ibm.com>

diff --git a/kernel/rcutree.c b/kernel/rcutree.c
index c0e0454..cd0076e 100644
--- a/kernel/rcutree.c
+++ b/kernel/rcutree.c
@@ -1262,6 +1262,15 @@ static int rcu_gp_kthread(void *arg)
 }
 
 /*
+ * Signal the RCU core to wake up the grace-period kthread, as we might
+ * not be in a context where it is safe to do the wakeup directly.
+ */
+static void rcu_wake_gp_kthread(struct rcu_state *rsp)
+{
+	ACCESS_ONCE(rsp->wake_gp_kthread) = 1;
+}
+
+/*
  * Start a new RCU grace period if warranted, re-initializing the hierarchy
  * in preparation for detecting the next grace period.  The caller must hold
  * the root node's ->lock, which is released before return.  Hard irqs must
@@ -1291,7 +1300,7 @@ rcu_start_gp(struct rcu_state *rsp, unsigned long flags)
 
 	rsp->gp_flags = 1;
 	raw_spin_unlock_irqrestore(&rnp->lock, flags);
-	wake_up(&rsp->gp_wq);
+	rcu_wake_gp_kthread(rsp);
 }
 
 /*
@@ -1306,7 +1315,7 @@ static void rcu_report_qs_rsp(struct rcu_state *rsp, unsigned long flags)
 {
 	WARN_ON_ONCE(!rcu_gp_in_progress(rsp));
 	raw_spin_unlock_irqrestore(&rcu_get_root(rsp)->lock, flags);
-	wake_up(&rsp->gp_wq);  /* Memory barrier implied by wake_up() path. */
+	rcu_wake_gp_kthread(rsp); /* smp_mb() implied by wakeup. */
 }
 
 /*
@@ -1889,6 +1898,12 @@ __rcu_process_callbacks(struct rcu_state *rsp)
 
 	WARN_ON_ONCE(rdp->beenonline == 0);
 
+	/* Awaken the grace-period kthread if needed. */
+	if (ACCESS_ONCE(rsp->wake_gp_kthread)) {
+		if (xchg(&rsp->wake_gp_kthread, 0)) /* Prevent wakeup storms. */
+			wake_up(&rsp->gp_wq);
+	}
+
 	/*
 	 * Advance callbacks in response to end of earlier grace
 	 * period that some other CPU ended.
@@ -2304,6 +2319,12 @@ static int __rcu_pending(struct rcu_state *rsp, struct rcu_data *rdp)
 		return 1;
 	}
 
+	/* Does the grace-period kthread need to be awakened? */
+	if (ACCESS_ONCE(rsp->wake_gp_kthread)) {
+		rdp->n_wake_gp_kthread++;
+		return 1;
+	}
+
 	/* nothing to do */
 	rdp->n_rp_need_nothing++;
 	return 0;
diff --git a/kernel/rcutree.h b/kernel/rcutree.h
index 280ad7c..89a190c 100644
--- a/kernel/rcutree.h
+++ b/kernel/rcutree.h
@@ -313,6 +313,7 @@ struct rcu_data {
 	unsigned long n_rp_cpu_needs_gp;
 	unsigned long n_rp_gp_completed;
 	unsigned long n_rp_gp_started;
+	unsigned long n_wake_gp_kthread;
 	unsigned long n_rp_need_nothing;
 
 	/* 6) _rcu_barrier() callback. */
@@ -381,6 +382,7 @@ struct rcu_state {
 	struct task_struct *gp_kthread;		/* Task for grace periods. */
 	wait_queue_head_t gp_wq;		/* Where GP task waits. */
 	int gp_flags;				/* Commands for GP task. */
+	int wake_gp_kthread;			/* Awaken GP task? */
 
 	/* End of fields guarded by root rcu_node's lock. */
 

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ