[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200902194341.GK29330@paulmck-ThinkPad-P72>
Date: Wed, 2 Sep 2020 12:43:41 -0700
From: "Paul E. McKenney" <paulmck@...nel.org>
To: peterz@...radead.org
Cc: mingo@...nel.org, Will Deacon <will@...nel.org>,
linux-kernel@...r.kernel.org,
Masami Hiramatsu <mhiramat@...nel.org>,
Boqun Feng <boqun.feng@...il.com>, frederic@...nel.org,
joel@...lfernandes.org
Subject: Re: [PATCH] lockdep: Fix "USED" <- "IN-NMI" inversions
On Wed, Sep 02, 2020 at 06:03:23PM +0200, peterz@...radead.org wrote:
>
> During the LPC RCU BoF Paul asked how come the "USED" <- "IN-NMI"
> detector doesn't trip over rcu_read_lock()'s lockdep annotation.
>
> Looking into this I found a very embarrasing typo in
> verify_lock_unused():
>
> - if (!(class->usage_mask & LOCK_USED))
> + if (!(class->usage_mask & LOCKF_USED))
>
> fixing that will indeed cause rcu_read_lock() to insta-splat :/
>
> The above typo means that instead of testing for: 0x100 (1 <<
> LOCK_USED), we test for 8 (LOCK_USED), which corresponds to (1 <<
> LOCK_ENABLED_HARDIRQ).
>
> So instead of testing for _any_ used lock, it will only match any lock
> used with interrupts enabled.
>
> The rcu_read_lock() annotation uses .check=0, which means it will not
> set any of the interrupt bits and will thus never match.
>
> In order to properly fix the situation and allow rcu_read_lock() to
> correctly work, split LOCK_USED into LOCK_USED and LOCK_USED_READ and by
> having .read users set USED_READ and test USED, pure read-recursive
> locks are permitted.
>
> Fixes: f6f48e180404 ("lockdep: Teach lockdep about "USED" <- "IN-NMI" inversions")
> Signed-off-by: Peter Zijlstra (Intel) <peterz@...radead.org>
Thank you for looking into this!
Acked-by: Paul E. McKenney <paulmck@...nel.org>
> ---
> diff --git a/kernel/locking/lockdep.c b/kernel/locking/lockdep.c
> index cccf4bc759c6..454355c033d2 100644
> --- a/kernel/locking/lockdep.c
> +++ b/kernel/locking/lockdep.c
> @@ -4324,13 +4324,18 @@ static int separate_irq_context(struct task_struct *curr,
> static int mark_lock(struct task_struct *curr, struct held_lock *this,
> enum lock_usage_bit new_bit)
> {
> - unsigned int new_mask = 1 << new_bit, ret = 1;
> + unsigned int old_mask, new_mask, ret = 1;
>
> if (new_bit >= LOCK_USAGE_STATES) {
> DEBUG_LOCKS_WARN_ON(1);
> return 0;
> }
>
> + if (new_bit == LOCK_USED && this->read)
> + new_bit = LOCK_USED_READ;
> +
> + new_mask = 1 << new_bit;
> +
> /*
> * If already set then do not dirty the cacheline,
> * nor do any checks:
> @@ -4343,13 +4348,22 @@ static int mark_lock(struct task_struct *curr, struct held_lock *this,
> /*
> * Make sure we didn't race:
> */
> - if (unlikely(hlock_class(this)->usage_mask & new_mask)) {
> - graph_unlock();
> - return 1;
> - }
> + if (unlikely(hlock_class(this)->usage_mask & new_mask))
> + goto unlock;
>
> + old_mask = hlock_class(this)->usage_mask;
> hlock_class(this)->usage_mask |= new_mask;
>
> + /*
> + * Save one usage_traces[] entry and map both LOCK_USED and
> + * LOCK_USED_READ onto the same entry.
> + */
> + if (new_bit == LOCK_USED || new_bit == LOCK_USED_READ) {
> + if (old_mask & (LOCKF_USED | LOCKF_USED_READ))
> + goto unlock;
> + new_bit = LOCK_USED;
> + }
> +
> if (!(hlock_class(this)->usage_traces[new_bit] = save_trace()))
> return 0;
>
> @@ -4363,6 +4377,7 @@ static int mark_lock(struct task_struct *curr, struct held_lock *this,
> return 0;
> }
>
> +unlock:
> graph_unlock();
>
> /*
> @@ -5297,12 +5312,20 @@ static void verify_lock_unused(struct lockdep_map *lock, struct held_lock *hlock
> {
> #ifdef CONFIG_PROVE_LOCKING
> struct lock_class *class = look_up_lock_class(lock, subclass);
> + unsigned long mask = LOCKF_USED;
>
> /* if it doesn't have a class (yet), it certainly hasn't been used yet */
> if (!class)
> return;
>
> - if (!(class->usage_mask & LOCK_USED))
> + /*
> + * READ locks only conflict with USED, such that if we only ever use
> + * READ locks, there is no deadlock possible -- RCU.
> + */
> + if (!hlock->read)
> + mask |= LOCKF_USED_READ;
> +
> + if (!(class->usage_mask & mask))
> return;
>
> hlock->class_idx = class - lock_classes;
> diff --git a/kernel/locking/lockdep_internals.h b/kernel/locking/lockdep_internals.h
> index baca699b94e9..b0be1560ed17 100644
> --- a/kernel/locking/lockdep_internals.h
> +++ b/kernel/locking/lockdep_internals.h
> @@ -19,6 +19,7 @@ enum lock_usage_bit {
> #include "lockdep_states.h"
> #undef LOCKDEP_STATE
> LOCK_USED,
> + LOCK_USED_READ,
> LOCK_USAGE_STATES
> };
>
> @@ -40,6 +41,7 @@ enum {
> #include "lockdep_states.h"
> #undef LOCKDEP_STATE
> __LOCKF(USED)
> + __LOCKF(USED_READ)
> };
>
> #define LOCKDEP_STATE(__STATE) LOCKF_ENABLED_##__STATE |
Powered by blists - more mailing lists