[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20190212171423.8308-6-frederic@kernel.org>
Date: Tue, 12 Feb 2019 18:13:56 +0100
From: Frederic Weisbecker <frederic@...nel.org>
To: LKML <linux-kernel@...r.kernel.org>
Cc: Frederic Weisbecker <frederic@...nel.org>,
Sebastian Andrzej Siewior <bigeasy@...utronix.de>,
Peter Zijlstra <peterz@...radead.org>,
Mauro Carvalho Chehab <mchehab@...pensource.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
"David S . Miller" <davem@...emloft.net>,
Thomas Gleixner <tglx@...utronix.de>,
"Paul E . McKenney" <paulmck@...ux.vnet.ibm.com>,
Frederic Weisbecker <fweisbec@...il.com>,
Pavan Kondeti <pkondeti@...eaurora.org>,
Ingo Molnar <mingo@...nel.org>,
Joel Fernandes <joel@...lfernandes.org>
Subject: [PATCH 05/32] locking/lockdep: Prepare valid_state() to handle plain masks
mark_lock_irq() is going to deal with lock usages that gather multiple
softirq vectors at once. Therefore the validation through valid_state()
will need to handle expanded usage masks.
So enhance valid_state() to that purpose.
Signed-off-by: Frederic Weisbecker <frederic@...nel.org>
Cc: Mauro Carvalho Chehab <mchehab@...pensource.com>
Cc: Joel Fernandes <joel@...lfernandes.org>
Cc: Thomas Gleixner <tglx@...utronix.de>
Cc: Pavan Kondeti <pkondeti@...eaurora.org>
Cc: Paul E . McKenney <paulmck@...ux.vnet.ibm.com>
Cc: David S . Miller <davem@...emloft.net>
Cc: Ingo Molnar <mingo@...nel.org>
Cc: Sebastian Andrzej Siewior <bigeasy@...utronix.de>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>
Cc: Peter Zijlstra <peterz@...radead.org>
---
kernel/locking/lockdep.c | 44 +++++++++++++++++++++++++++++++++++-----
1 file changed, 39 insertions(+), 5 deletions(-)
diff --git a/kernel/locking/lockdep.c b/kernel/locking/lockdep.c
index 578dd57f70d3..6b625b70598a 100644
--- a/kernel/locking/lockdep.c
+++ b/kernel/locking/lockdep.c
@@ -2538,7 +2538,26 @@ print_usage_bug(struct task_struct *curr, struct held_lock *this,
static u64 lock_usage_mask(struct lock_usage *usage)
{
- return BIT(usage->bit);
+ int nr = 0;
+ u64 vectors = usage->vector;
+ u64 mask = 0ULL;
+
+ if (!vectors)
+ return lock_flag(usage->bit);
+
+ /* Only softirqs can have non-zero vectors */
+ WARN_ON_ONCE(usage->bit < LOCK_USED_IN_SOFTIRQ ||
+ usage->bit > LOCK_ENABLED_SOFTIRQ_READ);
+
+ while (vectors) {
+ long fs = __ffs64(vectors) + 1;
+
+ vectors >>= fs;
+ nr += fs;
+ mask |= lock_flag(usage->bit) << (4 * (nr - 1));
+ }
+
+ return mask;
}
/*
@@ -2546,10 +2565,23 @@ static u64 lock_usage_mask(struct lock_usage *usage)
*/
static inline int
valid_state(struct task_struct *curr, struct held_lock *this,
- enum lock_usage_bit new_bit, enum lock_usage_bit bad_bit)
+ u64 new_mask, u64 bad_mask)
{
- if (unlikely(hlock_class(this)->usage_mask & lock_flag(bad_bit)))
+ u64 bad_intersec;
+
+ bad_intersec = hlock_class(this)->usage_mask & bad_mask;
+
+ if (unlikely(bad_intersec)) {
+ enum lock_usage_bit new_bit, bad_bit;
+ int err;
+
+ err = find_exclusive_match(new_mask,
+ bad_intersec, &new_bit, &bad_bit);
+ if (WARN_ON_ONCE(err < 0))
+ return err;
+
return print_usage_bug(curr, this, bad_bit, new_bit);
+ }
return 1;
}
@@ -2746,7 +2778,8 @@ mark_lock_irq(struct task_struct *curr, struct held_lock *this,
* Validate that this particular lock does not have conflicting
* usage states.
*/
- if (!valid_state(curr, this, new_usage->bit, excl_usage.bit))
+ if (!valid_state(curr, this, lock_usage_mask(new_usage),
+ lock_usage_mask(&excl_usage)))
return 0;
/*
@@ -2762,7 +2795,8 @@ mark_lock_irq(struct task_struct *curr, struct held_lock *this,
*/
if (!read) {
excl_usage.bit += LOCK_USAGE_READ_MASK;
- if (!valid_state(curr, this, new_usage->bit, excl_usage.bit))
+ if (!valid_state(curr, this, lock_usage_mask(new_usage),
+ lock_usage_mask(&excl_usage)))
return 0;
if (STRICT_READ_CHECKS &&
--
2.17.1
Powered by blists - more mailing lists