[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CA+55aFx0NjiHM5Aw0N7xDwRcnHOiaceV2iYuGOU1uM3FUyf+Lg@mail.gmail.com>
Date: Sun, 27 Aug 2017 14:40:48 -0700
From: Linus Torvalds <torvalds@...ux-foundation.org>
To: Tim Chen <tim.c.chen@...ux.intel.com>
Cc: Mel Gorman <mgorman@...hsingularity.net>,
Peter Zijlstra <peterz@...radead.org>,
Ingo Molnar <mingo@...e.hu>, Andi Kleen <ak@...ux.intel.com>,
Kan Liang <kan.liang@...el.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Johannes Weiner <hannes@...xchg.org>, Jan Kara <jack@...e.cz>,
Christopher Lameter <cl@...ux.com>,
"Eric W . Biederman" <ebiederm@...ssion.com>,
Davidlohr Bueso <dave@...olabs.net>,
linux-mm <linux-mm@...ck.org>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH 2/2 v2] sched/wait: Introduce lock breaker in wake_up_page_bit
On Sat, Aug 26, 2017 at 11:15 AM, Linus Torvalds
<torvalds@...ux-foundation.org> wrote:
>
> So how about just this fairly trivial patch?
So I just committed that trivial patch, because I think it's right,
but more importantly because I think I found a real and non-trivial
fundamental problem.
The reason I found it is actually that I was thinking about this
patch, and how the WQ_FLAG_EXCLUSIVE ordering matters.
And I don't really think the WQ_FLAG_EXCLUSIVE ordering matters all
that much, but just *thinking* about it made me realize that the code
is broken.
In particular, this caller:
int __lock_page_killable(struct page *__page)
{
struct page *page = compound_head(__page);
wait_queue_head_t *q = page_waitqueue(page);
return wait_on_page_bit_common(q, page, PG_locked, TASK_KILLABLE, true);
}
EXPORT_SYMBOL_GPL(__lock_page_killable);
is completely broken crap.
Why?
It's the combination of "TASK_KILLABLE" and "true" that is broken.
Always has been broken, afaik.
The "true" is that "bool lock" argument, and when it is set, we set
the WQ_FLAG_EXCLUSIVE bit.
But that bit - by definition, that's the whole point - means that the
waking side only wakes up *one* waiter.
So there's a race in anybody who uses __lock_page_killable().
The race goes like this:
thread1 thread2 thread3
---- ---- ----
.. CPU1 ...
__lock_page_killable
wait_on_page_bit_common()
get wq lock
__add_wait_queue_entry_tail_exclusive()
set_current_state(TASK_KILLABLE);
release wq lock
io_schedule
... CPU2 ...
__lock_page[_killable]()
wait_on_page_bit_common()
get wq lock
__add_wait_queue_entry_tail_exclusive()
set_current_state(TASK_KILLABLE);
release wq lock
io_schedule
... CPU3 ...
unlock_page()
wake_up_page_bit(page, PG_Locked)
wakes up CPU1 _only_
... lethal signal for thread1 happens ...
if (unlikely(signal_pending_state(state, current))) {
ret = -EINTR;
break;
}
End result: page is unlocked, CPU3 is waiting, nothing will wake CPU3 up.
Of course, if we have multiple threads all locking that page
concurrently, we probably will have *another* thread lock it
(successfully), and then when it unlocks it thread3 does get woken up
eventually.
But the keyword is "eventually". It could be a long while,
particularly if we don't lock the page *all* the time, just
occasionally.
So it might be a while, and it might explain how some waiters might queue up.
And who does __lock_page_killable? Page faults.
And who does a lot of page faults and page locking? That NUMA load from hell.
Does it have lethal signals, though? Probably not. That lethal signal
case really is unusual.
So I'm not saying that this is actually THE BUG. In particular,
despite that race, the page actually *is* unlocked afterwards. It's
just that one of the threads that wanted the lock didn't get notified
of it. So it doesn't really explain how non-locking waiters (ie the
people who don't do migrations, just wait for the migration entry)
would queue up.
But it sure looks like a real bug to me.
Basically, if you ask for anm exclusive wakeup, you *have* to take the
resource you are waiting for. Youl can't just say "never mind, I'll
return -EINTR".
I don't see a simple fix for this yet other than perhaps adding a
wakeup to the "ret = -EINTR; break" case.
Does anybody else see anything? Or maybe see a reason why this
wouldn't be a bug in the first place?
Anyway, I am officially starting to hate that page wait code. I've
stared at it for days now, and I am not getting more fond of it.
Linus
Powered by blists - more mailing lists