[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20230119110220.kphftcehehhi5l5u@techsingularity.net>
Date: Thu, 19 Jan 2023 11:02:20 +0000
From: Mel Gorman <mgorman@...hsingularity.net>
To: Sebastian Andrzej Siewior <bigeasy@...utronix.de>
Cc: Peter Zijlstra <peterz@...radead.org>,
Thomas Gleixner <tglx@...utronix.de>,
Ingo Molnar <mingo@...nel.org>,
Davidlohr Bueso <dave@...olabs.net>,
Linux-RT <linux-rt-users@...r.kernel.org>,
LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v2] locking/rwbase: Prevent indefinite writer starvation
On Thu, Jan 19, 2023 at 09:25:56AM +0100, Sebastian Andrzej Siewior wrote:
> On 2023-01-18 17:31:30 [+0000], Mel Gorman wrote:
> > If we drop that "we prefer the RT reader" then it would block on the
> > > RTmutex. It will _still_ be preferred over the writer because it will be
> > > enqueued before the writer in the queue due to its RT priority. The only
> > > downside is that it has to wait until all readers are left.
> >
> > The writer has to wait until all the readers have left anyway.
>
> I meant the READER in case it has RT priority. It will enqueue itself on
> the RTmutex, first in line, and wait until all other READER leave.
>
Ah.
> > If I understand you correctly, the patch becomes this;
>
> exactly.
>
> > --8<--
> ???
> > This patch records a timestamp when the first writer is blocked. DT /
>
> s/DT/DL
>
Fixed.
> > RT tasks can continue to take the lock for read as long as readers exist
> > indefinitely. Other readers can acquire the read lock unless a writer
> > has been blocked for a minimum of 4ms. This is sufficient to allow the
> > dio_truncate test case to complete within the 30 minutes timeout.
> >
> > Signed-off-by: Mel Gorman <mgorman@...hsingularity.net>
> > ---
> ???
> > diff --git a/kernel/locking/rwbase_rt.c b/kernel/locking/rwbase_rt.c
> > index c201aadb9301..84c5e4e4d25b 100644
> > --- a/kernel/locking/rwbase_rt.c
> > +++ b/kernel/locking/rwbase_rt.c
> > @@ -74,9 +106,11 @@ static int __sched __rwbase_read_lock(struct rwbase_rt *rwb,
> > raw_spin_lock_irq(&rtm->wait_lock);
> > /*
> > * Allow readers, as long as the writer has not completely
> > - * acquired the semaphore for write.
> > + * acquired the semaphore for write and reader bias is still
> > + * allowed.
> > */
> > - if (atomic_read(&rwb->readers) != WRITER_BIAS) {
> > + if (atomic_read(&rwb->readers) != WRITER_BIAS &&
> > + rwbase_allow_reader_bias(rwb)) {
> > atomic_inc(&rwb->readers);
> > raw_spin_unlock_irq(&rtm->wait_lock);
> > return 0;
> > @@ -264,12 +298,17 @@ static int __sched rwbase_write_lock(struct rwbase_rt *rwb,
> > if (__rwbase_write_trylock(rwb))
> > break;
> >
> > + /* Record first new read/write contention. */
> > + set_writer_blocked(rwb);
> > +
> > raw_spin_unlock_irqrestore(&rtm->wait_lock, flags);
> > rwbase_schedule();
> > raw_spin_lock_irqsave(&rtm->wait_lock, flags);
> >
> > set_current_state(state);
> > }
> > +
> > + rwb->waiter_timeout = 0;
>
> Regarding memory ordering and ordering in general:
> - Should the writer leave from rwbase_schedule() due to a signal then
> set_writer_blocked() sets a timeout but it is not cleared on the
> signal leave.
>
You're correct, it should be reset in the signal check block before the
wait_lock is released by __rwbase_write_unlock. I considered different
ways to avoid multiple reset points but it was untidy.
> - There is only writer in that for loop within rwbase_write_lock()
> because only one writer can own the rtmutex at a time. (A second
> writer blocks on the RTmutex and needs to wait, I may have spread some
> confusion earler). Therefore it should be okay to unconditionally set
> the timeout (instead of checking for zero).
>
Ah ok, I see now or at least I think I do.
> - Once the writer removes READER_BIAS, it forces the reader into the
> slowpath.
Removed in __rwbase_write_trylock IIUC
> At that time the writer does not own the wait_lock meaning
> the reader _could_ check the timeout before writer had a chance to set
> it. The worst thing is probably that if jiffies does not have the
> highest bit set then it will always disable the reader bias here.
> The easiest thing is probably to check timeout vs 0 and ensure on the
> writer side that the lowest bit is always set (in the unlikely case it
> will end up as zero).
>
I am missing something important. On the read side, we have
raw_spin_lock_irq(&rtm->wait_lock);
/*
* Allow readers, as long as the writer has not completely
* acquired the semaphore for write and reader bias is still
* allowed.
*/
if (atomic_read(&rwb->readers) != WRITER_BIAS &&
rwbase_allow_reader_bias(rwb)) {
atomic_inc(&rwb->readers);
raw_spin_unlock_irq(&rtm->wait_lock);
return 0;
}
So rtm->wait_lock is held and both the WRITER_BASE and timeout are checked
under the lock. On the write side it's also held when the timeout is
updated here
raw_spin_lock_irqsave(&rtm->wait_lock, flags);
...
for (;;) {
...
/* Record timeout when reader bias is ignored. */
rwb->waiter_timeout = jiffies + RWBASE_RT_WAIT_TIMEOUT;
raw_spin_unlock_irqrestore(&rtm->wait_lock, flags);
rwbase_schedule();
raw_spin_lock_irqsave(&rtm->wait_lock, flags);
set_current_state(state);
}
rwb->waiter_timeout = 0;
...
out_unlock:
raw_spin_unlock_irqrestore(&rtm->wait_lock, flags);
(it's also now held in the signal block when it is cleared)
I'm not seeing exactly what the problem is unless it's an issue in the
fast path but I think the atomic_try_cmpxchg_acquire works there.
In the absense of wait_lock, I guess there would be a potential
race between the atomic_read(&rwb->readers) != WRITER_BIAS and
rwbase_allow_reader_bias(rwb) but that shouldn't be the case now. Even
if it wasn't locked, it might allow a new reader through but it'd be
a transient problem and writer starvation should be prevented once the
timeout is observed.
--
Mel Gorman
SUSE Labs
Powered by blists - more mailing lists