lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:   Wed, 22 Jun 2022 16:07:42 -0400
From:   Waiman Long <longman@...hat.com>
To:     john.p.donnelly@...cle.com, Hillf Danton <hdanton@...a.com>
Cc:     linux-kernel@...r.kernel.org
Subject: Re: [PATCH] locking/rwsem: Allow slowpath writer to ignore handoff
 bit if not set by first waiter

On 6/22/22 13:48, john.p.donnelly@...cle.com wrote:
> On 4/27/22 8:23 PM, Hillf Danton wrote:
>> On  Wed, 27 Apr 2022 13:31:24 -0400 Waiman Long wrote:
>>> With commit d257cc8cb8d5 ("locking/rwsem: Make handoff bit handling 
>>> more
>>> consistent"), the writer that sets the handoff bit can be interrupted
>>> out without clearing the bit if the wait queue isn't empty. This 
>>> disables
>>> reader and writer optimistic lock spinning and stealing.
>>>
>>> Now if a non-first writer in the queue is somehow woken up or first
>>> entering the waiting loop, it can't acquire the lock.  This is not
>>> the case before that commit as the writer that set the handoff bit
>>> will clear it when exiting out via the out_nolock path. This is less
>>> efficient as the busy rwsem stays in an unlock state for a longer time.
>>>
>>> This patch allows a non-first writer to ignore the handoff bit if it
>>> is not originally set or initiated by the first waiter.
>>>
>>> Fixes: d257cc8cb8d5 ("locking/rwsem: Make handoff bit handling more 
>>> consistent")
>>> Signed-off-by: Waiman Long <longman@...hat.com>
>>> ---
>>>   kernel/locking/rwsem.c | 30 ++++++++++++++++++++----------
>>>   1 file changed, 20 insertions(+), 10 deletions(-)
>>>
>>> diff --git a/kernel/locking/rwsem.c b/kernel/locking/rwsem.c
>>> index 9d1db4a54d34..65f0262f635e 100644
>>> --- a/kernel/locking/rwsem.c
>>> +++ b/kernel/locking/rwsem.c
>>> @@ -335,8 +335,6 @@ struct rwsem_waiter {
>>>       struct task_struct *task;
>>>       enum rwsem_waiter_type type;
>>>       unsigned long timeout;
>>> -
>>> -    /* Writer only, not initialized in reader */
>>>       bool handoff_set;
>>>   };
>>>   #define rwsem_first_waiter(sem) \
>>> @@ -459,10 +457,12 @@ static void rwsem_mark_wake(struct 
>>> rw_semaphore *sem,
>>>                * to give up the lock), request a HANDOFF to
>>>                * force the issue.
>>>                */
>>> -            if (!(oldcount & RWSEM_FLAG_HANDOFF) &&
>>> -                time_after(jiffies, waiter->timeout)) {
>>> -                adjustment -= RWSEM_FLAG_HANDOFF;
>>> -                lockevent_inc(rwsem_rlock_handoff);
>>> +            if (time_after(jiffies, waiter->timeout)) {
>>> +                if (!(oldcount & RWSEM_FLAG_HANDOFF)) {
>>> +                    adjustment -= RWSEM_FLAG_HANDOFF;
>>> +                    lockevent_inc(rwsem_rlock_handoff);
>>> +                }
>>> +                waiter->handoff_set = true;
>>>               }
>>
>> Handoff is tracked in both sem->count and waiter->handoff_set,
>>
>>> atomic_long_add(-adjustment, &sem->count);
>>> @@ -599,7 +599,7 @@ rwsem_del_wake_waiter(struct rw_semaphore *sem, 
>>> struct rwsem_waiter *waiter,
>>>   static inline bool rwsem_try_write_lock(struct rw_semaphore *sem,
>>>                       struct rwsem_waiter *waiter)
>>>   {
>>> -    bool first = rwsem_first_waiter(sem) == waiter;
>>> +    struct rwsem_waiter *first = rwsem_first_waiter(sem);
>>>       long count, new;
>>>         lockdep_assert_held(&sem->wait_lock);
>>> @@ -609,11 +609,20 @@ static inline bool rwsem_try_write_lock(struct 
>>> rw_semaphore *sem,
>>>           bool has_handoff = !!(count & RWSEM_FLAG_HANDOFF);
>>>             if (has_handoff) {
>>> -            if (!first)
>>> +            /*
>>> +             * Honor handoff bit and yield only when the first
>>> +             * waiter is the one that set it. Otherwisee, we
>>> +             * still try to acquire the rwsem.
>>> +             */
>>> +            if (first->handoff_set && (waiter != first))
>>>                   return false;
>>
>> and checked against both parties, thus in a simpler manner 
>> RWSEM_FLAG_HANDOFF
>> in sem->count means the first waiter has been waiting for lock long 
>> enough.
>>
>> Feel free to ignore the comment given the Fixes tag above.
>>
>> Hillf
>>>   -            /* First waiter inherits a previously set handoff bit */
>>> -            waiter->handoff_set = true;
>>> +            /*
>>> +             * First waiter can inherit a previously set handoff
>>> +             * bit and spin on rwsem if lock acquisition fails.
>>> +             */
>>> +            if (waiter == first)
>>> +                waiter->handoff_set = true;
>>>           }
>>>             new = count;
>>> @@ -1027,6 +1036,7 @@ rwsem_down_read_slowpath(struct rw_semaphore 
>>> *sem, long count, unsigned int stat
>>>       waiter.task = current;
>>>       waiter.type = RWSEM_WAITING_FOR_READ;
>>>       waiter.timeout = jiffies + RWSEM_WAIT_TIMEOUT;
>>> +    waiter.handoff_set = false;
>>>         raw_spin_lock_irq(&sem->wait_lock);
>>>       if (list_empty(&sem->wait_list)) {
>>> -- 
>>> 2.27.0
>
>
> Was this ever added ?
>
> I don't see it in
>
>
> a111daf0c53ae 2022-06-19 | Linux 5.19-rc3

This patch hasn't been taken up by upstream yet. I have reposted a v2 
with update to the patch description.

Cheers,
Longman

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ