lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <PH0PR11MB5880282CC101743255D68136DA249@PH0PR11MB5880.namprd11.prod.outlook.com>
Date:   Sun, 30 Jan 2022 02:18:20 +0000
From:   "Zhang, Qiang1" <qiang1.zhang@...el.com>
To:     "paulmck@...nel.org" <paulmck@...nel.org>
CC:     "frederic@...nel.org" <frederic@...nel.org>,
        "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>
Subject: RE: [PATCH] rcu: When rcuog kthreads is in polling mode, wakeup
 waitqueue is not requried


On Sat, Jan 29, 2022 at 05:55:34AM +0000, Zhang, Qiang1 wrote:
> 
> On Fri, Jan 28, 2022 at 11:13:46AM +0800, Zqiang wrote:
> > When grace period cleanup, the rcuog kthreads that waiting in sq 
> > waitqueue will be awakened, however if the 'rcu_nocb_poll' is set, 
> > the sq waitqueue always empty, so if 'rcu_nocb_poll' is set, return 
> > directly.
> 
> >This does decrease grace-period-cleanup overhead in kernels built with CONFIG_RCU_NOCB_CPU=y and booted with the rcu_nocb_poll kernel boot parameter set.  On the other hand, it increases grace-period-cleanup overhead in kernels built with CONFIG_RCU_NOCB_CPU=y but booted without the rcu_nocb_poll kernel boot parameter set.
> >
> >Last I checked, more kernels were booted without the rcu_nocb_poll kernel boot parameter set.  If this is still the case, this patch would slow things down for most systems.
> >
> >Or are there now lots of systems booted with rcu_nocb_poll?
> 
> Hi Paul
> I found that some of our customers configured rcu_nocb_poll startup parameters under Preempt-RT kernel.
> at each grace period cleanup, we'll all wakeup sq waitqueue, however 
> when rcuog kthreads is in polling mode, the wakeup operation doesn't required, because the sq waitqueue always empty.

>>>OK, fair enough.  But was there any difference in performance measurable at the system level?  Let's take a look at swake_up_all():>>>
>>>
>>>	void swake_up_all(struct swait_queue_head *q)
>>>	{
>>>		struct swait_queue *curr;
>>>		LIST_HEAD(tmp);
>>>
>>>		raw_spin_lock_irq(&q->lock);
>>>		list_splice_init(&q->task_list, &tmp);
>>>		while (!list_empty(&tmp)) {
>>>			curr = list_first_entry(&tmp, typeof(*curr), task_list);
>>>
>>>			wake_up_state(curr->task, TASK_NORMAL);
>>>			list_del_init(&curr->task_list);
>>>
>>>			if (list_empty(&tmp))
>>>				break;
>>>
>>>			raw_spin_unlock_irq(&q->lock);
>>>			raw_spin_lock_irq(&q->lock);
>>>		}
>>>		raw_spin_unlock_irq(&q->lock);
>>>	}
>>>
>>>If the list is empty, we acquire an uncontended lock, splice an empty list, check a pair of pointers for equality, and release that lock.
>>>We do this once per 16 CPUs per grace period, which normally will be every few milliseconds or less frequently.
>>>
>>>What is the system-level performance difference and how did you measure it?

Sorry  I ignored that. I don't measure performance differences at the system level,  

>>>
>>>Please don't get me wrong.  If this really is causing your users trouble, we clearly do need to fix it.  But if so, let's do so in a way that doesn't risk penalizing the many users who do not set rcu_nocb_poll.

Thank you for detailed analysis .   the original intention of my modification is avoid  unnecessary wake-up operations, 
when rcu_nocb_poll is set. In polling mode, the rcuog kthreads don't use sq waitqueue,  however, every time the grace period cleanup, 
all rnp nodes must be traversed for wake-up operation.  after that, I will do the test.

Thanks
Zqiang

>>>
>>>							Thanx, Paul
>>>
> Thanks,
> Zqiang
> 
> >
> >							Thanx, Paul
> 
> > Signed-off-by: Zqiang <qiang1.zhang@...el.com>
> > ---
> >  kernel/rcu/tree_nocb.h | 2 ++
> >  1 file changed, 2 insertions(+)
> > 
> > diff --git a/kernel/rcu/tree_nocb.h b/kernel/rcu/tree_nocb.h index
> > 636d0546a4e9..9e106c590e56 100644
> > --- a/kernel/rcu/tree_nocb.h
> > +++ b/kernel/rcu/tree_nocb.h
> > @@ -201,6 +201,8 @@ static void rcu_lockdep_assert_cblist_protected(struct rcu_data *rdp)
> >   */
> >  static void rcu_nocb_gp_cleanup(struct swait_queue_head *sq)   
> > +	if (rcu_nocb_poll)
> > +		return;
> >  	swake_up_all(sq);
> >  }
> >  
> > --
> > 2.25.1
> > 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ