lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190807003501.GX28441@linux.ibm.com>
Date:   Tue, 6 Aug 2019 17:35:01 -0700
From:   "Paul E. McKenney" <paulmck@...ux.ibm.com>
To:     Joel Fernandes <joel@...lfernandes.org>
Cc:     rcu@...r.kernel.org, linux-kernel@...r.kernel.org,
        mingo@...nel.org, jiangshanlai@...il.com, dipankar@...ibm.com,
        akpm@...ux-foundation.org, mathieu.desnoyers@...icios.com,
        josh@...htriplett.org, tglx@...utronix.de, peterz@...radead.org,
        rostedt@...dmis.org, dhowells@...hat.com, edumazet@...gle.com,
        fweisbec@...il.com, oleg@...hat.com
Subject: Re: [PATCH RFC tip/core/rcu 02/14] rcu/nocb: Add bypass callback
 queueing

On Tue, Aug 06, 2019 at 08:03:13PM -0400, Joel Fernandes wrote:
> On Fri, Aug 02, 2019 at 08:14:49AM -0700, Paul E. McKenney wrote:
> > Use of the rcu_data structure's segmented ->cblist for no-CBs CPUs
> > takes advantage of unrelated grace periods, thus reducing the memory
> > footprint in the face of floods of call_rcu() invocations.  However,
> > the ->cblist field is a more-complex rcu_segcblist structure which must
> > be protected via locking.  Even though there are only three entities
> > which can acquire this lock (the CPU invoking call_rcu(), the no-CBs
> > grace-period kthread, and the no-CBs callbacks kthread), the contention
> > on this lock is excessive under heavy stress.
> > 
> > This commit therefore greatly reduces contention by provisioning
> > an rcu_cblist structure field named ->nocb_bypass within the
> > rcu_data structure.  Each no-CBs CPU is permitted only a limited
> > number of enqueues onto the ->cblist per jiffy, controlled by a new
> > nocb_nobypass_lim_per_jiffy kernel boot parameter that defaults to
> > about 16 enqueues per millisecond (16 * 1000 / HZ).  When that limit is
> > exceeded, the CPU instead enqueues onto the new ->nocb_bypass.
> 
> Looks quite interesting. I am guessing the not-no-CB (regular) enqueues don't
> need to use the same technique because both enqueues / callback execution are
> happening on same CPU..

That is the theory!  ;-)

> Still looking through patch but I understood the basic idea. Some nits below:
> 
> [snip]
> > diff --git a/kernel/rcu/tree.h b/kernel/rcu/tree.h
> > index 2c3e9068671c..e4df86db8137 100644
> > --- a/kernel/rcu/tree.h
> > +++ b/kernel/rcu/tree.h
> > @@ -200,18 +200,26 @@ struct rcu_data {
> >  	atomic_t nocb_lock_contended;	/* Contention experienced. */
> >  	int nocb_defer_wakeup;		/* Defer wakeup of nocb_kthread. */
> >  	struct timer_list nocb_timer;	/* Enforce finite deferral. */
> > +	unsigned long nocb_gp_adv_time;	/* Last call_rcu() CB adv (jiffies). */
> > +
> > +	/* The following fields are used by call_rcu, hence own cacheline. */
> > +	raw_spinlock_t nocb_bypass_lock ____cacheline_internodealigned_in_smp;
> > +	struct rcu_cblist nocb_bypass;	/* Lock-contention-bypass CB list. */
> > +	unsigned long nocb_bypass_first; /* Time (jiffies) of first enqueue. */
> > +	unsigned long nocb_nobypass_last; /* Last ->cblist enqueue (jiffies). */
> > +	int nocb_nobypass_count;	/* # ->cblist enqueues at ^^^ time. */
> 
> Can these and below fields be ifdef'd out if !CONFIG_RCU_NOCB_CPU so as to
> keep the size of struct smaller for benefit of systems that don't use NOCB?

Please see below...

> >  	/* The following fields are used by GP kthread, hence own cacheline. */
> >  	raw_spinlock_t nocb_gp_lock ____cacheline_internodealigned_in_smp;
> > -	bool nocb_gp_sleep;
> > -					/* Is the nocb GP thread asleep? */
> > +	struct timer_list nocb_bypass_timer; /* Force nocb_bypass flush. */
> > +	bool nocb_gp_sleep;		/* Is the nocb GP thread asleep? */
> 
> And these too, I think.
> 
> 
> >  	struct swait_queue_head nocb_gp_wq; /* For nocb kthreads to sleep on. */
> >  	bool nocb_cb_sleep;		/* Is the nocb CB thread asleep? */
> >  	struct task_struct *nocb_cb_kthread;
> >  	struct rcu_data *nocb_next_cb_rdp;
> >  					/* Next rcu_data in wakeup chain. */
> >  
> > -	/* The following fields are used by CB kthread, hence new cachline. */
> > +	/* The following fields are used by CB kthread, hence new cacheline. */
> >  	struct rcu_data *nocb_gp_rdp ____cacheline_internodealigned_in_smp;
> >  					/* GP rdp takes GP-end wakeups. */
> >  #endif /* #ifdef CONFIG_RCU_NOCB_CPU */

I believe that they in fact are all under CONFIG_RCU_NOCB_CPU.

> [snip]
> > +static void rcu_nocb_try_flush_bypass(struct rcu_data *rdp, unsigned long j)
> > +{
> > +	rcu_lockdep_assert_cblist_protected(rdp);
> > +	if (!rcu_segcblist_is_offloaded(&rdp->cblist) ||
> > +	    !rcu_nocb_bypass_trylock(rdp))
> > +		return;
> > +	WARN_ON_ONCE(!rcu_nocb_do_flush_bypass(rdp, NULL, j));
> > +}
> > +
> > +/*
> > + * See whether it is appropriate to use the ->nocb_bypass list in order
> > + * to control contention on ->nocb_lock.  A limited number of direct
> > + * enqueues are permitted into ->cblist per jiffy.  If ->nocb_bypass
> > + * is non-empty, further callbacks must be placed into ->nocb_bypass,
> > + * otherwise rcu_barrier() breaks.  Use rcu_nocb_flush_bypass() to switch
> > + * back to direct use of ->cblist.  However, ->nocb_bypass should not be
> > + * used if ->cblist is empty, because otherwise callbacks can be stranded
> > + * on ->nocb_bypass because we cannot count on the current CPU ever again
> > + * invoking call_rcu().  The general rule is that if ->nocb_bypass is
> > + * non-empty, the corresponding no-CBs grace-period kthread must not be
> > + * in an indefinite sleep state.
> > + *
> > + * Finally, it is not permitted to use the bypass during early boot,
> > + * as doing so would confuse the auto-initialization code.  Besides
> > + * which, there is no point in worrying about lock contention while
> > + * there is only one CPU in operation.
> > + */
> > +static bool rcu_nocb_try_bypass(struct rcu_data *rdp, struct rcu_head *rhp,
> > +				bool *was_alldone, unsigned long flags)
> > +{
> > +	unsigned long c;
> > +	unsigned long cur_gp_seq;
> > +	unsigned long j = jiffies;
> > +	long ncbs = rcu_cblist_n_cbs(&rdp->nocb_bypass);
> > +
> > +	if (!rcu_segcblist_is_offloaded(&rdp->cblist)) {
> > +		*was_alldone = !rcu_segcblist_pend_cbs(&rdp->cblist);
> > +		return false; /* Not offloaded, no bypassing. */
> > +	}
> > +	lockdep_assert_irqs_disabled();
> > +
> > +	// Don't use ->nocb_bypass during early boot.
> 
> Very minor nit: comment style should be /* */

I thought that Linus said that "//" was now OK.  Am I confused?

							Thanx, Paul

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ