lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-Id: <20160719030045.GR7094@linux.vnet.ibm.com>
Date:	Mon, 18 Jul 2016 20:00:45 -0700
From:	"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
To:	Stephen Rothwell <sfr@...b.auug.org.au>
Cc:	Thomas Gleixner <tglx@...utronix.de>, Ingo Molnar <mingo@...e.hu>,
	"H. Peter Anvin" <hpa@...or.com>,
	Peter Zijlstra <peterz@...radead.org>,
	linux-next@...r.kernel.org, linux-kernel@...r.kernel.org,
	Anna-Maria Gleixner <anna-maria@...utronix.de>
Subject: Re: linux-next: manual merge of the rcu tree with the tip tree

On Mon, Jul 18, 2016 at 03:26:28PM +1000, Stephen Rothwell wrote:
> Hi Paul,
> 
> Today's linux-next merge of the rcu tree got a conflict in:
> 
>   kernel/rcu/tree.c
> 
> between commit:
> 
>   4df8374254ea ("rcu: Convert rcutree to hotplug state machine")
> 
> from the tip tree and commit:
> 
>   2a84cde733b0 ("rcu: Exact CPU-online tracking for RCU")
> 
> from the rcu tree.
> 
> I fixed it up (see below) and can carry the fix as necessary. This
> is now fixed as far as linux-next is concerned, but any non trivial
> conflicts should be mentioned to your upstream maintainer when your tree
> is submitted for merging.  You may also want to consider cooperating
> with the maintainer of the conflicting tree to minimise any particularly
> complex conflicts.

Ah, that is in -tip now?  Easier to test, then!

Thank you, the patch looks good at first glance, but will beat it up
a bit.

							Thanx, Paul

> -- 
> Cheers,
> Stephen Rothwell
> 
> diff --cc kernel/rcu/tree.c
> index e5164deb51e1,5663d1e899d3..000000000000
> --- a/kernel/rcu/tree.c
> +++ b/kernel/rcu/tree.c
> @@@ -3812,54 -3809,34 +3809,80 @@@ int rcutree_prepare_cpu(unsigned int cp
>   
>   	for_each_rcu_flavor(rsp)
>   		rcu_init_percpu_data(cpu, rsp);
>  +
>  +	rcu_prepare_kthreads(cpu);
>  +	rcu_spawn_all_nocb_kthreads(cpu);
>  +
>  +	return 0;
>  +}
>  +
>  +static void rcutree_affinity_setting(unsigned int cpu, int outgoing)
>  +{
>  +	struct rcu_data *rdp = per_cpu_ptr(rcu_state_p->rda, cpu);
>  +
>  +	rcu_boost_kthread_setaffinity(rdp->mynode, outgoing);
>  +}
>  +
>  +int rcutree_online_cpu(unsigned int cpu)
>  +{
>  +	sync_sched_exp_online_cleanup(cpu);
>  +	rcutree_affinity_setting(cpu, -1);
>  +	return 0;
>  +}
>  +
>  +int rcutree_offline_cpu(unsigned int cpu)
>  +{
>  +	rcutree_affinity_setting(cpu, cpu);
>  +	return 0;
>  +}
>  +
>  +
>  +int rcutree_dying_cpu(unsigned int cpu)
>  +{
>  +	struct rcu_state *rsp;
>  +
>  +	for_each_rcu_flavor(rsp)
>  +		rcu_cleanup_dying_cpu(rsp);
>  +	return 0;
>  +}
>  +
>  +int rcutree_dead_cpu(unsigned int cpu)
>  +{
>  +	struct rcu_state *rsp;
>  +
>  +	for_each_rcu_flavor(rsp) {
>  +		rcu_cleanup_dead_cpu(cpu, rsp);
>  +		do_nocb_deferred_wakeup(per_cpu_ptr(rsp->rda, cpu));
>  +	}
>  +	return 0;
>   }
>   
> + /*
> +  * Mark the specified CPU as being online so that subsequent grace periods
> +  * (both expedited and normal) will wait on it.  Note that this means that
> +  * incoming CPUs are not allowed to use RCU read-side critical sections
> +  * until this function is called.  Failing to observe this restriction
> +  * will result in lockdep splats.
> +  */
> + void rcu_cpu_starting(unsigned int cpu)
> + {
> + 	unsigned long flags;
> + 	unsigned long mask;
> + 	struct rcu_data *rdp;
> + 	struct rcu_node *rnp;
> + 	struct rcu_state *rsp;
> + 
> + 	for_each_rcu_flavor(rsp) {
> + 		rdp = this_cpu_ptr(rsp->rda);
> + 		rnp = rdp->mynode;
> + 		mask = rdp->grpmask;
> + 		raw_spin_lock_irqsave_rcu_node(rnp, flags);
> + 		rnp->qsmaskinitnext |= mask;
> + 		rnp->expmaskinitnext |= mask;
> + 		raw_spin_unlock_irqrestore_rcu_node(rnp, flags);
> + 	}
> + }
> + 
>   #ifdef CONFIG_HOTPLUG_CPU
>   /*
>    * The CPU is exiting the idle loop into the arch_cpu_idle_dead()
> @@@ -4208,9 -4231,12 +4231,11 @@@ void __init rcu_init(void
>   	 * this is called early in boot, before either interrupts
>   	 * or the scheduler are operational.
>   	 */
>  -	cpu_notifier(rcu_cpu_notify, 0);
>   	pm_notifier(rcu_pm_notify, 0);
> - 	for_each_online_cpu(cpu)
> + 	for_each_online_cpu(cpu) {
>  -		rcu_cpu_notify(NULL, CPU_UP_PREPARE, (void *)(long)cpu);
>  +		rcutree_prepare_cpu(cpu);
> + 		rcu_cpu_starting(cpu);
> + 	}
>   }
>   
>   #include "tree_exp.h"
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ