lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <5a4acb9d-48ba-7f41-51c7-19f300a47f22@huawei.com>
Date:   Fri, 28 Oct 2022 11:16:55 +0800
From:   "Leizhen (ThunderTown)" <thunder.leizhen@...wei.com>
To:     <paulmck@...nel.org>
CC:     Frederic Weisbecker <frederic@...nel.org>,
        Neeraj Upadhyay <quic_neeraju@...cinc.com>,
        Josh Triplett <josh@...htriplett.org>,
        "Steven Rostedt" <rostedt@...dmis.org>,
        Mathieu Desnoyers <mathieu.desnoyers@...icios.com>,
        Lai Jiangshan <jiangshanlai@...il.com>,
        Joel Fernandes <joel@...lfernandes.org>, <rcu@...r.kernel.org>,
        <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v2 3/3] rcu: Add RCU stall diagnosis information



On 2022/10/28 1:33, Paul E. McKenney wrote:
> On Sat, Oct 22, 2022 at 08:45:25PM +0800, Zhen Lei wrote:
>> In some extreme cases, such as the I/O pressure test, the CPU usage may
>> be 100%, causing RCU stall. In this case, the printed information about
>> current is not useful. Displays the number and usage of hard interrupts,
>> soft interrupts, and context switches that are generated within half of
>> the CPU stall timeout, can help us make a general judgment. In other
>> cases, we can preliminarily determine whether an infinite loop occurs
>> when local_irq, local_bh or preempt is disabled.
>>
>> For example:
>> rcu: INFO: rcu_preempt self-detected stall on CPU
>> rcu:     0-....: (1250 ticks this GP) <omitted>
>> rcu:          hardirqs   softirqs   csw/system
>> rcu:  number:      624         45            0
>> rcu: cputime:       69          1         2425   ==> 2500(ms)
>>
>> The example above shows that the number of hard and soft interrupts is
>> small, there is zero context switching, and the system takes up a lot of
>> time. We can quickly conclude that the current task is infinitely looped
>> with preempt_disable().
>>
>> The impact on system performance is negligible because snapshot is
>> recorded only one time after 1/2 CPU stall timeout.
>>
>> Signed-off-by: Zhen Lei <thunder.leizhen@...wei.com>
>> ---
>>  kernel/rcu/tree.c       | 16 ++++++++++++++++
>>  kernel/rcu/tree.h       | 11 +++++++++++
>>  kernel/rcu/tree_stall.h | 28 ++++++++++++++++++++++++++++
>>  3 files changed, 55 insertions(+)
>>
>> diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
>> index 6bb8e72bc8151ef..56c49a3117e7a81 100644
>> --- a/kernel/rcu/tree.c
>> +++ b/kernel/rcu/tree.c
>> @@ -931,6 +931,22 @@ static int rcu_implicit_dynticks_qs(struct rcu_data *rdp)
>>  			rdp->rcu_iw_gp_seq = rnp->gp_seq;
>>  			irq_work_queue_on(&rdp->rcu_iw, rdp->cpu);
>>  		}
>> +
>> +		if (rdp->snap_record.gp_seq != rdp->gp_seq) {
>> +			u64 *cpustat;
>> +			struct rcu_snap_record *r;
>> +
>> +			cpustat = kcpustat_cpu(rdp->cpu).cpustat;
>> +
>> +			r = &rdp->snap_record;
>> +			r->cputime_irq     = cpustat[CPUTIME_IRQ];
>> +			r->cputime_softirq = cpustat[CPUTIME_SOFTIRQ];
>> +			r->cputime_system  = cpustat[CPUTIME_SYSTEM];
>> +			r->nr_hardirqs = kstat_cpu_irqs_sum(rdp->cpu);
>> +			r->nr_softirqs = kstat_cpu_softirqs_sum(rdp->cpu);
>> +			r->nr_csw = nr_context_switches_cpu(rdp->cpu);
>> +			r->gp_seq = rdp->gp_seq;
> 
> This needs to be optional.  Yes, it is normally rarely executed, but
> people who don't want the additional information should not pay the
> price for it.
> 
>> +		}
>>  	}
>>  
>>  	return 0;
>> diff --git a/kernel/rcu/tree.h b/kernel/rcu/tree.h
>> index d4a97e40ea9c3e2..fb3121d15cca6f8 100644
>> --- a/kernel/rcu/tree.h
>> +++ b/kernel/rcu/tree.h
>> @@ -158,6 +158,16 @@ union rcu_noqs {
>>  	u16 s; /* Set of bits, aggregate OR here. */
>>  };
>>  
>> +struct rcu_snap_record {
>> +	unsigned long	gp_seq;
>> +	u64		cputime_irq;
>> +	u64		cputime_softirq;
>> +	u64		cputime_system;
>> +	unsigned int	nr_hardirqs;
>> +	unsigned int	nr_softirqs;
>> +	unsigned long long nr_csw;
>> +};
> 
> Please add a comment saying what this is and what it is used for.

OK, I will do it.

> 
>> +
>>  /* Per-CPU data for read-copy update. */
>>  struct rcu_data {
>>  	/* 1) quiescent-state and grace-period handling : */
>> @@ -262,6 +272,7 @@ struct rcu_data {
>>  	short rcu_onl_gp_flags;		/* ->gp_flags at last online. */
>>  	unsigned long last_fqs_resched;	/* Time of last rcu_resched(). */
>>  	unsigned long last_sched_clock;	/* Jiffies of last rcu_sched_clock_irq(). */
>> +	struct rcu_snap_record snap_record;
> 
> And here as well, please.

OK

> 
>>  	int cpu;
>>  };
>> diff --git a/kernel/rcu/tree_stall.h b/kernel/rcu/tree_stall.h
>> index 5653560573e22d6..f8c9d0284d116a8 100644
>> --- a/kernel/rcu/tree_stall.h
>> +++ b/kernel/rcu/tree_stall.h
>> @@ -428,6 +428,32 @@ static bool rcu_is_rcuc_kthread_starving(struct rcu_data *rdp, unsigned long *jp
>>  	return j > 2 * HZ;
>>  }
>>  
>> +static void print_cpu_stat_info(int cpu)
>> +{
>> +	u64 *cpustat;
>> +	unsigned long half_timeout;
>> +	struct rcu_snap_record *r;
> 
> Let's please follow convention and call it "rsrp" rather than just "r".

OK

> 
>> +	struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
>> +
>> +	r = &rdp->snap_record;
>> +	if (r->gp_seq != rdp->gp_seq)
>> +		return;
>> +
>> +	cpustat = kcpustat_cpu(cpu).cpustat;
>> +	half_timeout = rcu_jiffies_till_stall_check() / 2;
>> +
>> +	pr_err("         hardirqs   softirqs   csw/system\n");
>> +	pr_err(" number: %8d %10d %12lld\n",
>> +		kstat_cpu_irqs_sum(cpu) - r->nr_hardirqs,
>> +		kstat_cpu_softirqs_sum(cpu) - r->nr_softirqs,
>> +		nr_context_switches_cpu(cpu) - r->nr_csw);
>> +	pr_err("cputime: %8lld %10lld %12lld   ==> %lld(ms)\n",
>> +		div_u64(cpustat[CPUTIME_IRQ] - r->cputime_irq, NSEC_PER_MSEC),
>> +		div_u64(cpustat[CPUTIME_SOFTIRQ] - r->cputime_softirq, NSEC_PER_MSEC),
>> +		div_u64(cpustat[CPUTIME_SYSTEM] - r->cputime_system, NSEC_PER_MSEC),
>> +		jiffies64_to_msecs(half_timeout));
>> +}
>> +
>>  /*
>>   * Print out diagnostic information for the specified stalled CPU.
>>   *
>> @@ -484,6 +510,8 @@ static void print_cpu_stall_info(int cpu)
>>  	       data_race(rcu_state.n_force_qs) - rcu_state.n_force_qs_gpstart,
>>  	       rcuc_starved ? buf : "",
>>  	       falsepositive ? " (false positive?)" : "");
>> +
>> +	print_cpu_stat_info(cpu);
> 
> Again, please make this conditional.  One way to do that is with a
> Kconfig option.  Another is with a kernel boot parameter, as is done
> wtih module_param() elsewhere in tree_stall.h.  Or if the parsing needs
> to be fancy (it shouldn't) using kernel_param_ops as is done in tree.c.
> Distros tend to like kernel boot parameters, while people dealing with
> large numbers of devices tend to like Kconfig options.  Choose wisely.  ;-)

OK. I will add both Kconfig option and boot parameter, let the user
choose freely.

> 
> Please make this initially default-off.  If enough people enable it for
> long enough, we can later switch it to default-on and maybe even later
> to unconditional.  But let's start carefully.

Right. You're so far-sighted.

> 
> 							Thanx, Paul
> 
>>  }
>>  
>>  /* Complain about starvation of grace-period kthread.  */
>> -- 
>> 2.25.1
>>
> .
> 

-- 
Regards,
  Zhen Lei

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ