lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:   Wed, 15 Sep 2021 08:39:38 -0700
From:   "Paul E. McKenney" <paulmck@...nel.org>
To:     Changbin Du <changbin.du@...il.com>
Cc:     Peter Zijlstra <peterz@...radead.org>,
        Ingo Molnar <mingo@...hat.com>, Will Deacon <will@...nel.org>,
        Josh Triplett <josh@...htriplett.org>,
        Thomas Gleixner <tglx@...utronix.de>,
        Petr Mladek <pmladek@...e.com>, linux-kernel@...r.kernel.org,
        rcu@...r.kernel.org
Subject: Re: [PATCH] kernel: in_irq() cleanup

On Wed, Sep 15, 2021 at 08:32:32AM +0800, Changbin Du wrote:
> Hi, could anyone pick up this cleanup?

If you send the RCU pieces separately, I will take them.

						Thanx, Paul

> On Fri, Aug 13, 2021 at 10:51:01PM +0800, Changbin Du wrote:
> > Replace the obsolete and ambiguos macro in_irq() with new
> > macro in_hardirq().
> > 
> > Signed-off-by: Changbin Du <changbin.du@...il.com>
> > ---
> >  include/linux/lockdep.h              |  2 +-
> >  include/linux/rcutiny.h              |  2 +-
> >  kernel/irq/irqdesc.c                 |  2 +-
> >  kernel/rcu/tree.c                    |  2 +-
> >  kernel/rcu/tree_plugin.h             |  2 +-
> >  kernel/softirq.c                     | 10 +++++-----
> >  kernel/time/timer.c                  |  4 ++--
> >  kernel/trace/trace.h                 |  2 +-
> >  kernel/trace/trace_functions_graph.c |  2 +-
> >  lib/locking-selftest.c               |  4 ++--
> >  lib/vsprintf.c                       |  2 +-
> >  11 files changed, 17 insertions(+), 17 deletions(-)
> > 
> > diff --git a/include/linux/lockdep.h b/include/linux/lockdep.h
> > index 5cf387813754..e7aa0050bfd8 100644
> > --- a/include/linux/lockdep.h
> > +++ b/include/linux/lockdep.h
> > @@ -633,7 +633,7 @@ do {									\
> >  #define lockdep_assert_in_softirq()					\
> >  do {									\
> >  	WARN_ON_ONCE(__lockdep_enabled			&&		\
> > -		     (!in_softirq() || in_irq() || in_nmi()));		\
> > +		     (!in_softirq() || in_hardirq() || in_nmi()));		\
> >  } while (0)
> >  
> >  #else
> > diff --git a/include/linux/rcutiny.h b/include/linux/rcutiny.h
> > index 953e70fafe38..7fedbd33d5d2 100644
> > --- a/include/linux/rcutiny.h
> > +++ b/include/linux/rcutiny.h
> > @@ -88,7 +88,7 @@ static inline void rcu_irq_enter_irqson(void) { }
> >  static inline void rcu_irq_exit(void) { }
> >  static inline void rcu_irq_exit_check_preempt(void) { }
> >  #define rcu_is_idle_cpu(cpu) \
> > -	(is_idle_task(current) && !in_nmi() && !in_irq() && !in_serving_softirq())
> > +	(is_idle_task(current) && !in_nmi() && !in_hardirq() && !in_serving_softirq())
> >  static inline void exit_rcu(void) { }
> >  static inline bool rcu_preempt_need_deferred_qs(struct task_struct *t)
> >  {
> > diff --git a/kernel/irq/irqdesc.c b/kernel/irq/irqdesc.c
> > index fadb93766020..0ffd6c19259f 100644
> > --- a/kernel/irq/irqdesc.c
> > +++ b/kernel/irq/irqdesc.c
> > @@ -640,7 +640,7 @@ int handle_irq_desc(struct irq_desc *desc)
> >  		return -EINVAL;
> >  
> >  	data = irq_desc_get_irq_data(desc);
> > -	if (WARN_ON_ONCE(!in_irq() && handle_enforce_irqctx(data)))
> > +	if (WARN_ON_ONCE(!in_hardirq() && handle_enforce_irqctx(data)))
> >  		return -EPERM;
> >  
> >  	generic_handle_irq_desc(desc);
> > diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
> > index 51f24ecd94b2..18f6658f0713 100644
> > --- a/kernel/rcu/tree.c
> > +++ b/kernel/rcu/tree.c
> > @@ -1510,7 +1510,7 @@ static void rcu_gp_kthread_wake(void)
> >  {
> >  	struct task_struct *t = READ_ONCE(rcu_state.gp_kthread);
> >  
> > -	if ((current == t && !in_irq() && !in_serving_softirq()) ||
> > +	if ((current == t && !in_hardirq() && !in_serving_softirq()) ||
> >  	    !READ_ONCE(rcu_state.gp_flags) || !t)
> >  		return;
> >  	WRITE_ONCE(rcu_state.gp_wake_time, jiffies);
> > diff --git a/kernel/rcu/tree_plugin.h b/kernel/rcu/tree_plugin.h
> > index de1dc3bb7f70..953171af6d0c 100644
> > --- a/kernel/rcu/tree_plugin.h
> > +++ b/kernel/rcu/tree_plugin.h
> > @@ -672,7 +672,7 @@ static void rcu_read_unlock_special(struct task_struct *t)
> >  			   (IS_ENABLED(CONFIG_RCU_BOOST) && irqs_were_disabled &&
> >  			    t->rcu_blocked_node);
> >  		// Need to defer quiescent state until everything is enabled.
> > -		if (use_softirq && (in_irq() || (expboost && !irqs_were_disabled))) {
> > +		if (use_softirq && (in_hardirq() || (expboost && !irqs_were_disabled))) {
> >  			// Using softirq, safe to awaken, and either the
> >  			// wakeup is free or there is either an expedited
> >  			// GP in flight or a potential need to deboost.
> > diff --git a/kernel/softirq.c b/kernel/softirq.c
> > index f3a012179f47..e8c62c4b1e5d 100644
> > --- a/kernel/softirq.c
> > +++ b/kernel/softirq.c
> > @@ -222,7 +222,7 @@ void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
> >  	u32 pending;
> >  	int curcnt;
> >  
> > -	WARN_ON_ONCE(in_irq());
> > +	WARN_ON_ONCE(in_hardirq());
> >  	lockdep_assert_irqs_enabled();
> >  
> >  	local_irq_save(flags);
> > @@ -305,7 +305,7 @@ void __local_bh_disable_ip(unsigned long ip, unsigned int cnt)
> >  {
> >  	unsigned long flags;
> >  
> > -	WARN_ON_ONCE(in_irq());
> > +	WARN_ON_ONCE(in_hardirq());
> >  
> >  	raw_local_irq_save(flags);
> >  	/*
> > @@ -352,14 +352,14 @@ static void __local_bh_enable(unsigned int cnt)
> >   */
> >  void _local_bh_enable(void)
> >  {
> > -	WARN_ON_ONCE(in_irq());
> > +	WARN_ON_ONCE(in_hardirq());
> >  	__local_bh_enable(SOFTIRQ_DISABLE_OFFSET);
> >  }
> >  EXPORT_SYMBOL(_local_bh_enable);
> >  
> >  void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
> >  {
> > -	WARN_ON_ONCE(in_irq());
> > +	WARN_ON_ONCE(in_hardirq());
> >  	lockdep_assert_irqs_enabled();
> >  #ifdef CONFIG_TRACE_IRQFLAGS
> >  	local_irq_disable();
> > @@ -617,7 +617,7 @@ static inline void tick_irq_exit(void)
> >  
> >  	/* Make sure that timer wheel updates are propagated */
> >  	if ((idle_cpu(cpu) && !need_resched()) || tick_nohz_full_cpu(cpu)) {
> > -		if (!in_irq())
> > +		if (!in_hardirq())
> >  			tick_nohz_irq_exit();
> >  	}
> >  #endif
> > diff --git a/kernel/time/timer.c b/kernel/time/timer.c
> > index e3d2c23c413d..dfdbb114a43d 100644
> > --- a/kernel/time/timer.c
> > +++ b/kernel/time/timer.c
> > @@ -1369,7 +1369,7 @@ int del_timer_sync(struct timer_list *timer)
> >  	 * don't use it in hardirq context, because it
> >  	 * could lead to deadlock.
> >  	 */
> > -	WARN_ON(in_irq() && !(timer->flags & TIMER_IRQSAFE));
> > +	WARN_ON(in_hardirq() && !(timer->flags & TIMER_IRQSAFE));
> >  
> >  	/*
> >  	 * Must be able to sleep on PREEMPT_RT because of the slowpath in
> > @@ -1784,7 +1784,7 @@ void update_process_times(int user_tick)
> >  	run_local_timers();
> >  	rcu_sched_clock_irq(user_tick);
> >  #ifdef CONFIG_IRQ_WORK
> > -	if (in_irq())
> > +	if (in_hardirq())
> >  		irq_work_tick();
> >  #endif
> >  	scheduler_tick();
> > diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> > index a180abf76d4e..2ed5b77b60ca 100644
> > --- a/kernel/trace/trace.h
> > +++ b/kernel/trace/trace.h
> > @@ -876,7 +876,7 @@ static inline int ftrace_graph_addr(struct ftrace_graph_ent *trace)
> >  		 * is set, and called by an interrupt handler, we still
> >  		 * want to trace it.
> >  		 */
> > -		if (in_irq())
> > +		if (in_hardirq())
> >  			trace_recursion_set(TRACE_IRQ_BIT);
> >  		else
> >  			trace_recursion_clear(TRACE_IRQ_BIT);
> > diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c
> > index 0de6837722da..b08d3923de98 100644
> > --- a/kernel/trace/trace_functions_graph.c
> > +++ b/kernel/trace/trace_functions_graph.c
> > @@ -120,7 +120,7 @@ static inline int ftrace_graph_ignore_irqs(void)
> >  	if (!ftrace_graph_skip_irqs || trace_recursion_test(TRACE_IRQ_BIT))
> >  		return 0;
> >  
> > -	return in_irq();
> > +	return in_hardirq();
> >  }
> >  
> >  int trace_graph_entry(struct ftrace_graph_ent *trace)
> > diff --git a/lib/locking-selftest.c b/lib/locking-selftest.c
> > index 161108e5d2fe..a4edff9ffc33 100644
> > --- a/lib/locking-selftest.c
> > +++ b/lib/locking-selftest.c
> > @@ -196,7 +196,7 @@ static void init_shared_classes(void)
> >  	local_irq_disable();			\
> >  	__irq_enter();				\
> >  	lockdep_hardirq_threaded();		\
> > -	WARN_ON(!in_irq());
> > +	WARN_ON(!in_hardirq());
> >  
> >  #define HARDIRQ_EXIT()				\
> >  	__irq_exit();				\
> > @@ -2456,7 +2456,7 @@ static void hardirq_exit(int *_)
> >  	int notthreaded_hardirq_guard_##name __guard(hardirq_exit);	\
> >  	local_irq_disable();						\
> >  	__irq_enter();							\
> > -	WARN_ON(!in_irq());
> > +	WARN_ON(!in_hardirq());
> >  
> >  static void softirq_exit(int *_)
> >  {
> > diff --git a/lib/vsprintf.c b/lib/vsprintf.c
> > index 26c83943748a..2c5b4351330c 100644
> > --- a/lib/vsprintf.c
> > +++ b/lib/vsprintf.c
> > @@ -865,7 +865,7 @@ char *restricted_pointer(char *buf, char *end, const void *ptr,
> >  		 * kptr_restrict==1 cannot be used in IRQ context
> >  		 * because its test for CAP_SYSLOG would be meaningless.
> >  		 */
> > -		if (in_irq() || in_serving_softirq() || in_nmi()) {
> > +		if (in_hardirq() || in_serving_softirq() || in_nmi()) {
> >  			if (spec.field_width == -1)
> >  				spec.field_width = 2 * sizeof(ptr);
> >  			return error_string(buf, end, "pK-error", spec);
> > -- 
> > 2.30.2
> > 
> 
> -- 
> Cheers,
> Changbin Du

Powered by blists - more mailing lists