LKML Archive on lore.kernel.org
help / color / mirror / Atom feed
* [PATCH] kernel: in_irq() cleanup
@ 2021-08-13 14:51 Changbin Du
  2021-08-14  2:26 ` Paul E. McKenney
                   ` (2 more replies)
  0 siblings, 3 replies; 6+ messages in thread
From: Changbin Du @ 2021-08-13 14:51 UTC (permalink / raw)
  To: Peter Zijlstra, Ingo Molnar
  Cc: Will Deacon, Paul E. McKenney, Josh Triplett, Thomas Gleixner,
	Petr Mladek, linux-kernel, rcu, Changbin Du

Replace the obsolete and ambiguos macro in_irq() with new
macro in_hardirq().

Signed-off-by: Changbin Du <changbin.du@gmail.com>
---
 include/linux/lockdep.h              |  2 +-
 include/linux/rcutiny.h              |  2 +-
 kernel/irq/irqdesc.c                 |  2 +-
 kernel/rcu/tree.c                    |  2 +-
 kernel/rcu/tree_plugin.h             |  2 +-
 kernel/softirq.c                     | 10 +++++-----
 kernel/time/timer.c                  |  4 ++--
 kernel/trace/trace.h                 |  2 +-
 kernel/trace/trace_functions_graph.c |  2 +-
 lib/locking-selftest.c               |  4 ++--
 lib/vsprintf.c                       |  2 +-
 11 files changed, 17 insertions(+), 17 deletions(-)

diff --git a/include/linux/lockdep.h b/include/linux/lockdep.h
index 5cf387813754..e7aa0050bfd8 100644
--- a/include/linux/lockdep.h
+++ b/include/linux/lockdep.h
@@ -633,7 +633,7 @@ do {									\
 #define lockdep_assert_in_softirq()					\
 do {									\
 	WARN_ON_ONCE(__lockdep_enabled			&&		\
-		     (!in_softirq() || in_irq() || in_nmi()));		\
+		     (!in_softirq() || in_hardirq() || in_nmi()));		\
 } while (0)
 
 #else
diff --git a/include/linux/rcutiny.h b/include/linux/rcutiny.h
index 953e70fafe38..7fedbd33d5d2 100644
--- a/include/linux/rcutiny.h
+++ b/include/linux/rcutiny.h
@@ -88,7 +88,7 @@ static inline void rcu_irq_enter_irqson(void) { }
 static inline void rcu_irq_exit(void) { }
 static inline void rcu_irq_exit_check_preempt(void) { }
 #define rcu_is_idle_cpu(cpu) \
-	(is_idle_task(current) && !in_nmi() && !in_irq() && !in_serving_softirq())
+	(is_idle_task(current) && !in_nmi() && !in_hardirq() && !in_serving_softirq())
 static inline void exit_rcu(void) { }
 static inline bool rcu_preempt_need_deferred_qs(struct task_struct *t)
 {
diff --git a/kernel/irq/irqdesc.c b/kernel/irq/irqdesc.c
index fadb93766020..0ffd6c19259f 100644
--- a/kernel/irq/irqdesc.c
+++ b/kernel/irq/irqdesc.c
@@ -640,7 +640,7 @@ int handle_irq_desc(struct irq_desc *desc)
 		return -EINVAL;
 
 	data = irq_desc_get_irq_data(desc);
-	if (WARN_ON_ONCE(!in_irq() && handle_enforce_irqctx(data)))
+	if (WARN_ON_ONCE(!in_hardirq() && handle_enforce_irqctx(data)))
 		return -EPERM;
 
 	generic_handle_irq_desc(desc);
diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
index 51f24ecd94b2..18f6658f0713 100644
--- a/kernel/rcu/tree.c
+++ b/kernel/rcu/tree.c
@@ -1510,7 +1510,7 @@ static void rcu_gp_kthread_wake(void)
 {
 	struct task_struct *t = READ_ONCE(rcu_state.gp_kthread);
 
-	if ((current == t && !in_irq() && !in_serving_softirq()) ||
+	if ((current == t && !in_hardirq() && !in_serving_softirq()) ||
 	    !READ_ONCE(rcu_state.gp_flags) || !t)
 		return;
 	WRITE_ONCE(rcu_state.gp_wake_time, jiffies);
diff --git a/kernel/rcu/tree_plugin.h b/kernel/rcu/tree_plugin.h
index de1dc3bb7f70..953171af6d0c 100644
--- a/kernel/rcu/tree_plugin.h
+++ b/kernel/rcu/tree_plugin.h
@@ -672,7 +672,7 @@ static void rcu_read_unlock_special(struct task_struct *t)
 			   (IS_ENABLED(CONFIG_RCU_BOOST) && irqs_were_disabled &&
 			    t->rcu_blocked_node);
 		// Need to defer quiescent state until everything is enabled.
-		if (use_softirq && (in_irq() || (expboost && !irqs_were_disabled))) {
+		if (use_softirq && (in_hardirq() || (expboost && !irqs_were_disabled))) {
 			// Using softirq, safe to awaken, and either the
 			// wakeup is free or there is either an expedited
 			// GP in flight or a potential need to deboost.
diff --git a/kernel/softirq.c b/kernel/softirq.c
index f3a012179f47..e8c62c4b1e5d 100644
--- a/kernel/softirq.c
+++ b/kernel/softirq.c
@@ -222,7 +222,7 @@ void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
 	u32 pending;
 	int curcnt;
 
-	WARN_ON_ONCE(in_irq());
+	WARN_ON_ONCE(in_hardirq());
 	lockdep_assert_irqs_enabled();
 
 	local_irq_save(flags);
@@ -305,7 +305,7 @@ void __local_bh_disable_ip(unsigned long ip, unsigned int cnt)
 {
 	unsigned long flags;
 
-	WARN_ON_ONCE(in_irq());
+	WARN_ON_ONCE(in_hardirq());
 
 	raw_local_irq_save(flags);
 	/*
@@ -352,14 +352,14 @@ static void __local_bh_enable(unsigned int cnt)
  */
 void _local_bh_enable(void)
 {
-	WARN_ON_ONCE(in_irq());
+	WARN_ON_ONCE(in_hardirq());
 	__local_bh_enable(SOFTIRQ_DISABLE_OFFSET);
 }
 EXPORT_SYMBOL(_local_bh_enable);
 
 void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
 {
-	WARN_ON_ONCE(in_irq());
+	WARN_ON_ONCE(in_hardirq());
 	lockdep_assert_irqs_enabled();
 #ifdef CONFIG_TRACE_IRQFLAGS
 	local_irq_disable();
@@ -617,7 +617,7 @@ static inline void tick_irq_exit(void)
 
 	/* Make sure that timer wheel updates are propagated */
 	if ((idle_cpu(cpu) && !need_resched()) || tick_nohz_full_cpu(cpu)) {
-		if (!in_irq())
+		if (!in_hardirq())
 			tick_nohz_irq_exit();
 	}
 #endif
diff --git a/kernel/time/timer.c b/kernel/time/timer.c
index e3d2c23c413d..dfdbb114a43d 100644
--- a/kernel/time/timer.c
+++ b/kernel/time/timer.c
@@ -1369,7 +1369,7 @@ int del_timer_sync(struct timer_list *timer)
 	 * don't use it in hardirq context, because it
 	 * could lead to deadlock.
 	 */
-	WARN_ON(in_irq() && !(timer->flags & TIMER_IRQSAFE));
+	WARN_ON(in_hardirq() && !(timer->flags & TIMER_IRQSAFE));
 
 	/*
 	 * Must be able to sleep on PREEMPT_RT because of the slowpath in
@@ -1784,7 +1784,7 @@ void update_process_times(int user_tick)
 	run_local_timers();
 	rcu_sched_clock_irq(user_tick);
 #ifdef CONFIG_IRQ_WORK
-	if (in_irq())
+	if (in_hardirq())
 		irq_work_tick();
 #endif
 	scheduler_tick();
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
index a180abf76d4e..2ed5b77b60ca 100644
--- a/kernel/trace/trace.h
+++ b/kernel/trace/trace.h
@@ -876,7 +876,7 @@ static inline int ftrace_graph_addr(struct ftrace_graph_ent *trace)
 		 * is set, and called by an interrupt handler, we still
 		 * want to trace it.
 		 */
-		if (in_irq())
+		if (in_hardirq())
 			trace_recursion_set(TRACE_IRQ_BIT);
 		else
 			trace_recursion_clear(TRACE_IRQ_BIT);
diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c
index 0de6837722da..b08d3923de98 100644
--- a/kernel/trace/trace_functions_graph.c
+++ b/kernel/trace/trace_functions_graph.c
@@ -120,7 +120,7 @@ static inline int ftrace_graph_ignore_irqs(void)
 	if (!ftrace_graph_skip_irqs || trace_recursion_test(TRACE_IRQ_BIT))
 		return 0;
 
-	return in_irq();
+	return in_hardirq();
 }
 
 int trace_graph_entry(struct ftrace_graph_ent *trace)
diff --git a/lib/locking-selftest.c b/lib/locking-selftest.c
index 161108e5d2fe..a4edff9ffc33 100644
--- a/lib/locking-selftest.c
+++ b/lib/locking-selftest.c
@@ -196,7 +196,7 @@ static void init_shared_classes(void)
 	local_irq_disable();			\
 	__irq_enter();				\
 	lockdep_hardirq_threaded();		\
-	WARN_ON(!in_irq());
+	WARN_ON(!in_hardirq());
 
 #define HARDIRQ_EXIT()				\
 	__irq_exit();				\
@@ -2456,7 +2456,7 @@ static void hardirq_exit(int *_)
 	int notthreaded_hardirq_guard_##name __guard(hardirq_exit);	\
 	local_irq_disable();						\
 	__irq_enter();							\
-	WARN_ON(!in_irq());
+	WARN_ON(!in_hardirq());
 
 static void softirq_exit(int *_)
 {
diff --git a/lib/vsprintf.c b/lib/vsprintf.c
index 26c83943748a..2c5b4351330c 100644
--- a/lib/vsprintf.c
+++ b/lib/vsprintf.c
@@ -865,7 +865,7 @@ char *restricted_pointer(char *buf, char *end, const void *ptr,
 		 * kptr_restrict==1 cannot be used in IRQ context
 		 * because its test for CAP_SYSLOG would be meaningless.
 		 */
-		if (in_irq() || in_serving_softirq() || in_nmi()) {
+		if (in_hardirq() || in_serving_softirq() || in_nmi()) {
 			if (spec.field_width == -1)
 				spec.field_width = 2 * sizeof(ptr);
 			return error_string(buf, end, "pK-error", spec);
-- 
2.30.2


^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH] kernel: in_irq() cleanup
  2021-08-13 14:51 [PATCH] kernel: in_irq() cleanup Changbin Du
@ 2021-08-14  2:26 ` Paul E. McKenney
  2021-08-26 14:58 ` Petr Mladek
  2021-09-15  0:32 ` Changbin Du
  2 siblings, 0 replies; 6+ messages in thread
From: Paul E. McKenney @ 2021-08-14  2:26 UTC (permalink / raw)
  To: Changbin Du
  Cc: Peter Zijlstra, Ingo Molnar, Will Deacon, Josh Triplett,
	Thomas Gleixner, Petr Mladek, linux-kernel, rcu

On Fri, Aug 13, 2021 at 10:51:01PM +0800, Changbin Du wrote:
> Replace the obsolete and ambiguos macro in_irq() with new
> macro in_hardirq().
> 
> Signed-off-by: Changbin Du <changbin.du@gmail.com>

From an RCU viewpoint:

Reviewed-by: Paul E. McKenney <paulmck@kernel.org>

> ---
>  include/linux/lockdep.h              |  2 +-
>  include/linux/rcutiny.h              |  2 +-
>  kernel/irq/irqdesc.c                 |  2 +-
>  kernel/rcu/tree.c                    |  2 +-
>  kernel/rcu/tree_plugin.h             |  2 +-
>  kernel/softirq.c                     | 10 +++++-----
>  kernel/time/timer.c                  |  4 ++--
>  kernel/trace/trace.h                 |  2 +-
>  kernel/trace/trace_functions_graph.c |  2 +-
>  lib/locking-selftest.c               |  4 ++--
>  lib/vsprintf.c                       |  2 +-
>  11 files changed, 17 insertions(+), 17 deletions(-)
> 
> diff --git a/include/linux/lockdep.h b/include/linux/lockdep.h
> index 5cf387813754..e7aa0050bfd8 100644
> --- a/include/linux/lockdep.h
> +++ b/include/linux/lockdep.h
> @@ -633,7 +633,7 @@ do {									\
>  #define lockdep_assert_in_softirq()					\
>  do {									\
>  	WARN_ON_ONCE(__lockdep_enabled			&&		\
> -		     (!in_softirq() || in_irq() || in_nmi()));		\
> +		     (!in_softirq() || in_hardirq() || in_nmi()));		\
>  } while (0)
>  
>  #else
> diff --git a/include/linux/rcutiny.h b/include/linux/rcutiny.h
> index 953e70fafe38..7fedbd33d5d2 100644
> --- a/include/linux/rcutiny.h
> +++ b/include/linux/rcutiny.h
> @@ -88,7 +88,7 @@ static inline void rcu_irq_enter_irqson(void) { }
>  static inline void rcu_irq_exit(void) { }
>  static inline void rcu_irq_exit_check_preempt(void) { }
>  #define rcu_is_idle_cpu(cpu) \
> -	(is_idle_task(current) && !in_nmi() && !in_irq() && !in_serving_softirq())
> +	(is_idle_task(current) && !in_nmi() && !in_hardirq() && !in_serving_softirq())
>  static inline void exit_rcu(void) { }
>  static inline bool rcu_preempt_need_deferred_qs(struct task_struct *t)
>  {
> diff --git a/kernel/irq/irqdesc.c b/kernel/irq/irqdesc.c
> index fadb93766020..0ffd6c19259f 100644
> --- a/kernel/irq/irqdesc.c
> +++ b/kernel/irq/irqdesc.c
> @@ -640,7 +640,7 @@ int handle_irq_desc(struct irq_desc *desc)
>  		return -EINVAL;
>  
>  	data = irq_desc_get_irq_data(desc);
> -	if (WARN_ON_ONCE(!in_irq() && handle_enforce_irqctx(data)))
> +	if (WARN_ON_ONCE(!in_hardirq() && handle_enforce_irqctx(data)))
>  		return -EPERM;
>  
>  	generic_handle_irq_desc(desc);
> diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
> index 51f24ecd94b2..18f6658f0713 100644
> --- a/kernel/rcu/tree.c
> +++ b/kernel/rcu/tree.c
> @@ -1510,7 +1510,7 @@ static void rcu_gp_kthread_wake(void)
>  {
>  	struct task_struct *t = READ_ONCE(rcu_state.gp_kthread);
>  
> -	if ((current == t && !in_irq() && !in_serving_softirq()) ||
> +	if ((current == t && !in_hardirq() && !in_serving_softirq()) ||
>  	    !READ_ONCE(rcu_state.gp_flags) || !t)
>  		return;
>  	WRITE_ONCE(rcu_state.gp_wake_time, jiffies);
> diff --git a/kernel/rcu/tree_plugin.h b/kernel/rcu/tree_plugin.h
> index de1dc3bb7f70..953171af6d0c 100644
> --- a/kernel/rcu/tree_plugin.h
> +++ b/kernel/rcu/tree_plugin.h
> @@ -672,7 +672,7 @@ static void rcu_read_unlock_special(struct task_struct *t)
>  			   (IS_ENABLED(CONFIG_RCU_BOOST) && irqs_were_disabled &&
>  			    t->rcu_blocked_node);
>  		// Need to defer quiescent state until everything is enabled.
> -		if (use_softirq && (in_irq() || (expboost && !irqs_were_disabled))) {
> +		if (use_softirq && (in_hardirq() || (expboost && !irqs_were_disabled))) {
>  			// Using softirq, safe to awaken, and either the
>  			// wakeup is free or there is either an expedited
>  			// GP in flight or a potential need to deboost.
> diff --git a/kernel/softirq.c b/kernel/softirq.c
> index f3a012179f47..e8c62c4b1e5d 100644
> --- a/kernel/softirq.c
> +++ b/kernel/softirq.c
> @@ -222,7 +222,7 @@ void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
>  	u32 pending;
>  	int curcnt;
>  
> -	WARN_ON_ONCE(in_irq());
> +	WARN_ON_ONCE(in_hardirq());
>  	lockdep_assert_irqs_enabled();
>  
>  	local_irq_save(flags);
> @@ -305,7 +305,7 @@ void __local_bh_disable_ip(unsigned long ip, unsigned int cnt)
>  {
>  	unsigned long flags;
>  
> -	WARN_ON_ONCE(in_irq());
> +	WARN_ON_ONCE(in_hardirq());
>  
>  	raw_local_irq_save(flags);
>  	/*
> @@ -352,14 +352,14 @@ static void __local_bh_enable(unsigned int cnt)
>   */
>  void _local_bh_enable(void)
>  {
> -	WARN_ON_ONCE(in_irq());
> +	WARN_ON_ONCE(in_hardirq());
>  	__local_bh_enable(SOFTIRQ_DISABLE_OFFSET);
>  }
>  EXPORT_SYMBOL(_local_bh_enable);
>  
>  void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
>  {
> -	WARN_ON_ONCE(in_irq());
> +	WARN_ON_ONCE(in_hardirq());
>  	lockdep_assert_irqs_enabled();
>  #ifdef CONFIG_TRACE_IRQFLAGS
>  	local_irq_disable();
> @@ -617,7 +617,7 @@ static inline void tick_irq_exit(void)
>  
>  	/* Make sure that timer wheel updates are propagated */
>  	if ((idle_cpu(cpu) && !need_resched()) || tick_nohz_full_cpu(cpu)) {
> -		if (!in_irq())
> +		if (!in_hardirq())
>  			tick_nohz_irq_exit();
>  	}
>  #endif
> diff --git a/kernel/time/timer.c b/kernel/time/timer.c
> index e3d2c23c413d..dfdbb114a43d 100644
> --- a/kernel/time/timer.c
> +++ b/kernel/time/timer.c
> @@ -1369,7 +1369,7 @@ int del_timer_sync(struct timer_list *timer)
>  	 * don't use it in hardirq context, because it
>  	 * could lead to deadlock.
>  	 */
> -	WARN_ON(in_irq() && !(timer->flags & TIMER_IRQSAFE));
> +	WARN_ON(in_hardirq() && !(timer->flags & TIMER_IRQSAFE));
>  
>  	/*
>  	 * Must be able to sleep on PREEMPT_RT because of the slowpath in
> @@ -1784,7 +1784,7 @@ void update_process_times(int user_tick)
>  	run_local_timers();
>  	rcu_sched_clock_irq(user_tick);
>  #ifdef CONFIG_IRQ_WORK
> -	if (in_irq())
> +	if (in_hardirq())
>  		irq_work_tick();
>  #endif
>  	scheduler_tick();
> diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> index a180abf76d4e..2ed5b77b60ca 100644
> --- a/kernel/trace/trace.h
> +++ b/kernel/trace/trace.h
> @@ -876,7 +876,7 @@ static inline int ftrace_graph_addr(struct ftrace_graph_ent *trace)
>  		 * is set, and called by an interrupt handler, we still
>  		 * want to trace it.
>  		 */
> -		if (in_irq())
> +		if (in_hardirq())
>  			trace_recursion_set(TRACE_IRQ_BIT);
>  		else
>  			trace_recursion_clear(TRACE_IRQ_BIT);
> diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c
> index 0de6837722da..b08d3923de98 100644
> --- a/kernel/trace/trace_functions_graph.c
> +++ b/kernel/trace/trace_functions_graph.c
> @@ -120,7 +120,7 @@ static inline int ftrace_graph_ignore_irqs(void)
>  	if (!ftrace_graph_skip_irqs || trace_recursion_test(TRACE_IRQ_BIT))
>  		return 0;
>  
> -	return in_irq();
> +	return in_hardirq();
>  }
>  
>  int trace_graph_entry(struct ftrace_graph_ent *trace)
> diff --git a/lib/locking-selftest.c b/lib/locking-selftest.c
> index 161108e5d2fe..a4edff9ffc33 100644
> --- a/lib/locking-selftest.c
> +++ b/lib/locking-selftest.c
> @@ -196,7 +196,7 @@ static void init_shared_classes(void)
>  	local_irq_disable();			\
>  	__irq_enter();				\
>  	lockdep_hardirq_threaded();		\
> -	WARN_ON(!in_irq());
> +	WARN_ON(!in_hardirq());
>  
>  #define HARDIRQ_EXIT()				\
>  	__irq_exit();				\
> @@ -2456,7 +2456,7 @@ static void hardirq_exit(int *_)
>  	int notthreaded_hardirq_guard_##name __guard(hardirq_exit);	\
>  	local_irq_disable();						\
>  	__irq_enter();							\
> -	WARN_ON(!in_irq());
> +	WARN_ON(!in_hardirq());
>  
>  static void softirq_exit(int *_)
>  {
> diff --git a/lib/vsprintf.c b/lib/vsprintf.c
> index 26c83943748a..2c5b4351330c 100644
> --- a/lib/vsprintf.c
> +++ b/lib/vsprintf.c
> @@ -865,7 +865,7 @@ char *restricted_pointer(char *buf, char *end, const void *ptr,
>  		 * kptr_restrict==1 cannot be used in IRQ context
>  		 * because its test for CAP_SYSLOG would be meaningless.
>  		 */
> -		if (in_irq() || in_serving_softirq() || in_nmi()) {
> +		if (in_hardirq() || in_serving_softirq() || in_nmi()) {
>  			if (spec.field_width == -1)
>  				spec.field_width = 2 * sizeof(ptr);
>  			return error_string(buf, end, "pK-error", spec);
> -- 
> 2.30.2
> 

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH] kernel: in_irq() cleanup
  2021-08-13 14:51 [PATCH] kernel: in_irq() cleanup Changbin Du
  2021-08-14  2:26 ` Paul E. McKenney
@ 2021-08-26 14:58 ` Petr Mladek
  2021-09-15  0:32 ` Changbin Du
  2 siblings, 0 replies; 6+ messages in thread
From: Petr Mladek @ 2021-08-26 14:58 UTC (permalink / raw)
  To: Changbin Du
  Cc: Peter Zijlstra, Ingo Molnar, Will Deacon, Paul E. McKenney,
	Josh Triplett, Thomas Gleixner, linux-kernel, rcu

On Fri 2021-08-13 22:51:01, Changbin Du wrote:
> Replace the obsolete and ambiguos macro in_irq() with new
> macro in_hardirq().
> 
> Signed-off-by: Changbin Du <changbin.du@gmail.com>

Reviewed-by: Petr Mladek <pmladek@suse.com>   # for vsprintf.c

But the rest looks good as well.

Best Regards,
Petr

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH] kernel: in_irq() cleanup
  2021-08-13 14:51 [PATCH] kernel: in_irq() cleanup Changbin Du
  2021-08-14  2:26 ` Paul E. McKenney
  2021-08-26 14:58 ` Petr Mladek
@ 2021-09-15  0:32 ` Changbin Du
  2021-09-15 15:39   ` Paul E. McKenney
  2 siblings, 1 reply; 6+ messages in thread
From: Changbin Du @ 2021-09-15  0:32 UTC (permalink / raw)
  To: Changbin Du
  Cc: Peter Zijlstra, Ingo Molnar, Will Deacon, Paul E. McKenney,
	Josh Triplett, Thomas Gleixner, Petr Mladek, linux-kernel, rcu

Hi, could anyone pick up this cleanup?

On Fri, Aug 13, 2021 at 10:51:01PM +0800, Changbin Du wrote:
> Replace the obsolete and ambiguos macro in_irq() with new
> macro in_hardirq().
> 
> Signed-off-by: Changbin Du <changbin.du@gmail.com>
> ---
>  include/linux/lockdep.h              |  2 +-
>  include/linux/rcutiny.h              |  2 +-
>  kernel/irq/irqdesc.c                 |  2 +-
>  kernel/rcu/tree.c                    |  2 +-
>  kernel/rcu/tree_plugin.h             |  2 +-
>  kernel/softirq.c                     | 10 +++++-----
>  kernel/time/timer.c                  |  4 ++--
>  kernel/trace/trace.h                 |  2 +-
>  kernel/trace/trace_functions_graph.c |  2 +-
>  lib/locking-selftest.c               |  4 ++--
>  lib/vsprintf.c                       |  2 +-
>  11 files changed, 17 insertions(+), 17 deletions(-)
> 
> diff --git a/include/linux/lockdep.h b/include/linux/lockdep.h
> index 5cf387813754..e7aa0050bfd8 100644
> --- a/include/linux/lockdep.h
> +++ b/include/linux/lockdep.h
> @@ -633,7 +633,7 @@ do {									\
>  #define lockdep_assert_in_softirq()					\
>  do {									\
>  	WARN_ON_ONCE(__lockdep_enabled			&&		\
> -		     (!in_softirq() || in_irq() || in_nmi()));		\
> +		     (!in_softirq() || in_hardirq() || in_nmi()));		\
>  } while (0)
>  
>  #else
> diff --git a/include/linux/rcutiny.h b/include/linux/rcutiny.h
> index 953e70fafe38..7fedbd33d5d2 100644
> --- a/include/linux/rcutiny.h
> +++ b/include/linux/rcutiny.h
> @@ -88,7 +88,7 @@ static inline void rcu_irq_enter_irqson(void) { }
>  static inline void rcu_irq_exit(void) { }
>  static inline void rcu_irq_exit_check_preempt(void) { }
>  #define rcu_is_idle_cpu(cpu) \
> -	(is_idle_task(current) && !in_nmi() && !in_irq() && !in_serving_softirq())
> +	(is_idle_task(current) && !in_nmi() && !in_hardirq() && !in_serving_softirq())
>  static inline void exit_rcu(void) { }
>  static inline bool rcu_preempt_need_deferred_qs(struct task_struct *t)
>  {
> diff --git a/kernel/irq/irqdesc.c b/kernel/irq/irqdesc.c
> index fadb93766020..0ffd6c19259f 100644
> --- a/kernel/irq/irqdesc.c
> +++ b/kernel/irq/irqdesc.c
> @@ -640,7 +640,7 @@ int handle_irq_desc(struct irq_desc *desc)
>  		return -EINVAL;
>  
>  	data = irq_desc_get_irq_data(desc);
> -	if (WARN_ON_ONCE(!in_irq() && handle_enforce_irqctx(data)))
> +	if (WARN_ON_ONCE(!in_hardirq() && handle_enforce_irqctx(data)))
>  		return -EPERM;
>  
>  	generic_handle_irq_desc(desc);
> diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
> index 51f24ecd94b2..18f6658f0713 100644
> --- a/kernel/rcu/tree.c
> +++ b/kernel/rcu/tree.c
> @@ -1510,7 +1510,7 @@ static void rcu_gp_kthread_wake(void)
>  {
>  	struct task_struct *t = READ_ONCE(rcu_state.gp_kthread);
>  
> -	if ((current == t && !in_irq() && !in_serving_softirq()) ||
> +	if ((current == t && !in_hardirq() && !in_serving_softirq()) ||
>  	    !READ_ONCE(rcu_state.gp_flags) || !t)
>  		return;
>  	WRITE_ONCE(rcu_state.gp_wake_time, jiffies);
> diff --git a/kernel/rcu/tree_plugin.h b/kernel/rcu/tree_plugin.h
> index de1dc3bb7f70..953171af6d0c 100644
> --- a/kernel/rcu/tree_plugin.h
> +++ b/kernel/rcu/tree_plugin.h
> @@ -672,7 +672,7 @@ static void rcu_read_unlock_special(struct task_struct *t)
>  			   (IS_ENABLED(CONFIG_RCU_BOOST) && irqs_were_disabled &&
>  			    t->rcu_blocked_node);
>  		// Need to defer quiescent state until everything is enabled.
> -		if (use_softirq && (in_irq() || (expboost && !irqs_were_disabled))) {
> +		if (use_softirq && (in_hardirq() || (expboost && !irqs_were_disabled))) {
>  			// Using softirq, safe to awaken, and either the
>  			// wakeup is free or there is either an expedited
>  			// GP in flight or a potential need to deboost.
> diff --git a/kernel/softirq.c b/kernel/softirq.c
> index f3a012179f47..e8c62c4b1e5d 100644
> --- a/kernel/softirq.c
> +++ b/kernel/softirq.c
> @@ -222,7 +222,7 @@ void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
>  	u32 pending;
>  	int curcnt;
>  
> -	WARN_ON_ONCE(in_irq());
> +	WARN_ON_ONCE(in_hardirq());
>  	lockdep_assert_irqs_enabled();
>  
>  	local_irq_save(flags);
> @@ -305,7 +305,7 @@ void __local_bh_disable_ip(unsigned long ip, unsigned int cnt)
>  {
>  	unsigned long flags;
>  
> -	WARN_ON_ONCE(in_irq());
> +	WARN_ON_ONCE(in_hardirq());
>  
>  	raw_local_irq_save(flags);
>  	/*
> @@ -352,14 +352,14 @@ static void __local_bh_enable(unsigned int cnt)
>   */
>  void _local_bh_enable(void)
>  {
> -	WARN_ON_ONCE(in_irq());
> +	WARN_ON_ONCE(in_hardirq());
>  	__local_bh_enable(SOFTIRQ_DISABLE_OFFSET);
>  }
>  EXPORT_SYMBOL(_local_bh_enable);
>  
>  void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
>  {
> -	WARN_ON_ONCE(in_irq());
> +	WARN_ON_ONCE(in_hardirq());
>  	lockdep_assert_irqs_enabled();
>  #ifdef CONFIG_TRACE_IRQFLAGS
>  	local_irq_disable();
> @@ -617,7 +617,7 @@ static inline void tick_irq_exit(void)
>  
>  	/* Make sure that timer wheel updates are propagated */
>  	if ((idle_cpu(cpu) && !need_resched()) || tick_nohz_full_cpu(cpu)) {
> -		if (!in_irq())
> +		if (!in_hardirq())
>  			tick_nohz_irq_exit();
>  	}
>  #endif
> diff --git a/kernel/time/timer.c b/kernel/time/timer.c
> index e3d2c23c413d..dfdbb114a43d 100644
> --- a/kernel/time/timer.c
> +++ b/kernel/time/timer.c
> @@ -1369,7 +1369,7 @@ int del_timer_sync(struct timer_list *timer)
>  	 * don't use it in hardirq context, because it
>  	 * could lead to deadlock.
>  	 */
> -	WARN_ON(in_irq() && !(timer->flags & TIMER_IRQSAFE));
> +	WARN_ON(in_hardirq() && !(timer->flags & TIMER_IRQSAFE));
>  
>  	/*
>  	 * Must be able to sleep on PREEMPT_RT because of the slowpath in
> @@ -1784,7 +1784,7 @@ void update_process_times(int user_tick)
>  	run_local_timers();
>  	rcu_sched_clock_irq(user_tick);
>  #ifdef CONFIG_IRQ_WORK
> -	if (in_irq())
> +	if (in_hardirq())
>  		irq_work_tick();
>  #endif
>  	scheduler_tick();
> diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> index a180abf76d4e..2ed5b77b60ca 100644
> --- a/kernel/trace/trace.h
> +++ b/kernel/trace/trace.h
> @@ -876,7 +876,7 @@ static inline int ftrace_graph_addr(struct ftrace_graph_ent *trace)
>  		 * is set, and called by an interrupt handler, we still
>  		 * want to trace it.
>  		 */
> -		if (in_irq())
> +		if (in_hardirq())
>  			trace_recursion_set(TRACE_IRQ_BIT);
>  		else
>  			trace_recursion_clear(TRACE_IRQ_BIT);
> diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c
> index 0de6837722da..b08d3923de98 100644
> --- a/kernel/trace/trace_functions_graph.c
> +++ b/kernel/trace/trace_functions_graph.c
> @@ -120,7 +120,7 @@ static inline int ftrace_graph_ignore_irqs(void)
>  	if (!ftrace_graph_skip_irqs || trace_recursion_test(TRACE_IRQ_BIT))
>  		return 0;
>  
> -	return in_irq();
> +	return in_hardirq();
>  }
>  
>  int trace_graph_entry(struct ftrace_graph_ent *trace)
> diff --git a/lib/locking-selftest.c b/lib/locking-selftest.c
> index 161108e5d2fe..a4edff9ffc33 100644
> --- a/lib/locking-selftest.c
> +++ b/lib/locking-selftest.c
> @@ -196,7 +196,7 @@ static void init_shared_classes(void)
>  	local_irq_disable();			\
>  	__irq_enter();				\
>  	lockdep_hardirq_threaded();		\
> -	WARN_ON(!in_irq());
> +	WARN_ON(!in_hardirq());
>  
>  #define HARDIRQ_EXIT()				\
>  	__irq_exit();				\
> @@ -2456,7 +2456,7 @@ static void hardirq_exit(int *_)
>  	int notthreaded_hardirq_guard_##name __guard(hardirq_exit);	\
>  	local_irq_disable();						\
>  	__irq_enter();							\
> -	WARN_ON(!in_irq());
> +	WARN_ON(!in_hardirq());
>  
>  static void softirq_exit(int *_)
>  {
> diff --git a/lib/vsprintf.c b/lib/vsprintf.c
> index 26c83943748a..2c5b4351330c 100644
> --- a/lib/vsprintf.c
> +++ b/lib/vsprintf.c
> @@ -865,7 +865,7 @@ char *restricted_pointer(char *buf, char *end, const void *ptr,
>  		 * kptr_restrict==1 cannot be used in IRQ context
>  		 * because its test for CAP_SYSLOG would be meaningless.
>  		 */
> -		if (in_irq() || in_serving_softirq() || in_nmi()) {
> +		if (in_hardirq() || in_serving_softirq() || in_nmi()) {
>  			if (spec.field_width == -1)
>  				spec.field_width = 2 * sizeof(ptr);
>  			return error_string(buf, end, "pK-error", spec);
> -- 
> 2.30.2
> 

-- 
Cheers,
Changbin Du

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH] kernel: in_irq() cleanup
  2021-09-15  0:32 ` Changbin Du
@ 2021-09-15 15:39   ` Paul E. McKenney
  2021-09-28  0:14     ` Changbin Du
  0 siblings, 1 reply; 6+ messages in thread
From: Paul E. McKenney @ 2021-09-15 15:39 UTC (permalink / raw)
  To: Changbin Du
  Cc: Peter Zijlstra, Ingo Molnar, Will Deacon, Josh Triplett,
	Thomas Gleixner, Petr Mladek, linux-kernel, rcu

On Wed, Sep 15, 2021 at 08:32:32AM +0800, Changbin Du wrote:
> Hi, could anyone pick up this cleanup?

If you send the RCU pieces separately, I will take them.

						Thanx, Paul

> On Fri, Aug 13, 2021 at 10:51:01PM +0800, Changbin Du wrote:
> > Replace the obsolete and ambiguos macro in_irq() with new
> > macro in_hardirq().
> > 
> > Signed-off-by: Changbin Du <changbin.du@gmail.com>
> > ---
> >  include/linux/lockdep.h              |  2 +-
> >  include/linux/rcutiny.h              |  2 +-
> >  kernel/irq/irqdesc.c                 |  2 +-
> >  kernel/rcu/tree.c                    |  2 +-
> >  kernel/rcu/tree_plugin.h             |  2 +-
> >  kernel/softirq.c                     | 10 +++++-----
> >  kernel/time/timer.c                  |  4 ++--
> >  kernel/trace/trace.h                 |  2 +-
> >  kernel/trace/trace_functions_graph.c |  2 +-
> >  lib/locking-selftest.c               |  4 ++--
> >  lib/vsprintf.c                       |  2 +-
> >  11 files changed, 17 insertions(+), 17 deletions(-)
> > 
> > diff --git a/include/linux/lockdep.h b/include/linux/lockdep.h
> > index 5cf387813754..e7aa0050bfd8 100644
> > --- a/include/linux/lockdep.h
> > +++ b/include/linux/lockdep.h
> > @@ -633,7 +633,7 @@ do {									\
> >  #define lockdep_assert_in_softirq()					\
> >  do {									\
> >  	WARN_ON_ONCE(__lockdep_enabled			&&		\
> > -		     (!in_softirq() || in_irq() || in_nmi()));		\
> > +		     (!in_softirq() || in_hardirq() || in_nmi()));		\
> >  } while (0)
> >  
> >  #else
> > diff --git a/include/linux/rcutiny.h b/include/linux/rcutiny.h
> > index 953e70fafe38..7fedbd33d5d2 100644
> > --- a/include/linux/rcutiny.h
> > +++ b/include/linux/rcutiny.h
> > @@ -88,7 +88,7 @@ static inline void rcu_irq_enter_irqson(void) { }
> >  static inline void rcu_irq_exit(void) { }
> >  static inline void rcu_irq_exit_check_preempt(void) { }
> >  #define rcu_is_idle_cpu(cpu) \
> > -	(is_idle_task(current) && !in_nmi() && !in_irq() && !in_serving_softirq())
> > +	(is_idle_task(current) && !in_nmi() && !in_hardirq() && !in_serving_softirq())
> >  static inline void exit_rcu(void) { }
> >  static inline bool rcu_preempt_need_deferred_qs(struct task_struct *t)
> >  {
> > diff --git a/kernel/irq/irqdesc.c b/kernel/irq/irqdesc.c
> > index fadb93766020..0ffd6c19259f 100644
> > --- a/kernel/irq/irqdesc.c
> > +++ b/kernel/irq/irqdesc.c
> > @@ -640,7 +640,7 @@ int handle_irq_desc(struct irq_desc *desc)
> >  		return -EINVAL;
> >  
> >  	data = irq_desc_get_irq_data(desc);
> > -	if (WARN_ON_ONCE(!in_irq() && handle_enforce_irqctx(data)))
> > +	if (WARN_ON_ONCE(!in_hardirq() && handle_enforce_irqctx(data)))
> >  		return -EPERM;
> >  
> >  	generic_handle_irq_desc(desc);
> > diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
> > index 51f24ecd94b2..18f6658f0713 100644
> > --- a/kernel/rcu/tree.c
> > +++ b/kernel/rcu/tree.c
> > @@ -1510,7 +1510,7 @@ static void rcu_gp_kthread_wake(void)
> >  {
> >  	struct task_struct *t = READ_ONCE(rcu_state.gp_kthread);
> >  
> > -	if ((current == t && !in_irq() && !in_serving_softirq()) ||
> > +	if ((current == t && !in_hardirq() && !in_serving_softirq()) ||
> >  	    !READ_ONCE(rcu_state.gp_flags) || !t)
> >  		return;
> >  	WRITE_ONCE(rcu_state.gp_wake_time, jiffies);
> > diff --git a/kernel/rcu/tree_plugin.h b/kernel/rcu/tree_plugin.h
> > index de1dc3bb7f70..953171af6d0c 100644
> > --- a/kernel/rcu/tree_plugin.h
> > +++ b/kernel/rcu/tree_plugin.h
> > @@ -672,7 +672,7 @@ static void rcu_read_unlock_special(struct task_struct *t)
> >  			   (IS_ENABLED(CONFIG_RCU_BOOST) && irqs_were_disabled &&
> >  			    t->rcu_blocked_node);
> >  		// Need to defer quiescent state until everything is enabled.
> > -		if (use_softirq && (in_irq() || (expboost && !irqs_were_disabled))) {
> > +		if (use_softirq && (in_hardirq() || (expboost && !irqs_were_disabled))) {
> >  			// Using softirq, safe to awaken, and either the
> >  			// wakeup is free or there is either an expedited
> >  			// GP in flight or a potential need to deboost.
> > diff --git a/kernel/softirq.c b/kernel/softirq.c
> > index f3a012179f47..e8c62c4b1e5d 100644
> > --- a/kernel/softirq.c
> > +++ b/kernel/softirq.c
> > @@ -222,7 +222,7 @@ void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
> >  	u32 pending;
> >  	int curcnt;
> >  
> > -	WARN_ON_ONCE(in_irq());
> > +	WARN_ON_ONCE(in_hardirq());
> >  	lockdep_assert_irqs_enabled();
> >  
> >  	local_irq_save(flags);
> > @@ -305,7 +305,7 @@ void __local_bh_disable_ip(unsigned long ip, unsigned int cnt)
> >  {
> >  	unsigned long flags;
> >  
> > -	WARN_ON_ONCE(in_irq());
> > +	WARN_ON_ONCE(in_hardirq());
> >  
> >  	raw_local_irq_save(flags);
> >  	/*
> > @@ -352,14 +352,14 @@ static void __local_bh_enable(unsigned int cnt)
> >   */
> >  void _local_bh_enable(void)
> >  {
> > -	WARN_ON_ONCE(in_irq());
> > +	WARN_ON_ONCE(in_hardirq());
> >  	__local_bh_enable(SOFTIRQ_DISABLE_OFFSET);
> >  }
> >  EXPORT_SYMBOL(_local_bh_enable);
> >  
> >  void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
> >  {
> > -	WARN_ON_ONCE(in_irq());
> > +	WARN_ON_ONCE(in_hardirq());
> >  	lockdep_assert_irqs_enabled();
> >  #ifdef CONFIG_TRACE_IRQFLAGS
> >  	local_irq_disable();
> > @@ -617,7 +617,7 @@ static inline void tick_irq_exit(void)
> >  
> >  	/* Make sure that timer wheel updates are propagated */
> >  	if ((idle_cpu(cpu) && !need_resched()) || tick_nohz_full_cpu(cpu)) {
> > -		if (!in_irq())
> > +		if (!in_hardirq())
> >  			tick_nohz_irq_exit();
> >  	}
> >  #endif
> > diff --git a/kernel/time/timer.c b/kernel/time/timer.c
> > index e3d2c23c413d..dfdbb114a43d 100644
> > --- a/kernel/time/timer.c
> > +++ b/kernel/time/timer.c
> > @@ -1369,7 +1369,7 @@ int del_timer_sync(struct timer_list *timer)
> >  	 * don't use it in hardirq context, because it
> >  	 * could lead to deadlock.
> >  	 */
> > -	WARN_ON(in_irq() && !(timer->flags & TIMER_IRQSAFE));
> > +	WARN_ON(in_hardirq() && !(timer->flags & TIMER_IRQSAFE));
> >  
> >  	/*
> >  	 * Must be able to sleep on PREEMPT_RT because of the slowpath in
> > @@ -1784,7 +1784,7 @@ void update_process_times(int user_tick)
> >  	run_local_timers();
> >  	rcu_sched_clock_irq(user_tick);
> >  #ifdef CONFIG_IRQ_WORK
> > -	if (in_irq())
> > +	if (in_hardirq())
> >  		irq_work_tick();
> >  #endif
> >  	scheduler_tick();
> > diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> > index a180abf76d4e..2ed5b77b60ca 100644
> > --- a/kernel/trace/trace.h
> > +++ b/kernel/trace/trace.h
> > @@ -876,7 +876,7 @@ static inline int ftrace_graph_addr(struct ftrace_graph_ent *trace)
> >  		 * is set, and called by an interrupt handler, we still
> >  		 * want to trace it.
> >  		 */
> > -		if (in_irq())
> > +		if (in_hardirq())
> >  			trace_recursion_set(TRACE_IRQ_BIT);
> >  		else
> >  			trace_recursion_clear(TRACE_IRQ_BIT);
> > diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c
> > index 0de6837722da..b08d3923de98 100644
> > --- a/kernel/trace/trace_functions_graph.c
> > +++ b/kernel/trace/trace_functions_graph.c
> > @@ -120,7 +120,7 @@ static inline int ftrace_graph_ignore_irqs(void)
> >  	if (!ftrace_graph_skip_irqs || trace_recursion_test(TRACE_IRQ_BIT))
> >  		return 0;
> >  
> > -	return in_irq();
> > +	return in_hardirq();
> >  }
> >  
> >  int trace_graph_entry(struct ftrace_graph_ent *trace)
> > diff --git a/lib/locking-selftest.c b/lib/locking-selftest.c
> > index 161108e5d2fe..a4edff9ffc33 100644
> > --- a/lib/locking-selftest.c
> > +++ b/lib/locking-selftest.c
> > @@ -196,7 +196,7 @@ static void init_shared_classes(void)
> >  	local_irq_disable();			\
> >  	__irq_enter();				\
> >  	lockdep_hardirq_threaded();		\
> > -	WARN_ON(!in_irq());
> > +	WARN_ON(!in_hardirq());
> >  
> >  #define HARDIRQ_EXIT()				\
> >  	__irq_exit();				\
> > @@ -2456,7 +2456,7 @@ static void hardirq_exit(int *_)
> >  	int notthreaded_hardirq_guard_##name __guard(hardirq_exit);	\
> >  	local_irq_disable();						\
> >  	__irq_enter();							\
> > -	WARN_ON(!in_irq());
> > +	WARN_ON(!in_hardirq());
> >  
> >  static void softirq_exit(int *_)
> >  {
> > diff --git a/lib/vsprintf.c b/lib/vsprintf.c
> > index 26c83943748a..2c5b4351330c 100644
> > --- a/lib/vsprintf.c
> > +++ b/lib/vsprintf.c
> > @@ -865,7 +865,7 @@ char *restricted_pointer(char *buf, char *end, const void *ptr,
> >  		 * kptr_restrict==1 cannot be used in IRQ context
> >  		 * because its test for CAP_SYSLOG would be meaningless.
> >  		 */
> > -		if (in_irq() || in_serving_softirq() || in_nmi()) {
> > +		if (in_hardirq() || in_serving_softirq() || in_nmi()) {
> >  			if (spec.field_width == -1)
> >  				spec.field_width = 2 * sizeof(ptr);
> >  			return error_string(buf, end, "pK-error", spec);
> > -- 
> > 2.30.2
> > 
> 
> -- 
> Cheers,
> Changbin Du

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH] kernel: in_irq() cleanup
  2021-09-15 15:39   ` Paul E. McKenney
@ 2021-09-28  0:14     ` Changbin Du
  0 siblings, 0 replies; 6+ messages in thread
From: Changbin Du @ 2021-09-28  0:14 UTC (permalink / raw)
  To: Paul E. McKenney
  Cc: Changbin Du, Peter Zijlstra, Ingo Molnar, Will Deacon,
	Josh Triplett, Thomas Gleixner, Petr Mladek, linux-kernel, rcu

On Wed, Sep 15, 2021 at 08:39:38AM -0700, Paul E. McKenney wrote:
> On Wed, Sep 15, 2021 at 08:32:32AM +0800, Changbin Du wrote:
> > Hi, could anyone pick up this cleanup?
> 
> If you send the RCU pieces separately, I will take them.
> 
> 						Thanx, Paul
> 
sure, I'll do it. Thanks.

-- 
Cheers,
Changbin Du

^ permalink raw reply	[flat|nested] 6+ messages in thread

end of thread, other threads:[~2021-09-28  0:14 UTC | newest]

Thread overview: 6+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-08-13 14:51 [PATCH] kernel: in_irq() cleanup Changbin Du
2021-08-14  2:26 ` Paul E. McKenney
2021-08-26 14:58 ` Petr Mladek
2021-09-15  0:32 ` Changbin Du
2021-09-15 15:39   ` Paul E. McKenney
2021-09-28  0:14     ` Changbin Du

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).