lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Fri, 27 Sep 2013 17:30:03 +0200
From:	Peter Zijlstra <peterz@...radead.org>
To:	Yuanhan Liu <yuanhan.liu@...ux.intel.com>
Cc:	mingo@...nel.org, hpa@...or.com, linux-kernel@...r.kernel.org,
	tglx@...utronix.de, linux-tip-commits@...r.kernel.org,
	Fengguang Wu <fengguang.wu@...el.com>,
	Huang Ying <ying.huang@...el.com>, lkp@...ux.intel.com
Subject: [PATCH] sched: Revert need_resched() to look at TIF_NEED_RESCHED

Subject: sched: Revert need_resched() to look at TIF_NEED_RESCHED
From: Peter Zijlstra <peterz@...radead.org>
Date: Fri Sep 27 17:20:30 CEST 2013

Yuanhan reported a serious throughput regression in his pigz
benchmark. Using the ftrace patch I found that several idle paths
need more TLC before we can switch the generic need_resched() over to
preempt_need_resched.

The preemption paths benefit most from preempt_need_resched and do
indeed use it; all other need_resched() users don't really care that
much so reverting need_resched() back to tif_need_resched() is the
simple and safe solution.

Reported-by: Yuanhan Liu <yuanhan.liu@...ux.intel.com>
Signed-off-by: Peter Zijlstra <peterz@...radead.org>
---
 arch/x86/include/asm/preempt.h |    8 --------
 include/asm-generic/preempt.h  |    8 --------
 include/linux/sched.h          |    5 +++++
 3 files changed, 5 insertions(+), 16 deletions(-)

--- a/arch/x86/include/asm/preempt.h
+++ b/arch/x86/include/asm/preempt.h
@@ -80,14 +80,6 @@ static __always_inline bool __preempt_co
 }
 
 /*
- * Returns true when we need to resched -- even if we can not.
- */
-static __always_inline bool need_resched(void)
-{
-	return unlikely(test_preempt_need_resched());
-}
-
-/*
  * Returns true when we need to resched and can (barring IRQ state).
  */
 static __always_inline bool should_resched(void)
--- a/include/asm-generic/preempt.h
+++ b/include/asm-generic/preempt.h
@@ -85,14 +85,6 @@ static __always_inline bool __preempt_co
 }
 
 /*
- * Returns true when we need to resched -- even if we can not.
- */
-static __always_inline bool need_resched(void)
-{
-	return unlikely(test_preempt_need_resched());
-}
-
-/*
  * Returns true when we need to resched and can (barring IRQ state).
  */
 static __always_inline bool should_resched(void)
--- a/include/linux/sched.h
+++ b/include/linux/sched.h
@@ -2577,6 +2577,11 @@ static inline bool __must_check current_
 }
 #endif
 
+static __always_inline bool need_resched(void)
+{
+	return unlikely(tif_need_resched());
+}
+
 /*
  * Thread group CPU time accounting.
  */
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists