[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <tip-bdd50d7421b2f8fd99f953e1f747e0cb3f3bed64@git.kernel.org>
Date: Tue, 23 Jul 2019 04:49:15 -0700
From: tip-bot for Masahiro Yamada <tipbot@...or.com>
To: linux-tip-commits@...r.kernel.org
Cc: yamada.masahiro@...ionext.com, mingo@...nel.org, hpa@...or.com,
linux-kernel@...r.kernel.org, tglx@...utronix.de
Subject: [tip:x86/cleanups] x86/bitops: Use __builtin_constant_p() directly
instead of IS_IMMEDIATE()
Commit-ID: bdd50d7421b2f8fd99f953e1f747e0cb3f3bed64
Gitweb: https://git.kernel.org/tip/bdd50d7421b2f8fd99f953e1f747e0cb3f3bed64
Author: Masahiro Yamada <yamada.masahiro@...ionext.com>
AuthorDate: Tue, 23 Jul 2019 16:44:15 +0900
Committer: Thomas Gleixner <tglx@...utronix.de>
CommitDate: Tue, 23 Jul 2019 13:44:18 +0200
x86/bitops: Use __builtin_constant_p() directly instead of IS_IMMEDIATE()
__builtin_constant_p(nr) is used everywhere now. It does not make much
sense to define IS_IMMEDIATE() as its alias.
Signed-off-by: Masahiro Yamada <yamada.masahiro@...ionext.com>
Signed-off-by: Thomas Gleixner <tglx@...utronix.de>
Link: https://lkml.kernel.org/r/20190723074415.26811-1-yamada.masahiro@socionext.com
---
arch/x86/include/asm/bitops.h | 7 +++----
1 file changed, 3 insertions(+), 4 deletions(-)
diff --git a/arch/x86/include/asm/bitops.h b/arch/x86/include/asm/bitops.h
index ba15d53c1ca7..7d1f6a49bfae 100644
--- a/arch/x86/include/asm/bitops.h
+++ b/arch/x86/include/asm/bitops.h
@@ -45,14 +45,13 @@
* We do the locked ops that don't return the old value as
* a mask operation on a byte.
*/
-#define IS_IMMEDIATE(nr) (__builtin_constant_p(nr))
#define CONST_MASK_ADDR(nr, addr) WBYTE_ADDR((void *)(addr) + ((nr)>>3))
#define CONST_MASK(nr) (1 << ((nr) & 7))
static __always_inline void
arch_set_bit(long nr, volatile unsigned long *addr)
{
- if (IS_IMMEDIATE(nr)) {
+ if (__builtin_constant_p(nr)) {
asm volatile(LOCK_PREFIX "orb %1,%0"
: CONST_MASK_ADDR(nr, addr)
: "iq" ((u8)CONST_MASK(nr))
@@ -72,7 +71,7 @@ arch___set_bit(long nr, volatile unsigned long *addr)
static __always_inline void
arch_clear_bit(long nr, volatile unsigned long *addr)
{
- if (IS_IMMEDIATE(nr)) {
+ if (__builtin_constant_p(nr)) {
asm volatile(LOCK_PREFIX "andb %1,%0"
: CONST_MASK_ADDR(nr, addr)
: "iq" ((u8)~CONST_MASK(nr)));
@@ -123,7 +122,7 @@ arch___change_bit(long nr, volatile unsigned long *addr)
static __always_inline void
arch_change_bit(long nr, volatile unsigned long *addr)
{
- if (IS_IMMEDIATE(nr)) {
+ if (__builtin_constant_p(nr)) {
asm volatile(LOCK_PREFIX "xorb %1,%0"
: CONST_MASK_ADDR(nr, addr)
: "iq" ((u8)CONST_MASK(nr)));
Powered by blists - more mailing lists