[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <tip-679cd5ce3bc7755bfe29ec22fa8d2cebede0d7c0@git.kernel.org>
Date: Sat, 23 Feb 2019 02:16:06 -0800
From: tip-bot for Kees Cook <tipbot@...or.com>
To: linux-tip-commits@...r.kernel.org
Cc: kernel-hardening@...ts.openwall.com, keescook@...omium.org,
mingo@...nel.org, jannh@...gle.com, hpa@...or.com,
sean.j.christopherson@...el.com, tglx@...utronix.de,
linux-kernel@...r.kernel.org, linux@...inikbrodowski.net
Subject: [tip:x86/asm] x86/asm: Pin sensitive CR4 bits
Commit-ID: 679cd5ce3bc7755bfe29ec22fa8d2cebede0d7c0
Gitweb: https://git.kernel.org/tip/679cd5ce3bc7755bfe29ec22fa8d2cebede0d7c0
Author: Kees Cook <keescook@...omium.org>
AuthorDate: Thu, 21 Feb 2019 10:09:47 -0800
Committer: Thomas Gleixner <tglx@...utronix.de>
CommitDate: Sat, 23 Feb 2019 11:07:11 +0100
x86/asm: Pin sensitive CR4 bits
Several recent exploits have used direct calls to the native_write_cr4()
function to disable SMEP and SMAP before then continuing their exploits
using userspace memory access. This pins bits of cr4 so that they cannot
be changed through a common function. This is not intended to be general
ROP protection (which would require CFI to defend against properly), but
rather a way to avoid trivial direct function calling (or CFI bypassing
via a matching function prototype) as seen in:
https://googleprojectzero.blogspot.com/2017/05/exploiting-linux-kernel-via-packet.html
(https://github.com/xairy/kernel-exploits/tree/master/CVE-2017-7308)
The goals of this change:
- pin specific bits (SMEP, SMAP, and UMIP) when writing cr4.
- avoid setting the bits too early (they must become pinned only after
first being used).
- pinning mask needs to be read-only during normal runtime.
- pinning needs to be rechecked after set to avoid jumps into the middle
of the function.
Using __ro_after_init on the mask is done so it can't be first disabled
with a malicious write. And since it becomes read-only, it must be avoided
writing to it later (hence the check for bits already having been set
instead of unconditionally writing to the mask).
The use of volatile is done to force the compiler to perform a full reload
of the mask after setting cr4 (to protect against just jumping into the
function past where the masking happens; it must check that the mask was
applied after the set). Due to how this function can be built by the
compiler (especially due to the removal of frame pointers), jumping into
the middle of the function frequently doesn't require stack manipulation to
construct a stack frame (there may only a retq without pops, which is
sufficient for use with exploits like timer overwrites mentioned above).
For example, without the recheck, the function may appear as:
native_write_cr4:
mov [pin], %rbx
or %rbx, %rdi
1: mov %rdi, %cr4
retq
The masking "or" could be trivially bypassed by just calling to label "1"
instead of "native_write_cr4". (CFI will force calls to only be able to
call into native_write_cr4, but CFI and CET are uncommon currently.)
Signed-off-by: Kees Cook <keescook@...omium.org>
Signed-off-by: Thomas Gleixner <tglx@...utronix.de>
Cc: Jann Horn <jannh@...gle.com>
Cc: Sean Christopherson <sean.j.christopherson@...el.com>
Cc: Dominik Brodowski <linux@...inikbrodowski.net>
Cc: Kernel Hardening <kernel-hardening@...ts.openwall.com>
Link: https://lkml.kernel.org/r/20190221180947.GA24138@beast
---
arch/x86/include/asm/special_insns.h | 13 +++++++++++++
arch/x86/kernel/cpu/common.c | 12 +++++++++++-
2 files changed, 24 insertions(+), 1 deletion(-)
diff --git a/arch/x86/include/asm/special_insns.h b/arch/x86/include/asm/special_insns.h
index 43c029cdc3fe..fabda1400137 100644
--- a/arch/x86/include/asm/special_insns.h
+++ b/arch/x86/include/asm/special_insns.h
@@ -72,9 +72,22 @@ static inline unsigned long native_read_cr4(void)
return val;
}
+extern volatile unsigned long cr4_pin;
+
static inline void native_write_cr4(unsigned long val)
{
+again:
+ val |= cr4_pin;
asm volatile("mov %0,%%cr4": : "r" (val), "m" (__force_order));
+ /*
+ * If the MOV above was used directly as a ROP gadget we can
+ * notice the lack of pinned bits in "val" and start the function
+ * from the beginning to gain the cr4_pin bits for sure.
+ */
+ if (WARN_ONCE((val & cr4_pin) != cr4_pin,
+ "Attempt to unpin cr4 bits: %lx, cr4 bypass attack?!",
+ ~val & cr4_pin))
+ goto again;
}
#ifdef CONFIG_X86_64
diff --git a/arch/x86/kernel/cpu/common.c b/arch/x86/kernel/cpu/common.c
index cb28e98a0659..7e0ea4470f8e 100644
--- a/arch/x86/kernel/cpu/common.c
+++ b/arch/x86/kernel/cpu/common.c
@@ -312,10 +312,16 @@ static __init int setup_disable_smep(char *arg)
}
__setup("nosmep", setup_disable_smep);
+volatile unsigned long cr4_pin __ro_after_init;
+EXPORT_SYMBOL_GPL(cr4_pin);
+
static __always_inline void setup_smep(struct cpuinfo_x86 *c)
{
- if (cpu_has(c, X86_FEATURE_SMEP))
+ if (cpu_has(c, X86_FEATURE_SMEP)) {
+ if (!(cr4_pin & X86_CR4_SMEP))
+ cr4_pin |= X86_CR4_SMEP;
cr4_set_bits(X86_CR4_SMEP);
+ }
}
static __init int setup_disable_smap(char *arg)
@@ -334,6 +340,8 @@ static __always_inline void setup_smap(struct cpuinfo_x86 *c)
if (cpu_has(c, X86_FEATURE_SMAP)) {
#ifdef CONFIG_X86_SMAP
+ if (!(cr4_pin & X86_CR4_SMAP))
+ cr4_pin |= X86_CR4_SMAP;
cr4_set_bits(X86_CR4_SMAP);
#else
cr4_clear_bits(X86_CR4_SMAP);
@@ -351,6 +359,8 @@ static __always_inline void setup_umip(struct cpuinfo_x86 *c)
if (!cpu_has(c, X86_FEATURE_UMIP))
goto out;
+ if (!(cr4_pin & X86_CR4_UMIP))
+ cr4_pin |= X86_CR4_UMIP;
cr4_set_bits(X86_CR4_UMIP);
pr_info_once("x86/cpu: User Mode Instruction Prevention (UMIP) activated\n");
Powered by blists - more mailing lists