[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200526231423.qcsolcpll534sgro@two.firstfloor.org>
Date: Tue, 26 May 2020 16:14:25 -0700
From: Andi Kleen <andi@...stfloor.org>
To: Kees Cook <keescook@...omium.org>
Cc: Andi Kleen <andi@...stfloor.org>,
Peter Zijlstra <peterz@...radead.org>,
Greg KH <gregkh@...uxfoundation.org>, x86@...nel.org,
linux-kernel@...r.kernel.org, sashal@...nel.org,
Andi Kleen <ak@...ux.intel.com>, stable@...r.kernel.org
Subject: Re: [PATCH v1] x86: Pin cr4 FSGSBASE
> And if this is going to be more permanent, we can separate the mask
> (untested):
The FSGSBASE one should not be permanent, it will be replaced
with the full FSGSBASE patches that set that bit correctly.
I was a bit wary of enforcing it for all bits, there might be other
CR4 bits which have benign uses. But I guess the risk of breaking
something there is low.
-Andi
>
>
> diff --git a/arch/x86/kernel/cpu/common.c b/arch/x86/kernel/cpu/common.c
> index bed0cb83fe24..ead64f7420a5 100644
> --- a/arch/x86/kernel/cpu/common.c
> +++ b/arch/x86/kernel/cpu/common.c
> @@ -347,6 +347,8 @@ static __always_inline void setup_umip(struct cpuinfo_x86 *c)
> cr4_clear_bits(X86_CR4_UMIP);
> }
>
> +static const unsigned long cr4_pinned_mask =
> + X86_CR4_SMEP | X86_CR4_SMAP | X86_CR4_UMIP | X86_CR4_FSGSBASE;
> static DEFINE_STATIC_KEY_FALSE_RO(cr_pinning);
> static unsigned long cr4_pinned_bits __ro_after_init;
>
> @@ -371,20 +373,20 @@ EXPORT_SYMBOL(native_write_cr0);
>
> void native_write_cr4(unsigned long val)
> {
> - unsigned long bits_missing = 0;
> + unsigned long bits_changed = 0;
>
> set_register:
> asm volatile("mov %0,%%cr4": "+r" (val), "+m" (cr4_pinned_bits));
>
> if (static_branch_likely(&cr_pinning)) {
> - if (unlikely((val & cr4_pinned_bits) != cr4_pinned_bits)) {
> - bits_missing = ~val & cr4_pinned_bits;
> - val |= bits_missing;
> + if (unlikely((val & cr4_pinned_mask) != cr4_pinned_bits)) {
> + bits_changed = ~val & cr4_pinned_mask;
> + val = (val & ~cr4_pinned_mask) | cr4_pinned_bits;
> goto set_register;
> }
> /* Warn after we've set the missing bits. */
> - WARN_ONCE(bits_missing, "CR4 bits went missing: %lx!?\n",
> - bits_missing);
> + WARN_ONCE(bits_changed, "pinned CR4 bits changed: %lx!?\n",
> + bits_changed);
> }
> }
> EXPORT_SYMBOL(native_write_cr4);
> @@ -396,7 +398,7 @@ void cr4_init(void)
> if (boot_cpu_has(X86_FEATURE_PCID))
> cr4 |= X86_CR4_PCIDE;
> if (static_branch_likely(&cr_pinning))
> - cr4 |= cr4_pinned_bits;
> + cr4 = (cr4 & ~cr4_pinned_mask) | cr4_pinned_bits;
>
> __write_cr4(cr4);
>
> @@ -411,10 +413,7 @@ void cr4_init(void)
> */
> static void __init setup_cr_pinning(void)
> {
> - unsigned long mask;
> -
> - mask = (X86_CR4_SMEP | X86_CR4_SMAP | X86_CR4_UMIP);
> - cr4_pinned_bits = this_cpu_read(cpu_tlbstate.cr4) & mask;
> + cr4_pinned_bits = this_cpu_read(cpu_tlbstate.cr4) & cr4_pinned_mask;
> static_key_enable(&cr_pinning.key);
> }
>
>
> --
> Kees Cook
>
Powered by blists - more mailing lists