[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAMzpN2iJHO01499VsarF1a3ZyjCLUn36FpNqEZeAaXuCoc9GwA@mail.gmail.com>
Date: Mon, 1 Jul 2024 15:22:44 -0400
From: Brian Gerst <brgerst@...il.com>
To: Pawan Gupta <pawan.kumar.gupta@...ux.intel.com>
Cc: Borislav Petkov <bp@...en8.de>, Dave Hansen <dave.hansen@...ux.intel.com>,
linux-kernel@...r.kernel.org, x86@...nel.org,
Robert Gill <rtgill82@...il.com>, Jari Ruusu <jariruusu@...tonmail.com>,
"Linux regression tracking (Thorsten Leemhuis)" <regressions@...mhuis.info>, antonio.gomez.iglesias@...ux.intel.com,
daniel.sneddon@...ux.intel.com
Subject: Re: [PATCH v3] x86/entry_32: Move CLEAR_CPU_BUFFERS before restoring segments
On Mon, Jul 1, 2024 at 2:34 PM Pawan Gupta
<pawan.kumar.gupta@...ux.intel.com> wrote:
>
> Robert Gill reported below #GP when dosemu software was executing vm86()
> system call:
>
> general protection fault: 0000 [#1] PREEMPT SMP
> CPU: 4 PID: 4610 Comm: dosemu.bin Not tainted 6.6.21-gentoo-x86 #1
> Hardware name: Dell Inc. PowerEdge 1950/0H723K, BIOS 2.7.0 10/30/2010
> EIP: restore_all_switch_stack+0xbe/0xcf
> EAX: 00000000 EBX: 00000000 ECX: 00000000 EDX: 00000000
> ESI: 00000000 EDI: 00000000 EBP: 00000000 ESP: ff8affdc
> DS: 0000 ES: 0000 FS: 0000 GS: 0033 SS: 0068 EFLAGS: 00010046
> CR0: 80050033 CR2: 00c2101c CR3: 04b6d000 CR4: 000406d0
> Call Trace:
> show_regs+0x70/0x78
> die_addr+0x29/0x70
> exc_general_protection+0x13c/0x348
> exc_bounds+0x98/0x98
> handle_exception+0x14d/0x14d
> exc_bounds+0x98/0x98
> restore_all_switch_stack+0xbe/0xcf
> exc_bounds+0x98/0x98
> restore_all_switch_stack+0xbe/0xcf
>
> This only happens when VERW based mitigations like MDS, RFDS are enabled.
> This is because segment registers with an arbitrary user value can result
> in #GP when executing VERW. Intel SDM vol. 2C documents the following
> behavior for VERW instruction:
>
> #GP(0) - If a memory operand effective address is outside the CS, DS, ES,
> FS, or GS segment limit.
>
> CLEAR_CPU_BUFFERS macro executes VERW instruction before returning to user
> space. Move CLEAR_CPU_BUFFERS to the macro RESTORE_REGS such that VERW is
> executed before user segments are restored. In the opportunistic SYSEXIT
> path use CLEAR_CPU_BUFFERS_SAFE that uses %ss instead.
>
> Below are the locations where CLEAR_CPU_BUFFERS* is currently being used:
>
> * entry_INT80_32(), entry_SYSENTER_32() and interrupts (via
> handle_exception_return) do:
>
> restore_all_switch_stack:
> [...]
> RESTORE_REGS pop=4 clear_cpu_buf=1
> pop %ebx
> pop %ecx
> pop %edx
> pop %esi
> pop %edi
> pop %ebp
> pop %eax
> verw 0xc0fc92c0 <-------------
> pop %ds
> pop %es
> pop %fs
>
> * Opportunistic SYSEXIT explicitly does CLEAR_CPU_BUFFERS_SAFE:
>
> [...]
> verw %ss:0xc0fc92c0 <-------------
> btrl $0x9,(%esp)
> popf
> pop %eax
> sti
> sysexit
>
> * NMIs use RESTORE_ALL_NMI -> RESTORE_REGS:
>
> nmi_return:
> [...]
> RESTORE_ALL_NMI cr3_reg=%edi
> jmp 0xc0fb22e0 <asm_exc_nmi+612>
> test $0x1000,%edi
> je 0xc0fb22e0 <asm_exc_nmi+612>
> mov %edi,%cr3
> pop %ebx
> pop %ecx
> pop %edx
> pop %esi
> pop %edi
> pop %ebp
> pop %eax
> verw 0xc0fc92c0 <-------------
> pop %ds
> pop %es
> pop %fs
>
> Fixes: a0e2dab44d22 ("x86/entry_32: Add VERW just before userspace transition")
> Cc: stable@...r.kernel.org # 5.10+
> Reported-by: Robert Gill <rtgill82@...il.com>
> Closes: https://bugzilla.kernel.org/show_bug.cgi?id=218707
> Closes: https://lore.kernel.org/all/8c77ccfd-d561-45a1-8ed5-6b75212c7a58@leemhuis.info/
> Suggested-by: Dave Hansen <dave.hansen@...ux.intel.com>
> Suggested-by: Brian Gerst <brgerst@...il.com> # Use %ss
> Signed-off-by: Pawan Gupta <pawan.kumar.gupta@...ux.intel.com>
> ---
> v3:
> - Simplify CLEAR_CPU_BUFFERS_SAFE by using %ss instead of %ds (Brian).
> - Do verw before popf in SYSEXIT path (Jari).
>
> v2: https://lore.kernel.org/r/20240627-fix-dosemu-vm86-v2-1-d5579f698e77@linux.intel.com
> - Safe guard against any other system calls like vm86() that might change %ds (Dave).
>
> v1: https://lore.kernel.org/r/20240426-fix-dosemu-vm86-v1-1-88c826a3f378@linux.intel.com
> ---
> arch/x86/entry/entry_32.S | 28 ++++++++++++++++++++++------
> 1 file changed, 22 insertions(+), 6 deletions(-)
>
> diff --git a/arch/x86/entry/entry_32.S b/arch/x86/entry/entry_32.S
> index d3a814efbff6..6a135478dc2e 100644
> --- a/arch/x86/entry/entry_32.S
> +++ b/arch/x86/entry/entry_32.S
> @@ -253,6 +253,17 @@
> .Lend_\@:
> .endm
>
> +/*
> + * Safer version of CLEAR_CPU_BUFFERS that uses %ss to reference VERW operand
> + * mds_verw_sel. This ensures VERW will not #GP for an arbitrary user %ds
> + * value.
> + */
> +.macro CLEAR_CPU_BUFFERS_SAFE
> + ALTERNATIVE "jmp .Lskip_verw\@", "", X86_FEATURE_CLEAR_CPU_BUF
> + verw %ss:_ASM_RIP(mds_verw_sel)
> +.Lskip_verw\@:
> +.endm
> +
> .macro RESTORE_INT_REGS
> popl %ebx
> popl %ecx
> @@ -263,8 +274,15 @@
> popl %eax
> .endm
>
> -.macro RESTORE_REGS pop=0
> +.macro RESTORE_REGS pop=0 clear_cpu_buf=0
> RESTORE_INT_REGS
> + /*
> + * CLEAR_CPU_BUFFERS must be done before restoring segment
> + * registers to avoid #GP when executing VERW with user %ds.
> + */
> + .if \clear_cpu_buf
> + CLEAR_CPU_BUFFERS
> + .endif
> 1: popl %ds
> 2: popl %es
> 3: popl %fs
> @@ -299,7 +317,7 @@
>
> BUG_IF_WRONG_CR3
>
> - RESTORE_REGS pop=\pop
> + RESTORE_REGS pop=\pop clear_cpu_buf=1
> .endm
>
> .macro CHECK_AND_APPLY_ESPFIX
> @@ -871,6 +889,7 @@ SYM_FUNC_START(entry_SYSENTER_32)
>
> /* Now ready to switch the cr3 */
> SWITCH_TO_USER_CR3 scratch_reg=%eax
> + CLEAR_CPU_BUFFERS_SAFE
>
> /*
> * Restore all flags except IF. (We restore IF separately because
> @@ -881,7 +900,6 @@ SYM_FUNC_START(entry_SYSENTER_32)
> BUG_IF_WRONG_CR3 no_user_check=1
> popfl
> popl %eax
> - CLEAR_CPU_BUFFERS
>
> /*
> * Return back to the vDSO, which will pop ecx and edx.
> @@ -950,8 +968,7 @@ restore_all_switch_stack:
> BUG_IF_WRONG_CR3
>
> /* Restore user state */
> - RESTORE_REGS pop=4 # skip orig_eax/error_code
> - CLEAR_CPU_BUFFERS
> + RESTORE_REGS pop=4 clear_cpu_buf=1 # skip orig_eax/error_code
> .Lirq_return:
> /*
> * ARCH_HAS_MEMBARRIER_SYNC_CORE rely on IRET core serialization
> @@ -1144,7 +1161,6 @@ SYM_CODE_START(asm_exc_nmi)
>
> /* Not on SYSENTER stack. */
> call exc_nmi
> - CLEAR_CPU_BUFFERS
> jmp .Lnmi_return
>
> .Lnmi_from_sysenter_stack:
>
> ---
> base-commit: f2661062f16b2de5d7b6a5c42a9a5c96326b8454
> change-id: 20240426-fix-dosemu-vm86-dd111a01737e
>
> Best regards,
> --
> Thanks,
> Pawan
>
>
Perhaps I should have been a bit clearer, but I meant adding the SS
override to the VERW instructions in their present locations, instead
of moving it into RESTORE_REGS. IIRC we don't want data reads
(including stack pops) between the VERW instruction and IRET/SYSEXIT.
Brian Gerst
Powered by blists - more mailing lists