[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20180721160619.GA19856@amd>
Date: Sat, 21 Jul 2018 18:06:19 +0200
From: Pavel Machek <pavel@....cz>
To: Joerg Roedel <joro@...tes.org>
Cc: Thomas Gleixner <tglx@...utronix.de>,
Ingo Molnar <mingo@...nel.org>,
"H . Peter Anvin" <hpa@...or.com>, x86@...nel.org,
linux-kernel@...r.kernel.org, linux-mm@...ck.org,
Linus Torvalds <torvalds@...ux-foundation.org>,
Andy Lutomirski <luto@...nel.org>,
Dave Hansen <dave.hansen@...el.com>,
Josh Poimboeuf <jpoimboe@...hat.com>,
Juergen Gross <jgross@...e.com>,
Peter Zijlstra <peterz@...radead.org>,
Borislav Petkov <bp@...en8.de>, Jiri Kosina <jkosina@...e.cz>,
Boris Ostrovsky <boris.ostrovsky@...cle.com>,
Brian Gerst <brgerst@...il.com>,
David Laight <David.Laight@...lab.com>,
Denys Vlasenko <dvlasenk@...hat.com>,
Eduardo Valentin <eduval@...zon.com>,
Greg KH <gregkh@...uxfoundation.org>,
Will Deacon <will.deacon@....com>, aliguori@...zon.com,
daniel.gruss@...k.tugraz.at, hughd@...gle.com, keescook@...gle.com,
Andrea Arcangeli <aarcange@...hat.com>,
Waiman Long <llong@...hat.com>,
"David H . Gutteridge" <dhgutteridge@...patico.ca>,
jroedel@...e.de, Arnaldo Carvalho de Melo <acme@...nel.org>,
Alexander Shishkin <alexander.shishkin@...ux.intel.com>,
Jiri Olsa <jolsa@...hat.com>,
Namhyung Kim <namhyung@...nel.org>
Subject: Re: [PATCH 2/3] x86/entry/32: Check for VM86 mode in slow-path check
Hi!
> The SWITCH_TO_KERNEL_STACK macro only checks for CPL == 0 to
> go down the slow and paranoid entry path. The problem is
> that this check also returns true when coming from VM86
> mode. This is not a problem by itself, as the paranoid path
> handles VM86 stack-frames just fine, but it is not necessary
> as the normal code path handles VM86 mode as well (and
> faster).
>
> Extend the check to include VM86 mode. This also makes an
> optimization of the paranoid path possible.
>
> Signed-off-by: Joerg Roedel <jroedel@...e.de>
> ---
> arch/x86/entry/entry_32.S | 12 ++++++++++--
> 1 file changed, 10 insertions(+), 2 deletions(-)
>
> diff --git a/arch/x86/entry/entry_32.S b/arch/x86/entry/entry_32.S
> index 010cdb4..2767c62 100644
> --- a/arch/x86/entry/entry_32.S
> +++ b/arch/x86/entry/entry_32.S
> @@ -414,8 +414,16 @@
> andl $(0x0000ffff), PT_CS(%esp)
>
> /* Special case - entry from kernel mode via entry stack */
> - testl $SEGMENT_RPL_MASK, PT_CS(%esp)
> - jz .Lentry_from_kernel_\@
> +#ifdef CONFIG_VM86
> + movl PT_EFLAGS(%esp), %ecx # mix EFLAGS and CS
> + movb PT_CS(%esp), %cl
> + andl $(X86_EFLAGS_VM | SEGMENT_RPL_MASK), %ecx
> +#else
> + movl PT_CS(%esp), %ecx
> + andl $SEGMENT_RPL_MASK, %ecx
> +#endif
> + cmpl $USER_RPL, %ecx
> + jb .Lentry_from_kernel_\@
Would it make sense to jump to the slow path as we did, and them jump
back if VM86 is detected?
Because VM86 is not really used often these days, and moving partial
registers results in short code but IIRC it will be rather slow.
Pavel
--
(english) http://www.livejournal.com/~pavelmachek
(cesky, pictures) http://atrey.karlin.mff.cuni.cz/~pavel/picture/horses/blog.html
Download attachment "signature.asc" of type "application/pgp-signature" (182 bytes)
Powered by blists - more mailing lists