[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190326225638.GQ18020@tassilo.jf.intel.com>
Date: Tue, 26 Mar 2019 15:56:38 -0700
From: Andi Kleen <ak@...ux.intel.com>
To: Thomas Gleixner <tglx@...utronix.de>
Cc: "Chang S. Bae" <chang.seok.bae@...el.com>,
Ingo Molnar <mingo@...nel.org>,
Andy Lutomirski <luto@...nel.org>,
"H . Peter Anvin" <hpa@...or.com>,
Ravi Shankar <ravi.v.shankar@...el.com>,
LKML <linux-kernel@...r.kernel.org>,
Andrew Cooper <andrew.cooper3@...rix.com>, x86@...nel.org,
Linus Torvalds <torvalds@...ux-foundation.org>,
Greg KH <gregkh@...uxfoundation.org>,
Arjan van de Ven <arjan@...ux.intel.com>
Subject: Re: New feature/ABI review process [was Re: [RESEND PATCH v6 04/12]
x86/fsgsbase/64:..]
>
> If you want to advocate the more complex design of mixed SWAPGS/FSGSBASE
> then provide numbers and not hand-waving. Numbers of real-world workloads,
> not numbers of artificial test cases which exercise the rare worst case.
Well you're proposing the much more complicated solution, not me.
SWAPGS is simple and it works everywhere except for paranoid.
> Yes, it's extra work and it's well spent. If the numbers are not
> significantly different then the simpler and consistent design is a clear
> win.
As long as everything is cache hot it's likely only a couple
of cycles difference (as Intel CPUs are very good executing
crappy code too), but if it's not then you end up with a huge cache miss
cost, causing jitter. That's a problem for real time for example.
> > Accessing user GSBASE needs a couple of SWAPGS operations. It is
> > avoidable if the user GSBASE is saved at kernel entry, being updated as
> > changes, and restored back at kernel exit. However, it seems to spend
> > more cycles for savings and restorations. Little or no benefit was
> > measured from experiments.
>
> So little or no benefit was measured. I don't see how that maps to your
> 'SWAPGS will be a lot faster' claim. One of those claims is obviously
> wrong.
If everything is cache hot it won't make much difference,
but if you have a cache miss you end up eating the cost.
>
> Aside of this needs more than numbers:
>
> 1) Proper documentation how the mixed bag is managed.
How SWAPGS is managed?
Like it always was since 20+ years when the x86_64
port was originally born.
The only case which has to do an two SWAPGS is the
context switch when it switches the base. Everything else
just does SWAPGS at the edges for kernel entries.
> You have a track record of not caring much about either of these, but I
> very much care for good reasons. I've been bitten by glued on and half
> baked patches from Intel in the past 10 years so many times, that I'm
> simply refusing to take anything which is not properly structured and
> documented.
In this case you're proposing the change, the Intel patch just leaves
SWAPGS alone. So you have to describe why it's a good idea.
At least what you proposed on this wasn't convincing
and would be rejected by a proper code review.
-Andi
Powered by blists - more mailing lists