[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <796a4b20-7fa3-3086-efa0-2f728f35ae06@linux.intel.com>
Date: Tue, 10 Aug 2021 12:23:48 -0700
From: Andi Kleen <ak@...ux.intel.com>
To: Dave Hansen <dave.hansen@...el.com>,
"Kirill A. Shutemov" <kirill@...temov.name>,
Borislav Petkov <bp@...en8.de>,
Andy Lutomirski <luto@...nel.org>,
Sean Christopherson <seanjc@...gle.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Joerg Roedel <jroedel@...e.de>
Cc: Kuppuswamy Sathyanarayanan
<sathyanarayanan.kuppuswamy@...ux.intel.com>,
David Rientjes <rientjes@...gle.com>,
Vlastimil Babka <vbabka@...e.cz>,
Tom Lendacky <thomas.lendacky@....com>,
Thomas Gleixner <tglx@...utronix.de>,
Peter Zijlstra <peterz@...radead.org>,
Paolo Bonzini <pbonzini@...hat.com>,
Ingo Molnar <mingo@...hat.com>,
Varad Gautam <varad.gautam@...e.com>,
Dario Faggioli <dfaggioli@...e.com>, x86@...nel.org,
linux-mm@...ck.org, linux-coco@...ts.linux.dev,
linux-kernel@...r.kernel.org,
"Kirill A. Shutemov" <kirill.shutemov@...ux.intel.com>
Subject: Re: [PATCH 1/5] mm: Add support for unaccepted memory
> But, not everyone is going to agree with me.
Both the Intel TDX and the AMD SEV side independently came to opposite
conclusions. In general people care a lot about boot time of VM guests.
>
> This also begs the question of how folks know when this "blip" is over.
> Do we have a counter for offline pages? Is there any way to force page
> acceptance? Or, are we just stuck allocating a bunch of memory to warm
> up the system?
>
> How do folks who care about these new blips avoid them?
It's not different than any other warmup. At warmup time you always have
lots of blips until the working set stabilizes. For example in
virtualization first touch of a new page is usually an EPT violation
handled to the host. Or in the native case you may need to do IO or free
memory. Everybody who based their critical latency percentiles around a
warming up process would be foolish, the picture would be completely
distorted.
So the basic operation is adding some overhead, but I don't think
anything is that unusual compared to the state of the art.
Now perhaps the locking might be a problem if the other operations all
run in parallel, causing unnecessary serialization If that's really a
problem I guess we can optimize later. I don't think there's anything
fundamental about the current locking.
-Andi
Powered by blists - more mailing lists