[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20240717212857.2ymcgikwn6rbulxp@amd.com>
Date: Wed, 17 Jul 2024 16:28:57 -0500
From: Michael Roth <michael.roth@....com>
To: Paolo Bonzini <pbonzini@...hat.com>
CC: <linux-kernel@...r.kernel.org>, <kvm@...r.kernel.org>, <seanjc@...gle.com>
Subject: Re: [PATCH 06/12] KVM: guest_memfd: delay kvm_gmem_prepare_folio()
until the memory is passed to the guest
On Thu, Jul 11, 2024 at 06:27:49PM -0400, Paolo Bonzini wrote:
> Initializing the contents of the folio on fallocate() is unnecessarily
> restrictive. It means that the page is registered with the firmware and
> then it cannot be touched anymore. In particular, this loses the
> possibility of using fallocate() to pre-allocate the page for SEV-SNP
> guests, because kvm_arch_gmem_prepare() then fails.
>
> It's only when the guest actually accesses the page (and therefore
> kvm_gmem_get_pfn() is called) that the page must be cleared from any
> stale host data and registered with the firmware. The up-to-date flag
> is clear if this has to be done (i.e. it is the first access and
> kvm_gmem_populate() has not been called).
>
> All in all, there are enough differences between kvm_gmem_get_pfn() and
> kvm_gmem_populate(), that it's better to separate the two flows completely.
> Extract the bulk of kvm_gmem_get_folio(), which take a folio and end up
> setting its up-to-date flag, to a new function kvm_gmem_prepare_folio();
> these are now done only by the non-__-prefixed kvm_gmem_get_pfn().
> As a bonus, __kvm_gmem_get_pfn() loses its ugly "bool prepare" argument.
>
> The previous semantics, where fallocate() could be used to prepare
> the pages in advance of running the guest, can be accessed with
> KVM_PRE_FAULT_MEMORY.
>
> For now, accessing a page in one VM will attempt to call
> kvm_arch_gmem_prepare() in all of those that have bound the guest_memfd.
> Cleaning this up is left to a separate patch.
>
> Suggested-by: Sean Christopherson <seanjc@...gle.com>
> Signed-off-by: Paolo Bonzini <pbonzini@...hat.com>
> ---
> virt/kvm/guest_memfd.c | 107 ++++++++++++++++++++++++-----------------
> 1 file changed, 63 insertions(+), 44 deletions(-)
>
> diff --git a/virt/kvm/guest_memfd.c b/virt/kvm/guest_memfd.c
> index 9271aba9b7b3..f637327ad8e1 100644
> --- a/virt/kvm/guest_memfd.c
> +++ b/virt/kvm/guest_memfd.c
> @@ -25,7 +25,7 @@ static inline kvm_pfn_t folio_file_pfn(struct folio *folio, pgoff_t index)
> return folio_pfn(folio) + (index & (folio_nr_pages(folio) - 1));
> }
>
> -static int kvm_gmem_prepare_folio(struct inode *inode, pgoff_t index, struct folio *folio)
> +static int __kvm_gmem_prepare_folio(struct inode *inode, pgoff_t index, struct folio *folio)
> {
> #ifdef CONFIG_HAVE_KVM_ARCH_GMEM_PREPARE
> struct list_head *gmem_list = &inode->i_mapping->i_private_list;
> @@ -59,49 +59,60 @@ static int kvm_gmem_prepare_folio(struct inode *inode, pgoff_t index, struct fol
> return 0;
> }
>
> -/* Returns a locked folio on success. */
> -static struct folio *kvm_gmem_get_folio(struct inode *inode, pgoff_t index, bool prepare)
> +/* Process @folio, which contains @gfn (which in turn is contained in @slot),
Need a line break after "/*"
> + * so that the guest can use it. On successful return the guest sees a zero
> + * page so as to avoid leaking host data and the up-to-date flag is set.
Might be worth noting that the folio must be locked so that the
up-to-date flag can be read/set. Or maybe WARN_ONCE() just to avoid any
subtle races slipping in.
> + */
> +static int kvm_gmem_prepare_folio(struct file *file, struct kvm_memory_slot *slot,
> + gfn_t gfn, struct folio *folio)
> {
> - struct folio *folio;
> + unsigned long nr_pages, i;
> + pgoff_t index;
> + int r;
>
> - /* TODO: Support huge pages. */
> - folio = filemap_grab_folio(inode->i_mapping, index);
> - if (IS_ERR(folio))
> - return folio;
> + if (folio_test_uptodate(folio))
> + return 0;
> +
> + nr_pages = folio_nr_pages(folio);
> + for (i = 0; i < nr_pages; i++)
> + clear_highpage(folio_page(folio, i));
>
> /*
> - * Use the up-to-date flag to track whether or not the memory has been
> - * zeroed before being handed off to the guest. There is no backing
> - * storage for the memory, so the folio will remain up-to-date until
> - * it's removed.
> + * Right now the folio order is always going to be zero,
> + * but the code is ready for huge folios, the only
> + * assumption being that the base pgoff of memslots is
> + * naturally aligned according to the folio's order.
> + * The desired order will be passed when creating the
> + * guest_memfd and checked when creating memslots.
> *
> - * TODO: Skip clearing pages when trusted firmware will do it when
> - * assigning memory to the guest.
> + * By making the base pgoff of the memslot naturally aligned
> + * to the folio's order, huge folios are safe to prepare, no
> + * matter what page size is used to map memory into the guest.
> */
> - if (!folio_test_uptodate(folio)) {
> - unsigned long nr_pages = folio_nr_pages(folio);
> - unsigned long i;
> -
> - for (i = 0; i < nr_pages; i++)
> - clear_highpage(folio_page(folio, i));
> - }
> -
> - if (prepare) {
> - int r = kvm_gmem_prepare_folio(inode, index, folio);
> - if (r < 0) {
> - folio_unlock(folio);
> - folio_put(folio);
> - return ERR_PTR(r);
> - }
> + WARN_ON(!IS_ALIGNED(slot->gmem.pgoff, 1 << folio_order(folio)));
> + index = gfn - slot->base_gfn + slot->gmem.pgoff;
> + index = ALIGN_DOWN(index, 1 << folio_order(folio));
I think, when huge folio support is added, we'd also need to make sure
that the huge folio is completely contained within the GPA range
covered by the memslot. For instance, Sean's original THP patch does
this to ensure the order that KVM MMU uses does not result in it
trying to map GFNs beyond the end of the memslot:
@@ -533,9 +577,24 @@ int kvm_gmem_get_pfn(struct kvm *kvm, struct kvm_memory_slot *slot,
page = folio_file_page(folio, index);
*pfn = page_to_pfn(page);
- if (max_order)
+ if (!max_order)
+ goto success;
+
+ *max_order = compound_order(compound_head(page));
+ if (!*max_order)
+ goto success;
+
+ /*
+ * The folio can be mapped with a hugepage if and only if the folio is
+ * fully contained by the range the memslot is bound to. Note, the
+ * caller is responsible for handling gfn alignment, this only deals
+ * with the file binding.
+ */
+ huge_index = ALIGN(index, 1ull << *max_order);
+ if (huge_index < ALIGN(slot->gmem.pgoff, 1ull << *max_order) ||
+ huge_index + (1ull << *max_order) > slot->gmem.pgoff + slot->npages)
*max_order = 0;
https://lore.kernel.org/kvm/20231027182217.3615211-1-seanjc@google.com/T/#mccbd3e8bf9897f0ddbf864e6318d6f2f208b269c
so maybe we'd want to similarly limit the order that
kvm_gmem_prepare_folio() ends up using to match up with that...
...but maybe it's not necessary. I don't think this causes any
problems currently: if a 2nd memslot, e.g.:
________________________
|Slot2 |Slot1 |
|------------|-----------|
GPA |4MB |2.5MB |0MB
+------------+-----------+
overlaps with a folio allocated for another memslot using that same
gmemfd, and the pages are already in a prepared state, then it's
fair to skip re-preparing the range. As long as KVM MMU does not
create mappings for GFN ranges beyond the end of the memslot (e.g.
via the above snippet from Sean's patch), then the guest will
naturally end up with 4K mappings for the range in question, which
will trigger an RMP #NPF which will get quickly resolved with a
PSMASH to split the 2MB RMP entry if some other slot trigger the
preparation of the folio.
With mapping_large_folio_support() not being set yet these conditions
are always met, but when it gets enabled we'd need the above logic
to limit the max_order passed to KVM MMU, so it might make sense to
add a comment on that.
Looks good otherwise.
-Mike
>
> + r = __kvm_gmem_prepare_folio(file_inode(file), index, folio);
> + if (!r)
> folio_mark_uptodate(folio);
> - }
>
> - /*
> - * Ignore accessed, referenced, and dirty flags. The memory is
> - * unevictable and there is no storage to write back to.
> - */
> - return folio;
> + return r;
> +}
> +
> +/*
> + * Returns a locked folio on success. The caller is responsible for
> + * setting the up-to-date flag before the memory is mapped into the guest.
> + * There is no backing storage for the memory, so the folio will remain
> + * up-to-date until it's removed.
> + *
> + * Ignore accessed, referenced, and dirty flags. The memory is
> + * unevictable and there is no storage to write back to.
> + */
> +static struct folio *kvm_gmem_get_folio(struct inode *inode, pgoff_t index)
> +{
> + /* TODO: Support huge pages. */
> + return filemap_grab_folio(inode->i_mapping, index);
> }
>
> static void kvm_gmem_invalidate_begin(struct kvm_gmem *gmem, pgoff_t start,
> @@ -201,7 +212,7 @@ static long kvm_gmem_allocate(struct inode *inode, loff_t offset, loff_t len)
> break;
> }
>
> - folio = kvm_gmem_get_folio(inode, index, true);
> + folio = kvm_gmem_get_folio(inode, index);
> if (IS_ERR(folio)) {
> r = PTR_ERR(folio);
> break;
> @@ -555,7 +566,7 @@ void kvm_gmem_unbind(struct kvm_memory_slot *slot)
> /* Returns a locked folio on success. */
> static struct folio *
> __kvm_gmem_get_pfn(struct file *file, struct kvm_memory_slot *slot,
> - gfn_t gfn, kvm_pfn_t *pfn, int *max_order, bool prepare)
> + gfn_t gfn, kvm_pfn_t *pfn, int *max_order)
> {
> pgoff_t index = gfn - slot->base_gfn + slot->gmem.pgoff;
> struct kvm_gmem *gmem = file->private_data;
> @@ -572,7 +583,7 @@ __kvm_gmem_get_pfn(struct file *file, struct kvm_memory_slot *slot,
> return ERR_PTR(-EIO);
> }
>
> - folio = kvm_gmem_get_folio(file_inode(file), index, prepare);
> + folio = kvm_gmem_get_folio(file_inode(file), index);
> if (IS_ERR(folio))
> return folio;
>
> @@ -594,17 +605,25 @@ int kvm_gmem_get_pfn(struct kvm *kvm, struct kvm_memory_slot *slot,
> {
> struct file *file = kvm_gmem_get_file(slot);
> struct folio *folio;
> + int r = 0;
>
> if (!file)
> return -EFAULT;
>
> - folio = __kvm_gmem_get_pfn(file, slot, gfn, pfn, max_order, true);
> - fput(file);
> - if (IS_ERR(folio))
> - return PTR_ERR(folio);
> + folio = __kvm_gmem_get_pfn(file, slot, gfn, pfn, max_order);
> + if (IS_ERR(folio)) {
> + r = PTR_ERR(folio);
> + goto out;
> + }
>
> + r = kvm_gmem_prepare_folio(file, slot, gfn, folio);
> folio_unlock(folio);
> - return 0;
> + if (r < 0)
> + folio_put(folio);
> +
> +out:
> + fput(file);
> + return r;
> }
> EXPORT_SYMBOL_GPL(kvm_gmem_get_pfn);
>
> @@ -643,7 +662,7 @@ long kvm_gmem_populate(struct kvm *kvm, gfn_t start_gfn, void __user *src, long
> break;
> }
>
> - folio = __kvm_gmem_get_pfn(file, slot, gfn, &pfn, &max_order, false);
> + folio = __kvm_gmem_get_pfn(file, slot, gfn, &pfn, &max_order);
> if (IS_ERR(folio)) {
> ret = PTR_ERR(folio);
> break;
> --
> 2.43.0
>
>
>
Powered by blists - more mailing lists