[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAJu=L58L4XrACYieQuM412TJuJoD+QYBb=qOcN1MtwdVAPzn2Q@mail.gmail.com>
Date: Tue, 23 Sep 2014 10:04:24 -0700
From: Andres Lagar-Cavilla <andreslc@...gle.com>
To: Paolo Bonzini <pbonzini@...hat.com>
Cc: Gleb Natapov <gleb@...nel.org>, Radim Krcmar <rkrcmar@...hat.com>,
Rik van Riel <riel@...hat.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Andrea Arcangeli <aarcange@...hat.com>,
Peter Feiner <pfeiner@...gle.com>, kvm@...r.kernel.org,
linux-kernel@...r.kernel.org, linux-mm@...ck.org,
Andres Lagar-Cavilla <andreslc@...ogle.com>
Subject: Re: [PATCH v4] kvm: Fix page ageing bugs
On Tue, Sep 23, 2014 at 12:49 AM, Paolo Bonzini <pbonzini@...hat.com> wrote:
> Il 22/09/2014 23:54, Andres Lagar-Cavilla ha scritto:
>> @@ -1406,32 +1406,24 @@ static int kvm_age_rmapp(struct kvm *kvm, unsigned long *rmapp,
>> struct rmap_iterator uninitialized_var(iter);
>> int young = 0;
>>
>> - /*
>> - * In case of absence of EPT Access and Dirty Bits supports,
>> - * emulate the accessed bit for EPT, by checking if this page has
>> - * an EPT mapping, and clearing it if it does. On the next access,
>> - * a new EPT mapping will be established.
>> - * This has some overhead, but not as much as the cost of swapping
>> - * out actively used pages or breaking up actively used hugepages.
>> - */
>> - if (!shadow_accessed_mask) {
>> - young = kvm_unmap_rmapp(kvm, rmapp, slot, data);
>> - goto out;
>> - }
>> + BUG_ON(!shadow_accessed_mask);
>>
>> for (sptep = rmap_get_first(*rmapp, &iter); sptep;
>> sptep = rmap_get_next(&iter)) {
>> + struct kvm_mmu_page *sp;
>> + gfn_t gfn;
>> BUG_ON(!is_shadow_present_pte(*sptep));
>> + /* From spte to gfn. */
>> + sp = page_header(__pa(sptep));
>> + gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt);
>>
>> if (*sptep & shadow_accessed_mask) {
>> young = 1;
>> clear_bit((ffs(shadow_accessed_mask) - 1),
>> (unsigned long *)sptep);
>> }
>> + trace_kvm_age_page(gfn, slot, young);
>
> Yesterday I couldn't think of a way to avoid the
> page_header/kvm_mmu_page_get_gfn on every iteration, but it's actually
> not hard. Instead of passing hva as datum, you can pass (unsigned long)
> &start. Then you can add PAGE_SIZE to it at the end of every call to
> kvm_age_rmapp, and keep the old tracing logic.
I'm not sure. The addition is not always by PAGE_SIZE, since it
depends on the current level we are iterating at in the outer
kvm_handle_hva_range(). IOW, could be PMD_SIZE or even PUD_SIZE, and
is_large_pte() enough to tell?
This is probably worth a general fix, I can see all the callbacks
benefiting from knowing the gfn (passed down by
kvm_handle_hva_range()) without any additional computation, and adding
that to a tracing call if they don't already.
Even passing the level down to the callback would help by cutting down
to one arithmetic op (subtract rmapp from slot rmap base pointer for
that level)
Andres
>
>
> Paolo
--
Andres Lagar-Cavilla | Google Kernel Team | andreslc@...gle.com
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists