[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20180919162809.30b5c416@mschwideX1>
Date: Wed, 19 Sep 2018 16:28:09 +0200
From: Martin Schwidefsky <schwidefsky@...ibm.com>
To: Peter Zijlstra <peterz@...radead.org>
Cc: will.deacon@....com, aneesh.kumar@...ux.vnet.ibm.com,
akpm@...ux-foundation.org, npiggin@...il.com,
linux-arch@...r.kernel.org, linux-mm@...ck.org,
linux-kernel@...r.kernel.org, linux@...linux.org.uk,
heiko.carstens@...ibm.com,
Linus Torvalds <torvalds@...ux-foundation.org>
Subject: Re: [PATCH 2/2] s390/tlb: convert to generic mmu_gather
On Wed, 19 Sep 2018 14:38:49 +0200
Peter Zijlstra <peterz@...radead.org> wrote:
> On Tue, Sep 18, 2018 at 02:51:51PM +0200, Martin Schwidefsky wrote:
> > +#define pte_free_tlb pte_free_tlb
> > +#define pmd_free_tlb pmd_free_tlb
> > +#define p4d_free_tlb p4d_free_tlb
> > +#define pud_free_tlb pud_free_tlb
>
> > @@ -121,9 +62,18 @@ static inline void tlb_remove_page_size(struct mmu_gather *tlb,
> > * page table from the tlb.
> > */
> > static inline void pte_free_tlb(struct mmu_gather *tlb, pgtable_t pte,
> > + unsigned long address)
> > {
> > + __tlb_adjust_range(tlb, address, PAGE_SIZE);
> > + tlb->mm->context.flush_mm = 1;
> > + tlb->freed_tables = 1;
> > + tlb->cleared_ptes = 1;
> > + /*
> > + * page_table_free_rcu takes care of the allocation bit masks
> > + * of the 2K table fragments in the 4K page table page,
> > + * then calls tlb_remove_table.
> > + */
> > + page_table_free_rcu(tlb, (unsigned long *) pte, address);
>
> (whitespace damage, fixed)
>
> Also, could you perhaps explain the need for that
> page_table_alloc/page_table_free code? That is, I get the comment about
> using 2K page-table fragments out of 4k physical page, but why this
> custom allocator instead of kmem_cache? It feels like there's a little
> extra complication, but it's not immediately obvious what.
The kmem_cache code uses the fields of struct page for its tracking.
pgtable_page_ctor uses the same fields, e.g. for the ptl. Last time
I tried to convert the page_table_alloc/page_table_free to kmem_cache
it just crashed. Plus the split of 4K pages into 2 2K fragments is
done on a per mm basis, that should help a little bit with fragmentation.
> > }
>
> We _could_ use __pte_free_tlb() here I suppose, but...
>
> > /*
> > @@ -139,6 +89,10 @@ static inline void pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd,
> > if (tlb->mm->context.asce_limit <= _REGION3_SIZE)
> > return;
> > pgtable_pmd_page_dtor(virt_to_page(pmd));
> > + __tlb_adjust_range(tlb, address, PAGE_SIZE);
> > + tlb->mm->context.flush_mm = 1;
> > + tlb->freed_tables = 1;
> > + tlb->cleared_puds = 1;
> > tlb_remove_table(tlb, pmd);
> > }
> >
> > @@ -154,6 +108,10 @@ static inline void p4d_free_tlb(struct mmu_gather *tlb, p4d_t *p4d,
> > {
> > if (tlb->mm->context.asce_limit <= _REGION1_SIZE)
> > return;
> > + __tlb_adjust_range(tlb, address, PAGE_SIZE);
> > + tlb->mm->context.flush_mm = 1;
> > + tlb->freed_tables = 1;
> > + tlb->cleared_p4ds = 1;
> > tlb_remove_table(tlb, p4d);
> > }
> >
> > @@ -169,19 +127,11 @@ static inline void pud_free_tlb(struct mmu_gather *tlb, pud_t *pud,
> > {
> > if (tlb->mm->context.asce_limit <= _REGION2_SIZE)
> > return;
> > + tlb->mm->context.flush_mm = 1;
> > + tlb->freed_tables = 1;
> > + tlb->cleared_puds = 1;
> > tlb_remove_table(tlb, pud);
> > }
>
> It's that ASCE limit that makes it impossible to use the generic
> helpers, right?
There are two problems, one of them is related to the ASCE limit:
1) s390 supports 4 different page table layouts. 2-levels (2^31 bytes) for 31-bit compat,
3-levels (2^42 bytes) as the default for 64-bit, 4-levels (2^53) if 4 tera-bytes are
not enough and 5-levels (2^64) for the bragging rights.
The pxd_free_tlb() turn into nops if the number of page table levels require it.
2) The mm->context.flush_mm indication.
That goes back to this beauty in the architecture:
* "A valid table entry must not be changed while it is attached
* to any CPU and may be used for translation by that CPU except to
* (1) invalidate the entry by using INVALIDATE PAGE TABLE ENTRY,
* or INVALIDATE DAT TABLE ENTRY, (2) alter bits 56-63 of a page
* table entry, or (3) make a change by means of a COMPARE AND SWAP
* AND PURGE instruction that purges the TLB."
If one CPU is doing a mmu_gather page table operation on the only active thread
in the system the individual page table updates are done in a lazy fashion with
simple stores. If a second CPU picks up another thread for execution, the
attach_count is increased and the page table updates are done with IPTE/IDTE
from now on. But there might by TLBs of around that are not flushed yet.
We may *not* let the second CPU see these TLBs, otherwise the CPU may start an
instruction, then loose the TLB without being able to recreate it. Due to that
the CPU can end up with a half finished instruction it can not roll back nor
complete, ending in a check-stop. The simplest example is MVC with a length
of e.g. 256 bytes. The instruction has to complete with all 256 bytes moved,
or no bytes may have at all.
That is where the mm->context.flush_mm indication comes into play, if the
second CPU finds the bit set at the time it attaches a thread, it will to
an IDTE for flush all TLBs for the mm.
--
blue skies,
Martin.
"Reality continues to ruin my life." - Calvin.
Powered by blists - more mailing lists