[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <20190225132017.GO14858@MiWiFi-R3L-srv>
Date: Mon, 25 Feb 2019 21:20:17 +0800
From: Baoquan He <bhe@...hat.com>
To: "Kirill A. Shutemov" <kirill@...temov.name>
Cc: linux-kernel@...r.kernel.org, dave.hansen@...ux.intel.com,
luto@...nel.org, peterz@...radead.org, tglx@...utronix.de,
mingo@...hat.com, bp@...en8.de, hpa@...or.com, x86@...nel.org,
kirill.shutemov@...ux.intel.com, keescook@...omium.org,
thgarnie@...gle.com
Subject: Re: [PATCH 1/2] x86/mm/KASLR: Only build one PUD entry of area for
real mode trampoline
On 02/25/19 at 03:31pm, Kirill A. Shutemov wrote:
> On Sun, Feb 24, 2019 at 09:22:30PM +0800, Baoquan He wrote:
> > The current code builds identity mapping for real mode treampoline by
> > borrowing page tables from the direct mapping section if KASLR is
> > enabled. It will copy present entries of the first PUD table in 4-level
> > paging mode, or the first P4D table in 5-level paging mode.
> >
> > However, there's only a very small area under low 1 MB reserved
> > for real mode trampoline in reserve_real_mode(). Makes no sense
> > to build up so large area of mapping for it. Since the randomization
> > granularity in 4-level is 1 GB, and 512 GB in 5-level, only copying
> > one PUD entry is enough.
>
> Can we get more of this info into comments in code?
Sure, I will add this to above init_trampoline(). Thanks.
>
> > Hence, only copy one PUD entry of area where physical address 0
> > resides. And this is preparation for later changing the randomization
> > granularity of 5-level paging mode from 512 GB to 1 GB.
> >
> > Signed-off-by: Baoquan He <bhe@...hat.com>
> > ---
> > arch/x86/mm/kaslr.c | 72 ++++++++++++++++++---------------------------
> > 1 file changed, 28 insertions(+), 44 deletions(-)
> >
> > diff --git a/arch/x86/mm/kaslr.c b/arch/x86/mm/kaslr.c
> > index 754b5da91d43..6b2a06c36b6f 100644
> > --- a/arch/x86/mm/kaslr.c
> > +++ b/arch/x86/mm/kaslr.c
> > @@ -226,74 +226,58 @@ void __init kernel_randomize_memory(void)
> >
> > static void __meminit init_trampoline_pud(void)
> > {
> > - unsigned long paddr, paddr_next;
> > + unsigned long paddr, vaddr;
> > pgd_t *pgd;
> > - pud_t *pud_page, *pud_page_tramp;
> > - int i;
> >
> > + p4d_t *p4d_page, *p4d_page_tramp, *p4d, *p4d_tramp;
> > + pud_t *pud_page, *pud_page_tramp, *pud, *pud_tramp;
> > +
> > +
> > + p4d_page_tramp = alloc_low_page();
>
> I believe this line should be under
>
> if (pgtable_l5_enabled()) {
>
> Right?
Yeah, you are right. No need to waste one page in 4-level case.
Will see if there's any other comment, then repost to update.
Thanks
Baoquan
>
> > pud_page_tramp = alloc_low_page();
> >
> > paddr = 0;
> > + vaddr = (unsigned long)__va(paddr);
> > pgd = pgd_offset_k((unsigned long)__va(paddr));
> > - pud_page = (pud_t *) pgd_page_vaddr(*pgd);
> >
> > - for (i = pud_index(paddr); i < PTRS_PER_PUD; i++, paddr = paddr_next) {
> > - pud_t *pud, *pud_tramp;
> > - unsigned long vaddr = (unsigned long)__va(paddr);
> > + if (pgtable_l5_enabled()) {
> > + p4d_page = (p4d_t *) pgd_page_vaddr(*pgd);
> > + p4d = p4d_page + p4d_index(vaddr);
> >
> > - pud_tramp = pud_page_tramp + pud_index(paddr);
> > + pud_page = (pud_t *) p4d_page_vaddr(*p4d);
> > pud = pud_page + pud_index(vaddr);
> > - paddr_next = (paddr & PUD_MASK) + PUD_SIZE;
> > -
> > - *pud_tramp = *pud;
> > - }
> > -
> > - set_pgd(&trampoline_pgd_entry,
> > - __pgd(_KERNPG_TABLE | __pa(pud_page_tramp)));
> > -}
> > -
> > -static void __meminit init_trampoline_p4d(void)
> > -{
> > - unsigned long paddr, paddr_next;
> > - pgd_t *pgd;
> > - p4d_t *p4d_page, *p4d_page_tramp;
> > - int i;
> >
> > - p4d_page_tramp = alloc_low_page();
> > + p4d_tramp = p4d_page_tramp + p4d_index(paddr);
> > + pud_tramp = pud_page_tramp + pud_index(paddr);
> >
> > - paddr = 0;
> > - pgd = pgd_offset_k((unsigned long)__va(paddr));
> > - p4d_page = (p4d_t *) pgd_page_vaddr(*pgd);
> > + *pud_tramp = *pud;
> >
> > - for (i = p4d_index(paddr); i < PTRS_PER_P4D; i++, paddr = paddr_next) {
> > - p4d_t *p4d, *p4d_tramp;
> > - unsigned long vaddr = (unsigned long)__va(paddr);
> > + set_p4d(p4d_tramp,
> > + __p4d(_KERNPG_TABLE | __pa(pud_page_tramp)));
> >
> > - p4d_tramp = p4d_page_tramp + p4d_index(paddr);
> > - p4d = p4d_page + p4d_index(vaddr);
> > - paddr_next = (paddr & P4D_MASK) + P4D_SIZE;
> > + set_pgd(&trampoline_pgd_entry,
> > + __pgd(_KERNPG_TABLE | __pa(p4d_page_tramp)));
> > + } else {
> > + pud_page = (pud_t *) pgd_page_vaddr(*pgd);
> > + pud = pud_page + pud_index(vaddr);
> >
> > - *p4d_tramp = *p4d;
> > + pud_tramp = pud_page_tramp + pud_index(paddr);
> > + *pud_tramp = *pud;
> > + set_pgd(&trampoline_pgd_entry,
> > + __pgd(_KERNPG_TABLE | __pa(pud_page_tramp)));
> > }
> > -
> > - set_pgd(&trampoline_pgd_entry,
> > - __pgd(_KERNPG_TABLE | __pa(p4d_page_tramp)));
> > }
> >
> > /*
> > - * Create PGD aligned trampoline table to allow real mode initialization
> > - * of additional CPUs. Consume only 1 low memory page.
> > + * Create PUD aligned trampoline table to allow real mode initialization
> > + * of additional CPUs. Consume only 1 or 2 low memory pages.
> > */
> > void __meminit init_trampoline(void)
> > {
> > -
> > if (!kaslr_memory_enabled()) {
> > init_trampoline_default();
> > return;
> > }
> >
> > - if (pgtable_l5_enabled())
> > - init_trampoline_p4d();
> > - else
> > - init_trampoline_pud();
> > + init_trampoline_pud();
> > }
> > --
> > 2.17.2
> >
>
> --
> Kirill A. Shutemov
Powered by blists - more mailing lists