[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190225123150.muyzsycmyrbimzqd@kshutemo-mobl1>
Date: Mon, 25 Feb 2019 15:31:50 +0300
From: "Kirill A. Shutemov" <kirill@...temov.name>
To: Baoquan He <bhe@...hat.com>
Cc: linux-kernel@...r.kernel.org, dave.hansen@...ux.intel.com,
luto@...nel.org, peterz@...radead.org, tglx@...utronix.de,
mingo@...hat.com, bp@...en8.de, hpa@...or.com, x86@...nel.org,
kirill.shutemov@...ux.intel.com, keescook@...omium.org,
thgarnie@...gle.com
Subject: Re: [PATCH 1/2] x86/mm/KASLR: Only build one PUD entry of area for
real mode trampoline
On Sun, Feb 24, 2019 at 09:22:30PM +0800, Baoquan He wrote:
> The current code builds identity mapping for real mode treampoline by
> borrowing page tables from the direct mapping section if KASLR is
> enabled. It will copy present entries of the first PUD table in 4-level
> paging mode, or the first P4D table in 5-level paging mode.
>
> However, there's only a very small area under low 1 MB reserved
> for real mode trampoline in reserve_real_mode(). Makes no sense
> to build up so large area of mapping for it. Since the randomization
> granularity in 4-level is 1 GB, and 512 GB in 5-level, only copying
> one PUD entry is enough.
Can we get more of this info into comments in code?
> Hence, only copy one PUD entry of area where physical address 0
> resides. And this is preparation for later changing the randomization
> granularity of 5-level paging mode from 512 GB to 1 GB.
>
> Signed-off-by: Baoquan He <bhe@...hat.com>
> ---
> arch/x86/mm/kaslr.c | 72 ++++++++++++++++++---------------------------
> 1 file changed, 28 insertions(+), 44 deletions(-)
>
> diff --git a/arch/x86/mm/kaslr.c b/arch/x86/mm/kaslr.c
> index 754b5da91d43..6b2a06c36b6f 100644
> --- a/arch/x86/mm/kaslr.c
> +++ b/arch/x86/mm/kaslr.c
> @@ -226,74 +226,58 @@ void __init kernel_randomize_memory(void)
>
> static void __meminit init_trampoline_pud(void)
> {
> - unsigned long paddr, paddr_next;
> + unsigned long paddr, vaddr;
> pgd_t *pgd;
> - pud_t *pud_page, *pud_page_tramp;
> - int i;
>
> + p4d_t *p4d_page, *p4d_page_tramp, *p4d, *p4d_tramp;
> + pud_t *pud_page, *pud_page_tramp, *pud, *pud_tramp;
> +
> +
> + p4d_page_tramp = alloc_low_page();
I believe this line should be under
if (pgtable_l5_enabled()) {
Right?
> pud_page_tramp = alloc_low_page();
>
> paddr = 0;
> + vaddr = (unsigned long)__va(paddr);
> pgd = pgd_offset_k((unsigned long)__va(paddr));
> - pud_page = (pud_t *) pgd_page_vaddr(*pgd);
>
> - for (i = pud_index(paddr); i < PTRS_PER_PUD; i++, paddr = paddr_next) {
> - pud_t *pud, *pud_tramp;
> - unsigned long vaddr = (unsigned long)__va(paddr);
> + if (pgtable_l5_enabled()) {
> + p4d_page = (p4d_t *) pgd_page_vaddr(*pgd);
> + p4d = p4d_page + p4d_index(vaddr);
>
> - pud_tramp = pud_page_tramp + pud_index(paddr);
> + pud_page = (pud_t *) p4d_page_vaddr(*p4d);
> pud = pud_page + pud_index(vaddr);
> - paddr_next = (paddr & PUD_MASK) + PUD_SIZE;
> -
> - *pud_tramp = *pud;
> - }
> -
> - set_pgd(&trampoline_pgd_entry,
> - __pgd(_KERNPG_TABLE | __pa(pud_page_tramp)));
> -}
> -
> -static void __meminit init_trampoline_p4d(void)
> -{
> - unsigned long paddr, paddr_next;
> - pgd_t *pgd;
> - p4d_t *p4d_page, *p4d_page_tramp;
> - int i;
>
> - p4d_page_tramp = alloc_low_page();
> + p4d_tramp = p4d_page_tramp + p4d_index(paddr);
> + pud_tramp = pud_page_tramp + pud_index(paddr);
>
> - paddr = 0;
> - pgd = pgd_offset_k((unsigned long)__va(paddr));
> - p4d_page = (p4d_t *) pgd_page_vaddr(*pgd);
> + *pud_tramp = *pud;
>
> - for (i = p4d_index(paddr); i < PTRS_PER_P4D; i++, paddr = paddr_next) {
> - p4d_t *p4d, *p4d_tramp;
> - unsigned long vaddr = (unsigned long)__va(paddr);
> + set_p4d(p4d_tramp,
> + __p4d(_KERNPG_TABLE | __pa(pud_page_tramp)));
>
> - p4d_tramp = p4d_page_tramp + p4d_index(paddr);
> - p4d = p4d_page + p4d_index(vaddr);
> - paddr_next = (paddr & P4D_MASK) + P4D_SIZE;
> + set_pgd(&trampoline_pgd_entry,
> + __pgd(_KERNPG_TABLE | __pa(p4d_page_tramp)));
> + } else {
> + pud_page = (pud_t *) pgd_page_vaddr(*pgd);
> + pud = pud_page + pud_index(vaddr);
>
> - *p4d_tramp = *p4d;
> + pud_tramp = pud_page_tramp + pud_index(paddr);
> + *pud_tramp = *pud;
> + set_pgd(&trampoline_pgd_entry,
> + __pgd(_KERNPG_TABLE | __pa(pud_page_tramp)));
> }
> -
> - set_pgd(&trampoline_pgd_entry,
> - __pgd(_KERNPG_TABLE | __pa(p4d_page_tramp)));
> }
>
> /*
> - * Create PGD aligned trampoline table to allow real mode initialization
> - * of additional CPUs. Consume only 1 low memory page.
> + * Create PUD aligned trampoline table to allow real mode initialization
> + * of additional CPUs. Consume only 1 or 2 low memory pages.
> */
> void __meminit init_trampoline(void)
> {
> -
> if (!kaslr_memory_enabled()) {
> init_trampoline_default();
> return;
> }
>
> - if (pgtable_l5_enabled())
> - init_trampoline_p4d();
> - else
> - init_trampoline_pud();
> + init_trampoline_pud();
> }
> --
> 2.17.2
>
--
Kirill A. Shutemov
Powered by blists - more mailing lists