lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:   Mon, 20 Apr 2020 14:38:02 +0100
From:   Steven Price <steven.price@....com>
To:     Mark Rutland <mark.rutland@....com>
Cc:     Catalin Marinas <catalin.marinas@....com>,
        Will Deacon <will@...nel.org>,
        Anshuman Khandual <anshuman.khandual@....com>,
        Greg Kroah-Hartman <gregkh@...uxfoundation.org>,
        David Hildenbrand <david@...hat.com>,
        linux-kernel@...r.kernel.org,
        Alexios Zavras <alexios.zavras@...el.com>,
        Thomas Gleixner <tglx@...utronix.de>,
        Pan Zhang <zhangpan26@...wei.com>,
        Hsin-Yi Wang <hsinyi@...omium.org>,
        Andrew Morton <akpm@...ux-foundation.org>,
        Logan Gunthorpe <logang@...tatee.com>,
        Ard Biesheuvel <ardb@...nel.org>,
        linux-arm-kernel@...ts.infradead.org,
        Steve Capper <steve.capper@....com>
Subject: Re: [PATCH RESEND] arm64: Map page table of linear map readonly

On 17/04/2020 17:04, Mark Rutland wrote:
> Hi Steven,
> 
> On Fri, Apr 17, 2020 at 04:26:19PM +0100, Steven Price wrote:
>> It's fairly rare that linear mappings need to be updated, so to improve
>> security we can map the leaf page table entries as read-only, this makes
>> it harder for an attacker to modify the permissions of the linear
>> mappings, while the overhead is low because the linear mappings don't
>> need to be changed frequently. When they do need to be updated we can
>> use fixmaps to create a temporary alternative mapping to do the update.
>>
>> Signed-off-by: Steven Price <steven.price@....com>
>> ---
>> Resending as I didn't receive any comments on my first posting[1].
>> Anyone have any views as to whether this is a good idea or not?
> 
> TBH, I'm not too keen on this as the direction of travel seems to be
> that we're going to be messing with the linear map more often for things
> like non-coherent DMA (and potentially for carving out VMs etc), and
> this doesn't have a specific threat in mind.

Fair enough - there may indeed be more security benefit modifying the 
linear map to protect parts of the memory rather than protecting the 
page tables themselves. This is why I made it optional because it 
affects the performance trade-offs of other features.

> I do agree that minimizing the necessary permissions seems nice in
> abstract, but I'm not sure that the complexity it adds is worthwhile.
> I'm similarly not all that keen on the stuff we did for KSMA with the
> root of the page tables, but the bulk of that was nice cleanup
> regardless, and it was unconditional so we only had one behaviour to
> maintain.

Obviously this could be simplified by making it unconditional, but I am 
aware that it has performance implications so it seems sensible to at 
least start with a config option in case it affects a particular 
platform/workload excessively.

Thanks for taking a look.

Steve

> Thanks,
> Mark.
> 
>>
>> [1] https://lore.kernel.org/linux-arm-kernel/20200306173217.44372-1-steven.price@arm.com/
>>
>>   arch/arm64/Kconfig              |  9 ++++++++
>>   arch/arm64/include/asm/fixmap.h | 20 ++++++++++++++++-
>>   arch/arm64/mm/mmu.c             | 38 +++++++++++++++++++++++++++++++--
>>   arch/arm64/mm/pageattr.c        | 36 ++++++++++++++++++++++++++++++-
>>   4 files changed, 99 insertions(+), 4 deletions(-)
>>
>> diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig
>> index 40fb05d96c60..138d8b50d894 100644
>> --- a/arch/arm64/Kconfig
>> +++ b/arch/arm64/Kconfig
>> @@ -1213,6 +1213,15 @@ config RODATA_FULL_DEFAULT_ENABLED
>>   	  This requires the linear region to be mapped down to pages,
>>   	  which may adversely affect performance in some cases.
>>   
>> +config LINEAR_PGTABLE_RO
>> +	bool "Apply r/o permissions to page tables of linear mapping"
>> +	help
>> +	  Apply read-only attributes to the page tables that make up the
>> +	  the linear mapping. This prevents the linear page tables from being
>> +	  inadvertently modified.
>> +
>> +	  This requires rodata=on (or RODATA_FULL_DEFAULT_ENABLED).
>> +
>>   config ARM64_SW_TTBR0_PAN
>>   	bool "Emulate Privileged Access Never using TTBR0_EL1 switching"
>>   	help
>> diff --git a/arch/arm64/include/asm/fixmap.h b/arch/arm64/include/asm/fixmap.h
>> index f987b8a8f325..55bf08151d73 100644
>> --- a/arch/arm64/include/asm/fixmap.h
>> +++ b/arch/arm64/include/asm/fixmap.h
>> @@ -67,6 +67,12 @@ enum fixed_addresses {
>>   	FIX_ENTRY_TRAMP_TEXT,
>>   #define TRAMP_VALIAS		(__fix_to_virt(FIX_ENTRY_TRAMP_TEXT))
>>   #endif /* CONFIG_UNMAP_KERNEL_AT_EL0 */
>> +
>> +#ifdef CONFIG_LINEAR_PGTABLE_RO
>> +	FIX_LINEAR_RO_BEGIN,
>> +	FIX_LINEAR_RO_END = FIX_LINEAR_RO_BEGIN + NR_CPUS - 1,
>> +#endif
>> +
>>   	__end_of_permanent_fixed_addresses,
>>   
>>   	/*
>> @@ -77,7 +83,15 @@ enum fixed_addresses {
>>   #define FIX_BTMAPS_SLOTS	7
>>   #define TOTAL_FIX_BTMAPS	(NR_FIX_BTMAPS * FIX_BTMAPS_SLOTS)
>>   
>> +#ifdef CONFIG_LINEAR_PGTABLE_RO
>> +	/*
>> +	 * Share the space with the LINEAR_RO area as during early boot, the
>> +	 * LINEAR_RO area isn't needed
>> +	 */
>> +	FIX_BTMAP_END = FIX_LINEAR_RO_BEGIN,
>> +#else
>>   	FIX_BTMAP_END = __end_of_permanent_fixed_addresses,
>> +#endif
>>   	FIX_BTMAP_BEGIN = FIX_BTMAP_END + TOTAL_FIX_BTMAPS - 1,
>>   
>>   	/*
>> @@ -89,9 +103,13 @@ enum fixed_addresses {
>>   	FIX_PUD,
>>   	FIX_PGD,
>>   
>> -	__end_of_fixed_addresses
>> +	___end_of_fixed_addresses
>>   };
>>   
>> +static const enum fixed_addresses __end_of_fixed_addresses =
>> +	___end_of_fixed_addresses > __end_of_permanent_fixed_addresses ?
>> +	___end_of_fixed_addresses : __end_of_permanent_fixed_addresses;
>> +
>>   #define FIXADDR_SIZE	(__end_of_permanent_fixed_addresses << PAGE_SHIFT)
>>   #define FIXADDR_START	(FIXADDR_TOP - FIXADDR_SIZE)
>>   
>> diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c
>> index a374e4f51a62..4f9d14b584a5 100644
>> --- a/arch/arm64/mm/mmu.c
>> +++ b/arch/arm64/mm/mmu.c
>> @@ -21,6 +21,7 @@
>>   #include <linux/fs.h>
>>   #include <linux/io.h>
>>   #include <linux/mm.h>
>> +#include <linux/pagewalk.h>
>>   #include <linux/vmalloc.h>
>>   
>>   #include <asm/barrier.h>
>> @@ -456,6 +457,34 @@ void __init mark_linear_text_alias_ro(void)
>>   			    PAGE_KERNEL_RO);
>>   }
>>   
>> +#ifdef CONFIG_LINEAR_PGTABLE_RO
>> +static int __init mark_linear_pmd_ro(pmd_t *pmd, unsigned long addr,
>> +				     unsigned long next, struct mm_walk *walk)
>> +{
>> +	phys_addr_t pmd_phys = pte_offset_phys(pmd, 0);
>> +
>> +	__map_memblock(walk->mm->pgd, pmd_phys, pmd_phys + PAGE_SIZE,
>> +		       PAGE_KERNEL_RO, NO_BLOCK_MAPPINGS | NO_CONT_MAPPINGS);
>> +
>> +	return 0;
>> +}
>> +
>> +static const struct mm_walk_ops mark_linear_pg_ro_ops __initconst = {
>> +	.pmd_entry = mark_linear_pmd_ro,
>> +};
>> +
>> +/*
>> + * Remove the write permission from the leaf page tables of the linear map
>> + */
>> +void __init mark_linear_pg_ro(void)
>> +{
>> +	down_read(&init_mm.mmap_sem);
>> +	walk_page_range_novma(&init_mm, PAGE_OFFSET, PAGE_END,
>> +			      &mark_linear_pg_ro_ops, init_mm.pgd, NULL);
>> +	up_read(&init_mm.mmap_sem);
>> +}
>> +#endif
>> +
>>   static void __init map_mem(pgd_t *pgdp)
>>   {
>>   	phys_addr_t kernel_start = __pa_symbol(_text);
>> @@ -503,7 +532,7 @@ static void __init map_mem(pgd_t *pgdp)
>>   	 * so we should avoid them here.
>>   	 */
>>   	__map_memblock(pgdp, kernel_start, kernel_end,
>> -		       PAGE_KERNEL, NO_CONT_MAPPINGS);
>> +		       PAGE_KERNEL, flags | NO_CONT_MAPPINGS);
>>   	memblock_clear_nomap(kernel_start, kernel_end - kernel_start);
>>   
>>   #ifdef CONFIG_KEXEC_CORE
>> @@ -679,6 +708,11 @@ void __init paging_init(void)
>>   	cpu_replace_ttbr1(lm_alias(swapper_pg_dir));
>>   	init_mm.pgd = swapper_pg_dir;
>>   
>> +#ifdef CONFIG_LINEAR_PGTABLE_RO
>> +	if (rodata_full)
>> +		mark_linear_pg_ro();
>> +#endif
>> +
>>   	memblock_free(__pa_symbol(init_pg_dir),
>>   		      __pa_symbol(init_pg_end) - __pa_symbol(init_pg_dir));
>>   
>> @@ -1187,8 +1221,8 @@ void __set_fixmap(enum fixed_addresses idx,
>>   		set_pte(ptep, pfn_pte(phys >> PAGE_SHIFT, flags));
>>   	} else {
>>   		pte_clear(&init_mm, addr, ptep);
>> -		flush_tlb_kernel_range(addr, addr+PAGE_SIZE);
>>   	}
>> +	flush_tlb_kernel_range(addr, addr + PAGE_SIZE);
>>   }
>>   
>>   void *__init fixmap_remap_fdt(phys_addr_t dt_phys, int *size, pgprot_t prot)
>> diff --git a/arch/arm64/mm/pageattr.c b/arch/arm64/mm/pageattr.c
>> index 250c49008d73..52d34c06656c 100644
>> --- a/arch/arm64/mm/pageattr.c
>> +++ b/arch/arm64/mm/pageattr.c
>> @@ -7,6 +7,7 @@
>>   #include <linux/module.h>
>>   #include <linux/sched.h>
>>   #include <linux/vmalloc.h>
>> +#include <linux/uaccess.h>
>>   
>>   #include <asm/pgtable.h>
>>   #include <asm/set_memory.h>
>> @@ -19,6 +20,33 @@ struct page_change_data {
>>   
>>   bool rodata_full __ro_after_init = IS_ENABLED(CONFIG_RODATA_FULL_DEFAULT_ENABLED);
>>   
>> +#ifdef CONFIG_LINEAR_PGTABLE_RO
>> +static void set_linear_pte_range(pte_t *ptep, pte_t pte)
>> +{
>> +	unsigned long flags;
>> +	unsigned int idx;
>> +	unsigned long addr;
>> +
>> +	local_irq_save(flags);
>> +	preempt_disable();
>> +
>> +	/* During early boot we use FIX_PTE as we don't need a per-CPU slot */
>> +	if (system_state < SYSTEM_SCHEDULING)
>> +		idx = FIX_PTE;
>> +	else
>> +		idx = FIX_LINEAR_RO_BEGIN + smp_processor_id();
>> +
>> +	addr = virt_to_phys(ptep);
>> +	set_fixmap(idx, addr);
>> +	ptep = (pte_t *)(__fix_to_virt(idx) + (addr & ~PAGE_MASK));
>> +	set_pte(ptep, pte);
>> +	clear_fixmap(idx);
>> +
>> +	preempt_enable();
>> +	local_irq_restore(flags);
>> +}
>> +#endif
>> +
>>   static int change_page_range(pte_t *ptep, unsigned long addr, void *data)
>>   {
>>   	struct page_change_data *cdata = data;
>> @@ -27,7 +55,13 @@ static int change_page_range(pte_t *ptep, unsigned long addr, void *data)
>>   	pte = clear_pte_bit(pte, cdata->clear_mask);
>>   	pte = set_pte_bit(pte, cdata->set_mask);
>>   
>> -	set_pte(ptep, pte);
>> +#ifdef CONFIG_LINEAR_PGTABLE_RO
>> +	if (addr >= PAGE_OFFSET)
>> +		set_linear_pte_range(ptep, pte);
>> +	else
>> +#endif
>> +		set_pte(ptep, pte);
>> +
>>   	return 0;
>>   }
>>   
>> -- 
>> 2.20.1
>>
>>
>> _______________________________________________
>> linux-arm-kernel mailing list
>> linux-arm-kernel@...ts.infradead.org
>> http://lists.infradead.org/mailman/listinfo/linux-arm-kernel

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ