[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <5491098f-5508-4665-a8dc-b91a950bbc02@arm.com>
Date: Mon, 16 Jun 2025 10:09:10 +0100
From: Ryan Roberts <ryan.roberts@....com>
To: Yang Shi <yang@...amperecomputing.com>, will@...nel.org,
catalin.marinas@....com, Miko.Lenczewski@....com, dev.jain@....com,
scott@...amperecomputing.com, cl@...two.org
Cc: linux-arm-kernel@...ts.infradead.org, linux-kernel@...r.kernel.org
Subject: Re: [v4 PATCH 0/4] arm64: support FEAT_BBM level 2 and large block
mapping when rodata=full
On 13/06/2025 18:21, Yang Shi wrote:
> Hi Ryan,
>
> Gently ping... any comments for this version?
Hi Yang, yes sorry for slow response - It's been in my queue. I'm going to start
looking at it now and plan to get you some feedback in the next couple of days.
>
> It looks Dev's series is getting stable except some nits. I went through his
> patches and all the call sites for changing page permission. They are:
> 1. change_memory_common(): called by set_memory_{ro|rw|x|nx}. It iterates
> every single page mapped in the vm area then change permission on page basis. It
> depends on whether the vm area is block mapped or not if we want to change
> permission on block mapping.
> 2. set_memory_valid(): it looks it assumes the [addr, addr + size) range is
> mapped contiguously, but it depends on the callers pass in block size (nr > 1).
> There are two sub cases:
> 2.a kfence and debugalloc just work for PTE mapping, so they pass in single
> page.
> 2.b The execmem passes in large page on x86, arm64 has not supported huge
> execmem cache yet, so it should still pass in singe page for the time being. But
> my series + Dev's series can handle both single page mapping and block mapping well
> for this case. So changing permission for block mapping can be
> supported automatically once arm64 supports huge execmem cache.
> 3. set_direct_map_{invalid|default}_noflush(): it looks they are page basis.
> So Dev's series has no change to them.
> 4. realm: if I remember correctly, realm forces PTE mapping for linear address
> space all the time, so no impact.
Yes for realm, we currently force PTE mapping - that's because we need page
granularity for sharing certain portions back to the host. But with this work I
think we will be able to do the splitting on the fly and map using big blocks
even for realms.
>
> So it looks like just #1 may need some extra work. But it seems simple. I should
> just need advance the address range in (1 << vm's order) stride. So there should
> be just some minor changes when I rebase my patches on top of Dev's, mainly
> context changes. It has no impact to the split primitive and repainting linear
> mapping.
I haven't looked at your series yet, but I had assumed that the most convenient
(and only) integration point would be to call your split primitive from dev's
___change_memory_common() (note 3 underscores at beginning). Something like this:
___change_memory_common(unsigned long start, unsigned long size, ...)
{
// This will need to return error for case where splitting would have
// been required but system does not support BBML2_NOABORT
ret = split_mapping_granularity(start, start + size)
if (ret)
return ret;
...
}
>
> Thanks,
> Yang
>
>
> On 5/30/25 7:41 PM, Yang Shi wrote:
>> Changelog
>> =========
>> v4:
>> * Rebased to v6.15-rc4.
>> * Based on Miko's latest BBML2 cpufeature patch (https://lore.kernel.org/
>> linux-arm-kernel/20250428153514.55772-4-miko.lenczewski@....com/).
>> * Keep block mappings rather than splitting to PTEs if it is fully contained
>> per Ryan.
>> * Return -EINVAL if page table allocation failed instead of BUG_ON per Ryan.
>> * When page table allocation failed, return -1 instead of 0 per Ryan.
>> * Allocate page table with GFP_ATOMIC for repainting per Ryan.
>> * Use idmap to wait for repainting is done per Ryan.
>> * Some minor fixes per the discussion for v3.
>> * Some clean up to reduce redundant code.
>>
>> v3:
>> * Rebased to v6.14-rc4.
>> * Based on Miko's BBML2 cpufeature patch (https://lore.kernel.org/linux-
>> arm-kernel/20250228182403.6269-3-miko.lenczewski@....com/).
>> Also included in this series in order to have the complete patchset.
>> * Enhanced __create_pgd_mapping() to handle split as well per Ryan.
>> * Supported CONT mappings per Ryan.
>> * Supported asymmetric system by splitting kernel linear mapping if such
>> system is detected per Ryan. I don't have such system to test, so the
>> testing is done by hacking kernel to call linear mapping repainting
>> unconditionally. The linear mapping doesn't have any block and cont
>> mappings after booting.
>>
>> RFC v2:
>> * Used allowlist to advertise BBM lv2 on the CPUs which can handle TLB
>> conflict gracefully per Will Deacon
>> * Rebased onto v6.13-rc5
>> * https://lore.kernel.org/linux-arm-kernel/20250103011822.1257189-1-
>> yang@...amperecomputing.com/
>>
>> v3: https://lore.kernel.org/linux-arm-kernel/20250304222018.615808-1-
>> yang@...amperecomputing.com/
>> RFC v2: https://lore.kernel.org/linux-arm-kernel/20250103011822.1257189-1-
>> yang@...amperecomputing.com/
>> RFC v1: https://lore.kernel.org/lkml/20241118181711.962576-1-
>> yang@...amperecomputing.com/
>>
>> Description
>> ===========
>> When rodata=full kernel linear mapping is mapped by PTE due to arm's
>> break-before-make rule.
>>
>> A number of performance issues arise when the kernel linear map is using
>> PTE entries due to arm's break-before-make rule:
>> - performance degradation
>> - more TLB pressure
>> - memory waste for kernel page table
>>
>> These issues can be avoided by specifying rodata=on the kernel command
>> line but this disables the alias checks on page table permissions and
>> therefore compromises security somewhat.
>>
>> With FEAT_BBM level 2 support it is no longer necessary to invalidate the
>> page table entry when changing page sizes. This allows the kernel to
>> split large mappings after boot is complete.
>>
>> This patch adds support for splitting large mappings when FEAT_BBM level 2
>> is available and rodata=full is used. This functionality will be used
>> when modifying page permissions for individual page frames.
>>
>> Without FEAT_BBM level 2 we will keep the kernel linear map using PTEs
>> only.
>>
>> If the system is asymmetric, the kernel linear mapping may be repainted once
>> the BBML2 capability is finalized on all CPUs. See patch #4 for more details.
>>
>> We saw significant performance increases in some benchmarks with
>> rodata=full without compromising the security features of the kernel.
>>
>> Testing
>> =======
>> The test was done on AmpereOne machine (192 cores, 1P) with 256GB memory and
>> 4K page size + 48 bit VA.
>>
>> Function test (4K/16K/64K page size)
>> - Kernel boot. Kernel needs change kernel linear mapping permission at
>> boot stage, if the patch didn't work, kernel typically didn't boot.
>> - Module stress from stress-ng. Kernel module load change permission for
>> linear mapping.
>> - A test kernel module which allocates 80% of total memory via vmalloc(),
>> then change the vmalloc area permission to RO, this also change linear
>> mapping permission to RO, then change it back before vfree(). Then launch
>> a VM which consumes almost all physical memory.
>> - VM with the patchset applied in guest kernel too.
>> - Kernel build in VM with guest kernel which has this series applied.
>> - rodata=on. Make sure other rodata mode is not broken.
>> - Boot on the machine which doesn't support BBML2.
>>
>> Performance
>> ===========
>> Memory consumption
>> Before:
>> MemTotal: 258988984 kB
>> MemFree: 254821700 kB
>>
>> After:
>> MemTotal: 259505132 kB
>> MemFree: 255410264 kB
>>
>> Around 500MB more memory are free to use. The larger the machine, the
>> more memory saved.
>>
>> Performance benchmarking
>> * Memcached
>> We saw performance degradation when running Memcached benchmark with
>> rodata=full vs rodata=on. Our profiling pointed to kernel TLB pressure.
>> With this patchset we saw ops/sec is increased by around 3.5%, P99
>> latency is reduced by around 9.6%.
>> The gain mainly came from reduced kernel TLB misses. The kernel TLB
>> MPKI is reduced by 28.5%.
>>
>> The benchmark data is now on par with rodata=on too.
>>
>> * Disk encryption (dm-crypt) benchmark
>> Ran fio benchmark with the below command on a 128G ramdisk (ext4) with disk
>> encryption (by dm-crypt with no read/write workqueue).
>> fio --directory=/data --random_generator=lfsr --norandommap --randrepeat 1 \
>> --status-interval=999 --rw=write --bs=4k --loops=1 --ioengine=sync \
>> --iodepth=1 --numjobs=1 --fsync_on_close=1 --group_reporting --thread \
>> --name=iops-test-job --eta-newline=1 --size 100G
>>
>> The IOPS is increased by 90% - 150% (the variance is high, but the worst
>> number of good case is around 90% more than the best number of bad case).
>> The bandwidth is increased and the avg clat is reduced proportionally.
>>
>> * Sequential file read
>> Read 100G file sequentially on XFS (xfs_io read with page cache populated).
>> The bandwidth is increased by 150%.
>>
>>
>> Yang Shi (4):
>> arm64: cpufeature: add AmpereOne to BBML2 allow list
>> arm64: mm: make __create_pgd_mapping() and helpers non-void
>> arm64: mm: support large block mapping when rodata=full
>> arm64: mm: split linear mapping if BBML2 is not supported on secondary
>> CPUs
>>
>> arch/arm64/include/asm/cpufeature.h | 26 +++++++
>> arch/arm64/include/asm/mmu.h | 4 +
>> arch/arm64/include/asm/pgtable.h | 12 ++-
>> arch/arm64/kernel/cpufeature.c | 30 ++++++--
>> arch/arm64/mm/mmu.c | 505 ++++++++++++++++++++++++++++++++++
>> +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>> +---------------
>> arch/arm64/mm/pageattr.c | 37 +++++++--
>> arch/arm64/mm/proc.S | 41 ++++++++++
>> 7 files changed, 585 insertions(+), 70 deletions(-)
>>
>
Powered by blists - more mailing lists