[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <a40df914-18cf-75ae-4e4d-98332c3d1b56@ti.com>
Date: Mon, 2 Oct 2023 09:17:27 -0500
From: Andrew Davis <afd@...com>
To: Linus Walleij <linus.walleij@...aro.org>,
"Russell King (Oracle)" <linux@...linux.org.uk>,
Nishanth Menon <nm@...com>,
Zhizhou Zhang <zhizhou.zh@...il.com>, <rppt@...nel.org>,
<akpm@...ux-foundation.org>, <vishal.moola@...il.com>,
<arnd@...db.de>, <wangkefeng.wang@...wei.com>,
<willy@...radead.org>, <linux-arm-kernel@...ts.infradead.org>,
<linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] ARM: mm: fix stack corruption when CONFIG_ARM_PV_FIXUP=y
On 9/9/23 3:23 AM, Zhi-zhou Zhang wrote:
> On Fri, Sep 08, 2023 at 11:00:31PM +0200, Linus Walleij wrote:
>> On Fri, Sep 8, 2023 at 3:50 PM Russell King (Oracle)
>> <linux@...linux.org.uk> wrote:
>>
>>> However, it makes a total nonsense of the comment, which explains
>>> precisely why the flush_cache_all() is where it is. Moving it before
>>> that comment means that the comment is now rediculous.
>>
>> Zhizhou, can you look over the comment placement?
>
> Linus, I found the bug on a cortex-a55 cpu with high address memory.
> Since the lr is also corruptted, when flush_cache_all() is done, the
> program continues at the next instruction after fixup_pv_table(). So
> the disabling cache and flush_cache_all() is executed a secondary time.
> Then this time lr is correct so the kernel may boot up as usual.
>
> I read the comment carefully, I am not sure how "to ensure nothing is
> prefetched into the caches" affects the system. My patch doesn't
> prevent instrution prefetch though. But in my board everythings looks
> good.
>
> So I come up with a new fixup plan, that's keep the location of
> flush_cache_all() with adding a flush stack cache before disabling
> cache, the code is as follow, the fix is a bit ugly -- it makes
> assumption stack grow towards low address and flush_cache_all() will
> not occupy more than 32 bytes in the future. Comparing with move
> flush_cache_all() before disabling cache, Which one do you prefer?
> Thanks!
>
> diff --git a/arch/arm/mm/mmu.c b/arch/arm/mm/mmu.c
> index 03fb0fe926f3..83a54c61a86b 100644
> --- a/arch/arm/mm/mmu.c
> +++ b/arch/arm/mm/mmu.c
> @@ -1640,6 +1640,7 @@ static void __init early_paging_init(const struct machine_desc *mdesc)
> unsigned long pa_pgd;
> unsigned int cr, ttbcr;
> long long offset;
> + void *stack;
>
> if (!mdesc->pv_fixup)
> return;
> @@ -1675,7 +1676,14 @@ static void __init early_paging_init(const struct machine_desc *mdesc)
> /* Run the patch stub to update the constants */
> fixup_pv_table(&__pv_table_begin,
> (&__pv_table_end - &__pv_table_begin) << 2);
> - flush_cache_all();
> +
> + /*
> + * clean stack in cacheline that undering memory will be changed in
> + * the following flush_cache_all(). assuming 32 bytes is enough for
> + * flush_cache_all().
Adding this extra clean here seems reasonable, but this comment needs
fixed to give the exact reasoning and warn others to not dirty the stack
after this point. Maybe something like
/*
* The stack is currently in cacheable memory, after caching is disabled
* writes to the stack will bypass the cached stack. If this now stale
* cached stack is then evicted it will overwrite the updated stack in
* memory. Clean the stack's cache-line and then ensure no writes to the
* stack are made between here and disabling the cache below.
*/
Andrew
> + */
> + stack = (void *) (current_stack_pointer - 32);
> + __cpuc_flush_dcache_area(stack, 32);
>
> /*
> * We changing not only the virtual to physical mapping, but also
> @@ -1691,6 +1699,7 @@ static void __init early_paging_init(const struct machine_desc *mdesc)
> asm("mrc p15, 0, %0, c2, c0, 2" : "=r" (ttbcr));
> asm volatile("mcr p15, 0, %0, c2, c0, 2"
> : : "r" (ttbcr & ~(3 << 8 | 3 << 10)));
> + flush_cache_all();
>
> /*
> * Fixup the page tables - this must be in the idmap region as
>
>>
>>> So, please don't put it in the patch system.
>>>
>>> The patch certainly needs to be tested on TI Keystone which is the
>>> primary user of this code.
>>
>> Added Andrew Davis and Nishanth Menon to the thread:
>> can you folks review and test this for Keystone?
>>
>> Yours,
>> Linus Walleij
Powered by blists - more mailing lists