lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Thu, 20 May 2021 11:59:40 +0200
From:   Ard Biesheuvel <ardb@...nel.org>
To:     Greg Kroah-Hartman <gregkh@...uxfoundation.org>
Cc:     Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
        "# 3.4.x" <stable@...r.kernel.org>,
        Nicolas Pitre <nico@...xnic.net>,
        Russell King <rmk+kernel@...linux.org.uk>,
        Sasha Levin <sashal@...nel.org>
Subject: Re: [PATCH 5.12 06/45] ARM: 9058/1: cache-v7: refactor
 v7_invalidate_l1 to avoid clobbering r5/r6

On Thu, 20 May 2021 at 11:25, Greg Kroah-Hartman
<gregkh@...uxfoundation.org> wrote:
>
> From: Ard Biesheuvel <ardb@...nel.org>
>
> [ Upstream commit f9e7a99fb6b86aa6a00e53b34ee6973840e005aa ]
>
> The cache invalidation code in v7_invalidate_l1 can be tweaked to
> re-read the associativity from CCSIDR, and keep the way identifier
> component in a single register that is assigned in the outer loop. This
> way, we need 2 registers less.
>
> Given that the number of sets is typically much larger than the
> associativity, rearrange the code so that the outer loop has the fewer
> number of iterations, ensuring that the re-read of CCSIDR only occurs a
> handful of times in practice.
>
> Fix the whitespace while at it, and update the comment to indicate that
> this code is no longer a clone of anything else.
>
> Acked-by: Nicolas Pitre <nico@...xnic.net>
> Signed-off-by: Ard Biesheuvel <ardb@...nel.org>
> Signed-off-by: Russell King <rmk+kernel@...linux.org.uk>
> Signed-off-by: Sasha Levin <sashal@...nel.org>

Please do NOT backport this to any stable trees.

It has no cc:stable tag
It has no fixes: tag
It was part of a 3 part series, but only the middle patch was selected.
It touches ARM assembly that may assemble without problems but be
completely broken at runtime when used out of the original intended
context.

Tip: disregard patches touching .S files entirely unless they have a
cc:stable or fixes: tag. Or teach the bot never to AUTOSEL anything
authored by me: if it needs to go to -stable, I will ask you myself.

--
Ard.



> ---
>  arch/arm/mm/cache-v7.S | 51 +++++++++++++++++++++---------------------
>  1 file changed, 25 insertions(+), 26 deletions(-)
>
> diff --git a/arch/arm/mm/cache-v7.S b/arch/arm/mm/cache-v7.S
> index dc8f152f3556..e3bc1d6e13d0 100644
> --- a/arch/arm/mm/cache-v7.S
> +++ b/arch/arm/mm/cache-v7.S
> @@ -33,41 +33,40 @@ icache_size:
>   * processor.  We fix this by performing an invalidate, rather than a
>   * clean + invalidate, before jumping into the kernel.
>   *
> - * This function is cloned from arch/arm/mach-tegra/headsmp.S, and needs
> - * to be called for both secondary cores startup and primary core resume
> - * procedures.
> + * This function needs to be called for both secondary cores startup and
> + * primary core resume procedures.
>   */
>  ENTRY(v7_invalidate_l1)
>         mov     r0, #0
>         mcr     p15, 2, r0, c0, c0, 0
>         mrc     p15, 1, r0, c0, c0, 0
>
> -       movw    r1, #0x7fff
> -       and     r2, r1, r0, lsr #13
> +       movw    r3, #0x3ff
> +       and     r3, r3, r0, lsr #3      @ 'Associativity' in CCSIDR[12:3]
> +       clz     r1, r3                  @ WayShift
> +       mov     r2, #1
> +       mov     r3, r3, lsl r1          @ NumWays-1 shifted into bits [31:...]
> +       movs    r1, r2, lsl r1          @ #1 shifted left by same amount
> +       moveq   r1, #1                  @ r1 needs value > 0 even if only 1 way
>
> -       movw    r1, #0x3ff
> +       and     r2, r0, #0x7
> +       add     r2, r2, #4              @ SetShift
>
> -       and     r3, r1, r0, lsr #3      @ NumWays - 1
> -       add     r2, r2, #1              @ NumSets
> +1:     movw    r4, #0x7fff
> +       and     r0, r4, r0, lsr #13     @ 'NumSets' in CCSIDR[27:13]
>
> -       and     r0, r0, #0x7
> -       add     r0, r0, #4      @ SetShift
> -
> -       clz     r1, r3          @ WayShift
> -       add     r4, r3, #1      @ NumWays
> -1:     sub     r2, r2, #1      @ NumSets--
> -       mov     r3, r4          @ Temp = NumWays
> -2:     subs    r3, r3, #1      @ Temp--
> -       mov     r5, r3, lsl r1
> -       mov     r6, r2, lsl r0
> -       orr     r5, r5, r6      @ Reg = (Temp<<WayShift)|(NumSets<<SetShift)
> -       mcr     p15, 0, r5, c7, c6, 2
> -       bgt     2b
> -       cmp     r2, #0
> -       bgt     1b
> -       dsb     st
> -       isb
> -       ret     lr
> +2:     mov     r4, r0, lsl r2          @ NumSet << SetShift
> +       orr     r4, r4, r3              @ Reg = (Temp<<WayShift)|(NumSets<<SetShift)
> +       mcr     p15, 0, r4, c7, c6, 2
> +       subs    r0, r0, #1              @ Set--
> +       bpl     2b
> +       subs    r3, r3, r1              @ Way--
> +       bcc     3f
> +       mrc     p15, 1, r0, c0, c0, 0   @ re-read cache geometry from CCSIDR
> +       b       1b
> +3:     dsb     st
> +       isb
> +       ret     lr
>  ENDPROC(v7_invalidate_l1)
>
>  /*
> --
> 2.30.2
>
>
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ