lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:   Mon, 19 Nov 2018 08:11:38 -0800
From:   Ard Biesheuvel <ard.biesheuvel@...aro.org>
To:     sunrui26@...wei.com
Cc:     Catalin Marinas <catalin.marinas@....com>,
        Will Deacon <will.deacon@....com>,
        linux-arm-kernel <linux-arm-kernel@...ts.infradead.org>,
        Linux Kernel Mailing List <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] arm64:crc:accelerated-crc32-by-64bytes

On Sun, 18 Nov 2018 at 23:30, Rui Sun <sunrui26@...wei.com> wrote:
>
> add 64 bytes loop to acceleration calculation
>

Can you share some performance numbers please?

Also, we don't need 64 byte, 32 byte and 16 byte code paths: just make
the 8 byte one a loop as well, and drop the 32 byte and 16 byte ones.


> Signed-off-by: Rui Sun <sunrui26@...wei.com>
> ---
>  arch/arm64/lib/crc32.S | 54 ++++++++++++++++++++++++++++++++++++++++++++++----
>  1 file changed, 50 insertions(+), 4 deletions(-)
>
> diff --git a/arch/arm64/lib/crc32.S b/arch/arm64/lib/crc32.S
> index 5bc1e85..2b37009 100644
> --- a/arch/arm64/lib/crc32.S
> +++ b/arch/arm64/lib/crc32.S
> @@ -15,15 +15,61 @@
>         .cpu            generic+crc
>
>         .macro          __crc32, c
> -0:     subs            x2, x2, #16
> -       b.mi            8f
> +
> +64: cmp     x2, #64
> +    b.lt    32f
> +
> +    adds    x11, x1, #16
> +    adds    x12, x1, #32
> +    adds    x13, x1, #48
> +
> +0 : subs    x2, x2, #64
> +    b.mi    32f
> +
> +    ldp     x3, x4, [x1], #64
> +    ldp     x5, x6, [x11], #64
> +    ldp     x7, x8, [x12], #64
> +    ldp     x9, x10,[x13], #64
> +

Can we do this instead, and get rid of the temp variables?

    ldp     x3, x4, [x1], #64
    ldp     x5, x6, [x1, #-48]
    ldp     x7, x8, [x1, #-32]
    ldp     x9, x10,[x1, #-16]

> + CPU_BE( rev     x3, x3      )
> + CPU_BE( rev     x4, x4      )
> + CPU_BE( rev     x5, x5      )
> + CPU_BE( rev     x6, x6      )
> + CPU_BE( rev     x7, x7      )
> + CPU_BE( rev     x8, x8      )
> + CPU_BE( rev     x9, x9      )
> + CPU_BE( rev     x10,x10     )
> +
> +    crc32\c\()x w0, w0, x3
> +    crc32\c\()x w0, w0, x4
> +    crc32\c\()x w0, w0, x5
> +    crc32\c\()x w0, w0, x6
> +    crc32\c\()x w0, w0, x7
> +    crc32\c\()x w0, w0, x8
> +    crc32\c\()x w0, w0, x9
> +    crc32\c\()x w0, w0, x10
> +
> +    b.ne       0b
> +    ret
> +
> +32: tbz     x2, #5, 16f
> +    ldp     x3, x4, [x1], #16
> +    ldp     x5, x6, [x1], #16
> +CPU_BE( rev     x3, x3      )
> +CPU_BE( rev     x4, x4      )
> +CPU_BE( rev     x5, x5      )
> +CPU_BE( rev     x6, x6      )
> +    crc32\c\()x w0, w0, x3
> +    crc32\c\()x w0, w0, x4
> +    crc32\c\()x w0, w0, x5
> +    crc32\c\()x w0, w0, x6
> +
> +16: tbz     x2, #4, 8f
>         ldp             x3, x4, [x1], #16
>  CPU_BE(        rev             x3, x3          )
>  CPU_BE(        rev             x4, x4          )
>         crc32\c\()x     w0, w0, x3
>         crc32\c\()x     w0, w0, x4
> -       b.ne            0b
> -       ret
>
>  8:     tbz             x2, #3, 4f
>         ldr             x3, [x1], #8
> --
> 1.8.3.1
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ