lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  PHC 
Open Source and information security mailing list archives
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Tue, 12 Jan 2021 09:46:03 +0100
From:   Arnd Bergmann <>
To:     Zhen Lei <>
Cc:     Russell King <>,
        Greg Kroah-Hartman <>,
        Will Deacon <>,
        Haojian Zhuang <>,
        Arnd Bergmann <>,
        Rob Herring <>,
        Wei Xu <>,
        devicetree <>,
        linux-arm-kernel <>,
        linux-kernel <>
Subject: Re: [PATCH v3 2/3] dt-bindings: arm: hisilicon: Add binding for L3
 cache controller

On Tue, Jan 12, 2021 at 2:56 AM Zhen Lei <> wrote:

> +---
> +$id:
> +$schema:
> +
> +title: Hisilicon L3 cache controller
> +
> +maintainers:
> +  - Wei Xu <>
> +
> +description: |
> +  The Hisilicon L3 outer cache controller supports a maximum of 36-bit physical
> +  addresses. The data cached in the L3 outer cache can be operated based on the
> +  physical address range or the entire cache.
> +
> +properties:
> +  compatible:
> +    items:
> +      - const: hisilicon,l3cache
> +

The compatible string needs to be a little more specific, I'm sure
you cannot guarantee that this is the only L3 cache controller ever
designed in the past or future by HiSilicon.

Normally when you have an IP block that is itself unnamed but that is specific
to one or a few SoCs but that has no na, the convention is to include the name
of the first SoC that contained it.

Can you share which products actually use this L3 cache controller?

On a related note, what does the memory map look like on this chip?
Do you support more than 4GB of total installed memory? If you
do, this becomes a problem in the future as highmem support
winds down. In fact  anything more than 1GB on a 32-bit system
requires more work on the kernel to be completed before we remove
highmem, and will incur a slowdown. If the total is under 4GB but the
memory is not in a contiguous physical address range. See my
Linaro connect presentation[1] for further information on the topic.



Powered by blists - more mailing lists