lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <38ce39d2d4754d76934bb07370eff48b@hisilicon.com>
Date:   Wed, 3 Feb 2021 11:32:32 +0000
From:   "Song Bao Hua (Barry Song)" <song.bao.hua@...ilicon.com>
To:     Tim Chen <tim.c.chen@...ux.intel.com>,
        "valentin.schneider@....com" <valentin.schneider@....com>,
        "catalin.marinas@....com" <catalin.marinas@....com>,
        "will@...nel.org" <will@...nel.org>,
        "rjw@...ysocki.net" <rjw@...ysocki.net>,
        "vincent.guittot@...aro.org" <vincent.guittot@...aro.org>,
        "lenb@...nel.org" <lenb@...nel.org>,
        "gregkh@...uxfoundation.org" <gregkh@...uxfoundation.org>,
        Jonathan Cameron <jonathan.cameron@...wei.com>,
        "mingo@...hat.com" <mingo@...hat.com>,
        "peterz@...radead.org" <peterz@...radead.org>,
        "juri.lelli@...hat.com" <juri.lelli@...hat.com>,
        "dietmar.eggemann@....com" <dietmar.eggemann@....com>,
        "rostedt@...dmis.org" <rostedt@...dmis.org>,
        "bsegall@...gle.com" <bsegall@...gle.com>,
        "mgorman@...e.de" <mgorman@...e.de>,
        "mark.rutland@....com" <mark.rutland@....com>,
        "sudeep.holla@....com" <sudeep.holla@....com>,
        "aubrey.li@...ux.intel.com" <aubrey.li@...ux.intel.com>
CC:     "linux-arm-kernel@...ts.infradead.org" 
        <linux-arm-kernel@...ts.infradead.org>,
        "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
        "linux-acpi@...r.kernel.org" <linux-acpi@...r.kernel.org>,
        "linuxarm@...neuler.org" <linuxarm@...neuler.org>,
        "xuwei (O)" <xuwei5@...wei.com>,
        "Zengtao (B)" <prime.zeng@...ilicon.com>,
        "tiantao (H)" <tiantao6@...ilicon.com>
Subject: RE: [RFC PATCH v3 0/2] scheduler: expose the topology of clusters and
 add cluster scheduler



> -----Original Message-----
> From: Tim Chen [mailto:tim.c.chen@...ux.intel.com]
> Sent: Friday, January 8, 2021 12:17 PM
> To: Song Bao Hua (Barry Song) <song.bao.hua@...ilicon.com>;
> valentin.schneider@....com; catalin.marinas@....com; will@...nel.org;
> rjw@...ysocki.net; vincent.guittot@...aro.org; lenb@...nel.org;
> gregkh@...uxfoundation.org; Jonathan Cameron <jonathan.cameron@...wei.com>;
> mingo@...hat.com; peterz@...radead.org; juri.lelli@...hat.com;
> dietmar.eggemann@....com; rostedt@...dmis.org; bsegall@...gle.com;
> mgorman@...e.de; mark.rutland@....com; sudeep.holla@....com;
> aubrey.li@...ux.intel.com
> Cc: linux-arm-kernel@...ts.infradead.org; linux-kernel@...r.kernel.org;
> linux-acpi@...r.kernel.org; linuxarm@...neuler.org; xuwei (O)
> <xuwei5@...wei.com>; Zengtao (B) <prime.zeng@...ilicon.com>; tiantao (H)
> <tiantao6@...ilicon.com>
> Subject: Re: [RFC PATCH v3 0/2] scheduler: expose the topology of clusters and
> add cluster scheduler
> 
> 
> 
> On 1/6/21 12:30 AM, Barry Song wrote:
> > ARM64 server chip Kunpeng 920 has 6 clusters in each NUMA node, and each
> > cluster has 4 cpus. All clusters share L3 cache data while each cluster
> > has local L3 tag. On the other hand, each cluster will share some
> > internal system bus. This means cache is much more affine inside one cluster
> > than across clusters.
> >
> >     +-----------------------------------+                          +---------+
> >     |  +------+    +------+            +---------------------------+         |
> >     |  | CPU0 |    | cpu1 |             |    +-----------+         |         |
> >     |  +------+    +------+             |    |           |         |         |
> >     |                                   +----+    L3     |         |         |
> >     |  +------+    +------+   cluster   |    |    tag    |         |         |
> >     |  | CPU2 |    | CPU3 |             |    |           |         |         |
> >     |  +------+    +------+             |    +-----------+         |         |
> >     |                                   |                          |         |
> >     +-----------------------------------+                          |         |
> >     +-----------------------------------+                          |         |
> >     |  +------+    +------+             +--------------------------+         |
> >     |  |      |    |      |             |    +-----------+         |         |
> >     |  +------+    +------+             |    |           |         |         |
> >     |                                   |    |    L3     |         |         |
> >     |  +------+    +------+             +----+    tag    |         |         |
> >     |  |      |    |      |             |    |           |         |         |
> >     |  +------+    +------+             |    +-----------+         |         |
> >     |                                   |                          |         |
> >     +-----------------------------------+                          |   L3    |
> >                                                                    |   data  |
> >     +-----------------------------------+                          |         |
> >     |  +------+    +------+             |    +-----------+         |         |
> >     |  |      |    |      |             |    |           |         |         |
> >     |  +------+    +------+             +----+    L3     |         |         |
> >     |                                   |    |    tag    |         |         |
> >     |  +------+    +------+             |    |           |         |         |
> >     |  |      |    |      |            ++    +-----------+         |         |
> >     |  +------+    +------+            |---------------------------+         |
> >     +-----------------------------------|                          |         |
> >     +-----------------------------------|                          |         |
> >     |  +------+    +------+            +---------------------------+         |
> >     |  |      |    |      |             |    +-----------+         |         |
> >     |  +------+    +------+             |    |           |         |         |
> >     |                                   +----+    L3     |         |         |
> >     |  +------+    +------+             |    |    tag    |         |         |
> >     |  |      |    |      |             |    |           |         |         |
> >     |  +------+    +------+             |    +-----------+         |         |
> >     |                                   |                          |         |
> >     +-----------------------------------+                          |         |
> >     +-----------------------------------+                          |         |
> >     |  +------+    +------+             +--------------------------+         |
> >     |  |      |    |      |             |   +-----------+          |         |
> >     |  +------+    +------+             |   |           |          |         |
> >
> >
> 
> There is a similar need for clustering in x86.  Some x86 cores could share L2
> caches that
> is similar to the cluster in Kupeng 920 (e.g. on Jacobsville there are 6 clusters
> of 4 Atom cores, each cluster sharing a separate L2, and 24 cores sharing L3).
> Having a sched domain at the L2 cluster helps spread load among
> L2 domains.  This will reduce L2 cache contention and help with
> performance for low to moderate load scenarios.
> 
> The cluster detection mechanism will need
> to be based on L2 cache sharing in this case.  I suggest making the
> cluster detection to be CPU architecture dependent so both ARM64 and x86 use
> cases
> can be accommodated.
> 
> Attached below are two RFC patches for creating x86 L2
> cache sched domain, sans the idle cpu selection on wake up code.  It is
> similar enough in concept to Barry's patch that we should have a
> single patchset that accommodates both use cases.

Hi Tim, Agreed on this.
hopefully the RFC v4 I am preparing will cover your case.

> 
> Thanks.
> 
> Tim

Thanks
Barry

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ