[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190926121007.GB4581@hirez.programming.kicks-ass.net>
Date: Thu, 26 Sep 2019 14:10:07 +0200
From: Peter Zijlstra <peterz@...radead.org>
To: Michal Hocko <mhocko@...nel.org>
Cc: Yunsheng Lin <linyunsheng@...wei.com>, catalin.marinas@....com,
will@...nel.org, mingo@...hat.com, bp@...en8.de, rth@...ddle.net,
ink@...assic.park.msu.ru, mattst88@...il.com,
benh@...nel.crashing.org, paulus@...ba.org, mpe@...erman.id.au,
heiko.carstens@...ibm.com, gor@...ux.ibm.com,
borntraeger@...ibm.com, ysato@...rs.sourceforge.jp,
dalias@...c.org, davem@...emloft.net, ralf@...ux-mips.org,
paul.burton@...s.com, jhogan@...nel.org, jiaxun.yang@...goat.com,
chenhc@...ote.com, akpm@...ux-foundation.org, rppt@...ux.ibm.com,
anshuman.khandual@....com, tglx@...utronix.de, cai@....pw,
robin.murphy@....com, linux-arm-kernel@...ts.infradead.org,
linux-kernel@...r.kernel.org, hpa@...or.com, x86@...nel.org,
dave.hansen@...ux.intel.com, luto@...nel.org, len.brown@...el.com,
axboe@...nel.dk, dledford@...hat.com, jeffrey.t.kirsher@...el.com,
linux-alpha@...r.kernel.org, naveen.n.rao@...ux.vnet.ibm.com,
mwb@...ux.vnet.ibm.com, linuxppc-dev@...ts.ozlabs.org,
linux-s390@...r.kernel.org, linux-sh@...r.kernel.org,
sparclinux@...r.kernel.org, tbogendoerfer@...e.de,
linux-mips@...r.kernel.org, rafael@...nel.org,
gregkh@...uxfoundation.org
Subject: Re: [PATCH v6] numa: make node_to_cpumask_map() NUMA_NO_NODE aware
On Thu, Sep 26, 2019 at 11:05:59AM +0200, Peter Zijlstra wrote:
> On Wed, Sep 25, 2019 at 11:45:26PM +0200, Peter Zijlstra wrote:
> > [ 7.149889] [Firmware Bug]: device: 'pci0000:7f': no node assigned on NUMA capable HW
> > [ 7.882888] [Firmware Bug]: device: 'pci0000:ff': no node assigned on NUMA capable HW
>
> Going by the limited number of intel numa boxes I have, it looks like:
>
> socket = (~busid) >> (8-n)
Bah, I got my notes mixed up, it should be: busid >> (8-n)
> where 'n' is the number of bits required to encode the largest socket
> id, ie 1 for 2-socket and 2 for 4 socket.
>
> For 8 socket systems we start using pci domains, and things get more
> 'interesting' :/
Powered by blists - more mailing lists