[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <2272d95.4512.197f7b1354f.Coremail.00107082@163.com>
Date: Fri, 11 Jul 2025 12:14:35 +0800 (CST)
From: "David Wang" <00107082@....com>
To: "Casey Chen" <cachen@...estorage.com>
Cc: akpm@...ux-foundation.org, surenb@...gle.com, kent.overstreet@...ux.dev,
corbet@....net, dennis@...nel.org, tj@...nel.org, cl@...two.org,
vbabka@...e.cz, mhocko@...e.com, jackmanb@...gle.com,
hannes@...xchg.org, ziy@...dia.com, rientjes@...gle.com,
roman.gushchin@...ux.dev, harry.yoo@...cle.com, linux-mm@...ck.org,
linux-kernel@...r.kernel.org, linux-doc@...r.kernel.org,
yzhong@...estorage.com, souravpanda@...gle.com
Subject: Re: [PATCH v3] alloc_tag: add per-NUMA node stats
At 2025-07-11 08:42:05, "Casey Chen" <cachen@...estorage.com> wrote:
>Hi All,
>
>Thanks for reviewing my previous patches. I am replying some comments
>in our previous discussion
>https://lore.kernel.org/all/CAJuCfpHhSUhxer-6MP3503w6520YLfgBTGp7Q9Qm9kgN4TNsfw@mail.gmail.com/T/#u
>
>Most people care about the motivations and usages of this feature.
>Internally, we used to have systems having asymmetric memory to NUMA
>nodes. Node 0 uses a lot of memory but node 1 is pretty empty.
>Requests to allocate memory on node 0 always fail. With this patch, we
>can find the imbalance and optimize the memory usage. Also, David
>Rientjes and Sourav Panda provide their scenarios in which this patch
>would be very useful. It is easy to turn on an off so I think it is
>nice to have, enabling more scenarios in the future.
>
>Andrew / Kent,
>* I agree with Kent on using for_each_possible_cpu rather than
>for_each_online_cpu, considering CPU online/offline.
>* When failing to allocate counters for in-kernel alloc_tag, panic()
>is better than WARN(), eventually the kernel would panic at invalid
>memory access.
>* percpu stats would bloat data structures quite a bit.
>
>David Wang,
>I don't really understand what is 'granularity of calling sites'. If
>NUMA imbalance is found, the calling site could request memory
>allocation from different nodes. Other factors can affect NUMA
>balance, those information can be implemented in a different patch.
I think my concern mostly due to my lack of knowledge and experience of NUMA,
but I still wondering what action to take when " the calling site could request memory
allocation from different nodes", does the calling site needs to detect numa unbalance at runtime
or it should change to hard coded numa node?
By 'granularity of calling sites', i meant to emphasize that information is local per calling site,
not global. What if the numa nodes usage are almost balanced globally, but strangely unbalance locally for some calling site.
"what adjustment the calling site would make to solve numa unbalance" is the *big* question to me
Thanks
David
>
>Thanks,
>Casey
>
Powered by blists - more mailing lists