[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ZWjcA4BA5vZe57Eh@P9FQF9L96D.corp.robot.car>
Date: Thu, 30 Nov 2023 11:01:23 -0800
From: Roman Gushchin <roman.gushchin@...ux.dev>
To: Kent Overstreet <kent.overstreet@...ux.dev>
Cc: Qi Zheng <zhengqi.arch@...edance.com>,
Michal Hocko <mhocko@...e.com>,
Muchun Song <muchun.song@...ux.dev>,
Linux-MM <linux-mm@...ck.org>, linux-kernel@...r.kernel.org,
Andrew Morton <akpm@...ux-foundation.org>,
Dave Chinner <david@...morbit.com>
Subject: Re: [PATCH 2/7] mm: shrinker: Add a .to_text() method for shrinkers
On Wed, Nov 29, 2023 at 10:21:49PM -0500, Kent Overstreet wrote:
> On Thu, Nov 30, 2023 at 11:09:42AM +0800, Qi Zheng wrote:
> >
> >
> > On 2023/11/30 07:11, Kent Overstreet wrote:
> > > On Wed, Nov 29, 2023 at 10:14:54AM +0100, Michal Hocko wrote:
> > > > On Tue 28-11-23 16:34:35, Roman Gushchin wrote:
> > > > > On Tue, Nov 28, 2023 at 02:23:36PM +0800, Qi Zheng wrote:
> > > > [...]
> > > > > > Now I think adding this method might not be a good idea. If we allow
> > > > > > shrinkers to report thier own private information, OOM logs may become
> > > > > > cluttered. Most people only care about some general information when
> > > > > > troubleshooting OOM problem, but not the private information of a
> > > > > > shrinker.
> > > > >
> > > > > I agree with that.
> > > > >
> > > > > It seems that the feature is mostly useful for kernel developers and it's easily
> > > > > achievable by attaching a bpf program to the oom handler. If it requires a bit
> > > > > of work on the bpf side, we can do that instead, but probably not. And this
> > > > > solution can potentially provide way more information in a more flexible way.
> > > > >
> > > > > So I'm not convinced it's a good idea to make the generic oom handling code
> > > > > more complicated and fragile for everybody, as well as making oom reports differ
> > > > > more between kernel versions and configurations.
> > > >
> > > > Completely agreed! From my many years of experience of oom reports
> > > > analysing from production systems I would conclude the following categories
> > > > - clear runaways (and/or memory leaks)
> > > > - userspace consumers - either shmem or anonymous memory
> > > > predominantly consumes the memory, swap is either depleted
> > > > or not configured.
> > > > OOM report is usually useful to pinpoint those as we
> > > > have required counters available
> > > > - kernel memory consumers - if we are lucky they are
> > > > using slab allocator and unreclaimable slab is a huge
> > > > part of the memory consumption. If this is a page
> > > > allocator user the oom repport only helps to deduce
> > > > the fact by looking at how much user + slab + page
> > > > table etc. form. But identifying the root cause is
> > > > close to impossible without something like page_owner
> > > > or a crash dump.
> > > > - misbehaving memory reclaim
> > > > - minority of issues and the oom report is usually
> > > > insufficient to drill down to the root cause. If the
> > > > problem is reproducible then collecting vmstat data
> > > > can give a much better clue.
> > > > - high number of slab reclaimable objects or free swap
> > > > are good indicators. Shrinkers data could be
> > > > potentially helpful in the slab case but I really have
> > > > hard time to remember any such situation.
> > > > On non-production systems the situation is quite different. I can see
> > > > how it could be very beneficial to add a very specific debugging data
> > > > for subsystem/shrinker which is developed and could cause the OOM. For
> > > > that purpose the proposed scheme is rather inflexible AFAICS.
> > >
> > > Considering that you're an MM guy, and that shrinkers are pretty much
> > > universally used by _filesystem_ people - I'm not sure your experience
> > > is the most relevant here?
> > >
> > > The general attitude I've been seeing in this thread has been one of
> > > dismissiveness towards filesystem people. Roman too; back when he was
> >
> > Oh, please don't say that, it seems like you are the only one causing
> > the fight. We deeply respect the opinions of file system developers, so
> > I invited Dave to this thread from the beginning. And you didn't CC
> > linux-fsdevel@...r.kernel.org yourself.
> >
> > > working on his shrinker debug feature I reached out to him, explained
> > > that I was working on my own, and asked about collaborating - got
> > > crickets in response...
> > >
> > > Hmm..
> > >
> > > Besides that, I haven't seen anything what-so-ever out of you guys to
> > > make our lives easier, regarding OOM debugging, nor do you guys even
> > > seem interested in the needs and perspectives of the filesytem people.
> > > Roman, your feature didn't help one bit for OOM debuging - didn't even
> > > come with documentation or hints as to what it's for.
> > >
> > > BPF? Please.
> >
> > (Disclaimer, no intention to start a fight, here are some objective
> > views.)
> >
> > Why not? In addition to printk, there are many good debugging tools
> > worth trying, such as BPF related tools, drgn, etc.
> >
> > For non-bcachefs developers, who knows what those statistics mean?
> >
> > You can use BPF or drgn to traverse in advance to get the address of the
> > bcachefs shrinker structure, and then during OOM, find the bcachefs
> > private structure through the shrinker->private_data member, and then
> > dump the bcachefs private data. Is there any problem with this?
>
> No, BPF is not an excuse for improving our OOM/allocation failure
> reports. BPF/tracing are secondary tools; whenever we're logging
> information about a problem we should strive to log enough information
> to debug the issue.
Ok, a simple question then:
why can't you dump /proc/slabinfo after the OOM?
Unlike anon memory, slab memory (fs caches in particular) should not be heavily
affected by killing some userspace task.
Thanks.
Powered by blists - more mailing lists