lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <nubqzts4e6n3a5e7xljdsql7mxgzkobo7besgkfvnhn4thhxk3@reob3iac3psp>
Date: Wed, 7 May 2025 20:01:55 -0400
From: Kent Overstreet <kent.overstreet@...ux.dev>
To: Suren Baghdasaryan <surenb@...gle.com>
Cc: David Wang <00107082@....com>, akpm@...ux-foundation.org, 
	linux-mm@...ck.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH] alloc_tag: avoid mem alloc and iter reset when reading
 allocinfo

On Wed, May 07, 2025 at 11:42:56PM +0000, Suren Baghdasaryan wrote:
> On Wed, May 7, 2025 at 6:19 PM David Wang <00107082@....com> wrote:
> >
> > Hi,
> > Just want to share how I notice those memory allocation behaivors: the cumulative counters~!
> >
> > With cumulative counters, I can identify which module keeps alloc/free memory, by the ratio between
> >  cumulative calls and remaining calls, and maybe an optimization could be applied.
> > Following is top16 I got on my system:
> >
> > +-----------------------------------------+-------+------------------+--------------------+
> > |                  alloc                  | calls | cumulative calls |       ratio        |
> > +-----------------------------------------+-------+------------------+--------------------+
> > |            fs/seq_file.c:584            |   2   |     18064825     |     9032412.5      |
> > |             fs/seq_file.c:38            |   5   |     18148288     |     3629657.6      |
> > |             fs/seq_file.c:63            |   15  |     18153271     | 1210218.0666666667 |
> > |          net/core/skbuff.c:577          |   9   |     10679975     | 1186663.888888889  |
> > |          net/core/skbuff.c:658          |   21  |     11013437     |  524449.380952381  |
> > |             fs/select.c:168             |   7   |     2831226      | 404460.85714285716 |
> > |            lib/alloc_tag.c:51           |   1   |      340649      |      340649.0      |  <--- Here I started
> > |           kernel/signal.c:455           |   1   |      300730      |      300730.0      |
> > | fs/notify/inotify/inotify_fsnotify.c:96 |   1   |      249831      |      249831.0      |
> > |            fs/ext4/dir.c:675            |   3   |      519734      | 173244.66666666666 |
> > |       drivers/usb/host/xhci.c:1555      |   4   |      126402      |      31600.5       |
> > |              fs/locks.c:275             |   36  |      986957      | 27415.472222222223 |
> > |           fs/proc/inode.c:502           |   3   |      63753       |      21251.0       |
> > |              fs/pipe.c:125              |  123  |     2143378      | 17425.837398373984 |
> > |            net/core/scm.c:84            |   3   |      43267       | 14422.333333333334 |
> > |         fs/kernel_read_file.c:80        |   2   |      26910       |      13455.0       |
> > +-----------------------------------------+-------+------------------+--------------------+
> >
> > I think this is another "good" usage for cumulative counters: if a module just keeps alloc/free memory,
> > maybe it is good to move the memory alloc/free to somewhere less frequent.
> >
> > In the case of this patch, a memory allocation for each read-calls, can be moved to opan-calls.
> >
> > If interested, I can re-send the patch for cumulative counters for further discussions.
> 
> Yeah, my issue with cumulative counters is that while they might be
> useful for some analyses, most usecases would probably not benefit
> from them while sharing the performance overhead. OTOH making it
> optional with a separate CONFIG that affects the content of the
> /proc/allocinfo seems like a bad idea to me. Userspace parsers now
> would have to check not only the file version but also whether this
> kernel config is enabled, or handle a possibility of an additional
> column in the output. Does not seem like a good solution to me.

Yeah, I don't see much benefit for cumulative counters over just running
a profiler.

Running a profiler is always the first thing you should do when you care
about CPU usage, that's always the thing that will give you the best
overall picture. If memory allocations are an issue, they'll show up
there.

But generally they're not, because slub is _really damn fast_. People
generally worry about memory allocation overhead a bit too much.

(Memory _layout_, otoh, avoid pointer chasing - that's always worth
worrying about, but cumulative counters won't show you that).

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ