lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ZzNrIdiHCxTy1QId@x1>
Date: Tue, 12 Nov 2024 11:50:09 -0300
From: Arnaldo Carvalho de Melo <acme@...nel.org>
To: Vlastimil Babka <vbabka@...e.cz>
Cc: Namhyung Kim <namhyung@...nel.org>, Ian Rogers <irogers@...gle.com>,
	Kan Liang <kan.liang@...ux.intel.com>, Jiri Olsa <jolsa@...nel.org>,
	Adrian Hunter <adrian.hunter@...el.com>,
	Peter Zijlstra <peterz@...radead.org>,
	Ingo Molnar <mingo@...nel.org>, LKML <linux-kernel@...r.kernel.org>,
	linux-perf-users@...r.kernel.org,
	Andrii Nakryiko <andrii@...nel.org>, Song Liu <song@...nel.org>,
	bpf@...r.kernel.org, Stephane Eranian <eranian@...gle.com>,
	Roman Gushchin <roman.gushchin@...ux.dev>,
	Hyeonggon Yoo <42.hyeyoo@...il.com>, Kees Cook <kees@...nel.org>
Subject: Re: [PATCH v2 3/4] perf lock contention: Resolve slab object name
 using BPF

On Tue, Nov 12, 2024 at 12:09:24PM +0100, Vlastimil Babka wrote:
> On 11/8/24 07:14, Namhyung Kim wrote:
> > The bpf_get_kmem_cache() kfunc can return an address of the slab cache
> > (kmem_cache).  As it has the name of the slab cache from the iterator,
> > we can use it to symbolize some dynamic kernel locks in a slab.
> > 
> > Before:
> >   root@...tme-ng:/home/namhyung/project/linux# tools/perf/perf lock con -abl sleep 1
> >    contended   total wait     max wait     avg wait            address   symbol
> > 
> >            2      3.34 us      2.87 us      1.67 us   ffff9d7800ad9600    (mutex)
> >            2      2.16 us      1.93 us      1.08 us   ffff9d7804b992d8    (mutex)
> >            4      1.37 us       517 ns       343 ns   ffff9d78036e6e00    (mutex)
> >            1      1.27 us      1.27 us      1.27 us   ffff9d7804b99378    (mutex)
> >            2       845 ns       599 ns       422 ns   ffffffff9e1c3620   delayed_uprobe_lock (mutex)
> >            1       845 ns       845 ns       845 ns   ffffffff9da0b280   jiffies_lock (spinlock)
> >            2       377 ns       259 ns       188 ns   ffffffff9e1cf840   pcpu_alloc_mutex (mutex)
> >            1       305 ns       305 ns       305 ns   ffffffff9e1b4cf8   tracepoint_srcu_srcu_usage (mutex)
> >            1       295 ns       295 ns       295 ns   ffffffff9e1c0940   pack_mutex (mutex)
> >            1       232 ns       232 ns       232 ns   ffff9d7804b7d8d8    (mutex)
> >            1       180 ns       180 ns       180 ns   ffffffff9e1b4c28   tracepoint_srcu_srcu_usage (mutex)
> >            1       165 ns       165 ns       165 ns   ffffffff9da8b3a0   text_mutex (mutex)
> > 
> > After:
> >   root@...tme-ng:/home/namhyung/project/linux# tools/perf/perf lock con -abl sleep 1
> >    contended   total wait     max wait     avg wait            address   symbol
> > 
> >            2      1.95 us      1.77 us       975 ns   ffff9d5e852d3498   &task_struct (mutex)
> >            1      1.18 us      1.18 us      1.18 us   ffff9d5e852d3538   &task_struct (mutex)
> >            4      1.12 us       354 ns       279 ns   ffff9d5e841ca800   &kmalloc-cg-512 (mutex)
> >            2       859 ns       617 ns       429 ns   ffffffffa41c3620   delayed_uprobe_lock (mutex)
> >            3       691 ns       388 ns       230 ns   ffffffffa41c0940   pack_mutex (mutex)
> >            3       421 ns       164 ns       140 ns   ffffffffa3a8b3a0   text_mutex (mutex)
> >            1       409 ns       409 ns       409 ns   ffffffffa41b4cf8   tracepoint_srcu_srcu_usage (mutex)
> >            2       362 ns       239 ns       181 ns   ffffffffa41cf840   pcpu_alloc_mutex (mutex)
> >            1       220 ns       220 ns       220 ns   ffff9d5e82b534d8   &signal_cache (mutex)
> >            1       215 ns       215 ns       215 ns   ffffffffa41b4c28   tracepoint_srcu_srcu_usage (mutex)
> > 
> > Note that the name starts with '&' sign for slab objects to inform they
> > are dynamic locks.  It won't give the accurate lock or type names but
> > it's still useful.  We may add type info to the slab cache later to get
> > the exact name of the lock in the type later.
> > 
> > Signed-off-by: Namhyung Kim <namhyung@...nel.org>
> 
> <snip>
> 
> > diff --git a/tools/perf/util/bpf_skel/lock_contention.bpf.c b/tools/perf/util/bpf_skel/lock_contention.bpf.c
> > index fd24ccb00faec0ba..b5bc37955560a58e 100644
> > --- a/tools/perf/util/bpf_skel/lock_contention.bpf.c
> > +++ b/tools/perf/util/bpf_skel/lock_contention.bpf.c
> > @@ -123,6 +123,8 @@ struct mm_struct___new {
> >  	struct rw_semaphore mmap_lock;
> >  } __attribute__((preserve_access_index));
> >  
> > +extern struct kmem_cache *bpf_get_kmem_cache(u64 addr) __ksym __weak;
> > +
> >  /* control flags */
> >  const volatile int has_cpu;
> >  const volatile int has_task;
> > @@ -496,8 +498,23 @@ int contention_end(u64 *ctx)
> >  		};
> >  		int err;
> >  
> > -		if (aggr_mode == LOCK_AGGR_ADDR)
> > -			first.flags |= check_lock_type(pelem->lock, pelem->flags);
> > +		if (aggr_mode == LOCK_AGGR_ADDR) {
> > +			first.flags |= check_lock_type(pelem->lock,
> > +						       pelem->flags & LCB_F_TYPE_MASK);
> > +
> > +			/* Check if it's from a slab object */
> > +			if (bpf_get_kmem_cache) {
> > +				struct kmem_cache *s;
> > +				struct slab_cache_data *d;
> > +
> > +				s = bpf_get_kmem_cache(pelem->lock);
> > +				if (s != NULL) {
> > +					d = bpf_map_lookup_elem(&slab_caches, &s);
> > +					if (d != NULL)
> > +						first.flags |= d->id;
> > +				}
> 
> Is this being executed as part of obtaining a perf event record, or as part
> of a postprocessing pass? I'm not familiar enough with the code to be certain.
> 
> - if it's part of perf event record, can you just store 's' and defer
> resolving the cache by bpf_map_lookup_elem() to postprocessing?

Namhyung is in vacation this week, so lemme try to help (and learn more
about this patchset since we discussed about it back in LSFMM :-)):

tldr;: He wants to store a 10 bit cookie for the slab cache, to avoid
storing 64 bits per contention record.

My reading of his code:

'first' is a 'struct contention_data' instance, that he will use for
post processing in tools/perf/builtin-lock.c, the relevant part:

	if (use_bpf) {
                lock_contention_start();
                if (argc)
                        evlist__start_workload(con.evlist);

                /* wait for signal */
                pause();

                lock_contention_stop();
                lock_contention_read(&con);
	} else
		process records from a perf.data file with tons
		of lock:lock_contention_{begin,end}, which the use_bpf
		mode above "pre-processes" at begin+end pairs and
                turns into 'struct contention_data' records in a BPF
		map for later post processing in the common part after
		this if/else block.

The post processing is in lock_contention_read(), that is in
tools/perf/util/bpf_lock_contention.c, I stripped out prep steps, etc,
the "meat" is:

        struct contention_data data = {};
	struct lock_stat *st = NULL;
<SNIP>
        while (!bpf_map_get_next_key(fd, prev_key, &key)) {
                s64 ls_key;
                const char *name;

                bpf_map_lookup_elem(fd, &key, &data);

                name = lock_contention_get_name(con, &key, stack_trace, data.flags);
                st = lock_stat_findnew(ls_key, name, data.flags);

That 'lock_stat' struct is then filled up and later, in the common part
to using or not BPF, it gets printed out in the builtin-lock.c main tool
codebase.

The part we're interested here is that lock_contention_get_name(), that
before this patch series returns "(mutex)" and now resolves it to the
slab cache name "&task_struct (mutex)".

key is:

struct contention_key {
        s32 stack_id;
        u32 pid;
        u64 lock_addr_or_cgroup;
};

lock_contention_get_name() tries to resolve the name to the usual
suspects: 
                /* per-process locks set upper bits of the flags */
                if (flags & LCD_F_MMAP_LOCK)
                        return "mmap_lock";
                if (flags & LCD_F_SIGHAND_LOCK)
                        return "siglock";

                /* global locks with symbols */
                sym = machine__find_kernel_symbol(machine, key->lock_addr_or_cgroup, &kmap);
                if (sym)
                        return sym->name;

And then if all of the above (there is another case for rq_lock) it
then gets to look the the ID area of contention_data->flags:

+#define LCB_F_SLAB_ID_SHIFT    16
+#define LCB_F_SLAB_ID_START    (1U << 16)
+#define LCB_F_SLAB_ID_END	(1U << 26)
+#define LCB_F_SLAB_ID_MASK     0x03FF0000U

>>> bin(0x03FF0000)
'0b11111111110000000000000000'
>>>

+               /* look slab_hash for dynamic locks in a slab object */
+               if (hashmap__find(&slab_hash, flags & LCB_F_SLAB_ID_MASK, &slab_data)) {
+                       snprintf(name_buf, sizeof(name_buf), "&%s", slab_data->name);
+                       return name_buf;
+        	}

He wants to avoid storing 64 bytes (the slab cache pointer, 's'), instead
he wants to store a shorter 'id' and encode it in the upper bits of the
'struct contention_data' 'flags' field.

The iterator, at the beggining of the session attributes this id,
starting from zero, to each of the slab caches, so it needs to map it
back from the address at contention_end tracepoint.

At post processing time it converts the id back to the name of the slab
cache.

I hope this helps,

- Arnaldo

> - if it's postprocessing, it would be too late for bpf_get_kmem_cache() as
> the object might be gone already?
> 
> The second alternative would be worse as it could miss the cache or
> misattribute (in case page is reallocated by another cache), the first is
> just less efficient than possible.
> 
> > +			}
> > +		}
> >  
> >  		err = bpf_map_update_elem(&lock_stat, &key, &first, BPF_NOEXIST);
> >  		if (err < 0) {

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ