lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-Id: <20250924003215.365db154e1fc79163d9d80fe@kernel.org>
Date: Wed, 24 Sep 2025 00:32:15 +0900
From: Masami Hiramatsu (Google) <mhiramat@...nel.org>
To: Feng Yang <yangfeng59949@....com>
Cc: alexei.starovoitov@...il.com, andrii@...nel.org, ast@...nel.org,
 bpf@...r.kernel.org, daniel@...earbox.net, eddyz87@...il.com,
 haoluo@...gle.com, john.fastabend@...il.com, jolsa@...nel.org,
 kpsingh@...nel.org, linux-kernel@...r.kernel.org,
 linux-trace-kernel@...r.kernel.org, martin.lau@...ux.dev, sdf@...ichev.me,
 song@...nel.org, yonghong.song@...ux.dev
Subject: Re: [BUG] Failed to obtain stack trace via bpf_get_stackid on ARM64
 architecture

On Mon, 22 Sep 2025 10:15:31 +0800
Feng Yang <yangfeng59949@....com> wrote:

> On Sun, 21 Sep 2025 22:30:37 +0900 Masami Hiramatsu wrote:
> 
> > On Fri, 19 Sep 2025 19:56:20 -0700
> > Alexei Starovoitov <alexei.starovoitov@...il.com> wrote:
> > 
> > > On Fri, Sep 19, 2025 at 12:19 AM Feng Yang <yangfeng59949@....com> wrote:
> > > >
> > > > When I use bpf_program__attach_kprobe_multi_opts to hook a BPF program that contains the bpf_get_stackid function on the arm64 architecture,
> > > > I find that the stack trace cannot be obtained. The trace->nr in __bpf_get_stackid is 0, and the function returns -EFAULT.
> > > >
> > > > For example:
> > > > diff --git a/tools/testing/selftests/bpf/progs/kprobe_multi.c b/tools/testing/selftests/bpf/progs/kprobe_multi.c
> > > > index 9e1ca8e34913..844fa88cdc4c 100644
> > > > --- a/tools/testing/selftests/bpf/progs/kprobe_multi.c
> > > > +++ b/tools/testing/selftests/bpf/progs/kprobe_multi.c
> > > > @@ -36,6 +36,15 @@ __u64 kretprobe_test6_result = 0;
> > > >  __u64 kretprobe_test7_result = 0;
> > > >  __u64 kretprobe_test8_result = 0;
> > > >
> > > > +typedef __u64 stack_trace_t[2];
> > > > +
> > > > +struct {
> > > > +       __uint(type, BPF_MAP_TYPE_STACK_TRACE);
> > > > +       __uint(max_entries, 1024);
> > > > +       __type(key, __u32);
> > > > +       __type(value, stack_trace_t);
> > > > +} stacks SEC(".maps");
> > > > +
> > > >  static void kprobe_multi_check(void *ctx, bool is_return)
> > > >  {
> > > >         if (bpf_get_current_pid_tgid() >> 32 != pid)
> > > > @@ -100,7 +109,9 @@ int test_kretprobe(struct pt_regs *ctx)
> > > >  SEC("kprobe.multi")
> > > >  int test_kprobe_manual(struct pt_regs *ctx)
> > > >  {
> > > > +       int id = bpf_get_stackid(ctx, &stacks, 0);
> > > 
> > > ftrace_partial_regs() supposed to work on x86 and arm64,
> > > but since multi-kprobe is the only user...
> > 
> > It should be able to unwind stack. It saves sp, pc, lr, fp.
> > 
> > 	regs->sp = afregs->sp;
> > 	regs->pc = afregs->pc;
> > 	regs->regs[29] = afregs->fp;
> > 	regs->regs[30] = afregs->lr;
> > 
> > > I suspect the arm64 implementation wasn't really tested.
> > > Or maybe there is some other issue.
> > 
> > It depends on how bpf_get_stackid() works. Some registers for that
> > function may not be saved.
> > 
> > If it returns -EFAULT, the get_perf_callchain() returns NULL.
> > 
> 
> During my test, the reason for returning -EFAULT was that trace->nr was 0.
> 
> static long __bpf_get_stackid(struct bpf_map *map,
> 			      struct perf_callchain_entry *trace, u64 flags)
> {
> 	struct bpf_stack_map *smap = container_of(map, struct bpf_stack_map, map);
> 	struct stack_map_bucket *bucket, *new_bucket, *old_bucket;
> 	u32 skip = flags & BPF_F_SKIP_FIELD_MASK;
> 	u32 hash, id, trace_nr, trace_len;
> 	bool user = flags & BPF_F_USER_STACK;
> 	u64 *ips;
> 	bool hash_matches;
> 
> 	if (trace->nr <= skip)
> 		/* skipping more than usable stack trace */
> 		return -EFAULT;
> 	......

Hmm. The "trace" is returned from get_perf_callchain()

get_perf_callchain(struct pt_regs *regs, u32 init_nr, bool kernel, bool user,
		   u32 max_stack, bool crosstask, bool add_mark)
{
...

	if (kernel && !user_mode(regs)) {
		if (add_mark)
			perf_callchain_store_context(&ctx, PERF_CONTEXT_KERNEL);
		perf_callchain_kernel(&ctx, regs);
	}

So this means `perf_callchain_kernel(&ctx, regs);` fails to unwind stack.

perf_callchain_kernel() -> arch_stack_walk() -> kunwind_stack_walk()

That is `kunwind_init_from_regs()` and `do_kunwind()`.

	if (regs) {
		if (task != current)
			return -EINVAL;
		kunwind_init_from_regs(&state, regs);
	} else if (task == current) {
		kunwind_init_from_caller(&state);
	} else {
		kunwind_init_from_task(&state, task);
	}

	return do_kunwind(&state, consume_state, cookie);

For initialization, it should be OK because it only refers pc and 
fp(regs[29]), which are recovered by ftrace_partial_regs().

static __always_inline void
kunwind_init_from_regs(struct kunwind_state *state,
		       struct pt_regs *regs)
{
	kunwind_init(state, current);

	state->regs = regs;
	state->common.fp = regs->regs[29];
	state->common.pc = regs->pc;
	state->source = KUNWIND_SOURCE_REGS_PC;
}

And do_kunwind() should work increase trace->nr before return
unless `kunwind_recover_return_address()` fails.

static __always_inline int
do_kunwind(struct kunwind_state *state, kunwind_consume_fn consume_state,
	   void *cookie)
{
	int ret;

	ret = kunwind_recover_return_address(state);
	if (ret)
		return ret;

	while (1) {
		if (!consume_state(state, cookie)) <--- this increases trace->nr (*).
			return -EINVAL;
		ret = kunwind_next(state);
		if (ret == -ENOENT)
			return 0;
		if (ret < 0)
			return ret;
	}
}

(*) consume_state() == arch_kunwind_consume_entry() 
  ->  data->consume_entry == callchain_trace() -> perf_callchain_store().

Hmm, can you also dump the regs and insert pr_info() to find
which function fails?

Thanks,

-- 
Masami Hiramatsu (Google) <mhiramat@...nel.org>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ