[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20240711103341.18a4eb4e@gandalf.local.home>
Date: Thu, 11 Jul 2024 10:33:41 -0400
From: Steven Rostedt <rostedt@...dmis.org>
To: Jesper Dangaard Brouer <hawk@...nel.org>
Cc: syzbot <syzbot+16b6ab88e66b34d09014@...kaller.appspotmail.com>,
akpm@...ux-foundation.org, cgroups@...r.kernel.org, hannes@...xchg.org,
linux-kernel@...r.kernel.org, linux-mm@...ck.org,
linux-trace-kernel@...r.kernel.org, lizefan.x@...edance.com,
mathieu.desnoyers@...icios.com, mhiramat@...nel.org,
netdev@...r.kernel.org, syzkaller-bugs@...glegroups.com, tj@...nel.org
Subject: Re: [syzbot] [mm?] possible deadlock in
__mmap_lock_do_trace_released
On Thu, 4 Jul 2024 22:12:45 +0200
Jesper Dangaard Brouer <hawk@...nel.org> wrote:
> > ============================================
> > WARNING: possible recursive locking detected
> > 6.10.0-rc2-syzkaller-00797-ga12978712d90 #0 Not tainted
> > --------------------------------------------
> > syz-executor646/5097 is trying to acquire lock:
> > ffff8880b94387e8 (lock#9){+.+.}-{2:2}, at: local_lock_acquire include/linux/local_lock_internal.h:29 [inline]
> > ffff8880b94387e8 (lock#9){+.+.}-{2:2}, at: __mmap_lock_do_trace_released+0x83/0x620 mm/mmap_lock.c:243
> >
> > but task is already holding lock:
> > ffff8880b94387e8 (lock#9){+.+.}-{2:2}, at: local_lock_acquire include/linux/local_lock_internal.h:29 [inline]
> > ffff8880b94387e8 (lock#9){+.+.}-{2:2}, at: __mmap_lock_do_trace_released+0x83/0x620 mm/mmap_lock.c:243
> >
> > other info that might help us debug this:
> > Possible unsafe locking scenario:
> >
> > CPU0
> > ----
> > lock(lock#9);
> > lock(lock#9);
> >
> > *** DEADLOCK ***
Looks like it's trying to take the rwsem mm->mmap_lock recursively. And
rwsems are *not* allowed to be recursively taken, as once there's a writer,
all new acquires of the reader will block. Then you can have:
CPU0 CPU1
---- ----
down_read(lockA);
down_write(lockA); // blocks
down_read(lockA); //blocks
DEADLOCK!
> >
> > May be due to missing lock nesting notation
> >
>
> To me, this looks like a lockdep false-positive, but I might be wrong.
>
> Could someone with more LOCKDEP knowledge give their interpretation?
>
> The commit[1] adds a fairly standard trylock scheme.
> Do I need to lockdep annotate trylock's in some special way?
>
> [1] https://git.kernel.org/torvalds/c/21c38a3bd4ee3fb733
>
> Also notice change uses raw_spin_lock, which might be harder for lockdep?
> So, I also enabled CONFIG_PROVE_RAW_LOCK_NESTING in my testlab to help
> with this, and CONFIG_PROVE_LOCKING.
> (And obviously I also enabled LOCKDEP*)
>
> --Jesper
>
> > 5 locks held by syz-executor646/5097:
> > #0: ffff8880182eb118 (&mm->mmap_lock){++++}-{3:3}, at: mmap_read_lock include/linux/mmap_lock.h:144 [inline]
> > #0: ffff8880182eb118 (&mm->mmap_lock){++++}-{3:3}, at: acct_collect+0x1cf/0x830 kernel/acct.c:563
> > #1: ffff8880b94387e8 (lock#9){+.+.}-{2:2}, at: local_lock_acquire include/linux/local_lock_internal.h:29 [inline]
> > #1: ffff8880b94387e8 (lock#9){+.+.}-{2:2}, at: __mmap_lock_do_trace_released+0x83/0x620 mm/mmap_lock.c:243
> > #2: ffffffff8e333fa0 (rcu_read_lock){....}-{1:2}, at: rcu_lock_acquire include/linux/rcupdate.h:329 [inline]
> > #2: ffffffff8e333fa0 (rcu_read_lock){....}-{1:2}, at: rcu_read_lock include/linux/rcupdate.h:781 [inline]
> > #2: ffffffff8e333fa0 (rcu_read_lock){....}-{1:2}, at: get_memcg_path_buf mm/mmap_lock.c:139 [inline]
> > #2: ffffffff8e333fa0 (rcu_read_lock){....}-{1:2}, at: get_mm_memcg_path+0xb1/0x600 mm/mmap_lock.c:209
> > #3: ffffffff8e333fa0 (rcu_read_lock){....}-{1:2}, at: trace_call_bpf+0xbc/0x8a0
> > #4: ffff8880182eb118 (&mm->mmap_lock){++++}-{3:3}, at: mmap_read_trylock include/linux/mmap_lock.h:163 [inline]
> > #4: ffff8880182eb118 (&mm->mmap_lock){++++}-{3:3}, at: stack_map_get_build_id_offset+0x237/0x9d0 kernel/bpf/stackmap.c:141
> >
> > stack backtrace:
> > CPU: 0 PID: 5097 Comm: syz-executor646 Not tainted 6.10.0-rc2-syzkaller-00797-ga12978712d90 #0
> > Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 06/07/2024
> > Call Trace:
> > <TASK>
> > __dump_stack lib/dump_stack.c:88 [inline]
> > dump_stack_lvl+0x241/0x360 lib/dump_stack.c:114
> > check_deadlock kernel/locking/lockdep.c:3062 [inline]
> > validate_chain+0x15d3/0x5900 kernel/locking/lockdep.c:3856
> > __lock_acquire+0x1346/0x1fd0 kernel/locking/lockdep.c:5137
> > lock_acquire+0x1ed/0x550 kernel/locking/lockdep.c:5754
> > local_lock_acquire include/linux/local_lock_internal.h:29 [inline]
> > __mmap_lock_do_trace_released+0x9c/0x620 mm/mmap_lock.c:243
Here we have:
static inline void mmap_read_lock(struct mm_struct *mm)
{
__mmap_lock_trace_start_locking(mm, false);
down_read(&mm->mmap_lock);
__mmap_lock_trace_acquire_returned(mm, false, true);
}
Which is taking the mm->mmap_lock for read.
> > __mmap_lock_trace_released include/linux/mmap_lock.h:42 [inline]
> > mmap_read_unlock include/linux/mmap_lock.h:170 [inline]
> > bpf_mmap_unlock_mm kernel/bpf/mmap_unlock_work.h:52 [inline]
> > stack_map_get_build_id_offset+0x9c7/0x9d0 kernel/bpf/stackmap.c:173
> > __bpf_get_stack+0x4ad/0x5a0 kernel/bpf/stackmap.c:449
> > bpf_prog_e6cf5f9c69743609+0x42/0x46
> > bpf_dispatcher_nop_func include/linux/bpf.h:1243 [inline]
> > __bpf_prog_run include/linux/filter.h:691 [inline]
> > bpf_prog_run include/linux/filter.h:698 [inline]
> > bpf_prog_run_array include/linux/bpf.h:2104 [inline]
> > trace_call_bpf+0x369/0x8a0 kernel/trace/bpf_trace.c:147
> > perf_trace_run_bpf_submit+0x7c/0x1d0 kernel/events/core.c:10269
> > perf_trace_mmap_lock+0x3d7/0x510 include/trace/events/mmap_lock.h:16
I'm guessing a bpf program attached to something within the same code:
> > trace_mmap_lock_released include/trace/events/mmap_lock.h:50 [inline]
> > __mmap_lock_do_trace_released+0x5bb/0x620 mm/mmap_lock.c:243
Here is the same function as above where it took the mm->mmap_lock.
My guess is the bpf program that attached to this event ends up calling the
same function and it tries to take the rwsem again, and that poses a risk
for deadlock.
-- Steve
> > __mmap_lock_trace_released include/linux/mmap_lock.h:42 [inline]
> > mmap_read_unlock include/linux/mmap_lock.h:170 [inline]
> > acct_collect+0x81d/0x830 kernel/acct.c:566
> > do_exit+0x936/0x27e0 kernel/exit.c:853
> > do_group_exit+0x207/0x2c0 kernel/exit.c:1023
> > __do_sys_exit_group kernel/exit.c:1034 [inline]
> > __se_sys_exit_group kernel/exit.c:1032 [inline]
> > __x64_sys_exit_group+0x3f/0x40 kernel/exit.c:1032
> > do_syscall_x64 arch/x86/entry/common.c:52 [inline]
> > do_syscall_64+0xf3/0x230 arch/x86/entry/common.c:83
> > entry_SYSCALL_64_after_hwframe+0x77/0x7f
> > RIP: 0033:0x7f8fac26d039
> > Code: 90 49 c7 c0 b8 ff ff ff be e7 00 00 00 ba 3c 00 00 00 eb 12 0f 1f
> > 44 00 00 89 d0 0f 05 48 3d 00 f0 ff ff 77 1c f4 89 f0 0f 05 <48> 3d 00
> > f0 ff ff 76 e7 f7 d8 64 41 89 00 eb df 0f 1f 80 00 00 00 RSP:
> > 002b:00007ffd95d56e68 EFLAGS: 00000246 ORIG_RAX: 00000000000000e7 RAX:
> > ffffffffffffffda RBX: 0000000000000000 RCX: 00007f8fac26d039 RDX:
> > 000000000000003c RSI: 00000000000000e7 RDI: 0000000000000000 RBP:
> > 00007f8fac2e82b0 R08: ffffffffffffffb8 R09: 00000000000000a0 R10:
> > 0000000000000000 R11: 0000000000000246 R12: 00007f8fac2e82b0 R13:
> > 0000000000000000 R14: 00007f8fac2e8d20 R15: 00007f8fac23e1e0 </TASK>
> >
> >
> > ---
> > This report is generated by a bot. It may contain errors.
> > See https://goo.gl/tpsmEJ for more information about syzbot.
> > syzbot engineers can be reached at syzkaller@...glegroups.com.
> >
> > syzbot will keep track of this issue. See:
> > https://goo.gl/tpsmEJ#status for how to communicate with syzbot.
> > For information about bisection process
> > see: https://goo.gl/tpsmEJ#bisection
> >
> > If the report is already addressed, let syzbot know by replying with:
> > #syz fix: exact-commit-title
> >
> > If you want syzbot to run the reproducer, reply with:
> > #syz test: git://repo/address.git branch-or-commit-hash
> > If you attach or paste a git patch, syzbot will apply it before testing.
> >
> > If you want to overwrite report's subsystems, reply with:
> > #syz set subsystems: new-subsystem
> > (See the list of subsystem names on the web dashboard)
> >
> > If the report is a duplicate of another one, reply with:
> > #syz dup: exact-subject-of-another-report
> >
> > If you want to undo deduplication, reply with:
> > #syz undup
Powered by blists - more mailing lists