lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Message-ID: <CAO4S-mcOn9_dCGRfBTw6oKPek1yHRyX43ZW+aZBYBucV-Wb2CQ@mail.gmail.com> Date: Mon, 29 Aug 2022 14:55:35 +0800 From: Jiacheng Xu <578001344xu@...il.com> To: linux-kernel@...r.kernel.org, "jhs@...atatu.com" <jhs@...atatu.com>, "xiyou.wangcong@...il.com" <xiyou.wangcong@...il.com>, jiri@...nulli.us Cc: netdev@...r.kernel.org Subject: possible deadlock in sch_direct_xmit Hello, When using modified Syzkaller to fuzz the Linux kernel-5.19, the following crash was triggered. We would appreciate a CVE ID if this is a security issue. HEAD commit: 3d7cb6b04c3f Linux-5.19 git tree: upstream console output: https://drive.google.com/file/d/1ba6S6vbR1GdjkocR9Z14H5fAvrnXnc9T/view?usp=sharing kernel config: https://drive.google.com/file/d/1wgIUDwP5ho29AM-K7HhysSTfWFpfXYkG/view?usp=sharing syz repro: https://drive.google.com/file/d/1bz7Wga1bdlnlUrqfLi26IRGMoAqy6NAf/view?usp=sharing C reproducer: https://drive.google.com/file/d/1Qz_ChFTK3zJEssQ_CfgPGNTdgGHxVFOY/view?usp=sharing Description: When sending socket to device, HARD_TX_LOCK() in sch_direct_xmit() is called to ensure only one CPU can execute it. However, sch_direct_xmit() could be re-entrant through ops->ndo_start_xmit in __netdev_start_xmit(). Such call stack is: ip_send_skb(); HARD_TX_LOCK(); sch_direct_xmit(); __netdev_start_xmit(); ip_tunnel_xmit(); HARD_TX_LOCK(); sch_direct_xmit(); Environment: Ubuntu 20.04 on Linux 5.4.0 QEMU 4.2.1: qemu-system-x86_64 \ -m 2G \ -smp 2 \ -kernel /home/workdir/bzImage \ -append "console=ttyS0 root=/dev/sda earlyprintk=serial net.ifnames=0" \ -drive file=/home/workdir/stretch.img,format=raw \ -net user,host=10.0.2.10,hostfwd=tcp:127.0.0.1:10021-:22 \ -net nic,model=e1000 \ -enable-kvm \ -nographic \ -pidfile vm.pid \ 2>&1 | tee vm.log If you fix this issue, please add the following tag to the commit: Reported-by Jiacheng Xu<578001344xu@...il.com> ============================================ WARNING: possible recursive locking detected 5.19.0 #2 Not tainted -------------------------------------------- syz-executor/7687 is trying to acquire lock: ffff88810b142458 (_xmit_ETHER#2){+.-.}-{2:2}, at: spin_lock include/linux/spinlock.h:349 [inline] ffff88810b142458 (_xmit_ETHER#2){+.-.}-{2:2}, at: __netif_tx_lock include/linux/netdevice.h:4256 [inline] ffff88810b142458 (_xmit_ETHER#2){+.-.}-{2:2}, at: sch_direct_xmit+0x318/0xc70 net/sched/sch_generic.c:340 but task is already holding lock: ffff88801dba0cd8 (_xmit_ETHER#2){+.-.}-{2:2}, at: spin_lock include/linux/spinlock.h:349 [inline] ffff88801dba0cd8 (_xmit_ETHER#2){+.-.}-{2:2}, at: __netif_tx_lock include/linux/netdevice.h:4256 [inline] ffff88801dba0cd8 (_xmit_ETHER#2){+.-.}-{2:2}, at: sch_direct_xmit+0x318/0xc70 net/sched/sch_generic.c:340 other info that might help us debug this: Possible unsafe locking scenario: CPU0 ---- lock(_xmit_ETHER#2); lock(_xmit_ETHER#2); *** DEADLOCK *** May be due to missing lock nesting notation 7 locks held by syz-executor/7687: #0: ffffffff8bd86e00 (rcu_read_lock_bh){....}-{1:2}, at: lwtunnel_xmit_redirect include/net/lwtunnel.h:95 [inline] #0: ffffffff8bd86e00 (rcu_read_lock_bh){....}-{1:2}, at: ip_finish_output2+0x29b/0x2200 net/ipv4/ip_output.c:214 #1: ffffffff8bd86e00 (rcu_read_lock_bh){....}-{1:2}, at: __dev_queue_xmit+0x20a/0x3ad0 net/core/dev.c:4172 #2: ffff88801abac258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: spin_trylock include/linux/spinlock.h:359 [inline] #2: ffff88801abac258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: qdisc_run_begin include/net/sch_generic.h:187 [inline] #2: ffff88801abac258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: qdisc_run_begin include/net/sch_generic.h:184 [inline] #2: ffff88801abac258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: __dev_xmit_skb net/core/dev.c:3804 [inline] #2: ffff88801abac258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: __dev_queue_xmit+0x1384/0x3ad0 net/core/dev.c:4221 #3: ffff88801dba0cd8 (_xmit_ETHER#2){+.-.}-{2:2}, at: spin_lock include/linux/spinlock.h:349 [inline] #3: ffff88801dba0cd8 (_xmit_ETHER#2){+.-.}-{2:2}, at: __netif_tx_lock include/linux/netdevice.h:4256 [inline] #3: ffff88801dba0cd8 (_xmit_ETHER#2){+.-.}-{2:2}, at: sch_direct_xmit+0x318/0xc70 net/sched/sch_generic.c:340 #4: ffffffff8bd86e00 (rcu_read_lock_bh){....}-{1:2}, at: lwtunnel_xmit_redirect include/net/lwtunnel.h:95 [inline] #4: ffffffff8bd86e00 (rcu_read_lock_bh){....}-{1:2}, at: ip_finish_output2+0x29b/0x2200 net/ipv4/ip_output.c:214 #5: ffffffff8bd86e00 (rcu_read_lock_bh){....}-{1:2}, at: __dev_queue_xmit+0x20a/0x3ad0 net/core/dev.c:4172 #6: ffff88810b772258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: spin_trylock include/linux/spinlock.h:359 [inline] #6: ffff88810b772258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: qdisc_run_begin include/net/sch_generic.h:187 [inline] #6: ffff88810b772258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: qdisc_run_begin include/net/sch_generic.h:184 [inline] #6: ffff88810b772258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: __dev_xmit_skb net/core/dev.c:3804 [inline] #6: ffff88810b772258 (dev->qdisc_tx_busylock ?: &qdisc_tx_busylock){+...}-{2:2}, at: __dev_queue_xmit+0x1384/0x3ad0 net/core/dev.c:4221 stack backtrace: CPU: 0 PID: 7687 Comm: syz-executor Not tainted 5.19.0 #2 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.13.0-1ubuntu1.1 04/01/2014 Call Trace: <TASK> __dump_stack lib/dump_stack.c:88 [inline] dump_stack_lvl+0xcd/0x134 lib/dump_stack.c:106 print_deadlock_bug kernel/locking/lockdep.c:2988 [inline] check_deadlock kernel/locking/lockdep.c:3031 [inline] validate_chain kernel/locking/lockdep.c:3816 [inline] __lock_acquire.cold+0x152/0x3ca kernel/locking/lockdep.c:5053 lock_acquire kernel/locking/lockdep.c:5665 [inline] lock_acquire+0x1ab/0x580 kernel/locking/lockdep.c:5630 __raw_spin_lock include/linux/spinlock_api_smp.h:133 [inline] _raw_spin_lock+0x2a/0x40 kernel/locking/spinlock.c:154 spin_lock include/linux/spinlock.h:349 [inline] __netif_tx_lock include/linux/netdevice.h:4256 [inline] sch_direct_xmit+0x318/0xc70 net/sched/sch_generic.c:340 __dev_xmit_skb net/core/dev.c:3817 [inline] __dev_queue_xmit+0x15c6/0x3ad0 net/core/dev.c:4221 dev_queue_xmit include/linux/netdevice.h:2994 [inline] neigh_resolve_output net/core/neighbour.c:1528 [inline] neigh_resolve_output+0x53d/0x870 net/core/neighbour.c:1508 neigh_output include/net/neighbour.h:549 [inline] ip_finish_output2+0x7ab/0x2200 net/ipv4/ip_output.c:228 __ip_finish_output net/ipv4/ip_output.c:306 [inline] __ip_finish_output+0x85c/0x1450 net/ipv4/ip_output.c:288 ip_finish_output+0x32/0x280 net/ipv4/ip_output.c:316 NF_HOOK_COND include/linux/netfilter.h:296 [inline] ip_output+0x20a/0x620 net/ipv4/ip_output.c:430 dst_output include/net/dst.h:451 [inline] ip_local_out+0xaf/0x1a0 net/ipv4/ip_output.c:126 iptunnel_xmit+0x693/0xa80 net/ipv4/ip_tunnel_core.c:82 ip_tunnel_xmit+0xff9/0x2c40 net/ipv4/ip_tunnel.c:811 erspan_xmit+0x521/0x2b00 net/ipv4/ip_gre.c:715 __netdev_start_xmit include/linux/netdevice.h:4805 [inline] netdev_start_xmit include/linux/netdevice.h:4819 [inline] xmit_one net/core/dev.c:3590 [inline] dev_hard_start_xmit+0x19d/0x8b0 net/core/dev.c:3606 sch_direct_xmit+0x19f/0xc70 net/sched/sch_generic.c:342 __dev_xmit_skb net/core/dev.c:3817 [inline] __dev_queue_xmit+0x15c6/0x3ad0 net/core/dev.c:4221 dev_queue_xmit include/linux/netdevice.h:2994 [inline] neigh_resolve_output net/core/neighbour.c:1528 [inline] neigh_resolve_output+0x53d/0x870 net/core/neighbour.c:1508 neigh_output include/net/neighbour.h:549 [inline] ip_finish_output2+0x7ab/0x2200 net/ipv4/ip_output.c:228 __ip_finish_output net/ipv4/ip_output.c:306 [inline] __ip_finish_output+0x85c/0x1450 net/ipv4/ip_output.c:288 ip_finish_output+0x32/0x280 net/ipv4/ip_output.c:316 NF_HOOK_COND include/linux/netfilter.h:296 [inline] ip_output+0x20a/0x620 net/ipv4/ip_output.c:430 dst_output include/net/dst.h:451 [inline] ip_local_out+0xaf/0x1a0 net/ipv4/ip_output.c:126 ip_send_skb+0x3e/0xe0 net/ipv4/ip_output.c:1571 udp_send_skb.isra.0+0x731/0x1460 net/ipv4/udp.c:967 udp_sendmsg+0x1dc3/0x2830 net/ipv4/udp.c:1254 udpv6_sendmsg+0x1762/0x2bd0 net/ipv6/udp.c:1365 inet6_sendmsg+0x99/0xe0 net/ipv6/af_inet6.c:652 sock_sendmsg_nosec net/socket.c:714 [inline] sock_sendmsg net/socket.c:734 [inline] sock_sendmsg+0xc3/0x120 net/socket.c:729 ____sys_sendmsg+0x334/0x810 net/socket.c:2488 ___sys_sendmsg+0x100/0x170 net/socket.c:2542 __sys_sendmmsg+0x195/0x470 net/socket.c:2628 __do_sys_sendmmsg net/socket.c:2657 [inline] __se_sys_sendmmsg net/socket.c:2654 [inline] __x64_sys_sendmmsg+0x99/0x100 net/socket.c:2654 do_syscall_x64 arch/x86/entry/common.c:50 [inline] do_syscall_64+0x35/0xb0 arch/x86/entry/common.c:80 entry_SYSCALL_64_after_hwframe+0x63/0xcd RIP: 0033:0x7f545b495dfd Code: 02 b8 ff ff ff ff c3 66 0f 1f 44 00 00 f3 0f 1e fa 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 c7 c1 bc ff ff ff f7 d8 64 89 01 48 RSP: 002b:00007f545c6afc58 EFLAGS: 00000246 ORIG_RAX: 0000000000000133 RAX: ffffffffffffffda RBX: 00007f545b5bc0a0 RCX: 00007f545b495dfd RDX: 0000000000000001 RSI: 0000000020000140 RDI: 0000000000000004 RBP: 00007f545b4ff4c1 R08: 0000000000000000 R09: 0000000000000000 R10: 0000000004000000 R11: 0000000000000246 R12: 00007f545b5bc0a0 R13: 00007fffca7bf54f R14: 00007fffca7bf6f0 R15: 00007f545c6afdc0 </TASK>
Powered by blists - more mailing lists