[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-id: <174777547393.62796.12174520124267417436@noble.neil.brown.name>
Date: Wed, 21 May 2025 07:11:13 +1000
From: "NeilBrown" <neilb@...e.de>
To: "Jeff Layton" <jlayton@...nel.org>
Cc: "syzbot" <syzbot+a15182e1a4094a69cbec@...kaller.appspotmail.com>,
Dai.Ngo@...cle.com, anna@...nel.org, chuck.lever@...cle.com,
davem@...emloft.net, edumazet@...gle.com, horms@...nel.org, kuba@...nel.org,
linux-kernel@...r.kernel.org, linux-nfs@...r.kernel.org,
netdev@...r.kernel.org, okorniev@...hat.com, pabeni@...hat.com,
syzkaller-bugs@...glegroups.com, tom@...pey.com, trondmy@...nel.org
Subject:
Re: [syzbot] [net?] [nfs?] KASAN: slab-out-of-bounds Read in cache_clean
On Wed, 21 May 2025, Jeff Layton wrote:
> On Tue, 2025-05-20 at 00:44 -0700, syzbot wrote:
> > Hello,
> >
> > syzbot found the following issue on:
> >
> > HEAD commit: 088d13246a46 Merge tag 'kbuild-fixes-v6.15' of git://git.k..
> > git tree: upstream
> > console output: https://syzkaller.appspot.com/x/log.txt?x=1047ce70580000
> > kernel config: https://syzkaller.appspot.com/x/.config?x=30181bfb60dbb0a9
> > dashboard link: https://syzkaller.appspot.com/bug?extid=a15182e1a4094a69cbec
> > compiler: gcc (Debian 12.2.0-14) 12.2.0, GNU ld (GNU Binutils for Debian) 2.40
> >
> > Unfortunately, I don't have any reproducer for this issue yet.
> >
> > Downloadable assets:
> > disk image (non-bootable): https://storage.googleapis.com/syzbot-assets/d900f083ada3/non_bootable_disk-088d1324.raw.xz
> > vmlinux: https://storage.googleapis.com/syzbot-assets/273390076ad5/vmlinux-088d1324.xz
> > kernel image: https://storage.googleapis.com/syzbot-assets/3de2f79595a2/bzImage-088d1324.xz
> >
> > IMPORTANT: if you fix the issue, please add the following tag to the commit:
> > Reported-by: syzbot+a15182e1a4094a69cbec@...kaller.appspotmail.com
> >
> > ==================================================================
> > BUG: KASAN: slab-out-of-bounds in hlist_empty include/linux/list.h:972 [inline]
> > BUG: KASAN: slab-out-of-bounds in cache_clean+0x8a0/0x990 net/sunrpc/cache.c:468
> > Read of size 8 at addr ffff88802ae03800 by task kworker/1:2/1335
> >
> > CPU: 1 UID: 0 PID: 1335 Comm: kworker/1:2 Not tainted 6.15.0-rc6-syzkaller-00105-g088d13246a46 #0 PREEMPT(full)
> > Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.3-debian-1.16.3-2~bpo12+1 04/01/2014
> > Workqueue: events_power_efficient do_cache_clean
> > Call Trace:
> > <TASK>
> > __dump_stack lib/dump_stack.c:94 [inline]
> > dump_stack_lvl+0x116/0x1f0 lib/dump_stack.c:120
> > print_address_description mm/kasan/report.c:408 [inline]
> > print_report+0xc3/0x670 mm/kasan/report.c:521
> > kasan_report+0xe0/0x110 mm/kasan/report.c:634
> > hlist_empty include/linux/list.h:972 [inline]
> > cache_clean+0x8a0/0x990 net/sunrpc/cache.c:468
> > do_cache_clean net/sunrpc/cache.c:519 [inline]
> > do_cache_clean+0x29/0xa0 net/sunrpc/cache.c:512
> > process_one_work+0x9cf/0x1b70 kernel/workqueue.c:3238
> > process_scheduled_works kernel/workqueue.c:3319 [inline]
> > worker_thread+0x6c8/0xf10 kernel/workqueue.c:3400
> > kthread+0x3c2/0x780 kernel/kthread.c:464
> > ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:153
> > ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:245
> > </TASK>
> >
>
> This is puzzling:
>
> hlist_empty is just checking whether hlist_head->first is a NULL
> pointer, so this probably means we over or underran the hlist array
> somehow. The array is allocated in cache_create_net, and is of a fixed
> size. AFAICT, the accesses in cache_clean are safe and should never
> overrun the allocation (they're bounded by the same value that we use
> to allocate the array).
I agree.
It would be nice if current_index were unsigned so that the
"current_index < hash_size"
test was more obviously correct, but I cannot see it ever going
negative.
This:
> >
> > The buggy address belongs to the object at ffff88802ae03000
> > which belongs to the cache kmalloc-2k of size 2048
> > The buggy address is located 0 bytes to the right of
> > allocated 2048-byte region [ffff88802ae03000, ffff88802ae03800)
suggests that current_index was 256 and that hash_size was 256
(resulting in a 2k allocation). But that is impossible....
current_index is only every accessed under cache_list_lock so it can't
have changed between the test and the access.
The cache is always unregistered before it is freed so that can't result
in any confusion either.
I can easily imagine the CPU issuing a speculative read to that address,
but I cannot is a non-speculative read (is there a better term) doing
this!
NeilBrown
> >
> > The buggy address belongs to the physical page:
> > page: refcount:0 mapcount:0 mapping:0000000000000000 index:0x0 pfn:0x2ae00
> > head: order:3 mapcount:0 entire_mapcount:0 nr_pages_mapped:0 pincount:0
> > anon flags: 0xfff00000000040(head|node=0|zone=1|lastcpupid=0x7ff)
> > page_type: f5(slab)
> > raw: 00fff00000000040 ffff88801b442f00 0000000000000000 dead000000000001
> > raw: 0000000000000000 0000000000080008 00000000f5000000 0000000000000000
> > head: 00fff00000000040 ffff88801b442f00 0000000000000000 dead000000000001
> > head: 0000000000000000 0000000000080008 00000000f5000000 0000000000000000
> > head: 00fff00000000003 ffffea0000ab8001 00000000ffffffff 00000000ffffffff
> > head: ffffffffffffffff 0000000000000000 00000000ffffffff 0000000000000008
> > page dumped because: kasan: bad access detected
> > page_owner tracks the page as allocated
> > page last allocated via order 3, migratetype Unmovable, gfp_mask 0xd20c0(__GFP_IO|__GFP_FS|__GFP_NOWARN|__GFP_NORETRY|__GFP_COMP|__GFP_NOMEMALLOC), pid 837, tgid 837 (kworker/3:2), ts 32884542472, free_ts 32014442038
> > set_page_owner include/linux/page_owner.h:32 [inline]
> > post_alloc_hook+0x181/0x1b0 mm/page_alloc.c:1718
> > prep_new_page mm/page_alloc.c:1726 [inline]
> > get_page_from_freelist+0x135c/0x3920 mm/page_alloc.c:3688
> > __alloc_frozen_pages_noprof+0x263/0x23a0 mm/page_alloc.c:4970
> > alloc_pages_mpol+0x1fb/0x550 mm/mempolicy.c:2301
> > alloc_slab_page mm/slub.c:2450 [inline]
> > allocate_slab mm/slub.c:2618 [inline]
> > new_slab+0x244/0x340 mm/slub.c:2672
> > ___slab_alloc+0xd9c/0x1940 mm/slub.c:3858
> > __slab_alloc.constprop.0+0x56/0xb0 mm/slub.c:3948
> > __slab_alloc_node mm/slub.c:4023 [inline]
> > slab_alloc_node mm/slub.c:4184 [inline]
> > __do_kmalloc_node mm/slub.c:4326 [inline]
> > __kmalloc_node_track_caller_noprof+0x2ee/0x510 mm/slub.c:4346
> > kmalloc_reserve+0xef/0x2c0 net/core/skbuff.c:599
> > __alloc_skb+0x166/0x380 net/core/skbuff.c:668
> > alloc_skb include/linux/skbuff.h:1340 [inline]
> > mld_newpack.isra.0+0x18e/0xa20 net/ipv6/mcast.c:1788
> > add_grhead+0x299/0x340 net/ipv6/mcast.c:1899
> > add_grec+0x112a/0x1680 net/ipv6/mcast.c:2037
> > mld_send_cr net/ipv6/mcast.c:2163 [inline]
> > mld_ifc_work+0x41f/0xca0 net/ipv6/mcast.c:2702
> > process_one_work+0x9cf/0x1b70 kernel/workqueue.c:3238
> > process_scheduled_works kernel/workqueue.c:3319 [inline]
> > worker_thread+0x6c8/0xf10 kernel/workqueue.c:3400
> > page last free pid 5792 tgid 5792 stack trace:
> > reset_page_owner include/linux/page_owner.h:25 [inline]
> > free_pages_prepare mm/page_alloc.c:1262 [inline]
> > __free_frozen_pages+0x69d/0xff0 mm/page_alloc.c:2725
> > discard_slab mm/slub.c:2716 [inline]
> > __put_partials+0x16d/0x1c0 mm/slub.c:3185
> > qlink_free mm/kasan/quarantine.c:163 [inline]
> > qlist_free_all+0x4e/0x120 mm/kasan/quarantine.c:179
> > kasan_quarantine_reduce+0x195/0x1e0 mm/kasan/quarantine.c:286
> > __kasan_slab_alloc+0x69/0x90 mm/kasan/common.c:329
> > kasan_slab_alloc include/linux/kasan.h:250 [inline]
> > slab_post_alloc_hook mm/slub.c:4147 [inline]
> > slab_alloc_node mm/slub.c:4196 [inline]
> > __do_kmalloc_node mm/slub.c:4326 [inline]
> > __kmalloc_noprof+0x1d4/0x510 mm/slub.c:4339
> > kmalloc_noprof include/linux/slab.h:909 [inline]
> > tomoyo_realpath_from_path+0xc2/0x6e0 security/tomoyo/realpath.c:251
> > tomoyo_get_realpath security/tomoyo/file.c:151 [inline]
> > tomoyo_path_perm+0x274/0x460 security/tomoyo/file.c:822
> > security_inode_getattr+0x116/0x290 security/security.c:2377
> > vfs_getattr fs/stat.c:256 [inline]
> > vfs_fstat+0x4b/0xd0 fs/stat.c:278
> > __do_sys_newfstat+0x91/0x110 fs/stat.c:546
> > do_syscall_x64 arch/x86/entry/syscall_64.c:63 [inline]
> > do_syscall_64+0xcd/0x260 arch/x86/entry/syscall_64.c:94
> > entry_SYSCALL_64_after_hwframe+0x77/0x7f
> >
> > Memory state around the buggy address:
> > ffff88802ae03700: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
> > ffff88802ae03780: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
> > > ffff88802ae03800: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc
> > ^
> > ffff88802ae03880: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc
> > ffff88802ae03900: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc
> > ==================================================================
> >
> >
> > ---
> > This report is generated by a bot. It may contain errors.
> > See https://goo.gl/tpsmEJ for more information about syzbot.
> > syzbot engineers can be reached at syzkaller@...glegroups.com.
> >
> > syzbot will keep track of this issue. See:
> > https://goo.gl/tpsmEJ#status for how to communicate with syzbot.
> >
> > If the report is already addressed, let syzbot know by replying with:
> > #syz fix: exact-commit-title
> >
> > If you want to overwrite report's subsystems, reply with:
> > #syz set subsystems: new-subsystem
> > (See the list of subsystem names on the web dashboard)
> >
> > If the report is a duplicate of another one, reply with:
> > #syz dup: exact-subject-of-another-report
> >
> > If you want to undo deduplication, reply with:
> > #syz undup
>
> --
> Jeff Layton <jlayton@...nel.org>
>
Powered by blists - more mailing lists