[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20230804-abstieg-behilflich-eda2ce9c2c0f@brauner>
Date: Fri, 4 Aug 2023 15:20:45 +0200
From: Christian Brauner <brauner@...nel.org>
To: Christoph Hellwig <hch@....de>
Cc: syzbot <syzbot+2faac0423fdc9692822b@...kaller.appspotmail.com>,
jack@...e.cz, linux-fsdevel@...r.kernel.org,
linux-kernel@...r.kernel.org, syzkaller-bugs@...glegroups.com,
viro@...iv.linux.org.uk
Subject: Re: [syzbot] [fs?] KASAN: slab-use-after-free Read in
test_bdev_super_fc
On Fri, Aug 04, 2023 at 12:14:08PM +0200, Christoph Hellwig wrote:
> FYI, I can reproduce this trivially locally, but even after spending a
> significant time with the trace I'm still puzzled at what is going
> on. I've started trying to make sense of the lockdep report about
> returning to userspace with s_umount held, originall locked in
> get_tree_bdev and am still missing how it could happen.
So in the old scheme:
s = alloc_super()
-> down_write_nested(&s->s_umount, SINGLE_DEPTH_NESTING);
and assume you're not finding an old one immediately afterwards you'd
-> spin_lock(&sb_lock)
static int set_bdev_super(struct super_block *s, void *data)
{
s->s_bdev = data;
s->s_dev = s->s_bdev->bd_dev;
s->s_bdi = bdi_get(s->s_bdev->bd_disk->bdi);
if (bdev_stable_writes(s->s_bdev))
s->s_iflags |= SB_I_STABLE_WRITES;
return 0;
}
-> spin_unlock(&sb_lock)
in the new scheme you're doing:
s = alloc_super()
-> down_write_nested(&s->s_umount, SINGLE_DEPTH_NESTING);
and assume you're not finding an old one immediately afterwards you'd
up_write(&s->s_umount);
error = setup_bdev_super(s, fc->sb_flags, fc);
-> spin_lock(&sb_lock);
sb->s_bdev = bdev;
sb->s_bdi = bdi_get(bdev->bd_disk->bdi);
if (bdev_stable_writes(bdev))
sb->s_iflags |= SB_I_STABLE_WRITES;
-> spin_unlock(&sb_lock);
down_write(&s->s_umount);
Which looks like the lock ordering here is changed?
Powered by blists - more mailing lists