lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAHbLzkoTC3WoNa-dLBBmi15oBRXNbJzJuS6-GPr8XPWiHrgO4A@mail.gmail.com>
Date:   Mon, 13 Apr 2020 16:19:46 -0700
From:   Yang Shi <shy828301@...il.com>
To:     syzbot <syzbot+e27980339d305f2dbfd9@...kaller.appspotmail.com>
Cc:     Andrew Morton <akpm@...ux-foundation.org>,
        Hugh Dickins <hughd@...gle.com>,
        Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
        Linux MM <linux-mm@...ck.org>, syzkaller-bugs@...glegroups.com
Subject: Re: possible deadlock in shmem_mfill_atomic_pte

On Tue, Mar 31, 2020 at 10:21 AM syzbot
<syzbot+e27980339d305f2dbfd9@...kaller.appspotmail.com> wrote:
>
> Hello,
>
> syzbot found the following crash on:
>
> HEAD commit:    527630fb Merge tag 'clk-fixes-for-linus' of git://git.kern..
> git tree:       upstream
> console output: https://syzkaller.appspot.com/x/log.txt?x=1214875be00000
> kernel config:  https://syzkaller.appspot.com/x/.config?x=27392dd2975fd692
> dashboard link: https://syzkaller.appspot.com/bug?extid=e27980339d305f2dbfd9
> compiler:       gcc (GCC) 9.0.0 20181231 (experimental)
>
> Unfortunately, I don't have any reproducer for this crash yet.
>
> IMPORTANT: if you fix the bug, please add the following tag to the commit:
> Reported-by: syzbot+e27980339d305f2dbfd9@...kaller.appspotmail.com
>
> WARNING: possible irq lock inversion dependency detected
> 5.6.0-rc7-syzkaller #0 Not tainted
> --------------------------------------------------------
> syz-executor.0/10317 just changed the state of lock:
> ffff888021d16568 (&(&info->lock)->rlock){+.+.}, at: spin_lock include/linux/spinlock.h:338 [inline]
> ffff888021d16568 (&(&info->lock)->rlock){+.+.}, at: shmem_mfill_atomic_pte+0x1012/0x21c0 mm/shmem.c:2407
> but this lock was taken by another, SOFTIRQ-safe lock in the past:
>  (&(&xa->xa_lock)->rlock#5){..-.}
>
>
> and interrupts could create inverse lock ordering between them.
>
>
> other info that might help us debug this:
>  Possible interrupt unsafe locking scenario:
>
>        CPU0                    CPU1
>        ----                    ----
>   lock(&(&info->lock)->rlock);
>                                local_irq_disable();
>                                lock(&(&xa->xa_lock)->rlock#5);
>                                lock(&(&info->lock)->rlock);
>   <Interrupt>
>     lock(&(&xa->xa_lock)->rlock#5);
>
>  *** DEADLOCK ***

This looks possible. shmem_mfill_atomic_pte() acquires info->lock with
irq enabled.

The below patch should be able to fix it:

diff --git a/mm/shmem.c b/mm/shmem.c
index d722eb8..762da6a 100644
--- a/mm/shmem.c
+++ b/mm/shmem.c
@@ -2399,11 +2399,11 @@ static int shmem_mfill_atomic_pte(struct
mm_struct *dst_mm,

        lru_cache_add_anon(page);

-       spin_lock(&info->lock);
+       spin_lock_irq(&info->lock);
        info->alloced++;
        inode->i_blocks += BLOCKS_PER_PAGE;
        shmem_recalc_inode(inode);
-       spin_unlock(&info->lock);
+       spin_unlock_irq(&info->lock);

        inc_mm_counter(dst_mm, mm_counter_file(page));
        page_add_file_rmap(page, false);


>
> 2 locks held by syz-executor.0/10317:
>  #0: ffff888011721898 (&mm->mmap_sem#2){++++}, at: __mcopy_atomic mm/userfaultfd.c:474 [inline]
>  #0: ffff888011721898 (&mm->mmap_sem#2){++++}, at: mcopy_atomic+0x185/0x2510 mm/userfaultfd.c:607
>  #1: ffff888024eda280 (&(ptlock_ptr(page))->rlock#2){+.+.}, at: spin_lock include/linux/spinlock.h:338 [inline]
>  #1: ffff888024eda280 (&(ptlock_ptr(page))->rlock#2){+.+.}, at: shmem_mfill_atomic_pte+0xf76/0x21c0 mm/shmem.c:2394
>
> the shortest dependencies between 2nd lock and 1st lock:
>  -> (&(&xa->xa_lock)->rlock#5){..-.} {
>     IN-SOFTIRQ-W at:
>                       lock_acquire+0x197/0x420 kernel/locking/lockdep.c:4484
>                       __raw_spin_lock_irqsave include/linux/spinlock_api_smp.h:110 [inline]
>                       _raw_spin_lock_irqsave+0x8c/0xbf kernel/locking/spinlock.c:159
>                       test_clear_page_writeback+0x1d7/0x11e0 mm/page-writeback.c:2728
>                       end_page_writeback+0x239/0x520 mm/filemap.c:1317
>                       end_buffer_async_write+0x6a9/0xa30 fs/buffer.c:389
>                       end_bio_bh_io_sync+0xe2/0x140 fs/buffer.c:3018
>                       bio_endio+0x473/0x820 block/bio.c:1872
>                       req_bio_endio block/blk-core.c:245 [inline]
>                       blk_update_request+0x3e1/0xdc0 block/blk-core.c:1468
>                       scsi_end_request+0x80/0x7a0 drivers/scsi/scsi_lib.c:576
>                       scsi_io_completion+0x1e7/0x1300 drivers/scsi/scsi_lib.c:960
>                       scsi_softirq_done+0x31e/0x3b0 drivers/scsi/scsi_lib.c:1476
>                       blk_done_softirq+0x2db/0x440 block/blk-softirq.c:37
>                       __do_softirq+0x26c/0x99d kernel/softirq.c:292
>                       invoke_softirq kernel/softirq.c:373 [inline]
>                       irq_exit+0x192/0x1d0 kernel/softirq.c:413
>                       exiting_irq arch/x86/include/asm/apic.h:546 [inline]
>                       do_IRQ+0xde/0x280 arch/x86/kernel/irq.c:263
>                       ret_from_intr+0x0/0x36
>                       clear_page_erms+0x7/0x10 arch/x86/lib/clear_page_64.S:48
>                       clear_page arch/x86/include/asm/page_64.h:49 [inline]
>                       clear_highpage include/linux/highmem.h:214 [inline]
>                       kernel_init_free_pages+0x92/0x120 mm/page_alloc.c:1118
>                       prep_new_page+0x12e/0x1f0 mm/page_alloc.c:2160
>                       get_page_from_freelist+0x14c7/0x3ee0 mm/page_alloc.c:3684
>                       __alloc_pages_nodemask+0x2a5/0x820 mm/page_alloc.c:4731
>                       alloc_pages_current+0xff/0x200 mm/mempolicy.c:2211
>                       alloc_pages include/linux/gfp.h:532 [inline]
>                       __page_cache_alloc+0x298/0x480 mm/filemap.c:959
>                       __do_page_cache_readahead+0x1a7/0x570 mm/readahead.c:196
>                       ra_submit mm/internal.h:62 [inline]
>                       ondemand_readahead+0x566/0xd60 mm/readahead.c:492
>                       page_cache_async_readahead mm/readahead.c:574 [inline]
>                       page_cache_async_readahead+0x43d/0x7c0 mm/readahead.c:547
>                       generic_file_buffered_read mm/filemap.c:2037 [inline]
>                       generic_file_read_iter+0x124a/0x2b00 mm/filemap.c:2302
>                       ext4_file_read_iter fs/ext4/file.c:131 [inline]
>                       ext4_file_read_iter+0x1d1/0x600 fs/ext4/file.c:114
>                       call_read_iter include/linux/fs.h:1896 [inline]
>                       new_sync_read+0x4a2/0x790 fs/read_write.c:414
>                       __vfs_read+0xc9/0x100 fs/read_write.c:427
>                       integrity_kernel_read+0x143/0x200 security/integrity/iint.c:200
>                       ima_calc_file_hash_tfm+0x2aa/0x3b0 security/integrity/ima/ima_crypto.c:360
>                       ima_calc_file_shash security/integrity/ima/ima_crypto.c:391 [inline]
>                       ima_calc_file_hash+0x199/0x540 security/integrity/ima/ima_crypto.c:456
>                       ima_collect_measurement+0x4c4/0x570 security/integrity/ima/ima_api.c:249
>                       process_measurement+0xc6d/0x1740 security/integrity/ima/ima_main.c:326
>                       ima_bprm_check+0xde/0x210 security/integrity/ima/ima_main.c:417
>                       security_bprm_check+0x89/0xb0 security/security.c:819
>                       search_binary_handler+0x70/0x580 fs/exec.c:1649
>                       exec_binprm fs/exec.c:1705 [inline]
>                       __do_execve_file.isra.0+0x12fc/0x2270 fs/exec.c:1825
>                       do_execveat_common fs/exec.c:1871 [inline]
>                       do_execve fs/exec.c:1888 [inline]
>                       __do_sys_execve fs/exec.c:1964 [inline]
>                       __se_sys_execve fs/exec.c:1959 [inline]
>                       __x64_sys_execve+0x8a/0xb0 fs/exec.c:1959
>                       do_syscall_64+0xf6/0x7d0 arch/x86/entry/common.c:294
>                       entry_SYSCALL_64_after_hwframe+0x49/0xbe
>     INITIAL USE at:
>                      lock_acquire+0x197/0x420 kernel/locking/lockdep.c:4484
>                      __raw_spin_lock_irq include/linux/spinlock_api_smp.h:128 [inline]
>                      _raw_spin_lock_irq+0x5b/0x80 kernel/locking/spinlock.c:167
>                      spin_lock_irq include/linux/spinlock.h:363 [inline]
>                      clear_inode+0x1b/0x1e0 fs/inode.c:529
>                      shmem_evict_inode+0x1db/0x9f0 mm/shmem.c:1116
>                      evict+0x2ed/0x650 fs/inode.c:576
>                      iput_final fs/inode.c:1572 [inline]
>                      iput+0x536/0x8c0 fs/inode.c:1598
>                      dentry_unlink_inode+0x2c0/0x3e0 fs/dcache.c:374
>                      d_delete fs/dcache.c:2451 [inline]
>                      d_delete+0x117/0x150 fs/dcache.c:2440
>                      vfs_unlink+0x4d5/0x620 fs/namei.c:4087
>                      handle_remove+0x417/0x720 drivers/base/devtmpfs.c:332
>                      handle drivers/base/devtmpfs.c:378 [inline]
>                      devtmpfsd.part.0+0x302/0x750 drivers/base/devtmpfs.c:413
>                      devtmpfsd+0x107/0x120 drivers/base/devtmpfs.c:403
>                      kthread+0x357/0x430 kernel/kthread.c:255
>                      ret_from_fork+0x24/0x30 arch/x86/entry/entry_64.S:352
>   }
>   ... key      at: [<ffffffff8c402aa0>] __key.17910+0x0/0x40
>   ... acquired at:
>    __raw_spin_lock_irqsave include/linux/spinlock_api_smp.h:110 [inline]
>    _raw_spin_lock_irqsave+0x8c/0xbf kernel/locking/spinlock.c:159
>    shmem_uncharge+0x24/0x270 mm/shmem.c:341
>    __split_huge_page mm/huge_memory.c:2540 [inline]
>    split_huge_page_to_list+0x2751/0x33c0 mm/huge_memory.c:2813
>    split_huge_page include/linux/huge_mm.h:169 [inline]
>    shmem_unused_huge_shrink+0x7ba/0x13a0 mm/shmem.c:542
>    shmem_unused_huge_scan+0x7a/0xb0 mm/shmem.c:574
>    super_cache_scan+0x34f/0x480 fs/super.c:111
>    do_shrink_slab+0x3fc/0xab0 mm/vmscan.c:512
>    shrink_slab mm/vmscan.c:673 [inline]
>    shrink_slab+0x16f/0x5f0 mm/vmscan.c:646
>    shrink_node_memcgs mm/vmscan.c:2676 [inline]
>    shrink_node+0x477/0x1b20 mm/vmscan.c:2780
>    shrink_zones mm/vmscan.c:2983 [inline]
>    do_try_to_free_pages+0x38d/0x13a0 mm/vmscan.c:3036
>    try_to_free_pages+0x293/0x8d0 mm/vmscan.c:3275
>    __perform_reclaim mm/page_alloc.c:4113 [inline]
>    __alloc_pages_direct_reclaim mm/page_alloc.c:4134 [inline]
>    __alloc_pages_slowpath+0x919/0x26a0 mm/page_alloc.c:4537
>    __alloc_pages_nodemask+0x5e1/0x820 mm/page_alloc.c:4751
>    __alloc_pages include/linux/gfp.h:496 [inline]
>    __alloc_pages_node include/linux/gfp.h:509 [inline]
>    alloc_pages_vma+0x3bd/0x600 mm/mempolicy.c:2155
>    shmem_alloc_hugepage+0x122/0x210 mm/shmem.c:1484
>    shmem_alloc_and_acct_page+0x3ba/0x980 mm/shmem.c:1522
>    shmem_getpage_gfp+0xdb9/0x2860 mm/shmem.c:1835
>    shmem_getpage mm/shmem.c:154 [inline]
>    shmem_write_begin+0x102/0x1e0 mm/shmem.c:2488
>    generic_perform_write+0x20a/0x4e0 mm/filemap.c:3287
>    __generic_file_write_iter+0x24c/0x610 mm/filemap.c:3416
>    generic_file_write_iter+0x3f0/0x62d mm/filemap.c:3448
>    call_write_iter include/linux/fs.h:1902 [inline]
>    new_sync_write+0x49c/0x700 fs/read_write.c:483
>    __vfs_write+0xc9/0x100 fs/read_write.c:496
>    vfs_write+0x262/0x5c0 fs/read_write.c:558
>    ksys_write+0x127/0x250 fs/read_write.c:611
>    do_syscall_64+0xf6/0x7d0 arch/x86/entry/common.c:294
>    entry_SYSCALL_64_after_hwframe+0x49/0xbe
>
> -> (&(&info->lock)->rlock){+.+.} {
>    HARDIRQ-ON-W at:
>                     lock_acquire+0x197/0x420 kernel/locking/lockdep.c:4484
>                     __raw_spin_lock include/linux/spinlock_api_smp.h:142 [inline]
>                     _raw_spin_lock+0x2a/0x40 kernel/locking/spinlock.c:151
>                     spin_lock include/linux/spinlock.h:338 [inline]
>                     shmem_mfill_atomic_pte+0x1012/0x21c0 mm/shmem.c:2407
>                     shmem_mcopy_atomic_pte+0x3a/0x50 mm/shmem.c:2445
>                     mfill_atomic_pte mm/userfaultfd.c:434 [inline]
>                     __mcopy_atomic mm/userfaultfd.c:557 [inline]
>                     mcopy_atomic+0xac7/0x2510 mm/userfaultfd.c:607
>                     userfaultfd_copy fs/userfaultfd.c:1736 [inline]
>                     userfaultfd_ioctl+0x4d2/0x3b10 fs/userfaultfd.c:1886
>                     vfs_ioctl fs/ioctl.c:47 [inline]
>                     ksys_ioctl+0x11a/0x180 fs/ioctl.c:763
>                     __do_sys_ioctl fs/ioctl.c:772 [inline]
>                     __se_sys_ioctl fs/ioctl.c:770 [inline]
>                     __x64_sys_ioctl+0x6f/0xb0 fs/ioctl.c:770
>                     do_syscall_64+0xf6/0x7d0 arch/x86/entry/common.c:294
>                     entry_SYSCALL_64_after_hwframe+0x49/0xbe
>    SOFTIRQ-ON-W at:
>                     lock_acquire+0x197/0x420 kernel/locking/lockdep.c:4484
>                     __raw_spin_lock include/linux/spinlock_api_smp.h:142 [inline]
>                     _raw_spin_lock+0x2a/0x40 kernel/locking/spinlock.c:151
>                     spin_lock include/linux/spinlock.h:338 [inline]
>                     shmem_mfill_atomic_pte+0x1012/0x21c0 mm/shmem.c:2407
>                     shmem_mcopy_atomic_pte+0x3a/0x50 mm/shmem.c:2445
>                     mfill_atomic_pte mm/userfaultfd.c:434 [inline]
>                     __mcopy_atomic mm/userfaultfd.c:557 [inline]
>                     mcopy_atomic+0xac7/0x2510 mm/userfaultfd.c:607
>                     userfaultfd_copy fs/userfaultfd.c:1736 [inline]
>                     userfaultfd_ioctl+0x4d2/0x3b10 fs/userfaultfd.c:1886
>                     vfs_ioctl fs/ioctl.c:47 [inline]
>                     ksys_ioctl+0x11a/0x180 fs/ioctl.c:763
>                     __do_sys_ioctl fs/ioctl.c:772 [inline]
>                     __se_sys_ioctl fs/ioctl.c:770 [inline]
>                     __x64_sys_ioctl+0x6f/0xb0 fs/ioctl.c:770
>                     do_syscall_64+0xf6/0x7d0 arch/x86/entry/common.c:294
>                     entry_SYSCALL_64_after_hwframe+0x49/0xbe
>    INITIAL USE at:
>                    lock_acquire+0x197/0x420 kernel/locking/lockdep.c:4484
>                    __raw_spin_lock_irq include/linux/spinlock_api_smp.h:128 [inline]
>                    _raw_spin_lock_irq+0x5b/0x80 kernel/locking/spinlock.c:167
>                    spin_lock_irq include/linux/spinlock.h:363 [inline]
>                    shmem_getpage_gfp+0xf10/0x2860 mm/shmem.c:1887
>                    shmem_read_mapping_page_gfp+0xd3/0x170 mm/shmem.c:4218
>                    shmem_read_mapping_page include/linux/shmem_fs.h:101 [inline]
>                    drm_gem_get_pages+0x293/0x530 drivers/gpu/drm/drm_gem.c:578
>                    drm_gem_shmem_get_pages_locked drivers/gpu/drm/drm_gem_shmem_helper.c:146 [inline]
>                    drm_gem_shmem_get_pages+0x9d/0x160 drivers/gpu/drm/drm_gem_shmem_helper.c:175
>                    virtio_gpu_object_attach+0x121/0x950 drivers/gpu/drm/virtio/virtgpu_vq.c:1090
>                    virtio_gpu_object_create+0x26f/0x490 drivers/gpu/drm/virtio/virtgpu_object.c:150
>                    virtio_gpu_gem_create+0xaa/0x1d0 drivers/gpu/drm/virtio/virtgpu_gem.c:42
>                    virtio_gpu_mode_dumb_create+0x21e/0x360 drivers/gpu/drm/virtio/virtgpu_gem.c:82
>                    drm_mode_create_dumb+0x27c/0x300 drivers/gpu/drm/drm_dumb_buffers.c:94
>                    drm_client_buffer_create drivers/gpu/drm/drm_client.c:267 [inline]
>                    drm_client_framebuffer_create+0x1b7/0x770 drivers/gpu/drm/drm_client.c:412
>                    drm_fb_helper_generic_probe+0x1e4/0x810 drivers/gpu/drm/drm_fb_helper.c:2051
>                    drm_fb_helper_single_fb_probe drivers/gpu/drm/drm_fb_helper.c:1600 [inline]
>                    __drm_fb_helper_initial_config_and_unlock+0xb56/0x11e0 drivers/gpu/drm/drm_fb_helper.c:1758
>                    drm_fb_helper_initial_config drivers/gpu/drm/drm_fb_helper.c:1853 [inline]
>                    drm_fb_helper_initial_config drivers/gpu/drm/drm_fb_helper.c:1845 [inline]
>                    drm_fbdev_client_hotplug+0x30f/0x580 drivers/gpu/drm/drm_fb_helper.c:2145
>                    drm_fbdev_generic_setup drivers/gpu/drm/drm_fb_helper.c:2224 [inline]
>                    drm_fbdev_generic_setup+0x18b/0x295 drivers/gpu/drm/drm_fb_helper.c:2197
>                    virtio_gpu_probe+0x28f/0x2de drivers/gpu/drm/virtio/virtgpu_drv.c:126
>                    virtio_dev_probe+0x463/0x710 drivers/virtio/virtio.c:248
>                    really_probe+0x281/0x6d0 drivers/base/dd.c:551
>                    driver_probe_device+0x104/0x210 drivers/base/dd.c:724
>                    device_driver_attach+0x108/0x140 drivers/base/dd.c:998
>                    __driver_attach+0xda/0x240 drivers/base/dd.c:1075
>                    bus_for_each_dev+0x14b/0x1d0 drivers/base/bus.c:305
>                    bus_add_driver+0x4a2/0x5a0 drivers/base/bus.c:622
>                    driver_register+0x1c4/0x330 drivers/base/driver.c:171
>                    do_one_initcall+0x10a/0x7d0 init/main.c:1152
>                    do_initcall_level init/main.c:1225 [inline]
>                    do_initcalls init/main.c:1241 [inline]
>                    do_basic_setup init/main.c:1261 [inline]
>                    kernel_init_freeable+0x501/0x5ae init/main.c:1445
>                    kernel_init+0xd/0x1bb init/main.c:1352
>                    ret_from_fork+0x24/0x30 arch/x86/entry/entry_64.S:352
>  }
>  ... key      at: [<ffffffff8c3f0420>] __key.55978+0x0/0x40
>  ... acquired at:
>    mark_lock_irq kernel/locking/lockdep.c:3316 [inline]
>    mark_lock+0x50e/0x1220 kernel/locking/lockdep.c:3665
>    mark_usage kernel/locking/lockdep.c:3583 [inline]
>    __lock_acquire+0x1236/0x3ca0 kernel/locking/lockdep.c:3908
>    lock_acquire+0x197/0x420 kernel/locking/lockdep.c:4484
>    __raw_spin_lock include/linux/spinlock_api_smp.h:142 [inline]
>    _raw_spin_lock+0x2a/0x40 kernel/locking/spinlock.c:151
>    spin_lock include/linux/spinlock.h:338 [inline]
>    shmem_mfill_atomic_pte+0x1012/0x21c0 mm/shmem.c:2407
>    shmem_mcopy_atomic_pte+0x3a/0x50 mm/shmem.c:2445
>    mfill_atomic_pte mm/userfaultfd.c:434 [inline]
>    __mcopy_atomic mm/userfaultfd.c:557 [inline]
>    mcopy_atomic+0xac7/0x2510 mm/userfaultfd.c:607
>    userfaultfd_copy fs/userfaultfd.c:1736 [inline]
>    userfaultfd_ioctl+0x4d2/0x3b10 fs/userfaultfd.c:1886
>    vfs_ioctl fs/ioctl.c:47 [inline]
>    ksys_ioctl+0x11a/0x180 fs/ioctl.c:763
>    __do_sys_ioctl fs/ioctl.c:772 [inline]
>    __se_sys_ioctl fs/ioctl.c:770 [inline]
>    __x64_sys_ioctl+0x6f/0xb0 fs/ioctl.c:770
>    do_syscall_64+0xf6/0x7d0 arch/x86/entry/common.c:294
>    entry_SYSCALL_64_after_hwframe+0x49/0xbe
>
>
> stack backtrace:
> CPU: 0 PID: 10317 Comm: syz-executor.0 Not tainted 5.6.0-rc7-syzkaller #0
> Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS rel-1.12.0-59-gc9ba5276e321-prebuilt.qemu.org 04/01/2014
> Call Trace:
>  __dump_stack lib/dump_stack.c:77 [inline]
>  dump_stack+0x188/0x20d lib/dump_stack.c:118
>  print_irq_inversion_bug kernel/locking/lockdep.c:3179 [inline]
>  check_usage_backwards.cold+0x1d/0x26 kernel/locking/lockdep.c:3230
>  mark_lock_irq kernel/locking/lockdep.c:3316 [inline]
>  mark_lock+0x50e/0x1220 kernel/locking/lockdep.c:3665
>  mark_usage kernel/locking/lockdep.c:3583 [inline]
>  __lock_acquire+0x1236/0x3ca0 kernel/locking/lockdep.c:3908
>  lock_acquire+0x197/0x420 kernel/locking/lockdep.c:4484
>  __raw_spin_lock include/linux/spinlock_api_smp.h:142 [inline]
>  _raw_spin_lock+0x2a/0x40 kernel/locking/spinlock.c:151
>  spin_lock include/linux/spinlock.h:338 [inline]
>  shmem_mfill_atomic_pte+0x1012/0x21c0 mm/shmem.c:2407
>  shmem_mcopy_atomic_pte+0x3a/0x50 mm/shmem.c:2445
>  mfill_atomic_pte mm/userfaultfd.c:434 [inline]
>  __mcopy_atomic mm/userfaultfd.c:557 [inline]
>  mcopy_atomic+0xac7/0x2510 mm/userfaultfd.c:607
>  userfaultfd_copy fs/userfaultfd.c:1736 [inline]
>  userfaultfd_ioctl+0x4d2/0x3b10 fs/userfaultfd.c:1886
>  vfs_ioctl fs/ioctl.c:47 [inline]
>  ksys_ioctl+0x11a/0x180 fs/ioctl.c:763
>  __do_sys_ioctl fs/ioctl.c:772 [inline]
>  __se_sys_ioctl fs/ioctl.c:770 [inline]
>  __x64_sys_ioctl+0x6f/0xb0 fs/ioctl.c:770
>  do_syscall_64+0xf6/0x7d0 arch/x86/entry/common.c:294
>  entry_SYSCALL_64_after_hwframe+0x49/0xbe
> RIP: 0033:0x45c6e9
> Code: bd b1 fb ff c3 66 2e 0f 1f 84 00 00 00 00 00 66 90 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 0f 83 8b b1 fb ff c3 66 2e 0f 1f 84 00 00 00 00
> RSP: 002b:00007fcff1e50c88 EFLAGS: 00000246 ORIG_RAX: 0000000000000010
> RAX: ffffffffffffffda RBX: 000000000072bf00 RCX: 000000000045c6e9
> RDX: 00000000200a0fe0 RSI: 00000000c028aa03 RDI: 0000000000000003
> RBP: 00007fcff1e516d4 R08: 0000000000000000 R09: 0000000000000000
> R10: 0000000000000000 R11: 0000000000000246 R12: 00000000ffffffff
> R13: 00000000000005b3 R14: 00000000004af2cd R15: 00000000006ec420
>
>
> ---
> This bug is generated by a bot. It may contain errors.
> See https://goo.gl/tpsmEJ for more information about syzbot.
> syzbot engineers can be reached at syzkaller@...glegroups.com.
>
> syzbot will keep track of this bug report. See:
> https://goo.gl/tpsmEJ#status for how to communicate with syzbot.
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ