lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CACT4Y+aGjkmq4cEaQXagd_YqjE4a1HoNgcEzqeNj-g0Hg_hQHw@mail.gmail.com>
Date:   Thu, 6 Sep 2018 21:41:37 +0200
From:   Dmitry Vyukov <dvyukov@...gle.com>
To:     "Theodore Y. Ts'o" <tytso@....edu>,
        syzbot <syzbot+0eefc1e06a77d327a056@...kaller.appspotmail.com>,
        Andreas Dilger <adilger.kernel@...ger.ca>,
        linux-ext4@...r.kernel.org, LKML <linux-kernel@...r.kernel.org>,
        syzkaller-bugs <syzkaller-bugs@...glegroups.com>,
        Casey Schaufler <casey@...aufler-ca.com>,
        James Morris <jmorris@...ei.org>,
        "Serge E. Hallyn" <serge@...lyn.com>,
        linux-security-module@...r.kernel.org
Subject: Re: possible deadlock in ext4_evict_inode

On Thu, Sep 6, 2018 at 9:38 PM, Theodore Y. Ts'o <tytso@....edu> wrote:
> On Thu, Sep 06, 2018 at 09:41:04AM -0700, syzbot wrote:
>> Hello,
>>
>> syzbot found the following crash on:
>>
>> HEAD commit:    b36fdc6853a3 Merge tag 'gpio-v4.19-2' of git://git.kernel...
>> git tree:       upstream
>> console output: https://syzkaller.appspot.com/x/log.txt?x=1716bc9e400000
>> kernel config:  https://syzkaller.appspot.com/x/.config?x=6c9564cd177daf0c
>> dashboard link: https://syzkaller.appspot.com/bug?extid=0eefc1e06a77d327a056
>> compiler:       gcc (GCC) 8.0.1 20180413 (experimental)
>> syz repro:      https://syzkaller.appspot.com/x/repro.syz?x=13db48be400000
>>
>> IMPORTANT: if you fix the bug, please add the following tag to the commit:
>> Reported-by: syzbot+0eefc1e06a77d327a056@...kaller.appspotmail.com
>
> This looks like it's a Smack issue?


Looks like it.

Happens only on smack instance:
https://syzkaller.appspot.com/bug?extid=0eefc1e06a77d327a056

Let's add +smack maintainers.


>> 8021q: adding VLAN 0 to HW filter on device team0
>> 8021q: adding VLAN 0 to HW filter on device team0
>> syz-executor0 (11193) used greatest stack depth: 15352 bytes left
>>
>> ======================================================
>> WARNING: possible circular locking dependency detected
>> 4.19.0-rc2+ #1 Not tainted
>> ------------------------------------------------------
>> syz-executor3/11182 is trying to acquire lock:
>> 00000000c157b042 (sb_internal){.+.+}, at: sb_start_intwrite
>> include/linux/fs.h:1613 [inline]
>> 00000000c157b042 (sb_internal){.+.+}, at: ext4_evict_inode+0x588/0x19b0
>> fs/ext4/inode.c:250
>>
>> but task is already holding lock:
>> 00000000128cdd3b (fs_reclaim){+.+.}, at: fs_reclaim_acquire.part.98+0x0/0x30
>> mm/page_alloc.c:463
>>
>> which lock already depends on the new lock.
>>
>>
>> the existing dependency chain (in reverse order) is:
>>
>> -> #3 (fs_reclaim){+.+.}:
>>        __fs_reclaim_acquire mm/page_alloc.c:3728 [inline]
>>        fs_reclaim_acquire.part.98+0x24/0x30 mm/page_alloc.c:3739
>>        fs_reclaim_acquire+0x14/0x20 mm/page_alloc.c:3740
>>        slab_pre_alloc_hook mm/slab.h:418 [inline]
>>        slab_alloc mm/slab.c:3378 [inline]
>>        kmem_cache_alloc_trace+0x2d/0x730 mm/slab.c:3618
>>        kmalloc include/linux/slab.h:513 [inline]
>>        kzalloc include/linux/slab.h:707 [inline]
>>        smk_fetch.part.24+0x5a/0xf0 security/smack/smack_lsm.c:273
>>        smk_fetch security/smack/smack_lsm.c:3548 [inline]
>>        smack_d_instantiate+0x946/0xea0 security/smack/smack_lsm.c:3502
>>        security_d_instantiate+0x5c/0xf0 security/security.c:1287
>>        d_instantiate+0x5e/0xa0 fs/dcache.c:1870
>>        shmem_mknod+0x189/0x1f0 mm/shmem.c:2812
>>        vfs_mknod+0x447/0x800 fs/namei.c:3719
>>        handle_create+0x1ff/0x7c0 drivers/base/devtmpfs.c:211
>>        handle drivers/base/devtmpfs.c:374 [inline]
>>        devtmpfsd+0x27f/0x4c0 drivers/base/devtmpfs.c:400
>>        kthread+0x35a/0x420 kernel/kthread.c:246
>>        ret_from_fork+0x3a/0x50 arch/x86/entry/entry_64.S:413
>>
>> -> #2 (&isp->smk_lock){+.+.}:
>>        __mutex_lock_common kernel/locking/mutex.c:925 [inline]
>>        __mutex_lock+0x171/0x1700 kernel/locking/mutex.c:1073
>>        mutex_lock_nested+0x16/0x20 kernel/locking/mutex.c:1088
>>        smack_d_instantiate+0x130/0xea0 security/smack/smack_lsm.c:3369
>>        security_d_instantiate+0x5c/0xf0 security/security.c:1287
>>        d_instantiate_new+0x7e/0x160 fs/dcache.c:1889
>>        ext4_add_nondir+0x81/0x90 fs/ext4/namei.c:2415
>>        ext4_symlink+0x761/0x1170 fs/ext4/namei.c:3162
>>        vfs_symlink+0x37a/0x5d0 fs/namei.c:4127
>>        do_symlinkat+0x242/0x2d0 fs/namei.c:4154
>>        __do_sys_symlink fs/namei.c:4173 [inline]
>>        __se_sys_symlink fs/namei.c:4171 [inline]
>>        __x64_sys_symlink+0x59/0x80 fs/namei.c:4171
>>        do_syscall_64+0x1b9/0x820 arch/x86/entry/common.c:290
>>        entry_SYSCALL_64_after_hwframe+0x49/0xbe
>>
>> -> #1 (jbd2_handle){++++}:
>>        start_this_handle+0x5c0/0x1260 fs/jbd2/transaction.c:385
>>        jbd2__journal_start+0x3c9/0x9f0 fs/jbd2/transaction.c:439
>>        __ext4_journal_start_sb+0x18d/0x590 fs/ext4/ext4_jbd2.c:81
>>        ext4_sample_last_mounted fs/ext4/file.c:414 [inline]
>>        ext4_file_open+0x552/0x7b0 fs/ext4/file.c:439
>>        do_dentry_open+0x499/0x1250 fs/open.c:771
>>        vfs_open+0xa0/0xd0 fs/open.c:880
>>        do_last fs/namei.c:3418 [inline]
>>        path_openat+0x130f/0x5340 fs/namei.c:3534
>>        do_filp_open+0x255/0x380 fs/namei.c:3564
>>        do_open_execat+0x221/0x8e0 fs/exec.c:853
>>        __do_execve_file.isra.35+0x1707/0x2460 fs/exec.c:1755
>>        do_execveat_common fs/exec.c:1866 [inline]
>>        do_execve fs/exec.c:1883 [inline]
>>        __do_sys_execve fs/exec.c:1964 [inline]
>>        __se_sys_execve fs/exec.c:1959 [inline]
>>        __x64_sys_execve+0x8f/0xc0 fs/exec.c:1959
>>        do_syscall_64+0x1b9/0x820 arch/x86/entry/common.c:290
>>        entry_SYSCALL_64_after_hwframe+0x49/0xbe
>>
>> -> #0 (sb_internal){.+.+}:
>>        lock_acquire+0x1e4/0x4f0 kernel/locking/lockdep.c:3901
>>        percpu_down_read_preempt_disable include/linux/percpu-rwsem.h:36
>> [inline]
>>        percpu_down_read include/linux/percpu-rwsem.h:59 [inline]
>>        __sb_start_write+0x1e9/0x300 fs/super.c:1387
>>        sb_start_intwrite include/linux/fs.h:1613 [inline]
>>        ext4_evict_inode+0x588/0x19b0 fs/ext4/inode.c:250
>>        evict+0x4ae/0x990 fs/inode.c:558
>>        iput_final fs/inode.c:1547 [inline]
>>        iput+0x5fa/0xa00 fs/inode.c:1573
>>        dentry_unlink_inode+0x461/0x5e0 fs/dcache.c:374
>>        __dentry_kill+0x44c/0x7a0 fs/dcache.c:566
>>        dentry_kill+0xc9/0x5a0 fs/dcache.c:685
>>        shrink_dentry_list+0x36c/0x7c0 fs/dcache.c:1090
>>        prune_dcache_sb+0x12f/0x1c0 fs/dcache.c:1171
>>        super_cache_scan+0x270/0x480 fs/super.c:102
>>        do_shrink_slab+0x4ba/0xbb0 mm/vmscan.c:536
>>        shrink_slab_memcg mm/vmscan.c:601 [inline]
>>        shrink_slab+0x6fe/0x8c0 mm/vmscan.c:674
>>        shrink_node+0x429/0x16a0 mm/vmscan.c:2735
>>        shrink_zones mm/vmscan.c:2964 [inline]
>>        do_try_to_free_pages+0x3e7/0x1290 mm/vmscan.c:3026
>>        try_to_free_pages+0x4b2/0xa60 mm/vmscan.c:3241
>>        __perform_reclaim mm/page_alloc.c:3769 [inline]
>>        __alloc_pages_direct_reclaim mm/page_alloc.c:3790 [inline]
>>        __alloc_pages_slowpath+0x95a/0x2cb0 mm/page_alloc.c:4191
>>        __alloc_pages_nodemask+0xa1b/0xd10 mm/page_alloc.c:4390
>>        __alloc_pages include/linux/gfp.h:473 [inline]
>>        __alloc_pages_node include/linux/gfp.h:486 [inline]
>>        kmem_getpages mm/slab.c:1409 [inline]
>>        cache_grow_begin+0x91/0x710 mm/slab.c:2677
>>        fallback_alloc+0x203/0x2c0 mm/slab.c:3219
>>        ____cache_alloc_node+0x1c7/0x1e0 mm/slab.c:3287
>>        slab_alloc_node mm/slab.c:3327 [inline]
>>        kmem_cache_alloc_node_trace+0xe9/0x720 mm/slab.c:3661
>>        __do_kmalloc_node mm/slab.c:3681 [inline]
>>        __kmalloc_node+0x33/0x70 mm/slab.c:3689
>>        kmalloc_node include/linux/slab.h:555 [inline]
>>        kvmalloc_node+0xb9/0xf0 mm/util.c:423
>>        kvmalloc include/linux/mm.h:577 [inline]
>>        sem_alloc ipc/sem.c:497 [inline]
>>        newary+0x244/0xb50 ipc/sem.c:527
>>        ipcget_new ipc/util.c:315 [inline]
>>        ipcget+0x15d/0x11d0 ipc/util.c:614
>>        ksys_semget+0x1c0/0x280 ipc/sem.c:604
>>        __do_sys_semget ipc/sem.c:609 [inline]
>>        __se_sys_semget ipc/sem.c:607 [inline]
>>        __x64_sys_semget+0x73/0xb0 ipc/sem.c:607
>>        do_syscall_64+0x1b9/0x820 arch/x86/entry/common.c:290
>>        entry_SYSCALL_64_after_hwframe+0x49/0xbe
>>
>> other info that might help us debug this:
>>
>> Chain exists of:
>>   sb_internal --> &isp->smk_lock --> fs_reclaim
>>
>>  Possible unsafe locking scenario:
>>
>>        CPU0                    CPU1
>>        ----                    ----
>>   lock(fs_reclaim);
>>                                lock(&isp->smk_lock);
>>                                lock(fs_reclaim);
>>   lock(sb_internal);
>>
>>  *** DEADLOCK ***
>>
>> 4 locks held by syz-executor3/11182:
>>  #0: 000000000ed49aa7 (&ids->rwsem){+.+.}, at: ipcget_new ipc/util.c:314
>> [inline]
>>  #0: 000000000ed49aa7 (&ids->rwsem){+.+.}, at: ipcget+0x125/0x11d0
>> ipc/util.c:614
>>  #1: 00000000128cdd3b (fs_reclaim){+.+.}, at:
>> fs_reclaim_acquire.part.98+0x0/0x30 mm/page_alloc.c:463
>>  #2: 00000000c7d74038 (shrinker_rwsem){++++}, at: shrink_slab_memcg
>> mm/vmscan.c:578 [inline]
>>  #2: 00000000c7d74038 (shrinker_rwsem){++++}, at: shrink_slab+0x1d1/0x8c0
>> mm/vmscan.c:674
>>  #3: 00000000a3e33771 (&type->s_umount_key#28){++++}, at:
>> trylock_super+0x22/0x110 fs/super.c:412
>>
>> stack backtrace:
>> CPU: 1 PID: 11182 Comm: syz-executor3 Not tainted 4.19.0-rc2+ #1
>> Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS
>> Google 01/01/2011
>> Call Trace:
>>  __dump_stack lib/dump_stack.c:77 [inline]
>>  dump_stack+0x1c9/0x2b4 lib/dump_stack.c:113
>>  print_circular_bug.isra.34.cold.55+0x1bd/0x27d
>> kernel/locking/lockdep.c:1222
>>  check_prev_add kernel/locking/lockdep.c:1862 [inline]
>>  check_prevs_add kernel/locking/lockdep.c:1975 [inline]
>>  validate_chain kernel/locking/lockdep.c:2416 [inline]
>>  __lock_acquire+0x3449/0x5020 kernel/locking/lockdep.c:3412
>>  lock_acquire+0x1e4/0x4f0 kernel/locking/lockdep.c:3901
>>  percpu_down_read_preempt_disable include/linux/percpu-rwsem.h:36 [inline]
>>  percpu_down_read include/linux/percpu-rwsem.h:59 [inline]
>>  __sb_start_write+0x1e9/0x300 fs/super.c:1387
>>  sb_start_intwrite include/linux/fs.h:1613 [inline]
>>  ext4_evict_inode+0x588/0x19b0 fs/ext4/inode.c:250
>>  evict+0x4ae/0x990 fs/inode.c:558
>>  iput_final fs/inode.c:1547 [inline]
>>  iput+0x5fa/0xa00 fs/inode.c:1573
>>  dentry_unlink_inode+0x461/0x5e0 fs/dcache.c:374
>>  __dentry_kill+0x44c/0x7a0 fs/dcache.c:566
>>  dentry_kill+0xc9/0x5a0 fs/dcache.c:685
>>  shrink_dentry_list+0x36c/0x7c0 fs/dcache.c:1090
>>  prune_dcache_sb+0x12f/0x1c0 fs/dcache.c:1171
>>  super_cache_scan+0x270/0x480 fs/super.c:102
>>  do_shrink_slab+0x4ba/0xbb0 mm/vmscan.c:536
>>  shrink_slab_memcg mm/vmscan.c:601 [inline]
>>  shrink_slab+0x6fe/0x8c0 mm/vmscan.c:674
>>  shrink_node+0x429/0x16a0 mm/vmscan.c:2735
>>  shrink_zones mm/vmscan.c:2964 [inline]
>>  do_try_to_free_pages+0x3e7/0x1290 mm/vmscan.c:3026
>>  try_to_free_pages+0x4b2/0xa60 mm/vmscan.c:3241
>>  __perform_reclaim mm/page_alloc.c:3769 [inline]
>>  __alloc_pages_direct_reclaim mm/page_alloc.c:3790 [inline]
>>  __alloc_pages_slowpath+0x95a/0x2cb0 mm/page_alloc.c:4191
>>  __alloc_pages_nodemask+0xa1b/0xd10 mm/page_alloc.c:4390
>>  __alloc_pages include/linux/gfp.h:473 [inline]
>>  __alloc_pages_node include/linux/gfp.h:486 [inline]
>>  kmem_getpages mm/slab.c:1409 [inline]
>>  cache_grow_begin+0x91/0x710 mm/slab.c:2677
>>  fallback_alloc+0x203/0x2c0 mm/slab.c:3219
>>  ____cache_alloc_node+0x1c7/0x1e0 mm/slab.c:3287
>>  slab_alloc_node mm/slab.c:3327 [inline]
>>  kmem_cache_alloc_node_trace+0xe9/0x720 mm/slab.c:3661
>>  __do_kmalloc_node mm/slab.c:3681 [inline]
>>  __kmalloc_node+0x33/0x70 mm/slab.c:3689
>>  kmalloc_node include/linux/slab.h:555 [inline]
>>  kvmalloc_node+0xb9/0xf0 mm/util.c:423
>>  kvmalloc include/linux/mm.h:577 [inline]
>>  sem_alloc ipc/sem.c:497 [inline]
>>  newary+0x244/0xb50 ipc/sem.c:527
>>  ipcget_new ipc/util.c:315 [inline]
>>  ipcget+0x15d/0x11d0 ipc/util.c:614
>>  ksys_semget+0x1c0/0x280 ipc/sem.c:604
>>  __do_sys_semget ipc/sem.c:609 [inline]
>>  __se_sys_semget ipc/sem.c:607 [inline]
>>  __x64_sys_semget+0x73/0xb0 ipc/sem.c:607
>>  do_syscall_64+0x1b9/0x820 arch/x86/entry/common.c:290
>>  entry_SYSCALL_64_after_hwframe+0x49/0xbe
>> RIP: 0033:0x457099
>> Code: fd b4 fb ff c3 66 2e 0f 1f 84 00 00 00 00 00 66 90 48 89 f8 48 89 f7
>> 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff
>> 0f 83 cb b4 fb ff c3 66 2e 0f 1f 84 00 00 00 00
>> RSP: 002b:00007fbecf217c78 EFLAGS: 00000246 ORIG_RAX: 0000000000000040
>> RAX: ffffffffffffffda RBX: 00007fbecf2186d4 RCX: 0000000000457099
>> RDX: 0000000000000000 RSI: 0000000000004007 RDI: 0000000000000000
>> RBP: 00000000009300a0 R08: 0000000000000000 R09: 0000000000000000
>> R10: 0000000000000000 R11: 0000000000000246 R12: 00000000ffffffff
>> R13: 00000000004d4740 R14: 00000000004c8d3e R15: 0000000000000000
>> syz-executor4 (11186) used greatest stack depth: 15080 bytes left
>> syz-executor7 (11171) used greatest stack depth: 14344 bytes left
>> syz-executor2 (11260) used greatest stack depth: 14296 bytes left
>> syz-executor0 invoked oom-killer: gfp_mask=0x6200ca(GFP_HIGHUSER_MOVABLE),
>> nodemask=(null), order=0, oom_score_adj=0
>> syz-executor0 cpuset=syz0 mems_allowed=0
>> CPU: 0 PID: 11308 Comm: syz-executor0 Not tainted 4.19.0-rc2+ #1
>> Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS
>> Google 01/01/2011
>> Call Trace:
>>  __dump_stack lib/dump_stack.c:77 [inline]
>>  dump_stack+0x1c9/0x2b4 lib/dump_stack.c:113
>>  dump_header+0x27b/0xf70 mm/oom_kill.c:441
>>  oom_kill_process.cold.28+0x10/0x95a mm/oom_kill.c:953
>>  out_of_memory+0xa88/0x1430 mm/oom_kill.c:1120
>>  __alloc_pages_may_oom mm/page_alloc.c:3522 [inline]
>>  __alloc_pages_slowpath+0x223f/0x2cb0 mm/page_alloc.c:4235
>>  __alloc_pages_nodemask+0xa1b/0xd10 mm/page_alloc.c:4390
>>  alloc_pages_current+0x10c/0x210 mm/mempolicy.c:2093
>>  alloc_pages include/linux/gfp.h:509 [inline]
>>  __page_cache_alloc+0x398/0x5e0 mm/filemap.c:946
>>  page_cache_read mm/filemap.c:2385 [inline]
>>  filemap_fault+0x1458/0x2220 mm/filemap.c:2569
>>  ext4_filemap_fault+0x82/0xad fs/ext4/inode.c:6257
>>  __do_fault+0xee/0x450 mm/memory.c:3240
>>  do_read_fault mm/memory.c:3652 [inline]
>>  do_fault mm/memory.c:3752 [inline]
>>  handle_pte_fault mm/memory.c:3983 [inline]
>>  __handle_mm_fault+0x2b4a/0x4350 mm/memory.c:4107
>>  handle_mm_fault+0x53e/0xc80 mm/memory.c:4144
>>  __do_page_fault+0x620/0xe50 arch/x86/mm/fault.c:1395
>>  do_page_fault+0xf6/0x7a4 arch/x86/mm/fault.c:1470
>>  page_fault+0x1e/0x30 arch/x86/entry/entry_64.S:1161
>> RIP: 0033:0x40a056
>> Code: Bad RIP value.
>> RSP: 002b:00007ffd766513d0 EFLAGS: 00010283
>> RAX: 0000000000730a08 RBX: 00000000009300a0 RCX: 0000000000000001
>> RDX: 0000000000000001 RSI: ffffffffffffffff RDI: 00000000004bce88
>> RBP: 00000000009300a0 R08: 0000000000000009 R09: 0000000000000001
>> R10: 00007ffd766514d0 R11: 0000000000000000 R12: 0000000000930aa0
>> R13: 000000000093014c R14: 000000000001f9ab R15: 000000000001f97e
>> Mem-Info:
>> active_anon:4715 inactive_anon:367 isolated_anon:0
>>  active_file:31 inactive_file:10 isolated_file:0
>>  unevictable:0 dirty:0 writeback:0 unstable:0
>>  slab_reclaimable:11145 slab_unreclaimable:1533913
>>  mapped:177 shmem:377 pagetables:511 bounce:0
>>  free:24282 free_pcp:0 free_cma:0
>> Node 0 active_anon:18860kB inactive_anon:1468kB active_file:116kB
>> inactive_file:0kB unevictable:0kB isolated(anon):0kB isolated(file):56kB
>> mapped:708kB dirty:0kB writeback:0kB shmem:1508kB shmem_thp: 0kB
>> shmem_pmdmapped: 0kB anon_thp: 6144kB writeback_tmp:0kB unstable:0kB
>> all_unreclaimable? yes
>> Node 0 DMA free:15908kB min:164kB low:204kB high:244kB active_anon:0kB
>> inactive_anon:0kB active_file:0kB inactive_file:0kB unevictable:0kB
>> writepending:0kB present:15992kB managed:15908kB mlocked:0kB
>> kernel_stack:0kB pagetables:0kB bounce:0kB free_pcp:0kB local_pcp:0kB
>> free_cma:0kB
>> lowmem_reserve[]: 0 2842 6348 6348
>> Node 0 DMA32 free:44052kB min:30180kB low:37724kB high:45268kB
>> active_anon:2056kB inactive_anon:0kB active_file:40kB inactive_file:72kB
>> unevictable:0kB writepending:0kB present:3129292kB managed:2914192kB
>> mlocked:0kB kernel_stack:32kB pagetables:0kB bounce:0kB free_pcp:0kB
>> local_pcp:0kB free_cma:0kB
>> lowmem_reserve[]: 0 0 3506 3506
>> Node 0 Normal free:37168kB min:37236kB low:46544kB high:55852kB
>> active_anon:16804kB inactive_anon:1468kB active_file:176kB inactive_file:0kB
>> unevictable:0kB writepending:0kB present:4718592kB managed:3590864kB
>> mlocked:0kB kernel_stack:5216kB pagetables:2044kB bounce:0kB free_pcp:0kB
>> local_pcp:0kB free_cma:0kB
>> lowmem_reserve[]: 0 0 0 0
>> Node 0 DMA: 1*4kB (U) 0*8kB 0*16kB 1*32kB (U) 2*64kB (U) 1*128kB (U) 1*256kB
>> (U) 0*512kB 1*1024kB (U) 1*2048kB (M) 3*4096kB (M) = 15908kB
>> Node 0 DMA32: 5*4kB (ME) 7*8kB (UME) 4*16kB (ME) 4*32kB (ME) 4*64kB (ME)
>> 2*128kB (UE) 4*256kB (UME) 3*512kB (ME) 4*1024kB (UME) 4*2048kB (UME)
>> 7*4096kB (M) = 44300kB
>> Node 0 Normal: 867*4kB (UME) 448*8kB (M) 250*16kB (ME) 106*32kB (ME) 51*64kB
>> (ME) 46*128kB (ME) 20*256kB (M) 15*512kB (M) 1*1024kB (E) 0*2048kB 0*4096kB
>> = 37420kB
>> Node 0 hugepages_total=0 hugepages_free=0 hugepages_surp=0
>> hugepages_size=2048kB
>> 424 total pagecache pages
>> 0 pages in swap cache
>> Swap cache stats: add 0, delete 0, find 0/0
>> Free swap  = 0kB
>> Total swap = 0kB
>> 1965969 pages RAM
>> 0 pages HighMem/MovableOnly
>> 335728 pages reserved
>> Unreclaimable slab info:
>> Name                      Used          Total
>> pid_2                    131KB        156KB
>> TIPC                      12KB         14KB
>> SCTPv6                    18KB         24KB
>> DCCPv6                    21KB         29KB
>> DCCP                      20KB         27KB
>> bridge_fdb_cache          15KB         19KB
>> fib6_nodes               103KB        108KB
>> ip6_dst_cache            113KB        300KB
>> RAWv6                     87KB         91KB
>> UDPv6                      3KB          3KB
>> TCPv6                     29KB         29KB
>> nf_conntrack               2KB          3KB
>> sd_ext_cdb                 0KB          3KB
>> scsi_sense_cache        1056KB       1060KB
>> virtio_scsi_cmd           16KB         16KB
>> sgpool-128                 8KB          8KB
>> sgpool-64                  4KB          6KB
>> sgpool-32                  2KB          7KB
>> sgpool-16                  1KB          3KB
>> sgpool-8                   0KB          3KB
>> mqueue_inode_cache         12KB         21KB
>> bio_post_read_ctx         14KB         15KB
>> bio-2                     14KB         15KB
>> jfs_mp                     7KB          7KB
>> nfs_commit_data            3KB          7KB
>> nfs_write_data            34KB         37KB
>> jbd2_inode                 2KB          3KB
>> ext4_system_zone           0KB          3KB
>> bio-1                      1KB          3KB
>> pid_namespace              2KB          7KB
>> rpc_buffers               17KB         19KB
>> rpc_tasks                  2KB          3KB
>> UNIX                      12KB         47KB
>> tcp_bind_bucket            1KB          4KB
>> ip_fib_trie               14KB         19KB
>> ip_fib_alias              69KB         71KB
>> ip_dst_cache               0KB          8KB
>> RAW                       51KB         57KB
>> UDP                       19KB         39KB
>> TCP                        5KB          5KB
>> hugetlbfs_inode_cache          1KB          7KB
>> fscache_cookie_jar          0KB          7KB
>> eventpoll_pwq             51KB         51KB
>> eventpoll_epi             90KB         90KB
>> inotify_inode_mark         90KB         90KB
>> request_queue            159KB        159KB
>> blkdev_requests            1KB          3KB
>> blkdev_ioc                23KB         23KB
>> bio-0                    536KB        536KB
>> biovec-max              1658KB       1658KB
>> biovec-64                401KB        401KB
>> biovec-16                127KB        127KB
>> bio_integrity_payload          1KB          4KB
>> khugepaged_mm_slot         27KB         27KB
>> dmaengine-unmap-2          0KB          3KB
>> skbuff_fclone_cache          1KB         18KB
>> skbuff_head_cache        673KB       1916KB
>> configfs_dir_cache          0KB          4KB
>> file_lock_cache            1KB          7KB
>> file_lock_ctx              0KB          3KB
>> fsnotify_mark_connector         51KB         51KB
>> net_namespace             69KB         69KB
>> shmem_inode_cache       3043KB       3043KB
>> task_delay_info          125KB        371KB
>> taskstats                176KB        176KB
>> proc_dir_entry           680KB        682KB
>> pde_opener                 0KB          3KB
>> seq_file                 362KB        362KB
>> sigqueue                 126KB        401KB
>> kernfs_node_cache      11279KB      11284KB
>> mnt_cache                 85KB        108KB
>> filp                    4848KB       8073KB
>> names_cache           121567KB     121567KB
>> inode_smack             4817KB       4817KB
>> key_jar                    3KB          7KB
>> uts_namespace              3KB          7KB
>> nsproxy                    1KB          7KB
>> vm_area_struct          8830KB      12970KB
>> mm_struct               1462KB       3614KB
>> fs_cache                 161KB        496KB
>> files_cache              538KB       1256KB
>> signal_cache             915KB       2185KB
>> sighand_cache            395KB        395KB
>> task_struct             3249KB       3304KB
>> cred_jar                 795KB       2348KB
>> anon_vma_chain          4757KB       6000KB
>> anon_vma                 237KB        392KB
>> pid                       73KB        228KB
>> Acpi-Operand             312KB        772KB
>> Acpi-Namespace           102KB        104KB
>> numa_policy                0KB          3KB
>> debug_objects_cache       1125KB       1126KB
>> trace_event_file         243KB        247KB
>> ftrace_event_field        348KB        350KB
>> pool_workqueue            67KB         72KB
>> task_group                 7KB          7KB
>> page->ptl               1564KB       3300KB
>> kmalloc-4194304      5713920KB    5713920KB
>> kmalloc-2097152         2050KB       2050KB
>> kmalloc-524288          2056KB       2056KB
>> kmalloc-262144          1290KB       1290KB
>> kmalloc-131072           650KB        650KB
>> kmalloc-65536            330KB        330KB
>> kmalloc-32768            891KB        990KB
>> kmalloc-16384            412KB        462KB
>> kmalloc-8192            2004KB       2004KB
>> kmalloc-4096           18670KB      18708KB
>> kmalloc-2048            6464KB       7764KB
>> kmalloc-1024            4572KB       5244KB
>> kmalloc-512             3885KB       5812KB
>> kmalloc-256             4762KB       4762KB
>> kmalloc-128              901KB        901KB
>> kmalloc-96              1935KB       2112KB
>> kmalloc-64              1843KB       1896KB
>> kmalloc-32              1849KB       1956KB
>> kmalloc-192             2706KB       4076KB
>> kmem_cache               221KB        221KB
>> Tasks state (memory values in pages):
>> [  pid  ]   uid  tgid total_vm      rss pgtables_bytes swapents
>> oom_score_adj name
>> [   2347]     0  2347      278      186    32768        0             0 none
>> [   2539]     0  2539     5377      170    86016        0         -1000
>> udevd
>> [   2754]     0  2754     5376      171    81920        0         -1000
>> udevd
>> [   4381]     0  4381     2493      573    65536        0             0
>> dhclient
>> [   4544]     0  4544    14267      155   118784        0             0
>> rsyslogd
>> [   4587]     0  4587     4725       49    86016        0             0 cron
>> [   4613]     0  4613    12490      153   139264        0         -1000 sshd
>> [   4638]     0  4638     3694       42    69632        0             0
>> getty
>> [   4639]     0  4639     3694       39    77824        0             0
>> getty
>> [   4640]     0  4640     3694       41    77824        0             0
>> getty
>> [   4641]     0  4641     3694       41    77824        0             0
>> getty
>> [   4642]     0  4642     3694       39    73728        0             0
>> getty
>> [   4643]     0  4643     3694       40    77824        0             0
>> getty
>> [   4644]     0  4644     3649       41    73728        0             0
>> getty
>> [   4645]     0  4645     5310      115    81920        0         -1000
>> udevd
>> [   4663]     0  4663    17821      198   188416        0             0 sshd
>> [   4665]     0  4665    41970     1817   163840        0             0
>> syz-execprog
>> [   4676]     0  4676     9360       15    40960        0             0
>> syz-executor4
>> [   4678]     0  4678     9360       15    40960        0             0
>> syz-executor5
>> [   4677]     0  4677     9359       22    49152        0             0
>> syz-executor4
>> [   4681]     0  4681     9360       15    40960        0             0
>> syz-executor3
>> [   4682]     0  4682     9360       15    40960        0             0
>> syz-executor0
>> [   4683]     0  4683     9359       21    49152        0             0
>> syz-executor5
>> [   4685]     0  4685     9360       15    40960        0             0
>> syz-executor7
>> [   4686]     0  4686     9359       22    49152        0             0
>> syz-executor3
>> [   4687]     0  4687     9360       15    40960        0             0
>> syz-executor6
>> [   4688]     0  4688     9359       22    49152        0             0
>> syz-executor0
>> [   4689]     0  4689     9359       21    49152        0             0
>> syz-executor7
>> [   4690]     0  4690     9360       15    40960        0             0
>> syz-executor2
>> [   4691]     0  4691     9359       22    49152        0             0
>> syz-executor6
>> [   4692]     0  4692     9360       14    40960        0             0
>> syz-executor1
>> [   4693]     0  4693     9359       22    49152        0             0
>> syz-executor2
>> [   4694]     0  4694     9359       21    53248        0             0
>> syz-executor1
>> [   6704]     0  6704     5310      115    81920        0         -1000
>> udevd
>> [   6710]     0  6710     5310      115    81920        0         -1000
>> udevd
>> [   6737]     0  6737     5376      171    81920        0         -1000
>> udevd
>> [   6804]     0  6804     5376      172    81920        0         -1000
>> udevd
>> [   6819]     0  6819     5376      172    81920        0         -1000
>> udevd
>> [   7149]     0  7149     5376      172    81920        0         -1000
>> udevd
>> [  11308]     0 11308     9425       22    61440        0             0
>> syz-executor0
>> [  11311]     0 11311     9425      534    61440        0             0
>> syz-executor6
>> Out of memory: Kill process 4665 (syz-execprog) score 1 or sacrifice child
>> Killed process 4676 (syz-executor4) total-vm:37440kB, anon-rss:60kB,
>> file-rss:0kB, shmem-rss:0kB
>> syz-executor0 (11308) used greatest stack depth: 14040 bytes left
>> device bridge_slave_1 left promiscuous mode
>> bridge0: port 2(bridge_slave_1) entered disabled state
>> device bridge_slave_0 left promiscuous mode
>> bridge0: port 1(bridge_slave_0) entered disabled state
>> team0 (unregistering): Port device team_slave_1 removed
>> IPVS: ftp: loaded support on port[0] = 21
>> team0 (unregistering): Port device team_slave_0 removed
>> bond0 (unregistering): Releasing backup interface bond_slave_1
>> bond0 (unregistering): Releasing backup interface bond_slave_0
>> bond0 (unregistering): Released all slaves
>> bridge0: port 1(bridge_slave_0) entered blocking state
>> bridge0: port 1(bridge_slave_0) entered disabled state
>> device bridge_slave_0 entered promiscuous mode
>> bridge0: port 2(bridge_slave_1) entered blocking state
>> bridge0: port 2(bridge_slave_1) entered disabled state
>> device bridge_slave_1 entered promiscuous mode
>> IPv6: ADDRCONF(NETDEV_UP): veth0_to_bridge: link is not ready
>> IPv6: ADDRCONF(NETDEV_UP): veth1_to_bridge: link is not ready
>> bond0: Enslaving bond_slave_0 as an active interface with an up link
>> bond0: Enslaving bond_slave_1 as an active interface with an up link
>> IPv6: ADDRCONF(NETDEV_UP): team_slave_0: link is not ready
>> team0: Port device team_slave_0 added
>> IPv6: ADDRCONF(NETDEV_UP): team_slave_1: link is not ready
>> team0: Port device team_slave_1 added
>> IPv6: ADDRCONF(NETDEV_UP): veth0_to_team: link is not ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): veth0_to_team: link becomes ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): team_slave_0: link becomes ready
>> IPv6: ADDRCONF(NETDEV_UP): veth1_to_team: link is not ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): veth1_to_team: link becomes ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): team_slave_1: link becomes ready
>> IPv6: ADDRCONF(NETDEV_UP): bridge_slave_0: link is not ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): bridge_slave_0: link becomes ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): veth0_to_bridge: link becomes ready
>> IPv6: ADDRCONF(NETDEV_UP): bridge_slave_1: link is not ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): bridge_slave_1: link becomes ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): veth1_to_bridge: link becomes ready
>> bridge0: port 2(bridge_slave_1) entered blocking state
>> bridge0: port 2(bridge_slave_1) entered forwarding state
>> bridge0: port 1(bridge_slave_0) entered blocking state
>> bridge0: port 1(bridge_slave_0) entered forwarding state
>> IPv6: ADDRCONF(NETDEV_UP): bridge0: link is not ready
>> 8021q: adding VLAN 0 to HW filter on device bond0
>> IPv6: ADDRCONF(NETDEV_UP): veth0: link is not ready
>> IPv6: ADDRCONF(NETDEV_UP): veth1: link is not ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): veth1: link becomes ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): veth0: link becomes ready
>> IPv6: ADDRCONF(NETDEV_UP): team0: link is not ready
>> 8021q: adding VLAN 0 to HW filter on device team0
>> IPv6: ADDRCONF(NETDEV_CHANGE): team0: link becomes ready
>> IPv6: ADDRCONF(NETDEV_CHANGE): bridge0: link becomes ready
>>
>>
>> ---
>> This bug is generated by a bot. It may contain errors.
>> See https://goo.gl/tpsmEJ for more information about syzbot.
>> syzbot engineers can be reached at syzkaller@...glegroups.com.
>>
>> syzbot will keep track of this bug report. See:
>> https://goo.gl/tpsmEJ#bug-status-tracking for how to communicate with
>> syzbot.
>> syzbot can test patches for this bug, for details see:
>> https://goo.gl/tpsmEJ#testing-patches
>
> --
> You received this message because you are subscribed to the Google Groups "syzkaller-bugs" group.
> To unsubscribe from this group and stop receiving emails from it, send an email to syzkaller-bugs+unsubscribe@...glegroups.com.
> To view this discussion on the web visit https://groups.google.com/d/msgid/syzkaller-bugs/20180906193822.GG5098%40thunk.org.
> For more options, visit https://groups.google.com/d/optout.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ