lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <20110819215754.GA12746@redhat.com>
Date:	Fri, 19 Aug 2011 17:57:54 -0400
From:	Dave Jones <davej@...hat.com>
To:	Linux Kernel <linux-kernel@...r.kernel.org>
Cc:	jaxboe@...ionio.com
Subject: 3.1.0-rc2 block related lockdep report.

Just got this while running kvm. (this is from the host)

	Dave

=================================
[ INFO: inconsistent lock state ]
3.1.0-rc2+ #139
---------------------------------
inconsistent {RECLAIM_FS-ON-W} -> {IN-RECLAIM_FS-W} usage.
qemu-kvm/8194 [HC0[0]:SC0[0]:HE1:SE1] takes:
 (pcpu_alloc_mutex){+.+.?.}, at: [<ffffffff81110b78>] pcpu_alloc+0x6f/0x80b
{RECLAIM_FS-ON-W} state was registered at:
  [<ffffffff8109138e>] mark_held_locks+0x6d/0x95
  [<ffffffff810919a9>] lockdep_trace_alloc+0x9f/0xc2
  [<ffffffff8113168c>] slab_pre_alloc_hook+0x1e/0x4f
  [<ffffffff81133c16>] __kmalloc+0x64/0x12f
  [<ffffffff81110164>] pcpu_mem_alloc+0x5e/0x67
  [<ffffffff8111026b>] pcpu_extend_area_map+0x2b/0xd4
  [<ffffffff81110cc8>] pcpu_alloc+0x1bf/0x80b
  [<ffffffff81111324>] __alloc_percpu+0x10/0x12
  [<ffffffff81135651>] kmem_cache_open+0x2cc/0x2d6
  [<ffffffff81135834>] kmem_cache_create+0x1d9/0x281
  [<ffffffff812bcd01>] acpi_os_create_cache+0x1d/0x2d
  [<ffffffff812e4d26>] acpi_ut_create_caches+0x26/0xb0
  [<ffffffff812e76d2>] acpi_ut_init_globals+0xe/0x244
  [<ffffffff81d7881f>] acpi_initialize_subsystem+0x35/0xae
  [<ffffffff81d77539>] acpi_early_init+0x5c/0xf7
  [<ffffffff81d48b9e>] start_kernel+0x3dd/0x3f7
  [<ffffffff81d482c4>] x86_64_start_reservations+0xaf/0xb3
  [<ffffffff81d483ca>] x86_64_start_kernel+0x102/0x111
irq event stamp: 140939
hardirqs last  enabled at (140939): [<ffffffff814e1c55>] __slab_alloc+0x41c/0x43d
hardirqs last disabled at (140938): [<ffffffff814e187e>] __slab_alloc+0x45/0x43d
softirqs last  enabled at (140484): [<ffffffff8106481f>] __do_softirq+0x1fd/0x257
softirqs last disabled at (140461): [<ffffffff814f19fc>] call_softirq+0x1c/0x30

other info that might help us debug this:
 Possible unsafe locking scenario:

       CPU0
       ----
  lock(pcpu_alloc_mutex);
  <Interrupt>
    lock(pcpu_alloc_mutex);

 *** DEADLOCK ***

1 lock held by qemu-kvm/8194:
 #0:  (&mm->mmap_sem){++++++}, at: [<ffffffff814ec328>] do_page_fault+0x188/0x39d

stack backtrace:
Pid: 8194, comm: qemu-kvm Tainted: G        W   3.1.0-rc2+ #139
Call Trace:
 [<ffffffff81081031>] ? up+0x39/0x3e
 [<ffffffff814dea37>] print_usage_bug+0x1e7/0x1f8
 [<ffffffff8101bb8d>] ? save_stack_trace+0x2c/0x49
 [<ffffffff8108f65a>] ? print_irq_inversion_bug.part.19+0x1a0/0x1a0
 [<ffffffff8108fd88>] mark_lock+0x106/0x220
 [<ffffffff81090236>] __lock_acquire+0x394/0xcf7
 [<ffffffff814e9318>] ? _raw_spin_unlock+0x32/0x54
 [<ffffffff814e932d>] ? _raw_spin_unlock+0x47/0x54
 [<ffffffff811339d2>] ? deactivate_slab+0x293/0x2b9
 [<ffffffff81110b78>] ? pcpu_alloc+0x6f/0x80b
 [<ffffffff8109108f>] lock_acquire+0xf3/0x13e
 [<ffffffff81110b78>] ? pcpu_alloc+0x6f/0x80b
 [<ffffffff814e7c9d>] ? mutex_lock_nested+0x3b/0x40
 [<ffffffff81110b78>] ? pcpu_alloc+0x6f/0x80b
 [<ffffffff814e77bd>] __mutex_lock_common+0x65/0x44a
 [<ffffffff81110b78>] ? pcpu_alloc+0x6f/0x80b
 [<ffffffff810914e3>] ? trace_hardirqs_on_caller+0x12d/0x164
 [<ffffffff814e1c61>] ? __slab_alloc+0x428/0x43d
 [<ffffffff81266500>] ? kzalloc_node+0x14/0x16
 [<ffffffff814e7c9d>] mutex_lock_nested+0x3b/0x40
 [<ffffffff81110b78>] pcpu_alloc+0x6f/0x80b
 [<ffffffff81266500>] ? kzalloc_node+0x14/0x16
 [<ffffffff81133b8b>] ? __kmalloc_node+0x146/0x16d
 [<ffffffff81266500>] ? kzalloc_node+0x14/0x16
 [<ffffffff81111324>] __alloc_percpu+0x10/0x12
 [<ffffffff81264fab>] blkio_alloc_blkg_stats+0x1d/0x31
 [<ffffffff8126653c>] throtl_alloc_tg+0x3a/0xdf
 [<ffffffff81266f4e>] blk_throtl_bio+0x14b/0x38e
 [<ffffffff81016224>] ? __cycles_2_ns+0xe/0x3a
 [<ffffffff8108202e>] ? local_clock+0x14/0x4c
 [<ffffffff810164da>] ? native_sched_clock+0x34/0x36
 [<ffffffff81259e7a>] generic_make_request+0x2e8/0x419
 [<ffffffff814e937f>] ? _raw_spin_unlock_irqrestore+0x45/0x7a
 [<ffffffff810fd7a7>] ? test_set_page_writeback+0xcc/0xfd
 [<ffffffff8125a089>] submit_bio+0xde/0xfd
 [<ffffffff810fd441>] ? account_page_writeback+0x13/0x15
 [<ffffffff810fd7c6>] ? test_set_page_writeback+0xeb/0xfd
 [<ffffffff81122161>] swap_writepage+0x94/0x9f
 [<ffffffff811082b1>] shmem_writepage+0x192/0x1d8
 [<ffffffff811056b6>] shrink_page_list+0x402/0x795
 [<ffffffff81105e82>] shrink_inactive_list+0x22c/0x3e6
 [<ffffffff8109138e>] ? mark_held_locks+0x6d/0x95
 [<ffffffff81106755>] shrink_zone+0x445/0x588
 [<ffffffff81168463>] ? wakeup_flusher_threads+0xcf/0xd8
 [<ffffffff811683c6>] ? wakeup_flusher_threads+0x32/0xd8
 [<ffffffff81106c0d>] do_try_to_free_pages+0x107/0x318
 [<ffffffff811071a1>] try_to_free_pages+0xd5/0x175
 [<ffffffff810fcfff>] __alloc_pages_nodemask+0x501/0x7b7
 [<ffffffff810dc435>] ? trace_preempt_on+0x15/0x28
 [<ffffffff8108de9e>] ? lock_release_holdtime.part.10+0x59/0x62
 [<ffffffff8112bdbc>] alloc_pages_vma+0xf5/0xfa
 [<ffffffff8113b4be>] do_huge_pmd_anonymous_page+0xb3/0x274
 [<ffffffff8111226b>] ? pmd_offset+0x19/0x3f
 [<ffffffff8111547f>] handle_mm_fault+0xfd/0x1b8
 [<ffffffff814ec328>] ? do_page_fault+0x188/0x39d
 [<ffffffff814ec4f6>] do_page_fault+0x356/0x39d
 [<ffffffff8108fcaf>] ? mark_lock+0x2d/0x220
 [<ffffffff810dc40c>] ? time_hardirqs_off+0x1b/0x2f
 [<ffffffff8108d8ff>] ? trace_hardirqs_off_caller+0x3f/0x9c
 [<ffffffff81279dfd>] ? trace_hardirqs_off_thunk+0x3a/0x3c
 [<ffffffff814e9d25>] page_fault+0x25/0x30
 [<ffffffff810f4d23>] ? file_read_actor+0x39/0x12a
 [<ffffffff810f6d60>] generic_file_aio_read+0x3fd/0x655
 [<ffffffff81016224>] ? __cycles_2_ns+0xe/0x3a
 [<ffffffff8108202e>] ? local_clock+0x14/0x4c
 [<ffffffff81145359>] do_sync_read+0xbf/0xff
 [<ffffffff812282b9>] ? security_file_permission+0x2e/0x33
 [<ffffffff811456cc>] ? rw_verify_area+0xb6/0xd3
 [<ffffffff81145a59>] vfs_read+0xac/0xf3
 [<ffffffff81146fe1>] ? fget_light+0x97/0xa2
 [<ffffffff81145be5>] sys_pread64+0x5d/0x79
 [<ffffffff814ef702>] system_call_fastpath+0x16/0x1b
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ