lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Message-ID: <1485820466.2851367230407574.JavaMail.root@shiva> Date: Mon, 29 Apr 2013 11:13:27 +0100 (GMT+01:00) From: luvar@...intext.sk To: Dmitry Monakhov <dmonakhov@...nvz.org> Cc: linux-ext4@...r.kernel.org, Zheng Liu <gnehzuil.liu@...il.com> Subject: Re: sustained write to disk, frozen copy Hi, here are additional data: luvar@...cktroja ~ $ cat /proc/mounts rootfs / rootfs rw 0 0 proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0 sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0 udev /dev tmpfs rw,nosuid,relatime,size=10240k,mode=755 0 0 devpts /dev/pts devpts rw,relatime,gid=5,mode=620 0 0 /dev/md2 / ext3 rw,noatime,errors=continue,commit=5,barrier=1,data=ordered 0 0 tmpfs /run tmpfs rw,nosuid,nodev,relatime,mode=755 0 0 rc-svcdir /lib64/rc/init.d tmpfs rw,nosuid,nodev,noexec,relatime,size=1024k,mode=755 0 0 configfs /sys/kernel/config configfs rw,nosuid,nodev,noexec,relatime 0 0 cgroup_root /sys/fs/cgroup tmpfs rw,nosuid,nodev,noexec,relatime,size=10240k,mode=755 0 0 cpuset /sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0 cpu /sys/fs/cgroup/cpu cgroup rw,nosuid,nodev,noexec,relatime,cpu 0 0 cpuacct /sys/fs/cgroup/cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpuacct 0 0 fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0 shm /dev/shm tmpfs rw,nosuid,nodev,noexec,relatime 0 0 /dev/mapper/vg-portage /usr/portage ext2 rw,noatime,errors=continue 0 0 /dev/mapper/vg-distfiles /usr/portage/distfiles ext2 rw,noatime,errors=continue 0 0 /dev/mapper/vg-home /home ext3 rw,noatime,errors=continue,commit=5,barrier=1,data=ordered 0 0 none /var/tmp/portage tmpfs rw,noatime,size=8388608k,nr_inodes=2097152 0 0 /dev/mapper/vg-svrkiNfs /mnt/nfs/svrki ext4 rw,noatime,stripe=48,data=ordered 0 0 /dev/mapper/vg-music /var/lib/mpd/music/local ext4 rw,noatime,stripe=48,data=ordered 0 0 /dev/sda2 /home/luvar/.m2 btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspace btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspaceGanz btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspaceMisc btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspaceOpen btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/eclipseWorkspaceAndroid btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/.mozilla btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/documents/trackit btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/documents/plaintext btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/programs/nexus-oss-webapp-bundle btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/sda2 /home/luvar/programs/android-sdk-linux btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0 /dev/mapper/vg-skolaDokumenty /home/luvar/documents/skola btrfs rw,noatime,thread_pool=4,compress=zlib,nospace_cache 0 0 /dev/mapper/vg-laciNilfs /mnt/nfs/laciNilfs nilfs2 rw,noatime 0 0 /dev/mapper/vg-boincPartition /var/lib/boinc ext4 rw,noatime,stripe=64 0 0 /dev/mapper/vg-postgresql91 /var/lib/postgresql/9.1 reiserfs rw,noatime 0 0 /dev/mapper/vg-yacy /home/luvar/programs/yacy/yacy_partition btrfs rw,noatime,thread_pool=1,noacl,nospace_cache 0 0 rpc_pipefs /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0 nfsd /proc/fs/nfsd nfsd rw,nosuid,nodev,noexec,relatime 0 0 /dev/mapper/vg-fotodvdbackup /mnt/fotodvdbackup ext2 rw,relatime,errors=continue 0 0 /dev/mapper/vg-pogamut /mnt/pogamut ext4 rw,noatime,stripe=48,data=ordered 0 0 /dev/mapper/vg-films /mnt/films ext4 rw,noatime,stripe=48,data=ordered 0 0 /dev/loop0 /mnt/cdrom iso9660 ro,relatime 0 0 /dev/sdg1 /mnt/usbstick vfat rw,relatime,uid=1000,gid=1000,fmask=0022,dmask=0022,codepage=cp437,iocharset=utf8,shortname=mixed,errors=remount-ro 0 0 luvar@...cktroja ~ $ cat /proc/mdstat Personalities : [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] md1 : active raid1 sdc1[0] sdf1[3] sde1[2] sdd1[1] sdb1[4] 1469824 blocks [5/5] [UUUUU] md3 : active raid5 sdc3[0] sdf3[3] sde3[2] sdd3[1] sdb3[4] 3068960768 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU] bitmap: 2/6 pages [8KB], 65536KB chunk md2 : active raid10 sdc2[3] sdf2[2] sde2[1] sdd2[0] 25398272 blocks 512K chunks 2 near-copies [4/4] [UUUU] unused devices: <none> lvm config is available on http://dawn.ynet.sk/~luvar/asdf/vgcfgbackup Thanks, PS:blacktroja luvar # uptime 12:12:58 up 8 days, 16:45, 12 users, load average: 5.95, 5.64, 5.75 Still writing... ----- "Dmitry Monakhov" <dmonakhov@...nvz.org> wrote: > On Sun, 28 Apr 2013 20:17:03 +0100 (GMT+01:00), LuVar > <luvar@...intext.sk> wrote: > > Fuf... Here are my deadlock things: > > > Strange looks md3_raid5 stuck ? > Can you please post your /proc/mounts, /proc/mdstat and lvm config > > SysRq : Show Blocked State > > task PC stack pid father > > md3_raid5 D 0000000000000001 0 16779 2 0x00000000 > > ffff88032f76fb70 0000000000000046 ffff88032f76e000 > 0000000000010c80 > > ffff88032f144890 0000000000010c80 ffff88032f76ffd8 > 0000000000004000 > > ffff88032f76ffd8 0000000000010c80 ffff8803330bb470 > ffff88032f144890 > > Call Trace: > > [<ffffffff81355e2f>] ? __blk_run_queue+0x16/0x18 > > [<ffffffff81358abe>] ? blk_queue_bio+0x29a/0x2b4 > > [<ffffffff81356546>] ? generic_make_request+0x97/0xda > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff8143bfa5>] md_super_wait+0x68/0x80 > > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > > [<ffffffff8144160f>] write_page+0x1d5/0x2be > > [<ffffffff81441365>] bitmap_update_sb+0x115/0x117 > > [<ffffffff8143c27c>] md_update_sb+0x2bf/0x467 > > [<ffffffff814ebab1>] ? __schedule+0x6b8/0x7be > > [<ffffffff8143ca00>] md_check_recovery+0x26b/0x5ff > > [<ffffffffa04a3624>] raid5d+0x1f/0x4c8 [raid456] > > [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 > > [<ffffffff81034cee>] ? del_timer_sync+0x3c/0x48 > > [<ffffffff814e9fdd>] ? schedule_timeout+0x189/0x1a9 > > [<ffffffff8143a69c>] md_thread+0xfd/0x11b > > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > > [<ffffffff8143a59f>] ? md_register_thread+0xc8/0xc8 > > [<ffffffff8104189d>] kthread+0x84/0x8c > > [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 > > [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d > > [<ffffffff814ee310>] ? gs_change+0xb/0xb > > jbd2/dm-3-8 D 0000000000000002 0 17714 2 0x00000000 > > ffff88032f647bb0 0000000000000046 ffff88032f646000 > 0000000000010c80 > > ffff880330a70440 0000000000010c80 ffff88032f647fd8 > 0000000000004000 > > ffff88032f647fd8 0000000000010c80 ffff8801a6c3e100 > ffff880330a70440 > > Call Trace: > > [<ffffffff810dae71>] ? __find_get_block_slow+0x113/0x12a > > [<ffffffff81438706>] ? md_make_request+0xc4/0x1b9 > > [<ffffffff810597da>] ? ktime_get_ts+0xa9/0xb5 > > [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39 > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff814ebd15>] io_schedule+0x87/0xca > > [<ffffffff810db7b6>] sleep_on_buffer+0x9/0xd > > [<ffffffff814ea18f>] __wait_on_bit+0x43/0x76 > > [<ffffffff814ea22b>] out_of_line_wait_on_bit+0x69/0x74 > > [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39 > > [<ffffffff81041ce0>] ? autoremove_wake_function+0x34/0x34 > > [<ffffffff810db772>] __wait_on_buffer+0x21/0x23 > > [<ffffffff8118acfb>] jbd2_journal_commit_transaction+0xd19/0x1182 > > [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b > > [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 > > [<ffffffff8118daf5>] kjournald2+0xc6/0x22e > > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > > [<ffffffff8118da2f>] ? commit_timeout+0xb/0xb > > [<ffffffff8104189d>] kthread+0x84/0x8c > > [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 > > [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d > > [<ffffffff814ee310>] ? gs_change+0xb/0xb > > flush-253:3 D ffff8803314e0024 0 6471 2 0x00000000 > > ffff88011fedda50 0000000000000046 ffff88011fedc000 > 0000000000010c80 > > ffff880130e02b90 0000000000010c80 ffff88011feddfd8 > 0000000000004000 > > ffff88011feddfd8 0000000000010c80 ffffffff81671410 > ffff880130e02b90 > > Call Trace: > > [<ffffffff8104b325>] ? try_to_wake_up+0x20a/0x21c > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff8118d7ce>] jbd2_log_wait_commit+0xc1/0x113 > > [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25 > > [<ffffffff8118ebae>] jbd2_journal_force_commit_nested+0x6a/0x7c > > [<ffffffff8115dd54>] ext4_da_writepages+0x376/0x447 > > [<ffffffff81095d18>] do_writepages+0x1e/0x27 > > [<ffffffff810d5c56>] __writeback_single_inode.clone.24+0x3a/0xda > > [<ffffffff810d6132>] writeback_sb_inodes+0x1b8/0x2f7 > > [<ffffffff810ba680>] ? put_super+0x20/0x2b > > [<ffffffff810d62de>] __writeback_inodes_wb+0x6d/0xab > > [<ffffffff810d641f>] wb_writeback+0x103/0x194 > > [<ffffffff810d6b48>] wb_do_writeback+0x111/0x16d > > [<ffffffff810d6c29>] bdi_writeback_thread+0x85/0x14a > > [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d > > [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d > > [<ffffffff8104189d>] kthread+0x84/0x8c > > [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10 > > [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d > > [<ffffffff814ee310>] ? gs_change+0xb/0xb > > kio_file D ffffffffffffffff 0 6546 6443 0x00000000 > > ffff880270a87b98 0000000000000082 ffff880270a86000 > 0000000000010c80 > > ffff8800aa600c10 0000000000010c80 ffff880270a87fd8 > 0000000000004000 > > ffff880270a87fd8 0000000000010c80 ffff8803330bb470 > ffff8800aa600c10 > > Call Trace: > > [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b > > [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83 > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff814e9fd5>] schedule_timeout+0x181/0x1a9 > > [<ffffffff8103492a>] ? run_timer_softirq+0x1ef/0x1ef > > [<ffffffff814ebf5b>] io_schedule_timeout+0x93/0xe4 > > [<ffffffff8138176e>] ? __percpu_counter_sum+0x4d/0x63 > > [<ffffffff8109597c>] > balance_dirty_pages_ratelimited_nr+0x54d/0x615 > > [<ffffffff810d7fca>] generic_file_splice_write+0x11e/0x130 > > [<ffffffff810d7acc>] do_splice_from+0x7d/0x8a > > [<ffffffff810d7af4>] direct_splice_actor+0x1b/0x1d > > [<ffffffff810d7dfb>] splice_direct_to_actor+0xd5/0x186 > > [<ffffffff810d7ad9>] ? do_splice_from+0x8a/0x8a > > [<ffffffff810d8e01>] do_splice_direct+0x47/0x5a > > [<ffffffff810b8f63>] do_sendfile+0x12e/0x1c3 > > [<ffffffff810b9bee>] sys_sendfile64+0x54/0x92 > > [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b > > sync D ffff88033089f3f0 0 6707 6477 0x00000000 > > ffff8801b9651d08 0000000000000086 ffff8801b9650000 > 0000000000010c80 > > ffff88033089f3f0 0000000000010c80 ffff8801b9651fd8 > 0000000000004000 > > ffff8801b9651fd8 0000000000010c80 ffff8803330ba0c0 > ffff88033089f3f0 > > Call Trace: > > [<ffffffff8108dc75>] ? find_get_pages_tag+0xf3/0x12f > > [<ffffffff81096d23>] ? release_pages+0x19c/0x1ab > > [<ffffffff81096691>] ? pagevec_lookup_tag+0x20/0x29 > > [<ffffffff814ebc8c>] schedule+0x5f/0x61 > > [<ffffffff814e9e7a>] schedule_timeout+0x26/0x1a9 > > [<ffffffff81049017>] ? check_preempt_curr+0x3e/0x6c > > [<ffffffff814eb2e6>] wait_for_common+0xc8/0x13f > > [<ffffffff8104b337>] ? try_to_wake_up+0x21c/0x21c > > [<ffffffff810d9745>] ? __sync_filesystem+0x7a/0x7a > > [<ffffffff814eb3f7>] wait_for_completion+0x18/0x1a > > [<ffffffff810d657f>] writeback_inodes_sb_nr+0xb8/0xc1 > > [<ffffffff810d6602>] writeback_inodes_sb+0x22/0x29 > > [<ffffffff810d971c>] __sync_filesystem+0x51/0x7a > > [<ffffffff810d9756>] sync_one_sb+0x11/0x13 > > [<ffffffff810bb6fa>] iterate_supers+0x68/0xb8 > > [<ffffffff810d9695>] sync_filesystems+0x1b/0x1d > > [<ffffffff810d97ba>] sys_sync+0x17/0x33 > > [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b > > > > Is there something eslse, what should I do before reboot? > > > > LuVar > > > > ----- "Zheng Liu" <gnehzuil.liu@...il.com> wrote: > > > > > On Sat, Apr 27, 2013 at 08:16:02AM +0100, LuVar wrote: > > > > Hi, > > > > I have my desktop about 24 hours in "deadlock". I was copying > (as > > > root in krusader from USB key (mounted as [1]) data to filesystem > [2]) > > > some files from one point to another. Now it is more than 24 > hours > > > with sustained disk write, see [3]. > > > > > > > > How can I help and "debug" this problem? I have 3.5.7 gentoo > kernel > > > ([4]). > > > > > > Hi LuVar, > > > > > > You could use 'echo w >/proc/sysrq-trigger' to look at which > process > > > has > > > been deadlock. > > > > > > # echo w >/proc/sysrq-trigger (WARN: you need a root privilege) > > > # dmesg | vim - > > > > > > SysRq : Show Blocked State > > > task PC stack pid father > > > > > > Here is nothing because my system hasn't any deadlock. Then you > > > could > > > use 'echo t >/proc/sysrq-trigger' to dump current tasks and their > > > information. That would be great if you could paste these details > in > > > mailing list. They are very useful for us to dig this problem. > > > > > > Thanks, > > > - Zheng > > > > > > > > > > > PS: I am an average user, so please by verbose to me. > > > > > > > > [1] sudo mount -o rw,uid=luvar,gid=luvar,iocharset=utf8 > /dev/sdg1 > > > /mnt/usbstick/ > > > > > > > > [2]: > > > > luvar@...cktroja ~ $ mount | grep music > > > > /dev/mapper/vg-music on /var/lib/mpd/music/local type ext4 > > > (rw,noatime,commit=0) > > > > > > > > [3]: > > > > iotop, two first records : > > > > 17714 be/3 root 0.00 B/s 0.00 B/s 0.00 % 97.60 % > > > [jbd2/dm-3-8] > > > > 6546 be/4 root 0.00 B/s 0.00 B/s 0.00 % 93.48 % > > > kdeinit4: kio_file [kdeinit] file > local:/tmp/ksocket-root/~-socket > > > local:/tmp/ksocket-root/krusaderZz6431.slave-socket > > > > > > > > [4]: > > > > luvar@...cktroja ~ $ uname -a > > > > Linux blacktroja 3.5.7-gentoo #1 SMP Sun Oct 28 17:18:07 CET > 2012 > > > x86_64 Intel(R) Core(TM) i7 CPU 930 @ 2.80GHz GenuineIntel > GNU/Linux > > > > > > > > Thanks, LuVar > > > > -- > > > > To unsubscribe from this list: send the line "unsubscribe > > > linux-ext4" in > > > > the body of a message to majordomo@...r.kernel.org > > > > More majordomo info at > http://vger.kernel.org/majordomo-info.html > > -- > > To unsubscribe from this list: send the line "unsubscribe > linux-ext4" in > > the body of a message to majordomo@...r.kernel.org > > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-ext4" in the body of a message to majordomo@...r.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists