Re: sustained write to disk, frozen copy

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi, here are additional data:

luvar@blacktroja ~ $ cat /proc/mounts 
rootfs / rootfs rw 0 0
proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0
sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0
udev /dev tmpfs rw,nosuid,relatime,size=10240k,mode=755 0 0
devpts /dev/pts devpts rw,relatime,gid=5,mode=620 0 0
/dev/md2 / ext3 rw,noatime,errors=continue,commit=5,barrier=1,data=ordered 0 0
tmpfs /run tmpfs rw,nosuid,nodev,relatime,mode=755 0 0
rc-svcdir /lib64/rc/init.d tmpfs rw,nosuid,nodev,noexec,relatime,size=1024k,mode=755 0 0
configfs /sys/kernel/config configfs rw,nosuid,nodev,noexec,relatime 0 0
cgroup_root /sys/fs/cgroup tmpfs rw,nosuid,nodev,noexec,relatime,size=10240k,mode=755 0 0
cpuset /sys/fs/cgroup/cpuset cgroup rw,nosuid,nodev,noexec,relatime,cpuset 0 0
cpu /sys/fs/cgroup/cpu cgroup rw,nosuid,nodev,noexec,relatime,cpu 0 0
cpuacct /sys/fs/cgroup/cpuacct cgroup rw,nosuid,nodev,noexec,relatime,cpuacct 0 0
fusectl /sys/fs/fuse/connections fusectl rw,relatime 0 0
shm /dev/shm tmpfs rw,nosuid,nodev,noexec,relatime 0 0
/dev/mapper/vg-portage /usr/portage ext2 rw,noatime,errors=continue 0 0
/dev/mapper/vg-distfiles /usr/portage/distfiles ext2 rw,noatime,errors=continue 0 0
/dev/mapper/vg-home /home ext3 rw,noatime,errors=continue,commit=5,barrier=1,data=ordered 0 0
none /var/tmp/portage tmpfs rw,noatime,size=8388608k,nr_inodes=2097152 0 0
/dev/mapper/vg-svrkiNfs /mnt/nfs/svrki ext4 rw,noatime,stripe=48,data=ordered 0 0
/dev/mapper/vg-music /var/lib/mpd/music/local ext4 rw,noatime,stripe=48,data=ordered 0 0
/dev/sda2 /home/luvar/.m2 btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/eclipseWorkspace btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/eclipseWorkspaceGanz btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/eclipseWorkspaceMisc btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/eclipseWorkspaceOpen btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/eclipseWorkspaceAndroid btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/.mozilla btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/documents/trackit btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/documents/plaintext btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/programs/nexus-oss-webapp-bundle btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/sda2 /home/luvar/programs/android-sdk-linux btrfs rw,noatime,thread_pool=6,compress=zlib,ssd,nospace_cache 0 0
/dev/mapper/vg-skolaDokumenty /home/luvar/documents/skola btrfs rw,noatime,thread_pool=4,compress=zlib,nospace_cache 0 0                                                                                                                                       
/dev/mapper/vg-laciNilfs /mnt/nfs/laciNilfs nilfs2 rw,noatime 0 0                                                                                                                                                                                              
/dev/mapper/vg-boincPartition /var/lib/boinc ext4 rw,noatime,stripe=64 0 0                                                                                                                                                                                     
/dev/mapper/vg-postgresql91 /var/lib/postgresql/9.1 reiserfs rw,noatime 0 0                                                                                                                                                                                    
/dev/mapper/vg-yacy /home/luvar/programs/yacy/yacy_partition btrfs rw,noatime,thread_pool=1,noacl,nospace_cache 0 0                                                                                                                                            
rpc_pipefs /var/lib/nfs/rpc_pipefs rpc_pipefs rw,relatime 0 0                                                                                                                                                                                                  
nfsd /proc/fs/nfsd nfsd rw,nosuid,nodev,noexec,relatime 0 0                                                                                                                                                                                                    
/dev/mapper/vg-fotodvdbackup /mnt/fotodvdbackup ext2 rw,relatime,errors=continue 0 0                                                                                                                                                                           
/dev/mapper/vg-pogamut /mnt/pogamut ext4 rw,noatime,stripe=48,data=ordered 0 0                                                                                                                                                                                 
/dev/mapper/vg-films /mnt/films ext4 rw,noatime,stripe=48,data=ordered 0 0                                                                                                                                                                                     
/dev/loop0 /mnt/cdrom iso9660 ro,relatime 0 0                                                                                                                                                                                                                  
/dev/sdg1 /mnt/usbstick vfat rw,relatime,uid=1000,gid=1000,fmask=0022,dmask=0022,codepage=cp437,iocharset=utf8,shortname=mixed,errors=remount-ro 0 0   

luvar@blacktroja ~ $ cat /proc/mdstat                                                                                                                                                                                                                          
Personalities : [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]                                                                                                                                                                                               
md1 : active raid1 sdc1[0] sdf1[3] sde1[2] sdd1[1] sdb1[4]                                                                                                                                                                                                     
      1469824 blocks [5/5] [UUUUU]                                                                                                                                                                                                                             
                                                                                                                                                                                                                                                               
md3 : active raid5 sdc3[0] sdf3[3] sde3[2] sdd3[1] sdb3[4]
      3068960768 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU]
      bitmap: 2/6 pages [8KB], 65536KB chunk

md2 : active raid10 sdc2[3] sdf2[2] sde2[1] sdd2[0]
      25398272 blocks 512K chunks 2 near-copies [4/4] [UUUU]
      
unused devices: <none>


lvm config is available on http://dawn.ynet.sk/~luvar/asdf/vgcfgbackup

Thanks,
PS:blacktroja luvar # uptime 
 12:12:58 up 8 days, 16:45, 12 users,  load average: 5.95, 5.64, 5.75
Still writing...


----- "Dmitry Monakhov" <dmonakhov@xxxxxxxxxx> wrote:

> On Sun, 28 Apr 2013 20:17:03 +0100 (GMT+01:00), LuVar
> <luvar@xxxxxxxxxxxx> wrote:
> > Fuf... Here are my deadlock things:
> > 
> Strange looks md3_raid5 stuck ?
> Can you please post your /proc/mounts, /proc/mdstat  and lvm config
> > SysRq : Show Blocked State
> >   task                        PC stack   pid father
> > md3_raid5       D 0000000000000001     0 16779      2 0x00000000
> >  ffff88032f76fb70 0000000000000046 ffff88032f76e000
> 0000000000010c80
> >  ffff88032f144890 0000000000010c80 ffff88032f76ffd8
> 0000000000004000
> >  ffff88032f76ffd8 0000000000010c80 ffff8803330bb470
> ffff88032f144890
> > Call Trace:
> >  [<ffffffff81355e2f>] ? __blk_run_queue+0x16/0x18
> >  [<ffffffff81358abe>] ? blk_queue_bio+0x29a/0x2b4
> >  [<ffffffff81356546>] ? generic_make_request+0x97/0xda
> >  [<ffffffff814ebc8c>] schedule+0x5f/0x61
> >  [<ffffffff8143bfa5>] md_super_wait+0x68/0x80
> >  [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25
> >  [<ffffffff8144160f>] write_page+0x1d5/0x2be
> >  [<ffffffff81441365>] bitmap_update_sb+0x115/0x117
> >  [<ffffffff8143c27c>] md_update_sb+0x2bf/0x467
> >  [<ffffffff814ebab1>] ? __schedule+0x6b8/0x7be
> >  [<ffffffff8143ca00>] md_check_recovery+0x26b/0x5ff
> >  [<ffffffffa04a3624>] raid5d+0x1f/0x4c8 [raid456]
> >  [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83
> >  [<ffffffff81034cee>] ? del_timer_sync+0x3c/0x48
> >  [<ffffffff814e9fdd>] ? schedule_timeout+0x189/0x1a9
> >  [<ffffffff8143a69c>] md_thread+0xfd/0x11b
> >  [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25
> >  [<ffffffff8143a59f>] ? md_register_thread+0xc8/0xc8
> >  [<ffffffff8104189d>] kthread+0x84/0x8c
> >  [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10
> >  [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d
> >  [<ffffffff814ee310>] ? gs_change+0xb/0xb
> > jbd2/dm-3-8     D 0000000000000002     0 17714      2 0x00000000
> >  ffff88032f647bb0 0000000000000046 ffff88032f646000
> 0000000000010c80
> >  ffff880330a70440 0000000000010c80 ffff88032f647fd8
> 0000000000004000
> >  ffff88032f647fd8 0000000000010c80 ffff8801a6c3e100
> ffff880330a70440
> > Call Trace:
> >  [<ffffffff810dae71>] ? __find_get_block_slow+0x113/0x12a
> >  [<ffffffff81438706>] ? md_make_request+0xc4/0x1b9
> >  [<ffffffff810597da>] ? ktime_get_ts+0xa9/0xb5
> >  [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39
> >  [<ffffffff814ebc8c>] schedule+0x5f/0x61
> >  [<ffffffff814ebd15>] io_schedule+0x87/0xca
> >  [<ffffffff810db7b6>] sleep_on_buffer+0x9/0xd
> >  [<ffffffff814ea18f>] __wait_on_bit+0x43/0x76
> >  [<ffffffff814ea22b>] out_of_line_wait_on_bit+0x69/0x74
> >  [<ffffffff810db7ad>] ? unmap_underlying_metadata+0x39/0x39
> >  [<ffffffff81041ce0>] ? autoremove_wake_function+0x34/0x34
> >  [<ffffffff810db772>] __wait_on_buffer+0x21/0x23
> >  [<ffffffff8118acfb>] jbd2_journal_commit_transaction+0xd19/0x1182
> >  [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b
> >  [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83
> >  [<ffffffff8118daf5>] kjournald2+0xc6/0x22e
> >  [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25
> >  [<ffffffff8118da2f>] ? commit_timeout+0xb/0xb
> >  [<ffffffff8104189d>] kthread+0x84/0x8c
> >  [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10
> >  [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d
> >  [<ffffffff814ee310>] ? gs_change+0xb/0xb
> > flush-253:3     D ffff8803314e0024     0  6471      2 0x00000000
> >  ffff88011fedda50 0000000000000046 ffff88011fedc000
> 0000000000010c80
> >  ffff880130e02b90 0000000000010c80 ffff88011feddfd8
> 0000000000004000
> >  ffff88011feddfd8 0000000000010c80 ffffffff81671410
> ffff880130e02b90
> > Call Trace:
> >  [<ffffffff8104b325>] ? try_to_wake_up+0x20a/0x21c
> >  [<ffffffff814ebc8c>] schedule+0x5f/0x61
> >  [<ffffffff8118d7ce>] jbd2_log_wait_commit+0xc1/0x113
> >  [<ffffffff81041cac>] ? wake_up_bit+0x25/0x25
> >  [<ffffffff8118ebae>] jbd2_journal_force_commit_nested+0x6a/0x7c
> >  [<ffffffff8115dd54>] ext4_da_writepages+0x376/0x447
> >  [<ffffffff81095d18>] do_writepages+0x1e/0x27
> >  [<ffffffff810d5c56>] __writeback_single_inode.clone.24+0x3a/0xda
> >  [<ffffffff810d6132>] writeback_sb_inodes+0x1b8/0x2f7
> >  [<ffffffff810ba680>] ? put_super+0x20/0x2b
> >  [<ffffffff810d62de>] __writeback_inodes_wb+0x6d/0xab
> >  [<ffffffff810d641f>] wb_writeback+0x103/0x194
> >  [<ffffffff810d6b48>] wb_do_writeback+0x111/0x16d
> >  [<ffffffff810d6c29>] bdi_writeback_thread+0x85/0x14a
> >  [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d
> >  [<ffffffff810d6ba4>] ? wb_do_writeback+0x16d/0x16d
> >  [<ffffffff8104189d>] kthread+0x84/0x8c
> >  [<ffffffff814ee314>] kernel_thread_helper+0x4/0x10
> >  [<ffffffff81041819>] ? kthread_freezable_should_stop+0x4d/0x4d
> >  [<ffffffff814ee310>] ? gs_change+0xb/0xb
> > kio_file        D ffffffffffffffff     0  6546   6443 0x00000000
> >  ffff880270a87b98 0000000000000082 ffff880270a86000
> 0000000000010c80
> >  ffff8800aa600c10 0000000000010c80 ffff880270a87fd8
> 0000000000004000
> >  ffff880270a87fd8 0000000000010c80 ffff8803330bb470
> ffff8800aa600c10
> > Call Trace:
> >  [<ffffffff810349e6>] ? lock_timer_base.clone.28+0x26/0x4b
> >  [<ffffffff81034ca6>] ? try_to_del_timer_sync+0x77/0x83
> >  [<ffffffff814ebc8c>] schedule+0x5f/0x61
> >  [<ffffffff814e9fd5>] schedule_timeout+0x181/0x1a9
> >  [<ffffffff8103492a>] ? run_timer_softirq+0x1ef/0x1ef
> >  [<ffffffff814ebf5b>] io_schedule_timeout+0x93/0xe4
> >  [<ffffffff8138176e>] ? __percpu_counter_sum+0x4d/0x63
> >  [<ffffffff8109597c>]
> balance_dirty_pages_ratelimited_nr+0x54d/0x615
> >  [<ffffffff810d7fca>] generic_file_splice_write+0x11e/0x130
> >  [<ffffffff810d7acc>] do_splice_from+0x7d/0x8a
> >  [<ffffffff810d7af4>] direct_splice_actor+0x1b/0x1d
> >  [<ffffffff810d7dfb>] splice_direct_to_actor+0xd5/0x186
> >  [<ffffffff810d7ad9>] ? do_splice_from+0x8a/0x8a
> >  [<ffffffff810d8e01>] do_splice_direct+0x47/0x5a
> >  [<ffffffff810b8f63>] do_sendfile+0x12e/0x1c3
> >  [<ffffffff810b9bee>] sys_sendfile64+0x54/0x92
> >  [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b
> > sync            D ffff88033089f3f0     0  6707   6477 0x00000000
> >  ffff8801b9651d08 0000000000000086 ffff8801b9650000
> 0000000000010c80
> >  ffff88033089f3f0 0000000000010c80 ffff8801b9651fd8
> 0000000000004000
> >  ffff8801b9651fd8 0000000000010c80 ffff8803330ba0c0
> ffff88033089f3f0
> > Call Trace:
> >  [<ffffffff8108dc75>] ? find_get_pages_tag+0xf3/0x12f
> >  [<ffffffff81096d23>] ? release_pages+0x19c/0x1ab
> >  [<ffffffff81096691>] ? pagevec_lookup_tag+0x20/0x29
> >  [<ffffffff814ebc8c>] schedule+0x5f/0x61
> >  [<ffffffff814e9e7a>] schedule_timeout+0x26/0x1a9
> >  [<ffffffff81049017>] ? check_preempt_curr+0x3e/0x6c
> >  [<ffffffff814eb2e6>] wait_for_common+0xc8/0x13f
> >  [<ffffffff8104b337>] ? try_to_wake_up+0x21c/0x21c
> >  [<ffffffff810d9745>] ? __sync_filesystem+0x7a/0x7a
> >  [<ffffffff814eb3f7>] wait_for_completion+0x18/0x1a
> >  [<ffffffff810d657f>] writeback_inodes_sb_nr+0xb8/0xc1
> >  [<ffffffff810d6602>] writeback_inodes_sb+0x22/0x29
> >  [<ffffffff810d971c>] __sync_filesystem+0x51/0x7a
> >  [<ffffffff810d9756>] sync_one_sb+0x11/0x13
> >  [<ffffffff810bb6fa>] iterate_supers+0x68/0xb8
> >  [<ffffffff810d9695>] sync_filesystems+0x1b/0x1d
> >  [<ffffffff810d97ba>] sys_sync+0x17/0x33
> >  [<ffffffff814ed062>] system_call_fastpath+0x16/0x1b
> > 
> > Is there something eslse, what should I do before reboot?
> > 
> > LuVar
> > 
> > ----- "Zheng Liu" <gnehzuil.liu@xxxxxxxxx> wrote:
> > 
> > > On Sat, Apr 27, 2013 at 08:16:02AM +0100, LuVar wrote:
> > > > Hi,
> > > > I have my desktop about 24 hours in "deadlock". I was copying
> (as
> > > root in krusader from USB key (mounted as [1]) data to filesystem
> [2])
> > > some files from one point to another. Now it is more than 24
> hours
> > > with sustained disk write, see [3].
> > > > 
> > > > How can I help and "debug" this problem? I have 3.5.7 gentoo
> kernel
> > > ([4]).
> > > 
> > > Hi LuVar,
> > > 
> > > You could use 'echo w >/proc/sysrq-trigger' to look at which
> process
> > > has
> > > been deadlock.
> > > 
> > >  # echo w >/proc/sysrq-trigger (WARN: you need a root privilege)
> > >  # dmesg | vim -
> > > 
> > >  SysRq : Show Blocked State
> > >    task                        PC stack   pid father
> > > 
> > > Here is nothing because my system hasn't any deadlock.  Then you
> > > could
> > > use 'echo t >/proc/sysrq-trigger' to dump current tasks and their
> > > information.  That would be great if you could paste these details
> in
> > > mailing list.  They are very useful for us to dig this problem.
> > > 
> > > Thanks,
> > >                                                 - Zheng
> > > 
> > > > 
> > > > PS: I am an average user, so please by verbose to me.
> > > > 
> > > > [1] sudo mount -o rw,uid=luvar,gid=luvar,iocharset=utf8
> /dev/sdg1
> > > /mnt/usbstick/
> > > > 
> > > > [2]:
> > > > luvar@blacktroja ~ $ mount | grep music
> > > > /dev/mapper/vg-music on /var/lib/mpd/music/local type ext4
> > > (rw,noatime,commit=0)
> > > > 
> > > > [3]:
> > > > iotop, two first records :
> > > > 17714 be/3 root        0.00 B/s    0.00 B/s  0.00 % 97.60 %
> > > [jbd2/dm-3-8]
> > > >  6546 be/4 root        0.00 B/s    0.00 B/s  0.00 % 93.48 %
> > > kdeinit4: kio_file [kdeinit] file
> local:/tmp/ksocket-root/~-socket
> > > local:/tmp/ksocket-root/krusaderZz6431.slave-socket
> > > > 
> > > > [4]:
> > > > luvar@blacktroja ~ $ uname -a
> > > > Linux blacktroja 3.5.7-gentoo #1 SMP Sun Oct 28 17:18:07 CET
> 2012
> > > x86_64 Intel(R) Core(TM) i7 CPU 930 @ 2.80GHz GenuineIntel
> GNU/Linux
> > > > 
> > > > Thanks, LuVar
> > > > --
> > > > To unsubscribe from this list: send the line "unsubscribe
> > > linux-ext4" in
> > > > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > > > More majordomo info at 
> http://vger.kernel.org/majordomo-info.html
> > --
> > To unsubscribe from this list: send the line "unsubscribe
> linux-ext4" in
> > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-ext4" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Reiser Filesystem Development]     [Ceph FS]     [Kernel Newbies]     [Security]     [Netfilter]     [Bugtraq]     [Linux FS]     [Yosemite National Park]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Samba]     [Device Mapper]     [Linux Media]

  Powered by Linux