On 19.04.2017 14:12, Carlos Maiolino wrote: > Hi, > > On Wed, Apr 19, 2017 at 12:58:05PM +0200, Michael Weissenbacher wrote: >> Hi List! >> I have a storage server which primarily does around 15-20 parallel >> rsync's, nothing special. Sometimes (3-4 times a day) i notice that all >> I/O on the file system suddenly comes to a halt and the only process >> that continues to do any I/O (according to iotop) is the process >> xfsaild/md127. When this happens, xfsaild only does reads (according to >> iotop) and consistently in D State (according to top). >> Unfortunately this can sometimes stay like this for 5-15 minutes. During >> this time even a simple "ls" our "touch" would block and be stuck in D >> state. All other running processes accessing the fs are of course also >> stuck in D state. It is a XFS V5 filesystem. >> Then again, as sudden as it began, everything goes back to normal and >> I/O continues. The problem is accompanied with several "process blocked >> for xxx seconds" in dmesg and also some dropped connections due to >> network timeouts. >> >> I've tried several things to remedy the problem, including: >> - changing I/O schedulers (tried noop, deadline and cfq). Deadline >> seems to be best (the block goes away in less time compared with the >> others). >> - removing all mount options (defaults + usrquota, grpquota) >> - upgrading to the latest 4.11.0-rc kernel (before that i was on 4.9.x) >> >> Nothing of the above seemed to have made a significant change to the >> problem. >> >> xfs_info output of the fs in question: >> meta-data=/dev/md127 isize=512 agcount=33, >> agsize=268435440 blks >> = sectsz=4096 attr=2, projid32bit=1 >> = crc=1 finobt=1 spinodes=0 rmapbt=0 >> = reflink=0 >> data = bsize=4096 blocks=8789917696, imaxpct=10 >> = sunit=16 swidth=96 blks >> naming =version 2 bsize=4096 ascii-ci=0 ftype=1 >> log =internal bsize=4096 blocks=521728, version=2 >> = sectsz=4096 sunit=1 blks, lazy-count=1 >> realtime =none extsz=4096 blocks=0, rtextents=0 >> > > This is really not enough to give any idea of what might be happening, although > this looks more like a slow storage while xfsaild is flushing the log, but we > really need more information to try to give a better idea of what is going on, > please look at: > > http://xfs.org/index.php/XFS_FAQ#Q:_What_information_should_I_include_when_reporting_a_problem.3F > > Specially for: storage layout (RAID arrays, LVMs, thin provisioning, etc), and > the dmesg output with the traces from the hang tasks. > Ok, including all of the info above: * Kernel Vesion: 4.11.0-rc5 (currently), but also reproducable on 4.9.20+ * xfsprogs version: 4.9.0 * number of cpus: 1 socket, 6 cores * contents of /proc/meminfo, /proc/mounts, /proc/partitions: see below * RAID layout: Hardware RAID-10 + linear MD device (sdb == md127); XFS on top of md127 md127 : active linear sdb[0] 35159670784 blocks super 1.2 64k rounding * LVM: no LVM * type of disks 12x 6TB SAS disks (enterprise class) * write cache state of disks: disabled * write cache of BBWC controller: enabled, 2GB * xfs_info output: already included * dmesg output: samples see below * iostat output: samples see below * vmstat output: will capture when it occurs the next time * D state processes: will capture when it occurs the next time; also see top output below >> Storage Subsystem: Dell Perc H730P Controller 2GB NVCACHE, 12 6TB Disks, >> RAID-10, latest Firmware Updates >> >> I would be happy to dig out more information if needed. How can i find >> out if the RAID Controller itself gets stuck? Nothing bad shows up in >> the hardware and SCSI controller logs. >> > # cat /proc/meminfo MemTotal: 65886208 kB MemFree: 3490480 kB MemAvailable: 41531280 kB Buffers: 84 kB Cached: 26125688 kB SwapCached: 64 kB Active: 20010544 kB Inactive: 6142828 kB Active(anon): 856 kB Inactive(anon): 27576 kB Active(file): 20009688 kB Inactive(file): 6115252 kB Unevictable: 0 kB Mlocked: 0 kB SwapTotal: 64511996 kB SwapFree: 64505852 kB Dirty: 8 kB Writeback: 0 kB AnonPages: 27536 kB Mapped: 14000 kB Shmem: 832 kB Slab: 27037624 kB SReclaimable: 20569360 kB SUnreclaim: 6468264 kB KernelStack: 2976 kB PageTables: 6228 kB NFS_Unstable: 0 kB Bounce: 0 kB WritebackTmp: 0 kB CommitLimit: 97455100 kB Committed_AS: 210904 kB VmallocTotal: 34359738367 kB VmallocUsed: 0 kB VmallocChunk: 0 kB HardwareCorrupted: 0 kB AnonHugePages: 0 kB ShmemHugePages: 0 kB ShmemPmdMapped: 0 kB HugePages_Total: 0 HugePages_Free: 0 HugePages_Rsvd: 0 HugePages_Surp: 0 Hugepagesize: 2048 kB DirectMap4k: 88820 kB DirectMap2M: 7155712 kB DirectMap1G: 61865984 kB # cat /proc/mounts sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0 proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0 udev /dev devtmpfs rw,relatime,size=10240k,nr_inodes=8233000,mode=755 0 0 devpts /dev/pts devpts rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000 0 0 tmpfs /run tmpfs rw,nosuid,noexec,relatime,size=6588624k,mode=755 0 0 /dev/sda4 / xfs rw,relatime,attr2,inode64,usrquota,grpquota 0 0 tmpfs /run/lock tmpfs rw,nosuid,nodev,noexec,relatime,size=5120k 0 0 pstore /sys/fs/pstore pstore rw,relatime 0 0 tmpfs /run/shm tmpfs rw,nosuid,nodev,noexec,relatime,size=26079640k 0 0 /dev/sda2 /boot ext2 rw,relatime,block_validity,barrier,user_xattr,acl 0 0 -> /dev/md127 /backup xfs rw,noatime,attr2,inode64,sunit=128,swidth=768,usrquota,grpquota 0 0 none /sys/kernel/config configfs rw,relatime 0 0 debugfs /sys/kernel/debug debugfs rw,relatime 0 0 # cat /proc/partitions major minor #blocks name 8 0 292421632 sda 8 1 1024 sda1 8 2 1021952 sda2 8 3 64512000 sda3 8 4 226884608 sda4 8 16 35159801856 sdb 9 127 35159670784 md127 top - 12:30:03 up 11 days, 7:07, 1 user, load average: 22.89, 19.86, 12.53 Tasks: 247 total, 1 running, 246 sleeping, 0 stopped, 0 zombie %Cpu(s): 0.0 us, 0.4 sy, 0.0 ni, 99.6 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st KiB Mem: 65886208 total, 62923220 used, 2962988 free, 84 buffers KiB Swap: 64511996 total, 6144 used, 64505852 free. 38382400 cached Mem PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 1609 root 20 0 0 0 0 D 2.0 0.0 54:39.51 xfsaild/md127 1703883 root 20 0 0 0 0 D 0.0 0.0 2:28.31 kworker/u12:0 1704625 checker+ 20 0 37512 11700 2416 D 0.0 0.0 1:19.68 rsync 1704627 checker+ 20 0 67104 9396 1628 D 0.0 0.0 1:10.44 rsync 1704845 10003 20 0 9148 3836 1744 D 0.0 0.0 0:33.56 rm 1705051 nospam.+ 20 0 24392 2240 1616 D 0.0 0.0 0:22.67 rsync 1705262 atacama+ 20 0 20740 9120 2436 D 0.0 0.0 0:11.27 rsync 1705265 10021 20 0 38144 11404 3800 D 0.0 0.0 0:00.06 rsnapshot 1705371 smtp1.n+ 20 0 7088 1892 1604 D 0.0 0.0 0:06.39 rm 1705373 kalahar+ 20 0 38144 11440 3840 D 0.0 0.0 0:00.07 rsnapshot 1705475 cvs.net+ 20 0 17248 4996 2332 D 0.0 0.0 0:04.66 rsync 1705584 hoth.ne+ 20 0 38040 11488 3888 D 0.0 0.0 0:00.06 rsnapshot 1705880 10035 20 0 38120 11372 3768 D 0.0 0.0 0:00.07 rsnapshot 1705983 smtp2.n+ 20 0 38144 11444 3844 D 0.0 0.0 0:00.07 rsnapshot 1706080 endor.n+ 20 0 38124 11456 3844 D 0.0 0.0 0:00.07 rsnapshot 1706081 10031 20 0 38040 11400 3800 D 0.0 0.0 0:00.06 rsnapshot 1706181 10032 20 0 38144 11380 3780 D 0.0 0.0 0:00.07 rsnapshot 1706182 10004 20 0 38144 11384 3780 D 0.0 0.0 0:00.06 rsnapshot 1706388 dns2.ne+ 20 0 38120 11376 3768 D 0.0 0.0 0:00.06 rsnapshot 1706485 10034 20 0 38144 11392 3788 D 0.0 0.0 0:00.06 rsnapshot 1706491 10022 20 0 38124 11372 3768 D 0.0 0.0 0:00.06 rsnapshot 1706786 wiki.ne+ 20 0 38124 11404 3800 D 0.0 0.0 0:00.06 rsnapshot iostat output when the problem occurs: 04/19/2017 12:30:17 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.03 0.00 0.17 0.00 0.00 99.80 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 0.00 190.40 0.00 3046.40 0.00 32.00 0.98 5.16 5.16 0.00 5.14 97.92 04/19/2017 12:30:22 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.00 0.00 0.20 0.00 0.00 99.80 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 0.00 159.80 0.00 2556.80 0.00 32.00 0.98 6.15 6.15 0.00 6.14 98.16 04/19/2017 12:30:27 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.03 0.00 0.23 0.03 0.00 99.70 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 0.00 235.80 0.00 3772.80 0.00 32.00 0.97 4.13 4.13 0.00 4.13 97.36 04/19/2017 12:30:32 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.93 0.00 1.17 0.07 0.00 97.83 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 0.00 161.00 3.40 2576.00 54.40 32.00 0.98 5.98 6.10 0.24 5.98 98.24 04/19/2017 12:30:37 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.03 0.00 0.20 0.00 0.00 99.77 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 0.00 194.80 0.00 3116.80 0.00 32.00 0.98 5.02 5.02 0.00 5.02 97.84 04/19/2017 12:30:42 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.00 0.00 0.20 0.00 0.00 99.80 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 0.00 192.20 0.00 3075.20 0.00 32.00 0.99 5.14 5.14 0.00 5.14 98.72 ... (after it suddendly stops blocking and starts writing again) 04/19/2017 12:31:37 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.03 0.00 2.51 3.38 0.00 94.08 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 4911.80 55.60 4664.00 889.60 137985.60 58.85 122.10 25.18 8.95 25.37 0.20 94.80 04/19/2017 12:31:42 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.03 0.00 0.81 7.86 0.00 91.30 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 2239.80 0.00 2592.20 0.00 70144.80 54.12 271.25 105.12 0.00 105.12 0.39 100.00 04/19/2017 12:31:47 PM avg-cpu: %user %nice %system %iowait %steal %idle 1.14 0.00 1.57 7.90 0.00 89.39 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 2620.40 0.00 2632.20 0.00 77292.00 58.73 270.84 102.98 0.00 102.98 0.38 100.00 04/19/2017 12:31:52 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.00 0.00 1.11 7.82 0.00 91.08 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 2592.80 0.00 2661.40 0.00 76693.60 57.63 270.95 101.39 0.00 101.39 0.38 100.00 04/19/2017 12:32:32 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.03 0.00 0.87 7.43 0.00 91.66 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.00 2007.60 0.00 2321.80 0.00 58285.60 50.21 271.18 116.05 0.00 116.05 0.43 100.00 04/19/2017 12:32:37 PM avg-cpu: %user %nice %system %iowait %steal %idle 0.80 0.00 4.09 13.70 0.00 81.41 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0.20 1234.00 6.40 1912.40 43.20 66422.40 69.28 224.64 118.18 135.38 118.13 0.52 100.00 ... followed by "normal" I/O with both reads and writes 04/19/2017 12:32:42 PM avg-cpu: %user %nice %system %iowait %steal %idle 3.98 0.00 10.23 35.92 0.00 49.87 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 2.00 0.00 423.20 374.80 9460.00 66360.00 190.03 18.15 25.78 47.33 1.44 1.25 100.00 04/19/2017 12:32:47 PM avg-cpu: %user %nice %system %iowait %steal %idle 9.26 0.00 7.95 56.36 0.00 26.43 Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 24.60 0.00 1826.00 188.00 53184.80 5222.40 58.00 17.24 8.25 9.08 0.21 0.50 100.00 with kind regards, Michael -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html