On 6/16/17 12:10 PM, Brian Matheson wrote: > Hi all, > > I'm writing to get some information about a problem we're seeing on > our nfs servers. We're using XFS on an LVM volume backed by an LSI > RAID card (raid 6, with 24 SSDs). We're nfs exporting the volume to a > number of hypervisors. We're seeing messages like the following: > > Jun 16 09:22:30 ny2r3s1 kernel: [15259176.032579] XFS: nfsd(2301) > possible memory allocation deadlock size 68256 in kmem_alloc > (mode:0x2400240) > > These messages are followed by nfsd failures as indicated by log messages like: > > Jun 16 09:22:39 ny2r3s1 kernel: [15259184.933311] nfsd: peername > failed (err 107)! > > Dropping the caches on the box fixes the problem immediately. Based > on a little research, we thought that the problem could be occurring > due to file fragmentation, so we're running xfs_fsr periodically to > defragment. At the moment we're also periodically dropping the cache > in an attempt to prevent the problem from occurring. A better approach might be to set extent size hints on the fragmented files in question, to avoid the fragmentation in the first place. drop caches is a pretty big hammer, and xfs_fsr can have other side effects w.r.t. filesystem aging and freespace fragmentation. How badly fragmented were the files in question? -Eric > Any help appreciated, and if this query belongs on a different mailing > list, please let me know. > > > The systems are running ubuntu 14.04 with a 4.4.0 kernel (Linux > ny2r3s1 4.4.0-53-generic #74~14.04.1-Ubuntu SMP Fri Dec 2 03:43:31 UTC > 2016 x86_64 x86_64 x86_64 GNU/Linux). xfs_repair is version 3.1.9. > They have 64G of RAM, most of which is used by cache, and 12 cpu > cores. As mentioned we're using ssds connected to an lsi raid card. > xfs_info reports: > > meta-data=/dev/mapper/VMSTORAGE_SSD-XFS_VHD isize=256 agcount=62, > agsize=167772096 blks > = sectsz=512 attr=2 > data = bsize=4096 blocks=10311515136, imaxpct=5 > = sunit=64 swidth=256 blks > naming =version 2 bsize=4096 ascii-ci=0 > log =internal bsize=4096 blocks=521728, version=2 > = sectsz=512 sunit=64 blks, lazy-count=1 > realtime =none extsz=4096 blocks=0, rtextents=0 > > At the moment, slabtop reports this: > Active / Total Objects (% used) : 5543699 / 5668921 (97.8%) > Active / Total Slabs (% used) : 157822 / 157822 (100.0%) > Active / Total Caches (% used) : 77 / 144 (53.5%) > Active / Total Size (% used) : 1110436.20K / 1259304.73K (88.2%) > Minimum / Average / Maximum Object : 0.01K / 0.22K / 18.50K > > OBJS ACTIVE USE OBJ SIZE SLABS OBJ/SLAB CACHE SIZE NAME > 4382508 4382508 100% 0.10K 112372 39 449488K buffer_head > 348152 348152 100% 0.57K 12434 28 198944K radix_tree_node > 116880 83796 71% 4.00K 14610 8 467520K kmalloc-4096 > 114492 88855 77% 0.09K 2726 42 10904K kmalloc-96 > 108640 86238 79% 0.12K 3395 32 13580K kmalloc-128 > 51680 51680 100% 0.12K 1520 34 6080K kernfs_node_cache > 49536 29011 58% 0.06K 774 64 3096K kmalloc-64 > 46464 46214 99% 0.03K 363 128 1452K kmalloc-32 > 44394 34860 78% 0.19K 1057 42 8456K dentry > 40188 38679 96% 0.04K 394 102 1576K ext4_extent_status > 33150 31649 95% 0.05K 390 85 1560K ftrace_event_field > 26207 25842 98% 0.05K 359 73 1436K Acpi-Parse > 23142 20528 88% 0.38K 551 42 8816K mnt_cache > 21756 21515 98% 0.19K 518 42 4144K kmalloc-192 > 20160 20160 100% 0.07K 360 56 1440K Acpi-Operand > 19800 19800 100% 0.18K 450 44 3600K xfs_log_ticket > > Thanks much, > Brian Matheson > -- > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html