On Mon, Jun 20, 2011 at 08:18:52AM +1000, Dave Chinner wrote: > > % touch /d1/tmp/foo > > touch: cannot touch `/d1/tmp/foo': No space left on device > > % df /d1 > > Filesystem 1K-blocks Used Available Use% Mounted on > > /dev/mapper/vg0-d1 943616000 904690332 38925668 96% /d1 > > Problems like this will occur if you run your filesystem at > 85-90% > full for extented periods.... Ah, yes, that's definitely been the case. I grow the filesystem when it hits 95% utilization or thereabouts. Hadn't realized that's such an awful use case for xfs. > > % df -i /d1 > > Filesystem Inodes IUsed IFree IUse% Mounted on > > /dev/mapper/vg0-d1 167509008 11806336 155702672 8% /d1 > > % sudo xfs_growfs -n /d1 > > meta-data=/dev/mapper/vg0-d1 isize=256 agcount=18, agsize=13107200 blks > > = sectsz=512 attr=2 > > data = bsize=4096 blocks=235929600, imaxpct=25 > > = sunit=0 swidth=0 blks > > naming =version 2 bsize=4096 ascii-ci=0 > > log =internal bsize=4096 blocks=25600, version=2 > > = sectsz=512 sunit=0 blks, lazy-count=1 > > realtime =none extsz=4096 blocks=0, rtextents=0 > > % grep d1 /proc/mounts > > /dev/mapper/vg0-d1 /d1 xfs rw,relatime,attr2,noquota 0 0 > > > > Obviously I'm missing something, but what? > > Most likely is that you have no contiguous free space large enough > to create a new inode chunk. using xfs_db to dump the freespace > size histogram will tell you if this is the case or not. % sudo xfs_db -c freesp /dev/vg0/d1 from to extents blocks pct 1 1 168504 168504 1.71 2 3 446 1135 0.01 4 7 5550 37145 0.38 8 15 49159 524342 5.33 16 31 1383 29223 0.30 2097152 4194303 1 2931455 29.78 4194304 8388607 1 6150953 62.49 I don't really grok that output. -andy _______________________________________________ xfs mailing list xfs@xxxxxxxxxxx http://oss.sgi.com/mailman/listinfo/xfs