Am 21.09.2011 04:11, schrieb Dave Chinner:
How much memory does your test machine have? The performance will be
vastly different if there is enough RAM to hold the working set of
inodes and page cache (~20GB all up), and that could be one of the
factors contributing to the problems.
The livesystems which crash within hours have between 48GB and 64GB RAM.
But my testing system has only 8GB.
The above xfs_info output is from your 160GB SSD - what's the output
from the 1TB device?
The 1TB device is now doing something else and does not have XFS on it
anymore. But here are the layouts of two livesystems.
xfs_info /dev/sda6
meta-data=/dev/root isize=256 agcount=4, agsize=35767872
blks
= sectsz=512 attr=2
data = bsize=4096 blocks=143071488, imaxpct=25
= sunit=64 swidth=512 blks
naming =version 2 bsize=4096 ascii-ci=0
log =internal bsize=4096 blocks=69888, version=2
= sectsz=512 sunit=64 blks, lazy-count=1
realtime =none extsz=4096 blocks=0, rtextents=0
xfs_info /dev/sda6
meta-data=/dev/root isize=256 agcount=4, agsize=35768000
blks
= sectsz=512 attr=2
data = bsize=4096 blocks=143071774, imaxpct=25
= sunit=64 swidth=512 blks
naming =version 2 bsize=4096 ascii-ci=0
log =internal bsize=4096 blocks=32768, version=2
= sectsz=512 sunit=64 blks, lazy-count=1
realtime =none extsz=4096 blocks=0, rtextents=0
Also, what phase do you see it hanging in? the random stat phase is
terribly slow on spinning disks, so if I can avoid that it woul dbe
nice....
Creating or deleting files. never in the stat phase.
Stefan
_______________________________________________
xfs mailing list
xfs@xxxxxxxxxxx
http://oss.sgi.com/mailman/listinfo/xfs