On Wed, Aug 27, 2014 at 04:53:17PM +0800, Zhang Qiang wrote: > 2014-08-26 21:13 GMT+08:00 Dave Chinner <david@xxxxxxxxxxxxx>: > > > On Tue, Aug 26, 2014 at 06:04:52PM +0800, Zhang Qiang wrote: > > > Thanks Dave/Greg for your analysis and suggestions. > > > > > > I can summarize what I should do next: > > > > > > - backup my data using xfsdump > > > - rebuilt filesystem using mkfs with options: agcount=32 for 2T disk > > > - mount filesystem with option inode64,nobarrier > > > > Ok up to here. > > > > > - applied patches about adding free list inode on disk structure > > > > No, don't do that. You're almost certain to get it wrong and corrupt > > your filesysetms and lose data. > > > > > As we have about ~100 servers need back up, so that will take much > > effort, > > > do you have any other suggestion? > > > > Just remount them with inode64. Nothing else. Over time as you add > > and remove files the inodes will redistribute across all 4 AGs. > > > OK. > > How I can see the layout number of inodes on each AGs? Here's my checking > steps: > > 1) Check unmounted file system first: > [root@fstest data1]# xfs_db -c "sb 0" -c "p" /dev/sdb1 |egrep > 'icount|ifree' > icount = 421793920 > ifree = 41 > [root@fstest data1]# xfs_db -c "sb 1" -c "p" /dev/sdb1 |egrep > 'icount|ifree' > icount = 0 > ifree = 0 That's wrong. You need to check the AGI headers, not the superblock. Only the primary superblock gets updated, and it's the aggregated of all the AGI values, not the per AG values. And, BTW, that's *421 million* inodes in that filesystem. Almost twice as many as the filesystem you started showing problems on... > OK, this is a social networking website back end servers, actually the CDN > infrastructure, and different server located different cities. > We have a global sync script to make all these 100 servers have the same > data. > > For each server we use RAID10 and XFS (CentOS6.3). > > There are about 3M files (50K in size) generated every day, and we track > the path of each files in database. I'd suggest you are overestimating the size of the files being storedi by an order of magnitude: 200M files at 50k in size is 10TB, not 1.5TB. But you've confirmed exactly what I thought - you're using the filesystem as an anonymous object store for hundreds of millions of small objects and that's exactly the situation I'd expect to see these problems.... > Do you have any suggestions to improve our solution? TANSTAAFL. I've given you some stuff to try, worst case is reformating and recopying all the data around. I don't really have much time to do much more than that - talk to Red Hat (because you are using CentOS) if you want help with a more targeted solution to your problem... Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx _______________________________________________ xfs mailing list xfs@xxxxxxxxxxx http://oss.sgi.com/mailman/listinfo/xfs