Hi, On Sat 25-07-09 17:17:52, Sylvain Rochet wrote: > Sorry for the late answer, waiting for the problem to happen again ;) No problem. > On Thu, Jul 16, 2009 at 07:27:49PM +0200, Jan Kara wrote: > > Hi, > > > > > We(TuxFamily) are having some inodes corruptions on a NFS server. > > > > > > So, let's start with the facts. > > > > > > > > > ==== NFS Server > > > > > > Linux bazooka 2.6.28.9 #1 SMP Mon Mar 30 12:58:22 CEST 2009 x86_64 GNU/Linux > > > > Can you still see the corruption with 2.6.30 kernel? > Not upgraded yet, we'll give a try. > > > If you can still see this problem, could you run: debugfs /dev/md10 > > and send output of the command: > > stat <40420228> > > (or whatever the corrupted inode number will be) > > and also: > > dump <40420228> /tmp/corrupted_dir > > One inode get corrupted recently, here is the output: > > root@bazooka:/data/web/ed/90/48/walotux.walon.org/htdocs/tmp/cache/e# ls -lai > total 64 > 88539836 drwxr-sr-x 2 18804 23084 4096 2009-07-25 07:53 . > 88539821 drwxr-sr-x 20 18804 23084 4096 2008-08-20 10:14 .. > 88541578 -rw-rw-rw- 1 18804 23084 471 2009-07-25 04:55 -inc_forum-10-wa.3cb1921f > 88541465 -rw-rw-rw- 1 18804 23084 6693 2009-07-25 07:53 -inc_rss_item-32-wa.23d91cc2 > 88541471 -rw-rw-rw- 1 18804 23084 1625 2009-07-25 07:53 -inc_rubriques-17-wa.f2f152f0 > 88541549 -rw-rw-rw- 1 18804 23084 2813 2009-07-25 03:04 INDEX-.edfac52c > 88541366 -rw-rw-rw- 1 18804 23084 0 2008-08-17 20:44 .ok > ? ?--------- ? ? ? ? ? spip%3Farticle19.f8740dca > 88541671 -rw-rw-rw- 1 18804 23084 5619 2009-07-24 21:07 spip%3Fauteur1.c64f7f7e > 88541460 -rw-rw-rw- 1 18804 23084 5636 2009-07-24 19:30 spip%3Fmot5.f3e9adda > 88540284 -rw-rw-rw- 1 18804 23084 3802 2009-07-25 16:10 spip%3Fpage%3Dforum-30.63b2c1b1 > 88541539 -rw-rw-rw- 1 18804 23084 12972 2009-07-25 11:14 spip%3Fpage%3Djquery.cce608b6.gz OK, so we couldn't stat a directory... > root@bazooka:/data/web/ed/90/48/walotux.walon.org/htdocs/tmp/cache/e# cat spip%3Farticle19.f8740dca > cat: spip%3Farticle19.f8740dca: Stale NFS file handle This is probably the misleading output from ext3_iget(). It should give you EIO in the latest kernel. > root@bazooka:~# debugfs /dev/md10 > debugfs 1.40-WIP (14-Nov-2006) > > debugfs: stat <88539836> > > Inode: 88539836 Type: directory Mode: 0755 Flags: 0x0 Generation: 791796957 > User: 18804 Group: 23084 Size: 4096 > File ACL: 0 Directory ACL: 0 > Links: 2 Blockcount: 8 > Fragment: Address: 0 Number: 0 Size: 0 > ctime: 0x4a6a9dd5 -- Sat Jul 25 07:53:25 2009 > atime: 0x4a0de585 -- Fri May 15 23:58:29 2009 > mtime: 0x4a6a9dd5 -- Sat Jul 25 07:53:25 2009 > Size of extra inode fields: 4 > BLOCKS: > (0):177096928 > TOTAL: 1 > > > debugfs: ls <88539836> > > 88539836 (12) . 88539821 (32) .. 88541366 (12) .ok > 88541465 (56) -inc_rss_item-32-wa.23d91cc2 > 88541539 (40) spip%3Fpage%3Djquery.cce608b6.gz > 88540284 (40) spip%3Fpage%3Dforum-30.63b2c1b1 > 88541460 (28) spip%3Fmot5.f3e9adda > 88541471 (160) -inc_rubriques-17-wa.f2f152f0 > 88541549 (24) INDEX-.edfac52c 88541578 (284) -inc_forum-10-wa.3cb1921f > 88541562 (36) spip%3Farticle19.f8740dca > 88541671 (3372) spip%3Fauteur1.c64f7f7e The directory itself looks fine... > debugfs: stat <88541562> > > Inode: 88541562 Type: regular Mode: 0666 Flags: 0x0 Generation: 860068541 > User: 18804 Group: 23084 Size: 0 > File ACL: 0 Directory ACL: 0 > Links: 0 Blockcount: 0 > Fragment: Address: 0 Number: 0 Size: 0 > ctime: 0x4a6a8fac -- Sat Jul 25 06:53:00 2009 > atime: 0x4a6a612f -- Sat Jul 25 03:34:39 2009 > mtime: 0x4a6a8fac -- Sat Jul 25 06:53:00 2009 > dtime: 0x4a6a8fac -- Sat Jul 25 06:53:00 2009 > Size of extra inode fields: 4 > BLOCKS: Ah, OK, here's the problem. The directory points to a file which is obviously deleted (note the "Links: 0"). All the content of the inode seems to indicate that the file was correctly deleted (you might check that the corresponding bit in the bitmap is cleared via: "icheck 88541562"). The question is how it could happen the directory still points to the inode. Really strange. It looks as if we've lost a write to the directory but I don't see how. Are there any suspitious kernel messages in this case? > debugfs: dump <88539836> /tmp/corrupted_dir > > (file attached) > > > > You might want to try disabling the DIR_INDEX feature and see whether > > the corruption still occurs... > > We'll try. It probably won't help. This particular directory had just one block so DIR_INDEX had no effect on it. > > > Keeping inodes into servers' cache seems to prevent the problem to happen. > > > ( yeah, # while true ; do ionice -c3 find /data -size +0 > /dev/null ; done ) > > > > I'd guess just because they don't have to be read from disk where they > > get corrupted. > > Exactly. > > > > Interesting, but it may well be just by the way how these files get > > created / updated. > > Yes, this is only because of that. > > Additional data that may help, we replaced the storage server to > something slower (less number of CPU, less number of cores, ...). We are > still getting some corruption but with non-common sense with the former > server. > > The data are stored on two storage arrays of disks. The primary one is > made of fiber-channel disks used through a simple fiber-channel card, > RAID soft with md, raid6. The secondary one is made of SCSI disks used > through a RAID-hard card. We got corruption on both, depending on > the one currently used into production. OK, so it's probably not a storage device problem. Good to know. Honza -- Jan Kara <jack@xxxxxxx> SUSE Labs, CR -- To unsubscribe from this list: send the line "unsubscribe linux-ext4" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html