On Fri, 2006-03-03 at 09:13 +1100, oly wrote: > Hi there, > I would like to give an update to my ticket. That will maybe help people > who've got similar trouble : > I resolved my problem by doing: > - gfs_tool shrink /home (supposed to reclaim but did not) > - gfs_tool reclaim /home (still not enough ) > unmount the /home on all my nodes > -gfs_fsck -y /dev/etherd/e0.0 > -remount my /home > VICTORY = i lost all the broken inode files > ADVICE= avoid 1 million file folder in the future Sorry, this is probably a late reply but out of curiosity .. Look to me that the filesystem had been corrupted before you unlinked the file. Is there any other errors *before* the filesystem consistency errors ? Out of memory warning ? How much memory had you put on this machine ? And could I assume that was an i686 machine ? -- Wendy > > > On Wed, 2006-03-01 at 14:02 +1100, oly wrote: > > Hi there > > I've got a 4nodes RHEL4 cluster with GFS version 6.1.0 (built > > Jun 7 > > 2005 12:46:04). > > The shared disk is a NAS detected by aoe as /dev/etherd/e0.0. > > ANd i have problem on few files on teh file system : if i tried > > to > > modify the inodes o this files (delete the file, or unlink the > > inode) > > the cluster nodes where i launch the command lost the GFS and > > the GFS > > modules stay busy and cannot be remove from the kernel. my nodes > > is so > > stuck and the only solution is only to hardware restart this > > nodes. > > All the GFS journal seems to work fine ...i can even get stat > > of the > > DEAD file. > > Is GFS got problem to manipulate file in a 'more than 1 million > > files' > > folder ? > > IS anyone got a solution to remove this dead files or delete > > teh fodler > > that content all these dead files ? > > Is a gfs.fsck can resolv my problem ? > > Is there any later version that fix this problem ? > > > > Thanks in advance. > > PS = see below all the details > > > > The error i get when i try to unlink the file inode: > > ===========ERROR============ > > GFS: fsid=entcluster:sataide.2: fatal: filesystem consistency > > error > > GFS: fsid=entcluster:sataide.2: inode = 8516674/8516674 > > GFS: fsid=entcluster:sataide.2: function = gfs_change_nlink > > GFS: fsid=entcluster:sataide.2: file > > = /usr/src/build/574067-i686/BUILD/smp/src/gfs/inode.c, line = > > 843 > > GFS: fsid=entcluster:sataide.2: time = 1141080134 > > GFS: fsid=entcluster:sataide.2: about to withdraw from the > > cluster > > GFS: fsid=entcluster:sataide.2: waiting for outstanding I/O > > GFS: fsid=entcluster:sataide.2: telling LM to withdraw > > lock_dlm: withdraw abandoned memory > > GFS: fsid=entcluster:sataide.2: withdrawn > > mh_magic = 0x01161970 > > mh_type = 4 > > mh_generation = 68 > > mh_format = 400 > > mh_incarn = 6 > > no_formal_ino = 8516674 > > no_addr = 8516674 > > di_mode = 0664 > > di_uid = 500 > > di_gid = 500 > > di_nlink = 0 > > di_size = 0 > > di_blocks = 1 > > di_atime = 1141042636 > > di_mtime = 1140001370 > > di_ctime = 1140001370 > > di_major = 0 > > di_minor = 0 > > di_rgrp = 8513987 > > di_goal_rgrp = 8513987 > > di_goal_dblk = 2682 > > di_goal_mblk = 2682 > > di_flags = 0x00000004 > > di_payload_format = 0 > > di_type = 1 > > di_height = 0 > > di_incarn = 0 > > di_pad = 0 > > di_depth = 0 > > di_entries = 0 > > no_formal_ino = 0 > > no_addr = 0 > > di_eattr = 0 > > di_reserved = > > 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 > > 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 > > 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 > > 00 00 00 00 00 00 00 00 > > ========END OF ERROR========== > > > > My cman status: > > ==========STATUS============ > > Protocol version: 5.0.1 > > Config version: 4 > > Cluster name: entcluster > > Cluster ID: 42548 > > Cluster Member: Yes > > Membership state: Cluster-Member > > Nodes: 4 > > Expected_votes: 1 > > Total_votes: 4 > > Quorum: 3 > > Active subsystems: 5 > > Node name: XXX.domainX.tld > > Node addresses: x.x.x.x > > ========END CMAN========= > > > > My gfs_tool df : > > ============DF========= > > /home: > > SB lock proto = "lock_dlm" > > SB lock table = "entcluster:sataide" > > SB ondisk format = 1309 > > SB multihost format = 1401 > > Block size = 4096 > > Journals = 4 > > Resource Groups = 274 > > Mounted lock proto = "lock_dlm" > > Mounted lock table = "entcluster:sataide" > > Mounted host data = "" > > Journal number = 0 > > Lock module flags = > > Local flocks = FALSE > > Local caching = FALSE > > Oopses OK = FALSE > > > > Type Total Used Free > > use% > > > > ------------------------------------------------------------------------ > > inodes 100642 100642 0 > > 100% > > metadata 3842538 8527 3834011 0% > > data 13999476 2760327 11239149 > > 20% > > =============END DF ========= > > Version of my modules : > > ========modules======== > > CMAN 2.6.9-36.0 (built May 31 2005 12:15:02) installed > > DLM 2.6.9-34.0 (built Jun 2 2005 15:17:56) installed > > Lock_Harness 2.6.9-35.5 (built Jun 7 2005 12:42:30) installed > > GFS 2.6.9-35.5 (built Jun 7 2005 12:42:49) installed > > aoe: aoe_init: AoE v2.6-11 initialised. > > Lock_DLM (built Jun 7 2005 12:42:32) installed > > ========end modules======== > > > > > > > > -- > > Aurelien Lemaire (oly) > > http://www.squiz.net > > Sydney | Canberra | London > > 92 Jarrett St Leichhardt, Sydney, NSW 2040 > > T:+61 2 9568 6866 > > F:+61 2 9568 6733 > > > > -- > > > > Linux-cluster@xxxxxxxxxx > > https://www.redhat.com/mailman/listinfo/linux-cluster > > -- > > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster -- Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster