Re: GFS = filesystem consistency error

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, 2006-03-03 at 09:13 +1100, oly wrote:
> Hi there,
> I would like to give an update to my ticket. That will maybe help people
> who've got similar trouble :
> I resolved my problem by doing:
> - gfs_tool shrink /home (supposed to reclaim but did not)
> - gfs_tool reclaim /home (still not enough )
> unmount the /home on all my nodes
> -gfs_fsck -y /dev/etherd/e0.0
> -remount my /home 
> VICTORY = i lost all the broken inode files
> ADVICE= avoid 1 million file folder in the future

Sorry, this is probably a late reply but out of curiosity ..

Look to me that the filesystem had been corrupted before you unlinked
the file. Is there any other errors *before* the filesystem consistency
errors ? Out of memory warning ? How much memory had you put on this
machine ? And could I assume that was an i686 machine ?

-- Wendy 


> 
> 
> On Wed, 2006-03-01 at 14:02 +1100, oly wrote:
> > Hi there
> >         I've got a 4nodes RHEL4 cluster with GFS version  6.1.0 (built
> >         Jun  7
> >         2005 12:46:04).
> >         The shared disk is a NAS detected by aoe as /dev/etherd/e0.0.
> >         ANd i have problem on few files on teh file system : if i tried
> >         to
> >         modify the inodes o this files (delete the file, or unlink the
> >         inode)
> >         the cluster nodes where i launch the command lost the GFS and
> >         the GFS
> >         modules stay busy and cannot be remove from the kernel. my nodes
> >         is so
> >         stuck and the only solution is only to hardware restart this
> >         nodes.
> >          All the GFS journal seems to work fine ...i can even get stat
> >         of the
> >         DEAD file.
> >          Is GFS got problem to manipulate file in a 'more than 1 million
> >         files'
> >         folder ?
> >          IS anyone got a solution to remove this dead files or delete
> >         teh fodler
> >         that content all these dead files ?
> >          Is a gfs.fsck can resolv my problem ?
> >          Is there any later version that fix this problem ?
> >         
> >         Thanks in advance.
> >         PS = see below all the details
> >          
> >         The error i get when i try to unlink the file inode:
> >         ===========ERROR============
> >         GFS: fsid=entcluster:sataide.2: fatal: filesystem consistency
> >         error
> >         GFS: fsid=entcluster:sataide.2:   inode = 8516674/8516674
> >         GFS: fsid=entcluster:sataide.2:   function = gfs_change_nlink
> >         GFS: fsid=entcluster:sataide.2:   file
> >         = /usr/src/build/574067-i686/BUILD/smp/src/gfs/inode.c, line =
> >         843
> >         GFS: fsid=entcluster:sataide.2:   time = 1141080134
> >         GFS: fsid=entcluster:sataide.2: about to withdraw from the
> >         cluster
> >         GFS: fsid=entcluster:sataide.2: waiting for outstanding I/O
> >         GFS: fsid=entcluster:sataide.2: telling LM to withdraw
> >         lock_dlm: withdraw abandoned memory
> >         GFS: fsid=entcluster:sataide.2: withdrawn
> >           mh_magic = 0x01161970
> >           mh_type = 4
> >           mh_generation = 68
> >           mh_format = 400
> >           mh_incarn = 6
> >           no_formal_ino = 8516674
> >           no_addr = 8516674
> >           di_mode = 0664
> >           di_uid = 500
> >           di_gid = 500
> >           di_nlink = 0
> >           di_size = 0
> >           di_blocks = 1
> >           di_atime = 1141042636
> >           di_mtime = 1140001370
> >           di_ctime = 1140001370
> >           di_major = 0
> >           di_minor = 0
> >           di_rgrp = 8513987
> >           di_goal_rgrp = 8513987
> >           di_goal_dblk = 2682
> >           di_goal_mblk = 2682
> >           di_flags = 0x00000004
> >           di_payload_format = 0
> >           di_type = 1
> >           di_height = 0
> >           di_incarn = 0
> >           di_pad = 0
> >           di_depth = 0
> >           di_entries = 0
> >           no_formal_ino = 0
> >           no_addr = 0
> >           di_eattr = 0
> >           di_reserved =
> >         00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
> >         00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
> >         00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
> >         00 00 00 00 00 00 00 00
> >         ========END OF ERROR==========
> >         
> >         My cman status:
> >         ==========STATUS============
> >         Protocol version: 5.0.1
> >         Config version: 4
> >         Cluster name: entcluster
> >         Cluster ID: 42548
> >         Cluster Member: Yes
> >         Membership state: Cluster-Member
> >         Nodes: 4
> >         Expected_votes: 1
> >         Total_votes: 4
> >         Quorum: 3
> >         Active subsystems: 5
> >         Node name: XXX.domainX.tld
> >         Node addresses: x.x.x.x
> >         ========END CMAN=========
> >         
> >         My gfs_tool df :
> >         ============DF=========
> >         /home:
> >           SB lock proto = "lock_dlm"
> >           SB lock table = "entcluster:sataide"
> >           SB ondisk format = 1309
> >           SB multihost format = 1401
> >           Block size = 4096
> >           Journals = 4
> >           Resource Groups = 274
> >           Mounted lock proto = "lock_dlm"
> >           Mounted lock table = "entcluster:sataide"
> >           Mounted host data = ""
> >           Journal number = 0
> >           Lock module flags =
> >           Local flocks = FALSE
> >           Local caching = FALSE
> >           Oopses OK = FALSE
> >         
> >           Type           Total          Used           Free
> >         use%
> >         
> >         ------------------------------------------------------------------------
> >           inodes         100642         100642         0
> >         100%
> >           metadata       3842538        8527           3834011        0%
> >           data           13999476       2760327        11239149
> >         20%
> >         =============END DF =========
> >         Version of my modules :
> >         ========modules========
> >         CMAN 2.6.9-36.0 (built May 31 2005 12:15:02) installed
> >         DLM 2.6.9-34.0 (built Jun  2 2005 15:17:56) installed
> >         Lock_Harness 2.6.9-35.5 (built Jun  7 2005 12:42:30) installed
> >         GFS 2.6.9-35.5 (built Jun  7 2005 12:42:49) installed
> >         aoe: aoe_init: AoE v2.6-11 initialised.
> >         Lock_DLM (built Jun  7 2005 12:42:32) installed
> >         ========end modules========
> >         
> >         
> >         
> >         -- 
> >         Aurelien Lemaire (oly)
> >         http://www.squiz.net
> >         Sydney | Canberra | London
> >         92 Jarrett St Leichhardt, Sydney, NSW 2040
> >         T:+61 2 9568 6866 
> >         F:+61 2 9568 6733    
> > 
> > --
> > 
> > Linux-cluster@xxxxxxxxxx
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> 
> --
> 
> Linux-cluster@xxxxxxxxxx
> https://www.redhat.com/mailman/listinfo/linux-cluster

--

Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux