----- Original Message ----- > From: "Gaurav Garg" <ggarg@xxxxxxxxxx> > To: "RASTELLI Alessandro" <alessandro.rastelli@xxxxxxxx> > Cc: gluster-users@xxxxxxxxxxx > Sent: Monday, May 11, 2015 4:42:59 AM > Subject: Re: GlusterD errors > > Hi Rastelli, > > Could you tell us what steps you followed or what command you executed for > getting these log. Also, what about manually running xfs_info on your brick filesystems and looking for errors. From the logs it looks like gluster can't get the inode size and xfs_info is return a non zero return code. Is it mounted? Do you see signs of FS corruption? Look in /var/log/message for XFS related errors. -b > > ~ Gaurav > > ----- Original Message ----- > From: "RASTELLI Alessandro" <alessandro.rastelli@xxxxxxxx> > To: gluster-users@xxxxxxxxxxx > Sent: Monday, May 11, 2015 2:05:16 PM > Subject: GlusterD errors > > Signature electronique > > > Hi, > > we’ve got a lot of these errors in /etc-glusterfs-glusterd.vol.log in our > Glusterfs environment. > > Just wanted to know if I can do anything about that, or if I can ignore them. > > Thank you > > > > [2015-05-11 08:22:43.848305] E > [glusterd-utils.c:7364:glusterd_add_inode_size_to_dict] 0-management: > xfs_info exited with non-zero exit status > > [2015-05-11 08:22:43.848347] E > [glusterd-utils.c:7390:glusterd_add_inode_size_to_dict] 0-management: failed > to get inode size > > [2015-05-11 08:22:52.911718] E [glusterd-op-sm.c:207:glusterd_get_txn_opinfo] > 0-: Unable to get transaction opinfo for transaction ID : > ace2f066-1acb-4e00-9cca-721f88691dce > > [2015-05-11 08:23:53.266666] E > [glusterd-syncop.c:961:_gd_syncop_commit_op_cbk] 0-management: Failed to > aggregate response from node/brick > > > > Alessandro > > > > > From: gluster-users-bounces@xxxxxxxxxxx > [mailto:gluster-users-bounces@xxxxxxxxxxx] On Behalf Of Pierre Léonard > Sent: venerdì 10 aprile 2015 16:18 > To: gluster-users@xxxxxxxxxxx > Subject: Re: one node change uuid in the night > > > > > > Hi Atin and all, > > > > > have corrected with the data in glusterd.info and suppress the bad peers > file. > Could you clarify what steps did you perform here. Also could you try to > start glusterd with -LDEBUG and share the glusterd log file with us. > Also do you see any delta in glusterd.info file between node 10 and the > other nodes? > ~Atin > > > The problem is solved. It came from a miwe of uuid file and their contents on > the 10 node. > As we said here "Ouf !" because I have vacation on next week. > > May be It could be necessary to save the peers directory, as many problem > came from their contents. > > As the log the name volfile in an error line I search on the web and found > that page : > http://www.gluster.org/community/documentation/index.php/Understanding_vol-file > > I have added some section of the example file. Is that pertinent for our 14 > node cluster or do I have to forget or change notably for the number of > threads ? > > Many thank's for all , > > > -- > > > > > > > > > > > > > Pierre Léonard > > > Senior IT Manager > > > MetaGenoPolis > > > > Pierre.Leonard@xxxxxxxxxxxx > > > Tél. : +33 (0)1 34 65 29 78 > > > > Centre de recherche INRA > > > Domaine de Vilvert – Bât. 325 R+1 > > > 78 352 Jouy-en-Josas CEDEX > > > France > > > www.mgps.eu > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://www.gluster.org/mailman/listinfo/gluster-users > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://www.gluster.org/mailman/listinfo/gluster-users _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users