On 05/27/2015 07:33 PM, Ted Miller wrote: > responses below > Ted Miller > > On 5/26/2015 12:01 AM, Atin Mukherjee wrote: >> >> On 05/26/2015 03:12 AM, Ted Miller wrote: >>> From: Niels de Vos <ndevos@xxxxxxxxxx> >>> Sent: Monday, May 25, 2015 4:44 PM >>> >>> On Mon, May 25, 2015 at 06:49:26PM +0000, Ted Miller wrote: >>>> ________________________________ >>>> From: Humble Devassy Chirammal <humble.devassy@xxxxxxxxx> >>>> Sent: Monday, May 18, 2015 9:37 AM >>>> Hi All, >>>> >>>> GlusterFS 3.7.0 RPMs for RHEL, CentOS, Fedora and packages for >>>> Debian are available at >>>> download.gluster.org<http://download.gluster.org> [1]. >>>> >>>> [1] http://download.gluster.org/pub/gluster/glusterfs/3.7/3.7.0/ >>>> >>>> --Humble >>>> >>>> >>>> On Thu, May 14, 2015 at 2:49 PM, Vijay Bellur >>>> <vbellur@xxxxxxxxxx<mailto:vbellur@xxxxxxxxxx>> wrote: >>>> >>>> Hi All, >>>> >>>> I am happy to announce that Gluster 3.7.0 is now generally >>>> available. 3.7.0 contains several >>>> >>>> [snip] >>>> >>>> Cheers, >>>> Vijay >>>> >>>> [snip] >>>> >>>> What happened to packages for RHEL/Centos 5? I have the (probably >>>> unusual--added gluster to existing servers) setup of running a replica >>>> 3 cluster where two nodes run on Centos 6 and one is still on Centos >>>> 5. This is a personal setup, and I have been using >>>> http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-5/x86_64/repodata/repomod.xml >>>> >>>> as my repo. It has worked fine for a while, but this time the two >>>> Centos 6 nodes updated to 3.7, but the Centos 5 node got left behind >>>> at 3.6.3. >>> Packages for RHEL/CentOS-5 are not available yet. These will follow >>> later. Thare are some changes needed to be able to build the packages on >>> EL5. Because we are currently stabilizing our CI/regression tests, we do >>> not merge any other changes. Until we provide packages in our >>> repository, you could apply patch http://review.gluster.org/10803 >>> yourself and build the EL5 version. I expect that we will do a release >>> in 2-3 weeks which will have EL5 RPMs too. >>> >>> I have no idea about the problem below, it sounds like something the >>> GlusterD developers could help with. >>> >>> Niels >>> >>>> Command 'gluster volume status' on the C5 machine makes everything >>>> look fine: >>>> >>>> Status of volume: ISO2 >>>> Gluster process Port >>>> Online Pid >>>> ------------------------------------------------------------------------------ >>>> >>>> Brick 10.x.x.2:/bricks/01/iso2 49162 >>>> Y 4679 >>>> Brick 10.x.x.4:/bricks/01/iso2 49183 >>>> Y 6447 >>>> Brick 10.x.x.9:/bricks/01/iso2 49169 >>>> Y 1985 >>>> >>>> But the same command on either of the C6 machines shows the C5 machine >>>> (10.x.x.2) missing in action (though it does recognize that there are >>>> NFS and heal daemons there): >>>> >>>> Status of volume: ISO2 >>>> Gluster process TCP Port RDMA Port >>>> Online Pid >>>> ------------------------------------------------------------------------------ >>>> >>>> Brick 10.41.65.4:/bricks/01/iso2 49183 0 >>>> Y 6447 >>>> Brick 10.41.65.9:/bricks/01/iso2 49169 0 >>>> Y 1985 >>>> NFS Server on localhost 2049 0 >>>> Y 2279 >>>> Self-heal Daemon on localhost N/A N/A >>>> Y 2754 >>>> NFS Server on 10.41.65.2 2049 0 >>>> Y 4757 >>>> Self-heal Daemon on 10.41.65.2 N/A N/A >>>> Y 4764 >>>> NFS Server on 10.41.65.4 2049 0 >>>> Y 6543 >>>> Self-heal Daemon on 10.41.65.4 N/A N/A >>>> Y 6551 >>>> >>>> So, is this just an oversight (I hope), or has support for C5 been >>>> dropped? >>>> If support for C5 is gone, how do I downgrade my Centos6 machines back >>>> to 3.6.x? (I know how to change the repo, but the actual sequence of >>>> yum commands and gluster commands is unknown to me). >> Could you attach the glusterd log file of 10.x.x.2 machine > attached as etc-glusterfs-glusterd.vol.log.newer.2, starting from last > machine reboot >> and the node from where you triggered volume status. > attached as etc-glusterfs-glusterd.vol.log.newer4 starting same time as > .2 log >> Could you also share gluster volume info output of all the nodes? > I have several volumes, so I chose the one that shows up first on the > listings: > > *from 10.41.65.2:* > > [root@office2 /var/log/glusterfs]$ gluster volume info > > Volume Name: ISO2 > Type: Replicate > Volume ID: 090da4b3-c666-41fe-8283-2c029228b3f7 > Status: Started > Number of Bricks: 1 x 3 = 3 > Transport-type: tcp > Bricks: > Brick1: 10.41.65.2:/bricks/01/iso2 > Brick2: 10.41.65.4:/bricks/01/iso2 > Brick3: 10.41.65.9:/bricks/01/iso2 > > [root@office2 /var/log/glusterfs]$ gluster volume status ISO2 > Status of volume: ISO2 > Gluster process Port Online Pid > ------------------------------------------------------------------------------ > > Brick 10.41.65.2:/bricks/01/iso2 49162 Y 4463 > Brick 10.41.65.4:/bricks/01/iso2 49183 Y 6447 > Brick 10.41.65.9:/bricks/01/iso2 49169 Y 1985 > NFS Server on localhost 2049 Y 4536 > Self-heal Daemon on localhost N/A Y 4543 > NFS Server on 10.41.65.9 2049 Y 2279 > Self-heal Daemon on 10.41.65.9 N/A Y 2754 > NFS Server on 10.41.65.4 2049 Y 6543 > Self-heal Daemon on 10.41.65.4 N/A Y 6551 > > Task Status of Volume ISO2 > ------------------------------------------------------------------------------ > > There are no active volume tasks > > [root@office2 ~]$ gluster peer status > Number of Peers: 2 > > Hostname: 10.41.65.9 > Uuid: cf2ae9c7-833e-4a73-a996-e72158011c69 > State: Peer in Cluster (Connected) > > Hostname: 10.41.65.4 > Uuid: bd3ca8b7-f2da-44ce-8739-c0db5e40158c > State: Peer in Cluster (Connected) > > > *from 10.41.65.4:* > > [root@office4b ~]# gluster volume info ISO2 > > Volume Name: ISO2 > Type: Replicate > Volume ID: 090da4b3-c666-41fe-8283-2c029228b3f7 > Status: Started > Number of Bricks: 1 x 3 = 3 > Transport-type: tcp > Bricks: > Brick1: 10.41.65.2:/bricks/01/iso2 > Brick2: 10.41.65.4:/bricks/01/iso2 > Brick3: 10.41.65.9:/bricks/01/iso2 > > [root@office4b ~]# gluster volume status ISO2 > Status of volume: ISO2 > Gluster process TCP Port RDMA Port Online > Pid > ------------------------------------------------------------------------------ > > Brick 10.41.65.4:/bricks/01/iso2 49183 0 Y 6447 > Brick 10.41.65.9:/bricks/01/iso2 49169 0 Y 1985 > NFS Server on localhost 2049 0 Y 6543 > Self-heal Daemon on localhost N/A N/A Y 6551 > NFS Server on 10.41.65.2 2049 0 Y 4536 > Self-heal Daemon on 10.41.65.2 N/A N/A Y 4543 > NFS Server on 10.41.65.9 2049 0 Y 2279 > Self-heal Daemon on 10.41.65.9 N/A N/A Y 2754 > > Task Status of Volume ISO2 > ------------------------------------------------------------------------------ > > There are no active volume tasks > > [root@office4b ~]# gluster peer status > Number of Peers: 2 > > Hostname: 10.41.65.2 > Uuid: 4a53ed8b-2b41-4a3c-acf7-2dabec431f57 > State: Peer in Cluster (Connected) > > Hostname: 10.41.65.9 > Uuid: cf2ae9c7-833e-4a73-a996-e72158011c69 > State: Peer in Cluster (Connected) > > > *from 10.41.65.9:* > > [root@office9 ~]$ gluster volume info ISO2 > > Volume Name: ISO2 > Type: Replicate > Volume ID: 090da4b3-c666-41fe-8283-2c029228b3f7 > Status: Started > Number of Bricks: 1 x 3 = 3 > Transport-type: tcp > Bricks: > Brick1: 10.41.65.2:/bricks/01/iso2 > Brick2: 10.41.65.4:/bricks/01/iso2 > Brick3: 10.41.65.9:/bricks/01/iso2 > [root@office9 ~]$ gluster volume status ISO2 > Status of volume: ISO2 > Gluster process TCP Port RDMA Port Online Pid > ------------------------------------------------------------------------------ > > Brick 10.41.65.4:/bricks/01/iso2 49183 0 Y 6447 > Brick 10.41.65.9:/bricks/01/iso2 49169 0 Y 1985 > NFS Server on localhost 2049 0 Y 2279 > Self-heal Daemon on localhost N/A N/A Y 2754 > NFS Server on 10.41.65.2 2049 0 Y 4536 > Self-heal Daemon on 10.41.65.2 N/A N/A Y 4543 > NFS Server on 10.41.65.4 2049 0 Y 6543 > Self-heal Daemon on 10.41.65.4 N/A N/A Y 6551 > > Task Status of Volume ISO2 > ------------------------------------------------------------------------------ > > There are no active volume tasks > > [root@office9 ~]$ gluster peer status > Number of Peers: 2 > > Hostname: 10.41.65.2 > Uuid: 4a53ed8b-2b41-4a3c-acf7-2dabec431f57 > State: Peer in Cluster (Connected) > > Hostname: 10.41.65.4 > Uuid: bd3ca8b7-f2da-44ce-8739-c0db5e40158c > State: Peer in Cluster (Connected) I think you just pasted snippet of the log sequence, its hard to identify anything from it. Would it be possible for you to kill and restart glusterd with glusterd -LDEBUG at 2 & 4 and share the complete log file for both of them? >>>> Ted Miller >>>> Elkhart, IN, USA >>> >>> Thanks for the information. As long as I know it is coming, I can >>> improvise and hang on. >>> >>> I am assuming that the problem with the .2 machine not being seen is >>> a result of running a cluster with a version split. >>> >>> Ted Miller >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users@xxxxxxxxxxx >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> > -- ~Atin _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users