On 08/10/2015 03:35 PM, Kingsley wrote: > Sorry for the blind panic - restarting the volume seems to have fixed > it. > > But then my next question - why is this necessary? Surely it undermines > the whole point of a high availability system? > > Cheers, > Kingsley. > > On Mon, 2015-08-10 at 10:53 +0100, Kingsley wrote: >> Hi, >> >> We have a 4 way replicated volume using gluster 3.6.3 on CentOS 7. >> >> Over the weekend I did a yum update on each of the bricks in turn, but >> now when clients (using fuse mounts) try to access the volume, it hangs. What does mount log file say when you tried to access the volume? Can you attach the mount log file? >> Gluster itself wasn't updated (we've disabled that repo so that we keep >> to 3.6.3 for now). >> >> This was what I did: >> >> * on first brick, "yum update" >> * reboot brick >> * watch "gluster volume status" on another brick and wait for it >> to say all 4 bricks are online before proceeding to update the >> next brick >> >> I was expecting the clients might pause 30 seconds while they notice a >> brick is offline, but then recover. >> >> I've tried re-mounting clients, but that hasn't helped. >> >> I can't see much data in any of the log files. >> >> I've tried "gluster volume heal callrec" but it doesn't seem to have >> helped. >> >> What shall I do next? >> >> I've pasted some stuff below in case any of it helps. >> >> Cheers, >> Kingsley. >> >> [root@gluster1b-1 ~]# gluster volume info callrec >> >> Volume Name: callrec >> Type: Replicate >> Volume ID: a39830b7-eddb-4061-b381-39411274131a >> Status: Started >> Number of Bricks: 1 x 4 = 4 >> Transport-type: tcp >> Bricks: >> Brick1: gluster1a-1:/data/brick/callrec >> Brick2: gluster1b-1:/data/brick/callrec >> Brick3: gluster2a-1:/data/brick/callrec >> Brick4: gluster2b-1:/data/brick/callrec >> Options Reconfigured: >> performance.flush-behind: off >> [root@gluster1b-1 ~]# >> >> >> [root@gluster1b-1 ~]# gluster volume status callrec >> Status of volume: callrec >> Gluster process Port Online Pid >> ------------------------------------------------------------------------------ >> Brick gluster1a-1:/data/brick/callrec 49153 Y 6803 >> Brick gluster1b-1:/data/brick/callrec 49153 Y 2614 >> Brick gluster2a-1:/data/brick/callrec 49153 Y 2645 >> Brick gluster2b-1:/data/brick/callrec 49153 Y 4325 >> NFS Server on localhost 2049 Y 2769 >> Self-heal Daemon on localhost N/A Y 2789 >> NFS Server on gluster2a-1 2049 Y 2857 >> Self-heal Daemon on gluster2a-1 N/A Y 2814 >> NFS Server on 88.151.41.100 2049 Y 6833 >> Self-heal Daemon on 88.151.41.100 N/A Y 6824 >> NFS Server on gluster2b-1 2049 Y 4428 >> Self-heal Daemon on gluster2b-1 N/A Y 4387 >> >> Task Status of Volume callrec >> ------------------------------------------------------------------------------ >> There are no active volume tasks >> >> [root@gluster1b-1 ~]# >> >> >> [root@gluster1b-1 ~]# gluster volume heal callrec info >> Brick gluster1a-1.dns99.co.uk:/data/brick/callrec/ >> /to_process - Possibly undergoing heal >> >> Number of entries: 1 >> >> Brick gluster1b-1.dns99.co.uk:/data/brick/callrec/ >> Number of entries: 0 >> >> Brick gluster2a-1.dns99.co.uk:/data/brick/callrec/ >> /to_process - Possibly undergoing heal >> >> Number of entries: 1 >> >> Brick gluster2b-1.dns99.co.uk:/data/brick/callrec/ >> Number of entries: 0 >> >> [root@gluster1b-1 ~]# >> >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users@xxxxxxxxxxx >> http://www.gluster.org/mailman/listinfo/gluster-users >> > > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://www.gluster.org/mailman/listinfo/gluster-users > -- ~Atin _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users