Sorry for the blind panic - restarting the volume seems to have fixed it. But then my next question - why is this necessary? Surely it undermines the whole point of a high availability system? Cheers, Kingsley. On Mon, 2015-08-10 at 10:53 +0100, Kingsley wrote: > Hi, > > We have a 4 way replicated volume using gluster 3.6.3 on CentOS 7. > > Over the weekend I did a yum update on each of the bricks in turn, but > now when clients (using fuse mounts) try to access the volume, it hangs. > Gluster itself wasn't updated (we've disabled that repo so that we keep > to 3.6.3 for now). > > This was what I did: > > * on first brick, "yum update" > * reboot brick > * watch "gluster volume status" on another brick and wait for it > to say all 4 bricks are online before proceeding to update the > next brick > > I was expecting the clients might pause 30 seconds while they notice a > brick is offline, but then recover. > > I've tried re-mounting clients, but that hasn't helped. > > I can't see much data in any of the log files. > > I've tried "gluster volume heal callrec" but it doesn't seem to have > helped. > > What shall I do next? > > I've pasted some stuff below in case any of it helps. > > Cheers, > Kingsley. > > [root@gluster1b-1 ~]# gluster volume info callrec > > Volume Name: callrec > Type: Replicate > Volume ID: a39830b7-eddb-4061-b381-39411274131a > Status: Started > Number of Bricks: 1 x 4 = 4 > Transport-type: tcp > Bricks: > Brick1: gluster1a-1:/data/brick/callrec > Brick2: gluster1b-1:/data/brick/callrec > Brick3: gluster2a-1:/data/brick/callrec > Brick4: gluster2b-1:/data/brick/callrec > Options Reconfigured: > performance.flush-behind: off > [root@gluster1b-1 ~]# > > > [root@gluster1b-1 ~]# gluster volume status callrec > Status of volume: callrec > Gluster process Port Online Pid > ------------------------------------------------------------------------------ > Brick gluster1a-1:/data/brick/callrec 49153 Y 6803 > Brick gluster1b-1:/data/brick/callrec 49153 Y 2614 > Brick gluster2a-1:/data/brick/callrec 49153 Y 2645 > Brick gluster2b-1:/data/brick/callrec 49153 Y 4325 > NFS Server on localhost 2049 Y 2769 > Self-heal Daemon on localhost N/A Y 2789 > NFS Server on gluster2a-1 2049 Y 2857 > Self-heal Daemon on gluster2a-1 N/A Y 2814 > NFS Server on 88.151.41.100 2049 Y 6833 > Self-heal Daemon on 88.151.41.100 N/A Y 6824 > NFS Server on gluster2b-1 2049 Y 4428 > Self-heal Daemon on gluster2b-1 N/A Y 4387 > > Task Status of Volume callrec > ------------------------------------------------------------------------------ > There are no active volume tasks > > [root@gluster1b-1 ~]# > > > [root@gluster1b-1 ~]# gluster volume heal callrec info > Brick gluster1a-1.dns99.co.uk:/data/brick/callrec/ > /to_process - Possibly undergoing heal > > Number of entries: 1 > > Brick gluster1b-1.dns99.co.uk:/data/brick/callrec/ > Number of entries: 0 > > Brick gluster2a-1.dns99.co.uk:/data/brick/callrec/ > /to_process - Possibly undergoing heal > > Number of entries: 1 > > Brick gluster2b-1.dns99.co.uk:/data/brick/callrec/ > Number of entries: 0 > > [root@gluster1b-1 ~]# > > > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://www.gluster.org/mailman/listinfo/gluster-users > _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users