Re: Rebooting gluster nodes make VMs pause due to storage error

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Nicolas,

Here are my experiences with GlusterFS 3.6 & 3.7 on a dual-replica set up.

First of all, setting the "network.ping-timeout" to a low value (3-5 seconds) helps in avoiding a 42 second freeze on the clients as mentioned in other threads.  This value seems to matter even if the clients are currently connected to the glusterd running on the host that's not going down.

Second, I found that there's a particular order that gluster processes need to stop in order to avoid disrupting the volume and avoid freezing on the clients that aren't connected to the glusterd host that's going down.

1) kill the glusterfs process   (/usr/bin/killall glusterfs)
2) stop the glusterfsd service (/etc/init.d/glusterfsd stop)
3) stop the glusterd service (/etc/init.d/glusterd stop)

Stopping the glusterd service before killing glusterfs & glusterfsd seems to cause freezing on the clients for the duration of ping-timeout.
In our case we have a NFS client/server set up on both hosts so we don't have to worry about client side failover.

Tomo Takebe
Information Systems & Computing
University of Pennsylvania


> -----Original Message-----
> From: gluster-users-bounces@xxxxxxxxxxx [mailto:gluster-users-
> bounces@xxxxxxxxxxx] On Behalf Of Nicolas
> Sent: Tuesday, October 27, 2015 1:24 PM
> To: gluster-users@xxxxxxxxxxx
> Subject:  Rebooting gluster nodes make VMs pause due to
> storage error
> 
> Hi,
> 
> We're using ovirt 3.5.3.1, and as storage backend we use GlusterFS. We
> added a Storage Domain with the path "gluster.fqdn1:/volume", and as
> options, we used "backup-volfile-servers=gluster.fqdn2". We now need to
> restart both gluster.fqdn1 and gluster.fqdn2 machines due to system update
> (not at the same time, obviously). We're worried because in previous
> attempts, when restarted the main gluster node (gluster.fqdn1 in this case),
> all the VMs running against that storage backend got paused due to storage
> errors, and we couldn't resume them and finally had to power them off the
> hard way and start them again.
> 
> Gluster version on gluster.fqdn1 and gluster.fqdn2 is 3.6.3-1 (on CentOS7).
> 
> Gluster configuration for that volume is:
> 
> Volume Name: volume
> Type: Replicate
> Volume ID: a2d7e52c-2f63-4e72-9635-4e311baae6ff
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: gluster.fqdn1:/gluster/brick_01/brick
> Brick2: gluster.fqdn2:/gluster/brick_01/brick
> Options Reconfigured:
> storage.owner-gid: 36
> storage.owner-uid: 36
> cluster.server-quorum-type: server
> cluster.quorum-type: none
> network.remote-dio: enable
> cluster.eager-lock: enable
> performance.stat-prefetch: off
> performance.io-cache: off
> performance.read-ahead: off
> performance.quick-read: off
> 
> We would like to know if this configuration should work, or if there's
> something missing or some problem with the above specified version, as
> pausing the VMs is a way to make it not fail but is not affordable for us. Also,
> we've noted that the self-healing process takes *a lot* of time, the above
> specified volume is 6T and it might take hours to synchronize after a half-
> hour desynchronization.
> 
> Any hints are appreciated,
> 
> Thanks.
> _______________________________________________
> Gluster-users mailing list
> Gluster-users@xxxxxxxxxxx
> http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users



[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux