New to GlusterFS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Joe,

You mentioned "If you shut down a server through the normal kill process":

What do you mean by normal kill process. Is it just shutdown the server for maintanence after going a "service glusterd stop" and "service glusterfsd stop".

Thanks & Regards,
Bobby Jacob
P SAVE TREES. Please don't print this e-mail unless you really need to.

From: gluster-users-bounces at gluster.org [mailto:gluster-users-bounces at gluster.org] On Behalf Of Joe Julian
Sent: Tuesday, October 22, 2013 4:12 PM
To: gluster-users at gluster.org
Subject: Re: New to GlusterFS

The reason for the long (42 second) ping-timeout is because re-establishing fd's and locks can be a very expensive operation. Allowing a longer time to reestablish connections is logical, unless you have servers that frequently die.

If you shut down a server through the normal kill process, the TCP connections will be closed properly. The client will be aware that the server is going away and there will be no timeout. This allows server maintenance without encountering that issue.

One issue with a 42 second timeout is that ext4 may detect an error and remount itself read only should that happen while the VM is running. You can override this behavior by specifying the mount option, "errors=continue" in fstab ("errors=remount-ro" is the default). The default can be changed, as well, by changing the superblock option with tune2fs.
On 10/22/2013 03:12 AM, John Mark Walker wrote:

Hi JC,

Yes, the default is a 42-second timeout for failover. You can configure that to be a smaller window.

-JM
On Oct 22, 2013 10:57 AM, "JC Putter" <jcputter at gmail.com<mailto:jcputter at gmail.com>> wrote:
Hi,

I am new to GlusterFS, i am trying to accomplish something which i am
not 100% sure is the correct use case but hear me out.

I want to use GlusterFS to host KVM VM's, from what I've read this was
not recommended due to poor write performance however since
libgfapi/qemu 1.3  this is now viable ?


Currently i'am testing out GlusterFS with two nodes, both running as
server and client

i have the following Volume:

Volume Name: DATA
Type: Replicate
Volume ID: eaa7746b-a1c1-4959-ad7d-743ac519f86a
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: glusterfs1.example.com:/data
Brick2: glusterfs2.example.com:/data


and mounting the brick locally on each server as /mnt/gluster,
replication works and everything but as soon as i kill one node, the
directory /mnt/gluster/ becomes unavailable for 30/40 seconds

log shows

[2013-10-22 11:55:48.055571] W [socket.c:514:__socket_rwv]
0-DATA-client-0: readv failed (No data available)


Thanks in advance!
_______________________________________________
Gluster-users mailing list
Gluster-users at gluster.org<mailto:Gluster-users at gluster.org>
http://supercolony.gluster.org/mailman/listinfo/gluster-users




_______________________________________________

Gluster-users mailing list

Gluster-users at gluster.org<mailto:Gluster-users at gluster.org>

http://supercolony.gluster.org/mailman/listinfo/gluster-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131023/c47a8997/attachment.html>


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux