afr :2 HA setup question

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all -

I have a setup based on this :
 http://www.gluster.org/docs/index.php/GlusterFS_High_Availability_Storage_with_GlusterFS<http://www.gluster.org/docs/index.php/GlusterFS_High_Availability_Storage_with_GlusterFS>
but with only 2 machines. Effectively just a mirror (glusterfsd
configuration below). 1.3.1 client and server.

The problem I am seeing is that if I start up one side of the afr and the
other side is down, I can only read and write to files that are already
there. If I try to create a new file, the glusterfsd logs show this:

07-09-07 11:39:00 D [common-utils.c:194:gf_resolve_ip] resolver: flushing
DNS cache
2007-09-07 11:39:00 D [tcp-client.c:142:tcp_connect] brick-ds-bak: connect
on 9 in progress (non-blocking)
2007-09-07 11:39:00 E [tcp-client.c:171:tcp_connect] brick-ds-bak:
non-blocking connect() returned: 111 (Connection refused)
2007-09-07 11:39:00 W [client-protocol.c:344:client_protocol_xfer]
brick-ds-bak: not connected at the moment to submit frame type(0) op(34)
2007-09-07 11:39:00 D [inode.c:297:__destroy_inode] brick/inode: destroy
inode(0) [@0x148e8200]
2007-09-07 11:39:17 D [client-protocol.c:4211:client_protocol_reconnect]
brick-ds-bak: attempting reconnect

and the file creation just never returns. ie, if i mount this mirror setup
on /brick and do: touch /brick/non_exist, an strace of tha tprocess just
shows it hanging in :

  open("/brick/non_exist", O_WRONLY|O_CREAT|O_NOCTTY|O_NONBLOCK, 0666

So my question is whether this is indended behavior? The wiki page makes it
sound like this is an HA setup, but if one side goes down and that makes the
remaining node useless, it's not much use as an HA setup. Am I missing
something critical here?


thanks,
august

------------
client (mounted on /brick):


  volume brick
     type protocol/client
     option transport-type tcp/client
     option remote-host 192.168.16.126
     option remote-subvolume brick
     option remote-port 16
   end-volume


server :

 volume brick-ds
           type storage/posix
           option directory /.brick-ds
 end-volume

 volume brick-ns
           type storage/posix
           option directory /.brick-ns
 end-volume


    volume brick-ds-bak
             type protocol/client
             option transport-type tcp/client
             option remote-host 192.168.16.254
             option remote-port 16
             option remote-subvolume brick-ds
     end-volume

     volume brick-ns-bak
             type protocol/client
             option transport-type tcp/client
             option remote-host 192.168.16.254
             option remote-port 16
             option remote-subvolume brick-ns
     end-volume


    volume brick-ds-afr
            type cluster/afr
            subvolumes brick-ds brick-ds-bak
            option replicate *:2
    end-volume
    volume brick-ns-afr
            type cluster/afr
            subvolumes brick-ds brick-ds-bak
            option replicate *:2
    end-volume

   volume brick-unify
            type cluster/unify
            subvolumes brick-ds-afr
            option namespace brick-ns-afr
            option scheduler rr
    end-volume


   volume brick
            type performance/io-threads
            option thread-count 4
            option cache-size 64MB
            subvolumes brick-unify
    end-volume

   volume server
     type protocol/server
     option transport-type tcp/server
     option bind-address 192.168.16.126
     option auth.ip.brick-ds.allow 192.168.16.*
     option auth.ip.brick-ns.allow 192.168.16.*
     option auth.ip.brick.allow 192.168.16.*
     option listen-port 16
     subvolumes brick
   end-volume


[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux