Krishna Srinivas wrote:
Gerry, Christopher,
Here is what I tried to do. Two servers, one client, simple setup, afr on the
client side. I did "ls" on client mount point, it works, now I do
"ifconfig eth0 down"
on the server, next I do "ls" on client, it hangs for 10 secs (timeout value)
and fails over and starts working again without any problem.
I guess few users are facing the problem you guys are facing. Can you give us
your setup details and mention the exact steps to reproduce. Also try
to come up with minimal config details which can still reproduce the
problem
Thanks!
Krishna
Krishna and Anand:
Hope these will help. Here are my configs:
================================================
### file: glusterfs-server.vol
### GRP: this files goes on all storage bricks
##############################################
### GlusterFS Server Volume Specification ##
##############################################
#### CONFIG FILE RULES:
### "#" is comment character.
### - Config file is case sensitive
### - Options within a volume block can be in any order.
### - Spaces or tabs are used as delimitter within a line.
### - Multiple values to options will be : delimitted.
### - Each option should end within a line.
### - Missing or commented fields will assume default values.
### - Blank/commented lines are allowed.
### - Sub-volumes should already be defined above before referring.
### Export volume "brick" with the contents of "/home/export" directory.
volume brick
type storage/posix # POSIX FS translator
option directory /home/vmail/mailbrick # Export this directory
end-volume
### Add network serving capability to above brick.
volume server
type protocol/server
option transport-type tcp/server # For TCP/IP transport
# option ibv-send-work-request-size 131072
# option ibv-send-work-request-count 64
# option ibv-recv-work-request-size 131072
# option ibv-recv-work-request-count 64
# option transport-type ib-sdp/server # For Infiniband transport
# option transport-type ib-verbs/server # For ib-verbs transport
# option bind-address 127.0.0.1 # Default is to listen on all interfaces
option listen-port 6996 # Default is 6996
# option client-volume-filename /etc/glusterfs/glusterfs-client.vol
subvolumes brick
# NOTE: Access to any volume through protocol/server is denied by
# default. You need to explicitly grant access through # "auth"
# option.
# option auth.ip.brick.allow * # Allow full access to
"brick" volume
# option auth.ip.brick.allow 192.168.* # Allow subnet access to
"brick" volume
option auth.ip.brick.allow 127.0.0.1,192.168.1.220,192.168.1.221 #
Allow access to "brick" volume
end-volume
================================================
================================================
### file: glusterfs-client.vol
### GRP: this file goes on every client node in cluster
##############################################
### GlusterFS Client Volume Specification ##
##############################################
#### CONFIG FILE RULES:
### "#" is comment character.
### - Config file is case sensitive
### - Options within a volume block can be in any order.
### - Spaces or tabs are used as delimitter within a line.
### - Each option should end within a line.
### - Missing or commented fields will assume default values.
### - Blank/commented lines are allowed.
### - Sub-volumes should already be defined above before referring.
### Add client feature and declare local subvolume
volume client-local
type storage/posix
option directory /home/vmail/mailbrick
end-volume
### Add client feature and attach to remote subvolume
volume client1
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 192.168.1.220 # IP address of the remote brick
option remote-port 6996 # default server port is 6996
option remote-subvolume brick # name of the remote volume
end-volume
volume client2
type protocol/client
option transport-type tcp/client
option remote-host 192.168.1.221
option remote-port 6996
option remote-subvolume brick
end-volume
#volume bricks
# type cluster/unify
# subvolumes *
# option scheduler nufa
# # does this brick name must be in local server.vol
# option nufa.local-volume-name brick # note 'brick' is singular
#end-volume
### Add automatic file replication (AFR) feature
volume afr
type cluster/afr
subvolumes client1 client2
# option replicate:*.html 2
# option replicate:*.db 5
## ok, this would be RAID-1 on 2 nodes
# option replicate:* 2
## so how would you say RAID-1 on all nodes? with * ?
# option replicate *:2
# option replicate client1,client2:2
# option replicate is no longer supported:
http://www.mail-archive.com/gluster-devel@xxxxxxxxxx/msg02201.html
# pattern-matching translator will be provided later in 1.4
end-volume
### Add writeback feature
#volume writeback
# type performance/write-behind
# option aggregate-size 131072 # unit in bytes
# subvolumes client
#end-volume
### Add readahead feature
#volume readahead
# type performance/read-ahead
# option page-size 65536 # unit in bytes
# option page-count 16 # cache per file = (page-count x page-size)
# subvolumes writeback
#end-volume
================================================
Regards,
Gerry