Hi Dale, This is a recent change. You need to put following line in unify for it to work when one of the node goes down. "option readdir-force-success on" We discussed whether user should know if one of the node has gone down, coming to the conclusion that it is best to leave it as an option for the user to configure it himself. Regards Krishna On 3/22/07, Dale Dude <dale@xxxxxxxxxxxxxxx> wrote:
Using code pulled March 21st @ 1am EST. Kernel 2.6.20-10 on Ubuntu Feisty 32bit. I have 3 machines serving with glusterfsd and mounted the cluster from the first server. If I kill one of the glusterfsd's on any of the machines, the mount point becomes broken with the 'Transport' error below. Also, glusterfs will produce this error even if I unmount and remount with the 1 glusterfsd server still down. Is this expected results or shouldn't the mount continue to work even though one of the servers has "died"? ls: reading directory local/: Transport endpoint is not connected glusterfs.log produces this: [Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed [Mar 21 06:25:17] [ERROR/tcp-client.c:284/tcp_connect()] tcp/client:non-blocking connect() returned: 111 (Connection refused) [Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed [Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed [Mar 21 06:25:17] [ERROR/tcp-client.c:284/tcp_connect()] tcp/client:non-blocking connect() returned: 111 (Connection refused) [Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed [Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] protocol/client: client_protocol_xfer: :transport_submit failed ====================================== glusterfs-server.vol used by all the servers: (ignore my bad volume naming, was just testing) volume brick type storage/posix # POSIX FS translator option directory /home/export # Export this directory end-volume volume iothreads type performance/io-threads option thread-count 8 subvolumes brick end-volume volume server type protocol/server option transport-type tcp/server # For TCP/IP transport subvolumes iothreads option auth.ip.brick.allow * # Allow access to "brick" volume end-volume ====================================== glusterfs-client.vol used on server1: (ignore my bad volume naming, was just testing) volume client1 type protocol/client option transport-type tcp/client # for TCP/IP transport option remote-host 1.1.1.1 # IP address of the remote brick option remote-subvolume brick # name of the remote volume end-volume volume client2 type protocol/client option transport-type tcp/client # for TCP/IP transport option remote-host 2.2.2.2 # IP address of the remote brick option remote-subvolume brick # name of the remote volume end-volume volume client3 type protocol/client option transport-type tcp/client # for TCP/IP transport option remote-host 3.3.3.3 # IP address of the remote brick option remote-subvolume brick # name of the remote volume end-volume volume bricks type cluster/unify subvolumes client1 client2 client3 option scheduler alu option alu.limits.min-free-disk 60GB # Stop creating files when free-space lt 60GB option alu.limits.max-open-files 10000 option alu.order disk-usage:read-usage:write-usage:open-files-usage:disk-speed-usage option alu.disk-usage.entry-threshold 2GB # Units in KB, MB and GB are allowed option alu.disk-usage.exit-threshold 60MB # Units in KB, MB and GB are allowed option alu.open-files-usage.entry-threshold 1024 option alu.open-files-usage.exit-threshold 32 option alu.stat-refresh.interval 10sec end-volume volume writebehind #writebehind improves write performance a lot type performance/write-behind option aggregate-size 131072 # in bytes subvolumes bricks end-volume _______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxx http://lists.nongnu.org/mailman/listinfo/gluster-devel