Using code pulled March 21st @ 1am EST. Kernel 2.6.20-10 on Ubuntu
Feisty 32bit.
I have 3 machines serving with glusterfsd and mounted the cluster from
the first server. If I kill one of the glusterfsd's on any of the
machines, the mount point becomes broken with the 'Transport' error
below. Also, glusterfs will produce this error even if I unmount and
remount with the 1 glusterfsd server still down. Is this expected
results or shouldn't the mount continue to work even though one of the
servers has "died"?
ls: reading directory local/: Transport endpoint is not connected
glusterfs.log produces this:
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()]
protocol/client: client_protocol_xfer: :transport_submit failed
[Mar 21 06:25:17] [ERROR/tcp-client.c:284/tcp_connect()]
tcp/client:non-blocking connect() returned: 111 (Connection refused)
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()]
protocol/client: client_protocol_xfer: :transport_submit failed
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()]
protocol/client: client_protocol_xfer: :transport_submit failed
[Mar 21 06:25:17] [ERROR/tcp-client.c:284/tcp_connect()]
tcp/client:non-blocking connect() returned: 111 (Connection refused)
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()]
protocol/client: client_protocol_xfer: :transport_submit failed
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()]
protocol/client: client_protocol_xfer: :transport_submit failed
======================================
glusterfs-server.vol used by all the servers:
(ignore my bad volume naming, was just testing)
volume brick
type storage/posix # POSIX FS translator
option directory /home/export # Export this directory
end-volume
volume iothreads
type performance/io-threads
option thread-count 8
subvolumes brick
end-volume
volume server
type protocol/server
option transport-type tcp/server # For TCP/IP transport
subvolumes iothreads
option auth.ip.brick.allow * # Allow access to "brick" volume
end-volume
======================================
glusterfs-client.vol used on server1:
(ignore my bad volume naming, was just testing)
volume client1
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 1.1.1.1 # IP address of the remote brick
option remote-subvolume brick # name of the remote volume
end-volume
volume client2
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 2.2.2.2 # IP address of the remote brick
option remote-subvolume brick # name of the remote volume
end-volume
volume client3
type protocol/client
option transport-type tcp/client # for TCP/IP transport
option remote-host 3.3.3.3 # IP address of the remote brick
option remote-subvolume brick # name of the remote volume
end-volume
volume bricks
type cluster/unify
subvolumes client1 client2 client3
option scheduler alu
option alu.limits.min-free-disk 60GB # Stop creating
files when free-space lt 60GB
option alu.limits.max-open-files 10000
option alu.order
disk-usage:read-usage:write-usage:open-files-usage:disk-speed-usage
option alu.disk-usage.entry-threshold 2GB # Units in KB, MB
and GB are allowed
option alu.disk-usage.exit-threshold 60MB # Units in KB, MB
and GB are allowed
option alu.open-files-usage.entry-threshold 1024
option alu.open-files-usage.exit-threshold 32
option alu.stat-refresh.interval 10sec
end-volume
volume writebehind #writebehind improves write performance a lot
type performance/write-behind
option aggregate-size 131072 # in bytes
subvolumes bricks
end-volume