Forgot to add that I am using Debian 5.0.1 and glusterfs 2.0.0 final, not using the gluster patched fuse. On Fri, May 1, 2009 at 4:47 PM, Ate Poorthuis <atepoorthuis at gmail.com>wrote: > I have a fully working gluster setup (2 distributed nodes that are > replicated on the server side to 2 other nodes). When exporting the > mountpoint as samba share, the gigE link is fully saturated on writing over > samba. However, when reading, performance is seriously reduced. > > Client bandwidth: > > | iface Rx Tx > Total > > ============================================================================== > lo: 0.00 KB/s 0.00 KB/s 0.00 > KB/s > eth0: 255.29 KB/s 5546.58 KB/s 5801.86 > KB/s > eth1: 16890.96 KB/s 389.40 KB/s 17280.36 > KB/s > > ------------------------------------------------------------------------------ > total: 17146.25 KB/s 5935.98 KB/s 23082.23 > KB/s > > And the following errors shows up in the logs: > > 2009-05-01 16:31:25 E [fuse-bridge.c:2332:fuse_getlk_cbk] glusterfs-fuse: > 73272: ERR => -1 (Transport endpoint is not connected) > 2009-05-01 16:31:25 E [fuse-bridge.c:2332:fuse_getlk_cbk] glusterfs-fuse: > 73289: ERR => -1 (Transport endpoint is not connected) > 2009-05-01 16:31:25 E [fuse-bridge.c:2332:fuse_getlk_cbk] glusterfs-fuse: > 73306: ERR => -1 (Transport endpoint is not connected) > 2009-05-01 16:31:25 E [fuse-bridge.c:2332:fuse_getlk_cbk] glusterfs-fuse: > 73323: ERR => -1 (Transport endpoint is not connected) > > I hope someone can provide any pointers on how to solve this. Vol files can > be found below. > > Thanks a lot, > > Ate > > Server .vol > -------- > volume posix1 > type storage/posix # POSIX FS translator > option directory /srv/export/gfs1/ # Export this directory > end-volume > > volume locks1 > type features/locks > subvolumes posix1 > end-volume > > volume read-ahead1 > type performance/read-ahead > option page-count 4 > subvolumes locks1 > end-volume > > volume write-behind1 > type performance/write-behind > subvolumes read-ahead1 > end-volume > > volume afr_52 > type protocol/client > option transport-type tcp/client > option remote-host 192.168.5.52 > option remote-subvolume write-behind1 > end-volume > > volume afr > type cluster/replicate > subvolumes write-behind1 afr_52 > end-volume > > ### Add network serving capability to above brick. > volume server > type protocol/server > option transport-type tcp > subvolumes write-behind1 afr > option auth.addr.write-behind1.allow * # Allow access to "brick" volume > option auth.addr.afr.allow * # Allow access to "brick" volume > end-volume > > > Client .vol > ----------- > volume brick1_51 > type protocol/client > option transport-type tcp > option remote-host 192.168.5.51 # IP address of the remote brick > option remote-subvolume afr # name of the remote volume > end-volume > volume brick1_101 > type protocol/client > option transport-type tcp > option remote-host 192.168.5.101 # IP address of the remote brick > option remote-subvolume afr # name of the remote volume > end-volume > > volume ha1 > type /testing/cluster/ha > subvolumes brick1_51 brick1_101 > end-volume > > volume brick1_52 > type protocol/client > option transport-type tcp > option remote-host 192.168.5.52 # IP address of the remote brick > option remote-subvolume afr # name of the remote volume > end-volume > > volume brick1_102 > type protocol/client > option transport-type tcp > option remote-host 192.168.5.102 # IP address of the remote brick > option remote-subvolume afr # name of the remote volume > end-volume > > volume ha2 > type /testing/cluster/ha > subvolumes brick1_52 brick1_102 > end-volume > > volume bricks > type cluster/distribute > option min-free-disk 5% > subvolumes ha1 ha2 > end-volume > > ### Add readahead feature > volume readahead > type performance/read-ahead > option page-size 1MB # unit in bytes > option page-count 2 # cache per file = (page-count x page-size) > subvolumes bricks > end-volume > ### Add writeback feature > volume writeback > type performance/write-behind > option flush-behind off > subvolumes readahead > end-volume > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://zresearch.com/pipermail/gluster-users/attachments/20090501/533f7f9a/attachment.htm>