Ok, I tried to change over to this, but now I just get: [2010-02-24 09:30:41] E [authenticate.c:234:gf_authenticate] auth: no authentication module is interested in accepting remote-client 10.0.0.24:1007 [2010-02-24 09:30:41] E [server-protocol.c:5822:mop_setvolume] tcb_remote: Cannot authenticate client from 10.0.0.24:1007 I am sure it is something simple, I just don't know what. Is this a port problem? the port in the log is 1007, but the server is on 50001. Here are my config files: ----- server.vol: ----- volume tcb_posix-export type storage/posix option directory /mnt/tcb_data end-volume volume tcb_locks-export type features/locks subvolumes tcb_posix-export end-volume volume tcb_export type performance/io-threads option thread-count 8 subvolumes tcb_locks-export end-volume volume tcb_remote type protocol/server option transport-type tcp option transport.socket.listen-port 50001 option transport.socket.nodelay on subvolumes tcb_export tcb_locks-export option auth.ip.tcb_locks-export.allow 10.0.0.*,10.0.20.*,10.0.30.*,192.168.1.*,192.168.20.*,192.168.30.*,127.0.0.1 option auth.ip.tcb_export.allow 10.0.0.*,10.0.20.*,10.0.30.*,192.168.1.*,192.168.20.*,192.168.30.*,127.0.0.1 end-volume ----- client.vol ----- volume tcb_remote1 type protocol/client option transport-type tcp option remote-port 50001 option remote-host 10.0.0.24 option remote-subvolume tcb_remote end-volume volume tcb_remote2 type protocol/client option transport-type tcp option remote-port 50001 option remote-host 10.0.0.25 option remote-subvolume tcb_remote end-volume volume tcb_mirror type cluster/afr subvolumes tcb_remote1 tcb_remote2 end-volume volume tcb_wb type performance/write-behind option cache-size 1MB subvolumes tcb_mirror end-volume volume tcb_ioc type performance/io-cache option cache-size 32MB subvolumes tcb_wb end-volume volume tcb_iothreads type performance/io-threads option thread-count 16 subvolumes tcb_ioc end-volume ^C Chad wrote: > I finally got the servers transported 2000 miles, set-up, wired, and > booted. > Here are the vol files. > Just to reiterate, the issues are slow performance on read/write, and > clients hanging when 1 server goes down. > > > ### glusterfs.vol ### > ############################################ > # Start tcb_cluster > ############################################ > # the exported volume to mount # required! > volume tcb_cluster > type protocol/client > option transport-type tcp/client > option remote-host glustcluster > option remote-port 50001 > option remote-subvolume tcb_cluster > end-volume > > ############################################ > # Start cs_cluster > ############################################ > # the exported volume to mount # required! > volume cs_cluster > type protocol/client > option transport-type tcp/client > option remote-host glustcluster > option remote-port 50002 > option remote-subvolume cs_cluster > end-volume > > ############################################ > # Start pbx_cluster > ############################################ > # the exported volume to mount # required! > volume pbx_cluster > type protocol/client > option transport-type tcp/client > option remote-host glustcluster > option remote-port 50003 > option remote-subvolume pbx_cluster > end-volume > > > --------------------------------------------------- > ### glusterfsd.vol ### > ############################################# > # Start tcb_data cluster > ############################################# > volume tcb_local > type storage/posix > option directory /mnt/tcb_data > end-volume > > volume tcb_locks > type features/locks > option mandatory-locks on # enables mandatory locking > on all files > subvolumes tcb_local > end-volume > > # dataspace on remote machine, look in /etc/hosts to see that > volume tcb_locks_remote > type protocol/client > option transport-type tcp > option remote-port 50001 > option remote-host 192.168.1.25 > option remote-subvolume tcb_locks > end-volume > > # automatic file replication translator for dataspace > volume tcb_cluster_afr > type cluster/replicate > subvolumes tcb_locks tcb_locks_remote > end-volume > > # the actual exported volume > volume tcb_cluster > type performance/io-threads > option thread-count 256 > option cache-size 128MB > subvolumes tcb_cluster_afr > end-volume > > volume tcb_cluster_server > type protocol/server > option transport-type tcp > option transport.socket.listen-port 50001 > option auth.addr.tcb_locks.allow * > option auth.addr.tcb_cluster.allow * > option transport.socket.nodelay on > subvolumes tcb_cluster > end-volume > > ############################################# > # Start cs_data cluster > ############################################# > volume cs_local > type storage/posix > option directory /mnt/cs_data > end-volume > > volume cs_locks > type features/locks > option mandatory-locks on # enables mandatory locking > on all files > subvolumes cs_local > end-volume > > # dataspace on remote machine, look in /etc/hosts to see that > volume cs_locks_remote > type protocol/client > option transport-type tcp > option remote-port 50002 > option remote-host 192.168.1.25 > option remote-subvolume cs_locks > end-volume > > # automatic file replication translator for dataspace > volume cs_cluster_afr > type cluster/replicate > subvolumes cs_locks cs_locks_remote > end-volume > > # the actual exported volume > volume cs_cluster > type performance/io-threads > option thread-count 256 > option cache-size 128MB > subvolumes cs_cluster_afr > end-volume > > volume cs_cluster_server > type protocol/server > option transport-type tcp > option transport.socket.listen-port 50002 > option auth.addr.cs_locks.allow * > option auth.addr.cs_cluster.allow * > option transport.socket.nodelay on > subvolumes cs_cluster > end-volume > > ############################################# > # Start pbx_data cluster > ############################################# > volume pbx_local > type storage/posix > option directory /mnt/pbx_data > end-volume > > volume pbx_locks > type features/locks > option mandatory-locks on # enables mandatory locking > on all files > subvolumes pbx_local > end-volume > > # dataspace on remote machine, look in /etc/hosts to see that > volume pbx_locks_remote > type protocol/client > option transport-type tcp > option remote-port 50003 > option remote-host 192.168.1.25 > option remote-subvolume pbx_locks > end-volume > > # automatic file replication translator for dataspace > volume pbx_cluster_afr > type cluster/replicate > subvolumes pbx_locks pbx_locks_remote > end-volume > > # the actual exported volume > volume pbx_cluster > type performance/io-threads > option thread-count 256 > option cache-size 128MB > subvolumes pbx_cluster_afr > end-volume > > volume pbx_cluster_server > type protocol/server > option transport-type tcp > option transport.socket.listen-port 50003 > option auth.addr.pbx_locks.allow * > option auth.addr.pbx_cluster.allow * > option transport.socket.nodelay on > subvolumes pbx_cluster > end-volume > > > -- > ^C > > > > Smart Weblications GmbH - Florian Wiessner wrote: >> Hi, >> >> Am 16.02.2010 01:58, schrieb Chad: >>> I am new to glusterfs, and this list, please let me know if I have made >>> any mistakes in posting this to the list. >>> I am not sure what your standards are. >>> >>> I came across glusterfs last week, it was super easy to set-up and test >>> and is almost exactly what I want/need. >>> I set up 2 "glusterfs servers" that serve up a mirrored raid5 disk >>> partitioned into 3 5oogb partitions to 6 clients. >>> I am using round robin DNS, but I also tried to use heartbeat and >>> ldirectord (see details below). >>> Each server has 2 NICs: 1 for the clients, the other has a cross over >>> cable connecting the 2 servers. Both NICs are 1000mbps. >>> >>> There are only 2 issues. >>> #1. When one of the servers goes down the clients hang at least for a >>> little while (more testing is needed) I am not sure if the clients can >>> recover at all. >>> #2. The read/write tests I performed came in at 1.6 when using >>> glusterfs, NFS on all the same machines came in at 11, and a direct test >>> on the data server came >>> in at 111. How do I improve the performance? >>> >> >> please share your vol-files. i don't understand why you would need >> loadbalancers. >> >>> ############################################### >>> My glusterfs set-up: >>> 2 supermicro dual Xeon 3.0 ghz CPUs, 8gb ram, 4 @ 750gb seagate sata >>> HDs, 3 in raid5 with 1 hot spare. (data servers) >> >> why not use raid10? same capacity, better speed.. >> >>> 6 supermicro dual AMD 2.8 ghz CPUs, 4gb ram, 2 @ 250gb seagate sata HDs >>> in raid 1. (client machines) >>> glusterfs is set-up with round robin DNS to handle the load balancing of >>> the 2 data servers. >> >> afaik there is no need to setup dns rr nor loadbalancing for the gluster >> servers, glusterfs should take care of that itself. but without your >> volfiles i >> can't give any hints. >> >> > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > >