Brandon, which is the fuse kernel module version you are using? please with the fuse kernel module from - http://ftp.zresearch.com/pub/gluster/glusterfs/fuse/old/fuse-2.7.0.tar.gz we have been seeing some perf issues with small files while write-behind is on. can you try with an untar of large files, with write-behind on? thanks, avati 2008/1/18, Brandon Lamb <brandonlamb@xxxxxxxxx>: > > On Jan 17, 2008 6:15 PM, Brandon Lamb <brandonlamb@xxxxxxxxx> wrote: > > > > On Jan 17, 2008 6:13 PM, Anand Avati <avati@xxxxxxxxxxxxx > wrote: > > > Brandon, > > > For the sake of diagnosing, can you try with these changes - > > > > > > 1. with a simple client and standalone server (no clustering anywhere) > > > 2. remove unify in your setup. unify is not needed in this > configuration. > > > AFR on the client itself would be even better. > > > 3. try removing write-behind. > > > > > > We're interested in knowing your results from those changes. > > > > > > avati > > > > > > 2008/1/17, Brandon Lamb < brandonlamb@xxxxxxxxx>: > > > > > > > > > > > > > > > > On Jan 17, 2008 9:56 AM, Brandon Lamb < brandonlamb@xxxxxxxxx> > wrote: > > > > > http://ezopg.com/gfs/ > > > > > > > > > > I uploaded my client config and the server configs for the 2 > servers. > > > > > 3 seperate machines. > > > > > > > > > > I can get to mounting, then do cd /mnt/gfs (mounted gfs dir) and > then > > > > > i typed find . -type f -exec head -c 1 {} \; >/dev/null and got > > > > > > > > > > find: ./test: Transport endpoint is not connected > > > > > find: ./bak: Transport endpoint is not connected > > > > > > > > > > And then the crash on server1 > > > > > > > > > > Am I doing someting obviously wrong? > > > > > > > > > > > > > Ok I emptied the gfs and gfsns dirs on both servers (they had > existing > > > > files/dirs from a different gfs setup i was testing). > > > > > > > > Now I can create files and dirs. > > > > > > > > Now I am wondering about speed. I have a 82 megabyte tarball with > 337 > > > files. > > > > [root@client gfstest]# time tar xf yoda.tar > > > > real 0m22.567s > > > > user 0m0.042s > > > > sys 0m0.357s > > > > > > > > Now i changed to a dir on server2 that i have mounted over nfs > > > > [root@client nfstest]# time tar xf yoda.tar > > > > real 0m4.956s > > > > user 0m0.030s > > > > sys 0m0.827s > > > > > > > > Do I have some performance translators configured wrong or in the > > > > wrong place or is that really the speed I should be expecting? > > > > > > > > Server 1 is a 8 sata2 raid (8 seagate 250g es drives) > > > > Server 2 is a 16 scsi 160 raid > > > > > > > > From roughly 5 seconds to 22 is a huge increase, Im hoping im doing > > > > something horribly wrong. Using gigabit switch on its own 192 > network > > > > for all this > > > > > > > > > > > > _______________________________________________ > > > > Gluster-devel mailing list > > > > Gluster-devel@xxxxxxxxxx > > > > http://lists.nongnu.org/mailman/listinfo/gluster-devel > > > > > > > > > > > > > > > > -- > > > If I traveled to the end of the rainbow > > > As Dame Fortune did intend, > > > Murphy would be there to tell me > > > The pot's at the other end. > > > > Ok I will set this up in awhile and report back later tonight > > http://ezopg.com/gfssimpletest/ > > I posted gluster log files for server and client, spec files and > results.txt > -- If I traveled to the end of the rainbow As Dame Fortune did intend, Murphy would be there to tell me The pot's at the other end.