I haven't tried all performance options disabled yet- I can try that tomorrow when the resource frees up. I was actually asking first before blindly trying different configuration matrices in case there's a clear direction I should take with it. I'll let you know. Jeremy On 3/24/2010 2:54 AM, Stephan von Krawczynski wrote: > Hi Jeremy, > > have you tried to reproduce with all performance options disabled? They are > possibly no good idea on a local system. > What local fs do you use? > > > -- > Regards, > Stephan > > > On Tue, 23 Mar 2010 19:11:28 -0500 > Jeremy Enos<jenos at ncsa.uiuc.edu> wrote: > > >> Stephan is correct- I primarily did this test to show a demonstrable >> overhead example that I'm trying to eliminate. It's pronounced enough >> that it can be seen on a single disk / single node configuration, which >> is good in a way (so anyone can easily repro). >> >> My distributed/clustered solution would be ideal if it were fast enough >> for small block i/o as well as large block- I was hoping that single >> node systems would achieve that, hence the single node test. Because >> the single node test performed poorly, I eventually reduced down to >> single disk to see if it could still be seen, and it clearly can be. >> Perhaps it's something in my configuration? I've pasted my config files >> below. >> thx- >> >> Jeremy >> >> ######################glusterfsd.vol###################### >> volume posix >> type storage/posix >> option directory /export >> end-volume >> >> volume locks >> type features/locks >> subvolumes posix >> end-volume >> >> volume disk >> type performance/io-threads >> option thread-count 4 >> subvolumes locks >> end-volume >> >> volume server-ib >> type protocol/server >> option transport-type ib-verbs/server >> option auth.addr.disk.allow * >> subvolumes disk >> end-volume >> >> volume server-tcp >> type protocol/server >> option transport-type tcp/server >> option auth.addr.disk.allow * >> subvolumes disk >> end-volume >> >> ######################ghome.vol###################### >> >> #-----------IB remotes------------------ >> volume ghome >> type protocol/client >> option transport-type ib-verbs/client >> # option transport-type tcp/client >> option remote-host acfs >> option remote-subvolume raid >> end-volume >> >> #------------Performance Options------------------- >> >> volume readahead >> type performance/read-ahead >> option page-count 4 # 2 is default option >> option force-atime-update off # default is off >> subvolumes ghome >> end-volume >> >> volume writebehind >> type performance/write-behind >> option cache-size 1MB >> subvolumes readahead >> end-volume >> >> volume cache >> type performance/io-cache >> option cache-size 1GB >> subvolumes writebehind >> end-volume >> >> ######################END###################### >> >> >> >> On 3/23/2010 6:02 AM, Stephan von Krawczynski wrote: >> >>> On Tue, 23 Mar 2010 02:59:35 -0600 (CST) >>> "Tejas N. Bhise"<tejas at gluster.com> wrote: >>> >>> >>> >>>> Out of curiosity, if you want to do stuff only on one machine, >>>> why do you want to use a distributed, multi node, clustered, >>>> file system ? >>>> >>>> >>> Because what he does is a very good way to show the overhead produced only by >>> glusterfs and nothing else (i.e. no network involved). >>> A pretty relevant test scenario I would say. >>> >>> -- >>> Regards, >>> Stephan >>> >>> >>> >>> >>>> Am I missing something here ? >>>> >>>> Regards, >>>> Tejas. >>>> >>>> ----- Original Message ----- >>>> From: "Jeremy Enos"<jenos at ncsa.uiuc.edu> >>>> To: gluster-users at gluster.org >>>> Sent: Tuesday, March 23, 2010 2:07:06 PM GMT +05:30 Chennai, Kolkata, Mumbai, New Delhi >>>> Subject: gluster local vs local = gluster x4 slower >>>> >>>> This test is pretty easy to replicate anywhere- only takes 1 disk, one >>>> machine, one tarball. Untarring to local disk directly vs thru gluster >>>> is about 4.5x faster. At first I thought this may be due to a slow host >>>> (Opteron 2.4ghz). But it's not- same configuration, on a much faster >>>> machine (dual 3.33ghz Xeon) yields the performance below. >>>> >>>> ####THIS TEST WAS TO A LOCAL DISK THRU GLUSTER#### >>>> [root at ac33 jenos]# time tar xzf >>>> /scratch/jenos/intel/l_cproc_p_11.1.064_intel64.tgz >>>> >>>> real 0m41.290s >>>> user 0m14.246s >>>> sys 0m2.957s >>>> >>>> ####THIS TEST WAS TO A LOCAL DISK (BYPASS GLUSTER)#### >>>> [root at ac33 jenos]# cd /export/jenos/ >>>> [root at ac33 jenos]# time tar xzf >>>> /scratch/jenos/intel/l_cproc_p_11.1.064_intel64.tgz >>>> >>>> real 0m8.983s >>>> user 0m6.857s >>>> sys 0m1.844s >>>> >>>> ####THESE ARE TEST FILE DETAILS#### >>>> [root at ac33 jenos]# tar tzvf >>>> /scratch/jenos/intel/l_cproc_p_11.1.064_intel64.tgz |wc -l >>>> 109 >>>> [root at ac33 jenos]# ls -l >>>> /scratch/jenos/intel/l_cproc_p_11.1.064_intel64.tgz >>>> -rw-r--r-- 1 jenos ac 804385203 2010-02-07 06:32 >>>> /scratch/jenos/intel/l_cproc_p_11.1.064_intel64.tgz >>>> [root at ac33 jenos]# >>>> >>>> These are the relevant performance options I'm using in my .vol file: >>>> >>>> #------------Performance Options------------------- >>>> >>>> volume readahead >>>> type performance/read-ahead >>>> option page-count 4 # 2 is default option >>>> option force-atime-update off # default is off >>>> subvolumes ghome >>>> end-volume >>>> >>>> volume writebehind >>>> type performance/write-behind >>>> option cache-size 1MB >>>> subvolumes readahead >>>> end-volume >>>> >>>> volume cache >>>> type performance/io-cache >>>> option cache-size 1GB >>>> subvolumes writebehind >>>> end-volume >>>> >>>> What can I do to improve gluster's performance? >>>> >>>> Jeremy >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>> >>>> >>>> >>> >>> >> > >