Willdo, though I recently added in those lines to help be explicit about behaviour (I had no options set before at all, leaving it to the default of 16 threads). I will remove and specify the default of 16 to see if that helps. Im adding: volume trace type debug/trace subvolumes cache end-volume to both sides now as well, so next time (if any) it locks up perhaps there will be some more info. Thanks Shehjar On 18 Jun 2009, at 11:26, Shehjar Tikoo wrote: > Daniel Jordan Bambach wrote: >> I'm experiencing various locking up issues ranging from Gluster >> locking up ( 'ls'ing the mount hangs ), to the whole machine >> locking up under load. >> My current config is below (two servers, afring) >> I would love to be able to get to the bottom of this, because it >> seems very strange that we should see erratic behaviour on such a >> simple setup. >> There is approx 12Gb of files, and to stress test (and heal) i run >> ls -alR on the mount. This will run for a while and eventually lock >> up Gluster, and occasionally the machine. I have found that in some >> cases killing Gluster and re-mounting does not solve the problem >> (in that perhaps both servers have entered a locked state in some >> way). >> Im finding it very hard to collect and debug information of any >> use, as there is no crashlog, no errors in the volume log. >> Can anyone suggest what I migth be able to do to extract more >> information as to what is occuring at lock-up time? >> volume posix >> type storage/posix >> option directory /home/export >> end-volume >> volume locks >> type features/locks >> subvolumes posix >> end-volume >> volume brick >> type performance/io-threads >> subvolumes locks >> option autoscaling on >> option min-threads 8 >> option max-threads 32 >> end-volume > I see that the max-threads will never exceed 32 which is > a reasonable valueand should work fine in most cases but considering > some of the other reports we've been getting, could you please try > again > but without the autoscaling turned on? > > It is off by default, so you can simply set the number of threads > you need by: > > option thread-count <COUNT> > > ...instead of the three "option" lines above. > > Thanks > Shehjar > > >> volume server >> type protocol/server >> option transport-type tcp >> option auth.addr.brick.allow * >> subvolumes brick >> end-volume >> volume latsrv2 >> type protocol/client >> option transport-type tcp >> option remote-host latsrv2 >> option remote-subvolume brick >> end-volume >> volume afr >> type cluster/replicate >> subvolumes brick latsrv2 >> option read-subvolume brick >> end-volume >> volume writebehind >> type performance/write-behind >> option cache-size 2MB >> subvolumes afr >> end-volume >> volume cache >> type performance/io-cache >> option cache-size 32MB >> option priority *.pyc:4,*.html:3,*.php:2,*:1 >> option cache-timeout 5 >> subvolumes writebehind >> end-volume >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > >