Also, there's nothing happening in the log files beyond the initial mount so it looks like the filesystem aspect of it is fine. > -----Original Message----- > From: gluster-users-bounces at gluster.org > [mailto:gluster-users-bounces at gluster.org] On Behalf Of Hiren Joshi > Sent: 23 September 2009 14:02 > To: Pavan Vilas Sondur > Cc: gluster-users at gluster.org > Subject: Re: Rsync > > Bellow. > > I've found that I get a performance hit if I add read cache or > whitebehind. > > Server conf: > > ##Open vols > > volume posix1 > type storage/posix > option directory /gluster/export1 > end-volume > > volume posix2 > type storage/posix > option directory /gluster/export2 > end-volume > > volume posix3 > type storage/posix > option directory /gluster/export3 > end-volume > > volume posix4 > type storage/posix > option directory /gluster/export4 > end-volume > > volume posix5 > type storage/posix > option directory /gluster/export5 > end-volume > > volume posix6 > type storage/posix > option directory /gluster/export6 > end-volume > ## Add the ability to lock files etc > > volume locks1 > type features/locks > subvolumes posix1 > end-volume > > volume locks2 > type features/locks > subvolumes posix2 > end-volume > > volume locks3 > type features/locks > subvolumes posix3 > end-volume > > volume locks4 > type features/locks > subvolumes posix4 > end-volume > > volume locks5 > type features/locks > subvolumes posix5 > end-volume > > volume locks6 > type features/locks > subvolumes posix6 > end-volume > ## Preformance translators > > volume brick1 > type performance/io-threads > option thread-count 8 > subvolumes locks1 > end-volume > > volume brick2 > type performance/io-threads > option thread-count 8 > subvolumes locks2 > end-volume > > volume brick3 > type performance/io-threads > option thread-count 8 > subvolumes locks3 > end-volume > > volume brick4 > type performance/io-threads > option thread-count 8 > subvolumes locks4 > end-volume > > volume brick5 > type performance/io-threads > option thread-count 8 > subvolumes locks5 > end-volume > > volume brick6 > type performance/io-threads > option thread-count 8 > subvolumes locks6 > end-volume > ##export the lot > > #volume brick6 > # type debug/trace > # subvolumes trace_brick6 > #end-volume > > volume server > type protocol/server > option transport-type tcp/server > option auth.addr.brick1.allow * > option auth.addr.brick2.allow * > option auth.addr.brick3.allow * > option auth.addr.brick4.allow * > option auth.addr.brick5.allow * > option auth.addr.brick6.allow * > subvolumes brick1 brick2 brick3 brick4 brick5 brick6 > end-volume > > > Vol file: > ##Clent config > ##import all the briks from all the mirrors and mirror them > > volume glust1a_1 > type protocol/client > option transport-type tcp/client > option remote-host glust1a > option ping-timeout 30 > option remote-subvolume brick1 > end-volume > > volume glust1b_1 > type protocol/client > option transport-type tcp/client > option remote-host glust1b > option ping-timeout 30 > option remote-subvolume brick1 > end-volume > > volume mirror1_1 > type cluster/replicate > subvolumes glust1a_1 glust1b_1 > end-volume > > volume glust1a_2 > type protocol/client > option transport-type tcp/client > option remote-host glust1a > option ping-timeout 30 > option remote-subvolume brick2 > end-volume > > volume glust1b_2 > type protocol/client > option transport-type tcp/client > option remote-host glust1b > option ping-timeout 30 > option remote-subvolume brick2 > end-volume > > volume mirror1_2 > type cluster/replicate > subvolumes glust1a_2 glust1b_2 > end-volume > > volume glust1a_3 > type protocol/client > option transport-type tcp/client > option remote-host glust1a > option ping-timeout 30 > option remote-subvolume brick3 > end-volume > > volume glust1b_3 > type protocol/client > option transport-type tcp/client > option remote-host glust1b > option ping-timeout 30 > option remote-subvolume brick3 > end-volume > > volume mirror1_3 > type cluster/replicate > subvolumes glust1a_3 glust1b_3 > end-volume > > volume glust1a_4 > type protocol/client > option transport-type tcp/client > option remote-host glust1a > option ping-timeout 30 > option remote-subvolume brick4 > end-volume > > volume glust1b_4 > type protocol/client > option transport-type tcp/client > option remote-host glust1b > option ping-timeout 30 > option remote-subvolume brick4 > end-volume > > volume mirror1_4 > type cluster/replicate > subvolumes glust1a_4 glust1b_4 > end-volume > > volume glust1a_5 > type protocol/client > option transport-type tcp/client > option remote-host glust1a > option ping-timeout 30 > option remote-subvolume brick5 > end-volume > > volume glust1b_5 > type protocol/client > option transport-type tcp/client > option remote-host glust1b > option ping-timeout 30 > option remote-subvolume brick5 > end-volume > > volume mirror1_5 > type cluster/replicate > subvolumes glust1a_5 glust1b_5 > end-volume > > volume glust1a_6 > type protocol/client > option transport-type tcp/client > option remote-host glust1a > option ping-timeout 30 > option remote-subvolume brick6 > end-volume > > volume glust1b_6 > type protocol/client > option transport-type tcp/client > option remote-host glust1b > option ping-timeout 30 > option remote-subvolume brick6 > end-volume > > volume mirror1_6 > type cluster/replicate > subvolumes glust1a_6 glust1b_6 > end-volume > > ##place it all into a hash > > volume dist > type cluster/distribute > subvolumes mirror1_1 mirror1_2 mirror1_3 mirror1_4 > mirror1_5 mirror1_6 > end-volume > > > > -----Original Message----- > > From: Pavan Vilas Sondur [mailto:pavan at gluster.com] > > Sent: 23 September 2009 13:38 > > To: Hiren Joshi > > Cc: gluster-users at gluster.org > > Subject: Re: Rsync > > > > Can you send us the volfiles and logfiles as well? > > > > Pavan > > > > On 23/09/09 10:08 +0100, Hiren Joshi wrote: > > > 2.0.6-1.el5, it's a 6 brick distributed system which is > > mirrored over 2 > > > machines (each brick has a mirror on another machine, these > > mirrors are > > > put in the hash). > > > > > > An update, after running the rsync for a day, I killed it > > and remounted > > > all the disks (the underlying filesystem, not the gluster) > > with noatime, > > > the rsync completed in about 600 minutes. I'm now going to > > try one level > > > up (about 1,000,000,000 dirs). > > > > > > > -----Original Message----- > > > > From: Pavan Vilas Sondur [mailto:pavan at gluster.com] > > > > Sent: 23 September 2009 07:55 > > > > To: Hiren Joshi > > > > Cc: gluster-users at gluster.org > > > > Subject: Re: Rsync > > > > > > > > Hi Hiren, > > > > What glusterfs version are you using? Can you send us the > > > > volfiles and the log files. > > > > > > > > Pavan > > > > > > > > On 22/09/09 16:01 +0100, Hiren Joshi wrote: > > > > > I forgot to mention, the mount is mounted with > > direct-io, would this > > > > > make a difference? > > > > > > > > > > > -----Original Message----- > > > > > > From: gluster-users-bounces at gluster.org > > > > > > [mailto:gluster-users-bounces at gluster.org] On Behalf Of > > > > Hiren Joshi > > > > > > Sent: 22 September 2009 11:40 > > > > > > To: gluster-users at gluster.org > > > > > > Subject: Rsync > > > > > > > > > > > > Hello all, > > > > > > > > > > > > I'm getting what I think is bizarre behaviour.... I have > > > > about 400G to > > > > > > rsync (rsync -av) onto a gluster share, the data is > > in a directory > > > > > > structure which has about 1000 directories per parent and > > > > about 1000 > > > > > > directories in each of them. > > > > > > > > > > > > When I try to rsync an end leaf directory (this has about 4 > > > > > > dirs and 100 > > > > > > files in each) the operation takes about 10 seconds. When I > > > > > > go one level > > > > > > above (1000 dirs with about 4 dirs in each with about 100 > > > > > > files in each) > > > > > > the operation takes about 10 minutes. > > > > > > > > > > > > Now, if I then go one level above that (that's 1000 > dirs with > > > > > > 1000 dirs > > > > > > in each with about 4 dirs in each with about 100 files in > > > > each) the > > > > > > operation takes days! Top shows glusterfsd takes 300-600% > > > > cpu usage > > > > > > (2X4core), I have about 48G of memory (usage is 0% as > > expected). > > > > > > > > > > > > Has anyone seen anything like this? How can I speed it up? > > > > > > > > > > > > Thanks, > > > > > > > > > > > > Josh. > > > > > > > > > > > _______________________________________________ > > > > > Gluster-users mailing list > > > > > Gluster-users at gluster.org > > > > > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > > > > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >