Done..... but now have this error in /var/log/glusterfs/mnt-mail.log [2010-05-11 20:40:21] E [quick-read.c:2194:init] quickread: FATAL: volume (quickread) not configured with exactly one child [2010-05-11 20:40:21] E [xlator.c:839:xlator_init_rec] quickread: Initialization of volume 'quickread' failed, review your volfile again [2010-05-11 20:40:21] E [glusterfsd.c:591:_xlator_graph_init] glusterfs: initializing translator failed [2010-05-11 20:40:21] E [glusterfsd.c:1394:main] glusterfs: translator initialization failed. exiting I did change the option max-file-size but when I received the errors I put it back to 64kb. New glusterfs.vol ## file auto generated by /bin/glusterfs-volgen (mount.vol) # Cmd line: # $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs 192.168.1.16:/fs # RAID 1 # TRANSPORT-TYPE tcp volume 192.168.1.16-1 type protocol/client option transport-type tcp option remote-host 192.168.1.16 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume brick1 end-volume volume 192.168.1.15-1 type protocol/client option transport-type tcp option remote-host 192.168.1.15 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume brick1 end-volume volume mirror-0 type cluster/replicate subvolumes 192.168.1.15-1 192.168.1.16-1 end-volume #volume readahead # type performance/read-ahead # option page-count 4 # subvolumes mirror-0 #end-volume #volume iocache # type performance/io-cache # option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed 's/[^0-9]//g') / 5120 ))`MB # option cache-timeout 1 # subvolumes readahead #end-volume volume quickread type performance/quick-read option cache-timeout 10 option max-file-size 64kB # subvolumes iocache end-volume volume writebehind type performance/write-behind option cache-size 4MB subvolumes quickread end-volume volume statprefetch type performance/stat-prefetch subvolumes writebehind end-volume Bryan McGuire On May 11, 2010, at 7:37 PM, Craig Carl wrote: > Bryan - > Your server vol file isn't perfect for large rsync operations. The > changes I'm recommending will improve your rsync performance if you > are moving a lot of small files. Please backup the current file > before making any changes. You should comment out "readahead" and > "iocache". In the quickread section change the "option cache- > timeout" to 10 and change the "max-file-size" to the size of the > largest file of which you have many, rounded up to the nearest > factor of 4. > After you have made the changes across all the storage nodes > please restart Gluster and measure the throughput again. > > Thanks, > > Craig > > > > ----- Original Message ----- > From: "Bryan McGuire" <bmcguire at newnet66.org> > To: "Craig Carl" <craig at gluster.com> > Cc: gluster-users at gluster.org > Sent: Tuesday, May 11, 2010 2:31:48 PM GMT -08:00 US/Canada Pacific > Subject: Re: mnt-mail.log > > Here they are, > > > For msvr1 - 192.168.1.15 > > ## file auto generated by /bin/glusterfs-volgen (export.vol) > # Cmd line: > # $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs > 192.168.1.16:/fs > > volume posix1 > type storage/posix > option directory /fs > end-volume > > volume locks1 > type features/locks > subvolumes posix1 > end-volume > > volume brick1 > type performance/io-threads > option thread-count 8 > subvolumes locks1 > end-volume > > volume server-tcp > type protocol/server > option transport-type tcp > option auth.addr.brick1.allow * > option transport.socket.listen-port 6996 > option transport.socket.nodelay on > subvolumes brick1 > end-volume > > > ## file auto generated by /bin/glusterfs-volgen (mount.vol) > # Cmd line: > # $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs > 192.168.1.16:/fs > > # RAID 1 > # TRANSPORT-TYPE tcp > volume 192.168.1.16-1 > type protocol/client > option transport-type tcp > option remote-host 192.168.1.16 > option transport.socket.nodelay on > option transport.remote-port 6996 > option remote-subvolume brick1 > end-volume > > volume 192.168.1.15-1 > type protocol/client > option transport-type tcp > option remote-host 192.168.1.15 > option transport.socket.nodelay on > option transport.remote-port 6996 > option remote-subvolume brick1 > end-volume > > volume mirror-0 > type cluster/replicate > subvolumes 192.168.1.15-1 192.168.1.16-1 > end-volume > > volume readahead > type performance/read-ahead > option page-count 4 > subvolumes mirror-0 > end-volume > > volume iocache > type performance/io-cache > option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed > 's/[^0-9]//g') / 5120 ))`MB > option cache-timeout 1 > subvolumes readahead > end-volume > > volume quickread > type performance/quick-read > option cache-timeout 1 > option max-file-size 64kB > subvolumes iocache > end-volume > > volume writebehind > type performance/write-behind > option cache-size 4MB > subvolumes quickread > end-volume > > volume statprefetch > type performance/stat-prefetch > subvolumes writebehind > end-volume > > > For msvr2 192.168.1.16 > > ## file auto generated by /bin/glusterfs-volgen (export.vol) > # Cmd line: > # $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs > 192.168.1.16:/fs > > volume posix1 > type storage/posix > option directory /fs > end-volume > > volume locks1 > type features/locks > subvolumes posix1 > end-volume > > volume brick1 > type performance/io-threads > option thread-count 8 > subvolumes locks1 > end-volume > > volume server-tcp > type protocol/server > option transport-type tcp > option auth.addr.brick1.allow * > option transport.socket.listen-port 6996 > option transport.socket.nodelay on > subvolumes brick1 > end-volume > > > ## file auto generated by /bin/glusterfs-volgen (mount.vol) > # Cmd line: > # $ /bin/glusterfs-volgen --name repstore1 --raid 1 192.168.1.15:/fs > 192.168.1.16:/fs > > # RAID 1 > # TRANSPORT-TYPE tcp > volume 192.168.1.16-1 > type protocol/client > option transport-type tcp > option remote-host 192.168.1.16 > option transport.socket.nodelay on > option transport.remote-port 6996 > option remote-subvolume brick1 > end-volume > > volume 192.168.1.15-1 > type protocol/client > option transport-type tcp > option remote-host 192.168.1.15 > option transport.socket.nodelay on > option transport.remote-port 6996 > option remote-subvolume brick1 > end-volume > > volume mirror-0 > type cluster/replicate > subvolumes 192.168.1.15-1 192.168.1.16-1 > end-volume > > volume readahead > type performance/read-ahead > option page-count 4 > subvolumes mirror-0 > end-volume > > volume iocache > type performance/io-cache > option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed > 's/[^0-9]//g') / 5120 ))`MB > option cache-timeout 1 > subvolumes readahead > end-volume > > volume quickread > type performance/quick-read > option cache-timeout 1 > option max-file-size 64kB > subvolumes iocache > end-volume > > volume writebehind > type performance/write-behind > option cache-size 4MB > subvolumes quickread > end-volume > > volume statprefetch > type performance/stat-prefetch > subvolumes writebehind > end-volume > > > > > > > Bryan McGuire > Senior Network Engineer > NewNet 66 > > 918.231.8063 > bmcguire at newnet66.org > > > > > > On May 11, 2010, at 4:26 PM, Craig Carl wrote: > >> Bryan - >> Can you send your client and server vol files? >> >> Thanks, >> >> Craig >> >> -- >> Craig Carl >> Sales Engineer >> Gluster, Inc. >> Cell - (408) 829-9953 (California, USA) >> Office - (408) 770-1884 >> Gtalk - craig.carl at gmail.com >> Twitter - @gluster >> >> ----- Original Message ----- >> From: "Bryan McGuire" <bmcguire at newnet66.org> >> To: gluster-users at gluster.org >> Sent: Tuesday, May 11, 2010 2:12:13 PM GMT -08:00 US/Canada Pacific >> Subject: mnt-mail.log >> >> Hello, >> >> I have Glusterfs 3.0.4 setup in a two node replication. It appears to >> be working just fine. Although I am using rsync to move over 350 Gig >> of email files and the process is very slow. >> >> I have noticed the following in the file /var/log/gluserfs/mntl- >> mail.log....... could someone explain what the lines mean. Thanks >> >> [2010-05-11 15:41:51] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs- >> fuse: LOOKUP(/_outgoing/retry/201005100854105298-1273614110_8.tmp) >> inode (ptr=0xa235c70, ino=808124434, gen=5468694309383486383) found >> conflict (ptr=0x2aaaea26c290, ino=808124434, gen=5468694309383486383) >> [2010-05-11 15:46:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs- >> fuse: LOOKUP(/_outgoing/retry/201005101016464462-1273614395_8.tmp) >> inode (ptr=0x2aaaf07f5550, ino=808124438, gen=5468694309383486385) >> found conflict (ptr=0x82e4420, ino=808124438, >> gen=5468694309383486385) >> [2010-05-11 15:46:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs- >> fuse: LOOKUP(/_outgoing/retry/201005100830599960-1273614395_8.tmp) >> inode (ptr=0x2aaac01da520, ino=808124430, gen=5468694309383486381) >> found conflict (ptr=0x60d7a90, ino=808124430, >> gen=5468694309383486381) >> [2010-05-11 15:46:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs- >> fuse: LOOKUP(/_outgoing/retry/201005101417175132-1273614396_8.tmp) >> inode (ptr=0x2aaaf07f5550, ino=808124446, gen=5468694309383486389) >> found conflict (ptr=0x8eb16e0, ino=808124446, >> gen=5468694309383486389) >> [2010-05-11 15:51:53] W [fuse-bridge.c:491:fuse_entry_cbk] glusterfs- >> fuse: LOOKUP(/_outgoing/retry/201005100749045904-1273614665_8.tmp) >> inode (ptr=0x1ec11ee0, ino=808124420, gen=5468694309383486379) found >> conflict (ptr=0x2aaaea26bd30, ino=808124420, gen=5468694309383486379) >> >> >> >> Bryan >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > > > -- > > > > Craig Carl > > > > Join us for a Webinar on May 26; > Case Studies: Deploying Open Source Storage Clouds > Sales Engineer > Gluster, Inc. > Cell - (408) 829-9953 (California, USA) > Office - (408) 770-1884 > Gtalk - craig.carl at gmail.com > Twitter - @gluster > http://www.gluster.com/files/installation-demo/demo.html