No, the context is quite appropriate. Linus has made it clear for a very long time that the kernel is not to break userspace programs with no exceptions. That said, however, Avati and Ted have been working together to implement a fix in the kernel. No idea on an eta. On 03/11/2013 10:18 AM, Rodrigo Severo wrote: > I know where this statement came from. I believe you are both: > > * trying to apply some statement on a context it's not pertinent to and > * fouling yourself and/or others arguing that this issue will/should > be fixed in the kernel. > > The ext4 hash size change was applied in the kernel an year ago. I > don't believe it will be undone. Gluster developers could argue that > this change was hard on them, and that it shouldn't be backported to > Enterprise kernels but after one year not having fixed it is on > Gluster developers. Arguing otherwise seems rather foolish to me. > > > Regards, > > Rodrigo Severo > > > > On Mon, Mar 11, 2013 at 2:00 PM, Joe Julian <joe at julianfamily.org > <mailto:joe at julianfamily.org>> wrote: > > This bug is in the kernel. > > "If a change results in user programs breaking, it's a bug in the > kernel. We never EVER blame the user programs." - Linus Torvalds > > http://lkml.org/lkml/2012/12/23/75 > > On 03/08/2013 12:42 PM, Stephan von Krawczynski wrote: >> I really do wonder if this bug in _glusterfs_ is not fixed. It really makes no >> sense to do an implementation that breaks on the most used fs on linux. >> And just as you said: don't wait on btrfs, it will never be production-ready. >> And xfs is no solution, it is just a bad work-around. >> >> >> On Fri, 8 Mar 2013 10:43:41 -0800 >> Bryan Whitehead<driver at megahappy.net> <mailto:driver at megahappy.net> wrote: >> >>> Here are some details about ext4 changes in the kernel screwing up >>> glusterfs: >>> http://www.gluster.org/2012/08/glusterfs-bit-by-ext4-structure-change/ >>> https://bugzilla.redhat.com/show_bug.cgi?id=838784 >>> >>> I thought I read there was a work-around in recent versions of gluster but >>> I think it came at a cost somewhere. I'm not sure since I've been using xfs >>> since the 1.x days of gluster and only see random ext3/4 problems bubble up >>> on these maillinglist. In general, ext4 was just a stopgap for the wait on >>> btrfs getting flushed out. That said, I don't see ext4 going away for a >>> long long time. :-/ >>> >>> NOTE: I don't even know if this is your problem. You might try updating 2 >>> bricks that are replica pairs to use xfs then do some performance tests on >>> files living on them to confirm. Example, you have 20 some servers/bricks. >>> If hostD and hostE are replica pairs for some subset of files, shutdown >>> glusterd on HostD, change fs to xfs, fire glusterd back up - let it resync >>> and recover all the files, do the same on hostE (once hostD is good), then >>> see if there is a read speed improvement for files living on those two host >>> pairs. >>> >>> >>> On Fri, Mar 8, 2013 at 6:40 AM, Thomas Wakefield<twake at cola.iges.org> <mailto:twake at cola.iges.org>wrote: >>> >>>> I am still confused how ext4 is suddenly slow to read when it's behind >>>> Gluster, but plenty fast stand alone reading? >>>> >>>> And it writes really fast from both the server and client. >>>> >>>> >>>> >>>> On Mar 8, 2013, at 4:07 AM, "Jon Tegner"<tegner at renget.se> <mailto:tegner at renget.se> wrote: >>>> >>>> We had issues with ext4 about a bit less than a year ago, at that time I >>>> upgraded the servers to CentOS-6.2. But that gave us large problems (more >>>> than slow reads). Since I didn't want to reformat the disks at that time >>>> (and switch to XFS) I went back to CentOS-5.5 (which we had used before). >>>> On some link (think it was >>>> https://bugzilla.redhat.com/show_bug.cgi?id=713546 but can't seem to >>>> reach that now) it was stated that the ext4-issue was present even on later >>>> versions of CentOS-5 (I _think_ 5.8 was affected). >>>> >>>> Are there hope that the "ext4-issue" will be solved in later >>>> kernels/versions of gluster? If not, it seems one is eventually forced to >>>> switch to XFS. >>>> >>>> Regards, >>>> >>>> /jon >>>> >>>> On Mar 8, 2013 03:27 "Thomas Wakefield"<twake at iges.org> <mailto:twake at iges.org> <twake at iges.org> <mailto:twake at iges.org>wrote: >>>> >>>> >>>> inode size is 256. >>>> >>>> >>>> >>>> >>>> Pretty stuck with these settings and ext4. I missed the memo that >>>> Gluster started to prefer xfs, back in the 2.x days xfs was not the >>>> preferred filesystem. At this point it's a 340TB filesystem with 160TB >>>> used. I just added more space, and was doing some followup testing and >>>> wasn't impressed with the results. But I am sure I was happier before >>>> with the performance. >>>> >>>> >>>> >>>> >>>> Still running CentOS 5.8 >>>> >>>> >>>> >>>> >>>> Anything else I could look at? >>>> >>>> >>>> >>>> >>>> Thanks, Tom >>>> >>>> >>>> >>>> >>>> >>>> On Mar 7, 2013, at 5:04 PM, Bryan Whitehead <<driver at megahappy.net> <mailto:driver at megahappy.net>> >>>> wrote: >>>> >>>> >>>>> I'm sure you know, but xfs is the recommended filesystem for >>>>> glusterfs. Ext4 has a number of issues. (Particularly on >>>>> CentOS/Redhat6). >>>>> >>>>> >>>>> >>>>> The default inode size for ext4 (and xfs) is small for the number of >>>>> extended attributes glusterfs uses. This causes a minor hit in >>>>> performance on xfs if theextended attributes grow more than 265 (xfs >>>>> default size). In xfs, this is fixed by setting the size of an inode >>>>> to 512. How big the impact is on ext4 is something I don't know >>>>> offhand. But looking at a couple of boxes I have it looks like some >>>>> ext4 filesystems have 128 inode size and some have 256 inode size >>>>> (both of which are too small for glusterfs). The performance hit is >>>>> everytimeextended attributes need to be read several inodes need to be >>>>> seeked and found. >>>>> >>>>> >>>>> >>>>> run "dumpe2fs -h <blockdevice> | grep size" on your ext4 mountpoints. >>>>> >>>>> >>>>> >>>>> >>>>> If it is not too much of a bother - I'd try xfs as your filesystem for >>>>> the bricks >>>>> >>>>> >>>>> >>>>> >>>>> mkfs.xfs -i size=512 <blockdevice> >>>>> >>>>> >>>>> >>>>> >>>>> Please see this for more detailed info: >>>>> >>>>> <https://access.redhat.com/knowledge/docs/en-US/Red_Hat_Storage/2.0/ht >>>>> ml-single/Administration_Guide/index.html#chap-User_Guide-Setting_Volu >>>>> mes> <https://access.redhat.com/knowledge/docs/en-US/Red_Hat_Storage/2.0/html-single/Administration_Guide/index.html#chap-User_Guide-Setting_Volumes> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> On Thu, Mar 7, 2013 at 12:08 PM, Thomas Wakefield >>>>> <<twake at cola.iges.org> <mailto:twake at cola.iges.org>> wrote: >>>>> >>>>>> Everything is built as ext4, no options other than >>>>>> lazy_itable_init=1 when I built the filesystems. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Server mount example: >>>>>> >>>>>> LABEL=disk2a /storage/disk2a ext4defaults 0 0 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Client mount: >>>>>> >>>>>> fs-disk2:/shared /shared glusterfs defaults 0 0 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Remember, the slow reads are only from gluster clients, the disks >>>>>> are really fast when I am local on the server testing the disks. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -Tom >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Mar 7, 2013, at 1:09 PM, Bryan Whitehead <<driver at megahappy.net> <mailto:driver at megahappy.net>> >>>>>> wrote: >>>>>> >>>>>> >>>>>>> Was just thinking, what is your mount options for your bricks >>>>>>> (using inode64?)? Also, you are using xfs... right? >>>>>>> >>>>>>> >>>>>>> >>>>>>> When you created the filesystems did you allocate more inode >>>>>>> space? -i size=512 ? >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Thu, Mar 7, 2013 at 5:49 AM, Thomas Wakefield >>>>>>> <<twake at cola.iges.org> <mailto:twake at cola.iges.org>> wrote: >>>>>>> >>>>>>>> Still looking for help. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Mar 4, 2013, at 7:43 AM, Thomas Wakefield <<twake at iges.org> <mailto:twake at iges.org>> >>>>>>>> wrote: >>>>>>>> >>>>>>>> >>>>>>>>> Also, I tested an NFS mount over the same 10GB link, and was >>>>>>>>> able to pull almost 200MB/s. But Gluster is still much slower. >>>>>>>>> Also I tested running it for a longer test, 105GB of data, and >>>>>>>>> still showed that writing is MUCH faster. Which makes no sense >>>>>>>>> when the disks can read 2x as fast as they can write. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Any other thoughts? >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> [root at cpu_crew1 ~]# dd if=/dev/zero >>>>>>>>> of=/shared/working/benchmark/test.cpucrew1 bs=512k >>>>>>>>> count=200000 ; dd if=/shared/working/benchmark/test.cpucrew1 >>>>>>>>> of=/dev/null bs=512k >>>>>>>>> >>>>>>>>> 200000+0 records in >>>>>>>>> >>>>>>>>> 200000+0 records out >>>>>>>>> >>>>>>>>> 104857600000 bytes (105 GB) copied, 159.135 seconds, 659 MB/s >>>>>>>>> >>>>>>>>> 200000+0 records in >>>>>>>>> >>>>>>>>> 200000+0 records out >>>>>>>>> >>>>>>>>> 104857600000 bytes (105 GB) copied, 1916.87 seconds, 54.7 MB/s >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mar 1, 2013, at 9:58 AM, Thomas Wakefield >>>>>>>>> <<twake at iges.org> <mailto:twake at iges.org>> wrote: >>>>>>>>> >>>>>>>>> >>>>>>>>>> The max setting for performance.read-ahead-page-count is 16, >>>>>>>>>> which I did just try. No significant change. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Any other setting options? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Feb 28, 2013, at 10:18 PM, Anand Avati >>>>>>>>>> <<anand.avati at gmail.com> <mailto:anand.avati at gmail.com>> wrote: >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> Can you try "gluster volume set <volname> >>>>>>>>>>> performance.read-ahead-page-count 64" or some value higher >>>>>>>>>>> or lower? >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Avati >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Thu, Feb 28, 2013 at 7:15 PM, Thomas Wakefield >>>>>>>>>>> <<twake at iges.org> <mailto:twake at iges.org>> wrote: >>>>>>>>>>> >>>>>>>>>>>> Good point, forgot to set a blcoksize, here are the >>>>>>>>>>>> redone dd tests: >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> [root at cpu_crew1 ~]# dd >>>>>>>>>>>> if=/shared/working/benchmark/test.cpucrew1 of=/dev/null >>>>>>>>>>>> bs=128k >>>>>>>>>>>> >>>>>>>>>>>> 40000+0 records in >>>>>>>>>>>> >>>>>>>>>>>> 40000+0 records out >>>>>>>>>>>> >>>>>>>>>>>> 5242880000 bytes (5.2 GB) copied, 65.4928 seconds, 80.1 >>>>>>>>>>>> MB/s >>>>>>>>>>>> >>>>>>>>>>>> [root at cpu_crew1 ~]# dd >>>>>>>>>>>> if=/shared/working/benchmark/test.cpucrew1 of=/dev/null >>>>>>>>>>>> bs=1M >>>>>>>>>>>> >>>>>>>>>>>> 5000+0 records in >>>>>>>>>>>> >>>>>>>>>>>> 5000+0 records out >>>>>>>>>>>> >>>>>>>>>>>> 5242880000 bytes (5.2 GB) copied, 49.0907 seconds, 107 >>>>>>>>>>>> MB/s >>>>>>>>>>>> >>>>>>>>>>>> [root at cpu_crew1 ~]# dd >>>>>>>>>>>> if=/shared/working/benchmark/test.cpucrew1 of=/dev/null >>>>>>>>>>>> bs=4M >>>>>>>>>>>> >>>>>>>>>>>> 1250+0 records in >>>>>>>>>>>> >>>>>>>>>>>> 1250+0 records out >>>>>>>>>>>> >>>>>>>>>>>> 5242880000 bytes (5.2 GB) copied, 44.5724 seconds, 118 >>>>>>>>>>>> MB/s >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Still not impressive. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -Tom >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Feb 28, 2013, at 8:42 PM, Jeff Anderson-Lee >>>>>>>>>>>> <<jonah at eecs.berkeley.edu> <mailto:jonah at eecs.berkeley.edu>> wrote: >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> Thomas, >>>>>>>>>>>>> >>>>>>>>>>>>> You have not specified a block size, so you are doing >>>>>>>>>>>>> a huge number of small(ish) reads with associated >>>>>>>>>>>>> round trips. What happens with dd bs=128k ..? >>>>>>>>>>>>> >>>>>>>>>>>>> Jeff Anderson-Lee >>>>>>>>>>>>> >>>>>>>>>>>>> On 2/28/2013 5:30 PM, Thomas Wakefield wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> Did a fresh dd test just to confirm, same results: >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> [root at cpu_crew1 benchmark]# dd if=/dev/zero >>>>>>>>>>>>>> of=/shared/working/benchmark/test.cpucrew1 bs=512k >>>>>>>>>>>>>> count=10000 >>>>>>>>>>>>>> >>>>>>>>>>>>>> 10000+0 records in >>>>>>>>>>>>>> >>>>>>>>>>>>>> 10000+0 records out >>>>>>>>>>>>>> >>>>>>>>>>>>>> 5242880000 bytes (5.2 GB) copied, 7.43695 seconds, >>>>>>>>>>>>>> 705 MB/s >>>>>>>>>>>>>> >>>>>>>>>>>>>> [root at cpu_crew1 benchmark]# dd >>>>>>>>>>>>>> if=/shared/working/benchmark/test.cpucrew1 >>>>>>>>>>>>>> of=/dev/null >>>>>>>>>>>>>> >>>>>>>>>>>>>> 552126+0 records in >>>>>>>>>>>>>> >>>>>>>>>>>>>> 552125+0 records out >>>>>>>>>>>>>> >>>>>>>>>>>>>> 282688000 bytes (283 MB) copied, 37.8514 seconds, >>>>>>>>>>>>>> 7.5 MB/s >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Feb 28, 2013, at 8:14 PM, Bryan Whitehead >>>>>>>>>>>>>> <<driver at megahappy.net> <mailto:driver at megahappy.net>> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> How are you doing the reading? Is this still an >>>>>>>>>>>>>>> iozone benchmark? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> if you simply dd if=/glustermount/bigfile >>>>>>>>>>>>>>> of=/dev/null, is the speed better? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Thu, Feb 28, 2013 at 5:05 PM, Thomas Wakefield >>>>>>>>>>>>>>> <<twake at iges.org> <mailto:twake at iges.org>> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I get great speed locally, it's only when I add >>>>>>>>>>>>>>>> gluster in that it slows down. I get 2GB/s >>>>>>>>>>>>>>>> locally to the exact same brick. It's gluster >>>>>>>>>>>>>>>> that is having the read issue (80MB/s). But >>>>>>>>>>>>>>>> Gluster can write just fine, 800MB/s. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> The blockdev idea is a good one, and I have >>>>>>>>>>>>>>>> already done it. Thanks though. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -Tom >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Feb 28, 2013, at 7:53 PM, Ling Ho >>>>>>>>>>>>>>>> <<ling at slac.stanford.edu> <mailto:ling at slac.stanford.edu>> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Tom, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> What type of disks do you have? If they are >>>>>>>>>>>>>>>>> raid 5 or 6, have you try setting the >>>>>>>>>>>>>>>>> read-ahead size to 8192 or 16384 (blockdev >>>>>>>>>>>>>>>>> --setra 8192 /dev/<sd?> ? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>>>> ling >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On 02/28/2013 04:23 PM, Thomas Wakefield >>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>> Did anyone else have any ideas on >>>>>>>>>>>>>>>>>> performance tuning for reads? >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On Feb 27, 2013, at 9:29 PM, Thomas >>>>>>>>>>>>>>>>>> Wakefield <<twake at iges.org> <mailto:twake at iges.org>> wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Bryan- >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Yes I can write at 700-800MBytes/sec, but >>>>>>>>>>>>>>>>>>> i can only read at 70-80 MBytes/sec. I >>>>>>>>>>>>>>>>>>> would be very happy if I could get it to >>>>>>>>>>>>>>>>>>> read at the same speed it can write at. >>>>>>>>>>>>>>>>>>> And the 70-80 is sequential, not random >>>>>>>>>>>>>>>>>>> for reads, same exact test commands on the >>>>>>>>>>>>>>>>>>> disk server are in the 2+GB/s range, so I >>>>>>>>>>>>>>>>>>> know the disk server can do it. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -Tom >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Feb 27, 2013, at 7:41 PM, Bryan >>>>>>>>>>>>>>>>>>> Whitehead <<driver at megahappy.net> <mailto:driver at megahappy.net>> wrote: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Are your figures 700-800MByte/sec? >>>>>>>>>>>>>>>>>>>> Because that is probably as fast as your >>>>>>>>>>>>>>>>>>>> 10G nic cards are able to do. You can >>>>>>>>>>>>>>>>>>>> test that by trying to push a large >>>>>>>>>>>>>>>>>>>> amount of data over nc or ftp. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Might want to try Infiniband. 40G cards >>>>>>>>>>>>>>>>>>>> are pretty routine. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> On Wed, Feb 27, 2013 at 3:45 PM, Thomas >>>>>>>>>>>>>>>>>>>> Wakefield <<twake at iges.org> <mailto:twake at iges.org>> wrote: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> I also get the same performance >>>>>>>>>>>>>>>>>>>>> running iozone for large file >>>>>>>>>>>>>>>>>>>>> sizes, iozone -u 1 -r 512k -s 2G -I >>>>>>>>>>>>>>>>>>>>> -F. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Large file IO is what I need the >>>>>>>>>>>>>>>>>>>>> system to do. I am just shocked at the >>>>>>>>>>>>>>>>>>>>> huge difference between local IO and >>>>>>>>>>>>>>>>>>>>> gluster client IO. I know there should >>>>>>>>>>>>>>>>>>>>> be some difference, but 10x is >>>>>>>>>>>>>>>>>>>>> unacceptable. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> -Tom >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> On Feb 27, 2013, at 5:31 PM, Bryan >>>>>>>>>>>>>>>>>>>>> Whitehead <<driver at megahappy.net> <mailto:driver at megahappy.net>> >>>>>>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Every time you open/close a file or >>>>>>>>>>>>>>>>>>>>>> a directory you will have to wait >>>>>>>>>>>>>>>>>>>>>> for locks which take time. This is >>>>>>>>>>>>>>>>>>>>>> totally expected. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Why don't you share what you want to >>>>>>>>>>>>>>>>>>>>>> do? iozone benchmarks look like crap >>>>>>>>>>>>>>>>>>>>>> but serving qcow2 files to qemu >>>>>>>>>>>>>>>>>>>>>> works fantastic for me. What are you >>>>>>>>>>>>>>>>>>>>>> doing? Make a benchmark that does >>>>>>>>>>>>>>>>>>>>>> that. If you are going to have many >>>>>>>>>>>>>>>>>>>>>> files with a wide variety of sizes >>>>>>>>>>>>>>>>>>>>>> glusterfs/fuse might not be what you >>>>>>>>>>>>>>>>>>>>>> are looking for. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> On Wed, Feb 27, 2013 at 12:56 PM, >>>>>>>>>>>>>>>>>>>>>> Thomas Wakefield >>>>>>>>>>>>>>>>>>>>>> <<twake at cola.iges.org> <mailto:twake at cola.iges.org>> wrote: >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> I have tested everything, small >>>>>>>>>>>>>>>>>>>>>>> and large files. I have used file >>>>>>>>>>>>>>>>>>>>>>> sizes ranging from 128k up to >>>>>>>>>>>>>>>>>>>>>>> multiple GB files. All the reads >>>>>>>>>>>>>>>>>>>>>>> are bad. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Here is a fairly exhaustive iozone >>>>>>>>>>>>>>>>>>>>>>> auto test: >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> random random bkwd record stride >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> KB reclen write rewrite read >>>>>>>>>>>>>>>>>>>>>>> reread read write read rewrite >>>>>>>>>>>>>>>>>>>>>>> read fwrite frewrite fread freread >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 64 4 40222 63492 26868 30060 1620 >>>>>>>>>>>>>>>>>>>>>>> 71037 1572 70570 31294 77096 72475 >>>>>>>>>>>>>>>>>>>>>>> 14736 13928 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 64 8 99207 116366 13591 13513 3214 >>>>>>>>>>>>>>>>>>>>>>> 97690 3155 109978 >>>>>>>>>>>>>>>>>>>>>>> <tel:3155%20%C2%A0%20109978> 28920 >>>>>>>>>>>>>>>>>>>>>>> 152018 158480 18936 17625 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 64 16 230257 253766 25156 28713 >>>>>>>>>>>>>>>>>>>>>>> 10867 223732 8873 244297 54796 >>>>>>>>>>>>>>>>>>>>>>> 303383 312204 15062 13545 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 64 32 255943 234481 5735102 >>>>>>>>>>>>>>>>>>>>>>> 7100397 11897 318502 13681 347801 >>>>>>>>>>>>>>>>>>>>>>> 24214 695778 528618 25838 28094 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 64 64 214096 681644 6421025 >>>>>>>>>>>>>>>>>>>>>>> 7100397 27453 292156 28117 621657 >>>>>>>>>>>>>>>>>>>>>>> 27338 376062 512471 28569 32534 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 128 4 74329 75468 26428 >>>>>>>>>>>>>>>>>>>>>>> <tel:75468%20%C2%A0%20%C2%A026428> >>>>>>>>>>>>>>>>>>>>>>> 41089 1131 72857 1118 66976 1597 >>>>>>>>>>>>>>>>>>>>>>> 73778 78343 13351 13026 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 128 8 100862 135170 24966 >>>>>>>>>>>>>>>>>>>>>>> <tel:135170%20%C2%A0%20%C2%A024966 >>>>>>>>>>>>>>>>>>>>>>>> 16734 2617 118966 2560 120406 >>>>>>>>>>>>>>>>>>>>>>>> 39156 125121 146613 16177 16180 >>>>>>>>>>>>>>>>>>>>>>> 128 16 115114 253983 28212 17854 >>>>>>>>>>>>>>>>>>>>>>> 5307 246180 5431 229843 47335 >>>>>>>>>>>>>>>>>>>>>>> 255920 271173 27256 24445 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 128 32 256042 391360 39848 64258 >>>>>>>>>>>>>>>>>>>>>>> 11329 290230 9905 429563 38176 >>>>>>>>>>>>>>>>>>>>>>> 490380 463696 20917 19219 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 128 64 248573 592699 4557257 >>>>>>>>>>>>>>>>>>>>>>> 6812590 19583 452366 29263 603357 >>>>>>>>>>>>>>>>>>>>>>> 42967 814915 692017 76327 37604 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 128 128 921183 526444 5603747 >>>>>>>>>>>>>>>>>>>>>>> 5379161 45614 390222 65441 826202 >>>>>>>>>>>>>>>>>>>>>>> 41384 662962 1040839 78526 39023 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 256 4 76212 77337 40295 >>>>>>>>>>>>>>>>>>>>>>> <tel:77337%20%C2%A0%20%C2%A040295> >>>>>>>>>>>>>>>>>>>>>>> 32125 1289 71866 1261 64645 1436 >>>>>>>>>>>>>>>>>>>>>>> 57309 53048 23073 29550 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 256 8 126922 141976 26237 >>>>>>>>>>>>>>>>>>>>>>> <tel:141976%20%C2%A0%20%C2%A026237 >>>>>>>>>>>>>>>>>>>>>>>> 25130 2566 128058 2565 138981 >>>>>>>>>>>>>>>>>>>>>>>> <tel:2565%20%C2%A0%20138981> >>>>>>>>>>>>>>>>>>>>>>>> 2985 125060 133603 22840 24955 >>>>>>>>>>>>>>>>>>>>>>> 256 16 242883 263636 41850 24371 >>>>>>>>>>>>>>>>>>>>>>> 4902 250009 5290 248792 89353 >>>>>>>>>>>>>>>>>>>>>>> 243821 247303 26965 26199 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 256 32 409074 439732 40101 39335 >>>>>>>>>>>>>>>>>>>>>>> 11953 436870 11209 430218 83743 >>>>>>>>>>>>>>>>>>>>>>> 409542 479390 30821 27750 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 256 64 259935 571502 64840 71847 >>>>>>>>>>>>>>>>>>>>>>> 22537 <tel:71847%20%C2%A0%2022537> >>>>>>>>>>>>>>>>>>>>>>> 617161 23383 392047 91852 672010 >>>>>>>>>>>>>>>>>>>>>>> 802614 41673 53111 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 256 128 847597 812329 185517 83198 >>>>>>>>>>>>>>>>>>>>>>> 49383 <tel:83198%20%C2%A0%2049383> >>>>>>>>>>>>>>>>>>>>>>> 708831 44668 794889 74267 1180188 >>>>>>>>>>>>>>>>>>>>>>> 1662639 54303 41018 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 256 256 481324 709299 5217259 >>>>>>>>>>>>>>>>>>>>>>> 5320671 44668 719277 40954 808050 >>>>>>>>>>>>>>>>>>>>>>> 41302 790209 771473 62224 35754 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 512 4 77667 75226 35102 29696 1337 >>>>>>>>>>>>>>>>>>>>>>> 66262 1451 67680 1413 69265 69142 >>>>>>>>>>>>>>>>>>>>>>> 42084 27897 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 512 8 134311 144341 30144 >>>>>>>>>>>>>>>>>>>>>>> <tel:144341%20%C2%A0%20%C2%A030144 >>>>>>>>>>>>>>>>>>>>>>>> 24646 2102 134143 2209 134699 >>>>>>>>>>>>>>>>>>>>>>>> 2296 108110 128616 25104 29123 >>>>>>>>>>>>>>>>>>>>>>> 512 16 200085 248787 30235 25697 >>>>>>>>>>>>>>>>>>>>>>> 4196 247240 4179 256116 >>>>>>>>>>>>>>>>>>>>>>> <tel:4179%20%C2%A0%20256116> 4768 >>>>>>>>>>>>>>>>>>>>>>> 250003 226436 32351 28455 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 512 32 330341 439805 26440 39284 >>>>>>>>>>>>>>>>>>>>>>> 8744 457611 8006 424168 >>>>>>>>>>>>>>>>>>>>>>> <tel:8006%20%C2%A0%20424168> >>>>>>>>>>>>>>>>>>>>>>> 125953 425935 448813 27660 26951 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 512 64 483906 733729 48747 41121 >>>>>>>>>>>>>>>>>>>>>>> 16032 555938 17424 587256 187343 >>>>>>>>>>>>>>>>>>>>>>> 366977 735740 41700 41548 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 512 128 836636 907717 69359 94921 >>>>>>>>>>>>>>>>>>>>>>> 42443 <tel:94921%20%C2%A0%2042443> >>>>>>>>>>>>>>>>>>>>>>> 761031 36828 964378 123165 651383 >>>>>>>>>>>>>>>>>>>>>>> 695697 58368 44459 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 512 256 520879 860437 145534 >>>>>>>>>>>>>>>>>>>>>>> 135523 40267 847532 31585 663252 >>>>>>>>>>>>>>>>>>>>>>> 69696 1270846 1492545 48822 48092 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 512 512 782951 973118 3099691 >>>>>>>>>>>>>>>>>>>>>>> 2942541 42328 871966 46218 911184 >>>>>>>>>>>>>>>>>>>>>>> 49791 953248 1036527 52723 48347 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 1024 4 76218 69362 36431 28711 >>>>>>>>>>>>>>>>>>>>>>> 1137 66171 1174 68938 1125 70566 >>>>>>>>>>>>>>>>>>>>>>> 70845 34942 28914 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 1024 8 126045 140524 37836 >>>>>>>>>>>>>>>>>>>>>>> <tel:140524%20%C2%A0%20%C2%A037836 >>>>>>>>>>>>>>>>>>>>>>>> 15664 2698 126000 2557 125566 >>>>>>>>>>>>>>>>>>>>>>>> 2567 110858 127255 26764 27945 >>>>>>>>>>>>>>>>>>>>>>> 1024 16 243398 261429 40238 23263 >>>>>>>>>>>>>>>>>>>>>>> 3987 246400 3882 260746 4093 >>>>>>>>>>>>>>>>>>>>>>> 236652 236874 31429 25076 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 1024 32 383109 422076 41731 41605 >>>>>>>>>>>>>>>>>>>>>>> 8277 473441 7775 415261 8588 >>>>>>>>>>>>>>>>>>>>>>> 394765 407306 40089 28537 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 1024 64 590145 619156 39623 53267 >>>>>>>>>>>>>>>>>>>>>>> 15051 722717 14624 753000 257294 >>>>>>>>>>>>>>>>>>>>>>> 597784 620946 38619 44073 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 1024 128 1077836 1124099 56192 >>>>>>>>>>>>>>>>>>>>>>> <tel:1124099%20%C2%A0%20%C2%A05619 >>>>>>>>>>>>>>>>>>>>>>> 2> 64916 36851 1102176 37198 >>>>>>>>>>>>>>>>>>>>>>> 1082454 281548 829175 792604 47975 >>>>>>>>>>>>>>>>>>>>>>> 51913 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 1024 256 941918 1074331 72783 >>>>>>>>>>>>>>>>>>>>>>> 81450 26778 >>>>>>>>>>>>>>>>>>>>>>> <tel:81450%20%C2%A0%2026778> >>>>>>>>>>>>>>>>>>>>>>> 1099636 32395 1060013 183218 >>>>>>>>>>>>>>>>>>>>>>> 1024121 995171 44371 45448 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 1024 512 697483 1130312 100324 >>>>>>>>>>>>>>>>>>>>>>> 114682 48215 1041758 41480 1058967 >>>>>>>>>>>>>>>>>>>>>>> 90156 994020 1563622 56328 46370 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 1024 1024 931702 1087111 4609294 >>>>>>>>>>>>>>>>>>>>>>> 4199201 44191 949834 45594 970656 >>>>>>>>>>>>>>>>>>>>>>> 56674 933525 1075676 44876 46115 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 4 71438 67066 58319 38913 >>>>>>>>>>>>>>>>>>>>>>> 1147 44147 1043 42916 967 66416 >>>>>>>>>>>>>>>>>>>>>>> 67205 45953 96750 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 8 141926 134567 61101 55445 >>>>>>>>>>>>>>>>>>>>>>> 2596 77528 2564 80402 4258 124211 >>>>>>>>>>>>>>>>>>>>>>> 120747 53888 100337 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 16 254344 255585 71550 74500 >>>>>>>>>>>>>>>>>>>>>>> 5410 139365 5201 141484 5171 >>>>>>>>>>>>>>>>>>>>>>> 205521 213113 67048 57304 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 32 397833 411261 56676 80027 >>>>>>>>>>>>>>>>>>>>>>> 10440 <tel:80027%20%C2%A0%2010440> >>>>>>>>>>>>>>>>>>>>>>> 260034 10126 230238 10814 391665 >>>>>>>>>>>>>>>>>>>>>>> 383379 79333 60877 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 64 595167 687205 64262 87327 >>>>>>>>>>>>>>>>>>>>>>> 20772 <tel:87327%20%C2%A0%2020772> >>>>>>>>>>>>>>>>>>>>>>> 456430 19960 477064 23190 540220 >>>>>>>>>>>>>>>>>>>>>>> 563096 86812 92565 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 128 833585 933403 121926 >>>>>>>>>>>>>>>>>>>>>>> 118621 37700 690020 37575 733254 >>>>>>>>>>>>>>>>>>>>>>> 567449 712337 734006 92011 104934 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 256 799003 949499 143688 >>>>>>>>>>>>>>>>>>>>>>> 125659 40871 >>>>>>>>>>>>>>>>>>>>>>> <tel:125659%20%C2%A0%2040871> >>>>>>>>>>>>>>>>>>>>>>> 892757 37977 880494 458281 836263 >>>>>>>>>>>>>>>>>>>>>>> 901375 131332 110237 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 512 979936 1040724 120896 >>>>>>>>>>>>>>>>>>>>>>> 138013 54381 859783 48721 780491 >>>>>>>>>>>>>>>>>>>>>>> 279203 1068824 1087085 97886 98078 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 1024 901754 987938 53352 >>>>>>>>>>>>>>>>>>>>>>> 53043 72727 >>>>>>>>>>>>>>>>>>>>>>> <tel:53043%20%C2%A0%2072727> >>>>>>>>>>>>>>>>>>>>>>> 1054522 68269 992275 181253 >>>>>>>>>>>>>>>>>>>>>>> 1309480 1524983 121600 95585 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 2048 2048 831890 1021540 4257067 >>>>>>>>>>>>>>>>>>>>>>> 3302797 75672 984203 80181 826209 >>>>>>>>>>>>>>>>>>>>>>> 94278 966920 1027159 111832 105921 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 4 66195 67316 62171 74785 >>>>>>>>>>>>>>>>>>>>>>> 1328 28963 1329 26397 1223 71470 >>>>>>>>>>>>>>>>>>>>>>> 69317 55903 84915 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 8 122221 120057 90537 60958 >>>>>>>>>>>>>>>>>>>>>>> 2598 47312 2468 59783 2640 128674 >>>>>>>>>>>>>>>>>>>>>>> 127872 41285 40422 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 16 238321 239251 29336 32121 >>>>>>>>>>>>>>>>>>>>>>> 4153 89262 3986 96930 4608 229970 >>>>>>>>>>>>>>>>>>>>>>> 237108 55039 56983 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 32 417110 421356 30974 50000 >>>>>>>>>>>>>>>>>>>>>>> 8382 156676 7886 153841 7900 >>>>>>>>>>>>>>>>>>>>>>> 359585 367288 26611 25952 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 64 648008 668066 32193 29389 >>>>>>>>>>>>>>>>>>>>>>> 14830 273265 14822 282211 19653 >>>>>>>>>>>>>>>>>>>>>>> 581898 620798 51281 50218 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 128 779422 848564 55594 60253 >>>>>>>>>>>>>>>>>>>>>>> 37108 <tel:60253%20%C2%A0%2037108> >>>>>>>>>>>>>>>>>>>>>>> 451296 35908 491361 37567 738163 >>>>>>>>>>>>>>>>>>>>>>> 728059 67681 66440 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 256 865623 886986 71368 63947 >>>>>>>>>>>>>>>>>>>>>>> 44255 <tel:63947%20%C2%A0%2044255> >>>>>>>>>>>>>>>>>>>>>>> 645961 42689 719491 736707 819696 >>>>>>>>>>>>>>>>>>>>>>> 837641 57059 60347 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 512 852099 889650 68870 73891 >>>>>>>>>>>>>>>>>>>>>>> 31185 845224 30259 830153 392334 >>>>>>>>>>>>>>>>>>>>>>> 910442 961983 60083 55558 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 1024 710357 867810 29377 >>>>>>>>>>>>>>>>>>>>>>> 29522 49954 846640 43665 926298 >>>>>>>>>>>>>>>>>>>>>>> 213677 986226 1115445 55130 59205 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 2048 826479 908420 43191 >>>>>>>>>>>>>>>>>>>>>>> 42075 59684 904022 58601 855664 >>>>>>>>>>>>>>>>>>>>>>> 115105 1418322 1524415 60548 66066 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 4096 4096 793351 855111 3232454 >>>>>>>>>>>>>>>>>>>>>>> 3673419 66018 861413 48833 847852 >>>>>>>>>>>>>>>>>>>>>>> 45914 852268 842075 42980 48374 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 4 67340 69421 42198 31740 994 >>>>>>>>>>>>>>>>>>>>>>> 23251 1166 16813 837 73827 73126 >>>>>>>>>>>>>>>>>>>>>>> 25169 29610 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 8 137150 125622 29131 36439 >>>>>>>>>>>>>>>>>>>>>>> 2051 44342 1988 48930 2315 134183 >>>>>>>>>>>>>>>>>>>>>>> 135367 31080 33573 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 16 237366 220826 24810 26584 >>>>>>>>>>>>>>>>>>>>>>> 3576 88004 3769 78717 4289 233751 >>>>>>>>>>>>>>>>>>>>>>> 235355 23302 28742 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 32 457447 454404 31594 27750 >>>>>>>>>>>>>>>>>>>>>>> 8141 142022 7846 143984 9322 >>>>>>>>>>>>>>>>>>>>>>> 353147 396188 34203 33265 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 64 670645 655259 28630 23255 >>>>>>>>>>>>>>>>>>>>>>> 16669 237476 16965 244968 15607 >>>>>>>>>>>>>>>>>>>>>>> 590365 575320 49998 43305 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 128 658676 760982 44197 47802 >>>>>>>>>>>>>>>>>>>>>>> 28693 379523 26614 378328 27184 >>>>>>>>>>>>>>>>>>>>>>> 720997 702038 51707 49733 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 256 643370 698683 56233 63165 >>>>>>>>>>>>>>>>>>>>>>> 28846 <tel:63165%20%C2%A0%2028846> >>>>>>>>>>>>>>>>>>>>>>> 543952 27745 576739 44014 701007 >>>>>>>>>>>>>>>>>>>>>>> 725534 59611 58985 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 512 696884 776793 67258 52705 >>>>>>>>>>>>>>>>>>>>>>> 18711 698854 21004 694124 621695 >>>>>>>>>>>>>>>>>>>>>>> 784812 773331 43101 47659 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 1024 729664 810451 15470 >>>>>>>>>>>>>>>>>>>>>>> 15875 31318 801490 38123 812944 >>>>>>>>>>>>>>>>>>>>>>> 301222 804323 832765 54308 53376 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 2048 749217 68757 21914 22667 >>>>>>>>>>>>>>>>>>>>>>> 48971 783309 48132 782738 172848 >>>>>>>>>>>>>>>>>>>>>>> 907408 929324 51156 50565 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 4096 707677 763960 32063 >>>>>>>>>>>>>>>>>>>>>>> 31928 47809 >>>>>>>>>>>>>>>>>>>>>>> <tel:31928%20%C2%A0%2047809> >>>>>>>>>>>>>>>>>>>>>>> 751692 49560 786339 93445 1046761 >>>>>>>>>>>>>>>>>>>>>>> 1297876 48037 51680 >>>>>>>>>>>>>>>>>>>>>>> <tel:48037%20%C2%A0%20%C2%A051680> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 8192 8192 623817 746288 2815955 >>>>>>>>>>>>>>>>>>>>>>> 3137358 48722 741633 35428 753787 >>>>>>>>>>>>>>>>>>>>>>> 49626 803683 823800 48977 52895 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 4 72372 73651 34471 30788 >>>>>>>>>>>>>>>>>>>>>>> 960 23610 903 22316 891 71445 >>>>>>>>>>>>>>>>>>>>>>> 71138 56451 55129 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 8 137920 141704 50830 33857 >>>>>>>>>>>>>>>>>>>>>>> 1935 41934 2275 35588 3608 130757 >>>>>>>>>>>>>>>>>>>>>>> 137801 51621 48525 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 16 245369 242460 41808 29770 >>>>>>>>>>>>>>>>>>>>>>> 3605 75682 4355 75315 4767 241100 >>>>>>>>>>>>>>>>>>>>>>> 239693 53263 30785 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 32 448877 433956 31846 35010 >>>>>>>>>>>>>>>>>>>>>>> 7973 118181 8819 112703 8177 >>>>>>>>>>>>>>>>>>>>>>> 381734 391651 57749 63417 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 64 710831 700712 66792 68864 >>>>>>>>>>>>>>>>>>>>>>> 20176 209806 19034 207852 >>>>>>>>>>>>>>>>>>>>>>> <tel:19034%20%C2%A0%20207852> >>>>>>>>>>>>>>>>>>>>>>> 21255 589503 601379 104567 105162 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 128 836901 860867 104226 >>>>>>>>>>>>>>>>>>>>>>> 100373 40899 358865 40946 360562 >>>>>>>>>>>>>>>>>>>>>>> 39415 675968 691538 96086 105695 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 256 798081 828146 107103 >>>>>>>>>>>>>>>>>>>>>>> 120433 39084 >>>>>>>>>>>>>>>>>>>>>>> <tel:120433%20%C2%A0%2039084> >>>>>>>>>>>>>>>>>>>>>>> 595325 39050 593110 56925 763466 >>>>>>>>>>>>>>>>>>>>>>> 797859 109645 113414 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 512 810851 843931 113564 >>>>>>>>>>>>>>>>>>>>>>> 106202 35111 714831 46244 745947 >>>>>>>>>>>>>>>>>>>>>>> 53636 802902 760172 110492 100879 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 1024 726399 820219 22106 >>>>>>>>>>>>>>>>>>>>>>> 22987 53087 >>>>>>>>>>>>>>>>>>>>>>> <tel:22987%20%C2%A0%2053087> >>>>>>>>>>>>>>>>>>>>>>> 749053 54781 777705 1075341 772686 >>>>>>>>>>>>>>>>>>>>>>> 809723 100349 96619 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 2048 807772 856458 23920 >>>>>>>>>>>>>>>>>>>>>>> 23617 66320 829576 72105 740848 >>>>>>>>>>>>>>>>>>>>>>> 656379 864539 835446 93499 101714 >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> 16384 4096 797470 840596 27270 >>>>>>>>>>>>>>>>>>>>>>> <tel:28132%20%C2%A0%2088784> >>>>>> ... >>>>>> >>>>>> [Message clipped] >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >> http://supercolony.gluster.org/mailman/listinfo/gluster-users > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> > http://supercolony.gluster.org/mailman/listinfo/gluster-users > > > > > -- > > *Rodrigo Severo* | DIRETOR DE TECNOLOGIA > Tel. *+55 61 3030-1515* > Siga a F?brica no twitter:*@empautaclipping* > > fabricadeideias.com <http://www.fabricadeideias.com/> > 12 ANOS DE TECNOLOGIA E COMUNICA??O > NUMA COMBINA??O PERFEITA > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130311/93d19f3e/attachment-0001.html>