Hi Jeff & Henry, I took your advice and used the following command to run iozone test /usr/bin/iozone -z -c -e -a -n 512M -g 2G -i 0 -i 1 -i 2 -f /mnt/ceph/file0517.dbf -Rb /excel0517.xls Attachment is the output result. Writer Report 4 8 16 32 64 128 256 512 1024 2048 4096 8192 16384 524288 97920 101778 97050 103121 103445 102137 101082 101993 98835 103674 102043 99424 102679 1048576 104306 104841 103616 105884 106044 102221 105666 105890 105811 106409 103957 104062 102115 2097152 106896 107989 108428 107500 108453 108444 109584 108618 109413 108102 107529 108357 107545 Reader Report 4 8 16 32 64 128 256 512 1024 2048 4096 8192 16384 524288 3756286 5029629 5049864 5810960 5577953 5554606 5388867 5636832 5709141 5253596 3805560 3640738 3442332 1048576 3709316 4733745 5400857 5655874 5788923 5565162 5865768 5546521 5886345 5167771 4485693 3531487 3473877 2097152 4305329 5103282 5713626 6399318 6180508 6392487 5804140 4621404 5873338 5837715 4181626 3420274 3506328 the file size grows from 512 MB to 2 GB, write performance is about 94 MB/s~ 106 MB/s and read performance still exceed 3 GB~6 GB, is it too fast ? Or we used the cache or buffer but we don't know? if yes, how could we avoid to use the cache to read ? Thanks very much. :-) Best Regards, Anny 2011/5/18 Jeff Wu <cpwu@xxxxxxxxxxxxx>: > > Hi AnnyRen, > > with the test results , run a 1G data at your ceph cluster , > should be about 100M/sec. > maybe,iozone test results didn't include iozone flush time. > > Could you list your hardware platform infos ? > network:1G,4G,8G,FC ...?,cpu,memory:size ? ,disk:PATA,SATA,SAS,SSD ?? > and > could you try other iozone commands ,for instance : > > 1) > add "-e" param to include flush(fsync,fflush) in the timing > calculations. > > /usr/bin/iozone -azcR -e -f /mnt/ceph/test0516.dbf \ > -g 1G -b /exceloutput0516.xls > > 2)run a large data which size is your host memory size*2: > > $./iozone -z -c -e -a -n 512M -g {memory_size}*2M -i 0 -i 1 -i 2 \ > -f /mnt/ceph/fio -Rb ./iozone.xls > > or > 2) > !/bin/sh > > for i in 32 64 128 256 > do > ./iozone -r ${i}k -t 10 -s 4096M -i 0 -i 1 -i 2 \ > -F /mnt/ceph/F1 /mnt/ceph/F2 /mnt/ceph/F3 /mnt/ceph/F4 /mnt/ceph/F5 /mnt/ceph/F6 /mnt/ceph/F7 /mnt/ceph/F8 /mnt/ceph/F9 /mnt/ceph/F10 > done > > > > > Jeff > > > > > > > On Tue, 2011-05-17 at 15:34 +0800, AnnyRen wrote: >> Hi, Jeff: >> >> I run "ceph osd tell osd_num bench" with 1 times per osd >> >> and use ceph -w to observe every osd performance, >> >> osd0: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 10.875844 >> sec at 96413 KB/sec >> osd1: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.784985 >> sec at 88975 KB/sec >> osd2: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.161067 >> sec at 93949 KB/sec >> osd3: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 10.798796 >> sec at 97101 KB/sec >> osd4: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 14.437141 >> sec at 72630 KB/sec >> osd5: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 14.451444 >> sec at 72558 KB/sec >> osd6: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.083872 >> sec at 94603 KB/sec >> osd7: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.062728 >> sec at 94784 KB/sec >> osd8: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 14.137312 >> sec at 74170 KB/sec >> osd9: [INF] bench: wrote 1024 MB in blocks of 4096 KB in 13.489992 >> sec at 77729 KB/sec >> >> >> and I run >> root@MDS2:/mnt/ceph# rados bench 60 write -p data >> >> the result is >> >> Total time run: 60.553247 >> Total writes made: 1689 >> Write size: 4194304 >> Bandwidth (MB/sec): 111.571 >> >> Average Latency: 0.573209 >> Max latency: 2.25691 >> Min latency: 0.218494 >> >> >> >> 2011/5/17 Jeff Wu <cpwu@xxxxxxxxxxxxx>: >> > Hi AnnyRen >> > >> > Could you run the following commands and give us the test results? >> > >> > $ceph osd tell OSD-N bench // OSD-N : osd number : 0,1,2 .... >> > $ceph -w >> > >> > $rados bench 60 write -p data // refer to "rados -h " >> > >> > Jeff >> > >> > >> > >> > On Tue, 2011-05-17 at 11:53 +0800, AnnyRen wrote: >> >> I'm running iozone on EXT4 with Ceph v0.26. >> >> But I got the weird result, most write performance exceed 1GB/s, even >> >> up to 3GB/s >> >> I think it's not normal to get the performance outpupt. >> >> >> >> Command line I used is: /usr/bin/iozone -azcR -f >> >> /mnt/ceph/test0516.dbf -g 1G -b /exceloutput0516.xls >> >> Attachment is the output file... >> >> >> >> and my environment is composed of 15 physical machines with >> >> >> >> 3 MON, 2MDS (1 active, 1 standby), 10 OSD (1 osd daemon (3T) /host) >> >> EXT4 format >> >> data replication size: 3 >> >> >> >> >> >> >> >> "Writer report" >> >> "4" "8" "16" "32" "64" "128" "256" "512" "1024" >> >> "2048" "4096" "8192" "16384" >> >> "64" 983980 1204796 1210227 1143223 1357066 >> >> "128" 1007629 1269781 1406136 1391557 1436229 1521718 >> >> "256" 1112909 1430119 1523457 1652404 1514860 1639786 1729594 >> >> "512" 1150351 1475116 1605228 1723770 1797349 1712772 1783912 1854787 >> >> "1024" 1213334 1471481 1679160 1828574 1888889 1899750 1885572 >> >> 1865912 1875690 >> >> "2048" 1229274 1540849 1708146 1843410 1903457 1980705 1930406 >> >> 1913634 1906837 1815744 >> >> "4096" 1213284 1528348 1674646 1762434 1872096 1882352 1881528 >> >> 1903416 1897949 1835102 1731177 >> >> "8192" 204560 155186 572387 238548 186597 429036 187327 >> >> 157205 553771 416512 299810 405842 >> >> "16384" 699749 559255 687450 541030 828776 555296 742561 >> >> 525483 604910 452423 564557 670539 970616 >> >> "32768" 532414 829610 812215 879441 863159 864794 865938 >> >> 804951 916352 879582 608132 860732 1239475 >> >> "65536" 994824 1096543 1095791 1317968 1280277 1390267 1259868 >> >> 1205214 1339111 1346927 1267888 863234 1190221 >> >> "131072" 1063429 1165115 1102650 1554828 1182128 1185731 >> >> 1190752 1195792 1277441 1211063 1237567 1226999 1336961 >> >> "262144" 1280619 1368554 1497837 1633397 1598255 1609212 >> >> 1607504 1665019 1590515 1548307 1591258 1505267 1625679 >> >> "524288" 1519583 1767928 1738523 1883151 2011216 1993877 >> >> 2023543 1867440 2106124 2055064 1906668 1778645 1838988 >> >> "1048576" 1580851 1887530 2044131 2166133 2236379 2283578 >> >> 2257454 2296612 2271066 2101274 1905829 1605923 2158238 >> >> >> >> >> >> >> >> "Reader report" >> >> "4" "8" "16" "32" "64" "128" "256" "512" "1024" >> >> "2048" "4096" "8192" "16384" >> >> "64" 1933893 2801873 3057153 3363612 3958892 >> >> "128" 2286447 3053774 2727923 3468030 4104338 4557257 >> >> "256" 2903529 3236056 3245838 3705040 3654598 4496299 5117791 >> >> "512" 2906696 3437042 3628697 3431550 4871723 4296637 6246213 6395018 >> >> "1024" 3229770 3483896 4609294 3791442 4614246 5536137 4550690 >> >> 5048117 4966395 >> >> "2048" 3554251 4310472 3885431 4096676 6401772 4842658 5080379 >> >> 5184636 5596757 5735012 >> >> "4096" 3416292 4691638 4321103 5728903 5475122 5171846 4819300 >> >> 5258919 6408472 5044289 4079948 >> >> "8192" 3233004 4615263 4536055 5618186 5414558 5025700 5553712 >> >> 4926264 5634770 5281396 4659702 3652258 >> >> "16384" 3141058 3704193 4567654 4395850 4568869 5387732 >> >> 4436432 5808029 5578420 4675810 3913007 3911225 3961277 >> >> "32768" 3704273 4598957 4088278 5133719 5896692 5537024 >> >> 5234412 5398271 4942992 4118662 3729099 3511757 3481511 >> >> "65536" 4131091 4210184 5341188 4647619 6077765 5852474 >> >> 5379762 5259330 5488249 5246682 4342682 3549202 3286487 >> >> "131072" 3582876 5251082 5332216 5269908 5303512 5574440 >> >> 5635064 5796372 5406363 4958839 4435918 3673443 3647874 >> >> "262144" 3659283 4551414 5746231 5433824 5876196 6011650 >> >> 5552000 5629260 5298830 4982226 4628902 4065823 3421924 >> >> "524288" 3905973 5488778 5219963 6047356 5916811 6180455 >> >> 5495733 5925628 5637512 5537123 3517132 3550861 3047013 >> >> "1048576" 3855595 5634571 5410298 6001809 6464713 6299610 >> >> 5894249 5516031 5800161 5209203 4295840 3724983 3641623 >> > >> > > > -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html