Re: IOZone Performance is very Strange

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi , 

Normally , read performance should be more than write performance ,about
20% ~ 30% . 

Could you run the following commands and attach all of the results ? 
 $rados bench 30 write -p data   //write performance
 $rados bench 30 seq  -p data    // read performance //



On Wed, 2011-05-18 at 13:54 +0800, AnnyRen wrote:
> Hi Jeff & Henry,
> 
> I took your advice and used the following command to run iozone test
> 
> /usr/bin/iozone -z -c -e -a -n 512M -g 2G -i 0 -i 1 -i 2 -f
> /mnt/ceph/file0517.dbf -Rb /excel0517.xls
> 
> Attachment is the output result.
> 
> Writer Report													
> 	4	8	16	32	64	128	256	512	1024	2048	4096	8192	16384
> 524288	97920	101778	97050	103121	103445	102137	101082	101993	98835	103674	102043	99424	102679
> 1048576	104306	104841	103616	105884	106044	102221	105666	105890	105811	106409	103957	104062	102115
> 2097152	106896	107989	108428	107500	108453	108444	109584	108618	109413	108102	107529	108357	107545
> 
> Reader Report													
> 	4	8	16	32	64	128	256	512	1024	2048	4096	8192	16384
> 524288	3756286	5029629	5049864	5810960	5577953	5554606	5388867	5636832	5709141	5253596	3805560	3640738	3442332
> 1048576	3709316	4733745	5400857	5655874	5788923	5565162	5865768	5546521	5886345	5167771	4485693	3531487	3473877
> 2097152	4305329	5103282	5713626	6399318	6180508	6392487	5804140	4621404	5873338	5837715	4181626	3420274	3506328
> 
> 
> the file size grows from 512 MB to 2 GB, write performance is about 94
> MB/s~ 106 MB/s
> and read performance still exceed 3 GB~6 GB, is it too fast ?
> 
> Or we used the cache or buffer but we don't know? if yes, how could we
> avoid to use the cache to read ?
> Thanks very much. :-)
> 
> Best Regards,
> Anny
> 
> 
> 
> 
> 2011/5/18 Jeff Wu <cpwu@xxxxxxxxxxxxx>:
> >
> > Hi AnnyRen,
> >
> > with the test results ,  run a 1G data at your ceph cluster ,
> > should be about 100M/sec.
> > maybe,iozone test results didn't include iozone flush time.
> >
> > Could you list your hardware platform infos ?
> > network:1G,4G,8G,FC ...?,cpu,memory:size ? ,disk:PATA,SATA,SAS,SSD ??
> > and
> > could you try other iozone commands ,for instance :
> >
> > 1)
> > add "-e" param to include flush(fsync,fflush) in the timing
> > calculations.
> >
> > /usr/bin/iozone -azcR -e -f /mnt/ceph/test0516.dbf  \
> > -g 1G -b /exceloutput0516.xls
> >
> > 2)run a large data which size is your host memory size*2:
> >
> > $./iozone -z -c -e -a -n 512M -g {memory_size}*2M -i 0 -i 1 -i 2  \
> > -f /mnt/ceph/fio -Rb ./iozone.xls
> >
> > or
> > 2)
> > !/bin/sh
> >
> > for i in 32 64 128 256
> > do
> > ./iozone -r ${i}k -t 10 -s 4096M -i 0 -i 1 -i 2   \
> > -F /mnt/ceph/F1 /mnt/ceph/F2 /mnt/ceph/F3 /mnt/ceph/F4 /mnt/ceph/F5 /mnt/ceph/F6 /mnt/ceph/F7 /mnt/ceph/F8 /mnt/ceph/F9 /mnt/ceph/F10
> > done
> >
> >
> >
> >
> > Jeff
> >
> >
> >
> >
> >
> >
> > On Tue, 2011-05-17 at 15:34 +0800, AnnyRen wrote:
> >> Hi, Jeff:
> >>
> >> I run "ceph osd tell osd_num bench" with 1 times per osd
> >>
> >> and use ceph -w to observe every osd performance,
> >>
> >> osd0:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 10.875844
> >> sec at 96413 KB/sec
> >> osd1:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.784985
> >> sec at 88975 KB/sec
> >> osd2:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.161067
> >> sec at 93949 KB/sec
> >> osd3:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 10.798796
> >> sec at 97101 KB/sec
> >> osd4:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 14.437141
> >> sec at 72630 KB/sec
> >> osd5:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 14.451444
> >> sec at 72558 KB/sec
> >> osd6:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.083872
> >> sec at 94603 KB/sec
> >> osd7:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.062728
> >> sec at 94784 KB/sec
> >> osd8:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 14.137312
> >> sec at 74170 KB/sec
> >> osd9:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 13.489992
> >> sec at 77729 KB/sec
> >>
> >>
> >> and I run
> >> root@MDS2:/mnt/ceph# rados bench 60 write -p data
> >>
> >> the result is
> >>
> >> Total time run:        60.553247
> >> Total writes made:     1689
> >> Write size:            4194304
> >> Bandwidth (MB/sec):    111.571
> >>
> >> Average Latency:       0.573209
> >> Max latency:           2.25691
> >> Min latency:           0.218494
> >>
> >>
> >>
> >> 2011/5/17 Jeff Wu <cpwu@xxxxxxxxxxxxx>:
> >> > Hi AnnyRen
> >> >
> >> > Could you run the following commands and give us the test results?
> >> >
> >> > $ceph osd tell OSD-N bench    // OSD-N : osd number : 0,1,2 ....
> >> > $ceph -w
> >> >
> >> > $rados bench 60 write -p data    // refer to "rados -h "
> >> >
> >> > Jeff
> >> >
> >> >
> >> >
> >> > On Tue, 2011-05-17 at 11:53 +0800, AnnyRen wrote:
> >> >> I'm running iozone on EXT4 with Ceph v0.26.
> >> >> But I got the weird result, most write performance exceed 1GB/s, even
> >> >> up to 3GB/s
> >> >> I think it's not normal to get the performance outpupt.
> >> >>
> >> >> Command line I used is: /usr/bin/iozone -azcR -f
> >> >> /mnt/ceph/test0516.dbf -g 1G -b /exceloutput0516.xls
> >> >> Attachment is the output file...
> >> >>
> >> >> and my environment is composed of 15 physical machines with
> >> >>
> >> >> 3 MON, 2MDS (1 active, 1 standby), 10 OSD (1 osd daemon (3T) /host)
> >> >> EXT4 format
> >> >> data replication size: 3
> >> >>
> >> >>
> >> >>
> >> >> "Writer report"
> >> >>         "4"  "8"  "16"  "32"  "64"  "128"  "256"  "512"  "1024"
> >> >> "2048"  "4096"  "8192"  "16384"
> >> >> "64"   983980  1204796  1210227  1143223  1357066
> >> >> "128"   1007629  1269781  1406136  1391557  1436229  1521718
> >> >> "256"   1112909  1430119  1523457  1652404  1514860  1639786  1729594
> >> >> "512"   1150351  1475116  1605228  1723770  1797349  1712772  1783912  1854787
> >> >> "1024"   1213334  1471481  1679160  1828574  1888889  1899750  1885572
> >> >>  1865912  1875690
> >> >> "2048"   1229274  1540849  1708146  1843410  1903457  1980705  1930406
> >> >>  1913634  1906837  1815744
> >> >> "4096"   1213284  1528348  1674646  1762434  1872096  1882352  1881528
> >> >>  1903416  1897949  1835102  1731177
> >> >> "8192"   204560  155186  572387  238548  186597  429036  187327
> >> >> 157205  553771  416512  299810  405842
> >> >> "16384"   699749  559255  687450  541030  828776  555296  742561
> >> >> 525483  604910  452423  564557  670539  970616
> >> >> "32768"   532414  829610  812215  879441  863159  864794  865938
> >> >> 804951  916352  879582  608132  860732  1239475
> >> >> "65536"   994824  1096543  1095791  1317968  1280277  1390267  1259868
> >> >>  1205214  1339111  1346927  1267888  863234  1190221
> >> >> "131072"   1063429  1165115  1102650  1554828  1182128  1185731
> >> >> 1190752  1195792  1277441  1211063  1237567  1226999  1336961
> >> >> "262144"   1280619  1368554  1497837  1633397  1598255  1609212
> >> >> 1607504  1665019  1590515  1548307  1591258  1505267  1625679
> >> >> "524288"   1519583  1767928  1738523  1883151  2011216  1993877
> >> >> 2023543  1867440  2106124  2055064  1906668  1778645  1838988
> >> >> "1048576"   1580851  1887530  2044131  2166133  2236379  2283578
> >> >> 2257454  2296612  2271066  2101274  1905829  1605923  2158238
> >> >>
> >> >>
> >> >>
> >> >> "Reader report"
> >> >>         "4"  "8"  "16"  "32"  "64"  "128"  "256"  "512"  "1024"
> >> >> "2048"  "4096"  "8192"  "16384"
> >> >> "64"   1933893  2801873  3057153  3363612  3958892
> >> >> "128"   2286447  3053774  2727923  3468030  4104338  4557257
> >> >> "256"   2903529  3236056  3245838  3705040  3654598  4496299  5117791
> >> >> "512"   2906696  3437042  3628697  3431550  4871723  4296637  6246213  6395018
> >> >> "1024"   3229770  3483896  4609294  3791442  4614246  5536137  4550690
> >> >>  5048117  4966395
> >> >> "2048"   3554251  4310472  3885431  4096676  6401772  4842658  5080379
> >> >>  5184636  5596757  5735012
> >> >> "4096"   3416292  4691638  4321103  5728903  5475122  5171846  4819300
> >> >>  5258919  6408472  5044289  4079948
> >> >> "8192"   3233004  4615263  4536055  5618186  5414558  5025700  5553712
> >> >>  4926264  5634770  5281396  4659702  3652258
> >> >> "16384"   3141058  3704193  4567654  4395850  4568869  5387732
> >> >> 4436432  5808029  5578420  4675810  3913007  3911225  3961277
> >> >> "32768"   3704273  4598957  4088278  5133719  5896692  5537024
> >> >> 5234412  5398271  4942992  4118662  3729099  3511757  3481511
> >> >> "65536"   4131091  4210184  5341188  4647619  6077765  5852474
> >> >> 5379762  5259330  5488249  5246682  4342682  3549202  3286487
> >> >> "131072"   3582876  5251082  5332216  5269908  5303512  5574440
> >> >> 5635064  5796372  5406363  4958839  4435918  3673443  3647874
> >> >> "262144"   3659283  4551414  5746231  5433824  5876196  6011650
> >> >> 5552000  5629260  5298830  4982226  4628902  4065823  3421924
> >> >> "524288"   3905973  5488778  5219963  6047356  5916811  6180455
> >> >> 5495733  5925628  5637512  5537123  3517132  3550861  3047013
> >> >> "1048576"   3855595  5634571  5410298  6001809  6464713  6299610
> >> >> 5894249  5516031  5800161  5209203  4295840  3724983  3641623
> >> >
> >> >
> >
> >

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux