Re: IOZone Performance is very Strange

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On Wed, 2011-05-18 at 14:24 +0800, AnnyRen wrote:
> root@MDS2:/mnt/ceph# rados bench 30 write -p data
> Maintaining 16 concurrent writes of 4194304 bytes for at least 30 seconds.
>   sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
>     0       0         0         0         0         0         -         0
>     1      16        37        21   83.9856        84  0.993284  0.567195
>     2      16        64        48   95.9862       108  0.470228  0.562306
>     3      16        92        76   101.314       112  0.320282  0.573153
>     4      16       119       103   102.983       108  0.645115  0.575708
>     5      16       147       131   104.782       112  0.595362   0.57243
>     6      16       176       160   106.649       116  0.320091  0.566513
>     7      16       204       188   107.412       112  0.764551  0.568444
>     8      16       230       214   106.983       104  0.519568  0.570681
>     9      16       261       245   108.872       124  0.572544  0.570245
>    10      16       287       271   108.383       104   1.06776  0.569834
>    11      16       317       301   109.437       120  0.450807  0.569928
>    12      16       342       326   108.649       100  0.336045  0.565149
>    13      16       371       355   109.213       116  0.846668  0.570469
>    14      16       400       384   109.696       116  0.758495  0.570053
>    15      16       429       413   110.115       116  0.325192  0.569528
>    16      16       455       439   109.732       104  0.454317  0.568965
>    17      16       485       469   110.335       120   1.36841  0.570114
>    18      16       513       497   110.427       112   1.26513   0.56956
>    19      16       541       525   110.509       112  0.650765   0.57068
> min lat: 0.237473 max lat: 1.6513 avg lat: 0.570027
>   sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
>    20      16       565       549   109.782        96  0.561031  0.570027
>    21      16       593       577   109.886       112  0.619525  0.570106
>    22      16       624       608   110.527       124  0.505926  0.572197
>    23      16       650       634   110.243       104  0.609987  0.570785
>    24      16       679       663   110.482       116  0.475714  0.569979
>    25      16       708       692   110.702       116  0.444257  0.568121
>    26      16       737       721   110.905       116  0.288582  0.569373
>    27      16       766       750   111.093       116   1.15793  0.567702
>    28      16       793       777   110.982       108  0.287232   0.56945
>    29      16       821       805   111.016       112  0.324029  0.571677
>    30      16       847       831   110.781       104  0.386532  0.569997
> Total time run:        30.413070
> Total writes made:     848
> Write size:            4194304
> Bandwidth (MB/sec):    111.531
> 
> Average Latency:       0.573732
> Max latency:           1.6513
> Min latency:           0.237473
> 
> 
> 
> root@MDS2:/mnt/ceph# rados bench 30 seq -p data
>   sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
>     0       0         0         0         0         0         -         0
>     1      16        39        23   91.9819        92  0.913858  0.441642
>     2      16        68        52   103.984       116  0.259619  0.472126
>     3      16        96        80   106.652       112   0.34457  0.511364
>     4      16       125       109   108.985       116  0.408573  0.533846
>     5      16       153       137   109.584       112   0.28345  0.550811
>     6      16       179       163    108.65       104  0.659187  0.549494
>     7      16       208       192   109.698       116  0.981796  0.553188
>     8      16       234       218   108.984       104  0.539604   0.55086
>     9      16       264       248   110.206       120  0.479113  0.547496
>    10      16       294       278   111.184       120   1.27273  0.552185
>    11      16       320       304   110.529       104   1.14992  0.557991
>    12      16       349       333   110.984       116  0.933907  0.559665
>    13      16       376       360   110.753       108   0.26945  0.559205
>    14      16       403       387   110.555       108   1.65993  0.559534
>    15      16       432       416   110.916       116  0.895407  0.563238
>    16      16       460       444   110.983       112  0.872572  0.560972
>    17      16       488       472   111.041       112  0.678441  0.561167
>    18      16       516       500   111.094       112  0.737577  0.563033
>    19      16       544       528    111.14       112  0.304587  0.565015
> min lat: 0.138238 max lat: 2.30221 avg lat: 0.564114
>   sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
>    20      16       573       557   111.381       116  0.413558  0.564114
>    21      16       602       586   111.601       116    1.0044  0.561393
>    22      16       627       611   111.072       100   1.33495  0.562511
>    23      16       658       642   111.633       124   1.28495  0.564387
>    24      16       685       669   111.482       108  0.255379  0.565283
>    25      16       714       698   111.662       116  0.979657  0.565571
>    26      16       742       726   111.674       112  0.296857  0.561789
>    27      16       770       754   111.685       112  0.205119   0.55878
>    28      16       796       780    111.41       104   0.44966  0.559254
>    29      16       826       810   111.706       120  0.432128  0.559298
> read got -2
> error during benchmark: -5
> error 5: Input/output error
> 

~~ 
should be a bug ,i also run into this issue.
filed it at 
http://tracker.newdream.net/issues/1095

with the log, your cluster platform read performance should be about
110M /sec. i'm not sure that your iozone read performance what's
wrong.maybe , read data from cache.



> 
> Do you know What happened to the final output of read performance test~?
> 
> Thank you, Jeff!!  :-)
> 
> 
> Anny
> 
> 
> 
> 
> 
> 
> 2011/5/18 Jeff Wu <cpwu@xxxxxxxxxxxxx>:
> > Hi ,
> >
> > Normally , read performance should be more than write performance ,about
> > 20% ~ 30% .
> >
> > Could you run the following commands and attach all of the results ?
> >  $rados bench 30 write -p data   //write performance
> >  $rados bench 30 seq  -p data    // read performance //
> >
> >
> >
> > On Wed, 2011-05-18 at 13:54 +0800, AnnyRen wrote:
> >> Hi Jeff & Henry,
> >>
> >> I took your advice and used the following command to run iozone test
> >>
> >> /usr/bin/iozone -z -c -e -a -n 512M -g 2G -i 0 -i 1 -i 2 -f
> >> /mnt/ceph/file0517.dbf -Rb /excel0517.xls
> >>
> >> Attachment is the output result.
> >>
> >> Writer Report
> >>       4       8       16      32      64      128     256     512     1024    2048    4096    8192    16384
> >> 524288        97920   101778  97050   103121  103445  102137  101082  101993  98835   103674  102043  99424   102679
> >> 1048576       104306  104841  103616  105884  106044  102221  105666  105890  105811  106409  103957  104062  102115
> >> 2097152       106896  107989  108428  107500  108453  108444  109584  108618  109413  108102  107529  108357  107545
> >>
> >> Reader Report
> >>       4       8       16      32      64      128     256     512     1024    2048    4096    8192    16384
> >> 524288        3756286 5029629 5049864 5810960 5577953 5554606 5388867 5636832 5709141 5253596 3805560 3640738 3442332
> >> 1048576       3709316 4733745 5400857 5655874 5788923 5565162 5865768 5546521 5886345 5167771 4485693 3531487 3473877
> >> 2097152       4305329 5103282 5713626 6399318 6180508 6392487 5804140 4621404 5873338 5837715 4181626 3420274 3506328
> >>
> >>
> >> the file size grows from 512 MB to 2 GB, write performance is about 94
> >> MB/s~ 106 MB/s
> >> and read performance still exceed 3 GB~6 GB, is it too fast ?
> >>
> >> Or we used the cache or buffer but we don't know? if yes, how could we
> >> avoid to use the cache to read ?
> >> Thanks very much. :-)
> >>
> >> Best Regards,
> >> Anny
> >>
> >>
> >>
> >>
> >> 2011/5/18 Jeff Wu <cpwu@xxxxxxxxxxxxx>:
> >> >
> >> > Hi AnnyRen,
> >> >
> >> > with the test results ,  run a 1G data at your ceph cluster ,
> >> > should be about 100M/sec.
> >> > maybe,iozone test results didn't include iozone flush time.
> >> >
> >> > Could you list your hardware platform infos ?
> >> > network:1G,4G,8G,FC ...?,cpu,memory:size ? ,disk:PATA,SATA,SAS,SSD ??
> >> > and
> >> > could you try other iozone commands ,for instance :
> >> >
> >> > 1)
> >> > add "-e" param to include flush(fsync,fflush) in the timing
> >> > calculations.
> >> >
> >> > /usr/bin/iozone -azcR -e -f /mnt/ceph/test0516.dbf  \
> >> > -g 1G -b /exceloutput0516.xls
> >> >
> >> > 2)run a large data which size is your host memory size*2:
> >> >
> >> > $./iozone -z -c -e -a -n 512M -g {memory_size}*2M -i 0 -i 1 -i 2  \
> >> > -f /mnt/ceph/fio -Rb ./iozone.xls
> >> >
> >> > or
> >> > 2)
> >> > !/bin/sh
> >> >
> >> > for i in 32 64 128 256
> >> > do
> >> > ./iozone -r ${i}k -t 10 -s 4096M -i 0 -i 1 -i 2   \
> >> > -F /mnt/ceph/F1 /mnt/ceph/F2 /mnt/ceph/F3 /mnt/ceph/F4 /mnt/ceph/F5 /mnt/ceph/F6 /mnt/ceph/F7 /mnt/ceph/F8 /mnt/ceph/F9 /mnt/ceph/F10
> >> > done
> >> >
> >> >
> >> >
> >> >
> >> > Jeff
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> > On Tue, 2011-05-17 at 15:34 +0800, AnnyRen wrote:
> >> >> Hi, Jeff:
> >> >>
> >> >> I run "ceph osd tell osd_num bench" with 1 times per osd
> >> >>
> >> >> and use ceph -w to observe every osd performance,
> >> >>
> >> >> osd0:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 10.875844
> >> >> sec at 96413 KB/sec
> >> >> osd1:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.784985
> >> >> sec at 88975 KB/sec
> >> >> osd2:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.161067
> >> >> sec at 93949 KB/sec
> >> >> osd3:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 10.798796
> >> >> sec at 97101 KB/sec
> >> >> osd4:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 14.437141
> >> >> sec at 72630 KB/sec
> >> >> osd5:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 14.451444
> >> >> sec at 72558 KB/sec
> >> >> osd6:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.083872
> >> >> sec at 94603 KB/sec
> >> >> osd7:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 11.062728
> >> >> sec at 94784 KB/sec
> >> >> osd8:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 14.137312
> >> >> sec at 74170 KB/sec
> >> >> osd9:  [INF] bench: wrote 1024 MB in blocks of 4096 KB in 13.489992
> >> >> sec at 77729 KB/sec
> >> >>
> >> >>
> >> >> and I run
> >> >> root@MDS2:/mnt/ceph# rados bench 60 write -p data
> >> >>
> >> >> the result is
> >> >>
> >> >> Total time run:        60.553247
> >> >> Total writes made:     1689
> >> >> Write size:            4194304
> >> >> Bandwidth (MB/sec):    111.571
> >> >>
> >> >> Average Latency:       0.573209
> >> >> Max latency:           2.25691
> >> >> Min latency:           0.218494
> >> >>
> >> >>
> >> >>
> >> >> 2011/5/17 Jeff Wu <cpwu@xxxxxxxxxxxxx>:
> >> >> > Hi AnnyRen
> >> >> >
> >> >> > Could you run the following commands and give us the test results?
> >> >> >
> >> >> > $ceph osd tell OSD-N bench    // OSD-N : osd number : 0,1,2 ....
> >> >> > $ceph -w
> >> >> >
> >> >> > $rados bench 60 write -p data    // refer to "rados -h "
> >> >> >
> >> >> > Jeff
> >> >> >
> >> >> >
> >> >> >
> >> >> > On Tue, 2011-05-17 at 11:53 +0800, AnnyRen wrote:
> >> >> >> I'm running iozone on EXT4 with Ceph v0.26.
> >> >> >> But I got the weird result, most write performance exceed 1GB/s, even
> >> >> >> up to 3GB/s
> >> >> >> I think it's not normal to get the performance outpupt.
> >> >> >>
> >> >> >> Command line I used is: /usr/bin/iozone -azcR -f
> >> >> >> /mnt/ceph/test0516.dbf -g 1G -b /exceloutput0516.xls
> >> >> >> Attachment is the output file...
> >> >> >>
> >> >> >> and my environment is composed of 15 physical machines with
> >> >> >>
> >> >> >> 3 MON, 2MDS (1 active, 1 standby), 10 OSD (1 osd daemon (3T) /host)
> >> >> >> EXT4 format
> >> >> >> data replication size: 3
> >> >> >>
> >> >> >>
> >> >> >>
> >> >> >> "Writer report"
> >> >> >>         "4"  "8"  "16"  "32"  "64"  "128"  "256"  "512"  "1024"
> >> >> >> "2048"  "4096"  "8192"  "16384"
> >> >> >> "64"   983980  1204796  1210227  1143223  1357066
> >> >> >> "128"   1007629  1269781  1406136  1391557  1436229  1521718
> >> >> >> "256"   1112909  1430119  1523457  1652404  1514860  1639786  1729594
> >> >> >> "512"   1150351  1475116  1605228  1723770  1797349  1712772  1783912  1854787
> >> >> >> "1024"   1213334  1471481  1679160  1828574  1888889  1899750  1885572
> >> >> >>  1865912  1875690
> >> >> >> "2048"   1229274  1540849  1708146  1843410  1903457  1980705  1930406
> >> >> >>  1913634  1906837  1815744
> >> >> >> "4096"   1213284  1528348  1674646  1762434  1872096  1882352  1881528
> >> >> >>  1903416  1897949  1835102  1731177
> >> >> >> "8192"   204560  155186  572387  238548  186597  429036  187327
> >> >> >> 157205  553771  416512  299810  405842
> >> >> >> "16384"   699749  559255  687450  541030  828776  555296  742561
> >> >> >> 525483  604910  452423  564557  670539  970616
> >> >> >> "32768"   532414  829610  812215  879441  863159  864794  865938
> >> >> >> 804951  916352  879582  608132  860732  1239475
> >> >> >> "65536"   994824  1096543  1095791  1317968  1280277  1390267  1259868
> >> >> >>  1205214  1339111  1346927  1267888  863234  1190221
> >> >> >> "131072"   1063429  1165115  1102650  1554828  1182128  1185731
> >> >> >> 1190752  1195792  1277441  1211063  1237567  1226999  1336961
> >> >> >> "262144"   1280619  1368554  1497837  1633397  1598255  1609212
> >> >> >> 1607504  1665019  1590515  1548307  1591258  1505267  1625679
> >> >> >> "524288"   1519583  1767928  1738523  1883151  2011216  1993877
> >> >> >> 2023543  1867440  2106124  2055064  1906668  1778645  1838988
> >> >> >> "1048576"   1580851  1887530  2044131  2166133  2236379  2283578
> >> >> >> 2257454  2296612  2271066  2101274  1905829  1605923  2158238
> >> >> >>
> >> >> >>
> >> >> >>
> >> >> >> "Reader report"
> >> >> >>         "4"  "8"  "16"  "32"  "64"  "128"  "256"  "512"  "1024"
> >> >> >> "2048"  "4096"  "8192"  "16384"
> >> >> >> "64"   1933893  2801873  3057153  3363612  3958892
> >> >> >> "128"   2286447  3053774  2727923  3468030  4104338  4557257
> >> >> >> "256"   2903529  3236056  3245838  3705040  3654598  4496299  5117791
> >> >> >> "512"   2906696  3437042  3628697  3431550  4871723  4296637  6246213  6395018
> >> >> >> "1024"   3229770  3483896  4609294  3791442  4614246  5536137  4550690
> >> >> >>  5048117  4966395
> >> >> >> "2048"   3554251  4310472  3885431  4096676  6401772  4842658  5080379
> >> >> >>  5184636  5596757  5735012
> >> >> >> "4096"   3416292  4691638  4321103  5728903  5475122  5171846  4819300
> >> >> >>  5258919  6408472  5044289  4079948
> >> >> >> "8192"   3233004  4615263  4536055  5618186  5414558  5025700  5553712
> >> >> >>  4926264  5634770  5281396  4659702  3652258
> >> >> >> "16384"   3141058  3704193  4567654  4395850  4568869  5387732
> >> >> >> 4436432  5808029  5578420  4675810  3913007  3911225  3961277
> >> >> >> "32768"   3704273  4598957  4088278  5133719  5896692  5537024
> >> >> >> 5234412  5398271  4942992  4118662  3729099  3511757  3481511
> >> >> >> "65536"   4131091  4210184  5341188  4647619  6077765  5852474
> >> >> >> 5379762  5259330  5488249  5246682  4342682  3549202  3286487
> >> >> >> "131072"   3582876  5251082  5332216  5269908  5303512  5574440
> >> >> >> 5635064  5796372  5406363  4958839  4435918  3673443  3647874
> >> >> >> "262144"   3659283  4551414  5746231  5433824  5876196  6011650
> >> >> >> 5552000  5629260  5298830  4982226  4628902  4065823  3421924
> >> >> >> "524288"   3905973  5488778  5219963  6047356  5916811  6180455
> >> >> >> 5495733  5925628  5637512  5537123  3517132  3550861  3047013
> >> >> >> "1048576"   3855595  5634571  5410298  6001809  6464713  6299610
> >> >> >> 5894249  5516031  5800161  5209203  4295840  3724983  3641623
> >> >> >
> >> >> >
> >> >
> >> >
> >
> >

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux