Re: Heavy speed difference between rbd and custom pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ok thanks but wouldn't it make sense to  set the default to the same as rbd has? How is the value for rbd calculated? I've also seen that rbd has a different crushmap. What's the difference between crushmap 0 and 2?

Stefan

Am 19.06.2012 um 00:41 schrieb Dan Mick <dan.mick@xxxxxxxxxxx>:

> Yes, this is almost certainly the problem.  When you create the pool, you can specify a pg count; the default is 8, which is quite low.
> The count can't currently be adjusted after pool-creation time (we're working on an enhancement for that).
> 
> http://ceph.com/docs/master/control/  shows
> 
> ceph osd pool create POOL [pg_num [pgp_num]]
> 
> You'll want to set pg_num the same for similar pools in order to get for similar pool performance.
> 
> I note also that you can get that field directlty:
> $ ceph osd pool get rbd pg_num
> PG_NUM: 448
> 
> I have a 'nova' pool that was created with "pool create":
> 
> $ ceph osd pool get nova pg_num
> PG_NUM: 8
> 
> 
> 
> On 06/18/2012 03:23 PM, Mark Nelson wrote:
>> On 06/18/2012 04:39 PM, Stefan Priebe wrote:
>>> Hello list,
>>> 
>>> i'm getting these rbd bench values for pool rbd. They're high and
>>> constant.
>>> ----------------------------- RBD pool
>>> # rados -p rbd bench 30 write -t 16
>>> Maintaining 16 concurrent writes of 4194304 bytes for at least 30
>>> seconds.
>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
>>> 0 0 0 0 0 0 - 0
>>> 1 16 274 258 1031.77 1032 0.043758 0.0602236
>>> 2 16 549 533 1065.82 1100 0.072168 0.0590944
>>> 3 16 825 809 1078.5 1104 0.040162 0.058682
>>> 4 16 1103 1087 1086.84 1112 0.052508 0.0584277
>>> 5 16 1385 1369 1095.04 1128 0.060233 0.0581288
>>> 6 16 1654 1638 1091.85 1076 0.050697 0.0583385
>>> 7 16 1939 1923 1098.71 1140 0.063716 0.057964
>>> 8 16 2219 2203 1101.35 1120 0.055435 0.0579105
>>> 9 16 2497 2481 1102.52 1112 0.060413 0.0578282
>>> 10 16 2773 2757 1102.66 1104 0.051134 0.0578561
>>> 11 16 3049 3033 1102.77 1104 0.057742 0.0578803
>>> 12 16 3326 3310 1103.19 1108 0.053769 0.0578627
>>> 13 16 3604 3588 1103.86 1112 0.064574 0.0578453
>>> 14 16 3883 3867 1104.72 1116 0.056524 0.0578018
>>> 15 16 4162 4146 1105.46 1116 0.054581 0.0577626
>>> 16 16 4440 4424 1105.86 1112 0.079015 0.057758
>>> 17 16 4725 4709 1107.86 1140 0.043511 0.0576647
>>> 18 16 5007 4991 1108.97 1128 0.053005 0.0576147
>>> 19 16 5292 5276 1110.6 1140 0.069004 0.057538
>>> 2012-06-18 23:36:19.124472min lat: 0.028568 max lat: 0.201941 avg lat:
>>> 0.0574953
>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
>>> 20 16 5574 5558 1111.46 1128 0.048482 0.0574953
>>> 21 16 5861 5845 1113.18 1148 0.051923 0.0574146
>>> 22 16 6147 6131 1114.58 1144 0.04461 0.0573461
>>> 23 16 6438 6422 1116.72 1164 0.050383 0.0572406
>>> 24 16 6724 6708 1117.85 1144 0.067827 0.0571864
>>> 25 16 7008 6992 1118.57 1136 0.049128 0.057147
>>> 26 16 7296 7280 1119.85 1152 0.050331 0.0570879
>>> 27 16 7573 7557 1119.4 1108 0.052711 0.0571132
>>> 28 16 7858 7842 1120.13 1140 0.056369 0.0570764
>>> 29 16 8143 8127 1120.81 1140 0.046558 0.0570438
>>> 30 16 8431 8415 1121.85 1152 0.049958 0.0569942
>>> Total time run: 30.045481
>>> Total writes made: 8431
>>> Write size: 4194304
>>> Bandwidth (MB/sec): 1122.432
>>> 
>>> Stddev Bandwidth: 26.0451
>>> Max bandwidth (MB/sec): 1164
>>> Min bandwidth (MB/sec): 1032
>>> Average Latency: 0.0570069
>>> Stddev Latency: 0.0128039
>>> Max latency: 0.235536
>>> Min latency: 0.028568
>>> -----------------------------
>>> 
>>> I created then a custom pool called kvmpool.
>>> 
>>> ~# ceph osd pool create kvmpool
>>> pool 'kvmpool' created
>>> 
>>> But with this one i get slow and jumping values:
>>> -------------------------------- kvmpool
>>> ~# rados -p kvmpool bench 30 write -t 16
>>> Maintaining 16 concurrent writes of 4194304 bytes for at least 30
>>> seconds.
>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
>>> 0 0 0 0 0 0 - 0
>>> 1 16 231 215 859.863 860 0.204867 0.069195
>>> 2 16 393 377 753.899 648 0.049444 0.0811933
>>> 3 16 535 519 691.908 568 0.232365 0.0899074
>>> 4 16 634 618 617.913 396 0.032758 0.0963399
>>> 5 16 806 790 631.913 688 0.075811 0.099529
>>> 6 16 948 932 621.249 568 0.156988 0.10179
>>> 7 16 1086 1070 611.348 552 0.036177 0.102064
>>> 8 16 1206 1190 594.922 480 0.028491 0.105235
>>> 9 16 1336 1320 586.589 520 0.041009 0.108735
>>> 10 16 1512 1496 598.32 704 0.258165 0.105086
>>> 11 16 1666 1650 599.921 616 0.040967 0.106146
>>> 12 15 1825 1810 603.255 640 0.198851 0.105463
>>> 13 16 1925 1909 587.309 396 0.042577 0.108449
>>> 14 16 2135 2119 605.352 840 0.035767 0.105219
>>> 15 16 2272 2256 601.523 548 0.246136 0.105357
>>> 16 16 2426 2410 602.424 616 0.19881 0.105692
>>> 17 16 2529 2513 591.22 412 0.031322 0.105463
>>> 18 16 2696 2680 595.48 668 0.028081 0.106749
>>> 19 16 2878 2862 602.449 728 0.044929 0.105856
>>> 2012-06-18 23:38:45.566094min lat: 0.023295 max lat: 0.763797 avg lat:
>>> 0.105597
>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
>>> 20 16 3041 3025 604.921 652 0.036028 0.105597
>>> 21 16 3182 3166 602.964 564 0.035072 0.104915
>>> 22 16 3349 3333 605.916 668 0.030493 0.105304
>>> 23 16 3512 3496 607.917 652 0.030523 0.10479
>>> 24 16 3668 3652 608.584 624 0.232933 0.10475
>>> 25 16 3821 3805 608.717 612 0.029881 0.104513
>>> 26 16 3963 3947 607.148 568 0.050244 0.10531
>>> 27 16 4112 4096 606.733 596 0.259069 0.105008
>>> 28 16 4261 4245 606.347 596 0.211877 0.105215
>>> 29 16 4437 4421 609.712 704 0.02802 0.104613
>>> 30 16 4566 4550 606.586 516 0.047076 0.105111
>>> Total time run: 30.062141
>>> Total writes made: 4566
>>> Write size: 4194304
>>> Bandwidth (MB/sec): 607.542
>>> 
>>> Stddev Bandwidth: 109.112
>>> Max bandwidth (MB/sec): 860
>>> Min bandwidth (MB/sec): 396
>>> Average Latency: 0.10532
>>> Stddev Latency: 0.108369
>>> Max latency: 0.763797
>>> Min latency: 0.023295
>>> --------------------------------
>>> 
>>> Why do these pools differ? Where is the difference?
>>> 
>>> Stefan
>> 
>> Are the number of placement groups the same for each pool?
>> 
>> try running "ceph osd dump -o - | grep <pool>" and looking for the
>> pg_num value.
>> 
>> Mark
>> 
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux