Re: Heavy speed difference between rbd and custom pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Sorry i meant crush_ruleset

Am 19.06.2012 08:06, schrieb Stefan Priebe:
Ok thanks but wouldn't it make sense to  set the default to the same as rbd has? How is the value for rbd calculated? I've also seen that rbd has a different crushmap. What's the difference between crushmap 0 and 2?

Stefan

Am 19.06.2012 um 00:41 schrieb Dan Mick<dan.mick@xxxxxxxxxxx>:

Yes, this is almost certainly the problem.  When you create the pool, you can specify a pg count; the default is 8, which is quite low.
The count can't currently be adjusted after pool-creation time (we're working on an enhancement for that).

http://ceph.com/docs/master/control/  shows

ceph osd pool create POOL [pg_num [pgp_num]]

You'll want to set pg_num the same for similar pools in order to get for similar pool performance.

I note also that you can get that field directlty:
$ ceph osd pool get rbd pg_num
PG_NUM: 448

I have a 'nova' pool that was created with "pool create":

$ ceph osd pool get nova pg_num
PG_NUM: 8



On 06/18/2012 03:23 PM, Mark Nelson wrote:
On 06/18/2012 04:39 PM, Stefan Priebe wrote:
Hello list,

i'm getting these rbd bench values for pool rbd. They're high and
constant.
----------------------------- RBD pool
# rados -p rbd bench 30 write -t 16
Maintaining 16 concurrent writes of 4194304 bytes for at least 30
seconds.
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
0 0 0 0 0 0 - 0
1 16 274 258 1031.77 1032 0.043758 0.0602236
2 16 549 533 1065.82 1100 0.072168 0.0590944
3 16 825 809 1078.5 1104 0.040162 0.058682
4 16 1103 1087 1086.84 1112 0.052508 0.0584277
5 16 1385 1369 1095.04 1128 0.060233 0.0581288
6 16 1654 1638 1091.85 1076 0.050697 0.0583385
7 16 1939 1923 1098.71 1140 0.063716 0.057964
8 16 2219 2203 1101.35 1120 0.055435 0.0579105
9 16 2497 2481 1102.52 1112 0.060413 0.0578282
10 16 2773 2757 1102.66 1104 0.051134 0.0578561
11 16 3049 3033 1102.77 1104 0.057742 0.0578803
12 16 3326 3310 1103.19 1108 0.053769 0.0578627
13 16 3604 3588 1103.86 1112 0.064574 0.0578453
14 16 3883 3867 1104.72 1116 0.056524 0.0578018
15 16 4162 4146 1105.46 1116 0.054581 0.0577626
16 16 4440 4424 1105.86 1112 0.079015 0.057758
17 16 4725 4709 1107.86 1140 0.043511 0.0576647
18 16 5007 4991 1108.97 1128 0.053005 0.0576147
19 16 5292 5276 1110.6 1140 0.069004 0.057538
2012-06-18 23:36:19.124472min lat: 0.028568 max lat: 0.201941 avg lat:
0.0574953
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
20 16 5574 5558 1111.46 1128 0.048482 0.0574953
21 16 5861 5845 1113.18 1148 0.051923 0.0574146
22 16 6147 6131 1114.58 1144 0.04461 0.0573461
23 16 6438 6422 1116.72 1164 0.050383 0.0572406
24 16 6724 6708 1117.85 1144 0.067827 0.0571864
25 16 7008 6992 1118.57 1136 0.049128 0.057147
26 16 7296 7280 1119.85 1152 0.050331 0.0570879
27 16 7573 7557 1119.4 1108 0.052711 0.0571132
28 16 7858 7842 1120.13 1140 0.056369 0.0570764
29 16 8143 8127 1120.81 1140 0.046558 0.0570438
30 16 8431 8415 1121.85 1152 0.049958 0.0569942
Total time run: 30.045481
Total writes made: 8431
Write size: 4194304
Bandwidth (MB/sec): 1122.432

Stddev Bandwidth: 26.0451
Max bandwidth (MB/sec): 1164
Min bandwidth (MB/sec): 1032
Average Latency: 0.0570069
Stddev Latency: 0.0128039
Max latency: 0.235536
Min latency: 0.028568
-----------------------------

I created then a custom pool called kvmpool.

~# ceph osd pool create kvmpool
pool 'kvmpool' created

But with this one i get slow and jumping values:
-------------------------------- kvmpool
~# rados -p kvmpool bench 30 write -t 16
Maintaining 16 concurrent writes of 4194304 bytes for at least 30
seconds.
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
0 0 0 0 0 0 - 0
1 16 231 215 859.863 860 0.204867 0.069195
2 16 393 377 753.899 648 0.049444 0.0811933
3 16 535 519 691.908 568 0.232365 0.0899074
4 16 634 618 617.913 396 0.032758 0.0963399
5 16 806 790 631.913 688 0.075811 0.099529
6 16 948 932 621.249 568 0.156988 0.10179
7 16 1086 1070 611.348 552 0.036177 0.102064
8 16 1206 1190 594.922 480 0.028491 0.105235
9 16 1336 1320 586.589 520 0.041009 0.108735
10 16 1512 1496 598.32 704 0.258165 0.105086
11 16 1666 1650 599.921 616 0.040967 0.106146
12 15 1825 1810 603.255 640 0.198851 0.105463
13 16 1925 1909 587.309 396 0.042577 0.108449
14 16 2135 2119 605.352 840 0.035767 0.105219
15 16 2272 2256 601.523 548 0.246136 0.105357
16 16 2426 2410 602.424 616 0.19881 0.105692
17 16 2529 2513 591.22 412 0.031322 0.105463
18 16 2696 2680 595.48 668 0.028081 0.106749
19 16 2878 2862 602.449 728 0.044929 0.105856
2012-06-18 23:38:45.566094min lat: 0.023295 max lat: 0.763797 avg lat:
0.105597
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
20 16 3041 3025 604.921 652 0.036028 0.105597
21 16 3182 3166 602.964 564 0.035072 0.104915
22 16 3349 3333 605.916 668 0.030493 0.105304
23 16 3512 3496 607.917 652 0.030523 0.10479
24 16 3668 3652 608.584 624 0.232933 0.10475
25 16 3821 3805 608.717 612 0.029881 0.104513
26 16 3963 3947 607.148 568 0.050244 0.10531
27 16 4112 4096 606.733 596 0.259069 0.105008
28 16 4261 4245 606.347 596 0.211877 0.105215
29 16 4437 4421 609.712 704 0.02802 0.104613
30 16 4566 4550 606.586 516 0.047076 0.105111
Total time run: 30.062141
Total writes made: 4566
Write size: 4194304
Bandwidth (MB/sec): 607.542

Stddev Bandwidth: 109.112
Max bandwidth (MB/sec): 860
Min bandwidth (MB/sec): 396
Average Latency: 0.10532
Stddev Latency: 0.108369
Max latency: 0.763797
Min latency: 0.023295
--------------------------------

Why do these pools differ? Where is the difference?

Stefan

Are the number of placement groups the same for each pool?

try running "ceph osd dump -o - | grep<pool>" and looking for the
pg_num value.

Mark

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux