Re: Heavy speed difference between rbd and custom pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Stephann
recommandations are 30-50 PGS by osd if I remember.

----- Mail original ----- 

De: "Mark Nelson" <mark.nelson@xxxxxxxxxxx> 
À: "Stefan Priebe" <s.priebe@xxxxxxxxxxxx> 
Cc: ceph-devel@xxxxxxxxxxxxxxx 
Envoyé: Mardi 19 Juin 2012 00:23:49 
Objet: Re: Heavy speed difference between rbd and custom pool 

On 06/18/2012 04:39 PM, Stefan Priebe wrote: 
> Hello list, 
> 
> i'm getting these rbd bench values for pool rbd. They're high and constant. 
> ----------------------------- RBD pool 
> # rados -p rbd bench 30 write -t 16 
> Maintaining 16 concurrent writes of 4194304 bytes for at least 30 seconds. 
> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 
> 0 0 0 0 0 0 - 0 
> 1 16 274 258 1031.77 1032 0.043758 0.0602236 
> 2 16 549 533 1065.82 1100 0.072168 0.0590944 
> 3 16 825 809 1078.5 1104 0.040162 0.058682 
> 4 16 1103 1087 1086.84 1112 0.052508 0.0584277 
> 5 16 1385 1369 1095.04 1128 0.060233 0.0581288 
> 6 16 1654 1638 1091.85 1076 0.050697 0.0583385 
> 7 16 1939 1923 1098.71 1140 0.063716 0.057964 
> 8 16 2219 2203 1101.35 1120 0.055435 0.0579105 
> 9 16 2497 2481 1102.52 1112 0.060413 0.0578282 
> 10 16 2773 2757 1102.66 1104 0.051134 0.0578561 
> 11 16 3049 3033 1102.77 1104 0.057742 0.0578803 
> 12 16 3326 3310 1103.19 1108 0.053769 0.0578627 
> 13 16 3604 3588 1103.86 1112 0.064574 0.0578453 
> 14 16 3883 3867 1104.72 1116 0.056524 0.0578018 
> 15 16 4162 4146 1105.46 1116 0.054581 0.0577626 
> 16 16 4440 4424 1105.86 1112 0.079015 0.057758 
> 17 16 4725 4709 1107.86 1140 0.043511 0.0576647 
> 18 16 5007 4991 1108.97 1128 0.053005 0.0576147 
> 19 16 5292 5276 1110.6 1140 0.069004 0.057538 
> 2012-06-18 23:36:19.124472min lat: 0.028568 max lat: 0.201941 avg lat: 
> 0.0574953 
> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 
> 20 16 5574 5558 1111.46 1128 0.048482 0.0574953 
> 21 16 5861 5845 1113.18 1148 0.051923 0.0574146 
> 22 16 6147 6131 1114.58 1144 0.04461 0.0573461 
> 23 16 6438 6422 1116.72 1164 0.050383 0.0572406 
> 24 16 6724 6708 1117.85 1144 0.067827 0.0571864 
> 25 16 7008 6992 1118.57 1136 0.049128 0.057147 
> 26 16 7296 7280 1119.85 1152 0.050331 0.0570879 
> 27 16 7573 7557 1119.4 1108 0.052711 0.0571132 
> 28 16 7858 7842 1120.13 1140 0.056369 0.0570764 
> 29 16 8143 8127 1120.81 1140 0.046558 0.0570438 
> 30 16 8431 8415 1121.85 1152 0.049958 0.0569942 
> Total time run: 30.045481 
> Total writes made: 8431 
> Write size: 4194304 
> Bandwidth (MB/sec): 1122.432 
> 
> Stddev Bandwidth: 26.0451 
> Max bandwidth (MB/sec): 1164 
> Min bandwidth (MB/sec): 1032 
> Average Latency: 0.0570069 
> Stddev Latency: 0.0128039 
> Max latency: 0.235536 
> Min latency: 0.028568 
> ----------------------------- 
> 
> I created then a custom pool called kvmpool. 
> 
> ~# ceph osd pool create kvmpool 
> pool 'kvmpool' created 
> 
> But with this one i get slow and jumping values: 
> -------------------------------- kvmpool 
> ~# rados -p kvmpool bench 30 write -t 16 
> Maintaining 16 concurrent writes of 4194304 bytes for at least 30 seconds. 
> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 
> 0 0 0 0 0 0 - 0 
> 1 16 231 215 859.863 860 0.204867 0.069195 
> 2 16 393 377 753.899 648 0.049444 0.0811933 
> 3 16 535 519 691.908 568 0.232365 0.0899074 
> 4 16 634 618 617.913 396 0.032758 0.0963399 
> 5 16 806 790 631.913 688 0.075811 0.099529 
> 6 16 948 932 621.249 568 0.156988 0.10179 
> 7 16 1086 1070 611.348 552 0.036177 0.102064 
> 8 16 1206 1190 594.922 480 0.028491 0.105235 
> 9 16 1336 1320 586.589 520 0.041009 0.108735 
> 10 16 1512 1496 598.32 704 0.258165 0.105086 
> 11 16 1666 1650 599.921 616 0.040967 0.106146 
> 12 15 1825 1810 603.255 640 0.198851 0.105463 
> 13 16 1925 1909 587.309 396 0.042577 0.108449 
> 14 16 2135 2119 605.352 840 0.035767 0.105219 
> 15 16 2272 2256 601.523 548 0.246136 0.105357 
> 16 16 2426 2410 602.424 616 0.19881 0.105692 
> 17 16 2529 2513 591.22 412 0.031322 0.105463 
> 18 16 2696 2680 595.48 668 0.028081 0.106749 
> 19 16 2878 2862 602.449 728 0.044929 0.105856 
> 2012-06-18 23:38:45.566094min lat: 0.023295 max lat: 0.763797 avg lat: 
> 0.105597 
> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 
> 20 16 3041 3025 604.921 652 0.036028 0.105597 
> 21 16 3182 3166 602.964 564 0.035072 0.104915 
> 22 16 3349 3333 605.916 668 0.030493 0.105304 
> 23 16 3512 3496 607.917 652 0.030523 0.10479 
> 24 16 3668 3652 608.584 624 0.232933 0.10475 
> 25 16 3821 3805 608.717 612 0.029881 0.104513 
> 26 16 3963 3947 607.148 568 0.050244 0.10531 
> 27 16 4112 4096 606.733 596 0.259069 0.105008 
> 28 16 4261 4245 606.347 596 0.211877 0.105215 
> 29 16 4437 4421 609.712 704 0.02802 0.104613 
> 30 16 4566 4550 606.586 516 0.047076 0.105111 
> Total time run: 30.062141 
> Total writes made: 4566 
> Write size: 4194304 
> Bandwidth (MB/sec): 607.542 
> 
> Stddev Bandwidth: 109.112 
> Max bandwidth (MB/sec): 860 
> Min bandwidth (MB/sec): 396 
> Average Latency: 0.10532 
> Stddev Latency: 0.108369 
> Max latency: 0.763797 
> Min latency: 0.023295 
> -------------------------------- 
> 
> Why do these pools differ? Where is the difference? 
> 
> Stefan 

Are the number of placement groups the same for each pool? 

try running "ceph osd dump -o - | grep <pool>" and looking for the 
pg_num value. 

Mark 

-- 
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in 
the body of a message to majordomo@xxxxxxxxxxxxxxx 
More majordomo info at http://vger.kernel.org/majordomo-info.html 



-- 

-- 




	Alexandre D erumier 
Ingénieur Système 
Fixe : 03 20 68 88 90 
Fax : 03 20 68 90 81 
45 Bvd du Général Leclerc 59100 Roubaix - France 
12 rue Marivaux 75002 Paris - France 
	
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux