On 09/02/2015 08:51 AM, Vickey Singh wrote:
Hello Ceph Experts
I have a strange problem , when i am reading or writing to Ceph pool ,
its not writing properly. Please notice Cur MB/s which is going up and down
--- Ceph Hammer 0.94.2
-- CentOS 6, 2.6
-- Ceph cluster is healthy
You might find that CentOS7 gives you better performance. In some cases
we were seeing nearly 2X.
One interesting thing is when every i start rados bench command for read
or write CPU Idle % goes down ~10 and System load is increasing like
anything.
Hardware
HpSL4540
Please make sure the controller is on the newest firmware. There used
to be a bug that would cause sequential write performance to bottleneck
when writeback cache was enabled on the RAID controller.
32Core CPU
196G Memory
10G Network
Be sure to check the network too. We've seen a lot of cases where folks
have been burned by one of the NICs acting funky.
I don't think hardware is a problem.
Please give me clues / pointers , how should i troubleshoot this problem.
# rados bench -p glance-test 60 write
Maintaining 16 concurrent writes of 4194304 bytes for up to 60 seconds
or 0 objects
Object prefix: benchmark_data_pouta-s01.pouta.csc.fi_2173350
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
0 0 0 0 0 0 - 0
1 16 20 4 15.99 16 0.12308 0.10001
2 16 37 21 41.9841 68 1.79104 0.827021
3 16 68 52 69.3122 124 0.084304 0.854829
4 16 114 98 97.9746 184 0.12285 0.614507
5 16 188 172 137.568 296 0.210669 0.449784
6 16 248 232 154.634 240 0.090418 0.390647
7 16 305 289 165.11 228 0.069769 0.347957
8 16 331 315 157.471 104 0.026247 0.3345
9 16 361 345 153.306 120 0.082861 0.320711
10 16 380 364 145.575 76 0.027964 0.310004
11 16 393 377 137.067 52 3.73332 0.393318
12 16 448 432 143.971 220 0.334664 0.415606
13 16 476 460 141.508 112 0.271096 0.406574
14 16 497 481 137.399 84 0.257794 0.412006
15 16 507 491 130.906 40 1.49351 0.428057
16 16 529 513 115.042 88 0.399384 0.48009
17 16 533 517 94.6286 16 5.50641 0.507804
18 16 537 521 83.405 16 4.42682 0.549951
19 16 538 522 80.349 4 11.2052 0.570363
2015-09-02 09:26:18.398641min lat: 0.023851 max lat: 11.2052 avg lat:
0.570363
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
20 16 538 522 77.3611 0 - 0.570363
21 16 540 524 74.8825 4 8.88847 0.591767
22 16 542 526 72.5748 8 1.41627 0.593555
23 16 543 527 70.2873 4 8.0856 0.607771
24 16 555 539 69.5674 48 0.145199 0.781685
25 16 560 544 68.0177 20 1.4342 0.787017
26 16 564 548 66.4241 16 0.451905 0.78765
27 16 566 550 64.7055 8 0.611129 0.787898
28 16 570 554 63.3138 16 2.51086 0.797067
29 16 570 554 61.5549 0 - 0.797067
30 16 572 556 60.1071 4 7.71382 0.830697
31 16 577 561 59.0515 20 23.3501 0.916368
32 16 590 574 58.8705 52 0.336684 0.956958
33 16 591 575 57.4986 4 1.92811 0.958647
34 16 591 575 56.0961 0 - 0.958647
35 16 591 575 54.7603 0 - 0.958647
36 16 597 581 54.0447 8 0.187351 1.00313
37 16 625 609 52.8394 112 2.12256 1.09256
38 16 631 615 52.227 24 1.57413 1.10206
39 16 638 622 51.7232 28 4.41663 1.15086
2015-09-02 09:26:40.510623min lat: 0.023851 max lat: 27.6704 avg lat:
1.15657
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
40 16 652 636 51.8102 56 0.113345 1.15657
41 16 682 666 53.1443 120 0.041251 1.17813
42 16 685 669 52.3395 12 0.501285 1.17421
43 15 690 675 51.7955 24 2.26605 1.18357
44 16 728 712 53.6062 148 0.589826 1.17478
45 16 728 712 52.6158 0 - 1.17478
46 16 728 712 51.6613 0 - 1.17478
47 16 728 712 50.7407 0 - 1.17478
48 16 772 756 52.9332 44 0.234811 1.1946
49 16 835 819 56.3577 252 5.67087 1.12063
50 16 890 874 59.1252 220 0.230806 1.06778
51 16 896 880 58.5409 24 0.382471 1.06121
52 16 896 880 57.5832 0 - 1.06121
53 16 896 880 56.6562 0 - 1.06121
54 16 896 880 55.7587 0 - 1.06121
55 16 897 881 54.9515 1 4.88333 1.06554
56 16 897 881 54.1077 0 - 1.06554
57 16 897 881 53.2894 0 - 1.06554
58 16 897 881 51.9335 0 - 1.06554
59 16 897 881 51.1792 0 - 1.06554
2015-09-02 09:27:01.267301min lat: 0.01405 max lat: 27.6704 avg lat: 1.06554
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
60 16 897 881 50.4445 0 - 1.06554
cluster 98d89661-f616-49eb-9ccf-84d720e179c0
health HEALTH_OK
monmap e3: 3 mons at
{s01=10.100.50.1:6789/0,s02=10.100.50.2:6789/0,s03=1
<http://10.100.50.1:6789/0,s02=10.100.50.2:6789/0,s03=1>
0.100.50.3:6789/0 <http://0.100.50.3:6789/0>}, election epoch 666,
quorum 0,1,2 s01,s02,s03
* osdmap e121039: 240 osds: 240 up, 240 in*
pgmap v850698: 7232 pgs, 31 pools, 439 GB data, 43090 kobjects
2635 GB used, 867 TB / 870 TB avail
7226 active+clean
6 active+clean+scrubbing+deep
Note the last line there. You'll likely want to try your test again
when scrubbing is complete. Also, you may want to try this script:
https://github.com/ceph/cbt/blob/master/tools/readpgdump.py
You can invoke it like:
ceph pg dump | ./readpgdump.py
That will give you a bunch of information about the pools on your
system. I'm a little concerned about how many PGs your glance-test pool
may have given your totals above.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com