Am I the only one who finds it funny that the "ceph problem" was fixed by an update to the disk controller firmware? :-)
IanOn Thu, Sep 3, 2015 at 11:13 AM, Vickey Singh <vickey.singh22693@xxxxxxxxx> wrote:
Hey Mark / CommunityThese are the sequences of changes that seems to have fixed the ceph problem
1# Upgrading Disk controller firmware from 6.34 to 6.64 ( latest )
2# Rebooting all nodes in order to make new firmware into effect
Read and write operations are now normal as well as system load and CPU utilization- Vickey -On Wed, Sep 2, 2015 at 11:28 PM, Vickey Singh <vickey.singh22693@xxxxxxxxx> wrote:Thank You Mark , please see my response below.On Wed, Sep 2, 2015 at 5:23 PM, Mark Nelson <mnelson@xxxxxxxxxx> wrote:On 09/02/2015 08:51 AM, Vickey Singh wrote:
Hello Ceph Experts
I have a strange problem , when i am reading or writing to Ceph pool ,
its not writing properly. Please notice Cur MB/s which is going up and down
--- Ceph Hammer 0.94.2
-- CentOS 6, 2.6
-- Ceph cluster is healthy
You might find that CentOS7 gives you better performance. In some cases we were seeing nearly 2X.Wooo 2X , i would definitely plan for upgrade. Thanks
One interesting thing is when every i start rados bench command for read
or write CPU Idle % goes down ~10 and System load is increasing like
anything.
Hardware
HpSL4540
Please make sure the controller is on the newest firmware. There used to be a bug that would cause sequential write performance to bottleneck when writeback cache was enabled on the RAID controller.Last month i have upgraded the firmwares for this hardware , so i hope they are up to date.
32Core CPU
196G Memory
10G Network
Be sure to check the network too. We've seen a lot of cases where folks have been burned by one of the NICs acting funky.At a first view , Interface looks good and they are pushing data nicely ( what ever they are getting )
<http://10.100.50.1:6789/0,s02=10.100.50.2:6789/0,s03=1>
I don't think hardware is a problem.
Please give me clues / pointers , how should i troubleshoot this problem.
# rados bench -p glance-test 60 write
Maintaining 16 concurrent writes of 4194304 bytes for up to 60 seconds
or 0 objects
Object prefix: benchmark_data_pouta-s01.pouta.csc.fi_2173350
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
0 0 0 0 0 0 - 0
1 16 20 4 15.99 16 0.12308 0.10001
2 16 37 21 41.9841 68 1.79104 0.827021
3 16 68 52 69.3122 124 0.084304 0.854829
4 16 114 98 97.9746 184 0.12285 0.614507
5 16 188 172 137.568 296 0.210669 0.449784
6 16 248 232 154.634 240 0.090418 0.390647
7 16 305 289 165.11 228 0.069769 0.347957
8 16 331 315 157.471 104 0.026247 0.3345
9 16 361 345 153.306 120 0.082861 0.320711
10 16 380 364 145.575 76 0.027964 0.310004
11 16 393 377 137.067 52 3.73332 0.393318
12 16 448 432 143.971 220 0.334664 0.415606
13 16 476 460 141.508 112 0.271096 0.406574
14 16 497 481 137.399 84 0.257794 0.412006
15 16 507 491 130.906 40 1.49351 0.428057
16 16 529 513 115.042 88 0.399384 0.48009
17 16 533 517 94.6286 16 5.50641 0.507804
18 16 537 521 83.405 16 4.42682 0.549951
19 16 538 522 80.349 4 11.2052 0.570363
2015-09-02 09:26:18.398641min lat: 0.023851 max lat: 11.2052 avg lat:
0.570363
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
20 16 538 522 77.3611 0 - 0.570363
21 16 540 524 74.8825 4 8.88847 0.591767
22 16 542 526 72.5748 8 1.41627 0.593555
23 16 543 527 70.2873 4 8.0856 0.607771
24 16 555 539 69.5674 48 0.145199 0.781685
25 16 560 544 68.0177 20 1.4342 0.787017
26 16 564 548 66.4241 16 0.451905 0.78765
27 16 566 550 64.7055 8 0.611129 0.787898
28 16 570 554 63.3138 16 2.51086 0.797067
29 16 570 554 61.5549 0 - 0.797067
30 16 572 556 60.1071 4 7.71382 0.830697
31 16 577 561 59.0515 20 23.3501 0.916368
32 16 590 574 58.8705 52 0.336684 0.956958
33 16 591 575 57.4986 4 1.92811 0.958647
34 16 591 575 56.0961 0 - 0.958647
35 16 591 575 54.7603 0 - 0.958647
36 16 597 581 54.0447 8 0.187351 1.00313
37 16 625 609 52.8394 112 2.12256 1.09256
38 16 631 615 52.227 24 1.57413 1.10206
39 16 638 622 51.7232 28 4.41663 1.15086
2015-09-02 09:26:40.510623min lat: 0.023851 max lat: 27.6704 avg lat:
1.15657
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
40 16 652 636 51.8102 56 0.113345 1.15657
41 16 682 666 53.1443 120 0.041251 1.17813
42 16 685 669 52.3395 12 0.501285 1.17421
43 15 690 675 51.7955 24 2.26605 1.18357
44 16 728 712 53.6062 148 0.589826 1.17478
45 16 728 712 52.6158 0 - 1.17478
46 16 728 712 51.6613 0 - 1.17478
47 16 728 712 50.7407 0 - 1.17478
48 16 772 756 52.9332 44 0.234811 1.1946
49 16 835 819 56.3577 252 5.67087 1.12063
50 16 890 874 59.1252 220 0.230806 1.06778
51 16 896 880 58.5409 24 0.382471 1.06121
52 16 896 880 57.5832 0 - 1.06121
53 16 896 880 56.6562 0 - 1.06121
54 16 896 880 55.7587 0 - 1.06121
55 16 897 881 54.9515 1 4.88333 1.06554
56 16 897 881 54.1077 0 - 1.06554
57 16 897 881 53.2894 0 - 1.06554
58 16 897 881 51.9335 0 - 1.06554
59 16 897 881 51.1792 0 - 1.06554
2015-09-02 09:27:01.267301min lat: 0.01405 max lat: 27.6704 avg lat: 1.06554
sec Cur ops started finished avg MB/s cur MB/s last lat avg lat
60 16 897 881 50.4445 0 - 1.06554
cluster 98d89661-f616-49eb-9ccf-84d720e179c0
health HEALTH_OK
monmap e3: 3 mons at
{s01=10.100.50.1:6789/0,s02=10.100.50.2:6789/0,s03=1
0.100.50.3:6789/0 <http://0.100.50.3:6789/0>}, election epoch 666,
quorum 0,1,2 s01,s02,s03
* osdmap e121039: 240 osds: 240 up, 240 in*
pgmap v850698: 7232 pgs, 31 pools, 439 GB data, 43090 kobjects
2635 GB used, 867 TB / 870 TB avail
7226 active+clean
6 active+clean+scrubbing+deep
Note the last line there. You'll likely want to try your test again when scrubbing is complete. Also, you may want to try this script:Yeah i have tried few times when cluster is perfectly healthy ( not doing scrubbing / repairs )
https://github.com/ceph/cbt/blob/master/tools/readpgdump.py
You can invoke it like:
ceph pg dump | ./readpgdump.py
That will give you a bunch of information about the pools on your system. I'm a little concerned about how many PGs your glance-test pool may have given your totals above.Thanks for the link i would do that and also run rados bench for other pools ( where PG is higher )Now here are my some observations1# When the cluster is not doing anything , Health_ok , with no background scrubbing / repairing. Also all system resources CPU/MEM/NET are mostly idle. In this Case when i start rados bench ( write / rand / seq ) , after suddenly a few seconds--- rados bench output drops from ~500M to few 10M--- At the same time CPU busy 90% and System load bumps UPOnce rados bench completes--- After few minutes System resources becomes IDLE2# Sometime some PG becomes unclean for a few minutes while rados bench runs and then then quickly they becomes active+cleanI am out of clues , so any help from community that leads me to think in right direction , would be helpful.- Vickey -
_______________________________________________
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
--
Ian R. Colle
Global Director of Software Engineering+1-303-601-7713
http://www.linkedin.com/in/ircolle
http://www.twitter.com/ircolle
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com