High reading IOPS in rgw gc pool since upgrade to Luminous

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I have a Ceph cluster with 6 hosts and 24 OSDs (mon and rgw colocated on these hosts) that is used just for RGW. Since upgrading to from Kraken (11.2.0) to Luminous (12.2.0) a week ago I can see lots of IOPS in the default.rgw.gc pool.

The cluster was created with Kraken two months ago and has the following status (yeah I know size=2 but data loss in this cluster is not problem):
root@ceph01:~# ceph status
  cluster:
    id:     14ff6ba0-6c70-4a01-ac80-5059c4956f2e
    health: HEALTH_OK
 
  services:
    mon: 3 daemons, quorum ceph04,ceph05,ceph06
    mgr: ceph04(active), standbys: ceph05, ceph06
    osd: 24 osds: 24 up, 24 in
    rgw: 4 daemons active
 
  data:
    pools:   13 pools, 1280 pgs
    objects: 16200k objects, 29355 GB
    usage:   59078 GB used, 44005 GB / 100 TB avail
    pgs:     1280 active+clean
 
  io:
    client:   177 MB/s rd, 24044 kB/s wr, 3260 op/s rd, 148 op/s wr
 
Normal load (before upgrade) is around 150 IOPS reading + 150 IOPS writing. Now the reading IOPS are between 2000 and 6000 all the time. All this excessive load is done in the default.rgw.gc pool. So my guess is there is some kind of garbage collection done. This started right after restarting the radosgw processes during upgrade. Otherwise the cluster is healthy and normal rgw work does not seem to be affected. The number of objects in the cluster went down by 1000k objects since the upgrade. The data size increased slightly which is expected by our workload.

When looking at the OSDs I can see a pattern of high CPU on a single OSD process in the cluster for about 60 to 90 minutes with lots of messages in the log and then after a break the load switches to another OSD.

2017-09-20 06:05:02.560587 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.560586174
2017-09-20 06:05:02.562030 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.562028463
2017-09-20 06:05:02.563489 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.563487625
2017-09-20 06:05:02.564869 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.564867424
2017-09-20 06:05:02.566284 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.566282739
2017-09-20 06:05:02.567652 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.567650712
2017-09-20 06:05:02.569123 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.569122362
2017-09-20 06:05:02.570587 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.570585626
2017-09-20 06:05:02.572017 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.572015899
2017-09-20 06:05:02.573471 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.573469906
2017-09-20 06:05:02.574912 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.574910343
2017-09-20 06:05:02.576363 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.576362151
2017-09-20 06:05:02.577807 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.577806002
2017-09-20 06:05:02.579190 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.579189306
2017-09-20 06:05:02.580642 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.580641131
2017-09-20 06:05:02.582086 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.582084531
2017-09-20 06:05:02.583506 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.583505141
2017-09-20 06:05:02.584996 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.584994743
2017-09-20 06:05:02.586429 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.586427627
2017-09-20 06:05:02.587885 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.587884152
2017-09-20 06:05:02.589272 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.589271242
2017-09-20 06:05:02.590766 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.590764473
2017-09-20 06:05:02.592184 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.592182375
2017-09-20 06:05:02.593651 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.593649812
2017-09-20 06:05:02.595083 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.595081484
2017-09-20 06:05:02.596534 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.596532864
2017-09-20 06:05:02.597924 7fd1bf114700  0 <cls> /build/ceph-12.2.0/src/cls/rgw/cls_rgw.cc:3251: gc_iterate_entries end_key=1_01505880302.597923234

As you can see this is a lot of stuff logged per second and these OSDs generate several GB of logs per day.

So my question is: Anybody else seeing this high read load since the upgrade? Is there some kind of data migration or intensiv initial GC done in Luminous when the cluster was upgraded from Kraken? Can we perhaps reduce the amount of logged messages in the next update? I remember reading about a fix in Luminous that was about leaked objects when doing multipart uploads. Maybe this is related?

Any more infos I can provide you with?

Thanks!

Uwe


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux