Re: Help with full osd and RGW not responsive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks for the response, we increased our pg count to something more reasonable (512 for now) and things are rebalancing.

 

Cheers,

-Bryan

 

From: Andreas Calminder [mailto:andreas.calminder@xxxxxxxxxx]
Sent: Tuesday, October 17, 2017 3:48 PM
To: Bryan Banister <bbanister@xxxxxxxxxxxxxxx>
Cc: Ceph Users <ceph-users@xxxxxxxxxxxxxx>
Subject: Re: [ceph-users] Help with full osd and RGW not responsive

 

Note: External Email


Hi,

You should most definitely look over number of pgs, there's a pg calculator available here: http://ceph.com/pgcalc/

 

You can increase pgs but not the other way around (http://docs.ceph.com/docs/jewel/rados/operations/placement-groups/)

 

To solve the immediate problem with your cluster being full you can reweight your osds, giving the full osd a lower weight will cause writes going to other osds and data on that osd being migrated to other osds in the cluster: ceph osd reweight $OSDNUM $WEIGHT, described here http://docs.ceph.com/docs/master/rados/operations/control/#osd-subsystem

 

When the osd isn't above the full threshold, default is 95%, the cluster will clear its full flag and your radosgw should start accepting write operations again, at least until another osd gets full, main problem here is probably the low pg count.

 

Regards,

Andreas

 

On 17 Oct 2017 19:08, "Bryan Banister" <bbanister@xxxxxxxxxxxxxxx> wrote:

Hi all,

 

Still a real novice here and we didn’t set up our initial RGW cluster very well.  We have 134 osds and set up our RGW pool with only 64 PGs, thus not all of our OSDs got data and now we have one that is 95% full. 

 

This apparently has put the cluster into a HEALTH_ERR condition:

[root@carf-ceph-osd01 ~]# ceph health detail

HEALTH_ERR full flag(s) set; 1 full osd(s); 1 pools have many more objects per pg than average; application not enabled on 6 pool(s); too few PGs per OSD (26 < min 30)

OSDMAP_FLAGS full flag(s) set

OSD_FULL 1 full osd(s)

    osd.5 is full

MANY_OBJECTS_PER_PG 1 pools have many more objects per pg than average

    pool carf01.rgw.buckets.data objects per pg (602762) is more than 18.3752 times cluster average (32803)

 

There is plenty of space on most of the OSDs and don’t know how to go about fixing this situation.  If we update the pg_num and pgp_num settings for this pool, can we rebalance the data across the OSDs?

 

Also, seems like this is causing a problem with the RGWs, which was reporting this error in the logs:

2017-10-16 16:36:47.534461 7fffe6c5c700  1 heartbeat_map is_healthy 'RGWAsyncRadosProcessor::m_tp thread 0x7fffdc447700' had timed out after 600

 

After trying to restart the RGW, we see this now:

2017-10-17 10:40:38.517002 7fffe6c5c700  1 heartbeat_map is_healthy 'RGWAsyncRadosProcessor::m_tp thread 0x7fffddc4a700' had timed out after 600

2017-10-17 10:40:42.124046 7ffff7fd4e00  0 deferred set uid:gid to 167:167 (ceph:ceph)

2017-10-17 10:40:42.124162 7ffff7fd4e00  0 ceph version 12.2.0 (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown), pid 65313

2017-10-17 10:40:42.245259 7ffff7fd4e00  0 client.769905.objecter  FULL, paused modify 0x55555662fb00 tid 0

2017-10-17 10:45:42.124283 7fffe7bcf700 -1 Initialization timeout, failed to initialize

2017-10-17 10:45:42.353496 7ffff7fd4e00  0 deferred set uid:gid to 167:167 (ceph:ceph)

2017-10-17 10:45:42.353618 7ffff7fd4e00  0 ceph version 12.2.0 (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown), pid 71842

2017-10-17 10:45:42.388621 7ffff7fd4e00  0 client.769986.objecter  FULL, paused modify 0x55555662fb00 tid 0

2017-10-17 10:50:42.353731 7fffe7bcf700 -1 Initialization timeout, failed to initialize

 

Seems pretty evident that the “FULL, paused” is a problem.  So if I fix the first issue the RGW should be ok after?

 

Thanks in advance,

-Bryan

 



Note: This email is for the confidential use of the named addressee(s) only and may contain proprietary, confidential or privileged information. If you are not the intended recipient, you are hereby notified that any review, dissemination or copying of this email is strictly prohibited, and to please notify the sender immediately and destroy this email and any attachments. Email transmission cannot be guaranteed to be secure or error-free. The Company, therefore, does not make any guarantees as to the completeness or accuracy of this email or any attachments. This email is for informational purposes only and does not constitute a recommendation, offer, request or solicitation of any kind to buy, sell, subscribe, redeem or perform any type of transaction of a financial product.


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

 




Note: This email is for the confidential use of the named addressee(s) only and may contain proprietary, confidential or privileged information. If you are not the intended recipient, you are hereby notified that any review, dissemination or copying of this email is strictly prohibited, and to please notify the sender immediately and destroy this email and any attachments. Email transmission cannot be guaranteed to be secure or error-free. The Company, therefore, does not make any guarantees as to the completeness or accuracy of this email or any attachments. This email is for informational purposes only and does not constitute a recommendation, offer, request or solicitation of any kind to buy, sell, subscribe, redeem or perform any type of transaction of a financial product.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux