On 11 January 2016 at 02:10, Rafael Lopez <rafael.lopez@xxxxxxxxxx> wrote: > @Steve, even when you remove due to failing, have you noticed that the cluster rebalances twice using the documented steps? You may not if you don't wait for the initial recovery after 'ceph osd out'. If you do 'ceph osd out' and immediately 'ceph osd crush remove', RH support has told me that this effectively 'cancels' the original move triggered from 'ceph osd out' and starts permanently remapping... which still doesn't really explain why we have to do the ceph osd out in the first place.. This topic was last discussed in December - the documentation for removing an OSD from the cluster is not helpful. Unfortunately it doesn't look like anyone is going to fix the documentation. http://comments.gmane.org/gmane.comp.file-systems.ceph.user/25627 Basically, when you want to remove an OSD, there's an alternative sequence of commands that avoids the double-rebalance. The better approach is to reweight the OSD to zero first, then wait for the (one and only) rebalance, then mark out and remove. Here's more details from the previous thread: http://permalink.gmane.org/gmane.comp.file-systems.ceph.user/25629 Thanks, Andy _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com