Try this: ceph osd crush reweight osd.XX 0 --Mike On 5/28/22 15:02, Nico Schottelius wrote:
Good evening dear fellow Ceph'ers, when removing OSDs from a cluster, we sometimes use ceph osd reweight osd.XX 0 and wait until the OSD's content has been redistributed. However, when then finally stopping and removing it, Ceph is again rebalancing. I assume this is due to a position that is removed in the CRUSH map and thus the logical placement is "wrong". (Am I wrong about that?) I wonder, is there a way to tell ceph properly that a particular OSD is planned to leave the cluster and to remove the data to the "correct new position" instead of doing the rebalance dance twice? Best regards, Nico -- Sustainable and modern Infrastructures by ungleich.ch _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx