If you are trying to reduce extra data movement, set and unset the nobackfill and norecover flags when you do the same for noout. You will want to
follow the instructions to fully remove the osd from the cluster including outing the osd, removing it from the crush map, removing it's auth from the cluster, and finally remove the osd from the cluster. After that, adding the osd back in should give it
the same osd id that the former one had. If you make sure that the id is the same and the weight in the crush map is the same (you can do this by saving your crush map before you remove the osd and uploading the same crush map after you add it back in with
the same id) then the only data movement will be onto the re-added osd and nothing else.
David Turner |
Cloud Operations Engineer |
StorageCraft
Technology Corporation 380 Data Drive Suite 300 | Draper | Utah | 84020 Office: 801.871.2760 | Mobile: 385.224.2943 |
If you are not the intended recipient of this message or received it erroneously, please notify the sender and delete it, together with any attachments, and be advised that any dissemination or copying of this message is prohibited. |
________________________________________
From: ceph-users [ceph-users-bounces@xxxxxxxxxxxxxx] on behalf of Goncalo Borges [goncalo.borges@xxxxxxxxxxxxx]
Sent: Sunday, August 14, 2016 5:47 AM
To: ceph-users@xxxxxxxx
Subject: Substitute a predicted failure (not yet failed) osd
Hi cephfers
I have a really simple question: the documentation always refers to the procedure to substitute failed disks. Currently I have a predicted failure in a raid 0 osd and I would like to substitute before it fails without having to go by replicating pgs once the osd is removed from crush map, and then, replicating again once I add the new drive.
Can I perform the following actions safely to achieve my goal?
# ceph osd set noout
# stop the osd
# unmount the osd
# remove it from crush map
# substitute the drive
# recreate the osd
# ceph osd unset noout
Cheers
Goncalo
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
From: ceph-users [ceph-users-bounces@xxxxxxxxxxxxxx] on behalf of Goncalo Borges [goncalo.borges@xxxxxxxxxxxxx]
Sent: Sunday, August 14, 2016 5:47 AM
To: ceph-users@xxxxxxxx
Subject: Substitute a predicted failure (not yet failed) osd
Hi cephfers
I have a really simple question: the documentation always refers to the procedure to substitute failed disks. Currently I have a predicted failure in a raid 0 osd and I would like to substitute before it fails without having to go by replicating pgs once the osd is removed from crush map, and then, replicating again once I add the new drive.
Can I perform the following actions safely to achieve my goal?
# ceph osd set noout
# stop the osd
# unmount the osd
# remove it from crush map
# substitute the drive
# recreate the osd
# ceph osd unset noout
Cheers
Goncalo
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com