Re: Substitute a predicted failure (not yet failed) osd

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



If you are trying to reduce extra data movement, set and unset the nobackfill and norecover flags when you do the same for noout.  You will want to follow the instructions to fully remove the osd from the cluster including outing the osd, removing it from the crush map, removing it's auth from the cluster, and finally remove the osd from the cluster.  After that, adding the osd back in should give it the same osd id that the former one had.  If you make sure that the id is the same and the weight in the crush map is the same (you can do this by saving your crush map before you remove the osd and uploading the same crush map after you add it back in with the same id) then the only data movement will be onto the re-added osd and nothing else.


David Turner | Cloud Operations Engineer | StorageCraft Technology Corporation
380 Data Drive Suite 300 | Draper | Utah | 84020
Office: 801.871.2760 | Mobile: 385.224.2943


If you are not the intended recipient of this message or received it erroneously, please notify the sender and delete it, together with any attachments, and be advised that any dissemination or copying of this message is prohibited.


________________________________________
From: ceph-users [ceph-users-bounces@xxxxxxxxxxxxxx] on behalf of Goncalo Borges [goncalo.borges@xxxxxxxxxxxxx]
Sent: Sunday, August 14, 2016 5:47 AM
To: ceph-users@xxxxxxxx
Subject: Substitute a predicted failure (not yet failed) osd

Hi cephfers

I have a really simple question: the documentation always refers to the procedure to substitute failed disks. Currently I have a predicted failure in a raid 0 osd and I would like to substitute before it fails without having to go by replicating pgs once the osd is removed from crush map, and then, replicating again once I add the new drive.

Can I perform the following actions safely  to achieve my goal?

# ceph osd set noout
# stop the osd
# unmount the osd
# remove it from crush map
# substitute the drive
# recreate the osd
# ceph osd unset noout

Cheers
Goncalo


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux