> On Oct 19, 2024, at 2:47 PM, Shain Miley <SMiley@xxxxxxx> wrote: > > We are running octopus but will be upgrading to reef or squid in the next few weeks. As part of that upgrade I am planning on switching over to using cephadm as well. > > Part of what I am doing right now is going through and replacing old drives and removing some of our oldest nodes and replacing them with new ones…then I will convert the rest of the filestore osd over to bluestore so that I can upgrade. > > One other question based on your suggestion below…my typical process of removing or replacing an osd involves the following: > > ceph osd crush reweight osd.id <http://osd.id/> 0.0 > ceph osd out osd.id <http://osd.id/> > service ceph stop osd.id <http://osd.id/> > ceph osd crush remove osd.id <http://osd.id/> > ceph auth del osd.id <http://osd.id/> > ceph osd rm id > > Does `ceph osd destroy` do something other than the last 3 commands above or am I just doing the same thing using multiple commands? If I need to start issuing the destroy command as well I can. > I don’t recall if it will stop the service if running, but it does leave the OSD in the CRUSH map marked as ‘destroyed’. I *think* it leaves the auth but I’m not sure. _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx