On Wed, Oct 3, 2018 at 3:52 PM Andras Pataki <apataki@xxxxxxxxxxxxxxxxxxxxx> wrote: > > Ok, understood (for next time). > > But just as an update/closure to my investigation - it seems this is a > feature of ceph-volume (that it can't just create an OSD from scratch > with a given ID), not of base ceph. The underlying ceph command (ceph > osd new) very happily accepts an osd-id as an extra optional argument > (after the fsid), and creates and osd with the given ID. In fact, a > quick change to ceph_volume (create_id function in prepare.py) will make > ceph-volume recreate the OSD with a given ID. I'm not a ceph-volume > expert, but a feature to create an OSD with a given ID from scratch > would be nice (given that the underlying raw ceph commands already > support it). That is something that I wasn't aware of, thanks for bringing it up. I've created an issue on the tracker to accommodate for that behavior: http://tracker.ceph.com/issues/36307 > > Andras > > On 10/3/18 11:41 AM, Alfredo Deza wrote: > > On Wed, Oct 3, 2018 at 11:23 AM Andras Pataki > > <apataki@xxxxxxxxxxxxxxxxxxxxx> wrote: > >> Thanks - I didn't realize that was such a recent fix. > >> > >> I've now tried 12.2.8, and perhaps I'm not clear on what I should have > >> done to the OSD that I'm replacing, since I'm getting the error "The osd > >> ID 747 is already in use or does not exist.". The case is clearly the > >> latter, since I've completely removed the old OSD (osd crush remove, > >> auth del, osd rm, wipe disk). Should I have done something different > >> (i.e. not remove the OSD completely)? > > Yeah, you completely removed it so now it can't be re-used. This is > > the proper way if wanting to re-use the ID: > > > > http://docs.ceph.com/docs/master/rados/operations/add-or-rm-osds/#rados-replacing-an-osd > > > > Basically: > > > > ceph osd destroy {id} --yes-i-really-mean-it > > > >> Searching the docs I see a command 'ceph osd destroy'. What does that > >> do (compared to my removal procedure, osd crush remove, auth del, osd rm)? > >> > >> Thanks, > >> > >> Andras > >> > >> > >> On 10/3/18 10:36 AM, Alfredo Deza wrote: > >>> On Wed, Oct 3, 2018 at 9:57 AM Andras Pataki > >>> <apataki@xxxxxxxxxxxxxxxxxxxxx> wrote: > >>>> After replacing failing drive I'd like to recreate the OSD with the same > >>>> osd-id using ceph-volume (now that we've moved to ceph-volume from > >>>> ceph-disk). However, I seem to not be successful. The command I'm using: > >>>> > >>>> ceph-volume lvm prepare --bluestore --osd-id 747 --data H901D44/H901D44 > >>>> --block.db /dev/disk/by-partlabel/H901J44 > >>>> > >>>> But it created an OSD the ID 601, which was the lowest it could allocate > >>>> and ignored the 747 apparently. This is with ceph 12.2.7. Any ideas? > >>> Yeah, this was a problem that was fixed and released as part of 12.2.8 > >>> > >>> The tracker issue is: http://tracker.ceph.com/issues/24044 > >>> > >>> The Luminous PR is https://github.com/ceph/ceph/pull/23102 > >>> > >>> Sorry for the trouble! > >>>> Andras > >>>> > >>>> _______________________________________________ > >>>> ceph-users mailing list > >>>> ceph-users@xxxxxxxxxxxxxx > >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com