This was dealt with on ceph-devel mailing list. On Thu, Dec 29, 2016 at 11:08 PM, Łukasz Chrustek <skidoo@xxxxxxx> wrote: > Hi, > > As I wrote at first mail - I have already done that, but I'm affraid > to load it back - is there any chance, that something will go wrong ? > > Thanks for answer ! > >> Hi, > > >> Hmm, could you try and dump the crush map - decompile it - modify >> it to remove the DNE osd's, compile it and load it back into ceph? > > >> http://docs.ceph.com/docs/master/rados/operations/crush-map/#get-a-crush-map > > > >> Thanks > >> On Thu, Dec 29, 2016 at 1:01 PM, Łukasz Chrustek <skidoo@xxxxxxx> wrote: > >> Hi, > >> ]# ceph osd tree >> ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY >> -7 16.89590 root ssd-disks >> -11 0 host ssd1 >> 598798032 0 osd.598798032 DNE 0 >> 21940 0 osd.21940 DNE 0 >> 71 0 osd.71 DNE 0 > >> ]# ceph osd rm osd.598798032 >> Error EINVAL: osd id 598798032 is too largeinvalid osd id-34 >> ]# ceph osd rm osd.21940 >> osd.21940 does not exist. >> ]# ceph osd rm osd.71 >> osd.71 does not exist. > > > >> ceph osd rm osd.$ID > > >> On Thu, Dec 29, 2016 at 10:44 AM, Łukasz Chrustek <skidoo@xxxxxxx> wrote: > > >> Hi, > > >> I was trying to delete 3 osds from cluster, deletion procces took very > >> long time and I interrupted it. mon process then crushed, and in ceph > >> osd tree (after restart ceph-mon) I saw: > > >> ~]# ceph osd tree > >> ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY > >> -7 16.89590 root ssd-disks > >> -11 0 host ssd1 > >> -231707408 0 > >> 22100 0 osd.22100 DNE 0 > >> 71 0 osd.71 DNE 0 > > > >> when I tried to delete osd.22100: > > >> [root@cc1 ~]# ceph osd crush remove osd.22100 > >> device 'osd.22100' does not appear in the crush map > > >> then I tried to delete osd.71 and mon proccess crushed: > > >> [root@cc1 ~]# ceph osd crush remove osd.71 > >> 2016-12-28 17:52:34.459668 7f426a862700 0 monclient: hunting for new mon > > >> after restart of ceph-mon in ceph osd tree it shows: > > >> # ceph osd tree > >> ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY > >> -7 16.89590 root ssd-disks > >> -11 0 host ssd1 > >> 598798032 0 osd.598798032 DNE 0 > >> 21940 0 osd.21940 DNE 0 > >> 71 0 osd.71 DNE 0 > > >> My question is how to delete this osds without direct editing crushmap > >> ? It is production system, I can't affort any service interruption :(, > >> when I try to ceph osd crush remove then ceph-mon crushes.... > > >> I dumped crushmap, but it took 19G (!!) after decompiling (compiled > >> file is very small). So, I cleaned this file with perl (it take very > >> long time), and I have now small txt crushmap, which I edited. But is > >> there any chance that ceph will still remember somewhere about this > >> huge numbers for osds ? Is it safe to apply this cleaned crushmap to > >> cluster ? Cluster now works OK, but there is over 23TB production data > >> which I can't loose. Please advice what to do. > > > >> -- > >> Regards > >> Luk > > >> _______________________________________________ > >> ceph-users mailing list > >> ceph-users@xxxxxxxxxxxxxx > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > >> -- >> Pozdrowienia, >> Łukasz Chrustek > > > > > > > -- > Pozdrowienia, > Łukasz Chrustek > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Cheers, Brad _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com