With four mons total then only one can be down... mon.osd01 is down already you're at the limit. It's possible that whichever reason is preventing this mon from joining will also prevent the new mon from joining. I think you should: 1. Investigate why mon.osd01 isn't coming back into the quorum... The logs on that mon or the others can help. 2. If you decide to give up on mon.osd01, then first you should rm it from the cluster before you add a mon from another host. .. Dan On Sun, 25 Jul 2021, 12:43 Ansgar Jazdzewski, <a.jazdzewski@xxxxxxxxxxxxxx> wrote: > hi folks > > I have a cluster running ceph 14.2.22 on ubuntu 18.04 and some hours > ago one of the mons stopped working and the on-call team rebooted the > node; not the mon is is not joining the ceph-cluster. > > TCP ports of mons are open and reachable! > > ceph health detail > HEALTH_WARN 1/3 mons down, quorum osd02,osd03 > MON_DOWN 1/3 mons down, quorum osd02,osd03 > mon.osd01 (rank 0) addr > [v2:10.152.28.171:3300/0,v1:10.152.28.171:6789/0] is down (out of > quorum) > > I like to add a new 3rd mon to the cluster on osd04 but I'm a bit > scared as it can result in 50% of the mons are not in reach!? > > Question: should I remove the mon on osd01 first and recreate the > demon before starting a new mon on osd04? > > > Thanks for your input! > Ansgar > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx