Re: frequent Monitor down

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Typically, the number of nodes is 2n+1 to cover n failures.
It's OK to have 4 nodes, from failure covering POV, it's the same
as 3 nodes. 4 nodes will cover 1 failure. If 2 nodes down, the
cluster is down. It works, just not make much sense.

Thanks!
Tony
> -----Original Message-----
> From: Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx>
> Sent: Thursday, October 29, 2020 1:42 AM
> To: andrei <andrei@xxxxxxxxxx>; eblock <eblock@xxxxxx>
> Cc: ceph-users <ceph-users@xxxxxxx>
> Subject:  Re: frequent Monitor down
> 
> Really? First time I read this here, afaik you can get a split brain
> like this.
> 
> 
> 
> -----Original Message-----
> Sent: Thursday, October 29, 2020 12:16 AM
> To: Eugen Block
> Cc: ceph-users
> Subject:  Re: frequent Monitor down
> 
> Eugen, I've got four physical servers and I've installed mon on all of
> them. I've discussed it with Wido and a few other chaps from ceph and
> there is no issue in doing it. The quorum issues would happen if you
> have 2 mons. If you've got more than 2 you should be fine.
> 
> Andrei
> 
> ----- Original Message -----
> > From: "Eugen Block" <eblock@xxxxxx>
> > To: "Andrei Mikhailovsky" <andrei@xxxxxxxxxx>
> > Cc: "ceph-users" <ceph-users@xxxxxxx>
> > Sent: Wednesday, 28 October, 2020 20:19:15
> > Subject: Re:  Re: frequent Monitor down
> 
> > Why do you have 4 MONs in the first place? That way a quorum is
> > difficult to achieve, could it be related to that?
> >
> > Zitat von Andrei Mikhailovsky <andrei@xxxxxxxxxx>:
> >
> >> Yes, I have, Eugen, I see no obvious reason / error / etc. I see a
> >> lot of entries relating to Compressing as well as monitor going down.
> >>
> >> Andrei
> >>
> >>
> >>
> >> ----- Original Message -----
> >>> From: "Eugen Block" <eblock@xxxxxx>
> >>> To: "ceph-users" <ceph-users@xxxxxxx>
> >>> Sent: Wednesday, 28 October, 2020 11:51:20
> >>> Subject:  Re: frequent Monitor down
> >>
> >>> Have you looked into syslog and mon logs?
> >>>
> >>>
> >>> Zitat von Andrei Mikhailovsky <andrei@xxxxxxxxxx>:
> >>>
> >>>> Hello everyone,
> >>>>
> >>>> I am having regular messages that the Monitors are going down and
> up:
> >>>>
> >>>> 2020-10-27T09:50:49.032431+0000 mon .arh-ibstorage2-ib ( mon .1)
> >>>> 2248 : cluster [WRN] Health check failed: 1/4 mons down, quorum
> >>>> arh-ibstorage2-ib,arh-ibstorage3-ib,arh-ibstorage4-ib (MON_DOWN)
> >>>> 2020-10-27T09:50:49.123511+0000 mon .arh-ibstorage2-ib ( mon .1)
> >>>> 2250 : cluster [WRN] overall HEALTH_WARN 23 OSD(s) experiencing
> >>>> BlueFS spillover; 3 large omap objects; 1/4 mons down, quorum
> >>>> arh-ibstorage2-ib,arh-ibstorage3-ib,arh-ibstorage4-ib; noout
> >>>> flag(s) set; 43 pgs not deep-scrubbed in time; 12 pgs not scrubbed
> >>>> in time 2020-10-27T09:50:52.735457+0000 mon .arh-ibstorage1-ib (
> >>>> mon .0)
> >>>> 31287 : cluster [INF] Health check cleared: MON_DOWN (was: 1/4 mons
> 
> >>>> down, quorum arh-ibstorage2-ib,arh-ibstorage3-ib,arh-ibstorage4-ib)
> >>>> 2020-10-27T12:35:20.556458+0000 mon .arh-ibstorage2-ib ( mon .1)
> >>>> 2260 : cluster [WRN] Health check failed: 1/4 mons down, quorum
> >>>> arh-ibstorage2-ib,arh-ibstorage3-ib,arh-ibstorage4-ib (MON_DOWN)
> >>>> 2020-10-27T12:35:20.643282+0000 mon .arh-ibstorage2-ib ( mon .1)
> >>>> 2262 : cluster [WRN] overall HEALTH_WARN 23 OSD(s) experiencing
> >>>> BlueFS spillover; 3 large omap objects; 1/4 mons down, quorum
> >>>> arh-ibstorage2-ib,arh-ibstorage3-ib,arh-ibstorage4-ib; noout
> >>>> flag(s) set; 47 pgs not deep-scrubbed in time; 14 pgs not scrubbed
> >>>> in time
> >>>>
> >>>>
> >>>> This happens on a daily basis several times a day.
> >>>>
> >>>> Could you please let me know how to fix this annoying problem?
> >>>>
> >>>> I am running ceph version 15.2.4
> >>>> (7447c15c6ff58d7fce91843b705a268a1917325c) octopus (stable) on
> >>>> Ubuntu 18.04 LTS with latest updates.
> >>>>
> >>>> Thanks
> >>>>
> >>>> Andrei
> >>>> _______________________________________________
> >>>> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send
> >>>> an email to ceph-users-leave@xxxxxxx
> >>>
> >>>
> >>> _______________________________________________
> >>> ceph-users mailing list -- ceph-users@xxxxxxx
> > >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an
> email to ceph-users-leave@xxxxxxx
> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an
> email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux