Re: Feature request: "max mon" setting

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Nov 3, 2011 at 05:02, Amon Ott <a.ott@xxxxxxxxxxxx> wrote:
> Documentation recommends three monitors. In our special cluster configuration,
> this would mean that if accidentially two nodes with monitors fail (e.g. one
> in maintenance and one crashes), the whole cluster dies. What I would really

If you feel two monitors going down is too likely, run a monitor
cluster of size 5. And if you feel 3 monitors going down is too
likely, run 7.

> like would be that I can define a monitor on each node and e.g. set "max mon
> = 3". Each monitor starting up can then check how many monitors are already
> up and go to standby, if the number has already been reached. Regular
> rechecking could allow another monitor to become active, if one of the
> previously active monitors has died. Just like "max mds" actually.

Unfortunately, that is fundamentally not wanted. That would let a so
called "split brain" situation occur, and the whole purpose of the
majority rule of monitors is to ensure that it does not happen. If we
didn't care about that, the monitors left standing would never need to
stop operating.

> A special case that gives me most headaches is the case of just two active
> nodes. According to documentation, the monitor problem means that one failing
> monitor kills the cluster whatever the number of defined monitors (1 or 2),
> even if we have all data safely placed on both nodes.

Yes, 1 or 2 total physical nodes in a Ceph cluster makes it hard to do
HA. You could run just one ceph-mon, then the other node failing
doesn't affect the cluster at all (but naturally the one running
ceph-mon must not fail).

Perhaps you can get a third machine to also be a monitor, even if it
doesn't participate in storage etc. ceph-mon is a very lightweight
process. Share a server that has other responsibilities, or buy the
cheapest atom netbook you can find -- it should do the job just fine.

Clusters of <=2 nodes are pretty far from what Ceph was designed for,
and while we use them all the time for testing, running a setup that
small for real is pretty rare. Sorry.
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux