Re: RH Cluster doesn't pass basic acceptance tests - bug in fenced?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



This isn't a bug, its working as expected.

IT People from the central bank doesn't think like that. I cannot blame
them, because it is strange to me, and to anybody who has seen this RH
cluster behaviour.
I have seen this behavior.  It is not strange to me.  This is only strange to people who do not understand how quorum systems work.

What you need in qdisk, set it up
with the proper hueristics and it will force the shutdown of the bad node before
the bad node has a chance to fence off the working node.

This is just a workaround for lack of communication between clurgmgrd and
fenced daemons, where first is aware of ethernet/network failure and is
trying to disable active service, and fenced which is fencing other node
without any good reason, because it doesn't know that it's node is faulty one.

This is *NOT* a workaround for lack of communication.  clurgmgrd is responsible for starting and stopping services.  Fencing is responsible for keeping nodes running.  clurgmgrd does not have the information and is not the right service to handle this.

The problem is that you have a two node cluster.  If you had three nodes, this would not be an issue.  In a two-node cluster, the two nodes are both capable of fencing each other even though they no longer have quorum.  There is mathematically no other way to have a majority of 2 nodes without both of them.

The Quorum Disk allows the running nodes to use a heuristic--like the ethernet link check you speak of (or a ping to the network gateway which would also be helpful).  This heuristic allows you to artificially reach quorum by giving extra votes to the node that can still determine that it is okay.

I have even better workaround (one bonding with native data ethernet and
tagged vlan for fence subnet) for this silly behaviour, but I will really
like to see this thing fixed, because people are laughing on us when
testing our cluster configurations (we are configuring Red Hat machines
and clusters).
The moment that a node fails for any reason other than an ethernet disconnection your workaround falls apart.

If some "Central Bank" is truly your customer, then you should be able to obtain a third node with no problems.  Otherwise, the Quorum Disk provides better behavior than your "workaround" by actually solving the problem in a generally applicable and sophisticated way.

This is a configuration problem.  If you desire not to be laughed at learn how to configure your software.  Also, for what its worth, I don't use bonding on my machines due to the switches I utilize (I use bridging instead), but I would recommend keeping this for reliability of the ethernet, as it is an important failure case.

-- 
Jayson Vantuyl
Systems Architect


--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux