Increasing the timeouts definitely helped. I also was using a
flawed test to simulate a failure. I was taking down the network; which left
qdisk in tack. Cman fenced the node before qdisk had a chance to do anything.
A hard power down produced the results I would expect. From:
linux-cluster-bounces@xxxxxxxxxx [mailto:linux-cluster-bounces@xxxxxxxxxx] On
Behalf Of Tomasz Sucharzewski I had the same issue and I solved it, Just increase quorum check interval. 2 seconds is to less to
inform cman about quorum status. I had to increase it to 7 seconds but remember it also
influence cman timeout which must be verified. Best regards, Tomek On Feb 17, 2009, at 9:12 PM, Hunt, Gary wrote:
Having an issue with my 2 node cluster. Think it is related
to the quorum disk. 2 node RHEL 5.3 cluster with quorum disk. Virtual servers
running on each node. Whenever node1 takes over the master role in qdisk it looses
quorum and restarts all the virtual servers. It does regain quorum a few
seconds later. If node1 is already the master and I fail node2; things
work as expected. Node2 doesn’t seem to have a problem taking over master
role. Whenever node1 needs to take over master role the cluster looses
quorum. Here is my cluster.conf. Any suggestions on what may be
causing this? <?xml version="1.0"?> <cluster alias="xencluster"
config_version="13" name="xencluster"> <fence_daemon
clean_start="0" post_fail_delay="0"
post_join_delay="3"/> <clusternodes>
<clusternode name="ricci2b.gallup.com" nodeid="2"
votes="1">
<fence>
<method name="1">
<device name="ricci2b"/>
</method>
</fence>
</clusternode>
<clusternode name="ricci1b.gallup.com" nodeid="1"
votes="1">
<fence>
<method name="1">
<device name="ricci1b"/>
</method>
</fence>
</clusternode> </clusternodes> <cman
expected_votes="3" two_node="0"/> <fencedevices>
<fencedevice agent="fence_ipmilan" ipaddr="172.30.3.110"
login="xxxx" name="ricci1b" passwd="xxxxxx"/>
<fencedevice
agent="fence_ipmilan" ipaddr="172.30.3.140"
login="xxxx" name="ricci2b" passwd="xxxxxx"/> </fencedevices> <rm>
<failoverdomains/>
<resources/>
<vm autostart="1" exclusive="0"
name="rhel_full" path="/xenconfigs"
recovery="restart"/>
<vm autostart="1" exclusive="0"
name="rhel_para" path="/xenconfigs"
recovery="restart"/> </rm> <quorumd
interval="2" label="quorum_disk_from_ricci1"
min_score="1" tko="3" votes="1"/> <totem
consensus="4800" join="60" token="12000"
token_retransmits_before_loss_const="20"/> </cluster> Thanks Gary -- |
-- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster