Hi Lon, thanks for the reply. My cluster.conf is as follows. coral1 and coral2 are my 2 nodes. <?xml version="1.0"?> <cluster config_version="264" name="MF_Cluster"> <fence_daemon clean_start="1" post_fail_delay="0" post_join_delay="3"/> <clusternodes> <clusternode name="coral1" votes="1"> <fence> <method name="1"> <device name="pdu1" option="off" port="13" switch="1"/> <device name="pdu2" option="off" port="13" switch="2"/> <device name="pdu1" option="on" port="13" switch="1"/> <device name="pdu2" option="on" port="13" switch="2"/> </method> </fence> </clusternode> <clusternode name="coral2" votes="1"> <fence> <method name="1"> <device name="pdu1" option="off" port="20" switch="1"/> <device name="pdu2" option="off" port="20" switch="2"/> <device name="pdu1" option="on" port="20" switch="1"/> <device name="pdu2" option="on" port="20" switch="2"/> </method> </fence> </clusternode> </clusternodes> <fencedevices> <fencedevice agent="fence_apc" ipaddr="10.10.50.100" login="apc" name="pdu1" passwd="apc"/> <fencedevice agent="fence_apc" ipaddr="10.10.50.101" login="apc" name="pdu2" passwd="apc"/> </fencedevices> <rm> <failoverdomains> <failoverdomain name="MF_Failover" ordered="0" restricted="1"> <failoverdomainnode name="coral2" priority="1"/> <failoverdomainnode name="coral1" priority="1"/> </failoverdomain> </failoverdomains> <resources> <fs device="/dev/sda1" force_fsck="0" force_unmount="1" fstype="ext3" mountpoint="/MF/MF_v1.1/shared" name="testmount" options="" self_fence="0"/> <script file="/etc/rc.d/init.d/msgfwd" name="Message Forwarder"/> <ip address="10.10.50.22" monitor_link="1"/> <script file="/etc/rc.d/init.d/namesvc" name="Name Service"/> </resources> <service autostart="1" domain="MF_Failover" name="msgfwd" recovery="relocate"> <fs ref="testmount"/> <script ref="Message Forwarder"/> <ip ref="10.10.50.22"/> </service> </rm> <cman expected_votes="1" two_node="1"/> </cluster> --- Lon Hohberger <lhh@xxxxxxxxxx> wrote: > On Wed, 2006-08-16 at 22:37 +0800, Gary Lua wrote: > > Hi, > > > > I'm currently configuring fencing devices for my 2 nodes on a > RHEL4 > > cluster. The problem is quite long, so please bear with me. > > > > I have 2 nodes (let's call them stone1 and stone2) and 2 APC > fencing > > devices (pdu1 and pdu2, both apc 7952 devices). Both stone1 and > stone2 > > has dual power supplies. Stone1's power supplies are connected to > outlet > > 13 of pdu1 and pdu2. Stone2's power supplies are connected to > outlet 20 > > of both the pdus. My question is: during the fencing configuration > for > > each node, i need to specify which fence device to add to the > fence > > level of each node. Is it correct to specify for stone1 as follows > : > > pdu1 -> port=13, switch=1, pdu2-> port=13, switch=2? The same > applies to > > stone 2 : pdu1-> port=20, switch=1, pdu2-> port=20, switch=2? > > > > After configuring as mentioned above, with both nodes on the > cluster > > running and my application running on stone1, i pull out the > ethernet > > cables for stone1 to simulate that the server is down. By right, > my > > application should fail over to stone2 and fencing should occur to > > > stone1 (ie, stone1 should be rebooted/shutdown). However, what > happened > > is that my application is started on stone2, and stone1 is not > fenced. > > In fact, when i reconnect by cables, my application is still > running on > > stone1! Seems that there are 2 instances of my application > running, each > > on stone1 and stone2. > > Post the cluster configuration. > > -- Lon > > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster