RHCS simple ip-failover problem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Dear List,

I've received today a evaluation License for RHEL 6.1[Red Hat Enterprise Linux Server release 6.1 (Santiago)] for learning purposes with the RHCS.

So I've setup 2 machines vbox1 and vbox2, with 2 Interfaces (Intranet + Hearbeat) and created a cluster.conf with a simple IP-fail-over Scenario.
Starting the cluster no node takes the IP.
After checking cluster state I'm a little bit confused, it doesn't list the service.
Maybe i understood something wrong from the documentation.

[root@vbox1 cluster]# clustat  -l
Cluster Status for vbox @ Sun Nov 27 14:38:00 2011
Member Status: Quorate

 Member Name                             ID   Status
 ------ ----                             ---- ------
 vbox1.example.local                         1 Online, Local
 vbox2.example.local                         2 Online

By testing the rules the system takes the ip-address.
[root@vbox1 cluster]# rg_test test /etc/cluster/cluster.conf start service webprod
Running in test mode.
..
Starting webprod...
<debug>  Link for eth0: Detected
Link for eth0: Detected
<info>   Adding IPv4 address 192.168.99.100/24 to eth0
Adding IPv4 address 192.168.99.100/24 to eth0
<debug>  Pinging addr 192.168.99.100 from dev eth0
Pinging addr 192.168.99.100 from dev eth0
<debug>  Sending gratuitous ARP: 192.168.99.100 00:0c:29:00:d1:05 brd ff:ff:ff:ff:ff:ff
Sending gratuitous ARP: 192.168.99.100 00:0c:29:00:d1:05 brd ff:ff:ff:ff:ff:ff
rdisc: no process killed
Start of webprod complete

[root@vbox1 cluster]# ip addr list eth0
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000
    link/ether 00:0c:29:00:d1:05 brd ff:ff:ff:ff:ff:ff
    inet 192.168.99.11/24 brd 192.168.99.255 scope global eth0
    inet 192.168.99.100/24 scope global secondary eth0
    inet6 fe80::20c:29ff:fe00:d105/64 scope link
       valid_lft forever preferred_lft forever

maybe someone give me a point to the right direction?

/etc/hosts
# Intranet
192.168.99.11 vbox1.example.local vbox1
192.168.99.12 vbox2.example.local vbox2
# Heartbeat
192.168.1.11 h-vbox1.example.local h-vbox1
192.168.1.12 h-vbox2.example.local h-vbox2
# Service-IP
192.168.99.100 vbox.example.local vbox

/etc/cluster/cluster.conf (for testing purposes only manual fencing. later i'l try it with fence_ipmilan)

<?xml version="1.0"?>
<cluster config_version="3" name="vbox">
<cman expected_votes="1" two_node="1"/>
<clusternodes>
<clusternode name="vbox1.example.local" nodeid="1">
<altname name="h-vbox1.example.local"/>
<fence>
<method name="n1">
<device name="human" nodename="vbox1.example.local"/>
</method>
</fence>
</clusternode>
<clusternode name="vbox2.example.local" nodeid="2">
<altname name="h-vbox2.example.local"/>
<fence>
<method name="n2">
<device name="human" nodename="vbox2.example.local"/>
</method>
</fence>
</clusternode>
</clusternodes>
<fencedevices>
<fencedevice agent="fence_manual" name="human"/>
</fencedevices>
<rm>
<resources>
<ip address="192.168.99.100" monitor_link="on" sleeptime="15"/>
</resources>
<service autostart="1" domain="web" exclusive="0" name="webprod" recovery="restart">
<ip ref="192.168.99.100"/>
</service>
<failoverdomains>
<failoverdomain name="web" nofailback="0" ordered="1" restricted="0">
<failoverdomainnode name="vbox1.example.local" priority="1"/>
<failoverdomainnode name="vbox2.example.local" priority="2"/>
</failoverdomain>
</failoverdomains>
</rm>
</cluster>

Thanks in advance
Chris
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux