Hello Digimer, Yes, I would agrre with you RHEL6.4 is old. We patched monthly, but I'm not sure why these servers are still at 6.4. Most of our system are 6.6. Here is my cluster config. All I want is using cluster to have BGFS2 mount via /etc/fstab. root@ustlvcmsp1955 ~]# cat /etc/cluster/cluster.conf <?xml version="1.0"?> <cluster config_version="15" name="p1954_to_p1958"> <clusternodes> <clusternode name="ustlvcmsp1954" nodeid="1"/> <clusternode name="ustlvcmsp1955" nodeid="2"/> <clusternode name="ustlvcmsp1956" nodeid="3"/> <clusternode name="ustlvcmsp1957" nodeid="4"/> <clusternode name="ustlvcmsp1958" nodeid="5"/> </clusternodes> <fencedevices> <fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.108" login="rhfence" name="p1954" passwd="xxxxxxxx"/> <fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.109" login="rhfence" name="p1955" passwd=" xxxxxxxx "/> <fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.110" login="rhfence" name="p1956" passwd=" xxxxxxxx "/> <fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.111" login="rhfence" name="p1957" passwd=" xxxxxxxx "/> <fencedevice agent="fence_vmware_soap" ipaddr="10.30.197.112" login="rhfence" name="p1958" passwd=" xxxxxxxx "/> </fencedevices> </cluster> clustat show: Cluster Status for p1954_to_p1958 @ Wed Jan 7 15:38:00 2015 Member Status: Quorate Member Name ID Status ------ ---- ---- ------ ustlvcmsp1954 1 Offline ustlvcmsp1955 2 Online, Local ustlvcmsp1956 3 Online ustlvcmsp1957 4 Offline ustlvcmsp1958 5 Online I need to make them all online, so I can use fencing for mounting shared disk. Thanks, Vinh -----Original Message----- From: linux-cluster-bounces@xxxxxxxxxx [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Digimer Sent: Wednesday, January 07, 2015 3:16 PM To: linux clustering Subject: Re: needs helps GFS2 on 5 nodes cluster My first though would be to set <fence_daemon post_join_delay="30" /> in cluster.conf. If that doesn't work, please share your configuration file. Then, with all nodes offline, open a terminal to each node and run 'tail -f -n 0 /var/log/messages'. With that running, start all the nodes and wait for things to settle down, then paste the five nodes' output as well. Also, 6.4 is pretty old, why not upgrade to 6.6? digimer On 07/01/15 03:10 PM, Cao, Vinh wrote: > Hello Cluster guru, > > I'm trying to setup Redhat 6.4 OS cluster with 5 nodes. With two nodes > I don't have any issue. > > But with 5 nodes, when I ran clustat I got 3 nodes online and the > other two off line. > > When I start the one that are off line. Service cman start. I got: > > [root@ustlvcmspxxx ~]# service cman status > > corosync is stopped > > [root@ustlvcmsp1954 ~]# service cman start > > Starting cluster: > > Checking if cluster has been disabled at boot... [ OK ] > > Checking Network Manager... [ OK ] > > Global setup... [ OK ] > > Loading kernel modules... [ OK ] > > Mounting configfs... [ OK ] > > Starting cman... [ OK ] > > Waiting for quorum... Timed-out waiting for cluster > > [FAILED] > > Stopping cluster: > > Leaving fence domain... [ OK ] > > Stopping gfs_controld... [ OK ] > > Stopping dlm_controld... [ OK ] > > Stopping fenced... [ OK ] > > Stopping cman... [ OK ] > > Waiting for corosync to shutdown: [ OK ] > > Unloading kernel modules... [ OK ] > > Unmounting configfs... [ OK ] > > Can you help? > > Thank you, > > Vinh > > > -- Digimer Papers and Projects: https://alteeve.ca/w/ What if the cure for cancer is trapped in the mind of a person without access to education? -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster