Ours are enabled.. I have verified that.. Thanks -----Original Message----- From: linux-cluster-bounces@xxxxxxxxxx [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of John Ruemker Sent: Thursday, March 27, 2008 10:41 AM To: linux clustering Subject: Re: 3 node cluster problems I believe some of the cisco switches do not have multicast enabled by default which would prevent some of the cluster communications from getting through properly. http://kbase.redhat.com/faq/FAQ_51_11755 John Bennie Thomas wrote: > Are you using a private vlan for your cluster communications. If not, > you should be. the communicatuions > between the clustered nodes is very chatty Just my opinion. > > These are my opinions and experiences. > > Any views or opinions presented are solely those of the author and do > not necessarily represent those of Raytheon unless specifically > stated. Electronic communications including email might be monitored > by Raytheon. for operational or business reasons. > > > Dalton, Maurice wrote: >> Cisco 3550 >> >> >> -----Original Message----- >> From: linux-cluster-bounces@xxxxxxxxxx >> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Bennie Thomas >> Sent: Thursday, March 27, 2008 9:53 AM >> To: linux clustering >> Subject: Re: 3 node cluster problems >> >> what is the switch brand. I have read where the RHCS has problems with >> >> certain switches >> >> Dalton, Maurice wrote: >> >>> Switches >>> >>> Storage is fiber >>> >>> >>> -----Original Message----- >>> From: linux-cluster-bounces@xxxxxxxxxx >>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Bennie Thomas >>> Sent: Thursday, March 27, 2008 9:04 AM >>> To: linux clustering >>> Subject: Re: 3 node cluster problems >>> >>> How is your Cluster connections connected. (ie. Are you using a >>> hub,switch or direct connecting the heartbeat cables) ? >>> >>> Dalton, Maurice wrote: >>> >>>> Still having the problem. I can't figure it out. >>>> I just upgraded to the latest 5.1 cman.. No help.!!!!!!!!! >>>> >>>> >>>> -----Original Message----- >>>> From: linux-cluster-bounces@xxxxxxxxxx >>>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Bennie Thomas >>>> Sent: Tuesday, March 25, 2008 10:57 AM >>>> To: linux clustering >>>> Subject: Re: 3 node cluster problems >>>> >>>> >>>> Glad they are working. I have not used lvm with our Clusters. You >>>> >> know >> >>>> >>> >>>> have peaked >>>> my curiosity and I will have to try building one. So were you also >>>> >>> using >>> >>>> GFS ? >>>> >>>> Dalton, Maurice wrote: >>>> >>>>> Sorry but security here will not allow me to send host files >>>>> >>>>> BUT. >>>>> >>>>> >>>>> I was getting this in /var/log/messages on csarcsys3 >>>>> >>>>> Mar 25 15:26:11 csarcsys3-eth0 ccsd[7448]: Cluster is not quorate. >>>>> Refusing connection. >>>>> Mar 25 15:26:11 csarcsys3-eth0 ccsd[7448]: Error while processing >>>>> connect: Connection refused >>>>> Mar 25 15:26:12 csarcsys3-eth0 dlm_controld[7476]: connect to ccs >>>>> >>>> error >>>> >>>>> -111, check ccsd or cluster status >>>>> Mar 25 15:26:12 csarcsys3-eth0 ccsd[7448]: Cluster is not quorate. >>>>> Refusing connection. >>>>> Mar 25 15:26:12 csarcsys3-eth0 ccsd[7448]: Error while processing >>>>> connect: Connection refused >>>>> >>>>> >>>>> I had /dev/vg0/gfsvol on these systems. >>>>> >>>>> I did a lvremove >>>>> Restarted cman on all systems and for some strange reason my >>>>> >> clusters >> >>>>> are working. >>>>> >>>>> It doesn't make any sense. >>>>> >>>>> I can't thank you enough for your help.......!!!!!! >>>>> >>>>> >>>>> Thanks. >>>>> >>>>> >>>>> -----Original Message----- >>>>> From: linux-cluster-bounces@xxxxxxxxxx >>>>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Bennie Thomas >>>>> Sent: Tuesday, March 25, 2008 10:27 AM >>>>> To: linux clustering >>>>> Subject: Re: 3 node cluster problems >>>>> >>>>> I am currently running several 3-node cluster without a quorum disk. >>>>> >> >> >>>>> However, If you want your cluster to run >>>>> if only one node is up then you will need a quorum disk. Can you >>>>> >> send >> >>>>> >>> >>>>> your /etc/hosts file >>>>> for all systems, Also, could there be another node name called >>>>> csarcsys3-eth0 in your NIS or DNS >>>>> >>>>> I configured some using Conga and some with system-config-cluster. >>>>> >>>> When >>>>> using the system-config-cluster >>>>> I basically run the config on all nodes; just adding the nodenames >>>>> >>> and >>> >>>>> >>>> >>>>> cluster name. I reboot all nodes >>>>> to make sure they see each other then go back and modify the config >>>>> files. >>>>> >>>>> The file /var/log/messages should also shed some light on the >>>>> >>> problem. >>> >>>>> Dalton, Maurice wrote: >>>>> >>>>>> Same problem. >>>>>> >>>>>> I now have qdiskd running. >>>>>> >>>>>> I have ran diff's on all three cluster.conf files.. all are the >>>>>> >> same >> >>>>>> [root@csarcsys1-eth0 cluster]# more cluster.conf >>>>>> >>>>>> <?xml version="1.0"?> >>>>>> >>>>>> <cluster config_version="6" name="csarcsys5"> >>>>>> >>>>>> <fence_daemon post_fail_delay="0" post_join_delay="3"/> >>>>>> >>>>>> <clusternodes> >>>>>> >>>>>> <clusternode name="csarcsys1-eth0" nodeid="1" votes="1"> >>>>>> >>>>>> <fence/> >>>>>> >>>>>> </clusternode> >>>>>> >>>>>> <clusternode name="csarcsys2-eth0" nodeid="2" votes="1"> >>>>>> >>>>>> <fence/> >>>>>> >>>>>> </clusternode> >>>>>> >>>>>> <clusternode name="csarcsys3-eth0" nodeid="3" votes="1"> >>>>>> >>>>>> <fence/> >>>>>> >>>>>> </clusternode> >>>>>> >>>>>> </clusternodes> >>>>>> >>>>>> <cman/> >>>>>> >>>>>> <fencedevices/> >>>>>> >>>>>> <rm> >>>>>> >>>>>> <failoverdomains> >>>>>> >>>>>> <failoverdomain name="csarcsysfo" ordered="0" restricted="1"> >>>>>> >>>>>> <failoverdomainnode name="csarcsys1-eth0" priority="1"/> >>>>>> >>>>>> <failoverdomainnode name="csarcsys2-eth0" priority="1"/> >>>>>> >>>>>> <failoverdomainnode name="csarcsys3-eth0" priority="1"/> >>>>>> >>>>>> </failoverdomain> >>>>>> >>>>>> </failoverdomains> >>>>>> >>>>>> <resources> >>>>>> >>>>>> <ip address="172.24.86.177" monitor_link="1"/> >>>>>> >>>>>> <fs device="/dev/sdc1" force_fsck="0" force_unmount="1" >>>>>> >> fsid="57739" >> >>>>>> >>> >>>>>> fstype="ext3" mountpo >>>>>> >>>>>> int="/csarc-test" name="csarcsys-fs" options="rw" self_fence="0"/> >>>>>> >>>>>> </resources> >>>>>> >>>>>> </rm> >>>>>> >>>>>> <quorumd interval="4" label="csarcsysQ" min_score="1" tko="30" >>>>>> >>>>> votes="2"/> >>>>> >>>>>> </cluster> >>>>>> >>>>>> More info from csarcsys3 >>>>>> >>>>>> [root@csarcsys3-eth0 cluster]# clustat >>>>>> >>>>>> msg_open: No such file or directory >>>>>> >>>>>> Member Status: Inquorate >>>>>> >>>>>> Member Name ID Status >>>>>> >>>>>> ------ ---- ---- ------ >>>>>> >>>>>> csarcsys1-eth0 1 Offline >>>>>> >>>>>> csarcsys2-eth0 2 Offline >>>>>> >>>>>> csarcsys3-eth0 3 Online, Local >>>>>> >>>>>> /dev/sdd1 0 Offline >>>>>> >>>>>> [root@csarcsys3-eth0 cluster]# mkqdisk -L >>>>>> >>>>>> mkqdisk v0.5.1 >>>>>> >>>>>> /dev/sdd1: >>>>>> >>>>>> Magic: eb7a62c2 >>>>>> >>>>>> Label: csarcsysQ >>>>>> >>>>>> Created: Wed Feb 13 13:44:35 2008 >>>>>> >>>>>> Host: csarcsys1-eth0.xxx.xxx.nasa.gov >>>>>> >>>>>> [root@csarcsys3-eth0 cluster]# ls -l /dev/sdd1 >>>>>> >>>>>> brw-r----- 1 root disk 8, 49 Mar 25 14:09 /dev/sdd1 >>>>>> >>>>>> clustat from csarcsys1 >>>>>> >>>>>> msg_open: No such file or directory >>>>>> >>>>>> Member Status: Quorate >>>>>> >>>>>> Member Name ID Status >>>>>> >>>>>> ------ ---- ---- ------ >>>>>> >>>>>> csarcsys1-eth0 1 Online, Local >>>>>> >>>>>> csarcsys2-eth0 2 Online >>>>>> >>>>>> csarcsys3-eth0 3 Offline >>>>>> >>>>>> /dev/sdd1 0 Offline, Quorum Disk >>>>>> >>>>>> [root@csarcsys1-eth0 cluster]# ls -l /dev/sdd1 >>>>>> >>>>>> brw-r----- 1 root disk 8, 49 Mar 25 14:19 /dev/sdd1 >>>>>> >>>>>> mkqdisk v0.5.1 >>>>>> >>>>>> /dev/sdd1: >>>>>> >>>>>> Magic: eb7a62c2 >>>>>> >>>>>> Label: csarcsysQ >>>>>> >>>>>> Created: Wed Feb 13 13:44:35 2008 >>>>>> >>>>>> Host: csarcsys1-eth0.xxx.xxx.nasa.gov >>>>>> >>>>>> Info from csarcsys2 >>>>>> >>>>>> root@csarcsys2-eth0 cluster]# clustat >>>>>> >>>>>> msg_open: No such file or directory >>>>>> >>>>>> Member Status: Quorate >>>>>> >>>>>> Member Name ID Status >>>>>> >>>>>> ------ ---- ---- ------ >>>>>> >>>>>> csarcsys1-eth0 1 Offline >>>>>> >>>>>> csarcsys2-eth0 2 Online, Local >>>>>> >>>>>> csarcsys3-eth0 3 Offline >>>>>> >>>>>> /dev/sdd1 0 Online, Quorum Disk >>>>>> >>>>>> *From:* linux-cluster-bounces@xxxxxxxxxx >>>>>> [mailto:linux-cluster-bounces@xxxxxxxxxx] *On Behalf Of *Panigrahi, >>>>>> >> >> >>>>>> Santosh Kumar >>>>>> *Sent:* Tuesday, March 25, 2008 7:33 AM >>>>>> *To:* linux clustering >>>>>> *Subject:* RE: 3 node cluster problems >>>>>> >>>>>> If you are configuring your cluster by system-config-cluster then >>>>>> >> no >> >>>>>> >>> >>>>>> need to run ricci/luci. Ricci/luci needed for configuring the >>>>>> >>> cluster >>> >>>>>> >>>> >>>>>> using conga. You can configure in either ways. >>>>>> >>>>>> On seeing your clustat command outputs, it seems cluster is >>>>>> partitioned (spilt brain) into 2 sub clusters [Sub1-* >>>>>> **(csarcsys1-eth0, csarcsys2-eth0*) 2-* **csarcsys3-eth0*]. Without >>>>>> >>> a >>> >>>>>> >>>> >>>>>> quorum device you can more often face this situation. To avoid this >>>>>> >> >> >>>>>> you can configure a quorum device with a heuristic like ping >>>>>> >>> message. >>> >>>>>> >>>> >>>>>> Use the link >>>>>> >> (http://www.redhatmagazine.com/2007/12/19/enhancing-cluster-quorum-with- >> >>> >>>> >>>>> qdisk/) >>>>>> for configuring a quorum disk in RHCS. >>>>>> >>>>>> Thanks, >>>>>> >>>>>> S >>>>>> >>>>>> -----Original Message----- >>>>>> From: linux-cluster-bounces@xxxxxxxxxx >>>>>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Dalton, >>>>>> >>>> Maurice >>>> >>>>>> Sent: Tuesday, March 25, 2008 5:18 PM >>>>>> To: linux clustering >>>>>> Subject: RE: 3 node cluster problems >>>>>> >>>>>> Still no change. Same as below. >>>>>> >>>>>> I completely rebuilt the cluster using system-config-cluster >>>>>> >>>>>> The Cluster software was installed from rhn, luci and ricci are >>>>>> >>>>> running. >>>>> >>>>>> This is the new config file and it has been copied to the 2 other >>>>>> >>>>>> systems >>>>>> >>>>>> [root@csarcsys1-eth0 cluster]# more cluster.conf >>>>>> >>>>>> <?xml version="1.0"?> >>>>>> >>>>>> <cluster config_version="5" name="csarcsys5"> >>>>>> >>>>>> <fence_daemon post_fail_delay="0" post_join_delay="3"/> >>>>>> >>>>>> <clusternodes> >>>>>> >>>>>> <clusternode name="csarcsys1-eth0" nodeid="1" votes="1"> >>>>>> >>>>>> <fence/> >>>>>> >>>>>> </clusternode> >>>>>> >>>>>> <clusternode name="csarcsys2-eth0" nodeid="2" votes="1"> >>>>>> >>>>>> <fence/> >>>>>> >>>>>> </clusternode> >>>>>> >>>>>> <clusternode name="csarcsys3-eth0" nodeid="3" votes="1"> >>>>>> >>>>>> <fence/> >>>>>> >>>>>> </clusternode> >>>>>> >>>>>> </clusternodes> >>>>>> >>>>>> <cman/> >>>>>> >>>>>> <fencedevices/> >>>>>> >>>>>> <rm> >>>>>> >>>>>> <failoverdomains> >>>>>> >>>>>> <failoverdomain name="csarcsysfo" ordered="0" >>>>>> >>>>>> restricted="1"> >>>>>> >>>>>> <failoverdomainnode >>>>>> >>>>>> name="csarcsys1-eth0" priority="1"/> >>>>>> >>>>>> <failoverdomainnode >>>>>> >>>>>> name="csarcsys2-eth0" priority="1"/> >>>>>> >>>>>> <failoverdomainnode >>>>>> >>>>>> name="csarcsys3-eth0" priority="1"/> >>>>>> >>>>>> </failoverdomain> >>>>>> >>>>>> </failoverdomains> >>>>>> >>>>>> <resources> >>>>>> >>>>>> <ip address="172.xx.xx.xxx" monitor_link="1"/> >>>>>> >>>>>> <fs device="/dev/sdc1" force_fsck="0" >>>>>> >>>>>> force_unmount="1" fsid="57739" fstype="ext3" mountpo >>>>>> >>>>>> int="/csarc-test" name="csarcsys-fs" options="rw" self_fence="0"/> >>>>>> >>>>>> </resources> >>>>>> >>>>>> </rm> >>>>>> >>>>>> </cluster> >>>>>> >>>>>> -----Original Message----- >>>>>> >>>>>> From: linux-cluster-bounces@xxxxxxxxxx >>>>>> >>>>>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Bennie >>>>>> >> Thomas >> >>>>>> Sent: Monday, March 24, 2008 4:17 PM >>>>>> >>>>>> To: linux clustering >>>>>> >>>>>> Subject: Re: 3 node cluster problems >>>>>> >>>>>> Did you load the Cluster software via Conga or manually ? You would >>>>>> >>>>> have >>>>> >>>>>> had to load >>>>>> >>>>>> luci on one node and ricci on all three. >>>>>> >>>>>> Try copying the modified /etc/cluster/cluster.conf from csarcsys1 >>>>>> >> to >> >>>>>> >>>>> the >>>>> >>>>>> other two nodes. >>>>>> >>>>>> Make sure you can ping the private interface to/from all nodes and >>>>>> >>>>>> reboot. If this does not work >>>>>> >>>>>> post your /etc/cluster/cluster.conf file again. >>>>>> >>>>>> Dalton, Maurice wrote: >>>>>> >>>>>> >>>>>>> Yes >>>>>>> I also rebooted again just now to be sure. >>>>>>> -----Original Message----- >>>>>>> From: linux-cluster-bounces@xxxxxxxxxx >>>>>>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Bennie >>>>>>> >>> Thomas >>> >>>>>>> Sent: Monday, March 24, 2008 3:33 PM >>>>>>> To: linux clustering >>>>>>> Subject: Re: 3 node cluster problems >>>>>>> When you changed the nodenames in the >>>>>>> /etc/lcuster/cluster.conf >>>>>>> >> and >> >>>>>>> >>>>>> made >>>>>> >>>>>> >>>>>>> sure the /etc/hosts >>>>>>> file had the correct nodenames (Ie. 10.0.0.100 csarcsys1-eth0 >>>>>>> csarcsys1-eth0.xxxx.xxxx.xxx.) >>>>>>> Did you reboot all the nodes at the sametime ? >>>>>>> Dalton, Maurice wrote: >>>>>>> >>>>>>>> No luck. It seems as if csarcsys3 thinks its in his own cluster >>>>>>>> I renamed all config files and rebuilt from >>>>>>>> system-config-cluster >>>>>>>> Clustat command from csarcsys3 >>>>>>>> [root@csarcsys3-eth0 cluster]# clustat >>>>>>>> msg_open: No such file or directory >>>>>>>> Member Status: Inquorate >>>>>>>> Member Name ID Status >>>>>>>> ------ ---- ---- ------ >>>>>>>> csarcsys1-eth0 1 Offline >>>>>>>> csarcsys2-eth0 2 Offline >>>>>>>> csarcsys3-eth0 3 Online, Local >>>>>>>> clustat command from csarcsys2 >>>>>>>> [root@csarcsys2-eth0 cluster]# clustat >>>>>>>> msg_open: No such file or directory >>>>>>>> Member Status: Quorate >>>>>>>> Member Name ID Status >>>>>>>> ------ ---- ---- ------ >>>>>>>> csarcsys1-eth0 1 Online >>>>>>>> csarcsys2-eth0 2 Online, Local >>>>>>>> csarcsys3-eth0 3 Offline >>>>>>>> -----Original Message----- >>>>>>>> From: linux-cluster-bounces@xxxxxxxxxx >>>>>>>> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of >>>>>>>> Bennie >>>>>>>> >>>>> Thomas >>>>> >>>>>>>> Sent: Monday, March 24, 2008 2:25 PM >>>>>>>> To: linux clustering >>>>>>>> Subject: Re: 3 node cluster problems >>>>>>>> You will also, need to make sure the clustered >>>>>>>> nodenames are in >>>>>>>> >>>>> your >>>>> >>>>>>>> /etc/hosts file. >>>>>>>> Also, make sure your cluster network interface is up on >>>>>>>> all nodes >>>>>>>> >>>>> and >>>>> >>>>>>>> that the >>>>>>>> /etc/cluster/cluster.conf are the same on all nodes. >>>>>>>> Dalton, Maurice wrote: >>>>>>>> >>>>>>>>> The last post is incorrect. >>>>>>>>> Fence is still hanging at start up. >>>>>>>>> >> >> >>>>>>>>> Here's another log message. >>>>>>>>> Mar 24 19:03:14 csarcsys3-eth0 ccsd[6425]: Error while >>>>>>>>> >> processing >> >>>>>>>>> connect: Connection refused >>>>>>>>> Mar 24 19:03:15 csarcsys3-eth0 dlm_controld[6453]: >>>>>>>>> connect to >>>>>>>>> >> ccs >> >>>>>>>>> error -111, check ccsd or cluster status >>>>>>>>> *From:* linux-cluster-bounces@xxxxxxxxxx >>>>>>>>> [mailto:linux-cluster-bounces@xxxxxxxxxx] *On Behalf >>>>>>>>> Of *Bennie >>>>>>>>> >>>>>>> Thomas >>>>>>> >>>>>>>>> *Sent:* Monday, March 24, 2008 11:22 AM >>>>>>>>> *To:* linux clustering >>>>>>>>> *Subject:* Re: 3 node cluster problems >>>>>>>>> try removing the fully qualified hostname from the >>>>>>>>> cluster.conf >>>>>>>>> >>>>>> file. >>>>>> >>>>>> >>>>>>>>> Dalton, Maurice wrote: >>>>>>>>> I have NO fencing equipment >>>>>>>>> I have been task to setup a 3 node cluster >>>>>>>>> Currently I have having problems getting cman(fence) >>>>>>>>> to start >>>>>>>>> Fence will try to start up during cman start up but >>>>>>>>> will fail >>>>>>>>> I tried to run /sbin/fenced -D - I get the following >>>>>>>>> 1206373475 cman_init error 0 111 >>>>>>>>> Here's my cluster.conf file >>>>>>>>> <?xml version="1.0"?> >>>>>>>>> <cluster alias="csarcsys51" config_version="26" >>>>>>>>> >>> name="csarcsys51"> >>> >>>>>>>>> <fence_daemon clean_start="0" post_fail_delay="0" >>>>>>>>> >>>>>>>> post_join_delay="3"/> >>>>>>>> >>>>>>>>> <clusternodes> >>>>>>>>> <clusternode name="csarcsys1-eth0.xxx.xxxx.nasa.gov" >>>>>>>>> nodeid="1" >>>>>>>>> >>>>>>>> votes="1"> >>>>>>>> >>>>>>>>> <fence/> >>>>>>>>> </clusternode> >>>>>>>>> <clusternode name="csarcsys2-eth0.xxx.xxxx.nasa.gov" >>>>>>>>> nodeid="2" >>>>>>>>> >>>>>>>> votes="1"> >>>>>>>> >>>>>>>>> <fence/> >>>>>>>>> </clusternode> >>>>>>>>> <clusternode name="csarcsys3-eth0.xxx.xxxxnasa.gov" >>>>>>>>> nodeid="3" >>>>>>>>> >>>>>>>> votes="1"> >>>>>>>> >>>>>>>>> <fence/> >>>>>>>>> </clusternode> >>>>>>>>> </clusternodes> >>>>>>>>> <cman/> >>>>>>>>> <fencedevices/> >>>>>>>>> <rm> >>>>>>>>> <failoverdomains> >>>>>>>>> <failoverdomain name="csarcsys-fo" ordered="1" >>>>>>>>> restricted="0"> >>>>>>>>> <failoverdomainnode >>>>>>>>> name="csarcsys1-eth0.xxx.xxxx.nasa.gov" >>>>>>>>> >>>>>>>> priority="1"/> >>>>>>>> >>>>>>>>> <failoverdomainnode name="csarcsys2-eth0.xxx.xxxx.nasa.gov" >>>>>>>>> >>>>>>>> priority="1"/> >>>>>>>> >>>>>>>>> <failoverdomainnode name="csarcsys2-eth0.xxx.xxxx.nasa.gov" >>>>>>>>> >>>>>>>> priority="1"/> >>>>>>>> >>>>>>>>> </failoverdomain> >>>>>>>>> </failoverdomains> >>>>>>>>> <resources> >>>>>>>>> <ip address="xxx.xxx.xxx.xxx" monitor_link="1"/> >>>>>>>>> <fs device="/dev/sdc1" force_fsck="0" force_unmount="1" >>>>>>>>> >>>>> fsid="57739" >>>>> >>>>>>>>> fstype="ext3" mountpo >>>>>>>>> int="/csarc-test" name="csarcsys-fs" options="rw" >>>>>>>>> >>> self_fence="0"/> >>> >>>>>>>>> <nfsexport name="csarcsys-export"/> >>>>>>>>> <nfsclient name="csarcsys-nfs-client" >>>>>>>>> >> options="no_root_squash,rw" >> >>>>>>>>> path="/csarc-test" targe >>>>>>>>> t="xxx.xxx.xxx.*"/> >>>>>>>>> </resources> >>>>>>>>> </rm> >>>>>>>>> </cluster> >>>>>>>>> Messages from the logs >>>>>>>>> ar 24 13:24:19 csarcsys2-eth0 ccsd[24888]: Cluster >>>>>>>>> is not >>>>>>>>> >>> quorate. >>> >>>>>>>>> Refusing connection. >>>>>>>>> Mar 24 13:24:19 csarcsys2-eth0 ccsd[24888]: Error while >>>>>>>>> >>> processing >>> >>>>>>>>> connect: Connection refused >>>>>>>>> Mar 24 13:24:20 csarcsys2-eth0 ccsd[24888]: Cluster >>>>>>>>> is not >>>>>>>>> >>>>> quorate. >>>>> >>>>>>>>> Refusing connection. >>>>>>>>> Mar 24 13:24:20 csarcsys2-eth0 ccsd[24888]: Error while >>>>>>>>> >>> processing >>> >>>>>>>>> connect: Connection refused >>>>>>>>> Mar 24 13:24:21 csarcsys2-eth0 ccsd[24888]: Cluster >>>>>>>>> is not >>>>>>>>> >>>>> quorate. >>>>> >>>>>>>>> Refusing connection. >>>>>>>>> Mar 24 13:24:21 csarcsys2-eth0 ccsd[24888]: Error while >>>>>>>>> >>> processing >>> >>>>>>>>> connect: Connection refused >>>>>>>>> Mar 24 13:24:22 csarcsys2-eth0 ccsd[24888]: Cluster >>>>>>>>> is not >>>>>>>>> >>>>> quorate. >>>>> >>>>>>>>> Refusing connection. >>>>>>>>> Mar 24 13:24:22 csarcsys2-eth0 ccsd[24888]: Error while >>>>>>>>> >>> processing >>> >>>>>>>>> connect: Connection refused >>>>>>>>> Mar 24 13:24:23 csarcsys2-eth0 ccsd[24888]: Cluster >>>>>>>>> is not >>>>>>>>> >>>>> quorate. >>>>> >>>>>>>>> Refusing connection. >>>>>>>>> Mar 24 13:24:23 csarcsys2-eth0 ccsd[24888]: Error while >>>>>>>>> >>> processing >>> >>>>>>>>> connect: Connection refused >>>>>>>>> >>>>>> >> ------------------------------------------------------------------------ >> >>> >>>> >>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>>>> <mailto:Linux-cluster@xxxxxxxxxx> >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>> >> ------------------------------------------------------------------------ >> >>> >>>> >>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> -- >>>>>> >>>>>> Linux-cluster mailing list >>>>>> >>>>>> Linux-cluster@xxxxxxxxxx >>>>>> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>>> -- >>>>>> >>>>>> Linux-cluster mailing list >>>>>> >>>>>> Linux-cluster@xxxxxxxxxx >>>>>> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>>> >>>>>> >> ------------------------------------------------------------------------ >> >>> >>>> >>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@xxxxxxxxxx >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@xxxxxxxxxx >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@xxxxxxxxxx >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@xxxxxxxxxx >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@xxxxxxxxxx >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster@xxxxxxxxxx >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster@xxxxxxxxxx >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster@xxxxxxxxxx >> https://www.redhat.com/mailman/listinfo/linux-cluster >> >> -- >> Linux-cluster mailing list >> Linux-cluster@xxxxxxxxxx >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster