Yes this is very strange. I don't know what to do now. May be re create the cluster? But not a good solution actually. Packages : Kernel: kernel-xen-2.6.18-164.el5 OS: Full updated of CentOS 5.3 except CMAN downgraded to cman-2.0.98-1.el5 Other packages related to cluster suite: rgmanager-2.0.52-1.el5.centos cman-2.0.98-1.el5 xen-3.0.3-80.el5_3.3 xen-libs-3.0.3-80.el5_3.3 kmod-gfs-xen-0.1.31-3.el5_3.1 kmod-gfs-xen-0.1.31-3.el5_3.1 kmod-gfs-0.1.31-3.el5_3.1 gfs-utils-0.1.18-1.el5 gfs2-utils-0.1.62-1.el5 lvm2-2.02.40-6.el5 lvm2-cluster-2.02.40-7.el5 openais-0.80.3-22.el5_3.9 Thanks! Paras. On Wed, Sep 30, 2009 at 10:02 AM, brem belguebli <brem.belguebli@xxxxxxxxx> wrote: > Hi Paras, > > Your cluster.conf file seems correct. If it is not a ntp issue, I > don't see anything except a bug that causes this, or some prerequisite > that is not respected. > > May be you could post the versions (os, kernel, packages etc...) you > are using, someone may have hit the same issue with your versions. > > Brem > > 2009/9/30, Paras pradhan <pradhanparas@xxxxxxxxx>: >> All of the nodes are synced with ntp server. So this is not the case with me. >> >> Thanks >> Paras. >> >> On Tue, Sep 29, 2009 at 6:29 PM, Johannes Rußek >> <johannes.russek@xxxxxxxxxxxxxxxxx> wrote: >> > make sure the time on the nodes is in sync, apparently when a node has too >> > much offset, you won't see rgmanager (even though the process is running). >> > this happened today and setting the time fixed it for me. afaicr there was >> > no sign of this in the logs though. >> > johannes >> > >> > Paras pradhan schrieb: >> >> >> >> I don't see rgmanager . >> >> >> >> Here is the o/p from clustat >> >> >> >> [root@cvtst1 cluster]# clustat >> >> Cluster Status for test @ Tue Sep 29 15:53:33 2009 >> >> Member Status: Quorate >> >> >> >> Member Name ID >> >> Status >> >> ------ ---- ---- >> >> ------ >> >> cvtst2 1 Online >> >> cvtst1 2 Online, >> >> Local >> >> cvtst3 3 Online >> >> >> >> >> >> Thanks >> >> Paras. >> >> >> >> On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli >> >> <brem.belguebli@xxxxxxxxx> wrote: >> >> >> >>> >> >>> It looks correct, rgmanager seems to start on all nodes >> >>> >> >>> what gives you clustat ? >> >>> >> >>> If rgmanager doesn't show, check out the logs something may have gone >> >>> wrong. >> >>> >> >>> >> >>> 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>> >> >>>> >> >>>> Change to 7 and i got this log >> >>>> >> >>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down >> >>>> Cluster Service Manager... >> >>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down >> >>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down >> >>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown complete, >> >>>> exiting >> >>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster Service >> >>>> Manager is stopped. >> >>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group >> >>>> Manager Starting >> >>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service Data >> >>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource Rules >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource Trees >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources defined >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover >> >>>> Domains >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing Services >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services Initialized >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port Opened >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: Local UP >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst2 UP >> >>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst3 UP >> >>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1) Processed >> >>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1) Processed >> >>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1) Processed >> >>>> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events processed >> >>>> >> >>>> >> >>>> Anything unusual here? >> >>>> >> >>>> Paras. >> >>>> >> >>>> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli >> >>>> <brem.belguebli@xxxxxxxxx> wrote: >> >>>> >> >>>>> >> >>>>> I use log_level=7 to have more debugging info. >> >>>>> >> >>>>> It seems 4 is not enough. >> >>>>> >> >>>>> Brem >> >>>>> >> >>>>> >> >>>>> 2009/9/29, Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>> >> >>>>>> >> >>>>>> Withe log_level of 3 I got only this >> >>>>>> >> >>>>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down >> >>>>>> Cluster Service Manager... >> >>>>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down >> >>>>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, >> >>>>>> exiting >> >>>>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service >> >>>>>> Manager is stopped. >> >>>>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group >> >>>>>> Manager Starting >> >>>>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down >> >>>>>> Cluster Service Manager... >> >>>>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service >> >>>>>> Manager is stopped. >> >>>>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group >> >>>>>> Manager Starting >> >>>>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting >> >>>>>> down >> >>>>>> >> >>>>>> I do not know what the last line means. >> >>>>>> >> >>>>>> rgmanager version I am running is: >> >>>>>> rgmanager-2.0.52-1.el5.centos >> >>>>>> >> >>>>>> I don't what has gone wrong. >> >>>>>> >> >>>>>> Thanks >> >>>>>> Paras. >> >>>>>> >> >>>>>> >> >>>>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli >> >>>>>> <brem.belguebli@xxxxxxxxx> wrote: >> >>>>>> >> >>>>>>> >> >>>>>>> you mean it stopped successfully on all the nodes but it is failing >> >>>>>>> to >> >>>>>>> start only on node cvtst1 ? >> >>>>>>> >> >>>>>>> look at the following page to make rgmanager more verbose. It 'll >> >>>>>>> help debug.... >> >>>>>>> >> >>>>>>> http://sources.redhat.com/cluster/wiki/RGManager >> >>>>>>> >> >>>>>>> at Logging Configuration section >> >>>>>>> >> >>>>>>> >> >>>>>>> >> >>>>>>> >> >>>>>>> 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>>>> >> >>>>>>>> >> >>>>>>>> Brem, >> >>>>>>>> >> >>>>>>>> When I try to restart rgmanager on all the nodes, this time i do not >> >>>>>>>> see rgmanager running on the first node. But I do see on other 2 >> >>>>>>>> nodes. >> >>>>>>>> >> >>>>>>>> Log on the first node: >> >>>>>>>> >> >>>>>>>> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group >> >>>>>>>> Manager Starting >> >>>>>>>> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down >> >>>>>>>> Cluster Service Manager... >> >>>>>>>> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down >> >>>>>>>> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, >> >>>>>>>> exiting >> >>>>>>>> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service >> >>>>>>>> Manager is stopped. >> >>>>>>>> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group >> >>>>>>>> Manager Starting >> >>>>>>>> >> >>>>>>>> - >> >>>>>>>> It seems service is running , but I do not see rgmanger running >> >>>>>>>> using clustat >> >>>>>>>> >> >>>>>>>> >> >>>>>>>> Don't know what is going on. >> >>>>>>>> >> >>>>>>>> Thanks >> >>>>>>>> Paras. >> >>>>>>>> >> >>>>>>>> >> >>>>>>>> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli >> >>>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >> >>>>>>>> >> >>>>>>>>> >> >>>>>>>>> Paras, >> >>>>>>>>> >> >>>>>>>>> Another thing, it would have been more interesting to have a start >> >>>>>>>>> DEBUG not a stop. >> >>>>>>>>> >> >>>>>>>>> That's why I was asking you to first stop the vm manually on all >> >>>>>>>>> your >> >>>>>>>>> nodes, stop eventually rgmanager on all the nodes to reset the >> >>>>>>>>> potential wrong states you may have, restart rgmanager. >> >>>>>>>>> >> >>>>>>>>> If your VM is configured to autostart, this will make it start. >> >>>>>>>>> >> >>>>>>>>> It should normally fail (as it does now). Send out your newly >> >>>>>>>>> created >> >>>>>>>>> DEBUG file. >> >>>>>>>>> >> >>>>>>>>> 2009/9/29 brem belguebli <brem.belguebli@xxxxxxxxx>: >> >>>>>>>>> >> >>>>>>>>>> >> >>>>>>>>>> Hi Paras, >> >>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>>>> I don't know the xen/cluster combination well, but if I do >> >>>>>>>>>> remember >> >>>>>>>>>> well, I think I've read somewhere that when using xen you have to >> >>>>>>>>>> declare the use_virsh=0 key in the VM definition in the >> >>>>>>>>>> cluster.conf. >> >>>>>>>>>> >> >>>>>>>>>> This would make rgmanager use xm commands instead of virsh >> >>>>>>>>>> The DEBUG output shows clearly that you are using virsh to manage >> >>>>>>>>>> your >> >>>>>>>>>> VM instead of xm commands. >> >>>>>>>>>> Check out the RH docs about virtualization >> >>>>>>>>>> >> >>>>>>>>>> I'm not a 100% sure about that, I may be completely wrong. >> >>>>>>>>>> >> >>>>>>>>>> Brem >> >>>>>>>>>> >> >>>>>>>>>> 2009/9/28 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>>>>>>> >> >>>>>>>>>>> >> >>>>>>>>>>> The only thing I noticed is the message after stopping the vm >> >>>>>>>>>>> using xm >> >>>>>>>>>>> in all nodes and starting using clusvcadm is >> >>>>>>>>>>> >> >>>>>>>>>>> "Virtual machine guest1 is blocked" >> >>>>>>>>>>> >> >>>>>>>>>>> The whole DEBUG file is attached. >> >>>>>>>>>>> >> >>>>>>>>>>> >> >>>>>>>>>>> Thanks >> >>>>>>>>>>> Paras. >> >>>>>>>>>>> >> >>>>>>>>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli >> >>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >> >>>>>>>>>>> >> >>>>>>>>>>>> >> >>>>>>>>>>>> There's a problem with the script that is called by rgmanager to >> >>>>>>>>>>>> start >> >>>>>>>>>>>> the VM, I don't know what causes it >> >>>>>>>>>>>> >> >>>>>>>>>>>> May be you should try something like : >> >>>>>>>>>>>> >> >>>>>>>>>>>> 1) stop the VM on all nodes with xm commands >> >>>>>>>>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the >> >>>>>>>>>>>> following >> >>>>>>>>>>>> lines (after the #!/bin/bash ): >> >>>>>>>>>>>> exec >/tmp/DEBUG 2>&1 >> >>>>>>>>>>>> set -x >> >>>>>>>>>>>> 3) start the VM with clusvcadm -e vm:guest1 >> >>>>>>>>>>>> >> >>>>>>>>>>>> It should fail as it did before. >> >>>>>>>>>>>> >> >>>>>>>>>>>> edit the the /tmp/DEBUG file and you will be able to see where >> >>>>>>>>>>>> it >> >>>>>>>>>>>> fails (it may generate a lot of debug) >> >>>>>>>>>>>> >> >>>>>>>>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh >> >>>>>>>>>>>> >> >>>>>>>>>>>> Post the DEBUG file if you're not able to see where it fails. >> >>>>>>>>>>>> >> >>>>>>>>>>>> Brem >> >>>>>>>>>>>> >> >>>>>>>>>>>> 2009/9/26 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> No I am not manually starting not using automatic init scripts. >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> I started the vm using: clusvcadm -e vm:guest1 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few >> >>>>>>>>>>>>> seconds it >> >>>>>>>>>>>>> says guest1 started . But after a while I can see the guest1 on >> >>>>>>>>>>>>> all >> >>>>>>>>>>>>> three nodes. >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> clustat says: >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> Service Name Owner >> >>>>>>>>>>>>> (Last) >> >>>>>>>>>>>>> State >> >>>>>>>>>>>>> ------- ---- ----- >> >>>>>>>>>>>>> ------ >> >>>>>>>>>>>>> ----- >> >>>>>>>>>>>>> vm:guest1 (none) >> >>>>>>>>>>>>> stopped >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> But I can see the vm from xm li. >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> This is what I can see from the log: >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm >> >>>>>>>>>>>>> "guest1" >> >>>>>>>>>>>>> returned 1 (generic error) >> >>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed >> >>>>>>>>>>>>> to start >> >>>>>>>>>>>>> vm:guest1; return value: 1 >> >>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping >> >>>>>>>>>>>>> service vm:guest1 >> >>>>>>>>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service >> >>>>>>>>>>>>> vm:guest1 is >> >>>>>>>>>>>>> recovering >> >>>>>>>>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering >> >>>>>>>>>>>>> failed >> >>>>>>>>>>>>> service vm:guest1 >> >>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm >> >>>>>>>>>>>>> "guest1" >> >>>>>>>>>>>>> returned 1 (generic error) >> >>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed >> >>>>>>>>>>>>> to start >> >>>>>>>>>>>>> vm:guest1; return value: 1 >> >>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping >> >>>>>>>>>>>>> service vm:guest1 >> >>>>>>>>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service >> >>>>>>>>>>>>> vm:guest1 is >> >>>>>>>>>>>>> recovering >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> Paras. >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli >> >>>>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >> >>>>>>>>>>>>> >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> Have you started your VM via rgmanager (clusvcadm -e >> >>>>>>>>>>>>>> vm:guest1) or >> >>>>>>>>>>>>>> using xm commands out of cluster control (or maybe a thru an >> >>>>>>>>>>>>>> automatic init script ?) >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> When clustered, you should never be starting services >> >>>>>>>>>>>>>> (manually or >> >>>>>>>>>>>>>> thru automatic init script) out of cluster control >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> The thing would be to stop your vm on all the nodes with the >> >>>>>>>>>>>>>> adequate >> >>>>>>>>>>>>>> xm command (not using xen myself) and try to start it with >> >>>>>>>>>>>>>> clusvcadm. >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> Then see if it is started on all nodes (send clustat output) >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> Ok. Please see below. my vm is running on all nodes though >> >>>>>>>>>>>>>>> clustat >> >>>>>>>>>>>>>>> says it is stopped. >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>>> [root@cvtst1 ~]# clustat >> >>>>>>>>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009 >> >>>>>>>>>>>>>>> Member Status: Quorate >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> Member Name >> >>>>>>>>>>>>>>> ID Status >> >>>>>>>>>>>>>>> ------ ---- >> >>>>>>>>>>>>>>> ---- ------ >> >>>>>>>>>>>>>>> cvtst2 1 >> >>>>>>>>>>>>>>> Online, rgmanager >> >>>>>>>>>>>>>>> cvtst1 2 >> >>>>>>>>>>>>>>> Online, >> >>>>>>>>>>>>>>> Local, rgmanager >> >>>>>>>>>>>>>>> cvtst3 3 >> >>>>>>>>>>>>>>> Online, rgmanager >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> Service Name >> >>>>>>>>>>>>>>> Owner (Last) >> >>>>>>>>>>>>>>> State >> >>>>>>>>>>>>>>> ------- ---- >> >>>>>>>>>>>>>>> ----- ------ >> >>>>>>>>>>>>>>> ----- >> >>>>>>>>>>>>>>> vm:guest1 >> >>>>>>>>>>>>>>> (none) >> >>>>>>>>>>>>>>> stopped >> >>>>>>>>>>>>>>> [root@cvtst1 ~]# >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> --- >> >>>>>>>>>>>>>>> o/p of xm li on cvtst1 >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>>> [root@cvtst1 ~]# xm li >> >>>>>>>>>>>>>>> Name ID Mem(MiB) VCPUs >> >>>>>>>>>>>>>>> State Time(s) >> >>>>>>>>>>>>>>> Domain-0 0 3470 2 >> >>>>>>>>>>>>>>> r----- 28939.4 >> >>>>>>>>>>>>>>> guest1 7 511 1 >> >>>>>>>>>>>>>>> -b---- 7727.8 >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> o/p of xm li on cvtst2 >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>>> [root@cvtst2 ~]# xm li >> >>>>>>>>>>>>>>> Name ID Mem(MiB) VCPUs >> >>>>>>>>>>>>>>> State Time(s) >> >>>>>>>>>>>>>>> Domain-0 0 3470 2 >> >>>>>>>>>>>>>>> r----- 31558.9 >> >>>>>>>>>>>>>>> guest1 21 511 1 >> >>>>>>>>>>>>>>> -b---- 7558.2 >> >>>>>>>>>>>>>>> --- >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> Thanks >> >>>>>>>>>>>>>>> Paras. >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli >> >>>>>>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> It looks like no. >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> can you send an output of clustat of when the VM is running >> >>>>>>>>>>>>>>>> on >> >>>>>>>>>>>>>>>> multiple nodes at the same time? >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> And by the way, another one after having stopped (clusvcadm >> >>>>>>>>>>>>>>>> -s vm:guest1) ? >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not >> >>>>>>>>>>>>>>>>> being >> >>>>>>>>>>>>>>>>> properly handled by the cluster. >> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> Thanks >> >>>>>>>>>>>>>>>>> Paras. >> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan >> >>>>>>>>>>>>>>>>> <pradhanparas@xxxxxxxxx> wrote: >> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>> Ok.. here is my cluster.conf file >> >>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>>>>>> [root@cvtst1 cluster]# more cluster.conf >> >>>>>>>>>>>>>>>>>> <?xml version="1.0"?> >> >>>>>>>>>>>>>>>>>> <cluster alias="test" config_version="9" name="test"> >> >>>>>>>>>>>>>>>>>> <fence_daemon clean_start="0" post_fail_delay="0" >> >>>>>>>>>>>>>>>>>> post_join_delay="3"/> >> >>>>>>>>>>>>>>>>>> <clusternodes> >> >>>>>>>>>>>>>>>>>> <clusternode name="cvtst2" nodeid="1" >> >>>>>>>>>>>>>>>>>> votes="1"> >> >>>>>>>>>>>>>>>>>> <fence/> >> >>>>>>>>>>>>>>>>>> </clusternode> >> >>>>>>>>>>>>>>>>>> <clusternode name="cvtst1" nodeid="2" >> >>>>>>>>>>>>>>>>>> votes="1"> >> >>>>>>>>>>>>>>>>>> <fence/> >> >>>>>>>>>>>>>>>>>> </clusternode> >> >>>>>>>>>>>>>>>>>> <clusternode name="cvtst3" nodeid="3" >> >>>>>>>>>>>>>>>>>> votes="1"> >> >>>>>>>>>>>>>>>>>> <fence/> >> >>>>>>>>>>>>>>>>>> </clusternode> >> >>>>>>>>>>>>>>>>>> </clusternodes> >> >>>>>>>>>>>>>>>>>> <cman/> >> >>>>>>>>>>>>>>>>>> <fencedevices/> >> >>>>>>>>>>>>>>>>>> <rm> >> >>>>>>>>>>>>>>>>>> <failoverdomains> >> >>>>>>>>>>>>>>>>>> <failoverdomain name="myfd1" >> >>>>>>>>>>>>>>>>>> nofailback="0" ordered="1" restricted="0"> >> >>>>>>>>>>>>>>>>>> <failoverdomainnode >> >>>>>>>>>>>>>>>>>> name="cvtst2" priority="3"/> >> >>>>>>>>>>>>>>>>>> <failoverdomainnode >> >>>>>>>>>>>>>>>>>> name="cvtst1" priority="1"/> >> >>>>>>>>>>>>>>>>>> <failoverdomainnode >> >>>>>>>>>>>>>>>>>> name="cvtst3" priority="2"/> >> >>>>>>>>>>>>>>>>>> </failoverdomain> >> >>>>>>>>>>>>>>>>>> </failoverdomains> >> >>>>>>>>>>>>>>>>>> <resources/> >> >>>>>>>>>>>>>>>>>> <vm autostart="1" domain="myfd1" >> >>>>>>>>>>>>>>>>>> exclusive="0" max_restarts="0" >> >>>>>>>>>>>>>>>>>> name="guest1" path="/vms" recovery="r >> >>>>>>>>>>>>>>>>>> estart" restart_expire_time="0"/> >> >>>>>>>>>>>>>>>>>> </rm> >> >>>>>>>>>>>>>>>>>> </cluster> >> >>>>>>>>>>>>>>>>>> [root@cvtst1 cluster]# >> >>>>>>>>>>>>>>>>>> ------ >> >>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>> Thanks! >> >>>>>>>>>>>>>>>>>> Paras. >> >>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer >> >>>>>>>>>>>>>>>>>> <volker@xxxxxxxxxxxx> wrote: >> >>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500, >> >>>>>>>>>>>>>>>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >> >>>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. >> >>>>>>>>>>>>>>>>>>>> Now I am >> >>>>>>>>>>>>>>>>>>>> having another problem. When I start the my xen vm in >> >>>>>>>>>>>>>>>>>>>> one node, it >> >>>>>>>>>>>>>>>>>>>> also starts on other nodes. Which daemon controls this? >> >>>>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the >> >>>>>>>>>>>>>>>>>>> rgmanager >> >>>>>>>>>>>>>>>>>>> package). To me, this sounds like a configuration >> >>>>>>>>>>>>>>>>>>> problem. Maybe, >> >>>>>>>>>>>>>>>>>>> you can post your cluster.conf? >> >>>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>>> Regards, >> >>>>>>>>>>>>>>>>>>> Volker >> >>>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> -- >> >>>>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>> >> >>>>>>>>>>>> -- >> >>>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>>> >> >>>>>>>>>>>> >> >>>>>>>>>>> >> >>>>>>>>>>> -- >> >>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>> >> >>>>>>>>>>> >> >>>>>>>>> >> >>>>>>>>> -- >> >>>>>>>>> Linux-cluster mailing list >> >>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>> >> >>>>>>>>> >> >>>>>>>> >> >>>>>>>> -- >> >>>>>>>> Linux-cluster mailing list >> >>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>> >> >>>>>>>> >> >>>>>>> >> >>>>>>> -- >> >>>>>>> Linux-cluster mailing list >> >>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>> >> >>>>>>> >> >>>>>> >> >>>>>> -- >> >>>>>> Linux-cluster mailing list >> >>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>> >> >>>>>> >> >>>>> >> >>>>> -- >> >>>>> Linux-cluster mailing list >> >>>>> Linux-cluster@xxxxxxxxxx >> >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>> >> >>>>> >> >>>> >> >>>> -- >> >>>> Linux-cluster mailing list >> >>>> Linux-cluster@xxxxxxxxxx >> >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>> >> >>>> >> >>> >> >>> -- >> >>> Linux-cluster mailing list >> >>> Linux-cluster@xxxxxxxxxx >> >>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>> >> >>> >> >> >> >> -- >> >> Linux-cluster mailing list >> >> Linux-cluster@xxxxxxxxxx >> >> https://www.redhat.com/mailman/listinfo/linux-cluster >> >> >> > >> > -- >> > Linux-cluster mailing list >> > Linux-cluster@xxxxxxxxxx >> > https://www.redhat.com/mailman/listinfo/linux-cluster >> > >> >> -- >> Linux-cluster mailing list >> Linux-cluster@xxxxxxxxxx >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster