Change to 7 and i got this log Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down Cluster Service Manager... Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown complete, exiting Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster Service Manager is stopped. Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group Manager Starting Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service Data Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource Rules Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource Trees Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources defined Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover Domains Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing Services Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services Initialized Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port Opened Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: Local UP Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst2 UP Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst3 UP Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1) Processed Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1) Processed Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1) Processed Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events processed Anything unusual here? Paras. On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli <brem.belguebli@xxxxxxxxx> wrote: > I use log_level=7 to have more debugging info. > > It seems 4 is not enough. > > Brem > > > 2009/9/29, Paras pradhan <pradhanparas@xxxxxxxxx>: >> Withe log_level of 3 I got only this >> >> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down >> Cluster Service Manager... >> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down >> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, exiting >> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service >> Manager is stopped. >> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group >> Manager Starting >> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down >> Cluster Service Manager... >> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service >> Manager is stopped. >> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group >> Manager Starting >> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting down >> >> I do not know what the last line means. >> >> rgmanager version I am running is: >> rgmanager-2.0.52-1.el5.centos >> >> I don't what has gone wrong. >> >> Thanks >> Paras. >> >> >> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli >> <brem.belguebli@xxxxxxxxx> wrote: >> > you mean it stopped successfully on all the nodes but it is failing to >> > start only on node cvtst1 ? >> > >> > look at the following page to make rgmanager more verbose. It 'll >> > help debug.... >> > >> > http://sources.redhat.com/cluster/wiki/RGManager >> > >> > at Logging Configuration section >> > >> > >> > >> > >> > 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >> Brem, >> >> >> >> When I try to restart rgmanager on all the nodes, this time i do not >> >> see rgmanager running on the first node. But I do see on other 2 >> >> nodes. >> >> >> >> Log on the first node: >> >> >> >> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group >> >> Manager Starting >> >> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down >> >> Cluster Service Manager... >> >> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down >> >> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, exiting >> >> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service >> >> Manager is stopped. >> >> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group >> >> Manager Starting >> >> >> >> - >> >> It seems service is running , but I do not see rgmanger running using clustat >> >> >> >> >> >> Don't know what is going on. >> >> >> >> Thanks >> >> Paras. >> >> >> >> >> >> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli >> >> <brem.belguebli@xxxxxxxxx> wrote: >> >>> Paras, >> >>> >> >>> Another thing, it would have been more interesting to have a start >> >>> DEBUG not a stop. >> >>> >> >>> That's why I was asking you to first stop the vm manually on all your >> >>> nodes, stop eventually rgmanager on all the nodes to reset the >> >>> potential wrong states you may have, restart rgmanager. >> >>> >> >>> If your VM is configured to autostart, this will make it start. >> >>> >> >>> It should normally fail (as it does now). Send out your newly created >> >>> DEBUG file. >> >>> >> >>> 2009/9/29 brem belguebli <brem.belguebli@xxxxxxxxx>: >> >>>> Hi Paras, >> >>>> >> >>>> >> >>>> I don't know the xen/cluster combination well, but if I do remember >> >>>> well, I think I've read somewhere that when using xen you have to >> >>>> declare the use_virsh=0 key in the VM definition in the cluster.conf. >> >>>> >> >>>> This would make rgmanager use xm commands instead of virsh >> >>>> The DEBUG output shows clearly that you are using virsh to manage your >> >>>> VM instead of xm commands. >> >>>> Check out the RH docs about virtualization >> >>>> >> >>>> I'm not a 100% sure about that, I may be completely wrong. >> >>>> >> >>>> Brem >> >>>> >> >>>> 2009/9/28 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>> The only thing I noticed is the message after stopping the vm using xm >> >>>>> in all nodes and starting using clusvcadm is >> >>>>> >> >>>>> "Virtual machine guest1 is blocked" >> >>>>> >> >>>>> The whole DEBUG file is attached. >> >>>>> >> >>>>> >> >>>>> Thanks >> >>>>> Paras. >> >>>>> >> >>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli >> >>>>> <brem.belguebli@xxxxxxxxx> wrote: >> >>>>>> There's a problem with the script that is called by rgmanager to start >> >>>>>> the VM, I don't know what causes it >> >>>>>> >> >>>>>> May be you should try something like : >> >>>>>> >> >>>>>> 1) stop the VM on all nodes with xm commands >> >>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the following >> >>>>>> lines (after the #!/bin/bash ): >> >>>>>> exec >/tmp/DEBUG 2>&1 >> >>>>>> set -x >> >>>>>> 3) start the VM with clusvcadm -e vm:guest1 >> >>>>>> >> >>>>>> It should fail as it did before. >> >>>>>> >> >>>>>> edit the the /tmp/DEBUG file and you will be able to see where it >> >>>>>> fails (it may generate a lot of debug) >> >>>>>> >> >>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh >> >>>>>> >> >>>>>> Post the DEBUG file if you're not able to see where it fails. >> >>>>>> >> >>>>>> Brem >> >>>>>> >> >>>>>> 2009/9/26 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>>>> No I am not manually starting not using automatic init scripts. >> >>>>>>> >> >>>>>>> I started the vm using: clusvcadm -e vm:guest1 >> >>>>>>> >> >>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few seconds it >> >>>>>>> says guest1 started . But after a while I can see the guest1 on all >> >>>>>>> three nodes. >> >>>>>>> >> >>>>>>> clustat says: >> >>>>>>> >> >>>>>>> Service Name Owner (Last) >> >>>>>>> State >> >>>>>>> ------- ---- ----- ------ >> >>>>>>> ----- >> >>>>>>> vm:guest1 (none) >> >>>>>>> stopped >> >>>>>>> >> >>>>>>> But I can see the vm from xm li. >> >>>>>>> >> >>>>>>> This is what I can see from the log: >> >>>>>>> >> >>>>>>> >> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" >> >>>>>>> returned 1 (generic error) >> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start >> >>>>>>> vm:guest1; return value: 1 >> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 >> >>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is >> >>>>>>> recovering >> >>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed >> >>>>>>> service vm:guest1 >> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" >> >>>>>>> returned 1 (generic error) >> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start >> >>>>>>> vm:guest1; return value: 1 >> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 >> >>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is >> >>>>>>> recovering >> >>>>>>> >> >>>>>>> >> >>>>>>> Paras. >> >>>>>>> >> >>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli >> >>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >> >>>>>>>> Have you started your VM via rgmanager (clusvcadm -e vm:guest1) or >> >>>>>>>> using xm commands out of cluster control (or maybe a thru an >> >>>>>>>> automatic init script ?) >> >>>>>>>> >> >>>>>>>> When clustered, you should never be starting services (manually or >> >>>>>>>> thru automatic init script) out of cluster control >> >>>>>>>> >> >>>>>>>> The thing would be to stop your vm on all the nodes with the adequate >> >>>>>>>> xm command (not using xen myself) and try to start it with clusvcadm. >> >>>>>>>> >> >>>>>>>> Then see if it is started on all nodes (send clustat output) >> >>>>>>>> >> >>>>>>>> >> >>>>>>>> >> >>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>>>>>> Ok. Please see below. my vm is running on all nodes though clustat >> >>>>>>>>> says it is stopped. >> >>>>>>>>> >> >>>>>>>>> -- >> >>>>>>>>> [root@cvtst1 ~]# clustat >> >>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009 >> >>>>>>>>> Member Status: Quorate >> >>>>>>>>> >> >>>>>>>>> Member Name ID Status >> >>>>>>>>> ------ ---- ---- ------ >> >>>>>>>>> cvtst2 1 Online, rgmanager >> >>>>>>>>> cvtst1 2 Online, >> >>>>>>>>> Local, rgmanager >> >>>>>>>>> cvtst3 3 Online, rgmanager >> >>>>>>>>> >> >>>>>>>>> Service Name Owner (Last) >> >>>>>>>>> State >> >>>>>>>>> ------- ---- ----- ------ >> >>>>>>>>> ----- >> >>>>>>>>> vm:guest1 (none) >> >>>>>>>>> stopped >> >>>>>>>>> [root@cvtst1 ~]# >> >>>>>>>>> >> >>>>>>>>> >> >>>>>>>>> --- >> >>>>>>>>> o/p of xm li on cvtst1 >> >>>>>>>>> >> >>>>>>>>> -- >> >>>>>>>>> [root@cvtst1 ~]# xm li >> >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) >> >>>>>>>>> Domain-0 0 3470 2 r----- 28939.4 >> >>>>>>>>> guest1 7 511 1 -b---- 7727.8 >> >>>>>>>>> >> >>>>>>>>> o/p of xm li on cvtst2 >> >>>>>>>>> >> >>>>>>>>> -- >> >>>>>>>>> [root@cvtst2 ~]# xm li >> >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) >> >>>>>>>>> Domain-0 0 3470 2 r----- 31558.9 >> >>>>>>>>> guest1 21 511 1 -b---- 7558.2 >> >>>>>>>>> --- >> >>>>>>>>> >> >>>>>>>>> Thanks >> >>>>>>>>> Paras. >> >>>>>>>>> >> >>>>>>>>> >> >>>>>>>>> >> >>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli >> >>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >> >>>>>>>>>> It looks like no. >> >>>>>>>>>> >> >>>>>>>>>> can you send an output of clustat of when the VM is running on >> >>>>>>>>>> multiple nodes at the same time? >> >>>>>>>>>> >> >>>>>>>>>> And by the way, another one after having stopped (clusvcadm -s vm:guest1) ? >> >>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >> >>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not being >> >>>>>>>>>>> properly handled by the cluster. >> >>>>>>>>>>> >> >>>>>>>>>>> >> >>>>>>>>>>> Thanks >> >>>>>>>>>>> Paras. >> >>>>>>>>>>> >> >>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >> >>>>>>>>>>>> Ok.. here is my cluster.conf file >> >>>>>>>>>>>> >> >>>>>>>>>>>> -- >> >>>>>>>>>>>> [root@cvtst1 cluster]# more cluster.conf >> >>>>>>>>>>>> <?xml version="1.0"?> >> >>>>>>>>>>>> <cluster alias="test" config_version="9" name="test"> >> >>>>>>>>>>>> <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/> >> >>>>>>>>>>>> <clusternodes> >> >>>>>>>>>>>> <clusternode name="cvtst2" nodeid="1" votes="1"> >> >>>>>>>>>>>> <fence/> >> >>>>>>>>>>>> </clusternode> >> >>>>>>>>>>>> <clusternode name="cvtst1" nodeid="2" votes="1"> >> >>>>>>>>>>>> <fence/> >> >>>>>>>>>>>> </clusternode> >> >>>>>>>>>>>> <clusternode name="cvtst3" nodeid="3" votes="1"> >> >>>>>>>>>>>> <fence/> >> >>>>>>>>>>>> </clusternode> >> >>>>>>>>>>>> </clusternodes> >> >>>>>>>>>>>> <cman/> >> >>>>>>>>>>>> <fencedevices/> >> >>>>>>>>>>>> <rm> >> >>>>>>>>>>>> <failoverdomains> >> >>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="0" ordered="1" restricted="0"> >> >>>>>>>>>>>> <failoverdomainnode name="cvtst2" priority="3"/> >> >>>>>>>>>>>> <failoverdomainnode name="cvtst1" priority="1"/> >> >>>>>>>>>>>> <failoverdomainnode name="cvtst3" priority="2"/> >> >>>>>>>>>>>> </failoverdomain> >> >>>>>>>>>>>> </failoverdomains> >> >>>>>>>>>>>> <resources/> >> >>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0" >> >>>>>>>>>>>> name="guest1" path="/vms" recovery="r >> >>>>>>>>>>>> estart" restart_expire_time="0"/> >> >>>>>>>>>>>> </rm> >> >>>>>>>>>>>> </cluster> >> >>>>>>>>>>>> [root@cvtst1 cluster]# >> >>>>>>>>>>>> ------ >> >>>>>>>>>>>> >> >>>>>>>>>>>> Thanks! >> >>>>>>>>>>>> Paras. >> >>>>>>>>>>>> >> >>>>>>>>>>>> >> >>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer <volker@xxxxxxxxxxxx> wrote: >> >>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500, >> >>>>>>>>>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >> >>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. Now I am >> >>>>>>>>>>>>>> having another problem. When I start the my xen vm in one node, it >> >>>>>>>>>>>>>> also starts on other nodes. Which daemon controls this? >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the rgmanager >> >>>>>>>>>>>>> package). To me, this sounds like a configuration problem. Maybe, >> >>>>>>>>>>>>> you can post your cluster.conf? >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> Regards, >> >>>>>>>>>>>>> Volker >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> -- >> >>>>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>>>> >> >>>>>>>>>>>> >> >>>>>>>>>>> >> >>>>>>>>>>> -- >> >>>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>>>> -- >> >>>>>>>>>> Linux-cluster mailing list >> >>>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>>> >> >>>>>>>>> >> >>>>>>>>> -- >> >>>>>>>>> Linux-cluster mailing list >> >>>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>>> >> >>>>>>>> >> >>>>>>>> -- >> >>>>>>>> Linux-cluster mailing list >> >>>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>>> >> >>>>>>> >> >>>>>>> -- >> >>>>>>> Linux-cluster mailing list >> >>>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>>> >> >>>>>> >> >>>>>> -- >> >>>>>> Linux-cluster mailing list >> >>>>>> Linux-cluster@xxxxxxxxxx >> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>>> >> >>>>> >> >>>>> -- >> >>>>> Linux-cluster mailing list >> >>>>> Linux-cluster@xxxxxxxxxx >> >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>>>> >> >>>> >> >>> >> >>> -- >> >>> Linux-cluster mailing list >> >>> Linux-cluster@xxxxxxxxxx >> >>> https://www.redhat.com/mailman/listinfo/linux-cluster >> >>> >> >> >> >> -- >> >> Linux-cluster mailing list >> >> Linux-cluster@xxxxxxxxxx >> >> https://www.redhat.com/mailman/listinfo/linux-cluster >> >> >> > >> > -- >> > Linux-cluster mailing list >> > Linux-cluster@xxxxxxxxxx >> > https://www.redhat.com/mailman/listinfo/linux-cluster >> > >> >> -- >> Linux-cluster mailing list >> Linux-cluster@xxxxxxxxxx >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster