Withe log_level of 3 I got only this Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down Cluster Service Manager... Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, exiting Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service Manager is stopped. Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group Manager Starting Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down Cluster Service Manager... Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service Manager is stopped. Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group Manager Starting Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting down I do not know what the last line means. rgmanager version I am running is: rgmanager-2.0.52-1.el5.centos I don't what has gone wrong. Thanks Paras. On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli <brem.belguebli@xxxxxxxxx> wrote: > you mean it stopped successfully on all the nodes but it is failing to > start only on node cvtst1 ? > > look at the following page to make rgmanager more verbose. It 'll > help debug.... > > http://sources.redhat.com/cluster/wiki/RGManager > > at Logging Configuration section > > > > > 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: >> Brem, >> >> When I try to restart rgmanager on all the nodes, this time i do not >> see rgmanager running on the first node. But I do see on other 2 >> nodes. >> >> Log on the first node: >> >> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group >> Manager Starting >> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down >> Cluster Service Manager... >> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down >> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, exiting >> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service >> Manager is stopped. >> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group >> Manager Starting >> >> - >> It seems service is running , but I do not see rgmanger running using clustat >> >> >> Don't know what is going on. >> >> Thanks >> Paras. >> >> >> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli >> <brem.belguebli@xxxxxxxxx> wrote: >>> Paras, >>> >>> Another thing, it would have been more interesting to have a start >>> DEBUG not a stop. >>> >>> That's why I was asking you to first stop the vm manually on all your >>> nodes, stop eventually rgmanager on all the nodes to reset the >>> potential wrong states you may have, restart rgmanager. >>> >>> If your VM is configured to autostart, this will make it start. >>> >>> It should normally fail (as it does now). Send out your newly created >>> DEBUG file. >>> >>> 2009/9/29 brem belguebli <brem.belguebli@xxxxxxxxx>: >>>> Hi Paras, >>>> >>>> >>>> I don't know the xen/cluster combination well, but if I do remember >>>> well, I think I've read somewhere that when using xen you have to >>>> declare the use_virsh=0 key in the VM definition in the cluster.conf. >>>> >>>> This would make rgmanager use xm commands instead of virsh >>>> The DEBUG output shows clearly that you are using virsh to manage your >>>> VM instead of xm commands. >>>> Check out the RH docs about virtualization >>>> >>>> I'm not a 100% sure about that, I may be completely wrong. >>>> >>>> Brem >>>> >>>> 2009/9/28 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>> The only thing I noticed is the message after stopping the vm using xm >>>>> in all nodes and starting using clusvcadm is >>>>> >>>>> "Virtual machine guest1 is blocked" >>>>> >>>>> The whole DEBUG file is attached. >>>>> >>>>> >>>>> Thanks >>>>> Paras. >>>>> >>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli >>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>>> There's a problem with the script that is called by rgmanager to start >>>>>> the VM, I don't know what causes it >>>>>> >>>>>> May be you should try something like : >>>>>> >>>>>> 1) stop the VM on all nodes with xm commands >>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the following >>>>>> lines (after the #!/bin/bash ): >>>>>> exec >/tmp/DEBUG 2>&1 >>>>>> set -x >>>>>> 3) start the VM with clusvcadm -e vm:guest1 >>>>>> >>>>>> It should fail as it did before. >>>>>> >>>>>> edit the the /tmp/DEBUG file and you will be able to see where it >>>>>> fails (it may generate a lot of debug) >>>>>> >>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh >>>>>> >>>>>> Post the DEBUG file if you're not able to see where it fails. >>>>>> >>>>>> Brem >>>>>> >>>>>> 2009/9/26 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>>>> No I am not manually starting not using automatic init scripts. >>>>>>> >>>>>>> I started the vm using: clusvcadm -e vm:guest1 >>>>>>> >>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few seconds it >>>>>>> says guest1 started . But after a while I can see the guest1 on all >>>>>>> three nodes. >>>>>>> >>>>>>> clustat says: >>>>>>> >>>>>>> Service Name Owner (Last) >>>>>>> State >>>>>>> ------- ---- ----- ------ >>>>>>> ----- >>>>>>> vm:guest1 (none) >>>>>>> stopped >>>>>>> >>>>>>> But I can see the vm from xm li. >>>>>>> >>>>>>> This is what I can see from the log: >>>>>>> >>>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" >>>>>>> returned 1 (generic error) >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start >>>>>>> vm:guest1; return value: 1 >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 >>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is >>>>>>> recovering >>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed >>>>>>> service vm:guest1 >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" >>>>>>> returned 1 (generic error) >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start >>>>>>> vm:guest1; return value: 1 >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 >>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is >>>>>>> recovering >>>>>>> >>>>>>> >>>>>>> Paras. >>>>>>> >>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli >>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>>>>> Have you started your VM via rgmanager (clusvcadm -e vm:guest1) or >>>>>>>> using xm commands out of cluster control (or maybe a thru an >>>>>>>> automatic init script ?) >>>>>>>> >>>>>>>> When clustered, you should never be starting services (manually or >>>>>>>> thru automatic init script) out of cluster control >>>>>>>> >>>>>>>> The thing would be to stop your vm on all the nodes with the adequate >>>>>>>> xm command (not using xen myself) and try to start it with clusvcadm. >>>>>>>> >>>>>>>> Then see if it is started on all nodes (send clustat output) >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>>>>>> Ok. Please see below. my vm is running on all nodes though clustat >>>>>>>>> says it is stopped. >>>>>>>>> >>>>>>>>> -- >>>>>>>>> [root@cvtst1 ~]# clustat >>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009 >>>>>>>>> Member Status: Quorate >>>>>>>>> >>>>>>>>> Member Name ID Status >>>>>>>>> ------ ---- ---- ------ >>>>>>>>> cvtst2 1 Online, rgmanager >>>>>>>>> cvtst1 2 Online, >>>>>>>>> Local, rgmanager >>>>>>>>> cvtst3 3 Online, rgmanager >>>>>>>>> >>>>>>>>> Service Name Owner (Last) >>>>>>>>> State >>>>>>>>> ------- ---- ----- ------ >>>>>>>>> ----- >>>>>>>>> vm:guest1 (none) >>>>>>>>> stopped >>>>>>>>> [root@cvtst1 ~]# >>>>>>>>> >>>>>>>>> >>>>>>>>> --- >>>>>>>>> o/p of xm li on cvtst1 >>>>>>>>> >>>>>>>>> -- >>>>>>>>> [root@cvtst1 ~]# xm li >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) >>>>>>>>> Domain-0 0 3470 2 r----- 28939.4 >>>>>>>>> guest1 7 511 1 -b---- 7727.8 >>>>>>>>> >>>>>>>>> o/p of xm li on cvtst2 >>>>>>>>> >>>>>>>>> -- >>>>>>>>> [root@cvtst2 ~]# xm li >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) >>>>>>>>> Domain-0 0 3470 2 r----- 31558.9 >>>>>>>>> guest1 21 511 1 -b---- 7558.2 >>>>>>>>> --- >>>>>>>>> >>>>>>>>> Thanks >>>>>>>>> Paras. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli >>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>>>>>>> It looks like no. >>>>>>>>>> >>>>>>>>>> can you send an output of clustat of when the VM is running on >>>>>>>>>> multiple nodes at the same time? >>>>>>>>>> >>>>>>>>>> And by the way, another one after having stopped (clusvcadm -s vm:guest1) ? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not being >>>>>>>>>>> properly handled by the cluster. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Thanks >>>>>>>>>>> Paras. >>>>>>>>>>> >>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >>>>>>>>>>>> Ok.. here is my cluster.conf file >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> [root@cvtst1 cluster]# more cluster.conf >>>>>>>>>>>> <?xml version="1.0"?> >>>>>>>>>>>> <cluster alias="test" config_version="9" name="test"> >>>>>>>>>>>> <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/> >>>>>>>>>>>> <clusternodes> >>>>>>>>>>>> <clusternode name="cvtst2" nodeid="1" votes="1"> >>>>>>>>>>>> <fence/> >>>>>>>>>>>> </clusternode> >>>>>>>>>>>> <clusternode name="cvtst1" nodeid="2" votes="1"> >>>>>>>>>>>> <fence/> >>>>>>>>>>>> </clusternode> >>>>>>>>>>>> <clusternode name="cvtst3" nodeid="3" votes="1"> >>>>>>>>>>>> <fence/> >>>>>>>>>>>> </clusternode> >>>>>>>>>>>> </clusternodes> >>>>>>>>>>>> <cman/> >>>>>>>>>>>> <fencedevices/> >>>>>>>>>>>> <rm> >>>>>>>>>>>> <failoverdomains> >>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="0" ordered="1" restricted="0"> >>>>>>>>>>>> <failoverdomainnode name="cvtst2" priority="3"/> >>>>>>>>>>>> <failoverdomainnode name="cvtst1" priority="1"/> >>>>>>>>>>>> <failoverdomainnode name="cvtst3" priority="2"/> >>>>>>>>>>>> </failoverdomain> >>>>>>>>>>>> </failoverdomains> >>>>>>>>>>>> <resources/> >>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0" >>>>>>>>>>>> name="guest1" path="/vms" recovery="r >>>>>>>>>>>> estart" restart_expire_time="0"/> >>>>>>>>>>>> </rm> >>>>>>>>>>>> </cluster> >>>>>>>>>>>> [root@cvtst1 cluster]# >>>>>>>>>>>> ------ >>>>>>>>>>>> >>>>>>>>>>>> Thanks! >>>>>>>>>>>> Paras. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer <volker@xxxxxxxxxxxx> wrote: >>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500, >>>>>>>>>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. Now I am >>>>>>>>>>>>>> having another problem. When I start the my xen vm in one node, it >>>>>>>>>>>>>> also starts on other nodes. Which daemon controls this? >>>>>>>>>>>>> >>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the rgmanager >>>>>>>>>>>>> package). To me, this sounds like a configuration problem. Maybe, >>>>>>>>>>>>> you can post your cluster.conf? >>>>>>>>>>>>> >>>>>>>>>>>>> Regards, >>>>>>>>>>>>> Volker >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@xxxxxxxxxx >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@xxxxxxxxxx >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster@xxxxxxxxxx >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> -- >> Linux-cluster mailing list >> Linux-cluster@xxxxxxxxxx >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster