I use log_level=7 to have more debugging info. It seems 4 is not enough. Brem 2009/9/29, Paras pradhan <pradhanparas@xxxxxxxxx>: > Withe log_level of 3 I got only this > > Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down > Cluster Service Manager... > Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down > Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, exiting > Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service > Manager is stopped. > Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group > Manager Starting > Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down > Cluster Service Manager... > Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service > Manager is stopped. > Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group > Manager Starting > Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting down > > I do not know what the last line means. > > rgmanager version I am running is: > rgmanager-2.0.52-1.el5.centos > > I don't what has gone wrong. > > Thanks > Paras. > > > On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli > <brem.belguebli@xxxxxxxxx> wrote: > > you mean it stopped successfully on all the nodes but it is failing to > > start only on node cvtst1 ? > > > > look at the following page to make rgmanager more verbose. It 'll > > help debug.... > > > > http://sources.redhat.com/cluster/wiki/RGManager > > > > at Logging Configuration section > > > > > > > > > > 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: > >> Brem, > >> > >> When I try to restart rgmanager on all the nodes, this time i do not > >> see rgmanager running on the first node. But I do see on other 2 > >> nodes. > >> > >> Log on the first node: > >> > >> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group > >> Manager Starting > >> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down > >> Cluster Service Manager... > >> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down > >> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, exiting > >> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service > >> Manager is stopped. > >> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group > >> Manager Starting > >> > >> - > >> It seems service is running , but I do not see rgmanger running using clustat > >> > >> > >> Don't know what is going on. > >> > >> Thanks > >> Paras. > >> > >> > >> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli > >> <brem.belguebli@xxxxxxxxx> wrote: > >>> Paras, > >>> > >>> Another thing, it would have been more interesting to have a start > >>> DEBUG not a stop. > >>> > >>> That's why I was asking you to first stop the vm manually on all your > >>> nodes, stop eventually rgmanager on all the nodes to reset the > >>> potential wrong states you may have, restart rgmanager. > >>> > >>> If your VM is configured to autostart, this will make it start. > >>> > >>> It should normally fail (as it does now). Send out your newly created > >>> DEBUG file. > >>> > >>> 2009/9/29 brem belguebli <brem.belguebli@xxxxxxxxx>: > >>>> Hi Paras, > >>>> > >>>> > >>>> I don't know the xen/cluster combination well, but if I do remember > >>>> well, I think I've read somewhere that when using xen you have to > >>>> declare the use_virsh=0 key in the VM definition in the cluster.conf. > >>>> > >>>> This would make rgmanager use xm commands instead of virsh > >>>> The DEBUG output shows clearly that you are using virsh to manage your > >>>> VM instead of xm commands. > >>>> Check out the RH docs about virtualization > >>>> > >>>> I'm not a 100% sure about that, I may be completely wrong. > >>>> > >>>> Brem > >>>> > >>>> 2009/9/28 Paras pradhan <pradhanparas@xxxxxxxxx>: > >>>>> The only thing I noticed is the message after stopping the vm using xm > >>>>> in all nodes and starting using clusvcadm is > >>>>> > >>>>> "Virtual machine guest1 is blocked" > >>>>> > >>>>> The whole DEBUG file is attached. > >>>>> > >>>>> > >>>>> Thanks > >>>>> Paras. > >>>>> > >>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli > >>>>> <brem.belguebli@xxxxxxxxx> wrote: > >>>>>> There's a problem with the script that is called by rgmanager to start > >>>>>> the VM, I don't know what causes it > >>>>>> > >>>>>> May be you should try something like : > >>>>>> > >>>>>> 1) stop the VM on all nodes with xm commands > >>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the following > >>>>>> lines (after the #!/bin/bash ): > >>>>>> exec >/tmp/DEBUG 2>&1 > >>>>>> set -x > >>>>>> 3) start the VM with clusvcadm -e vm:guest1 > >>>>>> > >>>>>> It should fail as it did before. > >>>>>> > >>>>>> edit the the /tmp/DEBUG file and you will be able to see where it > >>>>>> fails (it may generate a lot of debug) > >>>>>> > >>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh > >>>>>> > >>>>>> Post the DEBUG file if you're not able to see where it fails. > >>>>>> > >>>>>> Brem > >>>>>> > >>>>>> 2009/9/26 Paras pradhan <pradhanparas@xxxxxxxxx>: > >>>>>>> No I am not manually starting not using automatic init scripts. > >>>>>>> > >>>>>>> I started the vm using: clusvcadm -e vm:guest1 > >>>>>>> > >>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few seconds it > >>>>>>> says guest1 started . But after a while I can see the guest1 on all > >>>>>>> three nodes. > >>>>>>> > >>>>>>> clustat says: > >>>>>>> > >>>>>>> Service Name Owner (Last) > >>>>>>> State > >>>>>>> ------- ---- ----- ------ > >>>>>>> ----- > >>>>>>> vm:guest1 (none) > >>>>>>> stopped > >>>>>>> > >>>>>>> But I can see the vm from xm li. > >>>>>>> > >>>>>>> This is what I can see from the log: > >>>>>>> > >>>>>>> > >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" > >>>>>>> returned 1 (generic error) > >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start > >>>>>>> vm:guest1; return value: 1 > >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 > >>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is > >>>>>>> recovering > >>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed > >>>>>>> service vm:guest1 > >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" > >>>>>>> returned 1 (generic error) > >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start > >>>>>>> vm:guest1; return value: 1 > >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 > >>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is > >>>>>>> recovering > >>>>>>> > >>>>>>> > >>>>>>> Paras. > >>>>>>> > >>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli > >>>>>>> <brem.belguebli@xxxxxxxxx> wrote: > >>>>>>>> Have you started your VM via rgmanager (clusvcadm -e vm:guest1) or > >>>>>>>> using xm commands out of cluster control (or maybe a thru an > >>>>>>>> automatic init script ?) > >>>>>>>> > >>>>>>>> When clustered, you should never be starting services (manually or > >>>>>>>> thru automatic init script) out of cluster control > >>>>>>>> > >>>>>>>> The thing would be to stop your vm on all the nodes with the adequate > >>>>>>>> xm command (not using xen myself) and try to start it with clusvcadm. > >>>>>>>> > >>>>>>>> Then see if it is started on all nodes (send clustat output) > >>>>>>>> > >>>>>>>> > >>>>>>>> > >>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: > >>>>>>>>> Ok. Please see below. my vm is running on all nodes though clustat > >>>>>>>>> says it is stopped. > >>>>>>>>> > >>>>>>>>> -- > >>>>>>>>> [root@cvtst1 ~]# clustat > >>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009 > >>>>>>>>> Member Status: Quorate > >>>>>>>>> > >>>>>>>>> Member Name ID Status > >>>>>>>>> ------ ---- ---- ------ > >>>>>>>>> cvtst2 1 Online, rgmanager > >>>>>>>>> cvtst1 2 Online, > >>>>>>>>> Local, rgmanager > >>>>>>>>> cvtst3 3 Online, rgmanager > >>>>>>>>> > >>>>>>>>> Service Name Owner (Last) > >>>>>>>>> State > >>>>>>>>> ------- ---- ----- ------ > >>>>>>>>> ----- > >>>>>>>>> vm:guest1 (none) > >>>>>>>>> stopped > >>>>>>>>> [root@cvtst1 ~]# > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> --- > >>>>>>>>> o/p of xm li on cvtst1 > >>>>>>>>> > >>>>>>>>> -- > >>>>>>>>> [root@cvtst1 ~]# xm li > >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) > >>>>>>>>> Domain-0 0 3470 2 r----- 28939.4 > >>>>>>>>> guest1 7 511 1 -b---- 7727.8 > >>>>>>>>> > >>>>>>>>> o/p of xm li on cvtst2 > >>>>>>>>> > >>>>>>>>> -- > >>>>>>>>> [root@cvtst2 ~]# xm li > >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) > >>>>>>>>> Domain-0 0 3470 2 r----- 31558.9 > >>>>>>>>> guest1 21 511 1 -b---- 7558.2 > >>>>>>>>> --- > >>>>>>>>> > >>>>>>>>> Thanks > >>>>>>>>> Paras. > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli > >>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote: > >>>>>>>>>> It looks like no. > >>>>>>>>>> > >>>>>>>>>> can you send an output of clustat of when the VM is running on > >>>>>>>>>> multiple nodes at the same time? > >>>>>>>>>> > >>>>>>>>>> And by the way, another one after having stopped (clusvcadm -s vm:guest1) ? > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: > >>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not being > >>>>>>>>>>> properly handled by the cluster. > >>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>>> Thanks > >>>>>>>>>>> Paras. > >>>>>>>>>>> > >>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan <pradhanparas@xxxxxxxxx> wrote: > >>>>>>>>>>>> Ok.. here is my cluster.conf file > >>>>>>>>>>>> > >>>>>>>>>>>> -- > >>>>>>>>>>>> [root@cvtst1 cluster]# more cluster.conf > >>>>>>>>>>>> <?xml version="1.0"?> > >>>>>>>>>>>> <cluster alias="test" config_version="9" name="test"> > >>>>>>>>>>>> <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/> > >>>>>>>>>>>> <clusternodes> > >>>>>>>>>>>> <clusternode name="cvtst2" nodeid="1" votes="1"> > >>>>>>>>>>>> <fence/> > >>>>>>>>>>>> </clusternode> > >>>>>>>>>>>> <clusternode name="cvtst1" nodeid="2" votes="1"> > >>>>>>>>>>>> <fence/> > >>>>>>>>>>>> </clusternode> > >>>>>>>>>>>> <clusternode name="cvtst3" nodeid="3" votes="1"> > >>>>>>>>>>>> <fence/> > >>>>>>>>>>>> </clusternode> > >>>>>>>>>>>> </clusternodes> > >>>>>>>>>>>> <cman/> > >>>>>>>>>>>> <fencedevices/> > >>>>>>>>>>>> <rm> > >>>>>>>>>>>> <failoverdomains> > >>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="0" ordered="1" restricted="0"> > >>>>>>>>>>>> <failoverdomainnode name="cvtst2" priority="3"/> > >>>>>>>>>>>> <failoverdomainnode name="cvtst1" priority="1"/> > >>>>>>>>>>>> <failoverdomainnode name="cvtst3" priority="2"/> > >>>>>>>>>>>> </failoverdomain> > >>>>>>>>>>>> </failoverdomains> > >>>>>>>>>>>> <resources/> > >>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0" > >>>>>>>>>>>> name="guest1" path="/vms" recovery="r > >>>>>>>>>>>> estart" restart_expire_time="0"/> > >>>>>>>>>>>> </rm> > >>>>>>>>>>>> </cluster> > >>>>>>>>>>>> [root@cvtst1 cluster]# > >>>>>>>>>>>> ------ > >>>>>>>>>>>> > >>>>>>>>>>>> Thanks! > >>>>>>>>>>>> Paras. > >>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer <volker@xxxxxxxxxxxx> wrote: > >>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500, > >>>>>>>>>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote: > >>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. Now I am > >>>>>>>>>>>>>> having another problem. When I start the my xen vm in one node, it > >>>>>>>>>>>>>> also starts on other nodes. Which daemon controls this? > >>>>>>>>>>>>> > >>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the rgmanager > >>>>>>>>>>>>> package). To me, this sounds like a configuration problem. Maybe, > >>>>>>>>>>>>> you can post your cluster.conf? > >>>>>>>>>>>>> > >>>>>>>>>>>>> Regards, > >>>>>>>>>>>>> Volker > >>>>>>>>>>>>> > >>>>>>>>>>>>> -- > >>>>>>>>>>>>> Linux-cluster mailing list > >>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx > >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>>> -- > >>>>>>>>>>> Linux-cluster mailing list > >>>>>>>>>>> Linux-cluster@xxxxxxxxxx > >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> -- > >>>>>>>>>> Linux-cluster mailing list > >>>>>>>>>> Linux-cluster@xxxxxxxxxx > >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>>>>>>>>> > >>>>>>>>> > >>>>>>>>> -- > >>>>>>>>> Linux-cluster mailing list > >>>>>>>>> Linux-cluster@xxxxxxxxxx > >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>>>>>>>> > >>>>>>>> > >>>>>>>> -- > >>>>>>>> Linux-cluster mailing list > >>>>>>>> Linux-cluster@xxxxxxxxxx > >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>>>>>>> > >>>>>>> > >>>>>>> -- > >>>>>>> Linux-cluster mailing list > >>>>>>> Linux-cluster@xxxxxxxxxx > >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>>>>>> > >>>>>> > >>>>>> -- > >>>>>> Linux-cluster mailing list > >>>>>> Linux-cluster@xxxxxxxxxx > >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>>>>> > >>>>> > >>>>> -- > >>>>> Linux-cluster mailing list > >>>>> Linux-cluster@xxxxxxxxxx > >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>>>> > >>>> > >>> > >>> -- > >>> Linux-cluster mailing list > >>> Linux-cluster@xxxxxxxxxx > >>> https://www.redhat.com/mailman/listinfo/linux-cluster > >>> > >> > >> -- > >> Linux-cluster mailing list > >> Linux-cluster@xxxxxxxxxx > >> https://www.redhat.com/mailman/listinfo/linux-cluster > >> > > > > -- > > Linux-cluster mailing list > > Linux-cluster@xxxxxxxxxx > > https://www.redhat.com/mailman/listinfo/linux-cluster > > > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster