No I am not manually starting not using automatic init scripts. I started the vm using: clusvcadm -e vm:guest1 I have just stopped using clusvcadm -s vm:guest1. For few seconds it says guest1 started . But after a while I can see the guest1 on all three nodes. clustat says: Service Name Owner (Last) State ------- ---- ----- ------ ----- vm:guest1 (none) stopped But I can see the vm from xm li. This is what I can see from the log: Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" returned 1 (generic error) Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start vm:guest1; return value: 1 Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is recovering Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed service vm:guest1 Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" returned 1 (generic error) Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start vm:guest1; return value: 1 Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is recovering Paras. On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli <brem.belguebli@xxxxxxxxx> wrote: > Have you started your VM via rgmanager (clusvcadm -e vm:guest1) or > using xm commands out of cluster control (or maybe a thru an > automatic init script ?) > > When clustered, you should never be starting services (manually or > thru automatic init script) out of cluster control > > The thing would be to stop your vm on all the nodes with the adequate > xm command (not using xen myself) and try to start it with clusvcadm. > > Then see if it is started on all nodes (send clustat output) > > > > 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >> Ok. Please see below. my vm is running on all nodes though clustat >> says it is stopped. >> >> -- >> [root@cvtst1 ~]# clustat >> Cluster Status for test @ Fri Sep 25 16:52:34 2009 >> Member Status: Quorate >> >> Member Name ID Status >> ------ ---- ---- ------ >> cvtst2 1 Online, rgmanager >> cvtst1 2 Online, >> Local, rgmanager >> cvtst3 3 Online, rgmanager >> >> Service Name Owner (Last) >> State >> ------- ---- ----- ------ >> ----- >> vm:guest1 (none) >> stopped >> [root@cvtst1 ~]# >> >> >> --- >> o/p of xm li on cvtst1 >> >> -- >> [root@cvtst1 ~]# xm li >> Name ID Mem(MiB) VCPUs State Time(s) >> Domain-0 0 3470 2 r----- 28939.4 >> guest1 7 511 1 -b---- 7727.8 >> >> o/p of xm li on cvtst2 >> >> -- >> [root@cvtst2 ~]# xm li >> Name ID Mem(MiB) VCPUs State Time(s) >> Domain-0 0 3470 2 r----- 31558.9 >> guest1 21 511 1 -b---- 7558.2 >> --- >> >> Thanks >> Paras. >> >> >> >> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli >> <brem.belguebli@xxxxxxxxx> wrote: >>> It looks like no. >>> >>> can you send an output of clustat of when the VM is running on >>> multiple nodes at the same time? >>> >>> And by the way, another one after having stopped (clusvcadm -s vm:guest1) ? >>> >>> >>> >>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>> Anyone having issue as mine? Virtual machine service is not being >>>> properly handled by the cluster. >>>> >>>> >>>> Thanks >>>> Paras. >>>> >>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >>>>> Ok.. here is my cluster.conf file >>>>> >>>>> -- >>>>> [root@cvtst1 cluster]# more cluster.conf >>>>> <?xml version="1.0"?> >>>>> <cluster alias="test" config_version="9" name="test"> >>>>> <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/> >>>>> <clusternodes> >>>>> <clusternode name="cvtst2" nodeid="1" votes="1"> >>>>> <fence/> >>>>> </clusternode> >>>>> <clusternode name="cvtst1" nodeid="2" votes="1"> >>>>> <fence/> >>>>> </clusternode> >>>>> <clusternode name="cvtst3" nodeid="3" votes="1"> >>>>> <fence/> >>>>> </clusternode> >>>>> </clusternodes> >>>>> <cman/> >>>>> <fencedevices/> >>>>> <rm> >>>>> <failoverdomains> >>>>> <failoverdomain name="myfd1" nofailback="0" ordered="1" restricted="0"> >>>>> <failoverdomainnode name="cvtst2" priority="3"/> >>>>> <failoverdomainnode name="cvtst1" priority="1"/> >>>>> <failoverdomainnode name="cvtst3" priority="2"/> >>>>> </failoverdomain> >>>>> </failoverdomains> >>>>> <resources/> >>>>> <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0" >>>>> name="guest1" path="/vms" recovery="r >>>>> estart" restart_expire_time="0"/> >>>>> </rm> >>>>> </cluster> >>>>> [root@cvtst1 cluster]# >>>>> ------ >>>>> >>>>> Thanks! >>>>> Paras. >>>>> >>>>> >>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer <volker@xxxxxxxxxxxx> wrote: >>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500, >>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >>>>>>> I am using cluster suite for HA of xen virtual machines. Now I am >>>>>>> having another problem. When I start the my xen vm in one node, it >>>>>>> also starts on other nodes. Which daemon controls this? >>>>>> >>>>>> This is usually done bei clurgmgrd (which is part of the rgmanager >>>>>> package). To me, this sounds like a configuration problem. Maybe, >>>>>> you can post your cluster.conf? >>>>>> >>>>>> Regards, >>>>>> Volker >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@xxxxxxxxxx >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@xxxxxxxxxx >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster@xxxxxxxxxx >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> -- >> Linux-cluster mailing list >> Linux-cluster@xxxxxxxxxx >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster