the whole cluster.conf not only the vm entry. The problem seems to be rgmanager, 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: > Yes strange. > > cluster.conf 's entry is > > <vm autostart="0" domain="myfd1" exclusive="0" max_restarts="0" > name="guest1" path="/vms" recovery="restart" restart_expire_time="0"/> > > > Paras. > > > On Tue, Sep 29, 2009 at 4:02 PM, brem belguebli > <brem.belguebli@xxxxxxxxx> wrote: >> That looks strange, maybe your cluster.conf will help >> >> 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: >>> I don't see rgmanager . >>> >>> Here is the o/p from clustat >>> >>> [root@cvtst1 cluster]# clustat >>> Cluster Status for test @ Tue Sep 29 15:53:33 2009 >>> Member Status: Quorate >>> >>> Member Name ID Status >>> ------ ---- ---- ------ >>> cvtst2 1 Online >>> cvtst1 2 Online, Local >>> cvtst3 3 Online >>> >>> >>> Thanks >>> Paras. >>> >>> On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli >>> <brem.belguebli@xxxxxxxxx> wrote: >>>> It looks correct, rgmanager seems to start on all nodes >>>> >>>> what gives you clustat ? >>>> >>>> If rgmanager doesn't show, check out the logs something may have gone wrong. >>>> >>>> >>>> 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>> Change to 7 and i got this log >>>>> >>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down >>>>> Cluster Service Manager... >>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down >>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down >>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown complete, exiting >>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster Service >>>>> Manager is stopped. >>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group >>>>> Manager Starting >>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service Data >>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource Rules >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource Trees >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources defined >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover Domains >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing Services >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services Initialized >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port Opened >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: Local UP >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst2 UP >>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst3 UP >>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1) Processed >>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1) Processed >>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1) Processed >>>>> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events processed >>>>> >>>>> >>>>> Anything unusual here? >>>>> >>>>> Paras. >>>>> >>>>> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli >>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>>> I use log_level=7 to have more debugging info. >>>>>> >>>>>> It seems 4 is not enough. >>>>>> >>>>>> Brem >>>>>> >>>>>> >>>>>> 2009/9/29, Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>>>> Withe log_level of 3 I got only this >>>>>>> >>>>>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down >>>>>>> Cluster Service Manager... >>>>>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down >>>>>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, exiting >>>>>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service >>>>>>> Manager is stopped. >>>>>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group >>>>>>> Manager Starting >>>>>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down >>>>>>> Cluster Service Manager... >>>>>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service >>>>>>> Manager is stopped. >>>>>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group >>>>>>> Manager Starting >>>>>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting down >>>>>>> >>>>>>> I do not know what the last line means. >>>>>>> >>>>>>> rgmanager version I am running is: >>>>>>> rgmanager-2.0.52-1.el5.centos >>>>>>> >>>>>>> I don't what has gone wrong. >>>>>>> >>>>>>> Thanks >>>>>>> Paras. >>>>>>> >>>>>>> >>>>>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli >>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>>>> > you mean it stopped successfully on all the nodes but it is failing to >>>>>>> > start only on node cvtst1 ? >>>>>>> > >>>>>>> > look at the following page to make rgmanager more verbose. It 'll >>>>>>> > help debug.... >>>>>>> > >>>>>>> > http://sources.redhat.com/cluster/wiki/RGManager >>>>>>> > >>>>>>> > at Logging Configuration section >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>>>> >> Brem, >>>>>>> >> >>>>>>> >> When I try to restart rgmanager on all the nodes, this time i do not >>>>>>> >> see rgmanager running on the first node. But I do see on other 2 >>>>>>> >> nodes. >>>>>>> >> >>>>>>> >> Log on the first node: >>>>>>> >> >>>>>>> >> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group >>>>>>> >> Manager Starting >>>>>>> >> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down >>>>>>> >> Cluster Service Manager... >>>>>>> >> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down >>>>>>> >> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, exiting >>>>>>> >> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service >>>>>>> >> Manager is stopped. >>>>>>> >> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group >>>>>>> >> Manager Starting >>>>>>> >> >>>>>>> >> - >>>>>>> >> It seems service is running , but I do not see rgmanger running using clustat >>>>>>> >> >>>>>>> >> >>>>>>> >> Don't know what is going on. >>>>>>> >> >>>>>>> >> Thanks >>>>>>> >> Paras. >>>>>>> >> >>>>>>> >> >>>>>>> >> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli >>>>>>> >> <brem.belguebli@xxxxxxxxx> wrote: >>>>>>> >>> Paras, >>>>>>> >>> >>>>>>> >>> Another thing, it would have been more interesting to have a start >>>>>>> >>> DEBUG not a stop. >>>>>>> >>> >>>>>>> >>> That's why I was asking you to first stop the vm manually on all your >>>>>>> >>> nodes, stop eventually rgmanager on all the nodes to reset the >>>>>>> >>> potential wrong states you may have, restart rgmanager. >>>>>>> >>> >>>>>>> >>> If your VM is configured to autostart, this will make it start. >>>>>>> >>> >>>>>>> >>> It should normally fail (as it does now). Send out your newly created >>>>>>> >>> DEBUG file. >>>>>>> >>> >>>>>>> >>> 2009/9/29 brem belguebli <brem.belguebli@xxxxxxxxx>: >>>>>>> >>>> Hi Paras, >>>>>>> >>>> >>>>>>> >>>> >>>>>>> >>>> I don't know the xen/cluster combination well, but if I do remember >>>>>>> >>>> well, I think I've read somewhere that when using xen you have to >>>>>>> >>>> declare the use_virsh=0 key in the VM definition in the cluster.conf. >>>>>>> >>>> >>>>>>> >>>> This would make rgmanager use xm commands instead of virsh >>>>>>> >>>> The DEBUG output shows clearly that you are using virsh to manage your >>>>>>> >>>> VM instead of xm commands. >>>>>>> >>>> Check out the RH docs about virtualization >>>>>>> >>>> >>>>>>> >>>> I'm not a 100% sure about that, I may be completely wrong. >>>>>>> >>>> >>>>>>> >>>> Brem >>>>>>> >>>> >>>>>>> >>>> 2009/9/28 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>>>> >>>>> The only thing I noticed is the message after stopping the vm using xm >>>>>>> >>>>> in all nodes and starting using clusvcadm is >>>>>>> >>>>> >>>>>>> >>>>> "Virtual machine guest1 is blocked" >>>>>>> >>>>> >>>>>>> >>>>> The whole DEBUG file is attached. >>>>>>> >>>>> >>>>>>> >>>>> >>>>>>> >>>>> Thanks >>>>>>> >>>>> Paras. >>>>>>> >>>>> >>>>>>> >>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli >>>>>>> >>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>>>> >>>>>> There's a problem with the script that is called by rgmanager to start >>>>>>> >>>>>> the VM, I don't know what causes it >>>>>>> >>>>>> >>>>>>> >>>>>> May be you should try something like : >>>>>>> >>>>>> >>>>>>> >>>>>> 1) stop the VM on all nodes with xm commands >>>>>>> >>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the following >>>>>>> >>>>>> lines (after the #!/bin/bash ): >>>>>>> >>>>>> exec >/tmp/DEBUG 2>&1 >>>>>>> >>>>>> set -x >>>>>>> >>>>>> 3) start the VM with clusvcadm -e vm:guest1 >>>>>>> >>>>>> >>>>>>> >>>>>> It should fail as it did before. >>>>>>> >>>>>> >>>>>>> >>>>>> edit the the /tmp/DEBUG file and you will be able to see where it >>>>>>> >>>>>> fails (it may generate a lot of debug) >>>>>>> >>>>>> >>>>>>> >>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh >>>>>>> >>>>>> >>>>>>> >>>>>> Post the DEBUG file if you're not able to see where it fails. >>>>>>> >>>>>> >>>>>>> >>>>>> Brem >>>>>>> >>>>>> >>>>>>> >>>>>> 2009/9/26 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>>>> >>>>>>> No I am not manually starting not using automatic init scripts. >>>>>>> >>>>>>> >>>>>>> >>>>>>> I started the vm using: clusvcadm -e vm:guest1 >>>>>>> >>>>>>> >>>>>>> >>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few seconds it >>>>>>> >>>>>>> says guest1 started . But after a while I can see the guest1 on all >>>>>>> >>>>>>> three nodes. >>>>>>> >>>>>>> >>>>>>> >>>>>>> clustat says: >>>>>>> >>>>>>> >>>>>>> >>>>>>> Service Name Owner (Last) >>>>>>> >>>>>>> State >>>>>>> >>>>>>> ------- ---- ----- ------ >>>>>>> >>>>>>> ----- >>>>>>> >>>>>>> vm:guest1 (none) >>>>>>> >>>>>>> stopped >>>>>>> >>>>>>> >>>>>>> >>>>>>> But I can see the vm from xm li. >>>>>>> >>>>>>> >>>>>>> >>>>>>> This is what I can see from the log: >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" >>>>>>> >>>>>>> returned 1 (generic error) >>>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start >>>>>>> >>>>>>> vm:guest1; return value: 1 >>>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 >>>>>>> >>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is >>>>>>> >>>>>>> recovering >>>>>>> >>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed >>>>>>> >>>>>>> service vm:guest1 >>>>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" >>>>>>> >>>>>>> returned 1 (generic error) >>>>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start >>>>>>> >>>>>>> vm:guest1; return value: 1 >>>>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 >>>>>>> >>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is >>>>>>> >>>>>>> recovering >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> Paras. >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli >>>>>>> >>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>>>> >>>>>>>> Have you started your VM via rgmanager (clusvcadm -e vm:guest1) or >>>>>>> >>>>>>>> using xm commands out of cluster control (or maybe a thru an >>>>>>> >>>>>>>> automatic init script ?) >>>>>>> >>>>>>>> >>>>>>> >>>>>>>> When clustered, you should never be starting services (manually or >>>>>>> >>>>>>>> thru automatic init script) out of cluster control >>>>>>> >>>>>>>> >>>>>>> >>>>>>>> The thing would be to stop your vm on all the nodes with the adequate >>>>>>> >>>>>>>> xm command (not using xen myself) and try to start it with clusvcadm. >>>>>>> >>>>>>>> >>>>>>> >>>>>>>> Then see if it is started on all nodes (send clustat output) >>>>>>> >>>>>>>> >>>>>>> >>>>>>>> >>>>>>> >>>>>>>> >>>>>>> >>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>>>> >>>>>>>>> Ok. Please see below. my vm is running on all nodes though clustat >>>>>>> >>>>>>>>> says it is stopped. >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> -- >>>>>>> >>>>>>>>> [root@cvtst1 ~]# clustat >>>>>>> >>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009 >>>>>>> >>>>>>>>> Member Status: Quorate >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> Member Name ID Status >>>>>>> >>>>>>>>> ------ ---- ---- ------ >>>>>>> >>>>>>>>> cvtst2 1 Online, rgmanager >>>>>>> >>>>>>>>> cvtst1 2 Online, >>>>>>> >>>>>>>>> Local, rgmanager >>>>>>> >>>>>>>>> cvtst3 3 Online, rgmanager >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> Service Name Owner (Last) >>>>>>> >>>>>>>>> State >>>>>>> >>>>>>>>> ------- ---- ----- ------ >>>>>>> >>>>>>>>> ----- >>>>>>> >>>>>>>>> vm:guest1 (none) >>>>>>> >>>>>>>>> stopped >>>>>>> >>>>>>>>> [root@cvtst1 ~]# >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> --- >>>>>>> >>>>>>>>> o/p of xm li on cvtst1 >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> -- >>>>>>> >>>>>>>>> [root@cvtst1 ~]# xm li >>>>>>> >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) >>>>>>> >>>>>>>>> Domain-0 0 3470 2 r----- 28939.4 >>>>>>> >>>>>>>>> guest1 7 511 1 -b---- 7727.8 >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> o/p of xm li on cvtst2 >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> -- >>>>>>> >>>>>>>>> [root@cvtst2 ~]# xm li >>>>>>> >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) >>>>>>> >>>>>>>>> Domain-0 0 3470 2 r----- 31558.9 >>>>>>> >>>>>>>>> guest1 21 511 1 -b---- 7558.2 >>>>>>> >>>>>>>>> --- >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> Thanks >>>>>>> >>>>>>>>> Paras. >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli >>>>>>> >>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>>>> >>>>>>>>>> It looks like no. >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> can you send an output of clustat of when the VM is running on >>>>>>> >>>>>>>>>> multiple nodes at the same time? >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> And by the way, another one after having stopped (clusvcadm -s vm:guest1) ? >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>>>> >>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not being >>>>>>> >>>>>>>>>>> properly handled by the cluster. >>>>>>> >>>>>>>>>>> >>>>>>> >>>>>>>>>>> >>>>>>> >>>>>>>>>>> Thanks >>>>>>> >>>>>>>>>>> Paras. >>>>>>> >>>>>>>>>>> >>>>>>> >>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >>>>>>> >>>>>>>>>>>> Ok.. here is my cluster.conf file >>>>>>> >>>>>>>>>>>> >>>>>>> >>>>>>>>>>>> -- >>>>>>> >>>>>>>>>>>> [root@cvtst1 cluster]# more cluster.conf >>>>>>> >>>>>>>>>>>> <?xml version="1.0"?> >>>>>>> >>>>>>>>>>>> <cluster alias="test" config_version="9" name="test"> >>>>>>> >>>>>>>>>>>> <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/> >>>>>>> >>>>>>>>>>>> <clusternodes> >>>>>>> >>>>>>>>>>>> <clusternode name="cvtst2" nodeid="1" votes="1"> >>>>>>> >>>>>>>>>>>> <fence/> >>>>>>> >>>>>>>>>>>> </clusternode> >>>>>>> >>>>>>>>>>>> <clusternode name="cvtst1" nodeid="2" votes="1"> >>>>>>> >>>>>>>>>>>> <fence/> >>>>>>> >>>>>>>>>>>> </clusternode> >>>>>>> >>>>>>>>>>>> <clusternode name="cvtst3" nodeid="3" votes="1"> >>>>>>> >>>>>>>>>>>> <fence/> >>>>>>> >>>>>>>>>>>> </clusternode> >>>>>>> >>>>>>>>>>>> </clusternodes> >>>>>>> >>>>>>>>>>>> <cman/> >>>>>>> >>>>>>>>>>>> <fencedevices/> >>>>>>> >>>>>>>>>>>> <rm> >>>>>>> >>>>>>>>>>>> <failoverdomains> >>>>>>> >>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="0" ordered="1" restricted="0"> >>>>>>> >>>>>>>>>>>> <failoverdomainnode name="cvtst2" priority="3"/> >>>>>>> >>>>>>>>>>>> <failoverdomainnode name="cvtst1" priority="1"/> >>>>>>> >>>>>>>>>>>> <failoverdomainnode name="cvtst3" priority="2"/> >>>>>>> >>>>>>>>>>>> </failoverdomain> >>>>>>> >>>>>>>>>>>> </failoverdomains> >>>>>>> >>>>>>>>>>>> <resources/> >>>>>>> >>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0" >>>>>>> >>>>>>>>>>>> name="guest1" path="/vms" recovery="r >>>>>>> >>>>>>>>>>>> estart" restart_expire_time="0"/> >>>>>>> >>>>>>>>>>>> </rm> >>>>>>> >>>>>>>>>>>> </cluster> >>>>>>> >>>>>>>>>>>> [root@cvtst1 cluster]# >>>>>>> >>>>>>>>>>>> ------ >>>>>>> >>>>>>>>>>>> >>>>>>> >>>>>>>>>>>> Thanks! >>>>>>> >>>>>>>>>>>> Paras. >>>>>>> >>>>>>>>>>>> >>>>>>> >>>>>>>>>>>> >>>>>>> >>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer <volker@xxxxxxxxxxxx> wrote: >>>>>>> >>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500, >>>>>>> >>>>>>>>>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >>>>>>> >>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. Now I am >>>>>>> >>>>>>>>>>>>>> having another problem. When I start the my xen vm in one node, it >>>>>>> >>>>>>>>>>>>>> also starts on other nodes. Which daemon controls this? >>>>>>> >>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the rgmanager >>>>>>> >>>>>>>>>>>>> package). To me, this sounds like a configuration problem. Maybe, >>>>>>> >>>>>>>>>>>>> you can post your cluster.conf? >>>>>>> >>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>> Regards, >>>>>>> >>>>>>>>>>>>> Volker >>>>>>> >>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>> -- >>>>>>> >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>> >>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>> >>>>>>> >>>>>>>>>>> >>>>>>> >>>>>>>>>>> -- >>>>>>> >>>>>>>>>>> Linux-cluster mailing list >>>>>>> >>>>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>>>>>>> >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> -- >>>>>>> >>>>>>>>>> Linux-cluster mailing list >>>>>>> >>>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> -- >>>>>>> >>>>>>>>> Linux-cluster mailing list >>>>>>> >>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>> >>>>>>> >>>>>>>> -- >>>>>>> >>>>>>>> Linux-cluster mailing list >>>>>>> >>>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> Linux-cluster mailing list >>>>>>> >>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>>> >>>>>> -- >>>>>>> >>>>>> Linux-cluster mailing list >>>>>>> >>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>>> >>>>> >>>>>>> >>>>> -- >>>>>>> >>>>> Linux-cluster mailing list >>>>>>> >>>>> Linux-cluster@xxxxxxxxxx >>>>>>> >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>> >>>>>>> >>>> >>>>>>> >>> >>>>>>> >>> -- >>>>>>> >>> Linux-cluster mailing list >>>>>>> >>> Linux-cluster@xxxxxxxxxx >>>>>>> >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>> >>>>>>> >> >>>>>>> >> -- >>>>>>> >> Linux-cluster mailing list >>>>>>> >> Linux-cluster@xxxxxxxxxx >>>>>>> >> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >> >>>>>>> > >>>>>>> > -- >>>>>>> > Linux-cluster mailing list >>>>>>> > Linux-cluster@xxxxxxxxxx >>>>>>> > https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> > >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster@xxxxxxxxxx >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@xxxxxxxxxx >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@xxxxxxxxxx >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@xxxxxxxxxx >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster@xxxxxxxxxx >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> -- >> Linux-cluster mailing list >> Linux-cluster@xxxxxxxxxx >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster