Re: openais issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes I did that as well. Node1 of my cluster doesn't show (using clust)
vm service whereas others do. I guess the problem is with the
rgmanager but it looks to be running fine.


Paras.


On Tue, Oct 6, 2009 at 12:59 PM, Daniela Anzellotti
<daniela.anzellotti@xxxxxxxxxxxxx> wrote:
> Hi Paras,
>
> did you reboot all the cluster nodes? I needed a complete reboot (actually I
> was angry enough to switch everything off and on again): a restart of the
> cluster suite was not enough.
>
> As far as I understood, since the cluster was not able to bring a VM in a
> good running state, it decided that the VM has to migrate to another node...
> and at the end all the VMs was trying to migrate from one node to another
> with the result that I had all VMs starting on all the cluster nodes.
> Restarting the cluster suite didn't kill a lot of processes that was
> stubbornly trying to migrate virtual machines...
>
> Daniela
>
> Paras pradhan wrote:
>>
>> Adding use_virsh=0 works great. Now I do not have my vm starting at
>> all the nodes. This is a good fix. Thanks ..
>>
>> The only problem left is I do not see rgmanager running on my node1
>> and the clustat of node2 and node3 is reporting vm as migrating
>>
>> o/p
>>
>> Service Name                                          Owner (Last)
>>                                     State
>>  ------- ----                                          ----- ------
>>                                      -----
>>  vm:guest1                                             cvtst1
>>                             migrating
>> [root@cvtst3 vms]#
>>
>>
>> Thanks
>> Paras.
>>
>> On Tue, Oct 6, 2009 at 10:48 AM, Daniela Anzellotti
>> <daniela.anzellotti@xxxxxxxxxxxxx> wrote:
>>>
>>> Hi Paras,
>>>
>>> yes. At least it looks so...
>>>
>>> We have a cluster of two nodes + a quorum disk (it's not configured as a
>>> "two-node cluster")
>>>
>>> They are running Scientific Linux 5.x, kernel 2.6.18-128.7.1.el5xen and
>>>
>>>  openais-0.80.6-8.el5.x86_64
>>>  cman-2.0.115-1.el5.x86_64
>>>  rgmanager-2.0.52-1.el5.x86_64
>>>
>>> The XEN VMs access the disk as simple block devices.
>>> Disks are on a SAN, configured with Clustered LVM.
>>>
>>>  xen-3.0.3-94.el5_4.1.x86_64
>>>  xen-libs-3.0.3-94.el5_4.1.x86_64
>>>
>>> VM configuration files are as the following
>>>
>>>  name = "www1"
>>>  uuid = "3bd3e910-23c0-97ee-55ab-086260ef1e53"
>>>  memory = 1024
>>>  maxmem = 1024
>>>  vcpus = 1
>>>  bootloader = "/usr/bin/pygrub"
>>>  vfb = [ "type=vnc,vncunused=1,keymap=en-us" ]
>>>  disk = [ "phy:/dev/vg_cluster/www1.disk,xvda,w", \
>>>  "phy:/dev/vg_cluster/www1.swap,xvdb,w" ]
>>>  vif = [ "mac=00:16:3e:da:00:07,bridge=xenbr1" ]
>>>  on_poweroff = "destroy"
>>>  on_reboot = "restart"
>>>  on_crash = "restart"
>>>  extra = "xencons=tty0 console=tty0"
>>>
>>>
>>> I changed in /etc/cluster/cluster.conf all the VM directive from
>>>
>>>  <vm autostart="1" domain="rhcs1_dom" exclusive="0" \
>>>  migrate="live" name="www1" path="/etc/xen" recovery="restart"/>
>>>
>>> to
>>>
>>>  <vm autostart="1" use_virsh="0" domain="rhcs1_dom" exclusive="0" \
>>>  migrate="live" name="www1" path="/etc/xen" recovery="restart"/>
>>>
>>>
>>> Rebooted the cluster nodes and it started working again...
>>>
>>> As i said I hope I'll not have any other bad surprise (I tested a VM
>>> migration and it is working too), but at least cluster it's working now
>>> (it
>>> was not able to start a VM, before)!
>>>
>>> Ciao
>>> Daniela
>>>
>>>
>>> Paras pradhan wrote:
>>>>
>>>> So you mean your cluster is running fine with the CMAN
>>>> cman-2.0.115-1.el5.x86_64 ?
>>>>
>>>> Which version of openais are you running?
>>>>
>>>> Thanks
>>>> Paras.
>>>>
>>>>
>>>> On Mon, Oct 5, 2009 at 7:19 AM, Daniela Anzellotti
>>>> <daniela.anzellotti@xxxxxxxxxxxxx> wrote:
>>>>>
>>>>> Hi all,
>>>>>
>>>>> I had a problem similar to Paras's one today: yum updated the following
>>>>> rpms
>>>>> last week and today (I had to restart the cluster) the cluster was not
>>>>> able
>>>>> to start vm: services.
>>>>>
>>>>> Oct 02 05:31:05 Updated: openais-0.80.6-8.el5.x86_64
>>>>> Oct 02 05:31:07 Updated: cman-2.0.115-1.el5.x86_64
>>>>> Oct 02 05:31:10 Updated: rgmanager-2.0.52-1.el5.x86_64
>>>>>
>>>>> Oct 03 04:03:12 Updated: xen-libs-3.0.3-94.el5_4.1.x86_64
>>>>> Oct 03 04:03:12 Updated: xen-libs-3.0.3-94.el5_4.1.i386
>>>>> Oct 03 04:03:16 Updated: xen-3.0.3-94.el5_4.1.x86_64
>>>>>
>>>>>
>>>>> So, after checked the vm.sh script, I add the declaration use_virsh="0"
>>>>> in
>>>>> the VM definition in the cluster.conf (as suggested by Brem, thanks!)
>>>>> and
>>>>> everything is now working again.
>>>>>
>>>>>
>>>>> BTW I didn't understand if the problem was caused by the new XEN
>>>>> version
>>>>> or
>>>>> the new openais one, thus I disabled automatic updates for both.
>>>>>
>>>>> I hope I'll not have any other bad surprise...
>>>>>
>>>>> Thank you,
>>>>> cheers,
>>>>> Daniela
>>>>>
>>>>>
>>>>> Paras pradhan wrote:
>>>>>>
>>>>>> Yes this is very strange. I don't know what to do now. May be re
>>>>>> create the cluster? But not a good solution actually.
>>>>>>
>>>>>> Packages :
>>>>>>
>>>>>> Kernel: kernel-xen-2.6.18-164.el5
>>>>>> OS: Full updated of CentOS 5.3 except CMAN downgraded to
>>>>>> cman-2.0.98-1.el5
>>>>>>
>>>>>> Other packages related to cluster suite:
>>>>>>
>>>>>> rgmanager-2.0.52-1.el5.centos
>>>>>> cman-2.0.98-1.el5
>>>>>> xen-3.0.3-80.el5_3.3
>>>>>> xen-libs-3.0.3-80.el5_3.3
>>>>>> kmod-gfs-xen-0.1.31-3.el5_3.1
>>>>>> kmod-gfs-xen-0.1.31-3.el5_3.1
>>>>>> kmod-gfs-0.1.31-3.el5_3.1
>>>>>> gfs-utils-0.1.18-1.el5
>>>>>> gfs2-utils-0.1.62-1.el5
>>>>>> lvm2-2.02.40-6.el5
>>>>>> lvm2-cluster-2.02.40-7.el5
>>>>>> openais-0.80.3-22.el5_3.9
>>>>>>
>>>>>> Thanks!
>>>>>> Paras.
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, Sep 30, 2009 at 10:02 AM, brem belguebli
>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>>>>
>>>>>>> Hi Paras,
>>>>>>>
>>>>>>> Your cluster.conf file seems correct. If it is not a ntp issue, I
>>>>>>> don't see anything except a bug that causes this, or some
>>>>>>> prerequisite
>>>>>>> that is not respected.
>>>>>>>
>>>>>>> May be you could post the versions (os, kernel, packages etc...) you
>>>>>>> are using, someone may have hit the same issue with your versions.
>>>>>>>
>>>>>>> Brem
>>>>>>>
>>>>>>> 2009/9/30, Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>>>>>
>>>>>>>> All of the nodes are synced with ntp server. So this is not the case
>>>>>>>> with me.
>>>>>>>>
>>>>>>>> Thanks
>>>>>>>> Paras.
>>>>>>>>
>>>>>>>> On Tue, Sep 29, 2009 at 6:29 PM, Johannes Rußek
>>>>>>>> <johannes.russek@xxxxxxxxxxxxxxxxx> wrote:
>>>>>>>>>
>>>>>>>>> make sure the time on the nodes is in sync, apparently when a node
>>>>>>>>> has
>>>>>>>>> too
>>>>>>>>> much offset, you won't see rgmanager (even though the process is
>>>>>>>>> running).
>>>>>>>>> this happened today and setting the time fixed it for me. afaicr
>>>>>>>>> there
>>>>>>>>> was
>>>>>>>>> no sign of this in the logs though.
>>>>>>>>> johannes
>>>>>>>>>
>>>>>>>>> Paras pradhan schrieb:
>>>>>>>>>>
>>>>>>>>>> I don't see rgmanager .
>>>>>>>>>>
>>>>>>>>>> Here is the o/p from clustat
>>>>>>>>>>
>>>>>>>>>> [root@cvtst1 cluster]# clustat
>>>>>>>>>> Cluster Status for test @ Tue Sep 29 15:53:33 2009
>>>>>>>>>> Member Status: Quorate
>>>>>>>>>>
>>>>>>>>>>  Member Name
>>>>>>>>>> ID
>>>>>>>>>> Status
>>>>>>>>>>  ------ ----
>>>>>>>>>> ----
>>>>>>>>>> ------
>>>>>>>>>>  cvtst2                                                    1
>>>>>>>>>> Online
>>>>>>>>>>  cvtst1                                                     2
>>>>>>>>>> Online,
>>>>>>>>>> Local
>>>>>>>>>>  cvtst3                                                     3
>>>>>>>>>> Online
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Thanks
>>>>>>>>>> Paras.
>>>>>>>>>>
>>>>>>>>>> On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli
>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>>>>>>>
>>>>>>>>>>> It looks correct, rgmanager seems to start on all nodes
>>>>>>>>>>>
>>>>>>>>>>> what gives you clustat ?
>>>>>>>>>>>
>>>>>>>>>>> If rgmanager doesn't show, check out the logs something may have
>>>>>>>>>>> gone
>>>>>>>>>>> wrong.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>>>>>>>>
>>>>>>>>>>>> Change to 7 and i got this log
>>>>>>>>>>>>
>>>>>>>>>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting
>>>>>>>>>>>> down
>>>>>>>>>>>> Cluster Service Manager...
>>>>>>>>>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>>>>>>>>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>>>>>>>>>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown
>>>>>>>>>>>> complete,
>>>>>>>>>>>> exiting
>>>>>>>>>>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster
>>>>>>>>>>>> Service
>>>>>>>>>>>> Manager is stopped.
>>>>>>>>>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group
>>>>>>>>>>>> Manager Starting
>>>>>>>>>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service
>>>>>>>>>>>> Data
>>>>>>>>>>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading
>>>>>>>>>>>> Resource
>>>>>>>>>>>> Rules
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building
>>>>>>>>>>>> Resource
>>>>>>>>>>>> Trees
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources
>>>>>>>>>>>> defined
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading
>>>>>>>>>>>> Failover
>>>>>>>>>>>> Domains
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains
>>>>>>>>>>>> defined
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events
>>>>>>>>>>>> defined
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing
>>>>>>>>>>>> Services
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services
>>>>>>>>>>>> Initialized
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port
>>>>>>>>>>>> Opened
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change:
>>>>>>>>>>>> Local
>>>>>>>>>>>> UP
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change:
>>>>>>>>>>>> cvtst2
>>>>>>>>>>>> UP
>>>>>>>>>>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change:
>>>>>>>>>>>> cvtst3
>>>>>>>>>>>> UP
>>>>>>>>>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1)
>>>>>>>>>>>> Processed
>>>>>>>>>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1)
>>>>>>>>>>>> Processed
>>>>>>>>>>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1)
>>>>>>>>>>>> Processed
>>>>>>>>>>>> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events
>>>>>>>>>>>> processed
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Anything unusual here?
>>>>>>>>>>>>
>>>>>>>>>>>> Paras.
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli
>>>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> I use log_level=7 to have more debugging info.
>>>>>>>>>>>>>
>>>>>>>>>>>>> It seems 4 is not enough.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Brem
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2009/9/29, Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Withe log_level of 3 I got only this
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting
>>>>>>>>>>>>>> down
>>>>>>>>>>>>>> Cluster Service Manager...
>>>>>>>>>>>>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down
>>>>>>>>>>>>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown
>>>>>>>>>>>>>> complete,
>>>>>>>>>>>>>> exiting
>>>>>>>>>>>>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster
>>>>>>>>>>>>>> Service
>>>>>>>>>>>>>> Manager is stopped.
>>>>>>>>>>>>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource
>>>>>>>>>>>>>> Group
>>>>>>>>>>>>>> Manager Starting
>>>>>>>>>>>>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting
>>>>>>>>>>>>>> down
>>>>>>>>>>>>>> Cluster Service Manager...
>>>>>>>>>>>>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster
>>>>>>>>>>>>>> Service
>>>>>>>>>>>>>> Manager is stopped.
>>>>>>>>>>>>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource
>>>>>>>>>>>>>> Group
>>>>>>>>>>>>>> Manager Starting
>>>>>>>>>>>>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1
>>>>>>>>>>>>>> shutting
>>>>>>>>>>>>>> down
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I do not know what the last line means.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> rgmanager version I am running is:
>>>>>>>>>>>>>> rgmanager-2.0.52-1.el5.centos
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I don't what has gone wrong.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli
>>>>>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> you mean it stopped successfully on all the nodes but it is
>>>>>>>>>>>>>>> failing
>>>>>>>>>>>>>>> to
>>>>>>>>>>>>>>> start only on node cvtst1 ?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> look at the following page  to make rgmanager more verbose.
>>>>>>>>>>>>>>> It
>>>>>>>>>>>>>>> 'll
>>>>>>>>>>>>>>> help debug....
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> http://sources.redhat.com/cluster/wiki/RGManager
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> at Logging Configuration section
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Brem,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> When I try to restart rgmanager on all the nodes, this time
>>>>>>>>>>>>>>>> i
>>>>>>>>>>>>>>>> do
>>>>>>>>>>>>>>>> not
>>>>>>>>>>>>>>>> see rgmanager running on the first node. But I do see on
>>>>>>>>>>>>>>>> other
>>>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>>> nodes.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Log on the first node:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource
>>>>>>>>>>>>>>>> Group
>>>>>>>>>>>>>>>> Manager Starting
>>>>>>>>>>>>>>>> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting
>>>>>>>>>>>>>>>> down
>>>>>>>>>>>>>>>> Cluster Service Manager...
>>>>>>>>>>>>>>>> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting
>>>>>>>>>>>>>>>> down
>>>>>>>>>>>>>>>> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown
>>>>>>>>>>>>>>>> complete,
>>>>>>>>>>>>>>>> exiting
>>>>>>>>>>>>>>>> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster
>>>>>>>>>>>>>>>> Service
>>>>>>>>>>>>>>>> Manager is stopped.
>>>>>>>>>>>>>>>> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource
>>>>>>>>>>>>>>>> Group
>>>>>>>>>>>>>>>> Manager Starting
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> -
>>>>>>>>>>>>>>>> It seems service is running ,  but I do not see rgmanger
>>>>>>>>>>>>>>>> running
>>>>>>>>>>>>>>>> using clustat
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Don't know what is going on.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli
>>>>>>>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Paras,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Another thing, it would have been more interesting to have
>>>>>>>>>>>>>>>>> a
>>>>>>>>>>>>>>>>> start
>>>>>>>>>>>>>>>>> DEBUG not a stop.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> That's why I was asking you to first stop the vm manually
>>>>>>>>>>>>>>>>> on
>>>>>>>>>>>>>>>>> all
>>>>>>>>>>>>>>>>> your
>>>>>>>>>>>>>>>>> nodes, stop eventually rgmanager on all the nodes to reset
>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>> potential wrong states you may have, restart rgmanager.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> If your VM is configured to autostart, this will make it
>>>>>>>>>>>>>>>>> start.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> It should normally fail (as it does now). Send out your
>>>>>>>>>>>>>>>>> newly
>>>>>>>>>>>>>>>>> created
>>>>>>>>>>>>>>>>> DEBUG file.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> 2009/9/29 brem belguebli <brem.belguebli@xxxxxxxxx>:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi Paras,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I don't know the xen/cluster combination well, but if I do
>>>>>>>>>>>>>>>>>> remember
>>>>>>>>>>>>>>>>>> well, I think I've read somewhere that when using xen you
>>>>>>>>>>>>>>>>>> have
>>>>>>>>>>>>>>>>>> to
>>>>>>>>>>>>>>>>>> declare the use_virsh=0 key in the VM definition in the
>>>>>>>>>>>>>>>>>> cluster.conf.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> This would make rgmanager use xm commands instead of virsh
>>>>>>>>>>>>>>>>>> The DEBUG output shows clearly that you are using virsh to
>>>>>>>>>>>>>>>>>> manage
>>>>>>>>>>>>>>>>>> your
>>>>>>>>>>>>>>>>>> VM instead of xm commands.
>>>>>>>>>>>>>>>>>> Check out the RH docs about virtualization
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I'm not a 100% sure about that, I may be completely wrong.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Brem
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> 2009/9/28 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> The only thing I noticed is the message after stopping
>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>> vm
>>>>>>>>>>>>>>>>>>> using xm
>>>>>>>>>>>>>>>>>>> in all nodes and starting using clusvcadm is
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> "Virtual machine guest1 is blocked"
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> The whole DEBUG file is attached.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli
>>>>>>>>>>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> There's a problem with the script that is called by
>>>>>>>>>>>>>>>>>>>> rgmanager to
>>>>>>>>>>>>>>>>>>>> start
>>>>>>>>>>>>>>>>>>>> the VM, I don't know what causes it
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> May be you should try something like :
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> 1) stop the VM on all nodes with xm commands
>>>>>>>>>>>>>>>>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the
>>>>>>>>>>>>>>>>>>>> following
>>>>>>>>>>>>>>>>>>>> lines (after the #!/bin/bash ):
>>>>>>>>>>>>>>>>>>>>  exec >/tmp/DEBUG 2>&1
>>>>>>>>>>>>>>>>>>>>  set -x
>>>>>>>>>>>>>>>>>>>> 3) start the VM with clusvcadm -e vm:guest1
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> It should fail as it did before.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> edit the the /tmp/DEBUG file and you will be able to see
>>>>>>>>>>>>>>>>>>>> where
>>>>>>>>>>>>>>>>>>>> it
>>>>>>>>>>>>>>>>>>>> fails (it may generate a lot of debug)
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Post the DEBUG file if you're not able to see where it
>>>>>>>>>>>>>>>>>>>> fails.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Brem
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> 2009/9/26 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> No I am not manually starting not using automatic init
>>>>>>>>>>>>>>>>>>>>> scripts.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> I started the vm using: clusvcadm -e vm:guest1
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> I have just stopped using clusvcadm -s vm:guest1. For
>>>>>>>>>>>>>>>>>>>>> few
>>>>>>>>>>>>>>>>>>>>> seconds it
>>>>>>>>>>>>>>>>>>>>> says guest1 started . But after a while I can see the
>>>>>>>>>>>>>>>>>>>>> guest1 on
>>>>>>>>>>>>>>>>>>>>> all
>>>>>>>>>>>>>>>>>>>>> three nodes.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> clustat says:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>  Service Name
>>>>>>>>>>>>>>>>>>>>>  Owner
>>>>>>>>>>>>>>>>>>>>> (Last)
>>>>>>>>>>>>>>>>>>>>>                                      State
>>>>>>>>>>>>>>>>>>>>>  ------- ----
>>>>>>>>>>>>>>>>>>>>>  -----
>>>>>>>>>>>>>>>>>>>>> ------
>>>>>>>>>>>>>>>>>>>>>                                      -----
>>>>>>>>>>>>>>>>>>>>>  vm:guest1
>>>>>>>>>>>>>>>>>>>>> (none)
>>>>>>>>>>>>>>>>>>>>>                                      stopped
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> But I can see the vm from xm li.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> This is what I can see from the log:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start
>>>>>>>>>>>>>>>>>>>>> on
>>>>>>>>>>>>>>>>>>>>> vm
>>>>>>>>>>>>>>>>>>>>> "guest1"
>>>>>>>>>>>>>>>>>>>>> returned 1 (generic error)
>>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68:
>>>>>>>>>>>>>>>>>>>>> Failed
>>>>>>>>>>>>>>>>>>>>> to start
>>>>>>>>>>>>>>>>>>>>> vm:guest1; return value: 1
>>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice>
>>>>>>>>>>>>>>>>>>>>> Stopping
>>>>>>>>>>>>>>>>>>>>> service vm:guest1
>>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice>
>>>>>>>>>>>>>>>>>>>>> Service
>>>>>>>>>>>>>>>>>>>>> vm:guest1 is
>>>>>>>>>>>>>>>>>>>>> recovering
>>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice>
>>>>>>>>>>>>>>>>>>>>> Recovering
>>>>>>>>>>>>>>>>>>>>> failed
>>>>>>>>>>>>>>>>>>>>> service vm:guest1
>>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start
>>>>>>>>>>>>>>>>>>>>> on
>>>>>>>>>>>>>>>>>>>>> vm
>>>>>>>>>>>>>>>>>>>>> "guest1"
>>>>>>>>>>>>>>>>>>>>> returned 1 (generic error)
>>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68:
>>>>>>>>>>>>>>>>>>>>> Failed
>>>>>>>>>>>>>>>>>>>>> to start
>>>>>>>>>>>>>>>>>>>>> vm:guest1; return value: 1
>>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice>
>>>>>>>>>>>>>>>>>>>>> Stopping
>>>>>>>>>>>>>>>>>>>>> service vm:guest1
>>>>>>>>>>>>>>>>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice>
>>>>>>>>>>>>>>>>>>>>> Service
>>>>>>>>>>>>>>>>>>>>> vm:guest1 is
>>>>>>>>>>>>>>>>>>>>> recovering
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli
>>>>>>>>>>>>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Have you started  your VM via rgmanager (clusvcadm -e
>>>>>>>>>>>>>>>>>>>>>> vm:guest1) or
>>>>>>>>>>>>>>>>>>>>>> using xm commands out of cluster control  (or maybe a
>>>>>>>>>>>>>>>>>>>>>> thru
>>>>>>>>>>>>>>>>>>>>>> an
>>>>>>>>>>>>>>>>>>>>>> automatic init script ?)
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> When clustered, you should never be starting services
>>>>>>>>>>>>>>>>>>>>>> (manually or
>>>>>>>>>>>>>>>>>>>>>> thru automatic init script) out of cluster control
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> The thing would be to stop your vm on all the nodes
>>>>>>>>>>>>>>>>>>>>>> with
>>>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>> adequate
>>>>>>>>>>>>>>>>>>>>>> xm command (not using xen myself) and try to start it
>>>>>>>>>>>>>>>>>>>>>> with
>>>>>>>>>>>>>>>>>>>>>> clusvcadm.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> Then see if it is started on all nodes (send clustat
>>>>>>>>>>>>>>>>>>>>>> output)
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Ok. Please see below. my vm is running on all nodes
>>>>>>>>>>>>>>>>>>>>>>> though
>>>>>>>>>>>>>>>>>>>>>>> clustat
>>>>>>>>>>>>>>>>>>>>>>> says it is stopped.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>> [root@cvtst1 ~]# clustat
>>>>>>>>>>>>>>>>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009
>>>>>>>>>>>>>>>>>>>>>>> Member Status: Quorate
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>  Member Name
>>>>>>>>>>>>>>>>>>>>>>>  ID   Status
>>>>>>>>>>>>>>>>>>>>>>>  ------ ----
>>>>>>>>>>>>>>>>>>>>>>>  ---- ------
>>>>>>>>>>>>>>>>>>>>>>>  cvtst2
>>>>>>>>>>>>>>>>>>>>>>>  1
>>>>>>>>>>>>>>>>>>>>>>> Online, rgmanager
>>>>>>>>>>>>>>>>>>>>>>>  cvtst1
>>>>>>>>>>>>>>>>>>>>>>>  2
>>>>>>>>>>>>>>>>>>>>>>> Online,
>>>>>>>>>>>>>>>>>>>>>>> Local, rgmanager
>>>>>>>>>>>>>>>>>>>>>>>  cvtst3
>>>>>>>>>>>>>>>>>>>>>>>  3
>>>>>>>>>>>>>>>>>>>>>>> Online, rgmanager
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>  Service Name
>>>>>>>>>>>>>>>>>>>>>>>  Owner (Last)
>>>>>>>>>>>>>>>>>>>>>>>                                      State
>>>>>>>>>>>>>>>>>>>>>>>  ------- ----
>>>>>>>>>>>>>>>>>>>>>>>  ----- ------
>>>>>>>>>>>>>>>>>>>>>>>                                      -----
>>>>>>>>>>>>>>>>>>>>>>>  vm:guest1
>>>>>>>>>>>>>>>>>>>>>>> (none)
>>>>>>>>>>>>>>>>>>>>>>>                                      stopped
>>>>>>>>>>>>>>>>>>>>>>> [root@cvtst1 ~]#
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> ---
>>>>>>>>>>>>>>>>>>>>>>> o/p of xm li on cvtst1
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>> [root@cvtst1 ~]# xm li
>>>>>>>>>>>>>>>>>>>>>>> Name                                      ID Mem(MiB)
>>>>>>>>>>>>>>>>>>>>>>> VCPUs
>>>>>>>>>>>>>>>>>>>>>>> State   Time(s)
>>>>>>>>>>>>>>>>>>>>>>> Domain-0                                   0     3470
>>>>>>>>>>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>>>>>>>>>> r-----  28939.4
>>>>>>>>>>>>>>>>>>>>>>> guest1                                     7      511
>>>>>>>>>>>>>>>>>>>>>>> 1
>>>>>>>>>>>>>>>>>>>>>>> -b----   7727.8
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> o/p of xm li on cvtst2
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>> [root@cvtst2 ~]# xm li
>>>>>>>>>>>>>>>>>>>>>>> Name                                      ID Mem(MiB)
>>>>>>>>>>>>>>>>>>>>>>> VCPUs
>>>>>>>>>>>>>>>>>>>>>>> State   Time(s)
>>>>>>>>>>>>>>>>>>>>>>> Domain-0                                   0     3470
>>>>>>>>>>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>>>>>>>>>> r-----  31558.9
>>>>>>>>>>>>>>>>>>>>>>> guest1                                    21      511
>>>>>>>>>>>>>>>>>>>>>>> 1
>>>>>>>>>>>>>>>>>>>>>>> -b----   7558.2
>>>>>>>>>>>>>>>>>>>>>>> ---
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli
>>>>>>>>>>>>>>>>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> It looks like no.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> can you send an output of clustat  of when the VM is
>>>>>>>>>>>>>>>>>>>>>>>> running
>>>>>>>>>>>>>>>>>>>>>>>> on
>>>>>>>>>>>>>>>>>>>>>>>> multiple nodes at the same time?
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> And by the way, another one after having stopped
>>>>>>>>>>>>>>>>>>>>>>>> (clusvcadm
>>>>>>>>>>>>>>>>>>>>>>>> -s vm:guest1) ?
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Anyone having issue as mine? Virtual machine
>>>>>>>>>>>>>>>>>>>>>>>>> service
>>>>>>>>>>>>>>>>>>>>>>>>> is
>>>>>>>>>>>>>>>>>>>>>>>>> not
>>>>>>>>>>>>>>>>>>>>>>>>> being
>>>>>>>>>>>>>>>>>>>>>>>>> properly handled by the cluster.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan
>>>>>>>>>>>>>>>>>>>>>>>>> <pradhanparas@xxxxxxxxx> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> Ok.. here is my cluster.conf file
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>>>>> [root@cvtst1 cluster]# more cluster.conf
>>>>>>>>>>>>>>>>>>>>>>>>>> <?xml version="1.0"?>
>>>>>>>>>>>>>>>>>>>>>>>>>> <cluster alias="test" config_version="9"
>>>>>>>>>>>>>>>>>>>>>>>>>> name="test">
>>>>>>>>>>>>>>>>>>>>>>>>>>    <fence_daemon clean_start="0"
>>>>>>>>>>>>>>>>>>>>>>>>>> post_fail_delay="0"
>>>>>>>>>>>>>>>>>>>>>>>>>> post_join_delay="3"/>
>>>>>>>>>>>>>>>>>>>>>>>>>>    <clusternodes>
>>>>>>>>>>>>>>>>>>>>>>>>>>            <clusternode name="cvtst2" nodeid="1"
>>>>>>>>>>>>>>>>>>>>>>>>>> votes="1">
>>>>>>>>>>>>>>>>>>>>>>>>>>                    <fence/>
>>>>>>>>>>>>>>>>>>>>>>>>>>            </clusternode>
>>>>>>>>>>>>>>>>>>>>>>>>>>            <clusternode name="cvtst1" nodeid="2"
>>>>>>>>>>>>>>>>>>>>>>>>>> votes="1">
>>>>>>>>>>>>>>>>>>>>>>>>>>                    <fence/>
>>>>>>>>>>>>>>>>>>>>>>>>>>            </clusternode>
>>>>>>>>>>>>>>>>>>>>>>>>>>            <clusternode name="cvtst3" nodeid="3"
>>>>>>>>>>>>>>>>>>>>>>>>>> votes="1">
>>>>>>>>>>>>>>>>>>>>>>>>>>                    <fence/>
>>>>>>>>>>>>>>>>>>>>>>>>>>            </clusternode>
>>>>>>>>>>>>>>>>>>>>>>>>>>    </clusternodes>
>>>>>>>>>>>>>>>>>>>>>>>>>>    <cman/>
>>>>>>>>>>>>>>>>>>>>>>>>>>    <fencedevices/>
>>>>>>>>>>>>>>>>>>>>>>>>>>    <rm>
>>>>>>>>>>>>>>>>>>>>>>>>>>            <failoverdomains>
>>>>>>>>>>>>>>>>>>>>>>>>>>                    <failoverdomain name="myfd1"
>>>>>>>>>>>>>>>>>>>>>>>>>> nofailback="0" ordered="1" restricted="0">
>>>>>>>>>>>>>>>>>>>>>>>>>>                            <failoverdomainnode
>>>>>>>>>>>>>>>>>>>>>>>>>> name="cvtst2" priority="3"/>
>>>>>>>>>>>>>>>>>>>>>>>>>>                            <failoverdomainnode
>>>>>>>>>>>>>>>>>>>>>>>>>> name="cvtst1" priority="1"/>
>>>>>>>>>>>>>>>>>>>>>>>>>>                            <failoverdomainnode
>>>>>>>>>>>>>>>>>>>>>>>>>> name="cvtst3" priority="2"/>
>>>>>>>>>>>>>>>>>>>>>>>>>>                    </failoverdomain>
>>>>>>>>>>>>>>>>>>>>>>>>>>            </failoverdomains>
>>>>>>>>>>>>>>>>>>>>>>>>>>            <resources/>
>>>>>>>>>>>>>>>>>>>>>>>>>>            <vm autostart="1" domain="myfd1"
>>>>>>>>>>>>>>>>>>>>>>>>>> exclusive="0" max_restarts="0"
>>>>>>>>>>>>>>>>>>>>>>>>>> name="guest1" path="/vms" recovery="r
>>>>>>>>>>>>>>>>>>>>>>>>>> estart" restart_expire_time="0"/>
>>>>>>>>>>>>>>>>>>>>>>>>>>    </rm>
>>>>>>>>>>>>>>>>>>>>>>>>>> </cluster>
>>>>>>>>>>>>>>>>>>>>>>>>>> [root@cvtst1 cluster]#
>>>>>>>>>>>>>>>>>>>>>>>>>> ------
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>>>>>>>>>>>>> Paras.
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer
>>>>>>>>>>>>>>>>>>>>>>>>>> <volker@xxxxxxxxxxxx> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500,
>>>>>>>>>>>>>>>>>>>>>>>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual
>>>>>>>>>>>>>>>>>>>>>>>>>>>> machines.
>>>>>>>>>>>>>>>>>>>>>>>>>>>> Now I am
>>>>>>>>>>>>>>>>>>>>>>>>>>>> having another problem. When I start the my xen
>>>>>>>>>>>>>>>>>>>>>>>>>>>> vm
>>>>>>>>>>>>>>>>>>>>>>>>>>>> in
>>>>>>>>>>>>>>>>>>>>>>>>>>>> one node, it
>>>>>>>>>>>>>>>>>>>>>>>>>>>> also starts on other nodes. Which daemon
>>>>>>>>>>>>>>>>>>>>>>>>>>>> controls
>>>>>>>>>>>>>>>>>>>>>>>>>>>>  this?
>>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part
>>>>>>>>>>>>>>>>>>>>>>>>>>> of
>>>>>>>>>>>>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>>>>>>>>>>>>> rgmanager
>>>>>>>>>>>>>>>>>>>>>>>>>>> package). To me, this sounds like a configuration
>>>>>>>>>>>>>>>>>>>>>>>>>>> problem. Maybe,
>>>>>>>>>>>>>>>>>>>>>>>>>>> you can post your cluster.conf?
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> Regards,
>>>>>>>>>>>>>>>>>>>>>>>>>>> Volker
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Linux-cluster mailing list
>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Linux-cluster mailing list
>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>>
>>>>>>>> --
>>>>>>>> Linux-cluster mailing list
>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>
>>>>>>> --
>>>>>>> Linux-cluster mailing list
>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>
>>>>>> --
>>>>>> Linux-cluster mailing list
>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>
>>>>> --
>>>>> - Daniela Anzellotti ------------------------------------
>>>>>  INFN Roma - tel.: +39.06.49914282 - fax: +39.06.490354
>>>>>  e-mail: daniela.anzellotti@xxxxxxxxxxxxx
>>>>> ---------------------------------------------------------
>>>>>
>>>>> --
>>>>> Linux-cluster mailing list
>>>>> Linux-cluster@xxxxxxxxxx
>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>
>>>> --
>>>> Linux-cluster mailing list
>>>> Linux-cluster@xxxxxxxxxx
>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>
>>> --
>>> - Daniela Anzellotti ------------------------------------
>>>  INFN Roma - tel.: +39.06.49914282 - fax: +39.06.490354
>>>  e-mail: daniela.anzellotti@xxxxxxxxxxxxx
>>> ---------------------------------------------------------
>>>
>>> --
>>> Linux-cluster mailing list
>>> Linux-cluster@xxxxxxxxxx
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>
>>
>> --
>> Linux-cluster mailing list
>> Linux-cluster@xxxxxxxxxx
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
>
> --
> - Daniela Anzellotti ------------------------------------
>  INFN Roma - tel.: +39.06.49914282 - fax: +39.06.490354
>  e-mail: daniela.anzellotti@xxxxxxxxxxxxx
> ---------------------------------------------------------
>
> --
> Linux-cluster mailing list
> Linux-cluster@xxxxxxxxxx
> https://www.redhat.com/mailman/listinfo/linux-cluster
>

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux