Re: openais issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Good news,

the use_virsh=0 parameter is something that I have read somewhere. I
don't know if it was due to a bug, or anything else, and if it is
corrected.

As I said to Paras, I have no expertise on Xen setups.

Brem

2009/10/5, Daniela Anzellotti <daniela.anzellotti@xxxxxxxxxxxxx>:
> Hi all,
>
> I had a problem similar to Paras's one today: yum updated the following rpms
> last week and today (I had to restart the cluster) the cluster was not able
> to start vm: services.
>
> Oct 02 05:31:05 Updated: openais-0.80.6-8.el5.x86_64
> Oct 02 05:31:07 Updated: cman-2.0.115-1.el5.x86_64
> Oct 02 05:31:10 Updated: rgmanager-2.0.52-1.el5.x86_64
>
> Oct 03 04:03:12 Updated: xen-libs-3.0.3-94.el5_4.1.x86_64
> Oct 03 04:03:12 Updated: xen-libs-3.0.3-94.el5_4.1.i386
> Oct 03 04:03:16 Updated: xen-3.0.3-94.el5_4.1.x86_64
>
>
> So, after checked the vm.sh script, I add the declaration use_virsh="0" in
> the VM definition in the cluster.conf (as suggested by Brem, thanks!) and
> everything is now working again.
>
>
> BTW I didn't understand if the problem was caused by the new XEN version or
> the new openais one, thus I disabled automatic updates for both.
>
> I hope I'll not have any other bad surprise...
>
> Thank you,
> cheers,
> Daniela
>
>
>
> Paras pradhan wrote:
> > Yes this is very strange. I don't know what to do now. May be re
> > create the cluster? But not a good solution actually.
> >
> > Packages :
> >
> > Kernel: kernel-xen-2.6.18-164.el5
> > OS: Full updated of CentOS 5.3 except CMAN downgraded to cman-2.0.98-1.el5
> >
> > Other packages related to cluster suite:
> >
> > rgmanager-2.0.52-1.el5.centos
> > cman-2.0.98-1.el5
> > xen-3.0.3-80.el5_3.3
> > xen-libs-3.0.3-80.el5_3.3
> > kmod-gfs-xen-0.1.31-3.el5_3.1
> > kmod-gfs-xen-0.1.31-3.el5_3.1
> > kmod-gfs-0.1.31-3.el5_3.1
> > gfs-utils-0.1.18-1.el5
> > gfs2-utils-0.1.62-1.el5
> > lvm2-2.02.40-6.el5
> > lvm2-cluster-2.02.40-7.el5
> > openais-0.80.3-22.el5_3.9
> >
> > Thanks!
> > Paras.
> >
> >
> >
> >
> > On Wed, Sep 30, 2009 at 10:02 AM, brem belguebli
> > <brem.belguebli@xxxxxxxxx> wrote:
> >
> > > Hi Paras,
> > >
> > > Your cluster.conf file seems correct. If it is not a ntp issue, I
> > > don't see anything except a bug that causes this, or some prerequisite
> > > that is not respected.
> > >
> > > May be you could post the versions (os, kernel, packages etc...) you
> > > are using, someone may have hit the same issue with your versions.
> > >
> > > Brem
> > >
> > > 2009/9/30, Paras pradhan <pradhanparas@xxxxxxxxx>:
> > >
> > > > All of the nodes are synced with ntp server. So this is not the case
> with me.
> > > >
> > > > Thanks
> > > > Paras.
> > > >
> > > > On Tue, Sep 29, 2009 at 6:29 PM, Johannes Rußek
> > > > <johannes.russek@xxxxxxxxxxxxxxxxx> wrote:
> > > >
> > > > > make sure the time on the nodes is in sync, apparently when a node
> has too
> > > > > much offset, you won't see rgmanager (even though the process is
> running).
> > > > > this happened today and setting the time fixed it for me. afaicr
> there was
> > > > > no sign of this in the logs though.
> > > > > johannes
> > > > >
> > > > > Paras pradhan schrieb:
> > > > >
> > > > > > I don't see rgmanager .
> > > > > >
> > > > > > Here is the o/p from clustat
> > > > > >
> > > > > > [root@cvtst1 cluster]# clustat
> > > > > > Cluster Status for test @ Tue Sep 29 15:53:33 2009
> > > > > > Member Status: Quorate
> > > > > >
> > > > > >  Member Name
> ID
> > > > > > Status
> > > > > >  ------ ----
> ----
> > > > > > ------
> > > > > >  cvtst2                                                    1
> Online
> > > > > >  cvtst1                                                     2
> Online,
> > > > > > Local
> > > > > >  cvtst3                                                     3
> Online
> > > > > >
> > > > > >
> > > > > > Thanks
> > > > > > Paras.
> > > > > >
> > > > > > On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli
> > > > > > <brem.belguebli@xxxxxxxxx> wrote:
> > > > > >
> > > > > >
> > > > > > > It looks correct, rgmanager seems to start on all nodes
> > > > > > >
> > > > > > > what gives you clustat ?
> > > > > > >
> > > > > > > If rgmanager doesn't show, check out the logs something may have
> gone
> > > > > > > wrong.
> > > > > > >
> > > > > > >
> > > > > > > 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>:
> > > > > > >
> > > > > > >
> > > > > > > > Change to 7 and i got this log
> > > > > > > >
> > > > > > > > Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting
> down
> > > > > > > > Cluster Service Manager...
> > > > > > > > Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting
> down
> > > > > > > > Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting
> down
> > > > > > > > Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown
> complete,
> > > > > > > > exiting
> > > > > > > > Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster
> Service
> > > > > > > > Manager is stopped.
> > > > > > > > Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource
> Group
> > > > > > > > Manager Starting
> > > > > > > > Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading
> Service Data
> > > > > > > > Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading
> Resource Rules
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules
> loaded
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building
> Resource Trees
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources
> defined
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading
> Failover
> > > > > > > > Domains
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains
> defined
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events
> defined
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing
> Services
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services
> Initialized
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port
> Opened
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change:
> Local UP
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change:
> cvtst2 UP
> > > > > > > > Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change:
> cvtst3 UP
> > > > > > > > Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1)
> Processed
> > > > > > > > Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1)
> Processed
> > > > > > > > Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1)
> Processed
> > > > > > > > Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events
> processed
> > > > > > > >
> > > > > > > >
> > > > > > > > Anything unusual here?
> > > > > > > >
> > > > > > > > Paras.
> > > > > > > >
> > > > > > > > On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli
> > > > > > > > <brem.belguebli@xxxxxxxxx> wrote:
> > > > > > > >
> > > > > > > >
> > > > > > > > > I use log_level=7 to have more debugging info.
> > > > > > > > >
> > > > > > > > > It seems 4 is not enough.
> > > > > > > > >
> > > > > > > > > Brem
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > 2009/9/29, Paras pradhan <pradhanparas@xxxxxxxxx>:
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > > Withe log_level of 3 I got only this
> > > > > > > > > >
> > > > > > > > > > Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice>
> Shutting down
> > > > > > > > > > Cluster Service Manager...
> > > > > > > > > > Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting
> down
> > > > > > > > > > Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown
> complete,
> > > > > > > > > > exiting
> > > > > > > > > > Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster
> Service
> > > > > > > > > > Manager is stopped.
> > > > > > > > > > Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource
> Group
> > > > > > > > > > Manager Starting
> > > > > > > > > > Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice>
> Shutting down
> > > > > > > > > > Cluster Service Manager...
> > > > > > > > > > Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice>
> Cluster Service
> > > > > > > > > > Manager is stopped.
> > > > > > > > > > Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource
> Group
> > > > > > > > > > Manager Starting
> > > > > > > > > > Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1
> shutting
> > > > > > > > > > down
> > > > > > > > > >
> > > > > > > > > > I do not know what the last line means.
> > > > > > > > > >
> > > > > > > > > > rgmanager version I am running is:
> > > > > > > > > > rgmanager-2.0.52-1.el5.centos
> > > > > > > > > >
> > > > > > > > > > I don't what has gone wrong.
> > > > > > > > > >
> > > > > > > > > > Thanks
> > > > > > > > > > Paras.
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli
> > > > > > > > > > <brem.belguebli@xxxxxxxxx> wrote:
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > > you mean it stopped successfully on all the nodes but it
> is failing
> > > > > > > > > > > to
> > > > > > > > > > > start only on node cvtst1 ?
> > > > > > > > > > >
> > > > > > > > > > > look at the following page  to make rgmanager more
> verbose. It 'll
> > > > > > > > > > > help debug....
> > > > > > > > > > >
> > > > > > > > > > >
> http://sources.redhat.com/cluster/wiki/RGManager
> > > > > > > > > > >
> > > > > > > > > > > at Logging Configuration section
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>:
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > > Brem,
> > > > > > > > > > > >
> > > > > > > > > > > > When I try to restart rgmanager on all the nodes, this
> time i do not
> > > > > > > > > > > > see rgmanager running on the first node. But I do see
> on other 2
> > > > > > > > > > > > nodes.
> > > > > > > > > > > >
> > > > > > > > > > > > Log on the first node:
> > > > > > > > > > > >
> > > > > > > > > > > > Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice>
> Resource Group
> > > > > > > > > > > > Manager Starting
> > > > > > > > > > > > Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice>
> Shutting down
> > > > > > > > > > > > Cluster Service Manager...
> > > > > > > > > > > > Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice>
> Shutting down
> > > > > > > > > > > > Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice>
> Shutdown complete,
> > > > > > > > > > > > exiting
> > > > > > > > > > > > Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice>
> Cluster Service
> > > > > > > > > > > > Manager is stopped.
> > > > > > > > > > > > Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice>
> Resource Group
> > > > > > > > > > > > Manager Starting
> > > > > > > > > > > >
> > > > > > > > > > > > -
> > > > > > > > > > > > It seems service is running ,  but I do not see
> rgmanger running
> > > > > > > > > > > > using clustat
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > Don't know what is going on.
> > > > > > > > > > > >
> > > > > > > > > > > > Thanks
> > > > > > > > > > > > Paras.
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli
> > > > > > > > > > > > <brem.belguebli@xxxxxxxxx> wrote:
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > > Paras,
> > > > > > > > > > > > >
> > > > > > > > > > > > > Another thing, it would have been more interesting
> to have a start
> > > > > > > > > > > > > DEBUG not a stop.
> > > > > > > > > > > > >
> > > > > > > > > > > > > That's why I was asking you to first stop the vm
> manually on all
> > > > > > > > > > > > > your
> > > > > > > > > > > > > nodes, stop eventually rgmanager on all the nodes to
> reset the
> > > > > > > > > > > > > potential wrong states you may have, restart
> rgmanager.
> > > > > > > > > > > > >
> > > > > > > > > > > > > If your VM is configured to autostart, this will
> make it start.
> > > > > > > > > > > > >
> > > > > > > > > > > > > It should normally fail (as it does now). Send out
> your newly
> > > > > > > > > > > > > created
> > > > > > > > > > > > > DEBUG file.
> > > > > > > > > > > > >
> > > > > > > > > > > > > 2009/9/29 brem belguebli <brem.belguebli@xxxxxxxxx>:
> > > > > > > > > > > > >
> > > > > > > > > > > > >
> > > > > > > > > > > > > > Hi Paras,
> > > > > > > > > > > > > >
> > > > > > > > > > > > > >
> > > > > > > > > > > > > > I don't know the xen/cluster combination well, but
> if I do
> > > > > > > > > > > > > > remember
> > > > > > > > > > > > > > well, I think I've read somewhere that when using
> xen you have to
> > > > > > > > > > > > > > declare the use_virsh=0 key in the VM definition
> in the
> > > > > > > > > > > > > > cluster.conf.
> > > > > > > > > > > > > >
> > > > > > > > > > > > > > This would make rgmanager use xm commands instead
> of virsh
> > > > > > > > > > > > > > The DEBUG output shows clearly that you are using
> virsh to manage
> > > > > > > > > > > > > > your
> > > > > > > > > > > > > > VM instead of xm commands.
> > > > > > > > > > > > > > Check out the RH docs about virtualization
> > > > > > > > > > > > > >
> > > > > > > > > > > > > > I'm not a 100% sure about that, I may be
> completely wrong.
> > > > > > > > > > > > > >
> > > > > > > > > > > > > > Brem
> > > > > > > > > > > > > >
> > > > > > > > > > > > > > 2009/9/28 Paras pradhan <pradhanparas@xxxxxxxxx>:
> > > > > > > > > > > > > >
> > > > > > > > > > > > > >
> > > > > > > > > > > > > > > The only thing I noticed is the message after
> stopping the vm
> > > > > > > > > > > > > > > using xm
> > > > > > > > > > > > > > > in all nodes and starting using clusvcadm is
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > "Virtual machine guest1 is blocked"
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > The whole DEBUG file is attached.
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > Thanks
> > > > > > > > > > > > > > > Paras.
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli
> > > > > > > > > > > > > > > <brem.belguebli@xxxxxxxxx> wrote:
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > There's a problem with the script that is
> called by rgmanager to
> > > > > > > > > > > > > > > > start
> > > > > > > > > > > > > > > > the VM, I don't know what causes it
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > May be you should try something like :
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > 1) stop the VM on all nodes with xm commands
> > > > > > > > > > > > > > > > 2) edit the /usr/share/cluster/vm.sh script
> and add the
> > > > > > > > > > > > > > > > following
> > > > > > > > > > > > > > > > lines (after the #!/bin/bash ):
> > > > > > > > > > > > > > > >  exec >/tmp/DEBUG 2>&1
> > > > > > > > > > > > > > > >  set -x
> > > > > > > > > > > > > > > > 3) start the VM with clusvcadm -e vm:guest1
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > It should fail as it did before.
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > edit the the /tmp/DEBUG file and you will be
> able to see where
> > > > > > > > > > > > > > > > it
> > > > > > > > > > > > > > > > fails (it may generate a lot of debug)
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > 4) remove the debug lines from
> /usr/share/cluster/vm.sh
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > Post the DEBUG file if you're not able to see
> where it fails.
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > Brem
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > 2009/9/26 Paras pradhan
> <pradhanparas@xxxxxxxxx>:
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > No I am not manually starting not using
> automatic init scripts.
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > I started the vm using: clusvcadm -e
> vm:guest1
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > I have just stopped using clusvcadm -s
> vm:guest1. For few
> > > > > > > > > > > > > > > > > seconds it
> > > > > > > > > > > > > > > > > says guest1 started . But after a while I
> can see the guest1 on
> > > > > > > > > > > > > > > > > all
> > > > > > > > > > > > > > > > > three nodes.
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > clustat says:
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > >  Service Name
>             Owner
> > > > > > > > > > > > > > > > > (Last)
> > > > > > > > > > > > > > > > >                                        State
> > > > > > > > > > > > > > > > >  ------- ----
>             -----
> > > > > > > > > > > > > > > > > ------
> > > > > > > > > > > > > > > > >                                        -----
> > > > > > > > > > > > > > > > >  vm:guest1
>             (none)
> > > > > > > > > > > > > > > > >
> stopped
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > But I can see the vm from xm li.
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > This is what I can see from the log:
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > Sep 25 17:19:01 cvtst1 clurgmgrd[4298]:
> <notice> start on vm
> > > > > > > > > > > > > > > > > "guest1"
> > > > > > > > > > > > > > > > > returned 1 (generic error)
> > > > > > > > > > > > > > > > > Sep 25 17:19:01 cvtst1 clurgmgrd[4298]:
> <warning> #68: Failed
> > > > > > > > > > > > > > > > > to start
> > > > > > > > > > > > > > > > > vm:guest1; return value: 1
> > > > > > > > > > > > > > > > > Sep 25 17:19:01 cvtst1 clurgmgrd[4298]:
> <notice> Stopping
> > > > > > > > > > > > > > > > > service vm:guest1
> > > > > > > > > > > > > > > > > Sep 25 17:19:02 cvtst1 clurgmgrd[4298]:
> <notice> Service
> > > > > > > > > > > > > > > > > vm:guest1 is
> > > > > > > > > > > > > > > > > recovering
> > > > > > > > > > > > > > > > > Sep 25 17:19:15 cvtst1 clurgmgrd[4298]:
> <notice> Recovering
> > > > > > > > > > > > > > > > > failed
> > > > > > > > > > > > > > > > > service vm:guest1
> > > > > > > > > > > > > > > > > Sep 25 17:19:16 cvtst1 clurgmgrd[4298]:
> <notice> start on vm
> > > > > > > > > > > > > > > > > "guest1"
> > > > > > > > > > > > > > > > > returned 1 (generic error)
> > > > > > > > > > > > > > > > > Sep 25 17:19:16 cvtst1 clurgmgrd[4298]:
> <warning> #68: Failed
> > > > > > > > > > > > > > > > > to start
> > > > > > > > > > > > > > > > > vm:guest1; return value: 1
> > > > > > > > > > > > > > > > > Sep 25 17:19:16 cvtst1 clurgmgrd[4298]:
> <notice> Stopping
> > > > > > > > > > > > > > > > > service vm:guest1
> > > > > > > > > > > > > > > > > Sep 25 17:19:17 cvtst1 clurgmgrd[4298]:
> <notice> Service
> > > > > > > > > > > > > > > > > vm:guest1 is
> > > > > > > > > > > > > > > > > recovering
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > Paras.
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > On Fri, Sep 25, 2009 at 5:07 PM, brem
> belguebli
> > > > > > > > > > > > > > > > > <brem.belguebli@xxxxxxxxx> wrote:
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > Have you started  your VM via rgmanager
> (clusvcadm -e
> > > > > > > > > > > > > > > > > > vm:guest1) or
> > > > > > > > > > > > > > > > > > using xm commands out of cluster control
> (or maybe a thru an
> > > > > > > > > > > > > > > > > > automatic init script ?)
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > When clustered, you should never be
> starting services
> > > > > > > > > > > > > > > > > > (manually or
> > > > > > > > > > > > > > > > > > thru automatic init script) out of cluster
> control
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > The thing would be to stop your vm on all
> the nodes with the
> > > > > > > > > > > > > > > > > > adequate
> > > > > > > > > > > > > > > > > > xm command (not using xen myself) and try
> to start it with
> > > > > > > > > > > > > > > > > > clusvcadm.
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > Then see if it is started on all nodes
> (send clustat output)
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > 2009/9/25 Paras pradhan
> <pradhanparas@xxxxxxxxx>:
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > Ok. Please see below. my vm is running
> on all nodes though
> > > > > > > > > > > > > > > > > > > clustat
> > > > > > > > > > > > > > > > > > > says it is stopped.
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > > > [root@cvtst1 ~]# clustat
> > > > > > > > > > > > > > > > > > > Cluster Status for test @ Fri Sep 25
> 16:52:34 2009
> > > > > > > > > > > > > > > > > > > Member Status: Quorate
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > >  Member Name
> > > > > > > > > > > > > > > > > > >    ID   Status
> > > > > > > > > > > > > > > > > > >  ------ ----
> > > > > > > > > > > > > > > > > > >    ---- ------
> > > > > > > > > > > > > > > > > > >  cvtst2
>                   1
> > > > > > > > > > > > > > > > > > > Online, rgmanager
> > > > > > > > > > > > > > > > > > >  cvtst1
>                    2
> > > > > > > > > > > > > > > > > > > Online,
> > > > > > > > > > > > > > > > > > > Local, rgmanager
> > > > > > > > > > > > > > > > > > >  cvtst3
>                    3
> > > > > > > > > > > > > > > > > > > Online, rgmanager
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > >  Service Name
> > > > > > > > > > > > > > > > > > >  Owner (Last)
> > > > > > > > > > > > > > > > > > >
> State
> > > > > > > > > > > > > > > > > > >  ------- ----
> > > > > > > > > > > > > > > > > > >  ----- ------
> > > > > > > > > > > > > > > > > > >
> -----
> > > > > > > > > > > > > > > > > > >  vm:guest1
> > > > > > > > > > > > > > > > > > > (none)
> > > > > > > > > > > > > > > > > > >
> stopped
> > > > > > > > > > > > > > > > > > > [root@cvtst1 ~]#
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > ---
> > > > > > > > > > > > > > > > > > > o/p of xm li on cvtst1
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > > > [root@cvtst1 ~]# xm li
> > > > > > > > > > > > > > > > > > > Name
>  ID Mem(MiB) VCPUs
> > > > > > > > > > > > > > > > > > > State   Time(s)
> > > > > > > > > > > > > > > > > > > Domain-0
>   0     3470     2
> > > > > > > > > > > > > > > > > > > r-----  28939.4
> > > > > > > > > > > > > > > > > > > guest1
>   7      511     1
> > > > > > > > > > > > > > > > > > > -b----   7727.8
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > o/p of xm li on cvtst2
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > > > [root@cvtst2 ~]# xm li
> > > > > > > > > > > > > > > > > > > Name
>  ID Mem(MiB) VCPUs
> > > > > > > > > > > > > > > > > > > State   Time(s)
> > > > > > > > > > > > > > > > > > > Domain-0
>   0     3470     2
> > > > > > > > > > > > > > > > > > > r-----  31558.9
> > > > > > > > > > > > > > > > > > > guest1
>  21      511     1
> > > > > > > > > > > > > > > > > > > -b----   7558.2
> > > > > > > > > > > > > > > > > > > ---
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > Thanks
> > > > > > > > > > > > > > > > > > > Paras.
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > On Fri, Sep 25, 2009 at 4:22 PM, brem
> belguebli
> > > > > > > > > > > > > > > > > > > <brem.belguebli@xxxxxxxxx> wrote:
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > It looks like no.
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > can you send an output of clustat  of
> when the VM is running
> > > > > > > > > > > > > > > > > > > > on
> > > > > > > > > > > > > > > > > > > > multiple nodes at the same time?
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > And by the way, another one after
> having stopped (clusvcadm
> > > > > > > > > > > > > > > > > > > > -s vm:guest1) ?
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > 2009/9/25 Paras pradhan
> <pradhanparas@xxxxxxxxx>:
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > Anyone having issue as mine? Virtual
> machine service is not
> > > > > > > > > > > > > > > > > > > > > being
> > > > > > > > > > > > > > > > > > > > > properly handled by the cluster.
> > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > Thanks
> > > > > > > > > > > > > > > > > > > > > Paras.
> > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > On Mon, Sep 21, 2009 at 9:55 AM,
> Paras pradhan
> > > > > > > > > > > > > > > > > > > > > <pradhanparas@xxxxxxxxx> wrote:
> > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > Ok.. here is my cluster.conf file
> > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > > > > > > [root@cvtst1 cluster]# more
> cluster.conf
> > > > > > > > > > > > > > > > > > > > > > <?xml version="1.0"?>
> > > > > > > > > > > > > > > > > > > > > > <cluster alias="test"
> config_version="9" name="test">
> > > > > > > > > > > > > > > > > > > > > >      <fence_daemon clean_start="0"
> post_fail_delay="0"
> > > > > > > > > > > > > > > > > > > > > > post_join_delay="3"/>
> > > > > > > > > > > > > > > > > > > > > >      <clusternodes>
> > > > > > > > > > > > > > > > > > > > > >              <clusternode
> name="cvtst2" nodeid="1"
> > > > > > > > > > > > > > > > > > > > > > votes="1">
> > > > > > > > > > > > > > > > > > > > > >                      <fence/>
> > > > > > > > > > > > > > > > > > > > > >              </clusternode>
> > > > > > > > > > > > > > > > > > > > > >              <clusternode
> name="cvtst1" nodeid="2"
> > > > > > > > > > > > > > > > > > > > > > votes="1">
> > > > > > > > > > > > > > > > > > > > > >                      <fence/>
> > > > > > > > > > > > > > > > > > > > > >              </clusternode>
> > > > > > > > > > > > > > > > > > > > > >              <clusternode
> name="cvtst3" nodeid="3"
> > > > > > > > > > > > > > > > > > > > > > votes="1">
> > > > > > > > > > > > > > > > > > > > > >                      <fence/>
> > > > > > > > > > > > > > > > > > > > > >              </clusternode>
> > > > > > > > > > > > > > > > > > > > > >      </clusternodes>
> > > > > > > > > > > > > > > > > > > > > >      <cman/>
> > > > > > > > > > > > > > > > > > > > > >      <fencedevices/>
> > > > > > > > > > > > > > > > > > > > > >      <rm>
> > > > > > > > > > > > > > > > > > > > > >              <failoverdomains>
> > > > > > > > > > > > > > > > > > > > > >
> <failoverdomain name="myfd1"
> > > > > > > > > > > > > > > > > > > > > > nofailback="0" ordered="1"
> restricted="0">
> > > > > > > > > > > > > > > > > > > > > >
> <failoverdomainnode
> > > > > > > > > > > > > > > > > > > > > > name="cvtst2" priority="3"/>
> > > > > > > > > > > > > > > > > > > > > >
> <failoverdomainnode
> > > > > > > > > > > > > > > > > > > > > > name="cvtst1" priority="1"/>
> > > > > > > > > > > > > > > > > > > > > >
> <failoverdomainnode
> > > > > > > > > > > > > > > > > > > > > > name="cvtst3" priority="2"/>
> > > > > > > > > > > > > > > > > > > > > >
> </failoverdomain>
> > > > > > > > > > > > > > > > > > > > > >              </failoverdomains>
> > > > > > > > > > > > > > > > > > > > > >              <resources/>
> > > > > > > > > > > > > > > > > > > > > >              <vm autostart="1"
> domain="myfd1"
> > > > > > > > > > > > > > > > > > > > > > exclusive="0" max_restarts="0"
> > > > > > > > > > > > > > > > > > > > > > name="guest1" path="/vms"
> recovery="r
> > > > > > > > > > > > > > > > > > > > > > estart" restart_expire_time="0"/>
> > > > > > > > > > > > > > > > > > > > > >      </rm>
> > > > > > > > > > > > > > > > > > > > > > </cluster>
> > > > > > > > > > > > > > > > > > > > > > [root@cvtst1 cluster]#
> > > > > > > > > > > > > > > > > > > > > > ------
> > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > Thanks!
> > > > > > > > > > > > > > > > > > > > > > Paras.
> > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > On Sun, Sep 20, 2009 at 9:44 AM,
> Volker Dormeyer
> > > > > > > > > > > > > > > > > > > > > > <volker@xxxxxxxxxxxx> wrote:
> > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > > On Fri, Sep 18, 2009 at
> 05:08:57PM -0500,
> > > > > > > > > > > > > > > > > > > > > > > Paras pradhan
> <pradhanparas@xxxxxxxxx> wrote:
> > > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > > > I am using cluster suite for
> HA of xen virtual machines.
> > > > > > > > > > > > > > > > > > > > > > > > Now I am
> > > > > > > > > > > > > > > > > > > > > > > > having another problem. When I
> start the my xen vm in
> > > > > > > > > > > > > > > > > > > > > > > > one node, it
> > > > > > > > > > > > > > > > > > > > > > > > also starts on other nodes.
> Which daemon controls  this?
> > > > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > > This is usually done bei
> clurgmgrd (which is part of the
> > > > > > > > > > > > > > > > > > > > > > > rgmanager
> > > > > > > > > > > > > > > > > > > > > > > package). To me, this sounds
> like a configuration
> > > > > > > > > > > > > > > > > > > > > > > problem. Maybe,
> > > > > > > > > > > > > > > > > > > > > > > you can post your cluster.conf?
> > > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > > Regards,
> > > > > > > > > > > > > > > > > > > > > > > Volker
> > > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > > > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > --
> > > > > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > > >
> > > > > > > > > > > > > >
> > > > > > > > > > > > > --
> > > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > > >
> > > > > > > > > > > > >
> > > > > > > > > > > > >
> > > > > > > > > > > > --
> > > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > --
> > > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > --
> > > > > > > > > > Linux-cluster mailing list
> > > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > --
> > > > > > > > > Linux-cluster mailing list
> > > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > > >
> > > > > > > > >
> > > > > > > > >
> > > > > > > > --
> > > > > > > > Linux-cluster mailing list
> > > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > --
> > > > > > > Linux-cluster mailing list
> > > > > > > Linux-cluster@xxxxxxxxxx
> > > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > --
> > > > > > Linux-cluster mailing list
> > > > > > Linux-cluster@xxxxxxxxxx
> > > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > > >
> > > > > >
> > > > > --
> > > > > Linux-cluster mailing list
> > > > > Linux-cluster@xxxxxxxxxx
> > > > >
> https://www.redhat.com/mailman/listinfo/linux-cluster
> > > > >
> > > > >
> > > > --
> > > > Linux-cluster mailing list
> > > > Linux-cluster@xxxxxxxxxx
> > > > https://www.redhat.com/mailman/listinfo/linux-cluster
> > > >
> > > >
> > > --
> > > Linux-cluster mailing list
> > > Linux-cluster@xxxxxxxxxx
> > > https://www.redhat.com/mailman/listinfo/linux-cluster
> > >
> > >
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster@xxxxxxxxxx
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> >
> >
>
> --
> - Daniela Anzellotti ------------------------------------
>  INFN Roma - tel.: +39.06.49914282 - fax: +39.06.490354
>  e-mail: daniela.anzellotti@xxxxxxxxxxxxx
> ---------------------------------------------------------
>
>
> --
> Linux-cluster mailing list
> Linux-cluster@xxxxxxxxxx
> https://www.redhat.com/mailman/listinfo/linux-cluster
>

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux