Re: openais issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The only thing I noticed is the message after stopping the vm using xm
in all nodes and starting using clusvcadm is

"Virtual machine guest1 is blocked"

The whole DEBUG file is attached.


Thanks
Paras.

On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli
<brem.belguebli@xxxxxxxxx> wrote:
> There's a problem with the script that is called by rgmanager to start
> the VM, I don't know what causes it
>
> May be you should try something like :
>
> 1) stop the VM on all nodes with xm commands
> 2) edit the /usr/share/cluster/vm.sh script and add the following
> lines (after the #!/bin/bash ):
>   exec >/tmp/DEBUG 2>&1
>   set -x
> 3) start the VM with clusvcadm -e vm:guest1
>
> It should fail as it did before.
>
> edit the the /tmp/DEBUG file and you will be able to see where it
> fails (it may generate a lot of debug)
>
> 4) remove the debug lines from /usr/share/cluster/vm.sh
>
> Post the DEBUG file if you're not able to see where it fails.
>
> Brem
>
> 2009/9/26 Paras pradhan <pradhanparas@xxxxxxxxx>:
>> No I am not manually starting not using automatic init scripts.
>>
>> I started the vm using: clusvcadm -e vm:guest1
>>
>> I have just stopped using clusvcadm -s vm:guest1. For few seconds it
>> says guest1 started . But after a while I can see the guest1 on all
>> three nodes.
>>
>> clustat says:
>>
>>  Service Name                                            Owner (Last)
>>                                          State
>>  ------- ----                                            ----- ------
>>                                          -----
>>  vm:guest1                                               (none)
>>                                          stopped
>>
>> But I can see the vm from xm li.
>>
>> This is what I can see from the log:
>>
>>
>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1"
>> returned 1 (generic error)
>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start
>> vm:guest1; return value: 1
>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1
>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is
>> recovering
>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed
>> service vm:guest1
>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1"
>> returned 1 (generic error)
>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start
>> vm:guest1; return value: 1
>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1
>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is
>> recovering
>>
>>
>> Paras.
>>
>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli
>> <brem.belguebli@xxxxxxxxx> wrote:
>>> Have you started  your VM via rgmanager (clusvcadm -e vm:guest1) or
>>> using xm commands out of cluster control  (or maybe a thru an
>>> automatic init script ?)
>>>
>>> When clustered, you should never be starting services (manually or
>>> thru automatic init script) out of cluster control
>>>
>>> The thing would be to stop your vm on all the nodes with the adequate
>>> xm command (not using xen myself) and try to start it with clusvcadm.
>>>
>>> Then see if it is started on all nodes (send clustat output)
>>>
>>>
>>>
>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>> Ok. Please see below. my vm is running on all nodes though clustat
>>>> says it is stopped.
>>>>
>>>> --
>>>> [root@cvtst1 ~]# clustat
>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009
>>>> Member Status: Quorate
>>>>
>>>>  Member Name                                                     ID   Status
>>>>  ------ ----                                                     ---- ------
>>>>  cvtst2                                                    1 Online, rgmanager
>>>>  cvtst1                                                     2 Online,
>>>> Local, rgmanager
>>>>  cvtst3                                                     3 Online, rgmanager
>>>>
>>>>  Service Name                                            Owner (Last)
>>>>                                          State
>>>>  ------- ----                                            ----- ------
>>>>                                          -----
>>>>  vm:guest1                                               (none)
>>>>                                          stopped
>>>> [root@cvtst1 ~]#
>>>>
>>>>
>>>> ---
>>>> o/p of xm li on cvtst1
>>>>
>>>> --
>>>> [root@cvtst1 ~]# xm li
>>>> Name                                      ID Mem(MiB) VCPUs State   Time(s)
>>>> Domain-0                                   0     3470     2 r-----  28939.4
>>>> guest1                                     7      511     1 -b----   7727.8
>>>>
>>>> o/p of xm li on cvtst2
>>>>
>>>> --
>>>> [root@cvtst2 ~]# xm li
>>>> Name                                      ID Mem(MiB) VCPUs State   Time(s)
>>>> Domain-0                                   0     3470     2 r-----  31558.9
>>>> guest1                                    21      511     1 -b----   7558.2
>>>> ---
>>>>
>>>> Thanks
>>>> Paras.
>>>>
>>>>
>>>>
>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli
>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>> It looks like no.
>>>>>
>>>>> can you send an output of clustat  of when the VM is running on
>>>>> multiple nodes at the same time?
>>>>>
>>>>> And by the way, another one after having stopped (clusvcadm -s vm:guest1) ?
>>>>>
>>>>>
>>>>>
>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>>> Anyone having issue as mine? Virtual machine service is not being
>>>>>> properly handled by the cluster.
>>>>>>
>>>>>>
>>>>>> Thanks
>>>>>> Paras.
>>>>>>
>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan <pradhanparas@xxxxxxxxx> wrote:
>>>>>>> Ok.. here is my cluster.conf file
>>>>>>>
>>>>>>> --
>>>>>>> [root@cvtst1 cluster]# more cluster.conf
>>>>>>> <?xml version="1.0"?>
>>>>>>> <cluster alias="test" config_version="9" name="test">
>>>>>>>        <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/>
>>>>>>>        <clusternodes>
>>>>>>>                <clusternode name="cvtst2" nodeid="1" votes="1">
>>>>>>>                        <fence/>
>>>>>>>                </clusternode>
>>>>>>>                <clusternode name="cvtst1" nodeid="2" votes="1">
>>>>>>>                        <fence/>
>>>>>>>                </clusternode>
>>>>>>>                <clusternode name="cvtst3" nodeid="3" votes="1">
>>>>>>>                        <fence/>
>>>>>>>                </clusternode>
>>>>>>>        </clusternodes>
>>>>>>>        <cman/>
>>>>>>>        <fencedevices/>
>>>>>>>        <rm>
>>>>>>>                <failoverdomains>
>>>>>>>                        <failoverdomain name="myfd1" nofailback="0" ordered="1" restricted="0">
>>>>>>>                                <failoverdomainnode name="cvtst2" priority="3"/>
>>>>>>>                                <failoverdomainnode name="cvtst1" priority="1"/>
>>>>>>>                                <failoverdomainnode name="cvtst3" priority="2"/>
>>>>>>>                        </failoverdomain>
>>>>>>>                </failoverdomains>
>>>>>>>                <resources/>
>>>>>>>                <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0"
>>>>>>> name="guest1" path="/vms" recovery="r
>>>>>>> estart" restart_expire_time="0"/>
>>>>>>>        </rm>
>>>>>>> </cluster>
>>>>>>> [root@cvtst1 cluster]#
>>>>>>> ------
>>>>>>>
>>>>>>> Thanks!
>>>>>>> Paras.
>>>>>>>
>>>>>>>
>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer <volker@xxxxxxxxxxxx> wrote:
>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500,
>>>>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote:
>>>>>>>>> I am using cluster suite for HA of xen virtual machines. Now I am
>>>>>>>>> having another problem. When I start the my xen vm in one node, it
>>>>>>>>> also starts on other nodes. Which daemon controls  this?
>>>>>>>>
>>>>>>>> This is usually done bei clurgmgrd (which is part of the rgmanager
>>>>>>>> package). To me, this sounds like a configuration problem. Maybe,
>>>>>>>> you can post your cluster.conf?
>>>>>>>>
>>>>>>>> Regards,
>>>>>>>> Volker
>>>>>>>>
>>>>>>>> --
>>>>>>>> Linux-cluster mailing list
>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Linux-cluster mailing list
>>>>>> Linux-cluster@xxxxxxxxxx
>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>>
>>>>>
>>>>> --
>>>>> Linux-cluster mailing list
>>>>> Linux-cluster@xxxxxxxxxx
>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>
>>>>
>>>> --
>>>> Linux-cluster mailing list
>>>> Linux-cluster@xxxxxxxxxx
>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>
>>>
>>> --
>>> Linux-cluster mailing list
>>> Linux-cluster@xxxxxxxxxx
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>
>>
>> --
>> Linux-cluster mailing list
>> Linux-cluster@xxxxxxxxxx
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
>
> --
> Linux-cluster mailing list
> Linux-cluster@xxxxxxxxxx
> https://www.redhat.com/mailman/listinfo/linux-cluster
>

Attachment: DEBUG
Description: Binary data

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux