Re: openais issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



That looks strange, maybe your cluster.conf will help

2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>:
> I don't see rgmanager .
>
> Here is the o/p from clustat
>
> [root@cvtst1 cluster]# clustat
> Cluster Status for test @ Tue Sep 29 15:53:33 2009
> Member Status: Quorate
>
>  Member Name                                                     ID   Status
>  ------ ----                                                     ---- ------
>  cvtst2                                                    1 Online
>  cvtst1                                                     2 Online, Local
>  cvtst3                                                     3 Online
>
>
> Thanks
> Paras.
>
> On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli
> <brem.belguebli@xxxxxxxxx> wrote:
>> It looks correct, rgmanager seems to start on all nodes
>>
>> what gives you clustat ?
>>
>> If rgmanager doesn't show, check out the logs something may have gone wrong.
>>
>>
>> 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>> Change to 7 and i got this log
>>>
>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down
>>> Cluster Service Manager...
>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down
>>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown complete, exiting
>>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster Service
>>> Manager is stopped.
>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group
>>> Manager Starting
>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service Data
>>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource Rules
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource Trees
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources defined
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover Domains
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing Services
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services Initialized
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port Opened
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: Local UP
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst2 UP
>>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst3 UP
>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1) Processed
>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1) Processed
>>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1) Processed
>>> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events processed
>>>
>>>
>>> Anything unusual here?
>>>
>>> Paras.
>>>
>>> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli
>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>> I use log_level=7 to have more debugging info.
>>>>
>>>> It seems 4 is not enough.
>>>>
>>>> Brem
>>>>
>>>>
>>>> 2009/9/29, Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>> Withe log_level of 3 I got only this
>>>>>
>>>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down
>>>>> Cluster Service Manager...
>>>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down
>>>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, exiting
>>>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service
>>>>> Manager is stopped.
>>>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group
>>>>> Manager Starting
>>>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down
>>>>> Cluster Service Manager...
>>>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service
>>>>> Manager is stopped.
>>>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group
>>>>> Manager Starting
>>>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting down
>>>>>
>>>>> I do not know what the last line means.
>>>>>
>>>>> rgmanager version I am running is:
>>>>> rgmanager-2.0.52-1.el5.centos
>>>>>
>>>>> I don't what has gone wrong.
>>>>>
>>>>> Thanks
>>>>> Paras.
>>>>>
>>>>>
>>>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli
>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>> > you mean it stopped successfully on all the nodes but it is failing to
>>>>> > start only on node cvtst1 ?
>>>>> >
>>>>> > look at the following page  to make rgmanager more verbose. It 'll
>>>>> > help debug....
>>>>> >
>>>>> > http://sources.redhat.com/cluster/wiki/RGManager
>>>>> >
>>>>> > at Logging Configuration section
>>>>> >
>>>>> >
>>>>> >
>>>>> >
>>>>> > 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>> >> Brem,
>>>>> >>
>>>>> >> When I try to restart rgmanager on all the nodes, this time i do not
>>>>> >> see rgmanager running on the first node. But I do see on other 2
>>>>> >> nodes.
>>>>> >>
>>>>> >> Log on the first node:
>>>>> >>
>>>>> >> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group
>>>>> >> Manager Starting
>>>>> >> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down
>>>>> >> Cluster Service Manager...
>>>>> >> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down
>>>>> >> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, exiting
>>>>> >> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service
>>>>> >> Manager is stopped.
>>>>> >> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group
>>>>> >> Manager Starting
>>>>> >>
>>>>> >> -
>>>>> >> It seems service is running ,  but I do not see rgmanger running using clustat
>>>>> >>
>>>>> >>
>>>>> >> Don't know what is going on.
>>>>> >>
>>>>> >> Thanks
>>>>> >> Paras.
>>>>> >>
>>>>> >>
>>>>> >> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli
>>>>> >> <brem.belguebli@xxxxxxxxx> wrote:
>>>>> >>> Paras,
>>>>> >>>
>>>>> >>> Another thing, it would have been more interesting to have a start
>>>>> >>> DEBUG not a stop.
>>>>> >>>
>>>>> >>> That's why I was asking you to first stop the vm manually on all your
>>>>> >>> nodes, stop eventually rgmanager on all the nodes to reset the
>>>>> >>> potential wrong states you may have, restart rgmanager.
>>>>> >>>
>>>>> >>> If your VM is configured to autostart, this will make it start.
>>>>> >>>
>>>>> >>> It should normally fail (as it does now). Send out your newly created
>>>>> >>> DEBUG file.
>>>>> >>>
>>>>> >>> 2009/9/29 brem belguebli <brem.belguebli@xxxxxxxxx>:
>>>>> >>>> Hi Paras,
>>>>> >>>>
>>>>> >>>>
>>>>> >>>> I don't know the xen/cluster combination well, but if I do remember
>>>>> >>>> well, I think I've read somewhere that when using xen you have to
>>>>> >>>> declare the use_virsh=0 key in the VM definition in the cluster.conf.
>>>>> >>>>
>>>>> >>>> This would make rgmanager use xm commands instead of virsh
>>>>> >>>> The DEBUG output shows clearly that you are using virsh to manage your
>>>>> >>>> VM instead of xm commands.
>>>>> >>>> Check out the RH docs about virtualization
>>>>> >>>>
>>>>> >>>> I'm not a 100% sure about that, I may be completely wrong.
>>>>> >>>>
>>>>> >>>> Brem
>>>>> >>>>
>>>>> >>>> 2009/9/28 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>> >>>>> The only thing I noticed is the message after stopping the vm using xm
>>>>> >>>>> in all nodes and starting using clusvcadm is
>>>>> >>>>>
>>>>> >>>>> "Virtual machine guest1 is blocked"
>>>>> >>>>>
>>>>> >>>>> The whole DEBUG file is attached.
>>>>> >>>>>
>>>>> >>>>>
>>>>> >>>>> Thanks
>>>>> >>>>> Paras.
>>>>> >>>>>
>>>>> >>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli
>>>>> >>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>> >>>>>> There's a problem with the script that is called by rgmanager to start
>>>>> >>>>>> the VM, I don't know what causes it
>>>>> >>>>>>
>>>>> >>>>>> May be you should try something like :
>>>>> >>>>>>
>>>>> >>>>>> 1) stop the VM on all nodes with xm commands
>>>>> >>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the following
>>>>> >>>>>> lines (after the #!/bin/bash ):
>>>>> >>>>>>   exec >/tmp/DEBUG 2>&1
>>>>> >>>>>>   set -x
>>>>> >>>>>> 3) start the VM with clusvcadm -e vm:guest1
>>>>> >>>>>>
>>>>> >>>>>> It should fail as it did before.
>>>>> >>>>>>
>>>>> >>>>>> edit the the /tmp/DEBUG file and you will be able to see where it
>>>>> >>>>>> fails (it may generate a lot of debug)
>>>>> >>>>>>
>>>>> >>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh
>>>>> >>>>>>
>>>>> >>>>>> Post the DEBUG file if you're not able to see where it fails.
>>>>> >>>>>>
>>>>> >>>>>> Brem
>>>>> >>>>>>
>>>>> >>>>>> 2009/9/26 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>> >>>>>>> No I am not manually starting not using automatic init scripts.
>>>>> >>>>>>>
>>>>> >>>>>>> I started the vm using: clusvcadm -e vm:guest1
>>>>> >>>>>>>
>>>>> >>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few seconds it
>>>>> >>>>>>> says guest1 started . But after a while I can see the guest1 on all
>>>>> >>>>>>> three nodes.
>>>>> >>>>>>>
>>>>> >>>>>>> clustat says:
>>>>> >>>>>>>
>>>>> >>>>>>>  Service Name                                            Owner (Last)
>>>>> >>>>>>>                                          State
>>>>> >>>>>>>  ------- ----                                            ----- ------
>>>>> >>>>>>>                                          -----
>>>>> >>>>>>>  vm:guest1                                               (none)
>>>>> >>>>>>>                                          stopped
>>>>> >>>>>>>
>>>>> >>>>>>> But I can see the vm from xm li.
>>>>> >>>>>>>
>>>>> >>>>>>> This is what I can see from the log:
>>>>> >>>>>>>
>>>>> >>>>>>>
>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1"
>>>>> >>>>>>> returned 1 (generic error)
>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start
>>>>> >>>>>>> vm:guest1; return value: 1
>>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1
>>>>> >>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is
>>>>> >>>>>>> recovering
>>>>> >>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed
>>>>> >>>>>>> service vm:guest1
>>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1"
>>>>> >>>>>>> returned 1 (generic error)
>>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start
>>>>> >>>>>>> vm:guest1; return value: 1
>>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1
>>>>> >>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is
>>>>> >>>>>>> recovering
>>>>> >>>>>>>
>>>>> >>>>>>>
>>>>> >>>>>>> Paras.
>>>>> >>>>>>>
>>>>> >>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli
>>>>> >>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>> >>>>>>>> Have you started  your VM via rgmanager (clusvcadm -e vm:guest1) or
>>>>> >>>>>>>> using xm commands out of cluster control  (or maybe a thru an
>>>>> >>>>>>>> automatic init script ?)
>>>>> >>>>>>>>
>>>>> >>>>>>>> When clustered, you should never be starting services (manually or
>>>>> >>>>>>>> thru automatic init script) out of cluster control
>>>>> >>>>>>>>
>>>>> >>>>>>>> The thing would be to stop your vm on all the nodes with the adequate
>>>>> >>>>>>>> xm command (not using xen myself) and try to start it with clusvcadm.
>>>>> >>>>>>>>
>>>>> >>>>>>>> Then see if it is started on all nodes (send clustat output)
>>>>> >>>>>>>>
>>>>> >>>>>>>>
>>>>> >>>>>>>>
>>>>> >>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>> >>>>>>>>> Ok. Please see below. my vm is running on all nodes though clustat
>>>>> >>>>>>>>> says it is stopped.
>>>>> >>>>>>>>>
>>>>> >>>>>>>>> --
>>>>> >>>>>>>>> [root@cvtst1 ~]# clustat
>>>>> >>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009
>>>>> >>>>>>>>> Member Status: Quorate
>>>>> >>>>>>>>>
>>>>> >>>>>>>>>  Member Name                                                     ID   Status
>>>>> >>>>>>>>>  ------ ----                                                     ---- ------
>>>>> >>>>>>>>>  cvtst2                                                    1 Online, rgmanager
>>>>> >>>>>>>>>  cvtst1                                                     2 Online,
>>>>> >>>>>>>>> Local, rgmanager
>>>>> >>>>>>>>>  cvtst3                                                     3 Online, rgmanager
>>>>> >>>>>>>>>
>>>>> >>>>>>>>>  Service Name                                            Owner (Last)
>>>>> >>>>>>>>>                                          State
>>>>> >>>>>>>>>  ------- ----                                            ----- ------
>>>>> >>>>>>>>>                                          -----
>>>>> >>>>>>>>>  vm:guest1                                               (none)
>>>>> >>>>>>>>>                                          stopped
>>>>> >>>>>>>>> [root@cvtst1 ~]#
>>>>> >>>>>>>>>
>>>>> >>>>>>>>>
>>>>> >>>>>>>>> ---
>>>>> >>>>>>>>> o/p of xm li on cvtst1
>>>>> >>>>>>>>>
>>>>> >>>>>>>>> --
>>>>> >>>>>>>>> [root@cvtst1 ~]# xm li
>>>>> >>>>>>>>> Name                                      ID Mem(MiB) VCPUs State   Time(s)
>>>>> >>>>>>>>> Domain-0                                   0     3470     2 r-----  28939.4
>>>>> >>>>>>>>> guest1                                     7      511     1 -b----   7727.8
>>>>> >>>>>>>>>
>>>>> >>>>>>>>> o/p of xm li on cvtst2
>>>>> >>>>>>>>>
>>>>> >>>>>>>>> --
>>>>> >>>>>>>>> [root@cvtst2 ~]# xm li
>>>>> >>>>>>>>> Name                                      ID Mem(MiB) VCPUs State   Time(s)
>>>>> >>>>>>>>> Domain-0                                   0     3470     2 r-----  31558.9
>>>>> >>>>>>>>> guest1                                    21      511     1 -b----   7558.2
>>>>> >>>>>>>>> ---
>>>>> >>>>>>>>>
>>>>> >>>>>>>>> Thanks
>>>>> >>>>>>>>> Paras.
>>>>> >>>>>>>>>
>>>>> >>>>>>>>>
>>>>> >>>>>>>>>
>>>>> >>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli
>>>>> >>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote:
>>>>> >>>>>>>>>> It looks like no.
>>>>> >>>>>>>>>>
>>>>> >>>>>>>>>> can you send an output of clustat  of when the VM is running on
>>>>> >>>>>>>>>> multiple nodes at the same time?
>>>>> >>>>>>>>>>
>>>>> >>>>>>>>>> And by the way, another one after having stopped (clusvcadm -s vm:guest1) ?
>>>>> >>>>>>>>>>
>>>>> >>>>>>>>>>
>>>>> >>>>>>>>>>
>>>>> >>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>:
>>>>> >>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not being
>>>>> >>>>>>>>>>> properly handled by the cluster.
>>>>> >>>>>>>>>>>
>>>>> >>>>>>>>>>>
>>>>> >>>>>>>>>>> Thanks
>>>>> >>>>>>>>>>> Paras.
>>>>> >>>>>>>>>>>
>>>>> >>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan <pradhanparas@xxxxxxxxx> wrote:
>>>>> >>>>>>>>>>>> Ok.. here is my cluster.conf file
>>>>> >>>>>>>>>>>>
>>>>> >>>>>>>>>>>> --
>>>>> >>>>>>>>>>>> [root@cvtst1 cluster]# more cluster.conf
>>>>> >>>>>>>>>>>> <?xml version="1.0"?>
>>>>> >>>>>>>>>>>> <cluster alias="test" config_version="9" name="test">
>>>>> >>>>>>>>>>>>        <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/>
>>>>> >>>>>>>>>>>>        <clusternodes>
>>>>> >>>>>>>>>>>>                <clusternode name="cvtst2" nodeid="1" votes="1">
>>>>> >>>>>>>>>>>>                        <fence/>
>>>>> >>>>>>>>>>>>                </clusternode>
>>>>> >>>>>>>>>>>>                <clusternode name="cvtst1" nodeid="2" votes="1">
>>>>> >>>>>>>>>>>>                        <fence/>
>>>>> >>>>>>>>>>>>                </clusternode>
>>>>> >>>>>>>>>>>>                <clusternode name="cvtst3" nodeid="3" votes="1">
>>>>> >>>>>>>>>>>>                        <fence/>
>>>>> >>>>>>>>>>>>                </clusternode>
>>>>> >>>>>>>>>>>>        </clusternodes>
>>>>> >>>>>>>>>>>>        <cman/>
>>>>> >>>>>>>>>>>>        <fencedevices/>
>>>>> >>>>>>>>>>>>        <rm>
>>>>> >>>>>>>>>>>>                <failoverdomains>
>>>>> >>>>>>>>>>>>                        <failoverdomain name="myfd1" nofailback="0" ordered="1" restricted="0">
>>>>> >>>>>>>>>>>>                                <failoverdomainnode name="cvtst2" priority="3"/>
>>>>> >>>>>>>>>>>>                                <failoverdomainnode name="cvtst1" priority="1"/>
>>>>> >>>>>>>>>>>>                                <failoverdomainnode name="cvtst3" priority="2"/>
>>>>> >>>>>>>>>>>>                        </failoverdomain>
>>>>> >>>>>>>>>>>>                </failoverdomains>
>>>>> >>>>>>>>>>>>                <resources/>
>>>>> >>>>>>>>>>>>                <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0"
>>>>> >>>>>>>>>>>> name="guest1" path="/vms" recovery="r
>>>>> >>>>>>>>>>>> estart" restart_expire_time="0"/>
>>>>> >>>>>>>>>>>>        </rm>
>>>>> >>>>>>>>>>>> </cluster>
>>>>> >>>>>>>>>>>> [root@cvtst1 cluster]#
>>>>> >>>>>>>>>>>> ------
>>>>> >>>>>>>>>>>>
>>>>> >>>>>>>>>>>> Thanks!
>>>>> >>>>>>>>>>>> Paras.
>>>>> >>>>>>>>>>>>
>>>>> >>>>>>>>>>>>
>>>>> >>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer <volker@xxxxxxxxxxxx> wrote:
>>>>> >>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500,
>>>>> >>>>>>>>>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote:
>>>>> >>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. Now I am
>>>>> >>>>>>>>>>>>>> having another problem. When I start the my xen vm in one node, it
>>>>> >>>>>>>>>>>>>> also starts on other nodes. Which daemon controls  this?
>>>>> >>>>>>>>>>>>>
>>>>> >>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the rgmanager
>>>>> >>>>>>>>>>>>> package). To me, this sounds like a configuration problem. Maybe,
>>>>> >>>>>>>>>>>>> you can post your cluster.conf?
>>>>> >>>>>>>>>>>>>
>>>>> >>>>>>>>>>>>> Regards,
>>>>> >>>>>>>>>>>>> Volker
>>>>> >>>>>>>>>>>>>
>>>>> >>>>>>>>>>>>> --
>>>>> >>>>>>>>>>>>> Linux-cluster mailing list
>>>>> >>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>> >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>>>>>>>>>>>>
>>>>> >>>>>>>>>>>>
>>>>> >>>>>>>>>>>
>>>>> >>>>>>>>>>> --
>>>>> >>>>>>>>>>> Linux-cluster mailing list
>>>>> >>>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>> >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>>>>>>>>>>
>>>>> >>>>>>>>>>
>>>>> >>>>>>>>>> --
>>>>> >>>>>>>>>> Linux-cluster mailing list
>>>>> >>>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>> >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>>>>>>>>>
>>>>> >>>>>>>>>
>>>>> >>>>>>>>> --
>>>>> >>>>>>>>> Linux-cluster mailing list
>>>>> >>>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>> >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>>>>>>>>
>>>>> >>>>>>>>
>>>>> >>>>>>>> --
>>>>> >>>>>>>> Linux-cluster mailing list
>>>>> >>>>>>>> Linux-cluster@xxxxxxxxxx
>>>>> >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>>>>>>>
>>>>> >>>>>>>
>>>>> >>>>>>> --
>>>>> >>>>>>> Linux-cluster mailing list
>>>>> >>>>>>> Linux-cluster@xxxxxxxxxx
>>>>> >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>>>>>>
>>>>> >>>>>>
>>>>> >>>>>> --
>>>>> >>>>>> Linux-cluster mailing list
>>>>> >>>>>> Linux-cluster@xxxxxxxxxx
>>>>> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>>>>>
>>>>> >>>>>
>>>>> >>>>> --
>>>>> >>>>> Linux-cluster mailing list
>>>>> >>>>> Linux-cluster@xxxxxxxxxx
>>>>> >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>>>>
>>>>> >>>>
>>>>> >>>
>>>>> >>> --
>>>>> >>> Linux-cluster mailing list
>>>>> >>> Linux-cluster@xxxxxxxxxx
>>>>> >>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>>
>>>>> >>
>>>>> >> --
>>>>> >> Linux-cluster mailing list
>>>>> >> Linux-cluster@xxxxxxxxxx
>>>>> >> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >>
>>>>> >
>>>>> > --
>>>>> > Linux-cluster mailing list
>>>>> > Linux-cluster@xxxxxxxxxx
>>>>> > https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>> >
>>>>>
>>>>> --
>>>>> Linux-cluster mailing list
>>>>> Linux-cluster@xxxxxxxxxx
>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>>
>>>>
>>>> --
>>>> Linux-cluster mailing list
>>>> Linux-cluster@xxxxxxxxxx
>>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>>
>>>
>>> --
>>> Linux-cluster mailing list
>>> Linux-cluster@xxxxxxxxxx
>>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>>
>>
>> --
>> Linux-cluster mailing list
>> Linux-cluster@xxxxxxxxxx
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>>
>
> --
> Linux-cluster mailing list
> Linux-cluster@xxxxxxxxxx
> https://www.redhat.com/mailman/listinfo/linux-cluster
>

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux