That looks strange, maybe your cluster.conf will help 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: > I don't see rgmanager . > > Here is the o/p from clustat > > [root@cvtst1 cluster]# clustat > Cluster Status for test @ Tue Sep 29 15:53:33 2009 > Member Status: Quorate > > Member Name ID Status > ------ ---- ---- ------ > cvtst2 1 Online > cvtst1 2 Online, Local > cvtst3 3 Online > > > Thanks > Paras. > > On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli > <brem.belguebli@xxxxxxxxx> wrote: >> It looks correct, rgmanager seems to start on all nodes >> >> what gives you clustat ? >> >> If rgmanager doesn't show, check out the logs something may have gone wrong. >> >> >> 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: >>> Change to 7 and i got this log >>> >>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down >>> Cluster Service Manager... >>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down >>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down >>> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown complete, exiting >>> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster Service >>> Manager is stopped. >>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group >>> Manager Starting >>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service Data >>> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource Rules >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource Trees >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources defined >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover Domains >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing Services >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services Initialized >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port Opened >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: Local UP >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst2 UP >>> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst3 UP >>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1) Processed >>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1) Processed >>> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1) Processed >>> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events processed >>> >>> >>> Anything unusual here? >>> >>> Paras. >>> >>> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli >>> <brem.belguebli@xxxxxxxxx> wrote: >>>> I use log_level=7 to have more debugging info. >>>> >>>> It seems 4 is not enough. >>>> >>>> Brem >>>> >>>> >>>> 2009/9/29, Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>> Withe log_level of 3 I got only this >>>>> >>>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down >>>>> Cluster Service Manager... >>>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down >>>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, exiting >>>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service >>>>> Manager is stopped. >>>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group >>>>> Manager Starting >>>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down >>>>> Cluster Service Manager... >>>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service >>>>> Manager is stopped. >>>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group >>>>> Manager Starting >>>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting down >>>>> >>>>> I do not know what the last line means. >>>>> >>>>> rgmanager version I am running is: >>>>> rgmanager-2.0.52-1.el5.centos >>>>> >>>>> I don't what has gone wrong. >>>>> >>>>> Thanks >>>>> Paras. >>>>> >>>>> >>>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli >>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>> > you mean it stopped successfully on all the nodes but it is failing to >>>>> > start only on node cvtst1 ? >>>>> > >>>>> > look at the following page to make rgmanager more verbose. It 'll >>>>> > help debug.... >>>>> > >>>>> > http://sources.redhat.com/cluster/wiki/RGManager >>>>> > >>>>> > at Logging Configuration section >>>>> > >>>>> > >>>>> > >>>>> > >>>>> > 2009/9/29 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>> >> Brem, >>>>> >> >>>>> >> When I try to restart rgmanager on all the nodes, this time i do not >>>>> >> see rgmanager running on the first node. But I do see on other 2 >>>>> >> nodes. >>>>> >> >>>>> >> Log on the first node: >>>>> >> >>>>> >> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group >>>>> >> Manager Starting >>>>> >> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down >>>>> >> Cluster Service Manager... >>>>> >> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down >>>>> >> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, exiting >>>>> >> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service >>>>> >> Manager is stopped. >>>>> >> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group >>>>> >> Manager Starting >>>>> >> >>>>> >> - >>>>> >> It seems service is running , but I do not see rgmanger running using clustat >>>>> >> >>>>> >> >>>>> >> Don't know what is going on. >>>>> >> >>>>> >> Thanks >>>>> >> Paras. >>>>> >> >>>>> >> >>>>> >> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli >>>>> >> <brem.belguebli@xxxxxxxxx> wrote: >>>>> >>> Paras, >>>>> >>> >>>>> >>> Another thing, it would have been more interesting to have a start >>>>> >>> DEBUG not a stop. >>>>> >>> >>>>> >>> That's why I was asking you to first stop the vm manually on all your >>>>> >>> nodes, stop eventually rgmanager on all the nodes to reset the >>>>> >>> potential wrong states you may have, restart rgmanager. >>>>> >>> >>>>> >>> If your VM is configured to autostart, this will make it start. >>>>> >>> >>>>> >>> It should normally fail (as it does now). Send out your newly created >>>>> >>> DEBUG file. >>>>> >>> >>>>> >>> 2009/9/29 brem belguebli <brem.belguebli@xxxxxxxxx>: >>>>> >>>> Hi Paras, >>>>> >>>> >>>>> >>>> >>>>> >>>> I don't know the xen/cluster combination well, but if I do remember >>>>> >>>> well, I think I've read somewhere that when using xen you have to >>>>> >>>> declare the use_virsh=0 key in the VM definition in the cluster.conf. >>>>> >>>> >>>>> >>>> This would make rgmanager use xm commands instead of virsh >>>>> >>>> The DEBUG output shows clearly that you are using virsh to manage your >>>>> >>>> VM instead of xm commands. >>>>> >>>> Check out the RH docs about virtualization >>>>> >>>> >>>>> >>>> I'm not a 100% sure about that, I may be completely wrong. >>>>> >>>> >>>>> >>>> Brem >>>>> >>>> >>>>> >>>> 2009/9/28 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>> >>>>> The only thing I noticed is the message after stopping the vm using xm >>>>> >>>>> in all nodes and starting using clusvcadm is >>>>> >>>>> >>>>> >>>>> "Virtual machine guest1 is blocked" >>>>> >>>>> >>>>> >>>>> The whole DEBUG file is attached. >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> Thanks >>>>> >>>>> Paras. >>>>> >>>>> >>>>> >>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli >>>>> >>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>> >>>>>> There's a problem with the script that is called by rgmanager to start >>>>> >>>>>> the VM, I don't know what causes it >>>>> >>>>>> >>>>> >>>>>> May be you should try something like : >>>>> >>>>>> >>>>> >>>>>> 1) stop the VM on all nodes with xm commands >>>>> >>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the following >>>>> >>>>>> lines (after the #!/bin/bash ): >>>>> >>>>>> exec >/tmp/DEBUG 2>&1 >>>>> >>>>>> set -x >>>>> >>>>>> 3) start the VM with clusvcadm -e vm:guest1 >>>>> >>>>>> >>>>> >>>>>> It should fail as it did before. >>>>> >>>>>> >>>>> >>>>>> edit the the /tmp/DEBUG file and you will be able to see where it >>>>> >>>>>> fails (it may generate a lot of debug) >>>>> >>>>>> >>>>> >>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh >>>>> >>>>>> >>>>> >>>>>> Post the DEBUG file if you're not able to see where it fails. >>>>> >>>>>> >>>>> >>>>>> Brem >>>>> >>>>>> >>>>> >>>>>> 2009/9/26 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>> >>>>>>> No I am not manually starting not using automatic init scripts. >>>>> >>>>>>> >>>>> >>>>>>> I started the vm using: clusvcadm -e vm:guest1 >>>>> >>>>>>> >>>>> >>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few seconds it >>>>> >>>>>>> says guest1 started . But after a while I can see the guest1 on all >>>>> >>>>>>> three nodes. >>>>> >>>>>>> >>>>> >>>>>>> clustat says: >>>>> >>>>>>> >>>>> >>>>>>> Service Name Owner (Last) >>>>> >>>>>>> State >>>>> >>>>>>> ------- ---- ----- ------ >>>>> >>>>>>> ----- >>>>> >>>>>>> vm:guest1 (none) >>>>> >>>>>>> stopped >>>>> >>>>>>> >>>>> >>>>>>> But I can see the vm from xm li. >>>>> >>>>>>> >>>>> >>>>>>> This is what I can see from the log: >>>>> >>>>>>> >>>>> >>>>>>> >>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" >>>>> >>>>>>> returned 1 (generic error) >>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start >>>>> >>>>>>> vm:guest1; return value: 1 >>>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 >>>>> >>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is >>>>> >>>>>>> recovering >>>>> >>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed >>>>> >>>>>>> service vm:guest1 >>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm "guest1" >>>>> >>>>>>> returned 1 (generic error) >>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to start >>>>> >>>>>>> vm:guest1; return value: 1 >>>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service vm:guest1 >>>>> >>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 is >>>>> >>>>>>> recovering >>>>> >>>>>>> >>>>> >>>>>>> >>>>> >>>>>>> Paras. >>>>> >>>>>>> >>>>> >>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli >>>>> >>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>> >>>>>>>> Have you started your VM via rgmanager (clusvcadm -e vm:guest1) or >>>>> >>>>>>>> using xm commands out of cluster control (or maybe a thru an >>>>> >>>>>>>> automatic init script ?) >>>>> >>>>>>>> >>>>> >>>>>>>> When clustered, you should never be starting services (manually or >>>>> >>>>>>>> thru automatic init script) out of cluster control >>>>> >>>>>>>> >>>>> >>>>>>>> The thing would be to stop your vm on all the nodes with the adequate >>>>> >>>>>>>> xm command (not using xen myself) and try to start it with clusvcadm. >>>>> >>>>>>>> >>>>> >>>>>>>> Then see if it is started on all nodes (send clustat output) >>>>> >>>>>>>> >>>>> >>>>>>>> >>>>> >>>>>>>> >>>>> >>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>> >>>>>>>>> Ok. Please see below. my vm is running on all nodes though clustat >>>>> >>>>>>>>> says it is stopped. >>>>> >>>>>>>>> >>>>> >>>>>>>>> -- >>>>> >>>>>>>>> [root@cvtst1 ~]# clustat >>>>> >>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009 >>>>> >>>>>>>>> Member Status: Quorate >>>>> >>>>>>>>> >>>>> >>>>>>>>> Member Name ID Status >>>>> >>>>>>>>> ------ ---- ---- ------ >>>>> >>>>>>>>> cvtst2 1 Online, rgmanager >>>>> >>>>>>>>> cvtst1 2 Online, >>>>> >>>>>>>>> Local, rgmanager >>>>> >>>>>>>>> cvtst3 3 Online, rgmanager >>>>> >>>>>>>>> >>>>> >>>>>>>>> Service Name Owner (Last) >>>>> >>>>>>>>> State >>>>> >>>>>>>>> ------- ---- ----- ------ >>>>> >>>>>>>>> ----- >>>>> >>>>>>>>> vm:guest1 (none) >>>>> >>>>>>>>> stopped >>>>> >>>>>>>>> [root@cvtst1 ~]# >>>>> >>>>>>>>> >>>>> >>>>>>>>> >>>>> >>>>>>>>> --- >>>>> >>>>>>>>> o/p of xm li on cvtst1 >>>>> >>>>>>>>> >>>>> >>>>>>>>> -- >>>>> >>>>>>>>> [root@cvtst1 ~]# xm li >>>>> >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) >>>>> >>>>>>>>> Domain-0 0 3470 2 r----- 28939.4 >>>>> >>>>>>>>> guest1 7 511 1 -b---- 7727.8 >>>>> >>>>>>>>> >>>>> >>>>>>>>> o/p of xm li on cvtst2 >>>>> >>>>>>>>> >>>>> >>>>>>>>> -- >>>>> >>>>>>>>> [root@cvtst2 ~]# xm li >>>>> >>>>>>>>> Name ID Mem(MiB) VCPUs State Time(s) >>>>> >>>>>>>>> Domain-0 0 3470 2 r----- 31558.9 >>>>> >>>>>>>>> guest1 21 511 1 -b---- 7558.2 >>>>> >>>>>>>>> --- >>>>> >>>>>>>>> >>>>> >>>>>>>>> Thanks >>>>> >>>>>>>>> Paras. >>>>> >>>>>>>>> >>>>> >>>>>>>>> >>>>> >>>>>>>>> >>>>> >>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli >>>>> >>>>>>>>> <brem.belguebli@xxxxxxxxx> wrote: >>>>> >>>>>>>>>> It looks like no. >>>>> >>>>>>>>>> >>>>> >>>>>>>>>> can you send an output of clustat of when the VM is running on >>>>> >>>>>>>>>> multiple nodes at the same time? >>>>> >>>>>>>>>> >>>>> >>>>>>>>>> And by the way, another one after having stopped (clusvcadm -s vm:guest1) ? >>>>> >>>>>>>>>> >>>>> >>>>>>>>>> >>>>> >>>>>>>>>> >>>>> >>>>>>>>>> 2009/9/25 Paras pradhan <pradhanparas@xxxxxxxxx>: >>>>> >>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not being >>>>> >>>>>>>>>>> properly handled by the cluster. >>>>> >>>>>>>>>>> >>>>> >>>>>>>>>>> >>>>> >>>>>>>>>>> Thanks >>>>> >>>>>>>>>>> Paras. >>>>> >>>>>>>>>>> >>>>> >>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >>>>> >>>>>>>>>>>> Ok.. here is my cluster.conf file >>>>> >>>>>>>>>>>> >>>>> >>>>>>>>>>>> -- >>>>> >>>>>>>>>>>> [root@cvtst1 cluster]# more cluster.conf >>>>> >>>>>>>>>>>> <?xml version="1.0"?> >>>>> >>>>>>>>>>>> <cluster alias="test" config_version="9" name="test"> >>>>> >>>>>>>>>>>> <fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/> >>>>> >>>>>>>>>>>> <clusternodes> >>>>> >>>>>>>>>>>> <clusternode name="cvtst2" nodeid="1" votes="1"> >>>>> >>>>>>>>>>>> <fence/> >>>>> >>>>>>>>>>>> </clusternode> >>>>> >>>>>>>>>>>> <clusternode name="cvtst1" nodeid="2" votes="1"> >>>>> >>>>>>>>>>>> <fence/> >>>>> >>>>>>>>>>>> </clusternode> >>>>> >>>>>>>>>>>> <clusternode name="cvtst3" nodeid="3" votes="1"> >>>>> >>>>>>>>>>>> <fence/> >>>>> >>>>>>>>>>>> </clusternode> >>>>> >>>>>>>>>>>> </clusternodes> >>>>> >>>>>>>>>>>> <cman/> >>>>> >>>>>>>>>>>> <fencedevices/> >>>>> >>>>>>>>>>>> <rm> >>>>> >>>>>>>>>>>> <failoverdomains> >>>>> >>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="0" ordered="1" restricted="0"> >>>>> >>>>>>>>>>>> <failoverdomainnode name="cvtst2" priority="3"/> >>>>> >>>>>>>>>>>> <failoverdomainnode name="cvtst1" priority="1"/> >>>>> >>>>>>>>>>>> <failoverdomainnode name="cvtst3" priority="2"/> >>>>> >>>>>>>>>>>> </failoverdomain> >>>>> >>>>>>>>>>>> </failoverdomains> >>>>> >>>>>>>>>>>> <resources/> >>>>> >>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0" >>>>> >>>>>>>>>>>> name="guest1" path="/vms" recovery="r >>>>> >>>>>>>>>>>> estart" restart_expire_time="0"/> >>>>> >>>>>>>>>>>> </rm> >>>>> >>>>>>>>>>>> </cluster> >>>>> >>>>>>>>>>>> [root@cvtst1 cluster]# >>>>> >>>>>>>>>>>> ------ >>>>> >>>>>>>>>>>> >>>>> >>>>>>>>>>>> Thanks! >>>>> >>>>>>>>>>>> Paras. >>>>> >>>>>>>>>>>> >>>>> >>>>>>>>>>>> >>>>> >>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer <volker@xxxxxxxxxxxx> wrote: >>>>> >>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500, >>>>> >>>>>>>>>>>>> Paras pradhan <pradhanparas@xxxxxxxxx> wrote: >>>>> >>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. Now I am >>>>> >>>>>>>>>>>>>> having another problem. When I start the my xen vm in one node, it >>>>> >>>>>>>>>>>>>> also starts on other nodes. Which daemon controls this? >>>>> >>>>>>>>>>>>> >>>>> >>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the rgmanager >>>>> >>>>>>>>>>>>> package). To me, this sounds like a configuration problem. Maybe, >>>>> >>>>>>>>>>>>> you can post your cluster.conf? >>>>> >>>>>>>>>>>>> >>>>> >>>>>>>>>>>>> Regards, >>>>> >>>>>>>>>>>>> Volker >>>>> >>>>>>>>>>>>> >>>>> >>>>>>>>>>>>> -- >>>>> >>>>>>>>>>>>> Linux-cluster mailing list >>>>> >>>>>>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>> >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>>>>>>>>>> >>>>> >>>>>>>>>>>> >>>>> >>>>>>>>>>> >>>>> >>>>>>>>>>> -- >>>>> >>>>>>>>>>> Linux-cluster mailing list >>>>> >>>>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>> >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>>>>>>>> >>>>> >>>>>>>>>> >>>>> >>>>>>>>>> -- >>>>> >>>>>>>>>> Linux-cluster mailing list >>>>> >>>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>> >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>>>>>>> >>>>> >>>>>>>>> >>>>> >>>>>>>>> -- >>>>> >>>>>>>>> Linux-cluster mailing list >>>>> >>>>>>>>> Linux-cluster@xxxxxxxxxx >>>>> >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>>>>>> >>>>> >>>>>>>> >>>>> >>>>>>>> -- >>>>> >>>>>>>> Linux-cluster mailing list >>>>> >>>>>>>> Linux-cluster@xxxxxxxxxx >>>>> >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>>>>> >>>>> >>>>>>> >>>>> >>>>>>> -- >>>>> >>>>>>> Linux-cluster mailing list >>>>> >>>>>>> Linux-cluster@xxxxxxxxxx >>>>> >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>>>> >>>>> >>>>>> >>>>> >>>>>> -- >>>>> >>>>>> Linux-cluster mailing list >>>>> >>>>>> Linux-cluster@xxxxxxxxxx >>>>> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> >>>>> Linux-cluster mailing list >>>>> >>>>> Linux-cluster@xxxxxxxxxx >>>>> >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>>> >>>>> >>>> >>>>> >>> >>>>> >>> -- >>>>> >>> Linux-cluster mailing list >>>>> >>> Linux-cluster@xxxxxxxxxx >>>>> >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>> >>>>> >> >>>>> >> -- >>>>> >> Linux-cluster mailing list >>>>> >> Linux-cluster@xxxxxxxxxx >>>>> >> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >> >>>>> > >>>>> > -- >>>>> > Linux-cluster mailing list >>>>> > Linux-cluster@xxxxxxxxxx >>>>> > https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> > >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@xxxxxxxxxx >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@xxxxxxxxxx >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster@xxxxxxxxxx >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> -- >> Linux-cluster mailing list >> Linux-cluster@xxxxxxxxxx >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > -- > Linux-cluster mailing list > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster