Re: clusvcadm : Could not connect to resource group manager

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I was able to download RHEL5 32bit ISO evaluation copy so i went with it rather than downloading RHEL6 just for my PoC on two node cluster.

I was not blinding following RHEL4 doc which is why i did not look for installing those packages once after i did "Cluster" and "ClusterStorage" group install assuming everything must be there. 
I started having a doubt about any missing packages because of those dlm errors reported in the messages i sent earlier. 

Aug 23 05:02:59 server1 dlm_controld[27892]: group_init error (nil) 111
Aug 23 05:02:59 server1 fenced[27886]: group_init error (nil) 111

Searching on some forums always gives me old replies about RHEL4 and error which talks about some dlm rpm's, thats why.

-Param


On Sat, Aug 25, 2012 at 10:07 AM, Digimer <lists@xxxxxxxxxx> wrote:
I haven't used the rhcs stable 2 (version in RHEL 5) in some time, and I
don't remember for sure. I don't think I've seen magma though.

Why are you following the docs for RHEL 4, installing on RHEL 5 when
RHEL 6 is the current version?

On 08/25/2012 12:14 AM, PARAM KRISH wrote:
> Digimer,
>
> I just want to confirm if i am missing any rpm's in my setup.
>
> I referred this site
> http://www.centos.org/docs/4/html/rh-cs-en-4/ap-rhcs-sw-inst-cust.html
> to install the rpm's in red hat 5.6 but these packages *magma**, *dlm**
> are missing in the CD, does it mean these are really important ? I
> installed using 'yum group install' for "Cluster" and "ClusterStorage"
>
> This is what the yum.log says from server1. Please confirm if there is
> something missing.
>
> -Param
>
> On Fri, Aug 24, 2012 at 9:06 PM, PARAM KRISH <mkparam@xxxxxxxxx
> <mailto:mkparam@xxxxxxxxx>> wrote:
>
>     Please find below the details.
>
>     1. cluster.conf : Please find attached herewith.
>
>     2. Is fencing in any form mandatory for a setup as simple as this ?
>     I am just using two redhat5 VM's within VMware fusion in my Mac.
>     What kind of fencing is applicable to this kind of setup, that can
>     also benefit me ? All that i want to see from this PoC is to add
>     some services like Apache, MySQL to see how quick and reliable the
>     cluster to pick the service failures etc.,
>
>     3. Please find attached the messages from server1 and server2 when i
>     did "service cman stop and start" on both nodes one after the other.
>     In server1 both stop and start went fine but clustat showed "Could
>     not connect to CMAN: connection refused". If i do "cman_tool join"
>     on both nodes one after the other, things look green
>
>     Also, Am i missing any rpm's that are most important ? I used yum
>     group install "Clustering" and "ClusterStorage" to install all the
>     packages.
>
>     -Param
>
>     On Fri, Aug 24, 2012 at 7:28 PM, Digimer <lists@xxxxxxxxxx
>     <mailto:lists@xxxxxxxxxx>> wrote:
>
>         A few things;
>
>         1. Please repost your cluster.conf file with line wraps in plain
>         text.
>
>         2. Manual fencing is not supported in any way, please use real
>         fencing,
>         like IPMI, iLO, etc.
>
>         3. Please stop the cluster entirely, start 'tail -f -n 0
>         /var/log/messages' on both nodes, then start cman, then start
>         rgmanager.
>         Please share the output from the logs.
>
>         Digimer
>
>         On 08/24/2012 06:43 AM, PARAM KRISH wrote:
>         > Hi, Thanks for the help. I hope we are nearing to the problem.
>         >
>         > I enabled logging , this is how my cluster.conf looks like
>         >
>         > <?xml version="1.0"?>
>         > <cluster alias="newCluster" config_version="16" name="newCluster">
>         > <logging debug="on"/>
>         > <cman expected_votes="1" two_node="1"/>
>         > <clusternodes>
>         > <clusternode name="server1" nodeid="1" votes="1">
>         > <fence><method name="single"><device
>         > name="human"/></method></fence></clusternode><clusternode
>         name="server2"
>         > nodeid="2" votes="1"><fence><method name="single"><device
>         >
>         name="human"/></method></fence></clusternode></clusternodes><fencedevices>
>         >
>         >         </fencedevices><rm><failoverdomains><failoverdomain
>         > name="failOver" nofailback="0" ordered="1"
>         > restricted="0"><failoverdomainnode name="server1"
>         > priority="1"/><failoverdomainnode name="server2"
>         > priority="2"/></failoverdomain></failoverdomains><resources><ip
>         > address="192.168.61.130" monitor_link="1"/><apache
>         > config_file="conf/httpd.conf" name="httpd"
>         server_root="/etc/httpd"
>         > shutdown_wait="0"/></resources><service autostart="1"
>         domain="failOver"
>         > exclusive="1" name="Apache" recovery="relocate"><ip
>         > address="192.168.61.130" monitor_link="1"><apache
>         > config_file="conf/httpd.conf" name="Apache"
>         server_root="/etc/httpd"
>         > shutdown_wait="0"/></ip></service><service autostart="1"
>         > domain="failOver" exclusive="1" name="website"
>         recovery="relocate"><ip
>         > ref="192.168.61.130"><apache
>         > ref="httpd"/></ip></service></rm><fence_daemon clean_start="1"
>         > post_fail_delay="0" post_join_delay="3"/><logging
>         debug="on"/></cluster>
>         >
>         > There is no logging happening in /var/run/cluster/
>         >
>         > [root@server1 ~]# ls /var/run/cluster/
>         > apache  ccsd.pid  ccsd.sock  rgmanager.sk
>         <http://rgmanager.sk> <http://rgmanager.sk>
>         >
>         > I started resource manager in foreground and it says like ..
>         >
>         > failed acquiring lockspace: No such device
>         > Locks not working!
>         >
>         > What next i could do ?
>         >
>         > -Param
>         >
>         > On Fri, Aug 24, 2012 at 3:18 PM, emmanuel segura
>         <emi2fast@xxxxxxxxx <mailto:emi2fast@xxxxxxxxx>
>         > <mailto:emi2fast@xxxxxxxxx <mailto:emi2fast@xxxxxxxxx>>> wrote:
>         >
>         >     /etc/init.d/rgmanager start or service rgmanager start
>         >
>         >
>         >     2012/8/24 Heiko Nardmann <heiko.nardmann@xxxxxxxxxxxxx
>         <mailto:heiko.nardmann@xxxxxxxxxxxxx>
>         >     <mailto:heiko.nardmann@xxxxxxxxxxxxx
>         <mailto:heiko.nardmann@xxxxxxxxxxxxx>>>
>         >
>         >         It is strange that strace shows that
>         >         /var/run/cluster/rgmanager.sk <http://rgmanager.sk>
>         <http://rgmanager.sk> is missing.
>         >
>         >         Normally it is helpful to see the complete
>         cluster.conf. Could
>         >         you provide that one?
>         >
>         >         Also of interest is /var/log/cluster/rgmanager.log -
>         do you have
>         >         debug enabled inside cluster.conf?
>         >
>         >         Maybe it is possible to start rgmanager in the
>         foreground (-f)
>         >         with strace? That might also be a way to show why the
>         >         rgmanager.sk <http://rgmanager.sk>
>         <http://rgmanager.sk> is missing ...
>         >
>         >         Just some ideas ...
>         >
>         >
>         >         Kind regards,
>         >
>         >             Heiko
>         >
>         >         Am 24.08.2012 11 <tel:24.08.2012%2011>
>         <tel:24.08.2012%2011>:04, schrieb PARAM KRISH:
>         >
>         >             All,
>         >
>         >             I am trying to setup a simple two node cluster in
>         my laptop
>         >             using two RHEL VM's.
>         >
>         >             Everything looks just fine to me but i am unable
>         to enable a
>         >             apache service though it works beautifully when
>         tried with
>         >             "rg_test test" on both the nodes.
>         >
>         >             What could be the problem ? Please help. I am a
>         novice in
>         >             red hat cluster but learnt a bit of it in the last
>         few days
>         >             while trying to fix all the problems encountered.
>         >
>         >             Here are the details.
>         >
>         >             [root@server1 ~]# clustat
>         >             Cluster Status for newCluster @ Thu Aug 23
>         00:29:32 2012
>         >             Member Status: Quorate
>         >
>         >              Member Name                 ID   Status
>         >              ------ ----                 ---- ------
>         >              server1                     1 Online, Local
>         >              server2                     2 Online
>         >
>         >             [root@server1 ~]# clustat -x
>         >             <?xml version="1.0"?>
>         >             <clustat version="4.1.1">
>         >               <cluster name="newCluster" id="43188"
>         generation="250536"/>
>         >               <quorum quorate="1" groupmember="0"/>
>         >               <nodes>
>         >                 <node name="server1" state="1" local="1"
>         estranged="0"
>         >             rgmanager="0" rgmanager_master="0" qdisk="0"
>         >             nodeid="0x00000001"/>
>         >                 <node name="server2" state="1" local="0"
>         estranged="0"
>         >             rgmanager="0" rgmanager_master="0" qdisk="0"
>         >             nodeid="0x00000002"/>  </nodes>
>         >             </clustat>
>         >
>         >             [root@server2 ~]# clustat
>         >             Cluster Status for newCluster @ Thu Aug 23
>         03:13:34 2012
>         >             Member Status: Quorate
>         >
>         >              Member Name                 ID   Status
>         >              ------ ----                 ---- ------
>         >              server1                     1 Online
>         >              server2                     2 Online, Local
>         >
>         >             [root@server2 ~]# clustat -x
>         >             <?xml version="1.0"?>
>         >             <clustat version="4.1.1">
>         >               <cluster name="newCluster" id="43188"
>         generation="250536"/>
>         >               <quorum quorate="1" groupmember="0"/>
>         >               <nodes>
>         >                 <node name="server1" state="1" local="0"
>         estranged="0"
>         >             rgmanager="0" rgmanager_master="0" qdisk="0"
>         >             nodeid="0x00000001"/>
>         >                 <node name="server2" state="1" local="1"
>         estranged="0"
>         >             rgmanager="0" rgmanager_master="0" qdisk="0"
>         >             nodeid="0x00000002"/>
>         >               </nodes>
>         >             </clustat>
>         >
>         >
>         >             [root@server2 ~]# clusvcadm -e Apache
>         >             Local machine trying to enable
>         service:Apache...Could not
>         >             connect to resource group manager
>         >
>         >             strace cluvcsadm -e Apache
>         >             ...
>         >             stat64(1, {st_mode=S_IFCHR|0620,
>         st_rdev=makedev(136, 4),
>         >             ...}) = 0
>         >             mmap2(NULL, 4096, PROT_READ|PROT_WRITE,
>         >             MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0xb7fb5000
>         >             write(1, "Local machine trying to enable s"...,
>         48Local
>         >             machine trying to enable service:Apache...) = 48
>         >             socket(PF_FILE, SOCK_STREAM, 0)         = 5
>         >             connect(5, {sa_family=AF_FILE,
>         >             path="/var/run/cluster/rgmanag__er.sk
>         <http://rgmanag__er.sk> <http://rgmanager.sk>
>         >             <http://rgmanager.sk>"...}, 110) = -1 ENOENT (No
>         such file
>         >             or directory)
>         >
>         >             close(5)                                = 0
>         >             write(1, "Could not connect to resource gr"...,
>         44Could not
>         >             connect to resource group manager
>         >             ) = 44
>         >             exit_group(1)                           = ?
>         >
>         >
>         >             [root@server1 ~]# hostname
>         >             server1.localdomain
>         >
>         >             [root@server1 ~]# cat /etc/hosts
>         >             # Do not remove the following line, or various
>         programs
>         >             # that require network functionality will fail.
>         >             #127.0.0.1              server1.localdomain server1
>         >             localhost.localdomain localhost
>         >             192.168.61.132 server1.localdomain server1
>         >             192.168.61.133 server2.localdomain server2
>         >             ::1             localhost6.localdomain6 localhost6
>         >
>         >
>         >             Package versions :
>         >             luci-0.12.2-24.el5
>         >             ricci-0.12.2-24.el5
>         >             rgmanager-2.0.52-9.el5
>         >             modcluster-0.12.1-2.el5
>         >             cluster-cim-0.12.1-2.el5
>         >             system-config-cluster-1.0.57-7
>         >             lvm2-cluster-2.02.74-3.el5
>         >             cluster-snmp-0.12.1-2.el5
>         >
>         >             [root@server1 log]# cman_tool status
>         >             Version: 6.2.0
>         >             Config Version: 15
>         >             Cluster Name: newCluster
>         >             Cluster Id: 43188
>         >             Cluster Member: Yes
>         >             Cluster Generation: 250536
>         >             Membership state: Cluster-Member
>         >             Nodes: 2
>         >             Expected votes: 1
>         >             Total votes: 2
>         >             Quorum: 1
>         >             Active subsystems: 2
>         >             Flags: 2node
>         >             Ports Bound: 0
>         >             Node name: server1
>         >             Node ID: 1
>         >             Multicast addresses: 239.192.168.93
>         >             Node addresses: 192.168.61.132
>         >
>         >             Redhat :Red Hat Enterprise Linux Server release 5.6
>         >             (Tikanga)2.6.18-238.el5xen
>         >
>         >             [root@server1 log]# service rgmanager status
>         >             clurgmgrd (pid  9775) is running...
>         >
>         >             [root@server1 log]# netstat -na | grep 11111
>         >             tcp        0      0 0.0.0.0:11111
>         <http://0.0.0.0:11111> <http://0.0.0.0:11111>
>         >             <http://0.0.0.0:11111>         0.0.0.0:*
>         >             LISTEN
>         >
>         >
>         >             Please let me know if you can help. One thing i
>         noticed was
>         >             that in the "clustat" it does not show "rgmanager"
>         against
>         >             both the nodes but i see the service is just
>         running fine.
>         >
>         >             *Note : No iptables, no SELinux enabled.*
>         >             *
>         >
>         >             *
>         >             Hope i have given all the details required to help me
>         >             quickly. Thanks.
>         >
>         >             -Param
>         >
>         >
>         >
>         >         --
>         >         Linux-cluster mailing list
>         >         Linux-cluster@xxxxxxxxxx
>         <mailto:Linux-cluster@xxxxxxxxxx>
>         <mailto:Linux-cluster@xxxxxxxxxx <mailto:Linux-cluster@xxxxxxxxxx>>
>         >         https://www.redhat.com/__mailman/listinfo/linux-cluster
>         >         <https://www.redhat.com/mailman/listinfo/linux-cluster>
>         >
>         >
>         >
>         >
>         >     --
>         >     esta es mi vida e me la vivo hasta que dios quiera
>         >
>         >     --
>         >     Linux-cluster mailing list
>         >     Linux-cluster@xxxxxxxxxx <mailto:Linux-cluster@xxxxxxxxxx>
>         <mailto:Linux-cluster@xxxxxxxxxx <mailto:Linux-cluster@xxxxxxxxxx>>
>         >     https://www.redhat.com/mailman/listinfo/linux-cluster
>         >
>         >
>         >
>         >
>         > --
>         > Linux-cluster mailing list
>         > Linux-cluster@xxxxxxxxxx <mailto:Linux-cluster@xxxxxxxxxx>
>         > https://www.redhat.com/mailman/listinfo/linux-cluster
>         >
>
>
>         --
>         Digimer
>         Papers and Projects: https://alteeve.ca
>
>
>


--
Digimer
Papers and Projects: https://alteeve.ca

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux