I was not blinding following RHEL4 doc which is why i did not look for installing those packages once after i did "Cluster" and "ClusterStorage" group install assuming everything must be there.
I started having a doubt about any missing packages because of those dlm errors reported in the messages i sent earlier.
Aug 23 05:02:59 server1 dlm_controld[27892]: group_init error (nil) 111
Aug 23 05:02:59 server1 fenced[27886]: group_init error (nil) 111
Searching on some forums always gives me old replies about RHEL4 and error which talks about some dlm rpm's, thats why.
-Param
On Sat, Aug 25, 2012 at 10:07 AM, Digimer <lists@xxxxxxxxxx> wrote:
I haven't used the rhcs stable 2 (version in RHEL 5) in some time, and I
don't remember for sure. I don't think I've seen magma though.
Why are you following the docs for RHEL 4, installing on RHEL 5 when
RHEL 6 is the current version?
> to install the rpm's in red hat 5.6 but these packages *magma**, *dlm**
On 08/25/2012 12:14 AM, PARAM KRISH wrote:
> Digimer,
>
> I just want to confirm if i am missing any rpm's in my setup.
>
> I referred this site
> http://www.centos.org/docs/4/html/rh-cs-en-4/ap-rhcs-sw-inst-cust.html
> are missing in the CD, does it mean these are really important ? I
> installed using 'yum group install' for "Cluster" and "ClusterStorage"
>
> This is what the yum.log says from server1. Please confirm if there is
> something missing.
>
> -Param
>
> On Fri, Aug 24, 2012 at 9:06 PM, PARAM KRISH <mkparam@xxxxxxxxx
> <mailto:mkparam@xxxxxxxxx>> wrote:
>
> Please find below the details.
>
> 1. cluster.conf : Please find attached herewith.
>
> 2. Is fencing in any form mandatory for a setup as simple as this ?
> I am just using two redhat5 VM's within VMware fusion in my Mac.
> What kind of fencing is applicable to this kind of setup, that can
> also benefit me ? All that i want to see from this PoC is to add
> some services like Apache, MySQL to see how quick and reliable the
> cluster to pick the service failures etc.,
>
> 3. Please find attached the messages from server1 and server2 when i
> did "service cman stop and start" on both nodes one after the other.
> In server1 both stop and start went fine but clustat showed "Could
> not connect to CMAN: connection refused". If i do "cman_tool join"
> on both nodes one after the other, things look green
>
> Also, Am i missing any rpm's that are most important ? I used yum
> group install "Clustering" and "ClusterStorage" to install all the
> packages.
>
> -Param
>
> On Fri, Aug 24, 2012 at 7:28 PM, Digimer <lists@xxxxxxxxxx
> <http://rgmanager.sk> <http://rgmanager.sk>> <mailto:lists@xxxxxxxxxx>> wrote:
>
> A few things;
>
> 1. Please repost your cluster.conf file with line wraps in plain
> text.
>
> 2. Manual fencing is not supported in any way, please use real
> fencing,
> like IPMI, iLO, etc.
>
> 3. Please stop the cluster entirely, start 'tail -f -n 0
> /var/log/messages' on both nodes, then start cman, then start
> rgmanager.
> Please share the output from the logs.
>
> Digimer
>
> On 08/24/2012 06:43 AM, PARAM KRISH wrote:
> > Hi, Thanks for the help. I hope we are nearing to the problem.
> >
> > I enabled logging , this is how my cluster.conf looks like
> >
> > <?xml version="1.0"?>
> > <cluster alias="newCluster" config_version="16" name="newCluster">
> > <logging debug="on"/>
> > <cman expected_votes="1" two_node="1"/>
> > <clusternodes>
> > <clusternode name="server1" nodeid="1" votes="1">
> > <fence><method name="single"><device
> > name="human"/></method></fence></clusternode><clusternode
> name="server2"
> > nodeid="2" votes="1"><fence><method name="single"><device
> >
> name="human"/></method></fence></clusternode></clusternodes><fencedevices>
> >
> > </fencedevices><rm><failoverdomains><failoverdomain
> > name="failOver" nofailback="0" ordered="1"
> > restricted="0"><failoverdomainnode name="server1"
> > priority="1"/><failoverdomainnode name="server2"
> > priority="2"/></failoverdomain></failoverdomains><resources><ip
> > address="192.168.61.130" monitor_link="1"/><apache
> > config_file="conf/httpd.conf" name="httpd"
> server_root="/etc/httpd"
> > shutdown_wait="0"/></resources><service autostart="1"
> domain="failOver"
> > exclusive="1" name="Apache" recovery="relocate"><ip
> > address="192.168.61.130" monitor_link="1"><apache
> > config_file="conf/httpd.conf" name="Apache"
> server_root="/etc/httpd"
> > shutdown_wait="0"/></ip></service><service autostart="1"
> > domain="failOver" exclusive="1" name="website"
> recovery="relocate"><ip
> > ref="192.168.61.130"><apache
> > ref="httpd"/></ip></service></rm><fence_daemon clean_start="1"
> > post_fail_delay="0" post_join_delay="3"/><logging
> debug="on"/></cluster>
> >
> > There is no logging happening in /var/run/cluster/
> >
> > [root@server1 ~]# ls /var/run/cluster/
> > apache ccsd.pid ccsd.sock rgmanager.sk
> >
> > I started resource manager in foreground and it says like ..
> >
> > failed acquiring lockspace: No such device
> > Locks not working!
> >
> > What next i could do ?
> >
> > -Param
> >
> > On Fri, Aug 24, 2012 at 3:18 PM, emmanuel segura
> <emi2fast@xxxxxxxxx <mailto:emi2fast@xxxxxxxxx>
> > <mailto:emi2fast@xxxxxxxxx <mailto:emi2fast@xxxxxxxxx>>> wrote:> > <mailto:heiko.nardmann@xxxxxxxxxxxxx
> >
> > /etc/init.d/rgmanager start or service rgmanager start
> >
> >
> > 2012/8/24 Heiko Nardmann <heiko.nardmann@xxxxxxxxxxxxx
> <mailto:heiko.nardmann@xxxxxxxxxxxxx>
> <http://rgmanag__er.sk> <http://rgmanager.sk>> <mailto:heiko.nardmann@xxxxxxxxxxxxx>>>
> >
> > It is strange that strace shows that
> > /var/run/cluster/rgmanager.sk <http://rgmanager.sk>
> <http://rgmanager.sk> is missing.
> >
> > Normally it is helpful to see the complete
> cluster.conf. Could
> > you provide that one?
> >
> > Also of interest is /var/log/cluster/rgmanager.log -
> do you have
> > debug enabled inside cluster.conf?
> >
> > Maybe it is possible to start rgmanager in the
> foreground (-f)
> > with strace? That might also be a way to show why the
> > rgmanager.sk <http://rgmanager.sk>
> <http://rgmanager.sk> is missing ...
> >
> > Just some ideas ...
> >
> >
> > Kind regards,
> >
> > Heiko
> >
> > Am 24.08.2012 11 <tel:24.08.2012%2011>
> <tel:24.08.2012%2011>:04, schrieb PARAM KRISH:
> >
> > All,
> >
> > I am trying to setup a simple two node cluster in
> my laptop
> > using two RHEL VM's.
> >
> > Everything looks just fine to me but i am unable
> to enable a
> > apache service though it works beautifully when
> tried with
> > "rg_test test" on both the nodes.
> >
> > What could be the problem ? Please help. I am a
> novice in
> > red hat cluster but learnt a bit of it in the last
> few days
> > while trying to fix all the problems encountered.
> >
> > Here are the details.
> >
> > [root@server1 ~]# clustat
> > Cluster Status for newCluster @ Thu Aug 23
> 00:29:32 2012
> > Member Status: Quorate
> >
> > Member Name ID Status
> > ------ ---- ---- ------
> > server1 1 Online, Local
> > server2 2 Online
> >
> > [root@server1 ~]# clustat -x
> > <?xml version="1.0"?>
> > <clustat version="4.1.1">
> > <cluster name="newCluster" id="43188"
> generation="250536"/>
> > <quorum quorate="1" groupmember="0"/>
> > <nodes>
> > <node name="server1" state="1" local="1"
> estranged="0"
> > rgmanager="0" rgmanager_master="0" qdisk="0"
> > nodeid="0x00000001"/>
> > <node name="server2" state="1" local="0"
> estranged="0"
> > rgmanager="0" rgmanager_master="0" qdisk="0"
> > nodeid="0x00000002"/> </nodes>
> > </clustat>
> >
> > [root@server2 ~]# clustat
> > Cluster Status for newCluster @ Thu Aug 23
> 03:13:34 2012
> > Member Status: Quorate
> >
> > Member Name ID Status
> > ------ ---- ---- ------
> > server1 1 Online
> > server2 2 Online, Local
> >
> > [root@server2 ~]# clustat -x
> > <?xml version="1.0"?>
> > <clustat version="4.1.1">
> > <cluster name="newCluster" id="43188"
> generation="250536"/>
> > <quorum quorate="1" groupmember="0"/>
> > <nodes>
> > <node name="server1" state="1" local="0"
> estranged="0"
> > rgmanager="0" rgmanager_master="0" qdisk="0"
> > nodeid="0x00000001"/>
> > <node name="server2" state="1" local="1"
> estranged="0"
> > rgmanager="0" rgmanager_master="0" qdisk="0"
> > nodeid="0x00000002"/>
> > </nodes>
> > </clustat>
> >
> >
> > [root@server2 ~]# clusvcadm -e Apache
> > Local machine trying to enable
> service:Apache...Could not
> > connect to resource group manager
> >
> > strace cluvcsadm -e Apache
> > ...
> > stat64(1, {st_mode=S_IFCHR|0620,
> st_rdev=makedev(136, 4),
> > ...}) = 0
> > mmap2(NULL, 4096, PROT_READ|PROT_WRITE,
> > MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0xb7fb5000
> > write(1, "Local machine trying to enable s"...,
> 48Local
> > machine trying to enable service:Apache...) = 48
> > socket(PF_FILE, SOCK_STREAM, 0) = 5
> > connect(5, {sa_family=AF_FILE,
> > path="/var/run/cluster/rgmanag__er.sk
> <mailto:Linux-cluster@xxxxxxxxxx <mailto:Linux-cluster@xxxxxxxxxx>>> > <http://rgmanager.sk>"...}, 110) = -1 ENOENT (No
> such file
> > or directory)
> >
> > close(5) = 0
> > write(1, "Could not connect to resource gr"...,
> 44Could not
> > connect to resource group manager
> > ) = 44
> > exit_group(1) = ?
> >
> >
> > [root@server1 ~]# hostname
> > server1.localdomain
> >
> > [root@server1 ~]# cat /etc/hosts
> > # Do not remove the following line, or various
> programs
> > # that require network functionality will fail.
> > #127.0.0.1 server1.localdomain server1
> > localhost.localdomain localhost
> > 192.168.61.132 server1.localdomain server1
> > 192.168.61.133 server2.localdomain server2
> > ::1 localhost6.localdomain6 localhost6
> >
> >
> > Package versions :
> > luci-0.12.2-24.el5
> > ricci-0.12.2-24.el5
> > rgmanager-2.0.52-9.el5
> > modcluster-0.12.1-2.el5
> > cluster-cim-0.12.1-2.el5
> > system-config-cluster-1.0.57-7
> > lvm2-cluster-2.02.74-3.el5
> > cluster-snmp-0.12.1-2.el5
> >
> > [root@server1 log]# cman_tool status
> > Version: 6.2.0
> > Config Version: 15
> > Cluster Name: newCluster
> > Cluster Id: 43188
> > Cluster Member: Yes
> > Cluster Generation: 250536
> > Membership state: Cluster-Member
> > Nodes: 2
> > Expected votes: 1
> > Total votes: 2
> > Quorum: 1
> > Active subsystems: 2
> > Flags: 2node
> > Ports Bound: 0
> > Node name: server1
> > Node ID: 1
> > Multicast addresses: 239.192.168.93
> > Node addresses: 192.168.61.132
> >
> > Redhat :Red Hat Enterprise Linux Server release 5.6
> > (Tikanga)2.6.18-238.el5xen
> >
> > [root@server1 log]# service rgmanager status
> > clurgmgrd (pid 9775) is running...
> >
> > [root@server1 log]# netstat -na | grep 11111
> > tcp 0 0 0.0.0.0:11111
> <http://0.0.0.0:11111> <http://0.0.0.0:11111>
> > <http://0.0.0.0:11111> 0.0.0.0:*
> > LISTEN
> >
> >
> > Please let me know if you can help. One thing i
> noticed was
> > that in the "clustat" it does not show "rgmanager"
> against
> > both the nodes but i see the service is just
> running fine.
> >
> > *Note : No iptables, no SELinux enabled.*
> > *
> >
> > *
> > Hope i have given all the details required to help me
> > quickly. Thanks.
> >
> > -Param
> >
> >
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster@xxxxxxxxxx
> <mailto:Linux-cluster@xxxxxxxxxx>
> > https://www.redhat.com/__mailman/listinfo/linux-cluster> <mailto:Linux-cluster@xxxxxxxxxx <mailto:Linux-cluster@xxxxxxxxxx>>
> > <https://www.redhat.com/mailman/listinfo/linux-cluster>
> >
> >
> >
> >
> > --
> > esta es mi vida e me la vivo hasta que dios quiera
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster@xxxxxxxxxx <mailto:Linux-cluster@xxxxxxxxxx>
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> >
> >
> >
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster@xxxxxxxxxx <mailto:Linux-cluster@xxxxxxxxxx>
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> >
>
>
> --
> Digimer
> Papers and Projects: https://alteeve.ca
>
>
>
--
Digimer
Papers and Projects: https://alteeve.ca
-- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster