Re: Re: Starting up two of three nodes that compose a cluster

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



David Teigland wrote:
On Fri, Sep 21, 2007 at 05:02:18PM +0200, carlopmart wrote:
David Teigland wrote:
On Thu, Sep 20, 2007 at 11:40:55AM +0200, carlopmart wrote:
Please, any hints??

-------- Original Message --------
Subject: Starting up two of three nodes that compose a cluster
Date: Wed, 19 Sep 2007 14:51:46 +0200
From: carlopmart <carlopmart@xxxxxxxxx>
To: linux clustering <linux-cluster@xxxxxxxxxx>

Hi all,

I have setup a rhel5 based cluster with three nodes. Sometimes i need
to start only two of this three nodes, but cluster services that i
configured doesn't starts (fenced fail). Is it not possible to start up
only two nodes on a three node cluster?? Maybe I need to adjust votes
param to two instead of three??
Could you be more specific about what you run, where, what happens,
what messages you see, etc.

Dave


Yes,

First, I attached my cluster.conf. When /etc/init.d/cman starts, returns an ok, but when I try to mount my gfs partition returns this error:

[root@haldir cluster]# service mountgfs start
Mounting GFS filesystems: /sbin/mount.gfs: lock_dlm_join: gfs_controld join error: -22
/sbin/mount.gfs: error mounting lockproto lock_dlm

So an error is coming back from gfs_controld on mount.  Please do the
steps manually, without init scripts or other scripts, so we know exactly
what steps fail.  And look in /var/log/messages for anything from
gfs_controld.  If there are none, send the output of 'group_tool -v;
group_tool dump gfs' after the failed mount.

Dave


Hi Dave,

 When I try mount gfs patition fails:

 [root@thranduil log]# mount -t gfs /dev/xvdc1 /data
/sbin/mount.gfs: lock_dlm_join: gfs_controld join error: -22
/sbin/mount.gfs: error mounting lockproto lock_dlm
[root@thranduil log]#

Output of group_tool command:

[root@thranduil log]# group_tool -v; group_tool dump gfs
type             level name     id       state node id local_done
fence            0     default  00010001 JOIN_START_WAIT 1 100010001 0
[1]
1190386130 listen 1
1190386130 cpg 4
1190386130 groupd 6
1190386130 uevent 7
1190386130 plocks 10
1190386130 setup done
1190386167 client 6: join /data gfs lock_dlm XenDomUcluster:datavol01 rw /dev/xvdc1
1190386167 mount: /data gfs lock_dlm XenDomUcluster:datavol01 rw /dev/xvdc1
1190386167 datavol01 cluster name matches: XenDomUcluster
1190386167 mount: not in default fence domain
1190386167 datavol01 do_mount: rv -22
1190386167 client 6 fd 11 dead
1190386167 client 6 fd -1 dead
1190386228 client 6: join /data gfs lock_dlm XenDomUcluster:datavol01 rw /dev/xvdc1
1190386228 mount: /data gfs lock_dlm XenDomUcluster:datavol01 rw /dev/xvdc1
1190386228 datavol01 cluster name matches: XenDomUcluster
1190386228 mount: not in default fence domain
1190386228 datavol01 do_mount: rv -22
1190386228 client 6 fd 11 dead
1190386228 client 6 fd -1 dead
1190388485 client 6: join /data gfs lock_dlm XenDomUcluster:datavol01 rw /dev/xvdc1
1190388485 mount: /data gfs lock_dlm XenDomUcluster:datavol01 rw /dev/xvdc1
1190388485 datavol01 cluster name matches: XenDomUcluster
1190388485 mount: not in default fence domain
1190388485 datavol01 do_mount: rv -22
1190388485 client 6 fd 11 dead
1190388485 client 6 fd -1 dead
1190388530 client 6: dump
[root@thranduil log]#

 Thanks David.


--
CL Martinez
carlopmart {at} gmail {d0t} com

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux