I'm in trouble with adding a new node to an existing cluster of three nodes (so I want to have four), because it somehow doesn't let me access the cluster infrastructure.
These 3 nodes were set up as http servers, sharing a GFS2 volume (physical: vg_cluster, logical: lv_cluster) where data is stored.
I want to set up the new node to access the same GFS2 volume, with the idea of exporting the data via NFS, so that a remote backup library can be configured to backup nightly the data, by connecting to the new node.
I've tried a lot of things, always getting same kind of errors.
Running "cman_tool status" on any of the 3 nodes i get:
Version: 6.2.0
Config Version: 70
Cluster Name: campusvirtual
Cluster Id: 45794
Cluster Member: Yes
Cluster Generation: 1136
Membership state: Cluster-Member
Nodes: 3
Expected votes: 4
Total votes: 3
Quorum: 3
Active subsystems: 9
Flags: Dirty
Ports Bound: 0 11 177
Node name: cev01
Node ID: 2
Multicast addresses: 239.192.178.149
Node addresses: 150.214.243.20
while running "cman_tool status" on the new node:
Version: 6.2.0
Config Version: 70
Cluster Name: campusvirtual
Cluster Id: 45794
Cluster Member: Yes
Cluster Generation: 1124
Membership state: Cluster-Member
Nodes: 1
Expected votes: 4
Total votes: 1
Quorum: 3 Activity blocked
Active subsystems: 2
Flags:
Ports Bound: 0
Node name: cevstream.ugr.es
Node ID: 4
Multicast addresses: 239.192.178.149
Node addresses: 150.214.243.19
Running "fence_tool_dump" on the three nodes:
[root@cev01 ~]# fence_tool dump
dump read: Success
1260778939 our_nodeid 2 our_name cev01.ugr.es
1260778939 listen 4 member 5 groupd 7
1260778964 client 3: join default
1260778964 delay post_join 3s post_fail 0s
1260778964 added 4 nodes from ccs
1260778964 setid default 65538
1260778964 start default 1 members 2
1260778964 do_recovery stop 0 start 1 finish 0
1260778964 node "cevstream.ugr.es" not a cman member, cn 1
1260778964 add first victim cevstream.ugr.es
1260778965 node "cevstream.ugr.es" not a cman member, cn 1
1260778966 node "cevstream.ugr.es" not a cman member, cn 1
1260778967 node "cevstream.ugr.es" not a cman member, cn 1
1260778967 delay of 3s leaves 1 victims
1260778967 node "cevstream.ugr.es" not a cman member, cn 1
1260778967 node "cevstream.ugr.es" has not been fenced
1260778967 fencing node cevstream.ugr.es
1260778971 finish default 1
1260778971 stop default
1260778971 start default 2 members 3 2
1260778971 do_recovery stop 1 start 2 finish 1
1260778971 finish default 2
1260778971 stop default
1260778971 start default 3 members 1 3 2
1260778971 do_recovery stop 2 start 3 finish 2
1260778971 finish default 3
1260779876 client 3: dump
while running it in the new node:
[root@cevstream ~]# fence_tool dump
fence_tool: can't communicate with fenced
I get a lot of errors telling me that cluster is not quorate:
Dec 14 09:39:20 cevstream ccsd[3668]: Cluster is not quorate. Refusing connection.
Dec 14 09:39:20 cevstream ccsd[3668]: Error while processing connect: Connection refused
Printing the superblock on any of the three nodes:
[root@cev01 ~]# gfs2_tool sb /dev/vg_cluster/lv_cluster all
mh_magic = 0x01161970
mh_type = 1
mh_format = 100
sb_fs_format = 1801
sb_multihost_format = 1900
sb_bsize = 4096
sb_bsize_shift = 12
no_formal_ino = 2
no_addr = 23
no_formal_ino = 1
no_addr = 22
sb_lockproto = lock_dlm
sb_locktable = campusvirtual:gfs_cluster01
uuid = C6A9FBB4-A881-2128-2AB8-1AB8547C7F30
I've tried something i saw in some forums, deactivating and even removing the logical volume (with lvremove), because supposedly the new node could need this operation in order to access the gfs2 volume.
Running lvcreate on the new node, with all the other nodes deactivates and removed, i still get the error:
[root@cevstream ~]# lvcreate -l 100%FREE -n lv_cluster vg_cluster
connect() failed on local socket: Conexión rehusada
WARNING: Falling back to local file-based locking.
Volume Groups with the clustered attribute will be inaccessible.
Skipping clustered volume group vg_cluster
Find attached the configuration of cluster.conf.
I'm pretty desperate with this situation, i really don't know how to deal with the adition of a new node.
Best regards,
Arturo.
Attachment:
cluster.conf
Description: Binary data
-- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster