Re: clvmd problems with centos 6.3 or normal clvmd behaviour?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Aug 1, 2012 at 6:15 PM, Gianluca Cecchi wrote:
> On Wed, 1 Aug 2012 16:26:38 +0200 emmanuel segura wrote:
>> Why you don't remove expected_votes=3 and let the cluster automatic calculate that
>
> Thanks for your answer Emmanuel, but cman starts correctly, while the
> problem seems related to
> vgchange -aly
> command hanging.
> But I tried that option too and the cluster hangs at the same point as before.

Further testing shows that cluster is indeed quorated and problems are
related to lvm...

I also tried following a more used and clean configuration seen in
examples for 3 nodes + quorum daemon:

2 votes for each node
<clusternode name="nodeX" nodeid="X" votes="2">

3 votes for quorum disk
<quorumd device="/dev/mapper/mpathquorum" interval="5"
label="clrhevquorum" tko="24" votes="3">

with and without expected_votes="9" in <cman ... /> part

One node + its quorum only config should be ok (2+3 = 5 votes)

After cman starts and quorumd is not master yet:

# cman_tool status
Version: 6.2.0
Config Version: 51
Cluster Name: clrhev
Cluster Id: 43203
Cluster Member: Yes
Cluster Generation: 1428
Membership state: Cluster-Member
Nodes: 1
Expected votes: 9
Total votes: 2
Node votes: 2
Quorum: 5 Activity blocked
Active subsystems: 4
Flags:
Ports Bound: 0 178
Node name: intrarhev3
Node ID: 3
Multicast addresses: 239.192.168.108
Node addresses: 192.168.16.30

Then
# cman_tool status
Version: 6.2.0
Config Version: 51
Cluster Name: clrhev
Cluster Id: 43203
Cluster Member: Yes
Cluster Generation: 1428
Membership state: Cluster-Member
Nodes: 1
Expected votes: 9
Quorum device votes: 3
Total votes: 5
Node votes: 2
Quorum: 5
Active subsystems: 4
Flags:
Ports Bound: 0 178
Node name: intrarhev3
Node ID: 3
Multicast addresses: 239.192.168.108
Node addresses: 192.168.16.30

And startup continues up to clvmd step
In this phase, while clvmd startup hanges forever I have:

# dlm_tool ls
dlm lockspaces
name          clvmd
id            0x4104eefa
flags         0x00000000
change        member 1 joined 1 remove 0 failed 0 seq 1,1
members       3

# ps -ef|grep lv
root      3573  2593  0 01:05 ?        00:00:00 /bin/bash
/etc/rc3.d/S24clvmd start
root      3578     1  0 01:05 ?        00:00:00 clvmd -T30
root      3620     1  0 01:05 ?        00:00:00 /sbin/lvm pvscan
--cache --major 253 --minor 13
root      3804  3322  0 01:09 pts/0    00:00:00 grep lv

# ps -ef|grep vg
root      3601  3573  0 01:05 ?        00:00:00 /sbin/vgchange -ayl
root      3808  3322  0 01:09 pts/0    00:00:00 grep vg

# ps -ef|grep lv
root      3573  2593  0 01:05 ?        00:00:00 /bin/bash
/etc/rc3.d/S24clvmd start
root      3578     1  0 01:05 ?        00:00:00 clvmd -T30
root      4008  3322  0 01:13 pts/0    00:00:00 grep lv

# ps -ef|grep 3578
root      3578     1  0 01:05 ?        00:00:00 clvmd -T30
root      4017  3322  0 01:13 pts/0    00:00:00 grep 3578

It remains at
# service clvmd start
Starting clvmd:
Activating VG(s):   3 logical volume(s) in volume group "VG_VIRT02" now active

Is there any way to debug clvmd?
I suppose it communicates through intracluster, correct?
tcpdump output could be of any help?

Any one already passed to 6.3 (on rhel and/or CentOS) and having all
ok with clvmd?

BTW: I also tried lvmetad, that is tech preview in 6.3, enabling its
service and putting "use_lvmetad = 1" in lvm.conf but without luck...

Thanks in advance

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster


[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux