Re: Hassle with clvmd over external network

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Patrick, thanks for your comment.

I can't see any messages regarding dlm in the syslog. When I call
`strace clvmd -d`, the first two lines are as followed:
[root@compute-0-1 ~]# strace clvmd -d
execve("/usr/sbin/clvmd", ["clvmd", "-d"], [/* 34 vars */]) = 0
uname({sys="Linux", node="compute-0-1.local", ...}) = 0
[...]

So, obviously, clvmd doesn't use the hostnames from cluster.conf (where
it should be c0-1.external.dns.com), but the ones provided by uname.
Does anybody know how to change this behaviour? Any help is greatly
appreciated!

Regards,
Sebastian


Patrick Caulfield wrote:
> Sebastian Walter wrote:
>   
>> Dear list,
>>
>> I'm trying to set up RHCS and GFS in a cluster which has two network
>> memberships, one is internal (eth0, 10.1.0.0/16, dns-names: host.local),
>> the other external (eth1, our real-world subnet and dns names). Every
>> cluster node has these two interfaces and related ip-numbers in both
>> networks.
>>
>> When I set up the RHCS and GFS on the local subnet, everything works
>> fine (ccsd, cman, clvmd, ... and also gfs mountable volumes). But if I
>> try to change cluster.conf to use the real-world addresses (I want to
>> use the gfs volumes also outside of the cluster), clvmd always makes
>> problems. I followed the faq and changed /etc/init.d/cman to connect
>> with -n host.external.dns.com. All hosts are in /etc/hosts. ccsd starts
>> well on all clients, as does cman and fenced. But when I try to start
>> the clvmd service on all nodes simultaneously, I get errors (Starting
>> clvmd: clvmd startup timed out).
>>
>> This is what my /proc gives me:
>> [root@dtm ~]# cat /proc/cluster/services
>> Service          Name                              GID LID State     Code
>> Fence Domain:    "default"                          11   2 run       -
>> [8 7 6 5 4 3 2 9 10 11 1]
>>
>> DLM Lock Space:  "clvmd"                            14   3 join     
>> S-6,20,11
>> [8 7 6 5 4 3 2 9 10 11 1]
>>
>> [root@compute-0-1 ~]# cat /proc/cluster/services
>> Service          Name                              GID LID State     Code
>> Fence Domain:    "default"                          11   2 run       -
>> [2 3 4 5 6 7 9 10 8 11 1]
>>
>> DLM Lock Space:  "clvmd"                            14   3 update    U-4,1,1
>> [2 3 4 5 6 7 8 9 10 11 1]
>>
>> (the second output I get from all the other nodes, I think it depends on
>> which host i start the service first on)
>>
>> Has anybody an idea how clvmd communicates to each other? cman is doing
>> fine... any other experiences? Thanks for any advice...
>>
>>     
>
>
> It'll be waiting for the DLM lockspace creation to complete on all nodes. Have a
> look in syslog for DLM messages.
>
>   

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux