Re: Linux-cluster Digest, Vol 75, Issue 19

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




Rajkumar, Anoop wrote:
 
Hi

Cluster is using private subnet address as node name and cman shows
following information on thir node.

[root@usrylxap235 ~]# cman_tool status
Version: 6.1.0
Config Version: 44
Cluster Name: cluster1
Cluster Id: 26777
Cluster Member: Yes
Cluster Generation: 88
Membership state: Cluster-Member
Nodes: 1
Expected votes: 3
Total votes: 1
Quorum: 2 Activity blocked
Active subsystems: 5
Flags:
Ports Bound: 0
Node name: usrylxap235-1.merck.com
Node ID: 3
Multicast addresses: 239.192.104.2
Node addresses: 192.168.0.7

Two active nodes have following info.

[root@usrylxap237 net]# cman_tool status
Version: 6.1.0
Config Version: 44
Cluster Name: cluster1
Cluster Id: 26777
Cluster Member: Yes
Cluster Generation: 296
Membership state: Cluster-Member
Nodes: 2
Expected votes: 3
Total votes: 2
Quorum: 2
Active subsystems: 8
Flags: Dirty
Ports Bound: 0 177
Node name: usrylxap237-1.merck.com
Node ID: 1
Multicast addresses: 239.192.104.2
Node addresses: 192.168.0.2

All the nodes have private network connected to foundry switch. How I
make sure that multicast traffic is going through private network which
is eth5 for third node, eth0 for first and second node?
  
netstat -gn shows multicast groups and the interfaces they're linked to.
Ping the multicast address from one node and run "tcpdump -ni ethx icmp" on the other node to see if multicast traffic is reaching the node. Then repeat the process in the other direction.

Must have icmp traffic allowed between nodes first. If you don't want to test icmp and only want to test TCP/UDP, use netcat.

Regards,
Dan.
Thanks
Anoop 
-----Original Message-----
From: linux-cluster-bounces@xxxxxxxxxx
[mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of
linux-cluster-request@xxxxxxxxxx
Sent: Sunday, July 18, 2010 12:00 PM
To: linux-cluster@xxxxxxxxxx
Subject: Linux-cluster Digest, Vol 75, Issue 19

Send Linux-cluster mailing list submissions to
	linux-cluster@xxxxxxxxxx

To subscribe or unsubscribe via the World Wide Web, visit
	https://www.redhat.com/mailman/listinfo/linux-cluster
or, via email, send a message with subject or body 'help' to
	linux-cluster-request@xxxxxxxxxx

You can reach the person managing the list at
	linux-cluster-owner@xxxxxxxxxx

When replying, please edit your Subject line so it is more specific
than "Re: Contents of Linux-cluster digest..."


Today's Topics:

   1. Re: Adding third node to existing cluster (Marti, Robert)
   2. Re: How do i stop VM's on a failed node? (Volker Dormeyer)


----------------------------------------------------------------------

Message: 1
Date: Sat, 17 Jul 2010 22:10:52 -0500
From: "Marti, Robert" <RJM002@xxxxxxxx>
To: linux clustering <linux-cluster@xxxxxxxxxx>
Subject: Re:  Adding third node to existing cluster
Message-ID:
	<8FAC1E47484E43469AA28DBF35C955E4BDE0B47EAC@xxxxxxxxxxxxxx>
Content-Type: text/plain; charset="us-ascii"


  
On Fri, Jul 16, 2010 at 2:06 AM, Rajkumar, Anoop
    
<anoop_rajkumar@xxxxxxxxx> wrote:
  
Hi

Here are the firewall settings

#more /etc/sysconfig/iptables
# Generated by iptables-save v1.3.5 on Wed Jul 14 10:13:12 2010
*filter
:INPUT ACCEPT [2:186]
:FORWARD ACCEPT [0:0]
:OUTPUT ACCEPT [4:486]
-A INPUT -i eth5 -p udp -m udp --dport 5405 -j ACCEPT
-A INPUT -i eth5 -p udp -m udp --sport 5405 -j ACCEPT
-A INPUT -i eth0 -p tcp -m tcp --dport 14567 -j ACCEPT
-A INPUT -i eth0 -p tcp -m tcp --sport 14567 -j ACCEPT
-A INPUT -i eth0 -p tcp -m tcp --dport 16851 -j ACCEPT
-A INPUT -i eth0 -p tcp -m tcp --sport 16851 -j ACCEPT
-A INPUT -i eth5 -p udp -m udp --dport 50007 -j ACCEPT
-A INPUT -i eth5 -p udp -m udp --sport 50007 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --dport 11111 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --sport 11111 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --dport 21064 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --sport 21064 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --dport 50009 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --sport 50009 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --dport 50008 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --sport 50008 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --dport 50006 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --sport 50006 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --dport 41969 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --sport 41969 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --dport 41968 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --sport 41968 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --dport 41967 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --sport 41967 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --dport 41966 -j ACCEPT
-A INPUT -i eth5 -p tcp -m tcp --sport 41966 -j ACCEPT
-A OUTPUT -o eth5 -p udp -m udp --sport 5405 -j ACCEPT
-A OUTPUT -o eth5 -p udp -m udp --dport 5405 -j ACCEPT
-A OUTPUT -o eth0 -p tcp -m tcp --sport 14567 -j ACCEPT
-A OUTPUT -o eth0 -p tcp -m tcp --dport 14567 -j ACCEPT
-A OUTPUT -o eth0 -p tcp -m tcp --sport 16851 -j ACCEPT
-A OUTPUT -o eth0 -p tcp -m tcp --dport 16851 -j ACCEPT
-A OUTPUT -o eth5 -p udp -m udp --sport 50007 -j ACCEPT
-A OUTPUT -o eth5 -p udp -m udp --dport 50007 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --sport 11111 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --dport 11111 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --sport 21064 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --dport 21064 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --sport 50009 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --dport 50009 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --sport 50008 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --dport 50008 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --sport 50006 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --dport 50006 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --sport 41969 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --dport 41969 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --sport 41968 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --dport 41968 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --sport 41967 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --dport 41967 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --sport 41966 -j ACCEPT
-A OUTPUT -o eth5 -p tcp -m tcp --dport 41966 -j ACCEPT
COMMIT

Here are the IP in the system,I am using eth5 (Which is in private
    
network
  
with other two nodes, connected to switch)

[root@usrylxap235 ~]# ip addr list
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 16436 qdisc noqueue
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
    inet 127.0.0.1/8 scope host lo
    inet6 ::1/128 scope host
       valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
    
qlen
  
1000
    link/ether 00:1c:c4:f0:bd:d8 brd ff:ff:ff:ff:ff:ff
    inet 54.3.254.235/24 brd 54.3.254.255 scope global eth0
    inet6 fe80::21c:c4ff:fef0:bdd8/64 scope link
       valid_lft forever preferred_lft forever
3: eth1: <BROADCAST,MULTICAST> mtu 1500 qdisc noop qlen 1000
    link/ether 00:1c:c4:f0:bd:da brd ff:ff:ff:ff:ff:ff
4: eth2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop qlen 1000
    link/ether 00:1c:c4:5e:f8:d8 brd ff:ff:ff:ff:ff:ff
5: eth3: <BROADCAST,MULTICAST> mtu 1500 qdisc noop qlen 1000
    link/ether 00:1c:c4:5e:f8:da brd ff:ff:ff:ff:ff:ff
6: eth4: <BROADCAST,MULTICAST> mtu 1500 qdisc noop qlen 1000
    link/ether 00:1e:0b:71:ac:6c brd ff:ff:ff:ff:ff:ff
7: eth5: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
    
qlen
  
1000
    link/ether 00:1e:0b:71:ac:6e brd ff:ff:ff:ff:ff:ff
    inet 192.168.0.7/24 brd 192.168.0.255 scope global eth5
    inet6 fe80::21e:bff:fe71:ac6e/64 scope link
       valid_lft forever preferred_lft forever
8: sit0: <NOARP> mtu 1480 qdisc noop
    link/sit 0.0.0.0 brd 0.0.0.0

Thanks
Anoop
    

Hi,

Let's see if we can rule out security or network issues here. Are you
able to allow all traffic for eth5 on all nodes? In addition, you may
want to add a static multicast route to ensure the multicast traffic
is going through eth5 on all nodes.

Regards,
Bernard

-----------------------------------------------------------------------

He's essentially running a wide open firewall with those rules.  Default
ACCEPT on all chains, and no DROPs or REJECTs.  It's not a firewall
issue.  It may be multicast not going out the right interface, however.



------------------------------

Message: 2
Date: Sun, 18 Jul 2010 11:28:53 +0200
From: Volker Dormeyer <volker@xxxxxxxxxxxx>
To: linux clustering <linux-cluster@xxxxxxxxxx>
Subject: Re:  How do i stop VM's on a failed node?
Message-ID: <20100718092853.GA4159@dijkstra>
Content-Type: text/plain; charset=us-ascii

Hi,

On Fri, Jul 16, 2010 at 01:21:57PM -0400,
Nathan Lager <lagern@xxxxxxxxxxxxx> wrote:
  
I have a 4-node cluster, running KVM, and a host of VM's.

One of the nodes failed unexpectedly.  I'm having two issues.  First,
the VM's which were on that node are still reported as up and running
    
on
  
that node in clustat, and second, i'm unable to interact with the
cluster using clusvcadm.  Every command hangs.  I've tried disabling
these VM's, and the commands hang.

How can i clear out this dead node, without having to forceably
    
restart
  
my entire cluster?
    

To me, it sounds like fencing is not configured, properly.

What has been logged on the remaining nodes? I would assume, they tried
to
fence the failed node - but are not able to.

Log-Snippets and Config would be helpful.

Regards,
Volker



------------------------------

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

End of Linux-cluster Digest, Vol 75, Issue 19
*********************************************
Notice:  This e-mail message, together with any attachments, contains
information of Merck & Co., Inc. (One Merck Drive, Whitehouse Station,
New Jersey, USA 08889), and/or its affiliates Direct contact information
for affiliates is available at 
http://www.merck.com/contact/contacts.html) that may be confidential,
proprietary copyrighted and/or legally privileged. It is intended solely
for the use of the individual or entity named on this message. If you are
not the intended recipient, and have received this message in error,
please notify us immediately by reply e-mail and then delete it from 
your system.


--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster
  

-- 
Dan FRINCU
Systems Engineer
CCNA, RHCE
Streamwide Romania
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux