Re: Cluster environment issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, 30 May 2011 18:22:00 -0700 (PDT), Srija <swap_project@xxxxxxxxx>
wrote:
> Thanks for your quick reply.
> 
> I talked to the network people , but they are saying everything is good
at
> their end. Is there anyway at the server end, to figure it  for the
switch
> restart or multicast traffic? 
> 

If it is a switch restart you will have in your logs the interface going
down/up, but more problematic is to find a short drop of the multicast
traffic (even with a ping script you may miss it) which is more likely the
case, as your cluster is working fine, but suddenly looses connection to
all nodes at the same time.
You may ask the network people to check for STP changes and double check
the multicast configuration and you may also try to use broadcast instead
of multicast or use a dedicated switch.

> I think you have already checked the cluster.conf file.. Except quorum
> disk, do you think that the cluster configuration is sufficient for
> handling the sixteen node cluster!! 
> 

The config is OK ... probably add specific multicast address in the cman
section to avoid surprises, but the default is also fine.

To confirm it is a multicast drop (if you are lucky not ti miss it) - on
just one of the nodes enable icmp broadcasts:
	echo 0 >/proc/sys/net/ipv4/icmp_echo_ignore_broadcasts
then ping from another node, check if just a single one replies (change to
your interface and multicast address)
	ping -I ethX -b -L 239.x.x.x -c 1
and finaly run this script until the cluster gets broken

	while [ $((`ping -I ethX -w 1 -b -L 239.x.x.x -c 1 | grep -c ' 0% packet
loss'`)) -eq 1 ]; do sleep 1; done; echo "missed ping at `date`"

if you get 'missed ping' at the same when cluster goes down - it is
confirmed :)

> thanks again .
> regards
> 
> --- On Mon, 5/30/11, Kaloyan Kovachev <kkovachev@xxxxxxxxx> wrote:
> 
>> From: Kaloyan Kovachev <kkovachev@xxxxxxxxx>
>> Subject: Re:  Cluster environment issue
>> To: "linux clustering" <linux-cluster@xxxxxxxxxx>
>> Date: Monday, May 30, 2011, 4:05 PM
>> Hi,
>>  when your cluster gets broken, most likely the reason is,
>> there is a
>> network problem (switch restart or multicast traffic is
>> lost for a while)
>> on the interface where serverX-priv IPs are configured.
>> Having a quorum
>> disk may help by giving a quorum vote to one of the
>> servers, so it can
>> fence the others, but the best thing to do is to fix your
>> network and
>> preferably add a redundant link for the cluster
>> communication to avoid
>> breakage in the first place
>> 
>> On Mon, 30 May 2011 12:17:07 -0700 (PDT), Srija
<swap_project@xxxxxxxxx>
>> wrote:
>> > Hi,
>> > 
>> > I am very new to the redhat cluster. Need some help
>> and suggession for
>> the
>> > cluster configuration.
>> > We have sixteen node cluster of 
>> > 
>> >             OS
>> : Linux Server release 5.5 (Tikanga)
>> >         
>>    kernel :  2.6.18-194.3.1.el5xen.
>> > 
>> > The problem is sometimes the cluster is getting 
>> broken. The solution is
>> > (still yet)to reboot the 
>> > sixteen nodes. Otherwise the nodes are not joining
>> > 
>> > We are using  clvm and not using any quorum disk.
>> The quorum is by
>> default.
>> > 
>> > When it is getting broken, clustat commands
>> shows  evrything  offline
>> > except the node from where
>> > the clustat command executed.  If we execute vgs,
>> lvs command, those
>> > commands are getting hung.
>> > 
>> > Here is at present the clustat report
>> > -------------------------------------
>> > 
>> > [server1]# clustat
>> > Cluster Status for newcluster @ Mon May 30 14:55:10
>> 2011
>> > Member Status: Quorate
>> > 
>> >  Member Name         
>>            
>> ID   Status
>> >  ------ ----         
>>             ---- ------
>> >  server1           
>>               1 Online
>> >  server2           
>>               2 Online,
>> Local
>> >  server3           
>>               3 Online
>> >  server4           
>>               4 Online
>> >  server5           
>>               5 Online
>> >  server6           
>>               6 Online
>> >  server7           
>>               7 Online
>> >  server8           
>>               8 Online
>> >  server9           
>>               9 Online
>> >  server10         
>>            
>>    10 Online
>> >  server11         
>>            
>>    11 Online
>> >  server12         
>>            
>>    12 Online
>> >  server13         
>>            
>>    13 Online
>> >  server14         
>>            
>>    14 Online
>> >  server15         
>>            
>>    15 Online
>> >  server16         
>>            
>>    16 Online
>> > 
>> > Here the cman_tool status  output  from one
>> server
>> > --------------------------------------------------
>> > 
>> > [server1 ~]# cman_tool status
>> > Version: 6.2.0
>> > Config Version: 23
>> > Cluster Name: newcluster
>> > Cluster Id: 53322
>> > Cluster Member: Yes
>> > Cluster Generation: 11432
>> > Membership state: Cluster-Member
>> > Nodes: 16
>> > Expected votes: 16
>> > Total votes: 16
>> > Quorum: 9  
>> > Active subsystems: 8
>> > Flags: Dirty 
>> > Ports Bound: 0 11  
>> > Node name: server1
>> > Node ID: 1
>> > Multicast addresses: xxx.xxx.xxx.xx 
>> > Node addresses: 192.168.xxx.xx 
>> > 
>> > 
>> > Here is the cluster.conf file.
>> > ------------------------------
>> > 
>> > <?xml version="1.0"?>
>> > <cluster alias="newcluster" config_version="23"
>> name="newcluster">
>> > <fence_daemon clean_start="1" post_fail_delay="0"
>> post_join_delay="15"/>
>> > 
>> > <clusternodes>
>> > 
>> > <clusternode name="server1-priv" nodeid="1"
>> votes="1">
>> >               
>>   <fence><method name="1">
>> >               
>>   <device name="ilo-server1r"/></method>
>> >               
>>   </fence>
>> > </clusternode>
>> > 
>> > <clusternode name="server2-priv" nodeid="3"
>> votes="1">
>> >     
>>    <fence><method name="1">
>> >         <device
>> name="ilo-server2r"/></method>
>> >         </fence>
>> > </clusternode>
>> > 
>> > <clusternode name="server3-priv" nodeid="2"
>> votes="1">
>> >     
>>    <fence><method name="1">
>> >         <device
>> name="ilo-server3r"/></method>
>> >         </fence>
>> > </clusternode>
>> > 
>> > [ ... sinp .....]
>> > 
>> > <clusternode name="server16-priv" nodeid="16"
>> votes="1">
>> >        <fence><method
>> name="1">
>> >        <device
>> name="ilo-server16r"/></method>
>> >        </fence>
>> > </clusternode>
>> > 
>> > </clusternodes>
>> > <cman/>
>> > 
>> > <dlm plock_ownership="1" plock_rate_limit="0"/>
>> > <gfs_controld plock_rate_limit="0"/>
>> > 
>> > <fencedevices>
>> >         <fencedevice
>> agent="fence_ilo" hostname="server1r" login="Admin"
>> >     
>>    name="ilo-server1r" passwd="xxxxx"/>
>> >         ..........
>> >         <fencedevice
>> agent="fence_ilo" hostname="server16r"
>> login="Admin"
>> >     
>>    name="ilo-server16r" passwd="xxxxx"/>
>> > </fencedevices>
>> > <rm>
>> > <failoverdomains/>
>> > <resources/>
>> > </rm></cluster>
>> > 
>> > Here is the lvm.conf file
>> > --------------------------
>> > 
>> > devices {
>> > 
>> >     dir = "/dev"
>> >     scan = [ "/dev" ]
>> >     preferred_names = [ ]
>> >     filter = [
>> "r/scsi.*/","r/pci.*/","r/sd.*/","a/.*/" ]
>> >     cache_dir = "/etc/lvm/cache"
>> >     cache_file_prefix = ""
>> >     write_cache_state = 1
>> >     sysfs_scan = 1
>> >     md_component_detection = 1
>> >     md_chunk_alignment = 1
>> >     data_alignment_detection = 1
>> >     data_alignment = 0
>> > 
>>    data_alignment_offset_detection = 1
>> >     ignore_suspended_devices = 0
>> > }
>> > 
>> > log {
>> > 
>> >     verbose = 0
>> >     syslog = 1
>> >     overwrite = 0
>> >     level = 0
>> >     indent = 1
>> >     command_names = 0
>> >     prefix = "  "
>> > }
>> > 
>> > backup {
>> > 
>> >     backup = 1
>> >     backup_dir =
>> "/etc/lvm/backup"
>> >     archive = 1
>> >     archive_dir =
>> "/etc/lvm/archive"
>> >     retain_min = 10
>> >     retain_days = 30
>> > }
>> > 
>> > shell {
>> > 
>> >     history_size = 100
>> > }
>> > global {
>> >     library_dir = "/usr/lib64"
>> >     umask = 077
>> >     test = 0
>> >     units = "h"
>> >     si_unit_consistency = 0
>> >     activation = 1
>> >     proc = "/proc"
>> >     locking_type = 3
>> >     wait_for_locks = 1
>> >     fallback_to_clustered_locking
>> = 1
>> >     fallback_to_local_locking = 1
>> >     locking_dir = "/var/lock/lvm"
>> >     prioritise_write_locks = 1
>> > }
>> > 
>> > activation {
>> >     udev_sync = 1
>> >     missing_stripe_filler =
>> "error"
>> >     reserved_stack = 256
>> >     reserved_memory = 8192
>> >     process_priority = -18
>> >     mirror_region_size = 512
>> >     readahead = "auto"
>> >     mirror_log_fault_policy =
>> "allocate"
>> >     mirror_image_fault_policy =
>> "remove"
>> > }
>> > dmeventd {
>> > 
>> >     mirror_library =
>> "libdevmapper-event-lvm2mirror.so"
>> >     snapshot_library =
>> "libdevmapper-event-lvm2snapshot.so"
>> > }
>> > 
>> > 
>> > If you need more  information,  I can
>> provide ...
>> > 
>> > Thanks for your help
>> > Priya
>> > 
>> > --
>> > Linux-cluster mailing list
>> > Linux-cluster@xxxxxxxxxx
>> > https://www.redhat.com/mailman/listinfo/linux-cluster
>> 
>> --
>> Linux-cluster mailing list
>> Linux-cluster@xxxxxxxxxx
>> https://www.redhat.com/mailman/listinfo/linux-cluster
>> 
> 
> --
> Linux-cluster mailing list
> Linux-cluster@xxxxxxxxxx
> https://www.redhat.com/mailman/listinfo/linux-cluster

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster


[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux