Vladimir,
Hi, guys We have problems with corosync 1.4.x (1.4.6 and 1.4.7). This is the scenario: We have 3-node cluster: node-1 (10.108.2.3),node-2(10.108.2.4),node-3(10.108.2.5) in UDPU mode. We shut down the interface used for cluster communication on one of the nodes by issuing `ifdown eth2` command on node-1.
yes. This is known problem. Actually, there is a patch for corosync 2.x, but it's not yet reviewed and not yet backported for flatiron.
Just don't do ifdown or use NM in non server mode (does ifdown on cable unplug).
If you need to test failover ether block trafic on switch, unplug network cable (without NM or with NM but it had to be in server mode) or block traffic via iptables (but carefully, you have to block everything but not localhost...)
Regards, Honza
node-2 and node-3 find this and understand that node-1 left the cluster. runtime.totem.pg.mrp.srp.members.67267594.ip=r(0) ip(10.108.2.4) runtime.totem.pg.mrp.srp.members.67267594.join_count=1 runtime.totem.pg.mrp.srp.members.67267594.status=joined runtime.totem.pg.mrp.srp.members.84044810.ip=r(0) ip(10.108.2.5) runtime.totem.pg.mrp.srp.members.84044810.join_count=2 runtime.totem.pg.mrp.srp.members.84044810.status=joined runtime.totem.pg.mrp.srp.members.50490378.ip=r(0) ip(10.108.2.3) runtime.totem.pg.mrp.srp.members.50490378.join_count=1 runtime.totem.pg.mrp.srp.members.50490378.status=left But node-1 does not get it and thinks that node-2 and node-3 are still online: runtime.totem.pg.mrp.srp.firewall_enabled_or_nic_failure=1 runtime.totem.pg.mrp.srp.members.50490378.ip=r(0) ip(10.108.2.3) runtime.totem.pg.mrp.srp.members.50490378.join_count=1 runtime.totem.pg.mrp.srp.members.50490378.status=joined runtime.totem.pg.mrp.srp.members.67267594.ip=r(0) ip(10.108.2.4) runtime.totem.pg.mrp.srp.members.67267594.join_count=1 runtime.totem.pg.mrp.srp.members.67267594.status=joined runtime.totem.pg.mrp.srp.members.84044810.ip=r(0) ip(10.108.2.5) runtime.totem.pg.mrp.srp.members.84044810.join_count=1 runtime.totem.pg.mrp.srp.members.84044810.status=joined runtime.blackbox.dump_flight_data=no runtime.blackbox.dump_state=no In node-1 logs I see the following: 2014-08-13T15:46:18.848234+01:00 warning: [MAIN ] Totem is unable to form a cluster because of an operating system or network fault. The most common cause of this message is that the local firewall is configured improperly. 2014-08-13T15:46:18.866365+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.866799+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.866799+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.866799+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.866799+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.866799+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.935539+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.935932+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.935932+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.935932+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) 2014-08-13T15:46:18.935932+01:00 debug: [TOTEM ] sendmsg(mcast) failed (non-critical): Invalid argument (22) I am pretty sure that this is the bug as corosync should detect ring failure and mark dead nodes as dead on both sides. Can you help me fix it or provide with the clue where to look for the fix? _______________________________________________ discuss mailing list discuss@xxxxxxxxxxxx http://lists.corosync.org/mailman/listinfo/discuss
_______________________________________________ discuss mailing list discuss@xxxxxxxxxxxx http://lists.corosync.org/mailman/listinfo/discuss