Re: Problem with join 1.4.2 on multicast

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Jan,

2013/8/23 Jan Friesse <jfriesse@xxxxxxxxxx>
Andrew,
please make really sure that mcast works. You can use for example omping
for that
 
I used  typical ping and tcpdump:

root@nodeA:~# ping 224.0.0.187
PING 224.0.0.187 (224.0.0.187) 56(84) bytes of data.
^C
--- 224.0.0.187 ping statistics ---
5 packets transmitted, 0 received, 100% packet loss, time 4031ms


root@nodeB:~# tcpdump -n -i eth0 host  224.0.0.187
tcpdump: verbose output suppressed, use -v or -vv for full protocol decode
listening on eth0, link-type EN10MB (Ethernet), capture size 65535 bytes
11:48:52.665093 IP 10.40.129.209 > 224.0.0.187: ICMP echo request, id 7476, seq 1, length 64
11:48:53.672241 IP 10.40.129.209 > 224.0.0.187: ICMP echo request, id 7476, seq 2, length 64
11:48:54.680228 IP 10.40.129.209 > 224.0.0.187: ICMP echo request, id 7476, seq 3, length 64
11:48:55.688198 IP 10.40.129.209 > 224.0.0.187: ICMP echo request, id 7476, seq 4, length 64
11:48:56.696178 IP 10.40.129.209 > 224.0.0.187: ICMP echo request, id 7476, seq 5, length 64

trafic is visible, as well as in oposite direction

root@nodeB:~# ping 224.0.0.187
PING 224.0.0.187 (224.0.0.187) 56(84) bytes of data.
^C
--- 224.0.0.187 ping statistics ---
4 packets transmitted, 0 received, 100% packet loss, time 3025ms

root@nodeA:~# tcpdump -n -i eth0 host  224.0.0.187
tcpdump: verbose output suppressed, use -v or -vv for full protocol decode
listening on eth0, link-type EN10MB (Ethernet), capture size 65535 bytes
11:12:22.493138 IP 10.40.129.210 > 224.0.0.187: ICMP echo request, id 6855, seq 1, length 64
11:12:23.502201 IP 10.40.129.210 > 224.0.0.187: ICMP echo request, id 6855, seq 2, length 64
11:12:24.510161 IP 10.40.129.210 > 224.0.0.187: ICMP echo request, id 6855, seq 3, length 64
11:12:25.518215 IP 10.40.129.210 > 224.0.0.187: ICMP echo request, id 6855, seq 4, length 64


 
What seems to be strange is especially node A. If you compare your node
A output:

> 10:33:57.364852 IP 10.40.129.209.5464 > 224.0.0.187.5465: UDP, length 82
> 10:33:57.599081 IP 10.40.129.209.5464 > 224.0.0.187.5465: UDP, length 82

with node B output:

> 10:33:57.612370 IP 10.40.129.210.5464 > 224.0.0.187.5465: UDP, length 112
> 10:33:57.612473 IP 10.40.129.209.5464 > 224.0.0.187.5465: UDP, length 156

you can see node B sees both 10.40.129.210.5464 and 10.40.129.209.5464,
but node A sees only 10.40.129.209.5464

Sorry, I didn't write it cleary, all output from tcpdump on previous mail is from nodeA,  the first 9 packets with length 82 are captured before I started corosync on nodeB, and rest of this packet is captured when both instances of corosync are up. 

as you can see when I started corosync instance on nodeB, I captured packet from nodeB:

10:33:57.612370 IP 10.40.129.210.5464 > 224.0.0.187.5465: UDP, length 112

on nodeA, so multicast traffic is working well (i think).
 
You can also try to turn on debugging (logging debug: on in config file)
and see logs.


I turned on debug option, and done test:

1)started corosync on nodeA
2)clear logs on nodeA
3) started corosync on nodeB

There is no logs on nodeA since start NodeB(strange), there should be something if corosync on nodeA received pacekt from nodeB
below is log from nodeB, 

I cant see anything strange in that logs, below logs from nodeB

Aug 23 11:58:15 nodeB corosync[6874]:   [MAIN  ] Corosync Cluster Engine ('1.4.2'): started and ready to provide service.
Aug 23 11:58:15 nodeB corosync[6874]:   [MAIN  ] Corosync built-in features: nss
Aug 23 11:58:15 nodeB corosync[6874]:   [MAIN  ] Successfully read main configuration file '/etc/corosync/corosync.conf'.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Token Timeout (3000 ms) retransmit timeout (294 ms)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] token hold (225 ms) retransmits before loss (10 retrans)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] join (60 ms) send_join (0 ms) consensus (3600 ms) merge (200 ms)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] downcheck (1000 ms) fail to recv const (2500 msgs)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] seqno unchanged const (30 rotations) Maximum network MTU 1402
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] window size per rotation (50 messages) maximum messages per rotation (20 messages)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] missed count const (5 messages)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] send threads (0 threads)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] RRP token expired timeout (294 ms)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] RRP token problem counter (2000 ms)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] RRP threshold (10 problem count)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] RRP multicast threshold (100 problem count)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] RRP automatic recovery check timeout (1000 ms)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] RRP mode set to none.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] heartbeat_failures_allowed (0)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] max_network_delay (50 ms)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] HeartBeat is Disabled. To enable set heartbeat_failures_allowed > 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Initializing transport (UDP/IP Multicast).
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Initializing transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0).
Aug 23 11:58:15 nodeB corosync[6874]:   [IPC   ] you are using ipc api v2
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Receive multicast socket recv buffer size (262142 bytes).
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Transmit multicast socket send buffer size (262142 bytes).
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] The network interface [10.40.129.210] is now up.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Created or loaded sequence id a4.10.40.129.210 for this ring.
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: pcmk_user_lookup: Cluster user root has uid=0 gid=0
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: process_ais_conf: Reading configure
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: config_find_init: Local handle: 4730966301143465987 for logging
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: config_find_next: Processing additional logging options...
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: get_config_opt: Found 'on' for option: debug
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: get_config_opt: Found 'no' for option: to_logfile
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: get_config_opt: Found 'yes' for option: to_syslog
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: get_config_opt: Found 'daemon' for option: syslog_facility
Aug 23 11:58:15 nodeB cib: [6882]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/hacluster
Aug 23 11:58:15 nodeB cib: [6882]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.xml (digest: /var/lib/heartbeat/crm/cib.xml.sig)
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk] <cib epoch="5" num_updates="0" admin_epoch="0" validate-with="pacemaker-1.2" cib-last-written="Fri Aug 23 11:23:56 2013" crm_feature_set="3.0.6" update-origin="nodeA" update-client="crmd" have-quorum="1" >
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]   <configuration >
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]     <crm_config >
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]       <cluster_property_set id="cib-bootstrap-options" >
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]         <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff" />
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]         <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="openais" />
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]         <nvpair id="cib-bootstrap-options-expected-quorum-votes" name="expected-quorum-votes" value="2" />
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]       </cluster_property_set>
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]     </crm_config>
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]     <nodes >
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]       <node id="nodeA" uname="nodeA" type="normal" />
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]       <node id="nodeB" uname="nodeB" type="normal" />
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]     </nodes>
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]     <resources />
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]     <constraints />
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]   </configuration>
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk]   <status />
Aug 23 11:58:15 nodeB cib: [6882]: debug: readCibXmlFile: [on-disk] </cib>
Aug 23 11:58:15 nodeB cib: [6882]: info: validate_with_relaxng: Creating RNG parser context
Aug 23 11:58:15 nodeB cib: [6882]: debug: activateCibXml: Triggering CIB write for start op
Aug 23 11:58:15 nodeB cib: [6882]: info: startCib: CIB Initialization completed successfully
Aug 23 11:58:15 nodeB cib: [6882]: info: get_cluster_type: Cluster type is: 'openais'
Aug 23 11:58:15 nodeB cib: [6882]: notice: crm_cluster_connect: Connecting to cluster infrastructure: classic openais (with plugin)
Aug 23 11:58:15 nodeB cib: [6882]: info: init_ais_connection_classic: Creating connection to our Corosync plugin
Aug 23 11:58:15 nodeB stonith-ng: [6883]: info: Invoked: /usr/lib/pacemaker/stonithd
Aug 23 11:58:15 nodeB stonith-ng: [6883]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/root
Aug 23 11:58:15 nodeB stonith-ng: [6883]: info: get_cluster_type: Cluster type is: 'openais'
Aug 23 11:58:15 nodeB stonith-ng: [6883]: notice: crm_cluster_connect: Connecting to cluster infrastructure: classic openais (with plugin)
Aug 23 11:58:15 nodeB stonith-ng: [6883]: info: init_ais_connection_classic: Creating connection to our Corosync plugin
Aug 23 11:58:15 nodeB lrmd: [6884]: info: enabling coredumps
Aug 23 11:58:15 nodeB lrmd: [6884]: WARN: Core dumps could be lost if multiple dumps occur.
Aug 23 11:58:15 nodeB lrmd: [6884]: WARN: Consider setting non-default value in /proc/sys/kernel/core_pattern (or equivalent) for maximum supportability
Aug 23 11:58:15 nodeB lrmd: [6884]: WARN: Consider setting /proc/sys/kernel/core_uses_pid (or equivalent) to 1 for maximum supportability
Aug 23 11:58:15 nodeB lrmd: [6884]: debug: main: run the loop...
Aug 23 11:58:15 nodeB lrmd: [6884]: info: Started.
Aug 23 11:58:15 nodeB attrd: [6885]: info: Invoked: /usr/lib/pacemaker/attrd
Aug 23 11:58:15 nodeB attrd: [6885]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/hacluster
Aug 23 11:58:15 nodeB pengine: [6886]: info: Invoked: /usr/lib/pacemaker/pengine
Aug 23 11:58:15 nodeB attrd: [6885]: info: main: Starting up
Aug 23 11:58:15 nodeB attrd: [6885]: info: get_cluster_type: Cluster type is: 'openais'
Aug 23 11:58:15 nodeB attrd: [6885]: notice: crm_cluster_connect: Connecting to cluster infrastructure: classic openais (with plugin)
Aug 23 11:58:15 nodeB attrd: [6885]: info: init_ais_connection_classic: Creating connection to our Corosync plugin
Aug 23 11:58:15 nodeB pengine: [6886]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/hacluster
Aug 23 11:58:15 nodeB pengine: [6886]: debug: main: Checking for old instances of pengine
Aug 23 11:58:15 nodeB pengine: [6886]: debug: init_client_ipc_comms_nodispatch: Attempting to talk on: /var/run/crm/pengine
Aug 23 11:58:15 nodeB pengine: [6886]: debug: init_client_ipc_comms_nodispatch: Could not init comms on: /var/run/crm/pengine
Aug 23 11:58:15 nodeB pengine: [6886]: debug: main: Init server comms
Aug 23 11:58:15 nodeB pengine: [6886]: info: main: Starting pengine
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: config_find_init: Local handle: 7739444317642555396 for quorum
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: config_find_next: No additional configuration supplied for: quorum
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: get_config_opt: No default for option: provider
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: config_find_init: Local handle: 5650605097994944517 for service
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: config_find_next: Processing additional service options...
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: get_config_opt: Found '0' for option: ver
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: get_config_opt: Defaulting to 'pcmk' for option: clustername
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: get_config_opt: Defaulting to 'no' for option: use_logd
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: get_config_opt: Defaulting to 'no' for option: use_mgmtd
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_startup: CRM: Initialized
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] Logging: Initialized pcmk_startup
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_startup: Maximum core file size is: 18446744073709551615
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: pcmk_user_lookup: Cluster user hacluster has uid=106 gid=108
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_startup: Service: 9
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_startup: Local hostname: nodeB
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_update_nodeid: Local node id: 1384196106
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: update_member: Creating entry for node 1384196106 born on 0
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: update_member: 0x1f7fdc0 Node 1384196106 now known as nodeB (was: (null))
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: update_member: Node nodeB now has 1 quorum votes (was 0)
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: update_member: Node 1384196106/nodeB is now: member
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: pcmk_user_lookup: Cluster user hacluster has uid=106 gid=108
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: spawn_child: Forked child 6882 for process cib
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: spawn_child: Forked child 6883 for process stonith-ng
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: spawn_child: Forked child 6884 for process lrmd
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: pcmk_user_lookup: Cluster user hacluster has uid=106 gid=108
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: spawn_child: Forked child 6885 for process attrd
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: pcmk_user_lookup: Cluster user hacluster has uid=106 gid=108
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: spawn_child: Forked child 6886 for process pengine
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: pcmk_user_lookup: Cluster user hacluster has uid=106 gid=108
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: pcmk_user_lookup: Cluster user hacluster has uid=106 gid=108
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: spawn_child: Forked child 6887 for process crmd
Aug 23 11:58:15 nodeB corosync[6874]:   [SERV  ] Service engine loaded: Pacemaker Cluster Manager 1.1.7
Aug 23 11:58:15 nodeB corosync[6874]:   [SERV  ] Service engine loaded: corosync extended virtual synchrony service
Aug 23 11:58:15 nodeB corosync[6874]:   [SERV  ] Service engine loaded: corosync configuration service
Aug 23 11:58:15 nodeB corosync[6874]:   [SERV  ] Service engine loaded: corosync cluster closed process group service v1.01
Aug 23 11:58:15 nodeB corosync[6874]:   [SERV  ] Service engine loaded: corosync cluster config database access v1.01
Aug 23 11:58:15 nodeB corosync[6874]:   [SERV  ] Service engine loaded: corosync profile loading service
Aug 23 11:58:15 nodeB corosync[6874]:   [SERV  ] Service engine loaded: corosync cluster quorum service v0.1
Aug 23 11:58:15 nodeB corosync[6874]:   [MAIN  ] Compatibility mode set to whitetank.  Using V1 and V2 of the synchronization engine.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] entering GATHER state from 15.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Creating commit token because I am the rep.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Saving state aru 0 high seq received 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Storing new sequence id for ring a8
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] entering COMMIT state.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] got commit token
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] entering RECOVERY state.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] position [0] member 10.40.129.210:
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] previous ring seq a4 rep 10.40.129.210
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] aru 0 high delivered 0 received flag 1
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Did not need to originate any messages in recovery.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] got commit token
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Sending initial ORF token
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] token retrans flag is 0 my set retrans flag0 retrans queue empty 1 count 0, aru 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] install seq 0 aru 0 high seq received 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] token retrans flag is 0 my set retrans flag0 retrans queue empty 1 count 1, aru 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] install seq 0 aru 0 high seq received 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] token retrans flag is 0 my set retrans flag0 retrans queue empty 1 count 2, aru 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] install seq 0 aru 0 high seq received 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] token retrans flag is 0 my set retrans flag0 retrans queue empty 1 count 3, aru 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] install seq 0 aru 0 high seq received 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] retrans flag count 4 token aru 0 install seq 0 aru 0 0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Resetting old ring state
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] recovery to regular 1-0
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering to app 1 to 0
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] notice: pcmk_peer_update: Transitional membership event on ring 168: memb=0, new=0, lost=0
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] notice: pcmk_peer_update: Stable membership event on ring 168: memb=1, new=1, lost=0
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_peer_update: NEW:  nodeB 1384196106
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: pcmk_peer_update: Node 1384196106 has address r(0) ip(10.40.129.210)
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_peer_update: MEMB: nodeB 1384196106
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: send_cluster_id: Leaving born-on unset: 168
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: send_cluster_id: Local update: id=1384196106, born=0, seq=168
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: update_member: Node nodeB now has process list: 00000000000000000000000000111312 (1118994)
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] This node is within the primary component and will provide service.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] entering OPERATIONAL state.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] A processor joined or left the membership and a new membership was formed.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering 0 to 1
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq 1 to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: pcmk_cluster_id_callback: Node update: nodeB (1.1.7)
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] confchg entries 1
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier Start Received From 1384196106
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier completion status for nodeid 1384196106 = 1.
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization barrier completed
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization actions starting for (dummy CLM service)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering 1 to 2
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq 2 to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including 1
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering 2 to 3
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq 3 to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] confchg entries 1
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier Start Received From 1384196106
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier completion status for nodeid 1384196106 = 1.
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization barrier completed
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Committing synchronization for (dummy CLM service)
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization actions starting for (dummy AMF service)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including 2
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including 3
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering 3 to 4
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq 4 to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] confchg entries 1
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier Start Received From 1384196106
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier completion status for nodeid 1384196106 = 1.
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization barrier completed
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Committing synchronization for (dummy AMF service)
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization actions starting for (dummy CKPT service)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including 4
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering 4 to 5
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq 5 to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] confchg entries 1
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier Start Received From 1384196106
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier completion status for nodeid 1384196106 = 1.
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization barrier completed
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Committing synchronization for (dummy CKPT service)
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization actions starting for (dummy EVT service)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering 5 to 6
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq 6 to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including 5
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering 6 to 7
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq 7 to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including 6
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including 7
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering 7 to 8
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq 8 to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] confchg entries 1
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier Start Received From 1384196106
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier completion status for nodeid 1384196106 = 1.
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization barrier completed
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Committing synchronization for (dummy EVT service)
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization actions starting for (corosync cluster closed process group service v1.01)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering 8 to a
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq 9 to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq a to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [CPG   ] comparing: sender r(0) ip(10.40.129.210) ; members(old:0 left:0)
Aug 23 11:58:15 nodeB corosync[6874]:   [CPG   ] chosen downlist: sender r(0) ip(10.40.129.210) ; members(old:0 left:0)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including 8
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering a to b
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq b to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] confchg entries 1
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier Start Received From 1384196106
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Barrier completion status for nodeid 1384196106 = 1.
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Synchronization barrier completed
Aug 23 11:58:15 nodeB corosync[6874]:   [SYNC  ] Committing synchronization for (corosync cluster closed process group service v1.01)
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] mcasted message added to pending queue
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including a
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering b to c
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] Delivering MCAST message with seq c to pending delivery queue
Aug 23 11:58:15 nodeB corosync[6874]:   [MAIN  ] Completed service synchronization, ready to provide service.
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including b
Aug 23 11:58:15 nodeB corosync[6874]:   [TOTEM ] releasing messages up to and including c
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: init_ais_connection_classic: Adding fd=5 to mainloop
Aug 23 11:58:15 nodeB stonith-ng: [6883]: info: init_ais_connection_classic: AIS connection established
Aug 23 11:58:15 nodeB attrd: [6885]: debug: init_ais_connection_classic: Adding fd=5 to mainloop
Aug 23 11:58:15 nodeB attrd: [6885]: info: init_ais_connection_classic: AIS connection established
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_ipc: Recorded connection 0x1f90170 for stonith-ng/6883
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: process_ais_message: Msg[0] (dest=local:ais, from=nodeB:stonith-ng.6883, remote=true, size=5): 6883
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_ipc: Recorded connection 0x1f944d0 for attrd/6885
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: process_ais_message: Msg[0] (dest=local:ais, from=nodeB:attrd.6885, remote=true, size=5): 6885
Aug 23 11:58:15 nodeB stonith-ng: [6883]: info: get_ais_nodeid: Server details: id=1384196106 uname=nodeB cname=pcmk
Aug 23 11:58:15 nodeB stonith-ng: [6883]: info: init_ais_connection_once: Connection to 'classic openais (with plugin)': established
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: crm_new_peer: Creating entry for node nodeB/1384196106
Aug 23 11:58:15 nodeB stonith-ng: [6883]: info: crm_new_peer: Node nodeB now has id: 1384196106
Aug 23 11:58:15 nodeB stonith-ng: [6883]: info: crm_new_peer: Node 1384196106 is now known as nodeB
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: init_client_ipc_comms_nodispatch: Attempting to talk on: /var/run/crm/cib_rw
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: init_client_ipc_comms_nodispatch: Could not init comms on: /var/run/crm/cib_rw
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: cib_native_signon_raw: Connection to command channel failed
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: init_client_ipc_comms_nodispatch: Attempting to talk on: /var/run/crm/cib_callback
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: init_client_ipc_comms_nodispatch: Could not init comms on: /var/run/crm/cib_callback
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: cib_native_signon_raw: Connection to callback channel failed
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: cib_native_signon_raw: Connection to CIB failed: connection failed
Aug 23 11:58:15 nodeB stonith-ng: [6883]: debug: cib_native_signoff: Signing out of the CIB Service
Aug 23 11:58:15 nodeB attrd: [6885]: info: get_ais_nodeid: Server details: id=1384196106 uname=nodeB cname=pcmk
Aug 23 11:58:15 nodeB attrd: [6885]: info: init_ais_connection_once: Connection to 'classic openais (with plugin)': established
Aug 23 11:58:15 nodeB attrd: [6885]: debug: crm_new_peer: Creating entry for node nodeB/1384196106
Aug 23 11:58:15 nodeB attrd: [6885]: info: crm_new_peer: Node nodeB now has id: 1384196106
Aug 23 11:58:15 nodeB attrd: [6885]: info: crm_new_peer: Node 1384196106 is now known as nodeB
Aug 23 11:58:15 nodeB attrd: [6885]: info: main: Cluster connection active
Aug 23 11:58:15 nodeB crmd: [6887]: info: Invoked: /usr/lib/pacemaker/crmd
Aug 23 11:58:15 nodeB attrd: [6885]: info: main: Accepting attribute updates
Aug 23 11:58:15 nodeB attrd: [6885]: notice: main: Starting mainloop...
Aug 23 11:58:15 nodeB cib: [6882]: debug: init_ais_connection_classic: Adding fd=5 to mainloop
Aug 23 11:58:15 nodeB cib: [6882]: info: init_ais_connection_classic: AIS connection established
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_ipc: Recorded connection 0x1f8be10 for cib/6882
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] info: pcmk_ipc: Sending membership update 168 to cib
Aug 23 11:58:15 nodeB corosync[6874]:   [pcmk  ] debug: process_ais_message: Msg[0] (dest=local:ais, from=nodeB:cib.6882, remote=true, size=5): 6882
Aug 23 11:58:15 nodeB cib: [6882]: info: get_ais_nodeid: Server details: id=1384196106 uname=nodeB cname=pcmk
Aug 23 11:58:15 nodeB cib: [6882]: info: init_ais_connection_once: Connection to 'classic openais (with plugin)': established
Aug 23 11:58:15 nodeB cib: [6882]: debug: crm_new_peer: Creating entry for node nodeB/1384196106
Aug 23 11:58:15 nodeB cib: [6882]: info: crm_new_peer: Node nodeB now has id: 1384196106
Aug 23 11:58:15 nodeB cib: [6882]: info: crm_new_peer: Node 1384196106 is now known as nodeB
Aug 23 11:58:15 nodeB crmd: [6887]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/hacluster
Aug 23 11:58:15 nodeB cib: [6882]: info: cib_init: Starting cib mainloop
Aug 23 11:58:15 nodeB crmd: [6887]: notice: main: CRM Git Version: ee0730e13d124c3d58f00016c3376a1de5323cff
Aug 23 11:58:15 nodeB cib: [6882]: info: ais_dispatch_message: Membership 168: quorum still lost
Aug 23 11:58:15 nodeB cib: [6882]: info: crm_update_peer: Node nodeB: id=1384196106 state=member (new) addr=r(0) ip(10.40.129.210)  (new) votes=1 (new) born=0 seen=168 proc=00000000000000000000000000111312 (new)
Aug 23 11:58:15 nodeB cib: [6882]: debug: Forking temp process write_cib_contents
Aug 23 11:58:15 nodeB crmd: [6887]: debug: crmd_init: Starting crmd
Aug 23 11:58:15 nodeB crmd: [6887]: debug: s_crmd_fsa: Processing I_STARTUP: [ state=S_STARTING cause=C_STARTUP origin=crmd_init ]
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_LOG
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_log: FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_STARTUP
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_startup: Registering Signal Handlers
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_startup: Creating CIB and LRM objects
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_CIB_START
Aug 23 11:58:15 nodeB crmd: [6887]: debug: init_client_ipc_comms_nodispatch: Attempting to talk on: /var/run/crm/cib_rw
Aug 23 11:58:15 nodeB crmd: [6887]: debug: init_client_ipc_comms_nodispatch: Attempting to talk on: /var/run/crm/cib_callback
Aug 23 11:58:15 nodeB crmd: [6887]: debug: cib_native_signon_raw: Connection to CIB successful
Aug 23 11:58:15 nodeB crmd: [6887]: info: do_cib_control: CIB connection established
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_HA_CONNECT
Aug 23 11:58:15 nodeB crmd: [6887]: info: get_cluster_type: Cluster type is: 'openais'
Aug 23 11:58:15 nodeB crmd: [6887]: notice: crm_cluster_connect: Connecting to cluster infrastructure: classic openais (with plugin)
Aug 23 11:58:15 nodeB crmd: [6887]: info: init_ais_connection_classic: Creating connection to our Corosync plugin
Aug 23 11:58:15 nodeB cib: [6882]: debug: cib_common_callback_worker: Setting cib_refresh_notify callbacks for 6887 (0278a741-301f-4b7f-b80f-a8ca30aedd4a): on
Aug 23 11:58:15 nodeB cib: [6882]: debug: cib_common_callback_worker: Setting cib_diff_notify callbacks for 6887 (0278a741-301f-4b7f-b80f-a8ca30aedd4a): on
Aug 23 11:58:15 nodeB crmd: [6887]: debug: init_ais_connection_classic: Adding fd=7 to mainloop
Aug 23 11:58:15 nodeB crmd: [6887]: info: init_ais_connection_classic: AIS connection established
Aug 23 11:58:15 nodeB rsyslogd-2177: imuxsock begins to drop messages from pid 6874 due to rate-limiting
Aug 23 11:58:15 nodeB crmd: [6887]: info: get_ais_nodeid: Server details: id=1384196106 uname=nodeB cname=pcmk
Aug 23 11:58:15 nodeB crmd: [6887]: info: init_ais_connection_once: Connection to 'classic openais (with plugin)': established
Aug 23 11:58:15 nodeB crmd: [6887]: debug: crm_new_peer: Creating entry for node nodeB/1384196106
Aug 23 11:58:15 nodeB crmd: [6887]: info: crm_new_peer: Node nodeB now has id: 1384196106
Aug 23 11:58:15 nodeB crmd: [6887]: info: crm_new_peer: Node 1384196106 is now known as nodeB
Aug 23 11:58:15 nodeB crmd: [6887]: info: ais_status_callback: status: nodeB is now unknown
Aug 23 11:58:15 nodeB crmd: [6887]: info: do_ha_control: Connected to the cluster
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_READCONFIG
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_LRM_CONNECT
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_lrm_control: Connecting to the LRM
Aug 23 11:58:15 nodeB lrmd: [6884]: debug: on_msg_register:client crmd [6887] registered
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_lrm_control: LRM connection established
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_CCM_CONNECT
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_STARTED
Aug 23 11:58:15 nodeB crmd: [6887]: info: do_started: Delaying start, no membership data (0000000000100000)
Aug 23 11:58:15 nodeB crmd: [6887]: debug: register_fsa_input_adv: Stalling the FSA pending further input: cause=C_FSA_INTERNAL
Aug 23 11:58:15 nodeB crmd: [6887]: debug: s_crmd_fsa: Exiting the FSA: queue=0, fsa_actions=0x2, stalled=true
Aug 23 11:58:15 nodeB crmd: [6887]: debug: config_query_callback: Call 3 : Parsing CIB options
Aug 23 11:58:15 nodeB crmd: [6887]: debug: config_query_callback: Shutdown escalation occurs after: 1200000ms
Aug 23 11:58:15 nodeB crmd: [6887]: debug: config_query_callback: Checking for expired actions every 900000ms
Aug 23 11:58:15 nodeB crmd: [6887]: debug: config_query_callback: Sending expected-votes=2 to corosync
Aug 23 11:58:15 nodeB crmd: [6887]: info: ais_dispatch_message: Membership 168: quorum still lost
Aug 23 11:58:15 nodeB crmd: [6887]: notice: crmd_peer_update: Status update: Client nodeB/crmd now has status [online] (DC=<null>)
Aug 23 11:58:15 nodeB crmd: [6887]: info: ais_status_callback: status: nodeB is now member (was unknown)
Aug 23 11:58:15 nodeB crmd: [6887]: info: crm_update_peer: Node nodeB: id=1384196106 state=member (new) addr=r(0) ip(10.40.129.210)  (new) votes=1 (new) born=0 seen=168 proc=00000000000000000000000000111312 (new)
Aug 23 11:58:15 nodeB crmd: [6887]: debug: post_cache_update: Updated cache after membership event 168.
Aug 23 11:58:15 nodeB crmd: [6887]: debug: post_cache_update: post_cache_update added action A_ELECTION_CHECK to the FSA
Aug 23 11:58:15 nodeB crmd: [6887]: info: ais_dispatch_message: Membership 168: quorum still lost
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_STARTED
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_started: Init server comms
Aug 23 11:58:15 nodeB crmd: [6887]: notice: do_started: The local CRM is operational
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_ELECTION_CHECK
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_election_check: Ignore election check: we not in an election
Aug 23 11:58:15 nodeB crmd: [6887]: debug: s_crmd_fsa: Processing I_PENDING: [ state=S_STARTING cause=C_FSA_INTERNAL origin=do_started ]
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_LOG
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_log: FSA: Input I_PENDING from do_started() received in state S_STARTING
Aug 23 11:58:15 nodeB crmd: [6887]: notice: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_INTEGRATE_TIMER_STOP
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_FINALIZE_TIMER_STOP
Aug 23 11:58:15 nodeB crmd: [6887]: debug: do_fsa_action: actions:trace: #011// A_CL_JOIN_QUERY
Aug 23 11:58:15 nodeB cib: [6882]: info: Managed write_cib_contents process 6891 exited with return code 0.
Aug 23 11:58:16 nodeB stonith-ng: [6883]: debug: init_client_ipc_comms_nodispatch: Attempting to talk on: /var/run/crm/cib_rw






_______________________________________________
discuss mailing list
discuss@xxxxxxxxxxxx
http://lists.corosync.org/mailman/listinfo/discuss

[Index of Archives]     [Linux Clusters]     [Corosync Project]     [Linux USB Devel]     [Linux Audio Users]     [Photo]     [Yosemite News]    [Yosemite Photos]    [Linux Kernel]     [Linux SCSI]     [X.Org]

  Powered by Linux