Ah, awesome. Angus is back from vacation now, so he should see this soon. On Mon, Oct 15, 2012 at 7:15 AM, Grüninger, Andreas (LGL Extern) <Andreas.Grueninger@xxxxxxxxxx> wrote: > Problem solved. > It is an error in libqb. > In Solaris MSG_NOSIGNAL is not defined. In libqb SIGPIPE must be ignored or handled. > A pull request is waiting: https://github.com/asalkeld/libqb/pulls. > > Andreas > > -----Ursprüngliche Nachricht----- > Von: discuss-bounces@xxxxxxxxxxxx [mailto:discuss-bounces@xxxxxxxxxxxx] Im Auftrag von Grüninger, Andreas (LGL Extern) > Gesendet: Freitag, 12. Oktober 2012 16:29 > An: discuss@xxxxxxxxxxxx > Betreff: Re: shutdown of corosync-notifyd results in shutdown ofpacemaker > > I compiled the the current master of > - libqb > - pacemaker > - corosync > in Solaris 11U7 (gcc 4.5.2) and openSuse 12.2 (gcc 4.7.1). > > For the test a configuration with one node and no resources is used. > > The start of corosync-notifyd is handled nearly in the same way in linux and solaris. > See below the first two listings. The name of the linux host is "linux-t7bi" and the name of the solaris host is "zd-sol-s1". > In linux the MAIN module logs something. In solaris the same log entries are written from the MON module. > > When corosync-notifyd is killed linux handels this gracefully and there are no log entries of pacemaker. > In solaris the event is handled from pacemaker and not from corosync. There are no log entries for corosync. > Pacemaker shuts down and corosync is still running and healthy. If pacemaker is restarted it connects again to corosync. > > How can this be when the same source code is compiled? > > Andreas > > Log from linux after start of corosync-notifyd .... > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QB ] ipc_us.c:handle_new_connection:666 IPC credentials authenticated (20167-20287-30) > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QB ] ipc_us.c:qb_ipcs_us_connect:978 connecting to client (20167-20287-30) > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [MAIN ] ipc_glue.c:cs_ipcs_connection_created:269 connection created > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QB ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7febfe95a260 > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QB ] ipc_us.c:handle_new_connection:666 IPC credentials authenticated (20167-20287-32) > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QB ] ipc_us.c:qb_ipcs_us_connect:978 connecting to client (20167-20287-32) > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [MAIN ] ipc_glue.c:cs_ipcs_connection_created:269 connection created > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QUORUM] vsf_quorum.c:quorum_lib_init_fn:316 lib_init_fn: conn=0x7febfe95cd70 > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_gettype:471 got quorum_type request on 0x7febfe95cd70 > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:412 got trackstart request on 0x7febfe95cd70 > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:420 sending initial status to 0x7febfe95cd70 > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QUORUM] vsf_quorum.c:send_library_notification:359 sending quorum notification to 0x7febfe95cd70, length = 52 > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QB ] ipc_us.c:handle_new_connection:666 IPC credentials authenticated (20167-20287-34) > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QB ] ipc_us.c:qb_ipcs_us_connect:978 connecting to client (20167-20287-34) > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [MAIN ] ipc_glue.c:cs_ipcs_connection_created:269 connection created > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QB ] ipcs.c:qb_ipcs_dispatch_connection_request:716 HUP conn (20167-20287-34) > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [QB ] ipcs.c:qb_ipcs_disconnect:555 qb_ipcs_disconnect(20167-20287-34) state:2 > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [MAIN ] ipc_glue.c:cs_ipcs_connection_closed:414 cs_ipcs_connection_closed() > Oct 12 14:21:03 [20166] linux-t7bi corosync debug [MAIN ] ipc_glue.c:cs_ipcs_connection_destroyed:387 cs_ipcs_connection_destroyed() > .... > > Log from solaris after start of corosync-notifyd .... > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [QB ] ipc_us.c:handle_new_connection:666 IPC credentials authenticated (20153-20181-34) > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [QB ] ipc_us.c:qb_ipcs_us_connect:978 connecting to client (20153-20181-34) > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] ipc_glue.c:cs_ipcs_connection_created:269 connection created > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] cmap.c:cmap_lib_init_fn:181 lib_init_fn: conn=84433e8 > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [QB ] ipc_us.c:handle_new_connection:666 IPC credentials authenticated (20153-20181-36) > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [QB ] ipc_us.c:qb_ipcs_us_connect:978 connecting to client (20153-20181-36) > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] ipc_glue.c:cs_ipcs_connection_created:269 connection created > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] vsf_quorum.c:quorum_lib_init_fn:316 lib_init_fn: conn=840fcb8 > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] vsf_quorum.c:message_handler_req_lib_quorum_gettype:471 got quorum_type request on 840fcb8 > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:412 got trackstart request on 840fcb8 > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:420 sending initial status to 840fcb8 > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] vsf_quorum.c:send_library_notification:359 sending quorum notification to 840fcb8, length = 56 > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [QB ] ipc_us.c:handle_new_connection:666 IPC credentials authenticated (20153-20181-38) > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [QB ] ipc_us.c:qb_ipcs_us_connect:978 connecting to client (20153-20181-38) > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] ipc_glue.c:cs_ipcs_connection_created:269 connection created > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [QB ] ipc_us.c:qb_ipc_us_recv_at_most:326 recv(fd 38) got 0 bytes assuming ENOTCONN > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [QB ] ipcs.c:qb_ipcs_disconnect:555 qb_ipcs_disconnect(20153-20181-38) state:2 > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] ipc_glue.c:cs_ipcs_connection_closed:414 cs_ipcs_connection_closed() > Oct 12 14:20:35 [20152] zd-sol-s1 corosync debug [MON ] ipc_glue.c:cs_ipcs_connection_destroyed:387 cs_ipcs_connection_destroyed() > .... > > > > Log from linux after stop of corosync-notifyd .... > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [QB ] ipcs.c:qb_ipcs_dispatch_connection_request:716 HUP conn (20167-20287-32) > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [QB ] ipcs.c:qb_ipcs_disconnect:555 qb_ipcs_disconnect(20167-20287-32) state:2 > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [MAIN ] ipc_glue.c:cs_ipcs_connection_closed:414 cs_ipcs_connection_closed() > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [QUORUM] vsf_quorum.c:quorum_lib_exit_fn:328 lib_exit_fn: conn=0x7febfe95cd70 > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 340! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 344! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 340! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 344! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 344! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 352! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 344! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 340! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 344! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 344! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 344! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 344! > Oct 12 14:22:59 [20166] linux-t7bi corosync error [MAIN ] ipc_glue.c:msg_send_or_queue:526 event_send retuned -32, expected 340! > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [MAIN ] ipc_glue.c:cs_ipcs_connection_destroyed:387 cs_ipcs_connection_destroyed() > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [QB ] ipcs.c:qb_ipcs_dispatch_connection_request:716 HUP conn (20167-20287-30) > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [QB ] ipcs.c:qb_ipcs_disconnect:555 qb_ipcs_disconnect(20167-20287-30) state:2 > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [MAIN ] ipc_glue.c:cs_ipcs_connection_closed:414 cs_ipcs_connection_closed() > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [QB ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7febfe95a260 > Oct 12 14:22:59 [20166] linux-t7bi corosync debug [MAIN ] ipc_glue.c:cs_ipcs_connection_destroyed:387 cs_ipcs_connection_destroyed() > .... > > Log from solaris after stop of corosync-notifyd .... > Oct 12 14:24:05 [20157] pacemakerd: debug: qb_ipc_us_recv_at_most: recv(fd 7) got 0 bytes assuming ENOTCONN > Oct 12 14:24:05 [20157] pacemakerd: debug: _check_connection_state: interpreting result -134 as a disconnect: Transport endpoint is not connected (134) > Oct 12 14:24:05 [20157] pacemakerd: error: cfg_connection_destroy: Connection destroyed > Oct 12 14:24:05 [20157] pacemakerd: notice: pcmk_shutdown_worker: Shuting down Pacemaker > Oct 12 14:24:05 [20157] pacemakerd: notice: stop_child: Stopping crmd: Sent -15 to process 20163 > Oct 12 14:24:05 [20157] pacemakerd: debug: qb_ipc_us_recv_at_most: recv(fd 9) got 0 bytes assuming ENOTCONN > Oct 12 14:24:05 [20157] pacemakerd: debug: _check_connection_state: interpreting result -134 as a disconnect: Transport endpoint is not connected (134) > Oct 12 14:24:05 [20157] pacemakerd: error: cpg_connection_destroy: Connection destroyed > Oct 12 14:24:05 [20159] stonith-ng: debug: qb_ipc_us_recv_at_most: recv(fd 6) got 0 bytes assuming ENOTCONN > Oct 12 14:24:05 [20159] stonith-ng: debug: _check_connection_state: interpreting result -134 as a disconnect: Transport endpoint is not connected (134) > Oct 12 14:24:05 [20159] stonith-ng: error: pcmk_cpg_dispatch: Connection to the CPG API failed: 2 > Oct 12 14:24:05 [20163] crmd: info: crm_signal_dispatch: Invoking handler for signal 15: Terminated > Oct 12 14:24:05 [20159] stonith-ng: error: stonith_peer_ais_destroy: AIS connection terminated > Oct 12 14:24:05 [20163] crmd: notice: crm_shutdown: Requesting shutdown, upper limit is 1200000ms > Oct 12 14:24:05 [20159] stonith-ng: info: stonith_shutdown: Terminating with 1 clients > Oct 12 14:24:05 [20163] crmd: debug: crm_timer_start: Started Shutdown Escalation (I_STOP:1200000ms), src=21 > Oct 12 14:24:05 [20159] stonith-ng: debug: cib_native_signoff: Signing out of the CIB Service > Oct 12 14:24:05 [20161] attrd: debug: qb_ipc_us_recv_at_most: recv(fd 6) got 0 bytes assuming ENOTCONN > Oct 12 14:24:05 [20163] crmd: debug: s_crmd_fsa: Processing I_SHUTDOWN: [ state=S_NOT_DC cause=C_SHUTDOWN origin=crm_shutdown ] > Oct 12 14:24:05 [20159] stonith-ng: debug: qb_ipcc_disconnect: qb_ipcc_disconnect() > > > -----Ursprüngliche Nachricht----- > Von: Andrew Beekhof [mailto:andrew@xxxxxxxxxxx] > Gesendet: Freitag, 12. Oktober 2012 01:58 > An: Grüninger, Andreas (LGL Extern) > Cc: discuss@xxxxxxxxxxxx > Betreff: Re: shutdown of corosync-notifyd results in shutdown of pacemaker > > More specifically, stopping corosync-notifyd results in all Pacemaker's connections to Corosync being terminated. > Andreas: Did you test this on linux or solaris only? > > On Thu, Oct 11, 2012 at 11:45 PM, Grüninger, Andreas (LGL Extern) <Andreas.Grueninger@xxxxxxxxxx> wrote: >> When I start >> corosync-notifyd -f -l -s -m <MONITORINGSERVER> and close it with >> CTRL-C, pacemaker make a shutdown. >> Please see below for the details. >> >> I compiled the current master of corosync (tag 2.1.0) and the current master of pacemaker. >> The OS is Solaris 11U7. >> >> Is this a feature or a bug? >> In Solaris libqb must be patched to avoid errors. >> Please see >> https://lists.fedorahosted.org/pipermail/quarterback-devel/2012-September/000921.html "[PATCH] -ENOTCONN handled as error when client disconnects" >> Maybe this patch should not deliver -ESHUTDOWN when a client disconnects. >> IMHO this is the adaequate result. >> >> Andreas >> >> >> On Thu, Oct 4, 2012 at 5:57 PM, Grüninger, Andreas (LGL Extern) <Andreas.Grueninger@xxxxxxxxxx> wrote: >>>>> Is this an error or the desired result? >>> >>>>Based on the logs, pacemaker thinks corosync died. Did that happen? >>>>If so there is not much pacemaker can do :-( >>> >>> And that is absolutely ok when corosync dies. >>> Corosync does not die but is still healthy. >>> It is corosync-notifyd which is started additionally to corosync as a separate process and which is finished with kill as daemon or with ctrl-c as foreground process. >>> The job of corosync-notifyd is sending of SNMP traps. >>> This is the functionality of crm_mon -C .. -S ... for pacemaker. >>> >>> So corosync-notifyd sends the wrong signal or pacemaker does a little bit too much. >>> Pacemaker should just ignore this ending connection. >> >> All the Pacemaker daemons are being told, by Corosync itself, that their connections to Corosync are dead. >> Its a little difficult to ignore that. >> >>> Is there a chance in pacemaker or should should this better solved in corosync/corosync-notifyd? >> >> It needs to be addressed in corosync/corosync-notifyd. >> Corosync's CPG library is the one invoking our >> cpg_connection_destroy() callback. >> >>> >>> Andreas >>> >>> -----Ursprüngliche Nachricht----- >>> Von: Andrew Beekhof [mailto:andrew@xxxxxxxxxxx] >>> Gesendet: Mittwoch, 3. Oktober 2012 01:09 >>> An: The Pacemaker cluster resource manager >>> Betreff: Re: [Pacemaker] Exiting corosync-notifyd results in shutting >>> downof pacemakerd >>> >>> On Wed, Oct 3, 2012 at 2:51 AM, Grüninger, Andreas (LGL Extern) <Andreas.Grueninger@xxxxxxxxxx> wrote: >>>> I am currently investigating the monitoring of corosync/pacemaker with snmp. >>>> crm_mon used with the OCF resource ClusterMon works as it should. >>>> >>>> But corosync-notifyd can't be used in our case. >>>> I start corosync-notifyd in the foreground as follows >>>> corosync-notifyd -f -l -s -m 10.50.235.1 >>>> >>>> When I stop the running corosync-notifyd with CTRL-C, pacemaker shuts down with the following entries in the logfile. >>>> Is this an error or the desired result? >>> >>> Based on the logs, pacemaker thinks corosync died. Did that happen? >>> If so there is not much pacemaker can do :-( >>> >>>> >>>> .... >>>> Oct 02 18:42:19 [27126] pacemakerd: error: cfg_connection_destroy: Connection destroyed >>>> Oct 02 18:42:19 [27126] pacemakerd: notice: pcmk_shutdown_worker: Shuting down Pacemaker >>>> Oct 02 18:42:19 [27126] pacemakerd: notice: stop_child: Stopping crmd: Sent -15 to process 27177 >>>> Oct 02 18:42:19 [27126] pacemakerd: error: cpg_connection_destroy: Connection destroyed >>>> Oct 02 18:42:19 [27177] crmd: info: crm_signal_dispatch: Invoking handler for signal 15: Terminated >>>> Oct 02 18:42:19 [27177] crmd: notice: crm_shutdown: Requesting shutdown, upper limit is 1200000ms >>>> Oct 02 18:42:19 [27128] stonith-ng: error: pcmk_cpg_dispatch: Connection to the CPG API failed: 2 >>>> Oct 02 18:42:19 [27177] crmd: info: do_shutdown_req: Sending shutdown request to zd-sol-s1-v61 >>>> Oct 02 18:42:19 [27128] stonith-ng: error: stonith_peer_ais_destroy: AIS connection terminated >>>> Oct 02 18:42:19 [27128] stonith-ng: info: stonith_shutdown: Terminating with 1 clients >>>> Oct 02 18:42:19 [27130] attrd: error: pcmk_cpg_dispatch: Connection to the CPG API failed: 2 >>>> Oct 02 18:42:19 [27130] attrd: crit: attrd_ais_destroy: Lost connection to Corosync service! >>>> Oct 02 18:42:19 [27130] attrd: notice: main: Exiting... >>>> Oct 02 18:42:19 [27130] attrd: notice: main: Disconnecting client 81ffc38, pid=27177... >>>> Oct 02 18:42:19 [27128] stonith-ng: info: qb_ipcs_us_withdraw: withdrawing server sockets >>>> Oct 02 18:42:19 [27128] stonith-ng: info: crm_xml_cleanup: Cleaning up memory from libxml2 >>>> Oct 02 18:42:19 [27130] attrd: error: attrd_cib_connection_destroy: Connection to the CIB terminated... >>>> Oct 02 18:42:19 [27127] cib: error: pcmk_cpg_dispatch: Connection to the CPG API failed: 2 >>>> Oct 02 18:42:19 [27127] cib: error: cib_ais_destroy: Corosync connection lost! Exiting. >>>> Oct 02 18:42:19 [27129] lrmd: info: lrmd_ipc_destroy: LRMD client disconnecting 807e768 - name: crmd id: 1d659f61-d6e2-4ef3-f674-b9a8ba8029e8 >>>> Oct 02 18:42:19 [27127] cib: info: terminate_cib: cib_ais_destroy: Exiting fast... >>>> Oct 02 18:42:19 [27127] cib: info: qb_ipcs_us_withdraw: withdrawing server sockets >>>> Oct 02 18:42:19 [27127] cib: info: qb_ipcs_us_withdraw: withdrawing server sockets >>>> Oct 02 18:42:19 [27127] cib: info: qb_ipcs_us_withdraw: withdrawing server sockets >>>> Oct 02 18:42:19 [27126] pacemakerd: error: pcmk_child_exit: Child process attrd exited (pid=27130, rc=1) >>>> Oct 02 18:42:19 [27126] pacemakerd: error: send_cpg_message: Sending message via cpg FAILED: (rc=9) Bad handle >>>> Oct 02 18:42:19 [27126] pacemakerd: error: pcmk_child_exit: Child process cib exited (pid=27127, rc=64) >>>> Oct 02 18:42:19 [27126] pacemakerd: error: send_cpg_message: Sending message via cpg FAILED: (rc=9) Bad handle >>>> Oct 02 18:42:19 [27126] pacemakerd: notice: pcmk_child_exit: Child process crmd terminated with signal 13 (pid=27177, core=0) >>>> Oct 02 18:42:19 [27126] pacemakerd: error: send_cpg_message: Sending message via cpg FAILED: (rc=9) Bad handle >>>> Oct 02 18:42:19 [27126] pacemakerd: notice: stop_child: Stopping pengine: Sent -15 to process 27131 >>>> Oct 02 18:42:19 [27126] pacemakerd: info: pcmk_child_exit: Child process pengine exited (pid=27131, rc=0) >>>> Oct 02 18:42:19 [27126] pacemakerd: error: send_cpg_message: Sending message via cpg FAILED: (rc=9) Bad handle >>>> Oct 02 18:42:19 [27126] pacemakerd: notice: stop_child: Stopping lrmd: Sent -15 to process 27129 >>>> Oct 02 18:42:19 [27129] lrmd: info: crm_signal_dispatch: Invoking handler for signal 15: Terminated >>>> Oct 02 18:42:19 [27129] lrmd: info: lrmd_shutdown: Terminating with 0 clients >>>> Oct 02 18:42:19 [27129] lrmd: info: qb_ipcs_us_withdraw: withdrawing server sockets >>>> Oct 02 18:42:19 [27126] pacemakerd: info: pcmk_child_exit: Child process lrmd exited (pid=27129, rc=0) >>>> Oct 02 18:42:19 [27126] pacemakerd: error: send_cpg_message: Sending message via cpg FAILED: (rc=9) Bad handle >>>> Oct 02 18:42:19 [27126] pacemakerd: notice: stop_child: Stopping stonith-ng: Sent -15 to process 27128 >>>> Oct 02 18:42:19 [27126] pacemakerd: notice: pcmk_child_exit: Child process stonith-ng terminated with signal 11 (pid=27128, core=128) >>>> Oct 02 18:42:19 [27126] pacemakerd: error: send_cpg_message: Sending message via cpg FAILED: (rc=9) Bad handle >>>> Oct 02 18:42:19 [27126] pacemakerd: notice: pcmk_shutdown_worker: Shutdown complete >>>> Oct 02 18:42:19 [27126] pacemakerd: info: qb_ipcs_us_withdraw: withdrawing server sockets >>>> Oct 02 18:42:19 [27126] pacemakerd: info: main: Exiting pacemakerd >>>> >>>> Andreas >>>> >>>> _______________________________________________ >>>> Pacemaker mailing list: Pacemaker@xxxxxxxxxxxxxxxxxxx >>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>> >>>> Project Home: http://www.clusterlabs.org Getting started: >>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>> Bugs: http://bugs.clusterlabs.org >>> >>> _______________________________________________ >>> Pacemaker mailing list: Pacemaker@xxxxxxxxxxxxxxxxxxx >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org Getting started: >>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >>> >>> _______________________________________________ >>> Pacemaker mailing list: Pacemaker@xxxxxxxxxxxxxxxxxxx >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org Getting started: >>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@xxxxxxxxxxxxxxxxxxx >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org Getting started: >> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> >> >> _______________________________________________ >> discuss mailing list >> discuss@xxxxxxxxxxxx >> http://lists.corosync.org/mailman/listinfo/discuss > > _______________________________________________ > discuss mailing list > discuss@xxxxxxxxxxxx > http://lists.corosync.org/mailman/listinfo/discuss > _______________________________________________ > quarterback-devel mailing list > quarterback-devel@xxxxxxxxxxxxxxxxxxxxxx > https://lists.fedorahosted.org/mailman/listinfo/quarterback-devel _______________________________________________ discuss mailing list discuss@xxxxxxxxxxxx http://lists.corosync.org/mailman/listinfo/discuss