Hi list
I have a test ceph cluster include 3 nodes (node0: mon,
node1: osd and nfs server1, node2 osd and nfs server2).
os :centos6.6 ,kernel :3.10.94-1.el6.elrepo.x86_64, ceph
version 0.94.5
I followed the
http://www.sebastien-han.fr/blog/2012/07/06/nfs-over-rbd/
instructions to setup an active/standy NFS environment.
when using commands " #
service
corosync stop or # poweroff " on node1 , the fail
over switch situation went fine ( nfs server take over by
node2). But when I testing the situation of cutting off the
power of node1, the switch is failed.
1. [root@node1 ~]# crm status
Last updated: Fri Dec 18 17:14:19 2015
Last change: Fri Dec 18 17:13:29 2015
Stack: classic openais (with plugin)
Current DC: node1 - partition with quorum
Version: 1.1.11-97629de
2 Nodes configured, 3 expected votes
8 Resources configured
Online: [ node1 node2 ]
Resource Group: g_rbd_share_1
p_rbd_map_1 (ocf::ceph:rbd.in): Started
node1
p_fs_rbd_1 (ocf::heartbeat:Filesystem): Started
node1
p_export_rbd_1 (ocf::heartbeat:exportfs):
Started node1
p_vip_1 (ocf::heartbeat:IPaddr): Started
node1
Clone Set: clo_nfs [g_nfs]
Started: [ node1 node2 ]
2. [root@node1 ~]# service corosync stop
[root@node2 cluster]# crm status
Last updated: Fri Dec 18 17:14:59 2015
Last change: Fri Dec 18 17:13:29 2015
Stack: classic openais (with plugin)
Current DC: node2 - partition WITHOUT quorum
Version: 1.1.11-97629de
2 Nodes configured, 3 expected votes
8 Resources configured
Online: [ node2 ]
OFFLINE: [ node1 ]
Resource Group: g_rbd_share_1
p_rbd_map_1 (ocf::ceph:rbd.in): Started
node2
p_fs_rbd_1 (ocf::heartbeat:Filesystem): Started
node2
p_export_rbd_1 (ocf::heartbeat:exportfs):
Started node2
p_vip_1 (ocf::heartbeat:IPaddr): Started
node2
Clone Set: clo_nfs [g_nfs]
Started: [ node2 ]
Stopped: [ node1 ]
3. cut off node1 power manually
[root@node2 cluster]# crm status
Last updated: Fri Dec 18 17:23:06 2015
Last change: Fri Dec 18 17:13:29 2015
Stack: classic openais (with plugin)
Current DC: node2 - partition WITHOUT quorum
Version: 1.1.11-97629de
2 Nodes configured, 3 expected votes
8 Resources configured
Online: [ node2 ]
OFFLINE: [ node1 ]
Clone Set: clo_nfs [g_nfs]
Started: [ node2 ]
Stopped: [ node1 ]
Failed actions:
p_rbd_map_1_start_0 on node2 'unknown error' (1):
call=48, status=Timed Out, last-rc-change='Fri Dec 18
17:22:19 2015', queued=0ms, exec=20002ms
corosync.log:
Dec 18 17:22:19 corosync [pcmk ] notice: pcmk_peer_update:
Transitional membership event on ring 668: memb=1, new=0,
lost=1
Dec 18 17:22:19 corosync [pcmk ] info: pcmk_peer_update:
memb: node2 1211279552
Dec 18 17:22:19 corosync [pcmk ] info: pcmk_peer_update:
lost: node1 1194502336
Dec 18 17:22:19 corosync [pcmk ] notice: pcmk_peer_update:
Stable membership event on ring 668: memb=1, new=0, lost=0
Dec 18 17:22:19 corosync [pcmk ] info: pcmk_peer_update:
MEMB: node2 1211279552
Dec 18 17:22:19 corosync [pcmk ] info:
ais_mark_unseen_peer_dead: Node node1 was not seen in the
previous transition
Dec 18 17:22:19 corosync [pcmk ] info: update_member: Node
1194502336/node1 is now: lost
Dec 18 17:22:19 [2695] node2 crmd: notice:
plugin_handle_membership: Membership 668: quorum lost
Dec 18 17:22:19 [2690] node2 cib: notice:
plugin_handle_membership: Membership 668: quorum lost
Dec 18 17:22:19 corosync [pcmk ] info:
send_member_notification: Sending membership update 668 to 3
children
Dec 18 17:22:19 [2691] node2 stonith-ng: notice:
plugin_handle_membership: Membership 668: quorum lost
Dec 18 17:22:19 [2695] node2 crmd: info:
crm_update_peer_proc: plugin_handle_membership: Node
node1[1194502336] - unknown is now lost
Dec 18 17:22:19 [2690] node2 cib: info:
crm_update_peer_proc: plugin_handle_membership: Node
node1[1194502336] - unknown is now lost
Dec 18 17:22:19 corosync [TOTEM ] A processor joined or left
the membership and a new membership was formed.
Dec 18 17:22:19 [2695] node2 crmd: info:
peer_update_callback: Client node1/peer now has status
[offline] (DC=true, changed=111312)
Dec 18 17:22:19 [2691] node2 stonith-ng: info:
crm_update_peer_proc: plugin_handle_membership: Node
node1[1194502336] - unknown is now lost
Dec 18 17:22:19 [2695] node2 crmd: info:
peer_update_callback: Peer node1 left us
Dec 18 17:22:19 [2690] node2 cib: notice:
crm_update_peer_state: plugin_handle_membership: Node
node1[1194502336] - state is now lost (was member)
Dec 18 17:22:19 [2695] node2 crmd: warning:
match_down_event: No match for shutdown action on node1
Dec 18 17:22:19 [2691] node2 stonith-ng: notice:
crm_update_peer_state: plugin_handle_membership: Node
node1[1194502336] - state is now lost (was member)
Dec 18 17:22:19 [2695] node2 crmd: notice:
peer_update_callback: Stonith/shutdown of node1 not
matched
Dec 18 17:22:19 [2695] node2 crmd: info:
crm_update_peer_join: peer_update_callback: Node
node1[1194502336] - join-2 phase 4 -> 0
Dec 18 17:22:19 [2695] node2 crmd: info:
abort_transition_graph: Transition aborted: Node failure
(source=peer_update_callback:237, 1)
Dec 18 17:22:19 [2695] node2 crmd: notice:
crm_update_peer_state: plugin_handle_membership: Node
node1[1194502336] - state is now lost (was member)
Dec 18 17:22:19 [2695] node2 crmd: info:
peer_update_callback: node1 is now lost (was member)
Dec 18 17:22:19 [2695] node2 crmd: warning:
match_down_event: No match for shutdown action on node1
Dec 18 17:22:19 corosync [CPG ] chosen downlist: sender
r(0) ip(192.168.50.72) ; members(old:2 left:1)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section status to master (origin=local/crmd/88)
Dec 18 17:22:19 [2695] node2 crmd: notice:
peer_update_callback: Stonith/shutdown of node1 not
matched
Dec 18 17:22:19 [2695] node2 crmd: info:
abort_transition_graph: Transition aborted: Node failure
(source=peer_update_callback:237, 1)
Dec 18 17:22:19 corosync [MAIN ] Completed service
synchronization, ready to provide service.
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section status to master (origin=local/crmd/89)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: Diff: --- 0.69.157 2
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: Diff: +++ 0.69.158 (null)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: + /cib: @num_updates=158
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: + /cib/status/node_state[@id='node1']:
@crmd=offline, @crm-debug-origin=peer_update_callback
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section status: OK (rc=0, origin=node2/crmd/88,
version=0.69.158)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section status: OK (rc=0, origin=node2/crmd/89,
version=0.69.158)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section nodes to master (origin=local/crmd/90)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section status to master (origin=local/crmd/91)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section cib to master (origin=local/crmd/92)
Dec 18 17:22:19 [2695] node2 crmd: info:
crmd_cs_dispatch: Setting expected votes to 3
Dec 18 17:22:19 [2695] node2 crmd: notice:
do_state_transition: State transition S_IDLE ->
S_INTEGRATION [ input=I_NODE_JOIN cause=C_FSA_INTERNAL
origin=check_join_state ]
Dec 18 17:22:19 [2695] node2 crmd: info:
do_dc_join_offer_one: An unknown node joined -
(re-)offer to any unconfirmed nodes
Dec 18 17:22:19 [2695] node2 crmd: info:
join_make_offer: Making join offers based on membership
668
Dec 18 17:22:19 [2695] node2 crmd: info:
join_make_offer: Skipping node2: already known 4
Dec 18 17:22:19 [2695] node2 crmd: info:
join_make_offer: Not making an offer to node1: not
active (lost)
Dec 18 17:22:19 [2695] node2 crmd: info:
abort_transition_graph: Transition aborted: Peer Halt
(source=do_te_invoke:158, 1)
Dec 18 17:22:19 [2695] node2 crmd: info:
do_state_transition: State transition S_INTEGRATION
-> S_FINALIZE_JOIN [ input=I_INTEGRATED
cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 17:22:19 [2695] node2 crmd: info:
crmd_join_phase_log: join-2: node2=confirmed
Dec 18 17:22:19 [2695] node2 crmd: info:
crmd_join_phase_log: join-2: node1=none
Dec 18 17:22:19 [2695] node2 crmd: info:
do_state_transition: State transition S_FINALIZE_JOIN
-> S_INTEGRATION [ input=I_NODE_JOIN cause=C_FSA_INTERNAL
origin=peer_update_callback ]
Dec 18 17:22:19 [2695] node2 crmd: info:
do_dc_join_offer_one: An unknown node joined -
(re-)offer to any unconfirmed nodes
Dec 18 17:22:19 [2695] node2 crmd: info:
join_make_offer: Skipping node2: already known 4
Dec 18 17:22:19 [2695] node2 crmd: info:
join_make_offer: Not making an offer to node1: not
active (lost)
Dec 18 17:22:19 [2695] node2 crmd: info:
do_state_transition: State transition S_INTEGRATION
-> S_FINALIZE_JOIN [ input=I_INTEGRATED
cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 17:22:19 [2695] node2 crmd: info:
crmd_join_phase_log: join-2: node2=confirmed
Dec 18 17:22:19 [2695] node2 crmd: info:
crmd_join_phase_log: join-2: node1=none
Dec 18 17:22:19 [2695] node2 crmd: info:
do_state_transition: State transition S_FINALIZE_JOIN
-> S_POLICY_ENGINE [ input=I_FINALIZED
cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 17:22:19 [2695] node2 crmd: info:
abort_transition_graph: Transition aborted: Peer
Cancelled (source=do_te_invoke:151, 1)
Dec 18 17:22:19 [2693] node2 attrd: notice:
attrd_local_callback: Sending full refresh (origin=crmd)
Dec 18 17:22:19 [2693] node2 attrd: notice:
attrd_trigger_update: Sending flush op to all hosts for:
shutdown (0)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section nodes: OK (rc=0, origin=node2/crmd/90,
version=0.69.158)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: Diff: --- 0.69.158 2
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: Diff: +++ 0.69.159 (null)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: + /cib: @num_updates=159
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: + /cib/status/node_state[@id='node2']:
@crm-debug-origin=post_cache_update
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: + /cib/status/node_state[@id='node1']:
@in_ccm=false, @crm-debug-origin=post_cache_update
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section status: OK (rc=0, origin=node2/crmd/91,
version=0.69.159)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: Diff: --- 0.69.159 2
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: Diff: +++ 0.69.160 (null)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: + /cib: @num_updates=160,
@have-quorum=0
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section cib: OK (rc=0, origin=node2/crmd/92,
version=0.69.160)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section crm_config to master (origin=local/crmd/94)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section nodes to master (origin=local/crmd/95)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section status to master (origin=local/crmd/96)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section cib to master (origin=local/crmd/97)
Dec 18 17:22:19 [2693] node2 attrd: notice:
attrd_trigger_update: Sending flush op to all hosts for:
probe_complete (true)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section crm_config: OK (rc=0, origin=node2/crmd/94,
version=0.69.160)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section nodes: OK (rc=0, origin=node2/crmd/95,
version=0.69.160)
Dec 18 17:22:19 [2694] node2 pengine: notice:
unpack_config: On loss of CCM Quorum: Ignore
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: Diff: --- 0.69.160 2
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: Diff: +++ 0.69.161 (null)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: + /cib: @num_updates=161
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: + /cib/status/node_state[@id='node2']:
@crm-debug-origin=do_state_transition
Dec 18 17:22:19 [2690] node2 cib: info:
cib_perform_op: + /cib/status/node_state[@id='node1']:
@crm-debug-origin=do_state_transition, @join=down
Dec 18 17:22:19 [2694] node2 pengine: info:
determine_online_status: Node node2 is online
Dec 18 17:22:19 [2694] node2 pengine: info:
group_print: Resource Group: g_rbd_share_1
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section status: OK (rc=0, origin=node2/crmd/96,
version=0.69.161)
Dec 18 17:22:19 [2694] node2 pengine: info:
native_print: p_rbd_map_1 (ocf::ceph:rbd.in):
Stopped
Dec 18 17:22:19 [2694] node2 pengine: info:
native_print: p_fs_rbd_1
(ocf::heartbeat:Filesystem): Stopped
Dec 18 17:22:19 [2694] node2 pengine: info:
native_print: p_export_rbd_1
(ocf::heartbeat:exportfs): Stopped
Dec 18 17:22:19 [2694] node2 pengine: info:
native_print: p_vip_1
(ocf::heartbeat:IPaddr): Stopped
Dec 18 17:22:19 [2694] node2 pengine: info:
clone_print: Clone Set: clo_nfs [g_nfs]
Dec 18 17:22:19 [2694] node2 pengine: info:
short_print: Started: [ node2 ]
Dec 18 17:22:19 [2694] node2 pengine: info:
short_print: Stopped: [ node1 ]
Dec 18 17:22:19 [2694] node2 pengine: info:
rsc_merge_weights: p_rpcbind:1: Rolling back scores from
p_nfs_server:1
Dec 18 17:22:19 [2694] node2 pengine: info:
native_color: Resource p_rpcbind:1 cannot run anywhere
Dec 18 17:22:19 [2694] node2 pengine: info:
native_color: Resource p_nfs_server:1 cannot run
anywhere
Dec 18 17:22:19 [2694] node2 pengine: info:
RecurringOp: Start recurring monitor (10s) for
p_rbd_map_1 on node2
Dec 18 17:22:19 [2694] node2 pengine: info:
RecurringOp: Start recurring monitor (20s) for
p_fs_rbd_1 on node2
Dec 18 17:22:19 [2694] node2 pengine: info:
RecurringOp: Start recurring monitor (10s) for
p_export_rbd_1 on node2
Dec 18 17:22:19 [2694] node2 pengine: info:
RecurringOp: Start recurring monitor (5s) for p_vip_1
on node2
Dec 18 17:22:19 [2694] node2 pengine: notice:
LogActions: Start p_rbd_map_1 (node2)
Dec 18 17:22:19 [2694] node2 pengine: notice:
LogActions: Start p_fs_rbd_1 (node2)
Dec 18 17:22:19 [2694] node2 pengine: notice:
LogActions: Start p_export_rbd_1 (node2)
Dec 18 17:22:19 [2694] node2 pengine: notice:
LogActions: Start p_vip_1 (node2)
Dec 18 17:22:19 [2694] node2 pengine: info:
LogActions: Leave p_rpcbind:0 (Started node2)
Dec 18 17:22:19 [2694] node2 pengine: info:
LogActions: Leave p_nfs_server:0 (Started node2)
Dec 18 17:22:19 [2694] node2 pengine: info:
LogActions: Leave p_rpcbind:1 (Stopped)
Dec 18 17:22:19 [2694] node2 pengine: info:
LogActions: Leave p_nfs_server:1 (Stopped)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section cib: OK (rc=0, origin=node2/crmd/97,
version=0.69.161)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section status to master (origin=local/attrd/24)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section status: OK (rc=0, origin=node2/attrd/24,
version=0.69.161)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section status to master (origin=local/attrd/26)
Dec 18 17:22:19 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section status: OK (rc=0, origin=node2/attrd/26,
version=0.69.161)
Dec 18 17:22:19 [2694] node2 pengine: notice:
process_pe_message: Calculated Transition 3:
/var/lib/pacemaker/pengine/pe-input-234.bz2
Dec 18 17:22:19 [2695] node2 crmd: info:
do_state_transition: State transition S_POLICY_ENGINE
-> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
cause=C_IPC_MESSAGE origin=handle_response ]
Dec 18 17:22:19 [2695] node2 crmd: info:
do_te_invoke: Processing graph 3
(ref=pe_calc-dc-1450430539-52) derived from
/var/lib/pacemaker/pengine/pe-input-234.bz2
Dec 18 17:22:19 [2695] node2 crmd: notice:
te_rsc_command: Initiating action 6: start
p_rbd_map_1_start_0 on node2 (local)
Dec 18 17:22:19 [2695] node2 crmd: info:
do_lrm_rsc_op: Performing
key=6:3:0:1b17b95d-a029-4ea5-be6d-4e5d8add6ca9
op=p_rbd_map_1_start_0
Dec 18 17:22:19 [2692] node2 lrmd: info:
log_execute: executing - rsc:p_rbd_map_1 action:start
call_id:48
Dec 18 17:22:24 [2690] node2 cib: info:
cib_process_ping: Reporting our current digest to node2:
29875b5ae8dcbbb1d07f904ca3423de2 for 0.69.161 (0x12607b0 0)
Dec 18 17:22:39 [2692] node2 lrmd: warning:
child_timeout_callback: p_rbd_map_1_start_0 process (PID
11010) timed out
Dec 18 17:22:39 [2692] node2 lrmd: warning:
operation_finished: p_rbd_map_1_start_0:11010 - timed
out after 20000ms
Dec 18 17:22:39 [2692] node2 lrmd: notice:
operation_finished: p_rbd_map_1_start_0:11010:stderr [
libust[11019/11019]: Warning: HOME environment variable not
set. Disabling LTTng-UST per-user tracing. (in
setup_local_apps() at lttng-ust-comm.c:305) ]
Dec 18 17:22:39 [2692] node2 lrmd: info:
log_finished: finished - rsc:p_rbd_map_1 action:start
call_id:48 pid:11010 exit-code:1 exec-time:20002ms
queue-time:0ms
Dec 18 17:22:39 [2695] node2 crmd: info:
services_os_action_execute: Managed rbd.in_meta-data_0
process 11117 exited with rc=0
Dec 18 17:22:39 [2695] node2 crmd: error:
process_lrm_event: Operation p_rbd_map_1_start_0: Timed
Out (node=node2, call=48, timeout=20000ms)
Dec 18 17:22:39 [2692] node2 lrmd: warning:
child_timeout_callback: p_rbd_map_1_start_0 process (PID
11010) timed out
Dec 18 17:22:39 [2692] node2 lrmd: warning:
operation_finished: p_rbd_map_1_start_0:11010 - timed
out after 20000ms
Dec 18 17:22:39 [2692] node2 lrmd: notice:
operation_finished: p_rbd_map_1_start_0:11010:stderr [
libust[11019/11019]: Warning: HOME environment variable not
set. Disabling LTTng-UST per-user tracing. (in
setup_local_apps() at lttng-ust-comm.c:305) ]
Dec 18 17:22:39 [2692] node2 lrmd: info:
log_finished: finished - rsc:p_rbd_map_1 action:start
call_id:48 pid:11010 exit-code:1 exec-time:20002ms
queue-time:0ms
Dec 18 17:22:39 [2695] node2 crmd: info:
services_os_action_execute: Managed rbd.in_meta-data_0
process 11117 exited with rc=0
Dec 18 17:22:39 [2695] node2 crmd: error:
process_lrm_event: Operation p_rbd_map_1_start_0: Timed
Out (node=node2, call=48, timeout=20000ms)
Dec 18 17:22:39 [2695] node2 crmd: notice:
process_lrm_event: node2-p_rbd_map_1_start_0:48 [
libust[11019/11019]: Warning: HOME environment variable not
set. Disabling LTTng-UST per-user tracing. (in
setup_local_apps() at lttng-ust-comm.c:305)\n ]
Dec 18 17:22:39 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section status to master (origin=local/crmd/99)
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: Diff: --- 0.69.161 2
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: Diff: +++ 0.69.162 (null)
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: + /cib: @num_updates=162
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: + /cib/status/node_state[@id='node2']:
@crm-debug-origin=do_update_resource
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: +
/cib/status/node_state[@id='node2']/lrm[@id='node2']/lrm_resources/lrm_resource[@id='p_rbd_map_1']/lrm_rsc_op[@id='p_rbd_map_1_last_0']:
@operation_key=p_rbd_map_1_start_0, @operation=start,
@transition-key=6:3:0:1b17b95d-a029-4ea5-be6d-4e5d8add6ca9,
@transition-magic=2:1;6:3:0:1b17b95d-a029-4ea5-be6d-4e5d8add6ca9,
@call-id=48, @rc-code=1, @op-status=2, @last-run=1450430539,
@last-rc-change=1450430539, @exec-time=20002
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: ++
/cib/status/node_state[@id='node2']/lrm[@id='node2']/lrm_resources/lrm_resource[@id='p_rbd_map_1']:
<lrm_rsc_op id="p_rbd_map_1_last_failure_0"
operation_key="p_rbd_map_1_start_0" operation="start"
crm-debug-origin="do_update_resource"
crm_feature_set="3.0.9"
transition-key="6:3:0:1b17b95d-a029-4ea5-be6d-4e5d8add6ca9"
transition-magic="2:1;6:3:0:1b17b95d-a029-4ea5-be6d-4e5d8add6ca9"
call-id="48" rc-code="1" op-status="2" interval="0" l
Dec 18 17:22:39 [2690] node2 cib: info:
cib_process_request: Completed cib_modify operation for
section status: OK (rc=0, origin=node2/crmd/99,
version=0.69.162)
Dec 18 17:22:39 [2695] node2 crmd: warning:
status_from_rc: Action 6 (p_rbd_map_1_start_0) on node2
failed (target: 0 vs. rc: 1): Error
Dec 18 17:22:39 [2695] node2 crmd: warning:
update_failcount: Updating failcount for p_rbd_map_1 on
node2 after failed start: rc=1 (update=INFINITY,
time=1450430559)
Dec 18 17:22:39 [2695] node2 crmd: notice:
abort_transition_graph: Transition aborted by
p_rbd_map_1_start_0 'modify' on node2: Event failed
(magic=2:1;6:3:0:1b17b95d-a029-4ea5-be6d-4e5d8add6ca9,
cib=0.69.162, source=match_graph_event:344, 0)
Dec 18 17:22:39 [2695] node2 crmd: info:
match_graph_event: Action p_rbd_map_1_start_0 (6)
confirmed on node2 (rc=4)
Dec 18 17:22:39 [2693] node2 attrd: notice:
attrd_trigger_update: Sending flush op to all hosts for:
fail-count-p_rbd_map_1 (INFINITY)
Dec 18 17:22:39 [2695] node2 crmd: warning:
update_failcount: Updating failcount for p_rbd_map_1 on
node2 after failed start: rc=1 (update=INFINITY,
time=1450430559)
Dec 18 17:22:39 [2695] node2 crmd: info:
process_graph_event: Detected action (3.6)
p_rbd_map_1_start_0.48=unknown error: failed
Dec 18 17:22:39 [2695] node2 crmd: warning:
status_from_rc: Action 6 (p_rbd_map_1_start_0) on node2
failed (target: 0 vs. rc: 1): Error
Dec 18 17:22:39 [2695] node2 crmd: warning:
update_failcount: Updating failcount for p_rbd_map_1 on
node2 after failed start: rc=1 (update=INFINITY,
time=1450430559)
Dec 18 17:22:39 [2695] node2 crmd: info:
abort_transition_graph: Transition aborted by
p_rbd_map_1_start_0 'create' on (null): Event failed
(magic=2:1;6:3:0:1b17b95d-a029-4ea5-be6d-4e5d8add6ca9,
cib=0.69.162, source=match_graph_event:344, 0)
Dec 18 17:22:39 [2695] node2 crmd: info:
match_graph_event: Action p_rbd_map_1_start_0 (6)
confirmed on node2 (rc=4)
Dec 18 17:22:39 [2695] node2 crmd: warning:
update_failcount: Updating failcount for p_rbd_map_1 on
node2 after failed start: rc=1 (update=INFINITY,
time=1450430559)
Dec 18 17:22:39 [2695] node2 crmd: info:
process_graph_event: Detected action (3.6)
p_rbd_map_1_start_0.48=unknown error: failed
Dec 18 17:22:39 [2693] node2 attrd: notice:
attrd_perform_update: Sent update 28:
fail-count-p_rbd_map_1=INFINITY
Dec 18 17:22:39 [2690] node2 cib: info:
cib_process_request: Forwarding cib_modify operation for
section status to master (origin=local/attrd/28)
Dec 18 17:22:39 [2695] node2 crmd: notice:
run_graph: Transition 3 (Complete=2, Pending=0, Fired=0,
Skipped=8, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-234.bz2): Stopped
Dec 18 17:22:39 [2695] node2 crmd: info:
do_state_transition: State transition
S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC
cause=C_FSA_INTERNAL origin=notify_crmd ]
Dec 18 17:22:39 [2693] node2 attrd: notice:
attrd_trigger_update: Sending flush op to all hosts for:
last-failure-p_rbd_map_1 (1450430559)
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: Diff: --- 0.69.162 2
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: Diff: +++ 0.69.163 (null)
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: + /cib: @num_updates=163
Dec 18 17:22:39 [2690] node2 cib: info:
cib_perform_op: ++
/cib/status/node_state[@id='node2']/transient_attributes[@id='node2']/instance_attributes[@id='status-node2']:
<nvpair id="status-node2-fail-count-p_rbd_map_1"
name="fail-count-p_rbd_map_1" value="INFINITY"/>
.........
thanks