[Pacemaker] "crm_node -f -R NodeID" remove the wrong node?
Andrew Beekhof
andrew at beekhof.net
Mon Apr 14 07:08:44 UTC 2014
On 13 Apr 2014, at 11:36 am, Caizhifeng <bluewindow at h3c.com> wrote:
> Hi all,
> I'm kinda new to Pacemaker and have some questions I hope some of you could help me with, any idea(s) would be highly appreciated, thank you.
> I'am building the HA cluster with corosync2.3.3 + cluster-glue + pacemaker-1.1.11-rc5. The problem is, when I try to remove a node from HA cluster with "crm_node -f -R nodeID", it seems to remove the wrong node.
>
> My setup is as follows(and it can be reproduced):
> 1. The HA is ok with 2 resource and 2 nodes, and with stonith-enabled="false":
> root at h1:/opt/bin# crm status
> Last updated: Sat Apr 12 16:00:51 2014
> Last change: Sat Apr 12 16:00:43 2014 via cibadmin on h1
> Stack: corosync
> Current DC: h1 (1084752017) - partition with quorum
> Version: 1.1.10-33f9d09
> 2 Nodes configured
> 2 Resources configured
>
>
> Online: [ h0 h1 ]
>
> VMdos-1 (ocf::heartbeat:VirtualDomain): Started h1
> root at h1:/opt/bin#
>
> root at h1:/opt/bin# crm_node --list
> 1084752017 h1
> 1084752016 h0
> root at h1:/opt/bin#
>
> 2. remove node h0 from HA
> (1). Stop pacemakerd and corosync service on node h0 (2). Run the follow cmd on node h1 to Remove node info from cib:
> cibadmin --delete --obj_type status --crm_xml "<node_state id=\"1084752016\"/>"
> cibadmin --delete --obj_type nodes --crm_xml "<node id=\"1084752016\"/>"
>
> at this time; the result of "crm status" shows that node " h0 " is offline and there's no information of node h0 in CIB. But the "crm_node --list " still include node" h0"
>
> 3. crm_node -f -R 1084752016
Could you re-run this command but add -VVVVVVVV please?
I suspect the fault lies in this binary.
> After this cmd, node " h1" is removed, and result of "crm status " is as follow(with 0 node):
>
> root at h1:/opt/bin# crm status
> Last updated: Sat Apr 12 15:59:42 2014
> Last change: Sat Apr 12 15:59:37 2014 via crm_node on h1
> Stack: corosync
> Current DC: NONE
> 0 Nodes configured
> 2 Resources configured
> root at h1:/opt/bin
>
>
> The corosync.conf is as follow:
> .............
> quorum {
> provider: corosync_votequorum
> expected_votes: 2
> allow_downscale: 1
> two_node: 1
> }
> ...........
>
>
> The corosync.log is as follow, it seems node " h1" is removed due to fencing!
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-25004-14152-27-header
> Apr 12 16:22:35 [25026] h1 cib: debug: activateCibXml: Triggering CIB write for cib_delete op
> Apr 12 16:22:35 [25031] h1 crmd: notice: crm_reap_dead_member: Removing h0/1084752016 from the membership list
> Apr 12 16:22:35 [25031] h1 crmd: notice: reap_crm_member: Purged 1 peers with id=1084752016 and/or uname=(null) from the membership cache
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: log_cib_diff: Config update: Local-only Change: 0.12.1
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - <cib admin_epoch="0" epoch="11" num_updates="1">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - <configuration>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - <nodes>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <node id="1084752017" uname="h1"/>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - </nodes>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - </configuration>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - </cib>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: ++ <cib epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2" crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" update-origin="h1" update-client="crm_node" have-quorum="1" dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25026] h1 cib: notice: log_cib_diff: cib:diff: Local-only Change: 0.12.1
> Apr 12 16:22:35 [25026] h1 cib: notice: cib:diff: -- <node id="1084752017" uname="h1"/>
> Apr 12 16:22:35 [25026] h1 cib: notice: cib:diff: ++ <cib epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2" crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" update-origin="h1" update-client="crm_node" have-quorum="1" dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request: Completed cib_delete operation for section nodes: OK (rc=0, origin=local/crm_node/2, version=0.12.1)
> Apr 12 16:22:35 [25031] h1 crmd: debug: te_update_diff: Processing diff (cib_delete): 0.11.1 -> 0.12.1 (S_IDLE)
> Apr 12 16:22:35 [25031] h1 crmd: info: abort_transition_graph: te_update_diff:126 - Triggered transition abort (complete=1, node=, tag=diff, id=(null), magic=NA, cib=0.12.1) : Non-status change
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <diff crm_feature_set="3.0.8" digest="3cccef06483ac4dfeadfb562f6f8293a">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <diff-removed admin_epoch="0" epoch="11" num_updates="1">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <cib admin_epoch="0" epoch="11" num_updates="1">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <configuration>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <nodes>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <node id="1084752017" uname="h1" __crm_diff_marker__="removed:top"/>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause </nodes>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause </configuration>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause </cib>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause </diff-removed>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <diff-added>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <cib epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2" crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" update-origin="h1" update-client="crm_node" have-quorum="1" dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause </diff-added>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause </diff>
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing I_PE_CALC: [ state=S_IDLE cause=C_FSA_INTERNAL origin=abort_transition_graph ]
> Apr 12 16:22:35 [25031] h1 crmd: notice: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_state_transition: All 1 cluster nodes are eligible to run resources.
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke: Query 38: Requesting the current CIB: S_POLICY_ENGINE
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/38, version=0.12.1)
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke_callback: Invoking the PE: query=38, ref=pe_calc-dc-1397290955-21, seq=1024, quorate=1
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: Diff: --- 0.12.1
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: Diff: +++ 0.12.2 3d673c27c3c92939b41c7207edee9f46
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - <cib num_updates="1">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - <status>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <node_state id="1084752017" uname="h1" in_ccm="true" crmd="online" crm-debug-origin="post_cache_update" join="member" expected="member">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <transient_attributes id="1084752017">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <instance_attributes id="status-1084752017">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <nvpair id="status-1084752017-shutdown" name="shutdown" value="0"/>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <nvpair id="status-1084752017-probe_complete" name="probe_complete" value="true"/>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- </instance_attributes>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- </transient_attributes>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <lrm id="1084752017">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <lrm_resources>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <lrm_resource id="VMdos-1" type="VirtualDomain" class="ocf" provider="heartbeat">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <lrm_rsc_op id="VMdos-1_last_failure_0" operation_key="VMdos-1_monitor_0" operation="monitor" crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8" transition-key="6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c" transition-magic="0:0;6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="5" rc-code="0" op-status="0" interval="0" last-run="1397290607" last-rc-change="1397290607" exec-time="92" queue-time="0" op-digest="
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <lrm_rsc_op id="VMdos-1_monitor_30000" operation_key="VMdos-1_monitor_30000" operation="monitor" crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8" transition-key="7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c" transition-magic="0:0;7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="6" rc-code="0" op-status="0" interval="30000" last-rc-change="1397290607" exec-time="67" queue-time="0" op-digest="0874c7ce5f61a12
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- </lrm_resource>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <lrm_resource id="VMdos-2" type="VirtualDomain" class="ocf" provider="heartbeat">
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- <lrm_rsc_op id="VMdos-2_last_0" operation_key="VMdos-2_monitor_0" operation="monitor" crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8" transition-key="7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c" transition-magic="0:7;7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="10" rc-code="7" op-status="0" interval="0" last-run="1397290608" last-rc-change="1397290608" exec-time="61" queue-time="0" op-digest="c7d22be
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- </lrm_resource>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- </lrm_resources>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- </lrm>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: -- </node_state>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - </status>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: - </cib>
> Apr 12 16:22:35 [25027] h1 stonith-ng: debug: Config update: ++ <cib epoch="12" num_updates="2" admin_epoch="0" validate-with="pacemaker-1.2" crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" update-origin="h1" update-client="crm_node" have-quorum="1" dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: STONITH timeout: 60000
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: STONITH of failed nodes is disabled
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Stop all active resources: false
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Cluster is symmetric - resources can run anywhere by default
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Default stickiness: 0
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: On loss of CCM Quorum: Stop ALL resources
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_domains: Unpacking domains
> Apr 12 16:22:35 [25030] h1 pengine: warning: unpack_status: Node h1 in status section no longer exists
> Apr 12 16:22:35 [25030] h1 pengine: info: unpack_status: Node 1084752017 is unknown
> Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-1 (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-2 (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could not allocate a node for VMdos-1
> Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource VMdos-1 cannot run anywhere
> Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could not allocate a node for VMdos-2
> Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource VMdos-2 cannot run anywhere
> Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave VMdos-1 (Stopped)
> Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave VMdos-2 (Stopped)
> Apr 12 16:22:35 [25030] h1 pengine: notice: process_pe_message: Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-4.bz2
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE origin=handle_response ]
> Apr 12 16:22:35 [25031] h1 crmd: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: unpack_graph: Unpacked transition 4: 0 actions in 0 synapses
> Apr 12 16:22:35 [25031] h1 crmd: info: do_te_invoke: Processing graph 4 (ref=pe_calc-dc-1397290955-21) derived from /var/lib/pacemaker/pengine/pe-input-4.bz2
> Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty transition graph
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request: Completed cib_delete operation for section status: OK (rc=0, origin=local/crm_node/3, version=0.12.2)
> Apr 12 16:22:35 [25031] h1 crmd: debug: te_update_diff: Processing diff (cib_delete): 0.12.1 -> 0.12.2 (S_TRANSITION_ENGINE)
> Apr 12 16:22:35 [25031] h1 crmd: info: abort_transition_graph: te_update_diff:188 - Triggered transition abort (complete=0, node=h1, tag=transient_attributes, id=1084752017, magic=NA, cib=0.12.2) : Transient attribute: removal
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <transient_attributes id="1084752017">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <instance_attributes id="status-1084752017">
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <nvpair id="status-1084752017-shutdown" name="shutdown" value="0"/>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause <nvpair id="status-1084752017-probe_complete" name="probe_complete" value="true"/>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause </instance_attributes>
> Apr 12 16:22:35 [25031] h1 crmd: debug: abort_transition_graph: Cause </transient_attributes>
> Apr 12 16:22:35 [25031] h1 crmd: debug: update_abort_priority: Abort priority upgraded from 0 to 1000000
> Apr 12 16:22:35 [25031] h1 crmd: debug: update_abort_priority: Abort action done superceeded by restart
> Apr 12 16:22:35 [25031] h1 crmd: notice: run_graph: Transition 4 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-4.bz2): Complete
> Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty transition graph
> Apr 12 16:22:35 [25031] h1 crmd: debug: te_graph_trigger: Transition 4 is now complete
> Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Processing transition completion in state S_TRANSITION_ENGINE
> Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Transition 4 status: restart - Transient attribute: removal
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing I_PE_CALC: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1 crmd: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_state_transition: All 1 cluster nodes are eligible to run resources.
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke: Query 39: Requesting the current CIB: S_POLICY_ENGINE
> Apr 12 16:22:35 [25024] h1 pacemakerd: info: crm_client_new: Connecting 0x25b4ea0 for uid=0 gid=0 pid=14152 id=f3612e17-0806-4355-a3fc-2cf1feda1e6d
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: handle_new_connection: IPC credentials authenticated (25024-14152-10)
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_ipcs_shm_connect: connecting to client [14152]
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/39, version=0.12.2)
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_rb_open_2: shm size:131085; real_size:135168; rb->word_size:33792
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_dispatch_connection_request: HUP conn (25026-14152-13)
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_disconnect: qb_ipcs_disconnect(25026-14152-13) state:2
> Apr 12 16:22:35 [25026] h1 cib: info: crm_client_destroy: Destroying 0 events
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_rb_close: Free'ing ringbuffer: /dev/shm/qb-cib_rw-response-25026-14152-13-header
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_rb_close: Free'ing ringbuffer: /dev/shm/qb-cib_rw-event-25026-14152-13-header
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_rb_open_2: shm size:131085; real_size:135168; rb->word_size:33792
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_rb_open_2: shm size:131085; real_size:135168; rb->word_size:33792
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipc_setup.c:478 IPC credentials authenticated (25004-14152-27)
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipc_shm.c:294 connecting to client [14152]
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:236 shm size:1048589; real_size:1052672; rb->word_size:263168
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_pe_invoke_callback: Invoking the PE: query=39, ref=pe_calc-dc-1397290955-22, seq=1024, quorate=1
> Apr 12 16:22:35 [25031] h1 crmd: debug: qb_ipcs_dispatch_connection_request: HUP conn (25031-14152-14)
> Apr 12 16:22:35 [25031] h1 crmd: debug: qb_ipcs_disconnect: qb_ipcs_disconnect(25031-14152-14) state:2
> Apr 12 16:22:35 [25031] h1 crmd: info: crm_client_destroy: Destroying 0 events
> Apr 12 16:22:35 [25031] h1 crmd: debug: qb_rb_close: Free'ing ringbuffer: /dev/shm/qb-crmd-response-25031-14152-14-header
> ...
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: STONITH timeout: 60000
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: STONITH of failed nodes is disabled
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Stop all active resources: false
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Cluster is symmetric - resources can run anywhere by default
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Default stickiness: 0
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: On loss of CCM Quorum: Stop ALL resources
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
> Apr 12 16:22:35 [25030] h1 pengine: debug: unpack_domains: Unpacking domains
> Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-1 (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1 pengine: info: native_print: VMdos-2 (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could not allocate a node for VMdos-1
> Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource VMdos-1 cannot run anywhere
> Apr 12 16:22:35 [25030] h1 pengine: debug: native_assign_node: Could not allocate a node for VMdos-2
> Apr 12 16:22:35 [25030] h1 pengine: info: native_color: Resource VMdos-2 cannot run anywhere
> Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave VMdos-1 (Stopped)
> Apr 12 16:22:35 [25030] h1 pengine: info: LogActions: Leave VMdos-2 (Stopped)
> Apr 12 16:22:35 [25030] h1 pengine: notice: process_pe_message: Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-5.bz2
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE origin=handle_response ]
> Apr 12 16:22:35 [25031] h1 crmd: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: unpack_graph: Unpacked transition 5: 0 actions in 0 synapses
> Apr 12 16:22:35 [25031] h1 crmd: info: do_te_invoke: Processing graph 5 (ref=pe_calc-dc-1397290955-22) derived from /var/lib/pacemaker/pengine/pe-input-5.bz2
> Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty transition graph
> Apr 12 16:22:35 [25031] h1 crmd: notice: run_graph: Transition 5 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-5.bz2): Complete
> Apr 12 16:22:35 [25031] h1 crmd: debug: print_graph: Empty transition graph
> Apr 12 16:22:35 [25031] h1 crmd: debug: te_graph_trigger: Transition 5 is now complete
> Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Processing transition completion in state S_TRANSITION_ENGINE
> Apr 12 16:22:35 [25031] h1 crmd: debug: notify_crmd: Transition 5 status: done - <null>
> Apr 12 16:22:35 [25031] h1 crmd: debug: s_crmd_fsa: Processing I_TE_SUCCESS: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1 crmd: info: do_log: FSA: Input I_TE_SUCCESS from notify_crmd() received in state S_TRANSITION_ENGINE
> Apr 12 16:22:35 [25031] h1 crmd: notice: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1 crmd: debug: do_state_transition: Starting PEngine Recheck Timer
> Apr 12 16:22:35 [25031] h1 crmd: debug: crm_timer_start: Started PEngine Recheck Timer (I_PE_CALC:900000ms), src=62
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:236 shm size:1048589; real_size:1052672; rb->word_size:263168
> Apr 12 16:22:35 [25026] h1 cib: info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-9.raw
> Apr 12 16:22:35 [25026] h1 cib: debug: write_cib_contents: Writing CIB to disk
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_rb_close: Free'ing ringbuffer: /dev/shm/qb-cib_rw-request-25026-14152-13-header
> Apr 12 16:22:35 [25026] h1 cib: info: write_cib_contents: Wrote version 0.12.0 of the CIB to disk (digest: bd7d26226d6aa75f28b9eb670a67e944)
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ringbuffer.c:236 shm size:1048589; real_size:1052672; rb->word_size:263168
> Apr 12 16:22:35 [25003] h1 corosync debug [MAIN ] ipc_glue.c:272 connection created
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] cmap.c:306 lib_init_fn: conn=0x7f32d0f4f4b0
> Apr 12 16:22:35 [25024] h1 pacemakerd: notice: pcmk_ipc_dispatch: Instructing peers to remove references to node (null)/1084752016
> Apr 12 16:22:35 [25026] h1 cib: info: crm_client_new: Connecting 0xa23770 for uid=0 gid=0 pid=14152 id=701dfa3c-d585-49d2-bd1e-f44636823e1b
> Apr 12 16:22:35 [25026] h1 cib: debug: handle_new_connection: IPC credentials authenticated (25026-14152-13)
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_shm_connect: connecting to client [14152]
> Apr 12 16:22:35 [25026] h1 cib: debug: write_cib_contents: Wrote digest bd7d26226d6aa75f28b9eb670a67e944 to disk
> Apr 12 16:22:35 [25026] h1 cib: info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.y6wnAg (digest: /var/lib/pacemaker/cib/cib.ncXFvs)
> ...
> Apr 12 16:22:35 [25026] h1 cib: debug: write_cib_contents: Activating /var/lib/pacemaker/cib/cib.y6wnAg
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request: Completed cib_delete operation for section nodes: OK (rc=0, origin=local/crm_node/2, version=0.12.2)
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipcs.c:757 HUP conn (25004-14152-27)
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] ipcs.c:605 qb_ipcs_disconnect(25004-14152-27) state:2
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] loop_poll_epoll.c:117 epoll_ctl(del): Bad file descriptor (9)
> Apr 12 16:22:35 [25003] h1 corosync debug [MAIN ] ipc_glue.c:417 cs_ipcs_connection_closed()
> Apr 12 16:22:35 [25003] h1 corosync debug [QB ] cmap.c:325 exit_fn for conn=0x7f32d0f4f4b0
> Apr 12 16:22:35 [25003] h1 corosync debug [MAIN ] ipc_glue.c:390 cs_ipcs_connection_destroyed()
> Apr 12 16:22:35 [25026] h1 cib: info: cib_process_request: Completed cib_delete operation for section status: OK (rc=0, origin=local/crm_node/3, version=0.12.2)
> ...
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_ipcs_dispatch_connection_request: HUP conn (25024-14152-10)
> Apr 12 16:22:35 [25024] h1 pacemakerd: debug: qb_ipcs_disconnect: qb_ipcs_disconnect(25024-14152-10) state:2
> Apr 12 16:22:35 [25024] h1 pacemakerd: info: crm_client_destroy: Destroying 0 events
> ...
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_dispatch_connection_request: HUP conn (25026-14152-13)
> Apr 12 16:22:35 [25026] h1 cib: debug: qb_ipcs_disconnect: qb_ipcs_disconnect(25026-14152-13) state:2
> Apr 12 16:22:35 [25026] h1 cib: info: crm_client_destroy: Destroying 0 events
> ...
> Apr 12 16:22:35 [25024] h1 pacemakerd: notice: crm_reap_dead_member: Removing h0/1084752016 from the membership list
> Apr 12 16:22:35 [25024] h1 pacemakerd: notice: reap_crm_member: Purged 1 peers with id=1084752016 and/or uname=(null) from the membership cache
>
>
> -------------------------------------------------------------------------------------------------------------------------------------
> 本邮件及其附件含有杭州华三通信技术有限公司的保密信息,仅限于发送给上面地址中列出
> 的个人或群组。禁止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、
> 或散发)本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本
> 邮件!
> This e-mail and its attachments contain confidential information from H3C, which is
> intended only for the person or entity whose address is listed above. Any use of the
> information contained herein in any way (including, but not limited to, total or partial
> disclosure, reproduction, or dissemination) by persons other than the intended
> recipient(s) is prohibited. If you receive this e-mail in error, please notify the sender
> by phone or email immediately and delete it!
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 841 bytes
Desc: Message signed with OpenPGP using GPGMail
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20140414/08232fbb/attachment-0004.sig>
More information about the Pacemaker
mailing list