[Pacemaker] crmd restart due to internal error - pacemaker 1.1.8

pavan tc pavan.tc at gmail.com
Wed May 8 13:16:45 CEST 2013


Hi,

I have a two-node cluster with STONITH disabled. I am still running with
the pcmk plugin as opposed to the recommended CMAN plugin.

With 1.1.8, I see some messages (appended to this mail) once in a while. I
do not understand some keywords here - There is a "Leave" action. I am not
sure what that is. And, there is a CIB update failure that leads to a
RECOVER action. There is a message that says the RECOVER action is not
supported. Finally this leads to a stop and start of my resource. I can
copy the "crm configure show" output, but nothing special there.

Thanks much.
Pavan

PS: The resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb is stale. The
underlying device that represents this resource has been removed. However,
the resource is still part of the CIB. All errors related to that resource
can be ignored. But can this cause a node to be stopped/fenced?

----------------------------------------------------------------------------------------------------------------------------------------
May 07 05:15:24 [10845] vsan15    pengine:     info: short_print:
Masters: [ vsan15 ]
May 07 05:15:24 [10845] vsan15    pengine:     info: short_print:
Stopped: [ vha-090f26ed-5991-4f40-833e-02e76759dd41:1 ]
May 07 05:15:24 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan15
May 07 05:15:24 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:15:24 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan15
May 07 05:15:24 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:15:24 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan16
May 07 05:15:24 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:15:24 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan16
May 07 05:15:24 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:15:24 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 cannot run anywhere
May 07 05:15:24 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 cannot run anywhere
May 07 05:15:24 [10845] vsan15    pengine:     info: master_color:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb: Promoted 0 instances of a possible
1 to master
May 07 05:15:24 [10845] vsan15    pengine:     info: native_color:
Resource vha-090f26ed-5991-4f40-833e-02e76759dd41:1 cannot run anywhere
May 07 05:15:24 [10845] vsan15    pengine:     info: master_color:
Promoting vha-090f26ed-5991-4f40-833e-02e76759dd41:0 (Master vsan15)
May 07 05:15:24 [10845] vsan15    pengine:     info: master_color:
ms-090f26ed-5991-4f40-833e-02e76759dd41: Promoted 1 instances of a possible
1 to master
May 07 05:15:24 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (30s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:0 on vsan15
May 07 05:15:24 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (30s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:0 on vsan15
May 07 05:15:24 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0    (Stopped)
May 07 05:15:24 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1    (Stopped)
May 07 05:15:24 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-090f26ed-5991-4f40-833e-02e76759dd41:0    (Master vsan15)
May 07 05:15:24 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-090f26ed-5991-4f40-833e-02e76759dd41:1    (Stopped)
May 07 05:15:24 [10845] vsan15    pengine:   notice: process_pe_message:
    Calculated Transition 9: /var/lib/pacemaker/pengine/pe-input-189.bz2
May 07 05:15:24 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
May 07 05:15:24 [10846] vsan15       crmd:     info: do_te_invoke:
Processing graph 9 (ref=pe_calc-dc-1367928924-43) derived from
/var/lib/pacemaker/pengine/pe-input-189.bz2
May 07 05:15:24 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 16: monitor
vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_30000 on vsan15 (local)
May 07 05:15:24 [10846] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_30000
(call=50, rc=8, cib-update=66, confirmed=false) master
May 07 05:15:24 [10846] vsan15       crmd:   notice: run_graph:
Transition 9 (Complete=1, Pending=0, Fired=0, Skipped=0, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-189.bz2): Complete
May 07 05:15:24 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
cause=C_FSA_INTERNAL origin=notify_crmd ]
May 07 05:16:25 corosync [pcmk  ] notice: pcmk_peer_update: Transitional
membership event on ring 14136: memb=1, new=0, lost=0
May 07 05:16:25 corosync [pcmk  ] info: pcmk_peer_update: memb: vsan15
1682182316
May 07 05:16:25 corosync [pcmk  ] notice: pcmk_peer_update: Stable
membership event on ring 14136: memb=2, new=1, lost=0
May 07 05:16:25 corosync [pcmk  ] info: update_member: Node
1698959532/vsan16 is now: member
May 07 05:16:25 corosync [pcmk  ] info: pcmk_peer_update: NEW:  vsan16
1698959532
May 07 05:16:25 corosync [pcmk  ] info: pcmk_peer_update: MEMB: vsan15
1682182316
May 07 05:16:25 [10846] vsan15       crmd:   notice: ais_dispatch_message:
    Membership 14136: quorum acquired
May 07 05:16:25 corosync [pcmk  ] info: pcmk_peer_update: MEMB: vsan16
1698959532
May 07 05:16:25 [10846] vsan15       crmd:   notice: crm_update_peer_state:
    crm_update_ais_node: Node vsan16[1698959532] - state is now member
May 07 05:16:25 corosync [pcmk  ] info: send_member_notification: Sending
membership update 14136 to 2 children
May 07 05:16:25 [10846] vsan15       crmd:     info: peer_update_callback:
    vsan16 is now member (was lost)
May 07 05:16:25 corosync [TOTEM ] A processor joined or left the membership
and a new membership was formed.
May 07 05:16:25 [10841] vsan15        cib:   notice: ais_dispatch_message:
    Membership 14136: quorum acquired
May 07 05:16:25 [10841] vsan15        cib:   notice: crm_update_peer_state:
    crm_update_ais_node: Node vsan16[1698959532] - state is now member
May 07 05:16:25 corosync [pcmk  ] info: update_member: 0x15d4730 Node
1698959532 (vsan16) born on: 14136
May 07 05:16:25 corosync [pcmk  ] info: send_member_notification: Sending
membership update 14136 to 2 children
May 07 05:16:25 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/67, version=0.1344.91): OK (rc=0)
May 07 05:16:25 [10841] vsan15        cib:     info: ais_dispatch_message:
    Membership 14136: quorum retained
May 07 05:16:25 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section cib
(origin=local/crmd/69, version=0.1344.93): OK (rc=0)
May 07 05:16:25 [10846] vsan15       crmd:     info: crmd_ais_dispatch:
Setting expected votes to 2
May 07 05:16:25 [10846] vsan15       crmd:     info: ais_dispatch_message:
    Membership 14136: quorum retained
May 07 05:16:25 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/71, version=0.1344.94): OK (rc=0)
May 07 05:16:25 corosync [CPG   ] chosen downlist: sender r(0)
ip(172.16.68.100) ; members(old:1 left:0)
May 07 05:16:25 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/72, version=0.1344.95): OK (rc=0)
May 07 05:16:25 corosync [MAIN  ] Completed service synchronization, ready
to provide service.
May 07 05:16:25 [10846] vsan15       crmd:     info: crmd_ais_dispatch:
Setting expected votes to 2
May 07 05:16:25 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/75, version=0.1344.97): OK (rc=0)
May 07 05:16:28 [10842] vsan15 stonith-ng:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10841] vsan15        cib:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10842] vsan15 stonith-ng:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10841] vsan15        cib:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: peer_update_callback:
    Client vsan16/peer now has status [offline] (DC=true)
May 07 05:16:28 [10846] vsan15       crmd:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10841] vsan15        cib:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: peer_update_callback:
    Client vsan16/peer now has status [offline] (DC=true)
May 07 05:16:28 [10842] vsan15 stonith-ng:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: peer_update_callback:
    Client vsan16/peer now has status [offline] (DC=true)
May 07 05:16:28 [10846] vsan15       crmd:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: peer_update_callback:
    Client vsan16/peer now has status [offline] (DC=true)
May 07 05:16:28 [10842] vsan15 stonith-ng:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10841] vsan15        cib:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: peer_update_callback:
    Client vsan16/peer now has status [offline] (DC=true)
May 07 05:16:28 [10841] vsan15        cib:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10842] vsan15 stonith-ng:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info: peer_update_callback:
    Client vsan16/peer now has status [online] (DC=true)
May 07 05:16:28 [10841] vsan15        cib:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:  warning: match_down_event:
No match for shutdown action on vsan16
May 07 05:16:28 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_IDLE -> S_INTEGRATION [ input=I_NODE_JOIN
cause=C_FSA_INTERNAL origin=peer_update_callback ]
May 07 05:16:28 [10842] vsan15 stonith-ng:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:16:28 [10846] vsan15       crmd:     info:
abort_transition_graph:     do_te_invoke:163 - Triggered transition abort
(complete=1) : Peer Halt
May 07 05:16:28 [10846] vsan15       crmd:     info: join_make_offer:
Making join offers based on membership 14136
May 07 05:16:28 [10846] vsan15       crmd:     info: do_dc_join_offer_all:
    join-2: Waiting on 2 outstanding join acks
May 07 05:16:28 [10846] vsan15       crmd:     info: update_dc:     Set DC
to vsan15 (3.0.7)
May 07 05:16:29 [10841] vsan15        cib:  warning: cib_process_diff:
Diff 0.1344.0 -> 0.1344.1 from vsan16 not applied to 0.1344.98: current
"num_updates" is greater than required
May 07 05:16:30 [10846] vsan15       crmd:     info: do_dc_join_offer_all:
    A new node joined the cluster
May 07 05:16:30 [10846] vsan15       crmd:     info: do_dc_join_offer_all:
    join-3: Waiting on 2 outstanding join acks
May 07 05:16:30 [10846] vsan15       crmd:     info: update_dc:     Set DC
to vsan15 (3.0.7)
May 07 05:16:31 [10846] vsan15       crmd:     info:
crm_update_peer_expected:     do_dc_join_filter_offer: Node
vsan16[1698959532] - expected state is now member
May 07 05:16:31 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED
cause=C_FSA_INTERNAL origin=check_join_state ]
May 07 05:16:31 [10846] vsan15       crmd:     info: do_dc_join_finalize:
    join-3: Syncing the CIB from vsan15 to the rest of the cluster
May 07 05:16:31 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_sync for section 'all'
(origin=local/crmd/79, version=0.1344.98): OK (rc=0)
May 07 05:16:31 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/80, version=0.1344.99): OK (rc=0)
May 07 05:16:31 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/81, version=0.1344.100): OK (rc=0)
May 07 05:16:31 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_delete for section
//node_state[@uname='vsan16']/transient_attributes (origin=vsan16/crmd/7,
version=0.1344.101): OK (rc=0)
May 07 05:16:32 [10846] vsan15       crmd:     info:
services_os_action_execute:     Managed vgc-cm-agent.ocf_meta-data_0
process 14403 exited with rc=0
May 07 05:16:32 [10846] vsan15       crmd:     info: do_dc_join_ack:
join-3: Updating node state to member for vsan16
May 07 05:16:32 [10846] vsan15       crmd:     info: erase_status_tag:
Deleting xpath: //node_state[@uname='vsan16']/lrm
May 07 05:16:32 [10846] vsan15       crmd:     info: do_dc_join_ack:
join-3: Updating node state to member for vsan15
May 07 05:16:32 [10846] vsan15       crmd:     info: erase_status_tag:
Deleting xpath: //node_state[@uname='vsan15']/lrm
May 07 05:16:32 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_delete for section
//node_state[@uname='vsan16']/lrm (origin=local/crmd/82,
version=0.1344.102): OK (rc=0)
May 07 05:16:32 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED
cause=C_FSA_INTERNAL origin=check_join_state ]
May 07 05:16:32 [10846] vsan15       crmd:     info:
abort_transition_graph:     do_te_invoke:156 - Triggered transition abort
(complete=1) : Peer Cancelled
May 07 05:16:32 [10844] vsan15      attrd:   notice: attrd_local_callback:
    Sending full refresh (origin=crmd)
May 07 05:16:32 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
master-vha-090f26ed-5991-4f40-833e-02e76759dd41 (4)
May 07 05:16:32 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_delete for section
//node_state[@uname='vsan15']/lrm (origin=local/crmd/84,
version=0.1344.104): OK (rc=0)
May 07 05:16:32 [10846] vsan15       crmd:     info:
abort_transition_graph:     te_update_diff:271 - Triggered transition abort
(complete=1, tag=lrm_rsc_op,
id=vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb_last_0,
magic=0:0;1:2:0:ec7dd1c6-710a-4cee-b9d3-26e09c6ffb53, cib=0.1344.104) :
Resource op removal
May 07 05:16:32 [10846] vsan15       crmd:     info:
abort_transition_graph:     te_update_diff:227 - Triggered transition abort
(complete=1) : LRM Refresh
May 07 05:16:32 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/86, version=0.1344.106): OK (rc=0)
May 07 05:16:32 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section cib
(origin=local/crmd/88, version=0.1344.108): OK (rc=0)
May 07 05:16:32 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
fail-count-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb (INFINITY)
May 07 05:16:32 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
last-failure-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb (1367926970)
May 07 05:16:32 [10845] vsan15    pengine:     info: unpack_config:
Startup probes: enabled
May 07 05:16:32 [10845] vsan15    pengine:   notice: unpack_config:     On
loss of CCM Quorum: Ignore
May 07 05:16:32 [10845] vsan15    pengine:     info: unpack_config:
Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
May 07 05:16:32 [10845] vsan15    pengine:     info: unpack_domains:
Unpacking domains
May 07 05:16:32 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan15 is online
May 07 05:16:32 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan16 is online
May 07 05:16:32 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan15
to vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
May 07 05:16:32 [10845] vsan15    pengine:  warning: unpack_rsc_op:
Processing failed op start for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
on vsan15: not running (7)
May 07 05:16:32 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-090f26ed-5991-4f40-833e-02e76759dd41 on vsan15
to vha-090f26ed-5991-4f40-833e-02e76759dd41:0
May 07 05:16:32 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb
[vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb]
May 07 05:16:32 [10845] vsan15    pengine:     info: short_print:
Stopped: [ vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 ]
May 07 05:16:32 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-090f26ed-5991-4f40-833e-02e76759dd41
[vha-090f26ed-5991-4f40-833e-02e76759dd41]
May 07 05:16:32 [10845] vsan15    pengine:     info: short_print:
Masters: [ vsan15 ]
May 07 05:16:32 [10845] vsan15    pengine:     info: short_print:
Stopped: [ vha-090f26ed-5991-4f40-833e-02e76759dd41:1 ]
May 07 05:16:32 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan15
May 07 05:16:32 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:16:32 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan15
May 07 05:16:32 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:16:32 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 cannot run anywhere
May 07 05:16:32 [10845] vsan15    pengine:     info: master_color:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb: Promoted 0 instances of a possible
1 to master
May 07 05:16:32 [10845] vsan15    pengine:     info: master_color:
Promoting vha-090f26ed-5991-4f40-833e-02e76759dd41:0 (Master vsan15)
May 07 05:16:32 [10845] vsan15    pengine:     info: master_color:
ms-090f26ed-5991-4f40-833e-02e76759dd41: Promoted 1 instances of a possible
1 to master
May 07 05:16:32 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for: probe_complete (true)
May 07 05:16:32 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 on vsan16
May 07 05:16:32 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 on vsan16
May 07 05:16:32 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:1 on vsan16
May 07 05:16:32 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:1 on vsan16
May 07 05:16:32 [10845] vsan15    pengine:   notice: LogActions:
Start   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0    (vsan16)
May 07 05:16:32 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1    (Stopped)
May 07 05:16:32 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-090f26ed-5991-4f40-833e-02e76759dd41:0    (Master vsan15)
May 07 05:16:32 [10845] vsan15    pengine:   notice: LogActions:
Start   vha-090f26ed-5991-4f40-833e-02e76759dd41:1    (vsan16)
May 07 05:16:32 [10845] vsan15    pengine:   notice: process_pe_message:
    Calculated Transition 10: /var/lib/pacemaker/pengine/pe-input-190.bz2
May 07 05:16:32 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
May 07 05:16:32 [10846] vsan15       crmd:     info: do_te_invoke:
Processing graph 10 (ref=pe_calc-dc-1367928992-56) derived from
/var/lib/pacemaker/pengine/pe-input-190.bz2
May 07 05:16:32 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 6: monitor
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb_monitor_0 on vsan16
May 07 05:16:32 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 7: monitor
vha-090f26ed-5991-4f40-833e-02e76759dd41:1_monitor_0 on vsan16
May 07 05:16:33 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 5: probe_complete probe_complete on vsan16 - no waiting
May 07 05:16:33 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 8: start vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb_start_0
on vsan16
May 07 05:16:33 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 22: start
vha-090f26ed-5991-4f40-833e-02e76759dd41:1_start_0 on vsan16
May 07 05:16:33 [10846] vsan15       crmd:  warning: status_from_rc:
Action 8 (vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb_start_0) on vsan16
failed (target: 0 vs. rc: 7): Error
May 07 05:16:33 [10846] vsan15       crmd:  warning: update_failcount:
Updating failcount for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan16
after failed start: rc=7 (update=INFINITY, time=1367928993)
May 07 05:16:33 [10846] vsan15       crmd:     info:
abort_transition_graph:     match_graph_event:275 - Triggered transition
abort (complete=0, tag=lrm_rsc_op,
id=vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb_last_failure_0,
magic=0:7;8:10:0:635c50ca-8f47-4a94-a92c-eed5e2566766, cib=0.1344.117) :
Event failed
May 07 05:16:33 [10846] vsan15       crmd:  warning: update_failcount:
Updating failcount for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan16
after failed start: rc=7 (update=INFINITY, time=1367928993)
May 07 05:16:33 [10846] vsan15       crmd:     info: process_graph_event:
    Detected action (10.8)
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb_start_0.15=not running: failed
May 07 05:16:33 [10846] vsan15       crmd:     info:
abort_transition_graph:     te_update_diff:176 - Triggered transition abort
(complete=0, tag=nvpair,
id=status-vsan16-fail-count-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb,
name=fail-count-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb, value=INFINITY,
magic=NA, cib=0.1344.118) : Transient attribute: update
May 07 05:16:33 [10846] vsan15       crmd:     info:
abort_transition_graph:     te_update_diff:176 - Triggered transition abort
(complete=0, tag=nvpair,
id=status-vsan16-last-failure-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb,
name=last-failure-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb,
value=1367928993, magic=NA, cib=0.1344.119) : Transient attribute: update
May 07 05:16:34 [10846] vsan15       crmd:   notice: run_graph:
Transition 10 (Complete=10, Pending=0, Fired=0, Skipped=2, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-190.bz2): Stopped
May 07 05:16:34 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [
input=I_PE_CALC cause=C_FSA_INTERNAL origin=notify_crmd ]
May 07 05:16:34 [10845] vsan15    pengine:     info: unpack_config:
Startup probes: enabled
May 07 05:16:34 [10845] vsan15    pengine:   notice: unpack_config:     On
loss of CCM Quorum: Ignore
May 07 05:16:34 [10845] vsan15    pengine:     info: unpack_config:
Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
May 07 05:16:34 [10845] vsan15    pengine:     info: unpack_domains:
Unpacking domains
May 07 05:16:34 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan15 is online
May 07 05:16:34 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan16 is online
May 07 05:16:34 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan15
to vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
May 07 05:16:34 [10845] vsan15    pengine:  warning: unpack_rsc_op:
Processing failed op start for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
on vsan15: not running (7)
May 07 05:16:34 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-090f26ed-5991-4f40-833e-02e76759dd41 on vsan15
to vha-090f26ed-5991-4f40-833e-02e76759dd41:0
May 07 05:16:34 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan16
to vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
May 07 05:16:34 [10845] vsan15    pengine:  warning: unpack_rsc_op:
Processing failed op start for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
on vsan16: not running (7)
May 07 05:16:34 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-090f26ed-5991-4f40-833e-02e76759dd41 on vsan16
to vha-090f26ed-5991-4f40-833e-02e76759dd41:1
May 07 05:16:34 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb
[vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb]
May 07 05:16:34 [10845] vsan15    pengine:     info: native_print:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
(ocf::heartbeat:vgc-cm-agent.ocf):    Slave vsan16 FAILED
May 07 05:16:34 [10845] vsan15    pengine:     info: short_print:
Stopped: [ vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 ]
May 07 05:16:34 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-090f26ed-5991-4f40-833e-02e76759dd41
[vha-090f26ed-5991-4f40-833e-02e76759dd41]
May 07 05:16:34 [10845] vsan15    pengine:     info: short_print:
Masters: [ vsan15 ]
May 07 05:16:34 [10845] vsan15    pengine:     info: short_print:
Slaves: [ vsan16 ]
May 07 05:16:34 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan15
May 07 05:16:34 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:16:34 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan15
May 07 05:16:34 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:16:34 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan16
May 07 05:16:34 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:16:34 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan16
May 07 05:16:34 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:16:34 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 cannot run anywhere
May 07 05:16:34 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 cannot run anywhere
May 07 05:16:34 [10845] vsan15    pengine:     info: master_color:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb: Promoted 0 instances of a possible
1 to master
May 07 05:16:34 [10845] vsan15    pengine:     info: master_color:
Promoting vha-090f26ed-5991-4f40-833e-02e76759dd41:0 (Master vsan15)
May 07 05:16:34 [10845] vsan15    pengine:     info: master_color:
ms-090f26ed-5991-4f40-833e-02e76759dd41: Promoted 1 instances of a possible
1 to master
May 07 05:16:34 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:1 on vsan16
May 07 05:16:34 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:1 on vsan16
May 07 05:16:34 [10845] vsan15    pengine:   notice: LogActions:
Stop    vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0    (vsan16)
May 07 05:16:34 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1    (Stopped)
May 07 05:16:34 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-090f26ed-5991-4f40-833e-02e76759dd41:0    (Master vsan15)
May 07 05:16:34 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-090f26ed-5991-4f40-833e-02e76759dd41:1    (Slave vsan16)
May 07 05:16:34 [10845] vsan15    pengine:   notice: process_pe_message:
    Calculated Transition 11: /var/lib/pacemaker/pengine/pe-input-191.bz2
May 07 05:16:34 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
May 07 05:16:34 [10846] vsan15       crmd:     info: do_te_invoke:
Processing graph 11 (ref=pe_calc-dc-1367928994-62) derived from
/var/lib/pacemaker/pengine/pe-input-191.bz2
May 07 05:16:34 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 21: monitor
vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_31000 on vsan16
May 07 05:16:34 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 2: stop vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb_stop_0
on vsan16
May 07 05:16:34 [10846] vsan15       crmd:   notice: run_graph:
Transition 11 (Complete=5, Pending=0, Fired=0, Skipped=0, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-191.bz2): Complete
May 07 05:16:34 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
cause=C_FSA_INTERNAL origin=notify_crmd ]
May 07 05:16:36 [10841] vsan15        cib:     info: cib_replace_notify:
    Replaced: 0.1344.128 -> 0.1345.1 from vsan16
May 07 05:16:36 [10846] vsan15       crmd:     info:
abort_transition_graph:     te_update_diff:126 - Triggered transition abort
(complete=1, tag=diff, id=(null), magic=NA, cib=0.1345.1) : Non-status
change
May 07 05:16:36 [10841] vsan15        cib:   notice: cib:diff:     Diff:
--- 0.1344.128
May 07 05:16:36 [10841] vsan15        cib:   notice: cib:diff:     Diff:
+++ 0.1345.1 0d520a23bc38850f68e350ce2bd7f47a
May 07 05:16:36 [10841] vsan15        cib:   notice: cib:diff:     -- <cib
admin_epoch="0" epoch="1344" num_updates="128" />
May 07 05:16:36 [10841] vsan15        cib:   notice: cib:diff:     ++
<rsc_location id="ms_stop_res_on_node"
rsc="ms-090f26ed-5991-4f40-833e-02e76759dd41" >
May 07 05:16:36 [10841] vsan15        cib:   notice: cib:diff:
++         <rule id="ms_stop_res_on_node-rule" score="-INFINITY" >
May 07 05:16:36 [10841] vsan15        cib:   notice: cib:diff:
++           <expression attribute="#uname"
id="ms_stop_res_on_node-expression" operation="eq" value="vsan16" />
May 07 05:16:36 [10841] vsan15        cib:   notice: cib:diff:
++         </rule>
May 07 05:16:36 [10841] vsan15        cib:   notice: cib:diff:     ++
</rsc_location>
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_replace for section 'all'
(origin=vsan16/cibadmin/2, version=0.1345.1): OK (rc=0)
May 07 05:16:36 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC
cause=C_FSA_INTERNAL origin=abort_transition_graph ]
May 07 05:16:36 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_POLICY_ENGINE -> S_ELECTION [ input=I_ELECTION
cause=C_FSA_INTERNAL origin=do_cib_replaced ]
May 07 05:16:36 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC
cause=C_FSA_INTERNAL origin=do_election_check ]
May 07 05:16:36 [10846] vsan15       crmd:     info: do_dc_takeover:
Taking over DC status for this partition
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/93, version=0.1345.2): OK (rc=0)
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_master for section 'all'
(origin=local/crmd/96, version=0.1345.4): OK (rc=0)
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section cib
(origin=local/crmd/97, version=0.1345.5): OK (rc=0)
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/99, version=0.1345.6): OK (rc=0)
May 07 05:16:36 [10846] vsan15       crmd:     info: do_dc_join_offer_all:
    join-4: Waiting on 2 outstanding join acks
May 07 05:16:36 [10846] vsan15       crmd:     info: ais_dispatch_message:
    Membership 14136: quorum retained
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/101, version=0.1345.8): OK (rc=0)
May 07 05:16:36 [10846] vsan15       crmd:     info: crmd_ais_dispatch:
Setting expected votes to 2
May 07 05:16:36 [10846] vsan15       crmd:     info: update_dc:     Set DC
to vsan15 (3.0.7)
May 07 05:16:36 [10846] vsan15       crmd:     info: ais_dispatch_message:
    Membership 14136: quorum retained
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/104, version=0.1345.10): OK (rc=0)
May 07 05:16:36 [10846] vsan15       crmd:     info: crmd_ais_dispatch:
Setting expected votes to 2
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/107, version=0.1345.11): OK (rc=0)
May 07 05:16:36 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED
cause=C_FSA_INTERNAL origin=check_join_state ]
May 07 05:16:36 [10846] vsan15       crmd:     info: do_dc_join_finalize:
    join-4: Syncing the CIB from vsan16 to the rest of the cluster
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_replace:
    Digest matched on replace from vsan16: 2040dc845f64bd8da56d87369ad19883
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_replace:
    Replaced 0.1345.13 with 0.1345.13 from vsan16
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_sync for section 'all'
(origin=vsan16/vsan16/108, version=0.1345.13): OK (rc=0)
May 07 05:16:36 [10846] vsan15       crmd:     info: do_dc_join_ack:
join-4: Updating node state to member for vsan15
May 07 05:16:36 [10846] vsan15       crmd:     info: erase_status_tag:
Deleting xpath: //node_state[@uname='vsan15']/lrm
May 07 05:16:36 [10846] vsan15       crmd:     info: do_dc_join_ack:
join-4: Updating node state to member for vsan16
May 07 05:16:36 [10846] vsan15       crmd:     info: erase_status_tag:
Deleting xpath: //node_state[@uname='vsan16']/lrm
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/109, version=0.1345.14): OK (rc=0)
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/110, version=0.1345.15): OK (rc=0)
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_delete for section
//node_state[@uname='vsan15']/lrm (origin=local/crmd/111,
version=0.1345.16): OK (rc=0)
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_delete for section
//node_state[@uname='vsan16']/lrm (origin=local/crmd/113,
version=0.1345.21): OK (rc=0)
May 07 05:16:36 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED
cause=C_FSA_INTERNAL origin=check_join_state ]
May 07 05:16:36 [10846] vsan15       crmd:     info:
abort_transition_graph:     do_te_invoke:156 - Triggered transition abort
(complete=1) : Peer Cancelled
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/115, version=0.1345.23): OK (rc=0)
May 07 05:16:36 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section cib
(origin=local/crmd/117, version=0.1345.25): OK (rc=0)
May 07 05:16:36 [10844] vsan15      attrd:   notice: attrd_local_callback:
    Sending full refresh (origin=crmd)
May 07 05:16:36 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
master-vha-090f26ed-5991-4f40-833e-02e76759dd41 (4)
May 07 05:16:36 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
fail-count-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb (INFINITY)
May 07 05:16:36 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
last-failure-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb (1367926970)
May 07 05:16:36 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for: probe_complete (true)
May 07 05:16:37 [10845] vsan15    pengine:     info: unpack_config:
Startup probes: enabled
May 07 05:16:37 [10845] vsan15    pengine:   notice: unpack_config:     On
loss of CCM Quorum: Ignore
May 07 05:16:37 [10845] vsan15    pengine:     info: unpack_config:
Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
May 07 05:16:37 [10845] vsan15    pengine:     info: unpack_domains:
Unpacking domains
May 07 05:16:37 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan15 is online
May 07 05:16:37 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan16 is online
May 07 05:16:37 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan15
to vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
May 07 05:16:37 [10845] vsan15    pengine:  warning: unpack_rsc_op:
Processing failed op start for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
on vsan15: not running (7)
May 07 05:16:37 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-090f26ed-5991-4f40-833e-02e76759dd41 on vsan15
to vha-090f26ed-5991-4f40-833e-02e76759dd41:0
May 07 05:16:37 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan16
to vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
May 07 05:16:37 [10845] vsan15    pengine:  warning: unpack_rsc_op:
Processing failed op start for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
on vsan16: not running (7)
May 07 05:16:37 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-090f26ed-5991-4f40-833e-02e76759dd41 on vsan16
to vha-090f26ed-5991-4f40-833e-02e76759dd41:1
May 07 05:16:37 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb
[vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb]
May 07 05:16:37 [10845] vsan15    pengine:     info: short_print:
Stopped: [ vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 ]
May 07 05:16:37 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-090f26ed-5991-4f40-833e-02e76759dd41
[vha-090f26ed-5991-4f40-833e-02e76759dd41]
May 07 05:16:37 [10845] vsan15    pengine:     info: short_print:
Masters: [ vsan15 ]
May 07 05:16:37 [10845] vsan15    pengine:     info: short_print:
Slaves: [ vsan16 ]
May 07 05:16:37 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan15
May 07 05:16:37 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:16:37 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan15
May 07 05:16:37 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:16:37 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan16
May 07 05:16:37 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:16:37 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan16
May 07 05:16:37 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:16:37 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 cannot run anywhere
May 07 05:16:37 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 cannot run anywhere
May 07 05:16:37 [10845] vsan15    pengine:     info: master_color:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb: Promoted 0 instances of a possible
1 to master
May 07 05:16:37 [10845] vsan15    pengine:     info: native_color:
Resource vha-090f26ed-5991-4f40-833e-02e76759dd41:1 cannot run anywhere
May 07 05:16:37 [10845] vsan15    pengine:     info: master_color:
Promoting vha-090f26ed-5991-4f40-833e-02e76759dd41:0 (Master vsan15)
May 07 05:16:37 [10845] vsan15    pengine:     info: master_color:
ms-090f26ed-5991-4f40-833e-02e76759dd41: Promoted 1 instances of a possible
1 to master
May 07 05:16:37 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0    (Stopped)
May 07 05:16:37 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1    (Stopped)
May 07 05:16:37 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-090f26ed-5991-4f40-833e-02e76759dd41:0    (Master vsan15)
May 07 05:16:37 [10845] vsan15    pengine:   notice: LogActions:
Stop    vha-090f26ed-5991-4f40-833e-02e76759dd41:1    (vsan16)
May 07 05:16:37 [10845] vsan15    pengine:   notice: process_pe_message:
    Calculated Transition 12: /var/lib/pacemaker/pengine/pe-input-192.bz2
May 07 05:16:37 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
May 07 05:16:37 [10846] vsan15       crmd:     info: do_te_invoke:
Processing graph 12 (ref=pe_calc-dc-1367928997-72) derived from
/var/lib/pacemaker/pengine/pe-input-192.bz2
May 07 05:16:37 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 19: stop vha-090f26ed-5991-4f40-833e-02e76759dd41_stop_0
on vsan16
May 07 05:16:38 [10846] vsan15       crmd:   notice: run_graph:
Transition 12 (Complete=4, Pending=0, Fired=0, Skipped=0, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-192.bz2): Complete
May 07 05:16:38 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
cause=C_FSA_INTERNAL origin=notify_crmd ]
May 07 05:17:39 [10841] vsan15        cib:     info: cib_replace_notify:
    Replaced: 0.1345.36 -> 0.1346.1 from vsan16
May 07 05:17:39 [10841] vsan15        cib:   notice: cib:diff:     Diff:
--- 0.1345.36
May 07 05:17:39 [10841] vsan15        cib:   notice: cib:diff:     Diff:
+++ 0.1346.1 d5dcc8f04f6661a4137788e300500a61
May 07 05:17:39 [10841] vsan15        cib:   notice: cib:diff:     --
<rsc_location id="ms_stop_res_on_node"
rsc="ms-090f26ed-5991-4f40-833e-02e76759dd41" >
May 07 05:17:39 [10841] vsan15        cib:   notice: cib:diff:
--         <rule id="ms_stop_res_on_node-rule" score="-INFINITY" >
May 07 05:17:39 [10841] vsan15        cib:   notice: cib:diff:
--           <expression attribute="#uname"
id="ms_stop_res_on_node-expression" operation="eq" value="vsan16" />
May 07 05:17:39 [10846] vsan15       crmd:     info:
abort_transition_graph:     te_update_diff:126 - Triggered transition abort
(complete=1, tag=diff, id=(null), magic=NA, cib=0.1346.1) : Non-status
change
May 07 05:17:39 [10841] vsan15        cib:   notice: cib:diff:
--         </rule>
May 07 05:17:39 [10841] vsan15        cib:   notice: cib:diff:     --
</rsc_location>
May 07 05:17:39 [10841] vsan15        cib:   notice: cib:diff:     ++ <cib
epoch="1346" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2"
crm_feature_set="3.0.7" update-origin="vsan16" update-client="cibadmin"
cib-last-written="Tue May  7 05:16:36 2013" have-quorum="1"
dc-uuid="vsan15" />
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_replace for section 'all'
(origin=vsan16/cibadmin/2, version=0.1346.1): OK (rc=0)
May 07 05:17:39 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC
cause=C_FSA_INTERNAL origin=abort_transition_graph ]
May 07 05:17:39 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_POLICY_ENGINE -> S_ELECTION [ input=I_ELECTION
cause=C_FSA_INTERNAL origin=do_cib_replaced ]
May 07 05:17:39 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC
cause=C_FSA_INTERNAL origin=do_election_check ]
May 07 05:17:39 [10846] vsan15       crmd:     info: do_dc_takeover:
Taking over DC status for this partition
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/120, version=0.1346.2): OK (rc=0)
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_master for section 'all'
(origin=local/crmd/123, version=0.1346.4): OK (rc=0)
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section cib
(origin=local/crmd/124, version=0.1346.5): OK (rc=0)
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/126, version=0.1346.6): OK (rc=0)
May 07 05:17:39 [10846] vsan15       crmd:     info: do_dc_join_offer_all:
    join-5: Waiting on 2 outstanding join acks
May 07 05:17:39 [10846] vsan15       crmd:     info: ais_dispatch_message:
    Membership 14136: quorum retained
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/128, version=0.1346.8): OK (rc=0)
May 07 05:17:39 [10846] vsan15       crmd:     info: crmd_ais_dispatch:
Setting expected votes to 2
May 07 05:17:39 [10846] vsan15       crmd:     info: update_dc:     Set DC
to vsan15 (3.0.7)
May 07 05:17:39 [10846] vsan15       crmd:     info: ais_dispatch_message:
    Membership 14136: quorum retained
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/131, version=0.1346.10): OK (rc=0)
May 07 05:17:39 [10846] vsan15       crmd:     info: crmd_ais_dispatch:
Setting expected votes to 2
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section crm_config
(origin=local/crmd/134, version=0.1346.11): OK (rc=0)
May 07 05:17:39 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED
cause=C_FSA_INTERNAL origin=check_join_state ]
May 07 05:17:39 [10846] vsan15       crmd:     info: do_dc_join_finalize:
    join-5: Syncing the CIB from vsan16 to the rest of the cluster
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_replace:
    Digest matched on replace from vsan16: e06eba70e44824f664002c5d2e5a6dbd
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_replace:
    Replaced 0.1346.13 with 0.1346.13 from vsan16
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_sync for section 'all'
(origin=vsan16/vsan16/135, version=0.1346.13): OK (rc=0)
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/136, version=0.1346.14): OK (rc=0)
May 07 05:17:39 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/137, version=0.1346.15): OK (rc=0)
May 07 05:17:40 [10846] vsan15       crmd:     info:
services_os_action_execute:     Managed vgc-cm-agent.ocf_meta-data_0
process 14517 exited with rc=0
May 07 05:17:40 [10846] vsan15       crmd:     info: do_dc_join_ack:
join-5: Updating node state to member for vsan16
May 07 05:17:40 [10846] vsan15       crmd:     info: erase_status_tag:
Deleting xpath: //node_state[@uname='vsan16']/lrm
May 07 05:17:40 [10846] vsan15       crmd:     info: do_dc_join_ack:
join-5: Updating node state to member for vsan15
May 07 05:17:40 [10846] vsan15       crmd:     info: erase_status_tag:
Deleting xpath: //node_state[@uname='vsan15']/lrm
May 07 05:17:40 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_delete for section
//node_state[@uname='vsan16']/lrm (origin=local/crmd/138,
version=0.1346.16): OK (rc=0)
May 07 05:17:40 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_delete for section
//node_state[@uname='vsan15']/lrm (origin=local/crmd/140,
version=0.1346.21): OK (rc=0)
May 07 05:17:40 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED
cause=C_FSA_INTERNAL origin=check_join_state ]
May 07 05:17:40 [10846] vsan15       crmd:     info:
abort_transition_graph:     do_te_invoke:156 - Triggered transition abort
(complete=1) : Peer Cancelled
May 07 05:17:40 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/142, version=0.1346.23): OK (rc=0)
May 07 05:17:40 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section cib
(origin=local/crmd/144, version=0.1346.25): OK (rc=0)
May 07 05:17:40 [10844] vsan15      attrd:   notice: attrd_local_callback:
    Sending full refresh (origin=crmd)
May 07 05:17:40 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
master-vha-090f26ed-5991-4f40-833e-02e76759dd41 (4)
May 07 05:17:40 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
fail-count-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb (INFINITY)
May 07 05:17:40 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
last-failure-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb (1367926970)
May 07 05:17:40 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for: probe_complete (true)
May 07 05:17:41 [10845] vsan15    pengine:     info: unpack_config:
Startup probes: enabled
May 07 05:17:41 [10845] vsan15    pengine:   notice: unpack_config:     On
loss of CCM Quorum: Ignore
May 07 05:17:41 [10845] vsan15    pengine:     info: unpack_config:
Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
May 07 05:17:41 [10845] vsan15    pengine:     info: unpack_domains:
Unpacking domains
May 07 05:17:41 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan15 is online
May 07 05:17:41 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan16 is online
May 07 05:17:41 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan15
to vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
May 07 05:17:41 [10845] vsan15    pengine:  warning: unpack_rsc_op:
Processing failed op start for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
on vsan15: not running (7)
May 07 05:17:41 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-090f26ed-5991-4f40-833e-02e76759dd41 on vsan15
to vha-090f26ed-5991-4f40-833e-02e76759dd41:0
May 07 05:17:41 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan16
to vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
May 07 05:17:41 [10845] vsan15    pengine:  warning: unpack_rsc_op:
Processing failed op start for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
on vsan16: not running (7)
May 07 05:17:41 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-090f26ed-5991-4f40-833e-02e76759dd41 on vsan16
to vha-090f26ed-5991-4f40-833e-02e76759dd41:1
May 07 05:17:41 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb
[vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb]
May 07 05:17:41 [10845] vsan15    pengine:     info: short_print:
Stopped: [ vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 ]
May 07 05:17:41 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-090f26ed-5991-4f40-833e-02e76759dd41
[vha-090f26ed-5991-4f40-833e-02e76759dd41]
May 07 05:17:41 [10845] vsan15    pengine:     info: short_print:
Masters: [ vsan15 ]
May 07 05:17:41 [10845] vsan15    pengine:     info: short_print:
Stopped: [ vha-090f26ed-5991-4f40-833e-02e76759dd41:1 ]
May 07 05:17:41 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan15
May 07 05:17:41 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:17:41 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan15
May 07 05:17:41 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:17:41 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan16
May 07 05:17:41 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:17:41 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan16
May 07 05:17:41 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:17:41 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 cannot run anywhere
May 07 05:17:41 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 cannot run anywhere
May 07 05:17:41 [10845] vsan15    pengine:     info: master_color:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb: Promoted 0 instances of a possible
1 to master
May 07 05:17:41 [10845] vsan15    pengine:     info: master_color:
Promoting vha-090f26ed-5991-4f40-833e-02e76759dd41:0 (Master vsan15)
May 07 05:17:41 [10845] vsan15    pengine:     info: master_color:
ms-090f26ed-5991-4f40-833e-02e76759dd41: Promoted 1 instances of a possible
1 to master
May 07 05:17:41 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:1 on vsan16
May 07 05:17:41 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:1 on vsan16
May 07 05:17:41 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0    (Stopped)
May 07 05:17:41 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1    (Stopped)
May 07 05:17:41 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-090f26ed-5991-4f40-833e-02e76759dd41:0    (Master vsan15)
May 07 05:17:41 [10845] vsan15    pengine:   notice: LogActions:
Start   vha-090f26ed-5991-4f40-833e-02e76759dd41:1    (vsan16)
May 07 05:17:41 [10845] vsan15    pengine:   notice: process_pe_message:
    Calculated Transition 13: /var/lib/pacemaker/pengine/pe-input-193.bz2
May 07 05:17:41 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
May 07 05:17:41 [10846] vsan15       crmd:     info: do_te_invoke:
Processing graph 13 (ref=pe_calc-dc-1367929061-81) derived from
/var/lib/pacemaker/pengine/pe-input-193.bz2
May 07 05:17:41 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 18: start
vha-090f26ed-5991-4f40-833e-02e76759dd41_start_0 on vsan16
May 07 05:17:41 [10846] vsan15       crmd:     info:
abort_transition_graph:     te_update_diff:176 - Triggered transition abort
(complete=0, tag=nvpair,
id=status-vsan16-master-vha-090f26ed-5991-4f40-833e-02e76759dd41,
name=master-vha-090f26ed-5991-4f40-833e-02e76759dd41, value=3, magic=NA,
cib=0.1346.36) : Transient attribute: update
May 07 05:17:41 [10846] vsan15       crmd:   notice: run_graph:
Transition 13 (Complete=3, Pending=0, Fired=0, Skipped=1, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-193.bz2): Stopped
May 07 05:17:41 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [
input=I_PE_CALC cause=C_FSA_INTERNAL origin=notify_crmd ]
May 07 05:17:41 [10845] vsan15    pengine:     info: unpack_config:
Startup probes: enabled
May 07 05:17:41 [10845] vsan15    pengine:   notice: unpack_config:     On
loss of CCM Quorum: Ignore
May 07 05:17:41 [10845] vsan15    pengine:     info: unpack_config:
Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
May 07 05:17:41 [10845] vsan15    pengine:     info: unpack_domains:
Unpacking domains
May 07 05:17:41 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan15 is online
May 07 05:17:41 [10845] vsan15    pengine:     info:
determine_online_status:     Node vsan16 is online
May 07 05:17:41 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan15
to vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
May 07 05:17:41 [10845] vsan15    pengine:  warning: unpack_rsc_op:
Processing failed op start for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
on vsan15: not running (7)
May 07 05:17:41 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-090f26ed-5991-4f40-833e-02e76759dd41 on vsan15
to vha-090f26ed-5991-4f40-833e-02e76759dd41:0
May 07 05:17:41 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb on vsan16
to vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
May 07 05:17:41 [10845] vsan15    pengine:  warning: unpack_rsc_op:
Processing failed op start for vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
on vsan16: not running (7)
May 07 05:17:41 [10845] vsan15    pengine:     info: find_anonymous_clone:
    Internally renamed vha-090f26ed-5991-4f40-833e-02e76759dd41 on vsan16
to vha-090f26ed-5991-4f40-833e-02e76759dd41:1
May 07 05:17:41 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb
[vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb]
May 07 05:17:41 [10845] vsan15    pengine:     info: short_print:
Stopped: [ vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 ]
May 07 05:17:41 [10845] vsan15    pengine:     info: clone_print:
 Master/Slave Set: ms-090f26ed-5991-4f40-833e-02e76759dd41
[vha-090f26ed-5991-4f40-833e-02e76759dd41]
May 07 05:17:41 [10845] vsan15    pengine:     info: short_print:
Masters: [ vsan15 ]
May 07 05:17:41 [10845] vsan15    pengine:     info: short_print:
Slaves: [ vsan16 ]
May 07 05:17:41 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan15
May 07 05:17:41 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:17:41 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan15
May 07 05:17:41 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan15 after 1000000
failures (max=1000000)
May 07 05:17:41 [10845] vsan15    pengine:     info: get_failcount:
vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 has failed INFINITY times on
vsan16
May 07 05:17:41 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:17:41 [10845] vsan15    pengine:     info: get_failcount:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb has failed INFINITY times on vsan16
May 07 05:17:41 [10845] vsan15    pengine:  warning:
common_apply_stickiness:     Forcing
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb away from vsan16 after 1000000
failures (max=1000000)
May 07 05:17:41 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0 cannot run anywhere
May 07 05:17:41 [10845] vsan15    pengine:     info: native_color:
Resource vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1 cannot run anywhere
May 07 05:17:41 [10845] vsan15    pengine:     info: master_color:
ms-bcd94724-3ec0-4a8d-8951-9d27be3a6acb: Promoted 0 instances of a possible
1 to master
May 07 05:17:41 [10845] vsan15    pengine:     info: master_color:
Promoting vha-090f26ed-5991-4f40-833e-02e76759dd41:0 (Master vsan15)
May 07 05:17:41 [10845] vsan15    pengine:     info: master_color:
ms-090f26ed-5991-4f40-833e-02e76759dd41: Promoted 1 instances of a possible
1 to master
May 07 05:17:41 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:1 on vsan16
May 07 05:17:41 [10845] vsan15    pengine:     info: RecurringOp:
 Start recurring monitor (31s) for
vha-090f26ed-5991-4f40-833e-02e76759dd41:1 on vsan16
May 07 05:17:41 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:0    (Stopped)
May 07 05:17:41 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb:1    (Stopped)
May 07 05:17:41 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-090f26ed-5991-4f40-833e-02e76759dd41:0    (Master vsan15)
May 07 05:17:41 [10845] vsan15    pengine:     info: LogActions:
Leave   vha-090f26ed-5991-4f40-833e-02e76759dd41:1    (Slave vsan16)
May 07 05:17:41 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
May 07 05:17:41 [10845] vsan15    pengine:   notice: process_pe_message:
    Calculated Transition 14: /var/lib/pacemaker/pengine/pe-input-194.bz2
May 07 05:17:41 [10846] vsan15       crmd:     info: do_te_invoke:
Processing graph 14 (ref=pe_calc-dc-1367929061-83) derived from
/var/lib/pacemaker/pengine/pe-input-194.bz2
May 07 05:17:41 [10846] vsan15       crmd:     info: te_rsc_command:
Initiating action 20: monitor
vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_31000 on vsan16
May 07 05:17:42 [10846] vsan15       crmd:   notice: run_graph:
Transition 14 (Complete=1, Pending=0, Fired=0, Skipped=0, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-194.bz2): Complete
May 07 05:17:42 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
cause=C_FSA_INTERNAL origin=notify_crmd ]
May 07 05:18:25 [10846] vsan15       crmd:    error:
node_list_update_callback:     CIB Update 72 failed: Timer expired
May 07 05:18:25 [10846] vsan15       crmd:  warning:
node_list_update_callback:     update:failed: No data to dump as XML
May 07 05:18:25 [10846] vsan15       crmd:    error: do_log:     FSA: Input
I_ERROR from node_list_update_callback() received in state S_IDLE
May 07 05:18:25 [10846] vsan15       crmd:   notice: do_state_transition:
    State transition S_IDLE -> S_RECOVERY [ input=I_ERROR
cause=C_FSA_INTERNAL origin=node_list_update_callback ]
May 07 05:18:25 [10846] vsan15       crmd:    error: do_recover:     Action
A_RECOVER (0000000001000000) not supported
May 07 05:18:25 [10846] vsan15       crmd:  warning: do_election_vote:
Not voting in election, we're in state S_RECOVERY
May 07 05:18:25 [10846] vsan15       crmd:     info: do_dc_release:     DC
role released
May 07 05:18:25 [10846] vsan15       crmd:     info: pe_ipc_destroy:
Connection to the Policy Engine released
May 07 05:18:25 [10846] vsan15       crmd:     info: do_te_control:
Transitioner is now inactive
May 07 05:18:25 [10846] vsan15       crmd:    error: do_log:     FSA: Input
I_TERMINATE from do_recover() received in state S_RECOVERY
May 07 05:18:25 [10846] vsan15       crmd:     info: do_state_transition:
    State transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE
cause=C_FSA_INTERNAL origin=do_recover ]
May 07 05:18:25 [10846] vsan15       crmd:     info: do_shutdown:
Disconnecting STONITH...
May 07 05:18:25 [10846] vsan15       crmd:     info:
tengine_stonith_connection_destroy:     Fencing daemon disconnected
May 07 05:18:25 [10843] vsan15       lrmd:     info:
cancel_recurring_action:     Cancelling operation
vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_30000
May 07 05:18:25 [10846] vsan15       crmd:    error: verify_stopped:
Resource vha-090f26ed-5991-4f40-833e-02e76759dd41 was active at shutdown.
You may ignore this error if it is unmanaged.
May 07 05:18:25 [10846] vsan15       crmd:     info: lrmd_api_disconnect:
    Disconnecting from lrmd service
May 07 05:18:25 [10843] vsan15       lrmd:     info: lrmd_ipc_destroy:
LRMD client disconnecting 0x19129f0 - name: crmd id:
1a9d196e-7822-4895-b066-7e2702a75f45
May 07 05:18:25 [10846] vsan15       crmd:     info:
lrmd_connection_destroy:     connection destroyed
May 07 05:18:25 [10846] vsan15       crmd:     info:
lrm_connection_destroy:     LRM Connection disconnected
May 07 05:18:25 [10846] vsan15       crmd:     info: do_lrm_control:
Disconnected from the LRM
May 07 05:18:25 [10846] vsan15       crmd:     info:
crm_cluster_disconnect:     Disconnecting from cluster infrastructure:
classic openais (with plugin)
May 07 05:18:25 [10846] vsan15       crmd:   notice:
terminate_cs_connection:     Disconnecting from Corosync
May 07 05:18:25 [10846] vsan15       crmd:     info:
crm_cluster_disconnect:     Disconnected from classic openais (with plugin)
May 07 05:18:25 [10846] vsan15       crmd:     info: do_ha_control:
Disconnected from the cluster
May 07 05:18:25 [10846] vsan15       crmd:     info: do_cib_control:
Disconnecting CIB
May 07 05:18:25 [10841] vsan15        cib:     info: cib_process_readwrite:
    We are now in R/O mode
May 07 05:18:25 [10841] vsan15        cib:  warning: qb_ipcs_event_sendv:
    new_event_notification (10841-10846-12): Broken pipe (32)
May 07 05:18:25 [10841] vsan15        cib:     info: crm_ipcs_send:
Event 722 failed, size=162, to=0x1721890[10846], queue=1, retries=0,
rc=-32: <cib-reply t="cib" cib_op="cib_slave" cib_callid="148"
cib_clientid="0003ee04-6a88-496a-b0fd-1d0e24c1510b" cib_callopt="
May 07 05:18:25 [10841] vsan15        cib:  warning: do_local_notify:
A-Sync reply to crmd failed: No message of desired type
May 07 05:18:25 [10846] vsan15       crmd:     info:
crmd_cib_connection_destroy:     Connection to the CIB terminated...
May 07 05:18:25 [10846] vsan15       crmd:     info: qb_ipcs_us_withdraw:
    withdrawing server sockets
May 07 05:18:25 [10846] vsan15       crmd:     info: do_exit:
Performing A_EXIT_0 - gracefully exiting the CRMd
May 07 05:18:25 [10846] vsan15       crmd:    error: do_exit:     Could not
recover from internal error
May 07 05:18:25 [10846] vsan15       crmd:     info: do_exit:     [crmd]
stopped (2)
May 07 05:18:25 [10846] vsan15       crmd:     info: crmd_exit:
Dropping I_PENDING: [ state=S_TERMINATE cause=C_FSA_INTERNAL
origin=do_election_vote ]
May 07 05:18:25 [10846] vsan15       crmd:     info: crmd_exit:
Dropping I_RELEASE_SUCCESS: [ state=S_TERMINATE cause=C_FSA_INTERNAL
origin=do_dc_release ]
May 07 05:18:25 [10846] vsan15       crmd:     info: crmd_exit:
Dropping I_TERMINATE: [ state=S_TERMINATE cause=C_FSA_INTERNAL
origin=do_stop ]
May 07 05:18:25 [10846] vsan15       crmd:     info: lrmd_api_disconnect:
    Disconnecting from lrmd service
May 07 05:18:25 [10846] vsan15       crmd:     info: crm_xml_cleanup:
Cleaning up memory from libxml2
May 07 05:18:25 corosync [pcmk  ] info: pcmk_ipc_exit: Client crmd
(conn=0x15f4560, async-conn=0x15f4560) left
May 07 05:18:25 [10835] vsan15 pacemakerd:    error: pcmk_child_exit:
Child process crmd exited (pid=10846, rc=2)
May 07 05:18:25 [10842] vsan15 stonith-ng:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan15[1682182316] - unknown is now (null)
May 07 05:18:25 [10835] vsan15 pacemakerd:   notice: pcmk_process_exit:
Respawning failed child process: crmd
May 07 05:18:25 [10841] vsan15        cib:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan15[1682182316] - unknown is now (null)
May 07 05:18:25 [10835] vsan15 pacemakerd:     info: start_child:
Forked child 14640 for process crmd
May 07 05:18:25 [10842] vsan15 stonith-ng:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan15[1682182316] - unknown is now (null)
May 07 05:18:25 [10841] vsan15        cib:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan15[1682182316] - unknown is now (null)
May 07 05:18:25 corosync [pcmk  ] WARN: route_ais_message: Sending message
to local.crmd failed: ipc delivery failed (rc=-2)
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_log_init:
Cannot change active directory to /var/lib/pacemaker/cores/hacluster:
Permission denied (13)
May 07 05:18:25 [14640] vsan15       crmd:   notice: main:     CRM Git
Version: 394e906
May 07 05:18:25 [14640] vsan15       crmd:     info: get_cluster_type:
Cluster type is: 'openais'
May 07 05:18:25 [14640] vsan15       crmd:     info: do_cib_control:
CIB connection established
May 07 05:18:25 [14640] vsan15       crmd:   notice: crm_cluster_connect:
    Connecting to cluster infrastructure: classic openais (with plugin)
May 07 05:18:25 [14640] vsan15       crmd:     info:
init_cs_connection_classic:     Creating connection to our Corosync plugin
May 07 05:18:25 [14640] vsan15       crmd:     info:
init_cs_connection_classic:     AIS connection established
May 07 05:18:25 corosync [pcmk  ] info: pcmk_ipc: Recorded connection
0x15f4560 for crmd/0
May 07 05:18:25 corosync [pcmk  ] info: pcmk_ipc: Sending membership update
14136 to crmd
May 07 05:18:25 [14640] vsan15       crmd:     info: get_ais_nodeid:
Server details: id=1682182316 uname=vsan15 cname=pcmk
May 07 05:18:25 [14640] vsan15       crmd:     info:
init_cs_connection_once:     Connection to 'classic openais (with plugin)':
established
May 07 05:18:25 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_modify for section nodes
(origin=local/crmd/3, version=0.1346.40): OK (rc=0)
May 07 05:18:25 [14640] vsan15       crmd:     info: do_ha_control:
Connected to the cluster
May 07 05:18:25 [14640] vsan15       crmd:     info: lrmd_api_connect:
Connecting to lrmd
May 07 05:18:25 [10843] vsan15       lrmd:     info: lrmd_ipc_accept:
Accepting client connection: 0x191a860 pid=14640 for uid=495 gid=0
May 07 05:18:25 [14640] vsan15       crmd:     info: do_started:
Delaying start, no membership data (0000000000100000)
May 07 05:18:25 [14640] vsan15       crmd:     info: do_started:
Delaying start, no membership data (0000000000100000)
May 07 05:18:25 [14640] vsan15       crmd:   notice: ais_dispatch_message:
    Membership 14136: quorum acquired
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_get_peer:     Node
vsan15 now has id: 1682182316
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_get_peer:     Node
1682182316 is now known as vsan15
May 07 05:18:25 [14640] vsan15       crmd:     info: peer_update_callback:
    vsan15 is now (null)
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_get_peer:     Node
1682182316 has uuid vsan15
May 07 05:18:25 [14640] vsan15       crmd:   notice: crm_update_peer_state:
    crm_update_ais_node: Node vsan15[1682182316] - state is now member
May 07 05:18:25 [14640] vsan15       crmd:     info: peer_update_callback:
    vsan15 is now member (was (null))
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_update_peer:
crm_update_ais_node: Node vsan15: id=1682182316 state=member addr=r(0)
ip(172.16.68.100)  (new) votes=1 (new) born=14128 seen=14136
proc=00000000000000000000000000000000
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_get_peer:     Node
vsan16 now has id: 1698959532
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_get_peer:     Node
1698959532 is now known as vsan16
May 07 05:18:25 [14640] vsan15       crmd:     info: peer_update_callback:
    vsan16 is now (null)
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_get_peer:     Node
1698959532 has uuid vsan16
May 07 05:18:25 [14640] vsan15       crmd:   notice: crm_update_peer_state:
    crm_update_ais_node: Node vsan16[1698959532] - state is now member
May 07 05:18:25 [14640] vsan15       crmd:     info: peer_update_callback:
    vsan16 is now member (was (null))
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_update_peer:
crm_update_ais_node: Node vsan16: id=1698959532 state=member addr=r(0)
ip(172.16.68.101)  (new) votes=1 (new) born=14136 seen=14136
proc=00000000000000000000000000000000
May 07 05:18:25 [14640] vsan15       crmd:     info: ais_dispatch_message:
    Membership 14136: quorum retained
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan15[1682182316] - unknown is now (null)
May 07 05:18:25 [14640] vsan15       crmd:     info: peer_update_callback:
    Client vsan15/peer now has status [online] (DC=<null>)
May 07 05:18:25 [14640] vsan15       crmd:     info: crm_update_peer_proc:
    pcmk_mcp_dispatch: Node vsan16[1698959532] - unknown is now (null)
May 07 05:18:25 [14640] vsan15       crmd:     info: peer_update_callback:
    Client vsan16/peer now has status [online] (DC=<null>)
May 07 05:18:25 [14640] vsan15       crmd:     info: qb_ipcs_us_publish:
    server name: crmd
May 07 05:18:25 [14640] vsan15       crmd:   notice: do_started:     The
local CRM is operational
May 07 05:18:25 [14640] vsan15       crmd:     info: do_state_transition:
    State transition S_STARTING -> S_PENDING [ input=I_PENDING
cause=C_FSA_INTERNAL origin=do_started ]
May 07 05:18:27 [10842] vsan15 stonith-ng:     info: stonith_command:
Processed register from crmd.14640: OK (0)
May 07 05:18:27 [10842] vsan15 stonith-ng:     info: stonith_command:
Processed st_notify from crmd.14640: OK (0)
May 07 05:18:27 [10842] vsan15 stonith-ng:     info: stonith_command:
Processed st_notify from crmd.14640: OK (0)
May 07 05:18:46 [14640] vsan15       crmd:     info: crm_timer_popped:
Election Trigger (I_DC_TIMEOUT) just popped (20000ms)
May 07 05:18:46 [14640] vsan15       crmd:  warning: do_log:     FSA: Input
I_DC_TIMEOUT from crm_timer_popped() received in state S_PENDING
May 07 05:18:46 [14640] vsan15       crmd:     info: do_state_transition:
    State transition S_PENDING -> S_ELECTION [ input=I_DC_TIMEOUT
cause=C_TIMER_POPPED origin=crm_timer_popped ]
May 07 05:18:46 [14640] vsan15       crmd:     info:
do_election_count_vote:     Election 3 (owner: vsan16) lost: vote from
vsan16 (Uptime)
May 07 05:18:46 [14640] vsan15       crmd:   notice: do_state_transition:
    State transition S_ELECTION -> S_PENDING [ input=I_PENDING
cause=C_FSA_INTERNAL origin=do_election_count_vote ]
May 07 05:18:46 [14640] vsan15       crmd:     info: do_dc_release:     DC
role released
May 07 05:18:46 [14640] vsan15       crmd:     info: do_te_control:
Transitioner is now inactive
May 07 05:18:46 [14640] vsan15       crmd:     info: update_dc:     Set DC
to vsan16 (3.0.7)
May 07 05:18:46 [10841] vsan15        cib:     info: cib_process_request:
    Operation complete: op cib_sync for section 'all'
(origin=vsan16/crmd/35, version=0.1346.46): OK (rc=0)
May 07 05:18:46 [14640] vsan15       crmd:     info: erase_status_tag:
Deleting xpath: //node_state[@uname='vsan15']/transient_attributes
May 07 05:18:46 [14640] vsan15       crmd:     info: update_attrd:
Connecting to attrd... 5 retries remaining
May 07 05:18:46 [14640] vsan15       crmd:   notice: do_state_transition:
    State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC
cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
May 07 05:18:46 [10844] vsan15      attrd:   notice: attrd_local_callback:
    Sending full refresh (origin=crmd)
May 07 05:18:46 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
master-vha-090f26ed-5991-4f40-833e-02e76759dd41 (4)
May 07 05:18:46 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
fail-count-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb (INFINITY)
May 07 05:18:46 [10841] vsan15        cib:     info: cib_process_replace:
    Digest matched on replace from vsan16: 8eaadcc9996565180563eac16495c423
May 07 05:18:46 [10841] vsan15        cib:     info: cib_process_replace:
    Replaced 0.1346.46 with 0.1346.46 from vsan16
May 07 05:18:46 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
last-failure-vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb (1367926970)
May 07 05:18:46 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for: probe_complete (true)
May 07 05:18:46 [10844] vsan15      attrd:  warning: attrd_cib_callback:
    Update 144 for master-vha-090f26ed-5991-4f40-833e-02e76759dd41=4
failed: No such device or address
May 07 05:18:49 [14640] vsan15       crmd:     info:
services_os_action_execute:     Managed vgc-cm-agent.ocf_meta-data_0
process 14679 exited with rc=0
May 07 05:18:49 [14640] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-bcd94724-3ec0-4a8d-8951-9d27be3a6acb_monitor_0 (call=55,
rc=7, cib-update=10, confirmed=true) not running
May 07 05:18:49 [14640] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_0 (call=57,
rc=7, cib-update=11, confirmed=true) not running
May 07 05:18:50 [14640] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_start_0 (call=61,
rc=0, cib-update=12, confirmed=true) ok
May 07 05:18:50 [14640] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_31000
(call=64, rc=8, cib-update=13, confirmed=false) master
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_ais_dispatch:
    Update relayed from vsan16
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
fail-count-vha-090f26ed-5991-4f40-833e-02e76759dd41 (1)
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_perform_update:
    Sent update 165: fail-count-vha-090f26ed-5991-4f40-833e-02e76759dd41=1
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_ais_dispatch:
    Update relayed from vsan16
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
last-failure-vha-090f26ed-5991-4f40-833e-02e76759dd41 (1367929130)
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_perform_update:
    Sent update 168:
last-failure-vha-090f26ed-5991-4f40-833e-02e76759dd41=1367929130
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_ais_dispatch:
    Update relayed from vsan16
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
fail-count-vha-090f26ed-5991-4f40-833e-02e76759dd41 (2)
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_perform_update:
    Sent update 171: fail-count-vha-090f26ed-5991-4f40-833e-02e76759dd41=2
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_ais_dispatch:
    Update relayed from vsan16
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
last-failure-vha-090f26ed-5991-4f40-833e-02e76759dd41 (1367929130)
May 07 05:18:50 [10844] vsan15      attrd:   notice: attrd_perform_update:
    Sent update 174:
last-failure-vha-090f26ed-5991-4f40-833e-02e76759dd41=1367929130
May 07 05:18:50 [10843] vsan15       lrmd:     info:
cancel_recurring_action:     Cancelling operation
vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_31000
May 07 05:18:50 [14640] vsan15       crmd:     info: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_31000
(call=64, status=1, cib-update=0, confirmed=false) Cancelled
May 07 05:18:50 [14640] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_demote_0 (call=68,
rc=0, cib-update=14, confirmed=true) ok
May 07 05:18:51 [10844] vsan15      attrd:   notice: attrd_trigger_update:
    Sending flush op to all hosts for:
master-vha-090f26ed-5991-4f40-833e-02e76759dd41 (<null>)
May 07 05:18:51 [10844] vsan15      attrd:   notice: attrd_perform_update:
    Sent delete 178: node=vsan15,
attr=master-vha-090f26ed-5991-4f40-833e-02e76759dd41, id=<n/a>, set=(null),
section=status
May 07 05:18:51 [10844] vsan15      attrd:   notice: attrd_perform_update:
    Sent delete 180: node=vsan15,
attr=master-vha-090f26ed-5991-4f40-833e-02e76759dd41, id=<n/a>, set=(null),
section=status
May 07 05:18:51 [14640] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_stop_0 (call=72,
rc=0, cib-update=15, confirmed=true) ok
May 07 05:18:52 [14640] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_start_0 (call=75,
rc=0, cib-update=16, confirmed=true) ok
May 07 05:18:52 [14640] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_31000
(call=78, rc=0, cib-update=17, confirmed=false) ok
May 07 10:58:21 [14640] vsan15       crmd:   notice: process_lrm_event:
LRM operation vha-090f26ed-5991-4f40-833e-02e76759dd41_monitor_31000
(call=78, rc=7, cib-update=18, confirmed=false) not running
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20130508/554f4ca6/attachment-0001.html>


More information about the Pacemaker mailing list