Jan 10 18:23:25 dl380g5d heartbeat: [4277]: WARN: Logging daemon is disabled --enabling logging daemon is recommended
Jan 10 18:23:25 dl380g5d heartbeat: [4277]: info: **************************
Jan 10 18:23:25 dl380g5d heartbeat: [4277]: info: Configuration validated. Starting heartbeat 3.0.5
Jan 10 18:23:25 dl380g5d heartbeat: [4278]: info: heartbeat: version 3.0.5
Jan 10 18:23:25 dl380g5d heartbeat: [4278]: info: Heartbeat generation: 1337914273
Jan 10 18:23:25 dl380g5d heartbeat: [4278]: info: glib: UDP Broadcast heartbeat started on port 694 (694) interface eth2
Jan 10 18:23:25 dl380g5d heartbeat: [4278]: info: glib: UDP Broadcast heartbeat closed on port 694 interface eth2 - Status: 1
Jan 10 18:23:25 dl380g5d heartbeat: [4278]: notice: Using watchdog device: /dev/watchdog
Jan 10 18:23:25 dl380g5d heartbeat: [4278]: info: Local status now set to: 'up'
Jan 10 18:23:25 dl380g5d heartbeat: [4278]: info: Link dl380g5d:eth2 up.
Jan 10 18:23:25 dl380g5d heartbeat: [4278]: info: Link dl380g5c:eth2 up.
Jan 10 18:23:25 dl380g5d heartbeat: [4278]: debug: get_delnodelist: delnodelist= 
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: Status update for node dl380g5c: status active
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: Comm_now_up(): updating status to active
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: Local status now set to: 'active'
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: Starting child client "/usr/lib64/heartbeat/ccm" (495,489)
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: Starting child client "/usr/lib64/heartbeat/cib" (495,489)
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: Starting child client "/usr/lib64/heartbeat/lrmd -r" (0,0)
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: Starting child client "/usr/lib64/heartbeat/stonithd" (0,0)
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: Starting child client "/usr/lib64/heartbeat/attrd" (495,489)
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: Starting child client "/usr/lib64/heartbeat/crmd" (495,489)
Jan 10 18:23:26 dl380g5d heartbeat: [4287]: info: Starting "/usr/lib64/heartbeat/cib" as uid 495  gid 489 (pid 4287)
Jan 10 18:23:26 dl380g5d heartbeat: [4289]: info: Starting "/usr/lib64/heartbeat/stonithd" as uid 0  gid 0 (pid 4289)
Jan 10 18:23:26 dl380g5d heartbeat: [4286]: info: Starting "/usr/lib64/heartbeat/ccm" as uid 495  gid 489 (pid 4286)
Jan 10 18:23:26 dl380g5d heartbeat: [4288]: info: Starting "/usr/lib64/heartbeat/lrmd -r" as uid 0  gid 0 (pid 4288)
Jan 10 18:23:26 dl380g5d heartbeat: [4291]: info: Starting "/usr/lib64/heartbeat/crmd" as uid 495  gid 489 (pid 4291)
Jan 10 18:23:26 dl380g5d heartbeat: [4290]: info: Starting "/usr/lib64/heartbeat/attrd" as uid 495  gid 489 (pid 4290)
Jan 10 18:23:26 dl380g5d lrmd: [4288]: info: max-children set to 4 (4 processors online)
Jan 10 18:23:26 dl380g5d crmd: [4291]: info: Invoked: /usr/lib64/heartbeat/crmd 
Jan 10 18:23:26 dl380g5d crmd: [4291]: info: main: CRM Hg Version: 066152e
Jan 10 18:23:26 dl380g5d ccm: [4286]: info: Hostname: dl380g5d
Jan 10 18:23:26 dl380g5d crmd: [4291]: info: crmd_init: Starting crmd
Jan 10 18:23:26 dl380g5d cib: [4287]: info: Invoked: /usr/lib64/heartbeat/cib 
Jan 10 18:23:26 dl380g5d cib: [4287]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.xml (digest: /var/lib/heartbeat/crm/cib.xml.sig)
Jan 10 18:23:26 dl380g5d cib: [4287]: WARN: retrieveCib: Cluster configuration not found: /var/lib/heartbeat/crm/cib.xml
Jan 10 18:23:26 dl380g5d cib: [4287]: WARN: readCibXmlFile: Primary configuration corrupt or unusable, trying backup...
Jan 10 18:23:26 dl380g5d cib: [4287]: WARN: readCibXmlFile: Continuing with an empty configuration.
Jan 10 18:23:26 dl380g5d lrmd: [4288]: info: enabling coredumps
Jan 10 18:23:26 dl380g5d lrmd: [4288]: info: Started.
Jan 10 18:23:26 dl380g5d attrd: [4290]: info: Invoked: /usr/lib64/heartbeat/attrd 
Jan 10 18:23:26 dl380g5d attrd: [4290]: info: main: Starting up
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: the send queue length from heartbeat to client ccm is set to 1024
Jan 10 18:23:26 dl380g5d stonithd: [4289]: info: register_heartbeat_conn: Hostname: dl380g5d
Jan 10 18:23:26 dl380g5d stonithd: [4289]: info: register_heartbeat_conn: UUID: 3f6ec88d-ee47-4f63-bfeb-652b8dd96027
Jan 10 18:23:26 dl380g5d stonithd: [4289]: info: crm_cluster_connect: Connecting to Heartbeat
Jan 10 18:23:26 dl380g5d cib: [4287]: info: startCib: CIB Initialization completed successfully
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: the send queue length from heartbeat to client stonithd is set to 1024
Jan 10 18:23:26 dl380g5d attrd: [4290]: info: register_heartbeat_conn: Hostname: dl380g5d
Jan 10 18:23:26 dl380g5d attrd: [4290]: info: register_heartbeat_conn: UUID: 3f6ec88d-ee47-4f63-bfeb-652b8dd96027
Jan 10 18:23:26 dl380g5d attrd: [4290]: info: crm_cluster_connect: Connecting to Heartbeat
Jan 10 18:23:26 dl380g5d attrd: [4290]: info: main: Cluster connection active
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: the send queue length from heartbeat to client attrd is set to 1024
Jan 10 18:23:26 dl380g5d attrd: [4290]: info: main: Accepting attribute updates
Jan 10 18:23:26 dl380g5d attrd: [4290]: info: main: Starting mainloop...
Jan 10 18:23:26 dl380g5d stonithd: [4289]: notice: /usr/lib64/heartbeat/stonithd start up successfully.
Jan 10 18:23:26 dl380g5d cib: [4287]: info: register_heartbeat_conn: Hostname: dl380g5d
Jan 10 18:23:26 dl380g5d cib: [4287]: info: register_heartbeat_conn: UUID: 3f6ec88d-ee47-4f63-bfeb-652b8dd96027
Jan 10 18:23:26 dl380g5d cib: [4287]: info: crm_cluster_connect: Connecting to Heartbeat
Jan 10 18:23:26 dl380g5d cib: [4287]: info: ccm_connect: Registering with CCM...
Jan 10 18:23:26 dl380g5d cib: [4287]: WARN: ccm_connect: CCM Activation failed
Jan 10 18:23:26 dl380g5d cib: [4287]: WARN: ccm_connect: CCM Connection failed 1 times (30 max)
Jan 10 18:23:26 dl380g5d heartbeat: [4278]: info: the send queue length from heartbeat to client cib is set to 1024
Jan 10 18:23:27 dl380g5d heartbeat: [4278]: WARN: 1 lost packet(s) for [dl380g5c] [12:14]
Jan 10 18:23:27 dl380g5d heartbeat: [4278]: info: No pkts missing from dl380g5c!
Jan 10 18:23:27 dl380g5d crmd: [4291]: info: do_cib_control: Could not connect to the CIB service: connection failed
Jan 10 18:23:27 dl380g5d crmd: [4291]: WARN: do_cib_control: Couldn't complete CIB registration 1 times... pause and retry
Jan 10 18:23:27 dl380g5d crmd: [4291]: info: crmd_init: Starting crmd's mainloop
Jan 10 18:23:28 dl380g5d heartbeat: [4278]: WARN: 1 lost packet(s) for [dl380g5c] [17:19]
Jan 10 18:23:28 dl380g5d heartbeat: [4278]: info: No pkts missing from dl380g5c!
Jan 10 18:23:28 dl380g5d ccm: [4286]: debug: quorum plugin: majority
Jan 10 18:23:28 dl380g5d ccm: [4286]: debug: cluster:linux-ha, member_count=1, member_quorum_votes=100
Jan 10 18:23:28 dl380g5d ccm: [4286]: debug: total_node_count=2, total_quorum_votes=200
Jan 10 18:23:28 dl380g5d ccm: [4286]: debug: quorum plugin: twonodes
Jan 10 18:23:28 dl380g5d ccm: [4286]: debug: cluster:linux-ha, member_count=1, member_quorum_votes=100
Jan 10 18:23:28 dl380g5d ccm: [4286]: debug: total_node_count=2, total_quorum_votes=200
Jan 10 18:23:28 dl380g5d ccm: [4286]: info: Break tie for 2 nodes cluster
Jan 10 18:23:29 dl380g5d ccm: [4286]: debug: quorum plugin: majority
Jan 10 18:23:29 dl380g5d ccm: [4286]: debug: cluster:linux-ha, member_count=2, member_quorum_votes=200
Jan 10 18:23:29 dl380g5d ccm: [4286]: debug: total_node_count=2, total_quorum_votes=200
Jan 10 18:23:29 dl380g5d crmd: [4291]: info: crm_timer_popped: Wait Timer (I_NULL) just popped!
Jan 10 18:23:29 dl380g5d cib: [4287]: info: ccm_connect: Registering with CCM...
Jan 10 18:23:29 dl380g5d cib: [4287]: info: cib_init: Requesting the list of configured nodes
Jan 10 18:23:29 dl380g5d cib: [4287]: info: cib_init: Starting cib mainloop
Jan 10 18:23:29 dl380g5d cib: [4287]: info: cib_client_status_callback: Status update: Client dl380g5d/cib now has status [join]
Jan 10 18:23:29 dl380g5d cib: [4287]: info: crm_new_peer: Node 0 is now known as dl380g5d
Jan 10 18:23:29 dl380g5d cib: [4287]: info: crm_update_peer_proc: dl380g5d.cib is now online
Jan 10 18:23:29 dl380g5d cib: [4287]: info: mem_handle_event: Got an event OC_EV_MS_NEW_MEMBERSHIP from ccm
Jan 10 18:23:29 dl380g5d cib: [4287]: info: mem_handle_event: instance=2, nodes=2, new=2, lost=0, n_idx=0, new_idx=0, old_idx=4
Jan 10 18:23:29 dl380g5d cib: [4287]: info: cib_ccm_msg_callback: Processing CCM event=NEW MEMBERSHIP (id=2)
Jan 10 18:23:29 dl380g5d cib: [4287]: info: crm_get_peer: Node dl380g5d now has id: 1
Jan 10 18:23:29 dl380g5d cib: [4287]: info: crm_update_peer: Node dl380g5d: id=1 state=member (new) addr=(null) votes=-1 born=1 seen=2 proc=00000000000000000000000000000100
Jan 10 18:23:29 dl380g5d cib: [4287]: info: crm_update_peer_proc: dl380g5d.ais is now online
Jan 10 18:23:29 dl380g5d cib: [4287]: info: crm_update_peer_proc: dl380g5d.crmd is now online
Jan 10 18:23:30 dl380g5d cib: [4287]: info: crm_new_peer: Node 0 is now known as dl380g5c
Jan 10 18:23:30 dl380g5d cib: [4287]: info: crm_update_peer: Node dl380g5c: id=0 state=member (new) addr=(null) votes=-1 born=2 seen=2 proc=00000000000000000000000000000000
Jan 10 18:23:30 dl380g5d cib: [4287]: info: crm_update_peer_proc: dl380g5c.ais is now online
Jan 10 18:23:30 dl380g5d cib: [4287]: info: crm_update_peer_proc: dl380g5c.crmd is now online
Jan 10 18:23:30 dl380g5d cib: [4287]: info: cib_client_status_callback: Status update: Client dl380g5d/cib now has status [online]
Jan 10 18:23:30 dl380g5d cib: [4295]: info: write_cib_contents: Wrote version 0.0.0 of the CIB to disk (digest: 63ad14a4533cbde25732210894612651)
Jan 10 18:23:30 dl380g5d cib: [4295]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.Yi6J7y (digest: /var/lib/heartbeat/crm/cib.ybAqyI)
Jan 10 18:23:30 dl380g5d crmd: [4291]: info: do_cib_control: CIB connection established
Jan 10 18:23:30 dl380g5d cib: [4287]: info: cib_client_status_callback: Status update: Client dl380g5c/cib now has status [online]
Jan 10 18:23:30 dl380g5d cib: [4287]: info: crm_update_peer_proc: dl380g5c.cib is now online
Jan 10 18:23:30 dl380g5d crmd: [4291]: info: register_heartbeat_conn: Hostname: dl380g5d
Jan 10 18:23:30 dl380g5d crmd: [4291]: info: register_heartbeat_conn: UUID: 3f6ec88d-ee47-4f63-bfeb-652b8dd96027
Jan 10 18:23:30 dl380g5d crmd: [4291]: info: crm_cluster_connect: Connecting to Heartbeat
Jan 10 18:23:31 dl380g5d heartbeat: [4278]: info: the send queue length from heartbeat to client crmd is set to 1024
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: do_ha_control: Connected to the cluster
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: do_ccm_control: CCM connection established... waiting for first callback
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: do_started: Delaying start, CCM (0000000000100000) not connected
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: config_query_callback: Checking for expired actions every 900000ms
Jan 10 18:23:31 dl380g5d crmd: [4291]: notice: crmd_client_status_callback: Status update: Client dl380g5d/crmd now has status [online] (DC=false)
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crm_new_peer: Node 0 is now known as dl380g5d
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crm_update_peer_proc: dl380g5d.crmd is now online
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crmd_client_status_callback: Not the DC
Jan 10 18:23:31 dl380g5d crmd: [4291]: notice: crmd_client_status_callback: Status update: Client dl380g5d/crmd now has status [online] (DC=false)
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crmd_client_status_callback: Not the DC
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: mem_handle_event: Got an event OC_EV_MS_NEW_MEMBERSHIP from ccm
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: mem_handle_event: instance=2, nodes=2, new=2, lost=0, n_idx=0, new_idx=0, old_idx=4
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crmd_ccm_msg_callback: Quorum (re)attained after event=NEW MEMBERSHIP (id=2)
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: ccm_event_detail: NEW MEMBERSHIP: trans=2, nodes=2, new=2, lost=0 n_idx=0, new_idx=0, old_idx=4
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: ccm_event_detail: 	CURRENT: dl380g5d [nodeid=1, born=1]
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: ccm_event_detail: 	CURRENT: dl380g5c [nodeid=0, born=2]
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: ccm_event_detail: 	NEW:     dl380g5d [nodeid=1, born=1]
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: ccm_event_detail: 	NEW:     dl380g5c [nodeid=0, born=2]
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crm_get_peer: Node dl380g5d now has id: 1
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crm_update_peer: Node dl380g5d: id=1 state=member (new) addr=(null) votes=-1 born=1 seen=2 proc=00000000000000000000000000000200
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crm_update_peer_proc: dl380g5d.ais is now online
Jan 10 18:23:31 dl380g5d attrd: [4290]: info: cib_connect: Connected to the CIB after 1 signon attempts
Jan 10 18:23:31 dl380g5d attrd: [4290]: info: cib_connect: Sending full refresh
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crm_new_peer: Node 0 is now known as dl380g5c
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crm_update_peer: Node dl380g5c: id=0 state=member (new) addr=(null) votes=-1 born=2 seen=2 proc=00000000000000000000000000000000
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crm_update_peer_proc: dl380g5c.ais is now online
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: crm_update_peer_proc: dl380g5c.crmd is now online
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: do_started: The local CRM is operational
Jan 10 18:23:31 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
Jan 10 18:23:32 dl380g5d crmd: [4291]: notice: crmd_client_status_callback: Status update: Client dl380g5c/crmd now has status [online] (DC=false)
Jan 10 18:23:32 dl380g5d crmd: [4291]: info: crmd_client_status_callback: Not the DC
Jan 10 18:23:50 dl380g5d crmd: [4291]: info: do_election_count_vote: Election 2 (owner: 21c624bd-c426-43dc-9665-bbfb92054bcd) pass: vote from dl380g5c (Age)
Jan 10 18:23:50 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_PENDING -> S_ELECTION [ input=I_ELECTION cause=C_FSA_INTERNAL origin=do_election_count_vote ]
Jan 10 18:23:52 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC cause=C_FSA_INTERNAL origin=do_election_check ]
Jan 10 18:23:52 dl380g5d crmd: [4291]: info: do_te_control: Registering TE UUID: f619d469-9fad-4ee5-acd6-20cb09c4099c
Jan 10 18:23:52 dl380g5d crmd: [4291]: info: set_graph_functions: Setting custom graph functions
Jan 10 18:23:52 dl380g5d crmd: [4291]: info: unpack_graph: Unpacked transition -1: 0 actions in 0 synapses
Jan 10 18:23:52 dl380g5d crmd: [4291]: info: start_subsystem: Starting sub-system "pengine"
Jan 10 18:23:52 dl380g5d pengine: [4297]: info: Invoked: /usr/lib64/heartbeat/pengine 
Jan 10 18:23:52 dl380g5d pengine: [4297]: info: main: Starting pengine
Jan 10 18:23:56 dl380g5d crmd: [4291]: info: do_dc_takeover: Taking over DC status for this partition
Jan 10 18:23:56 dl380g5d cib: [4287]: info: cib_process_readwrite: We are now in R/W mode
Jan 10 18:23:56 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_master for section 'all' (origin=local/crmd/5, version=0.0.0): ok (rc=0)
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="0" num_updates="0" />
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + <cib crm_feature_set="3.0.1" admin_epoch="0" epoch="1" num_updates="1" />
Jan 10 18:23:56 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section cib (origin=local/crmd/6, version=0.1.1): ok (rc=0)
Jan 10 18:23:56 dl380g5d cib: [4304]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-0.raw
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="1" num_updates="1" />
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="2" num_updates="1" >
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   <configuration >
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     <crm_config >
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       <cluster_property_set id="cib-bootstrap-options" __crm_diff_marker__="added:top" >
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.0.12-066152e" />
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       </cluster_property_set>
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     </crm_config>
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   </configuration>
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + </cib>
Jan 10 18:23:56 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section crm_config (origin=local/crmd/9, version=0.2.1): ok (rc=0)
Jan 10 18:23:56 dl380g5d crmd: [4291]: info: join_make_offer: Making join offers based on membership 2
Jan 10 18:23:56 dl380g5d crmd: [4291]: info: do_dc_join_offer_all: join-1: Waiting on 2 outstanding join acks
Jan 10 18:23:56 dl380g5d crmd: [4291]: info: te_connect_stonith: Attempting connection to fencing daemon...
Jan 10 18:23:56 dl380g5d cib: [4304]: info: write_cib_contents: Wrote version 0.1.0 of the CIB to disk (digest: 00d3881d2d2a8f19337322a27749c792)
Jan 10 18:23:56 dl380g5d cib: [4304]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.A9aQSV (digest: /var/lib/heartbeat/crm/cib.YvTDxs)
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="2" num_updates="1" />
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="3" num_updates="1" >
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   <configuration >
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     <crm_config >
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       <cluster_property_set id="cib-bootstrap-options" >
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="Heartbeat" __crm_diff_marker__="added:top" />
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       </cluster_property_set>
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     </crm_config>
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   </configuration>
Jan 10 18:23:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + </cib>
Jan 10 18:23:56 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section crm_config (origin=local/crmd/12, version=0.3.1): ok (rc=0)
Jan 10 18:23:56 dl380g5d cib: [4305]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-1.raw
Jan 10 18:23:56 dl380g5d cib: [4305]: info: write_cib_contents: Wrote version 0.3.0 of the CIB to disk (digest: 9dc954be2be499cb3c773d84bd50a4e8)
Jan 10 18:23:56 dl380g5d cib: [4305]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.74oT6Y (digest: /var/lib/heartbeat/crm/cib.IylIPv)
Jan 10 18:23:57 dl380g5d crmd: [4291]: info: te_connect_stonith: Connected
Jan 10 18:23:57 dl380g5d crmd: [4291]: info: config_query_callback: Checking for expired actions every 900000ms
Jan 10 18:23:57 dl380g5d crmd: [4291]: info: config_query_callback: Checking for expired actions every 900000ms
Jan 10 18:23:57 dl380g5d crmd: [4291]: info: update_dc: Set DC to dl380g5d (3.0.1)
Jan 10 18:23:57 dl380g5d crmd: [4291]: info: config_query_callback: Checking for expired actions every 900000ms
Jan 10 18:23:57 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED cause=C_FSA_INTERNAL origin=check_join_state ]
Jan 10 18:23:57 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes responded to the join offer.
Jan 10 18:23:57 dl380g5d crmd: [4291]: info: do_dc_join_finalize: join-1: Syncing the CIB from dl380g5d to the rest of the cluster
Jan 10 18:23:57 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_sync for section 'all' (origin=local/crmd/17, version=0.3.1): ok (rc=0)
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="3" num_updates="1" />
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="4" num_updates="1" >
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   <configuration >
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     <nodes >
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       <node id="21c624bd-c426-43dc-9665-bbfb92054bcd" uname="dl380g5c" type="normal" __crm_diff_marker__="added:top" />
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     </nodes>
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   </configuration>
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + </cib>
Jan 10 18:23:58 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/18, version=0.4.1): ok (rc=0)
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="4" num_updates="1" />
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="5" num_updates="1" >
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   <configuration >
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     <nodes >
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       <node id="3f6ec88d-ee47-4f63-bfeb-652b8dd96027" uname="dl380g5d" type="normal" __crm_diff_marker__="added:top" />
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     </nodes>
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   </configuration>
Jan 10 18:23:58 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + </cib>
Jan 10 18:23:58 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/19, version=0.5.1): ok (rc=0)
Jan 10 18:23:58 dl380g5d cib: [4306]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-2.raw
Jan 10 18:23:58 dl380g5d cib: [4306]: info: write_cib_contents: Wrote version 0.5.0 of the CIB to disk (digest: 0e72e51ed24e43bb6de7ae4f3a31471b)
Jan 10 18:23:58 dl380g5d cib: [4306]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.aSAV51 (digest: /var/lib/heartbeat/crm/cib.1AptjF)
Jan 10 18:23:58 dl380g5d crmd: [4291]: info: update_attrd: Connecting to attrd...
Jan 10 18:23:58 dl380g5d attrd: [4290]: info: find_hash_entry: Creating hash entry for terminate
Jan 10 18:23:58 dl380g5d attrd: [4290]: info: find_hash_entry: Creating hash entry for shutdown
Jan 10 18:23:58 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='dl380g5d']/transient_attributes (origin=local/crmd/20, version=0.5.1): ok (rc=0)
Jan 10 18:23:58 dl380g5d crmd: [4291]: info: erase_xpath_callback: Deletion of "//node_state[@uname='dl380g5d']/transient_attributes": ok (rc=0)
Jan 10 18:23:59 dl380g5d crmd: [4291]: info: do_dc_join_ack: join-1: Updating node state to member for dl380g5c
Jan 10 18:23:59 dl380g5d crmd: [4291]: info: do_dc_join_ack: join-1: Updating node state to member for dl380g5d
Jan 10 18:23:59 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='dl380g5c']/transient_attributes (origin=dl380g5c/crmd/8, version=0.5.1): ok (rc=0)
Jan 10 18:23:59 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='dl380g5c']/lrm (origin=local/crmd/21, version=0.5.1): ok (rc=0)
Jan 10 18:23:59 dl380g5d crmd: [4291]: info: erase_xpath_callback: Deletion of "//node_state[@uname='dl380g5c']/lrm": ok (rc=0)
Jan 10 18:23:59 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED cause=C_FSA_INTERNAL origin=check_join_state ]
Jan 10 18:23:59 dl380g5d crmd: [4291]: info: populate_cib_nodes_ha: Requesting the list of configured nodes
Jan 10 18:23:59 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='dl380g5d']/lrm (origin=local/crmd/23, version=0.5.2): ok (rc=0)
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: do_dc_join_final: Ensuring DC, quorum and node attributes are up-to-date
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: crm_update_quorum: Updating quorum status to true (call=27)
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: abort_transition_graph: do_te_invoke:185 - Triggered transition abort (complete=1) : Peer Cancelled
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: do_pe_invoke: Query 28: Requesting the current CIB: S_POLICY_ENGINE
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: erase_xpath_callback: Deletion of "//node_state[@uname='dl380g5d']/lrm": ok (rc=0)
Jan 10 18:24:00 dl380g5d attrd: [4290]: info: attrd_local_callback: Sending full refresh (origin=crmd)
Jan 10 18:24:00 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/25, version=0.5.3): ok (rc=0)
Jan 10 18:24:00 dl380g5d cib: [4287]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="5" num_updates="3" />
Jan 10 18:24:00 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + <cib have-quorum="1" dc-uuid="3f6ec88d-ee47-4f63-bfeb-652b8dd96027" admin_epoch="0" epoch="6" num_updates="1" />
Jan 10 18:24:00 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section cib (origin=local/crmd/27, version=0.6.1): ok (rc=0)
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: abort_transition_graph: need_abort:59 - Triggered transition abort (complete=1) : Non-status change
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: need_abort: Aborting on change to have-quorum
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: do_pe_invoke: Query 29: Requesting the current CIB: S_POLICY_ENGINE
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: do_pe_invoke_callback: Invoking the PE: query=29, ref=pe_calc-dc-1357809840-9, seq=2, quorate=1
Jan 10 18:24:00 dl380g5d pengine: [4297]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jan 10 18:24:00 dl380g5d pengine: [4297]: ERROR: unpack_resources: Resource start-up disabled since no STONITH resources have been defined
Jan 10 18:24:00 dl380g5d pengine: [4297]: ERROR: unpack_resources: Either configure some or disable STONITH with the stonith-enabled option
Jan 10 18:24:00 dl380g5d pengine: [4297]: ERROR: unpack_resources: NOTE: Clusters with shared data need STONITH to ensure data integrity
Jan 10 18:24:00 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5c is online
Jan 10 18:24:00 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5d is online
Jan 10 18:24:00 dl380g5d pengine: [4297]: info: stage6: Delaying fencing operations until there are resources to manage
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: unpack_graph: Unpacked transition 0: 2 actions in 2 synapses
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: do_te_invoke: Processing graph 0 (ref=pe_calc-dc-1357809840-9) derived from /var/lib/pengine/pe-input-0.bz2
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 2: probe_complete probe_complete on dl380g5c - no waiting
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 3: probe_complete probe_complete on dl380g5d (local) - no waiting
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: run_graph: ====================================================
Jan 10 18:24:00 dl380g5d crmd: [4291]: notice: run_graph: Transition 0 (Complete=2, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pengine/pe-input-0.bz2): Complete
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: te_graph_trigger: Transition 0 is now complete
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: notify_crmd: Transition 0 status: done - <null>
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Jan 10 18:24:00 dl380g5d attrd: [4290]: info: find_hash_entry: Creating hash entry for probe_complete
Jan 10 18:24:00 dl380g5d attrd: [4290]: info: attrd_trigger_update: Sending flush op to all hosts for: probe_complete (true)
Jan 10 18:24:00 dl380g5d crmd: [4291]: info: do_state_transition: Starting PEngine Recheck Timer
Jan 10 18:24:00 dl380g5d pengine: [4297]: info: process_pe_message: Transition 0: PEngine Input stored in: /var/lib/pengine/pe-input-0.bz2
Jan 10 18:24:00 dl380g5d pengine: [4297]: info: process_pe_message: Configuration ERRORs found during PE processing.  Please run "crm_verify -L" to identify issues.
Jan 10 18:24:00 dl380g5d attrd: [4290]: info: attrd_perform_update: Sent update 4: probe_complete=true
Jan 10 18:24:00 dl380g5d cib: [4327]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-3.raw
Jan 10 18:24:00 dl380g5d cib: [4327]: info: write_cib_contents: Wrote version 0.6.0 of the CIB to disk (digest: 9de0b987f60d4b10a89b70c865b2b7fa)
Jan 10 18:24:00 dl380g5d cib: [4327]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.3n65vc (digest: /var/lib/heartbeat/crm/cib.ef3ekW)
Jan 10 18:24:01 dl380g5d ccm: [4286]: info: client (pid=4326) removed from ccm
Jan 10 18:24:01 dl380g5d attrd: [4290]: info: attrd_ha_callback: flush message from dl380g5c
Jan 10 18:24:57 dl380g5d cib: [4287]: info: cib_replace_notify: Replaced: 0.6.3 -> 0.7.1 from dl380g5c
Jan 10 18:24:57 dl380g5d attrd: [4290]: info: do_cib_replaced: Sending full refresh
Jan 10 18:24:57 dl380g5d attrd: [4290]: info: attrd_trigger_update: Sending flush op to all hosts for: probe_complete (true)
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="6" num_updates="3" />
Jan 10 18:24:57 dl380g5d crmd: [4291]: info: abort_transition_graph: need_abort:59 - Triggered transition abort (complete=1) : Non-status change
Jan 10 18:24:57 dl380g5d crmd: [4291]: info: need_abort: Aborting on change to admin_epoch
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="7" num_updates="1" >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   <configuration >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     <crm_config >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       <cluster_property_set id="cib-bootstrap-options" >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-no-quorum-policy" name="no-quorum-policy" value="ignore" __crm_diff_marker__="added:top" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="false" __crm_diff_marker__="added:top" />
Jan 10 18:24:57 dl380g5d crmd: [4291]: info: populate_cib_nodes_ha: Requesting the list of configured nodes
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-startup-fencing" name="startup-fencing" value="false" __crm_diff_marker__="added:top" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-crmd-transition-delay" name="crmd-transition-delay" value="2s" __crm_diff_marker__="added:top" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       </cluster_property_set>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     </crm_config>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     <resources >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       <master id="stateful" __crm_diff_marker__="added:top" >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <primitive class="ocf" id="dummy" provider="pacemaker" type="Stateful" >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +           <operations >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +             <op id="dummy-start-0s" interval="0s" name="start" on-fail="restart" timeout="100s" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +             <op id="dummy-monitor-10s" interval="10s" name="monitor" on-fail="restart" role="Master" timeout="100s" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +             <op id="dummy-monitor-20s" interval="20s" name="monitor" on-fail="restart" role="Slave" timeout="100s" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +             <op id="dummy-promote-0s" interval="0s" name="promote" on-fail="restart" timeout="100s" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +             <op id="dummy-demote-0s" interval="0s" name="demote" on-fail="block" timeout="100s" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +             <op id="dummy-stop-0s" interval="0s" name="stop" on-fail="block" timeout="100s" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +           </operations>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         </primitive>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       </master>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     </resources>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     <rsc_defaults __crm_diff_marker__="added:top" >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       <meta_attributes id="rsc-options" >
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <nvpair id="rsc-options-resource-stickiness" name="resource-stickiness" value="INFINITY" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <nvpair id="rsc-options-migration-threshold" name="migration-threshold" value="1" />
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       </meta_attributes>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     </rsc_defaults>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   </configuration>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + </cib>
Jan 10 18:24:57 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_replace for section 'all' (origin=dl380g5c/cibadmin/2, version=0.7.1): ok (rc=0)
Jan 10 18:24:57 dl380g5d cib: [4376]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-4.raw
Jan 10 18:24:57 dl380g5d cib: [4376]: info: write_cib_contents: Wrote version 0.7.0 of the CIB to disk (digest: cfce166174df5afa2e880f2417cf29f7)
Jan 10 18:24:57 dl380g5d cib: [4376]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.lv7WZL (digest: /var/lib/heartbeat/crm/cib.8V5m7x)
Jan 10 18:24:59 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
Jan 10 18:24:59 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jan 10 18:24:59 dl380g5d crmd: [4291]: info: do_pe_invoke: Query 32: Requesting the current CIB: S_POLICY_ENGINE
Jan 10 18:24:59 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_ELECTION [ input=I_ELECTION cause=C_FSA_INTERNAL origin=do_cib_replaced ]
Jan 10 18:24:59 dl380g5d crmd: [4291]: info: update_dc: Unset DC dl380g5d
Jan 10 18:24:59 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/30, version=0.7.1): ok (rc=0)
Jan 10 18:24:59 dl380g5d crmd: [4291]: info: config_query_callback: Checking for expired actions every 900000ms
Jan 10 18:25:00 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC cause=C_FSA_INTERNAL origin=do_election_check ]
Jan 10 18:25:00 dl380g5d crmd: [4291]: info: start_subsystem: Starting sub-system "pengine"
Jan 10 18:25:00 dl380g5d crmd: [4291]: WARN: start_subsystem: Client pengine already running as pid 4297
Jan 10 18:25:00 dl380g5d ccm: [4286]: info: client (pid=4396) removed from ccm
Jan 10 18:25:04 dl380g5d crmd: [4291]: info: do_dc_takeover: Taking over DC status for this partition
Jan 10 18:25:04 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_master for section 'all' (origin=local/crmd/34, version=0.7.1): ok (rc=0)
Jan 10 18:25:04 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section cib (origin=local/crmd/35, version=0.7.1): ok (rc=0)
Jan 10 18:25:04 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section crm_config (origin=local/crmd/37, version=0.7.1): ok (rc=0)
Jan 10 18:25:04 dl380g5d crmd: [4291]: info: do_dc_join_offer_all: join-2: Waiting on 2 outstanding join acks
Jan 10 18:25:04 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section crm_config (origin=local/crmd/39, version=0.7.1): ok (rc=0)
Jan 10 18:25:04 dl380g5d crmd: [4291]: info: config_query_callback: Checking for expired actions every 900000ms
Jan 10 18:25:04 dl380g5d crmd: [4291]: info: update_dc: Set DC to dl380g5d (3.0.1)
Jan 10 18:25:05 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED cause=C_FSA_INTERNAL origin=check_join_state ]
Jan 10 18:25:05 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes responded to the join offer.
Jan 10 18:25:05 dl380g5d crmd: [4291]: info: do_dc_join_finalize: join-2: Syncing the CIB from dl380g5d to the rest of the cluster
Jan 10 18:25:05 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_sync for section 'all' (origin=local/crmd/42, version=0.7.1): ok (rc=0)
Jan 10 18:25:05 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/43, version=0.7.1): ok (rc=0)
Jan 10 18:25:05 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/44, version=0.7.1): ok (rc=0)
Jan 10 18:25:05 dl380g5d crmd: [4291]: info: do_dc_join_ack: join-2: Updating node state to member for dl380g5d
Jan 10 18:25:05 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='dl380g5d']/lrm (origin=local/crmd/45, version=0.7.2): ok (rc=0)
Jan 10 18:25:05 dl380g5d crmd: [4291]: info: erase_xpath_callback: Deletion of "//node_state[@uname='dl380g5d']/lrm": ok (rc=0)
Jan 10 18:25:06 dl380g5d attrd: [4290]: info: attrd_ha_callback: flush message from dl380g5c
Jan 10 18:25:06 dl380g5d crmd: [4291]: info: do_dc_join_ack: join-2: Updating node state to member for dl380g5c
Jan 10 18:25:06 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='dl380g5c']/lrm (origin=local/crmd/47, version=0.7.4): ok (rc=0)
Jan 10 18:25:06 dl380g5d crmd: [4291]: info: erase_xpath_callback: Deletion of "//node_state[@uname='dl380g5c']/lrm": ok (rc=0)
Jan 10 18:25:06 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED cause=C_FSA_INTERNAL origin=check_join_state ]
Jan 10 18:25:06 dl380g5d crmd: [4291]: info: populate_cib_nodes_ha: Requesting the list of configured nodes
Jan 10 18:25:07 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jan 10 18:25:07 dl380g5d crmd: [4291]: info: do_dc_join_final: Ensuring DC, quorum and node attributes are up-to-date
Jan 10 18:25:07 dl380g5d crmd: [4291]: info: crm_update_quorum: Updating quorum status to true (call=51)
Jan 10 18:25:07 dl380g5d crmd: [4291]: info: abort_transition_graph: do_te_invoke:185 - Triggered transition abort (complete=1) : Peer Cancelled
Jan 10 18:25:07 dl380g5d attrd: [4290]: info: attrd_local_callback: Sending full refresh (origin=crmd)
Jan 10 18:25:07 dl380g5d attrd: [4290]: info: attrd_trigger_update: Sending flush op to all hosts for: probe_complete (true)
Jan 10 18:25:07 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/49, version=0.7.5): ok (rc=0)
Jan 10 18:25:07 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section cib (origin=local/crmd/51, version=0.7.5): ok (rc=0)
Jan 10 18:25:07 dl380g5d ccm: [4286]: info: client (pid=4456) removed from ccm
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped!
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: do_pe_invoke: Query 52: Requesting the current CIB: S_POLICY_ENGINE
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: do_pe_invoke_callback: Invoking the PE: query=52, ref=pe_calc-dc-1357809909-19, seq=2, quorate=1
Jan 10 18:25:09 dl380g5d pengine: [4297]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jan 10 18:25:09 dl380g5d pengine: [4297]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jan 10 18:25:09 dl380g5d pengine: [4297]: WARN: unpack_nodes: Blind faith: not fencing unseen nodes
Jan 10 18:25:09 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5c is online
Jan 10 18:25:09 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5d is online
Jan 10 18:25:09 dl380g5d pengine: [4297]: notice: clone_print:  Master/Slave Set: stateful
Jan 10 18:25:09 dl380g5d pengine: [4297]: notice: short_print:      Stopped: [ dummy:0 dummy:1 ]
Jan 10 18:25:09 dl380g5d pengine: [4297]: info: master_color: stateful: Promoted 0 instances of a possible 1 to master
Jan 10 18:25:09 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (20s) for dummy:0 on dl380g5c
Jan 10 18:25:09 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (20s) for dummy:1 on dl380g5d
Jan 10 18:25:09 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (20s) for dummy:0 on dl380g5c
Jan 10 18:25:09 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (20s) for dummy:1 on dl380g5d
Jan 10 18:25:09 dl380g5d pengine: [4297]: notice: LogActions: Start   dummy:0	(dl380g5c)
Jan 10 18:25:09 dl380g5d pengine: [4297]: notice: LogActions: Start   dummy:1	(dl380g5d)
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: unpack_graph: Unpacked transition 1: 12 actions in 12 synapses
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: do_te_invoke: Processing graph 1 (ref=pe_calc-dc-1357809909-19) derived from /var/lib/pengine/pe-input-1.bz2
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 4: monitor dummy:0_monitor_0 on dl380g5c
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 6: monitor dummy:1_monitor_0 on dl380g5d (local)
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: do_lrm_rsc_op: Performing key=6:1:7:f619d469-9fad-4ee5-acd6-20cb09c4099c op=dummy:1_monitor_0 )
Jan 10 18:25:09 dl380g5d lrmd: [4288]: info: rsc:dummy:1 probe[2] (pid 4497)
Jan 10 18:25:09 dl380g5d pengine: [4297]: info: process_pe_message: Transition 1: PEngine Input stored in: /var/lib/pengine/pe-input-1.bz2
Jan 10 18:25:09 dl380g5d lrmd: [4288]: info: operation monitor[2] on dummy:1 for client 4291: pid 4497 exited with return code 7
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: process_lrm_event: LRM operation dummy:1_monitor_0 (call=2, rc=7, cib-update=53, confirmed=true) not running
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: match_graph_event: Action dummy:1_monitor_0 (6) confirmed on dl380g5d (rc=0)
Jan 10 18:25:09 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 5: probe_complete probe_complete on dl380g5d (local) - no waiting
Jan 10 18:25:10 dl380g5d attrd: [4290]: info: attrd_ha_callback: flush message from dl380g5c
Jan 10 18:25:10 dl380g5d attrd: [4290]: info: find_hash_entry: Creating hash entry for master-dummy:0
Jan 10 18:25:10 dl380g5d crmd: [4291]: info: abort_transition_graph: te_update_diff:150 - Triggered transition abort (complete=0, tag=nvpair, id=status-21c624bd-c426-43dc-9665-bbfb92054bcd-master-dummy:0, name=master-dummy:0, value=10, magic=NA, cib=0.7.7) : Transient attribute: update
Jan 10 18:25:10 dl380g5d crmd: [4291]: info: update_abort_priority: Abort priority upgraded from 0 to 1000000
Jan 10 18:25:10 dl380g5d crmd: [4291]: info: update_abort_priority: Abort action done superceeded by restart
Jan 10 18:25:10 dl380g5d crmd: [4291]: WARN: status_from_rc: Action 4 (dummy:0_monitor_0) on dl380g5c failed (target: 7 vs. rc: 8): Error
Jan 10 18:25:10 dl380g5d crmd: [4291]: info: abort_transition_graph: match_graph_event:299 - Triggered transition abort (complete=0, tag=lrm_rsc_op, id=dummy:0_monitor_0, magic=0:8;4:1:7:f619d469-9fad-4ee5-acd6-20cb09c4099c, cib=0.7.8) : Event failed
Jan 10 18:25:10 dl380g5d crmd: [4291]: info: match_graph_event: Action dummy:0_monitor_0 (4) confirmed on dl380g5c (rc=4)
Jan 10 18:25:10 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 3: probe_complete probe_complete on dl380g5c - no waiting
Jan 10 18:25:10 dl380g5d crmd: [4291]: info: run_graph: ====================================================
Jan 10 18:25:10 dl380g5d crmd: [4291]: notice: run_graph: Transition 1 (Complete=4, Pending=0, Fired=0, Skipped=7, Incomplete=1, Source=/var/lib/pengine/pe-input-1.bz2): Stopped
Jan 10 18:25:10 dl380g5d crmd: [4291]: info: te_graph_trigger: Transition 1 is now complete
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped!
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: do_pe_invoke: Query 54: Requesting the current CIB: S_POLICY_ENGINE
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: do_pe_invoke_callback: Invoking the PE: query=54, ref=pe_calc-dc-1357809912-24, seq=2, quorate=1
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jan 10 18:25:12 dl380g5d pengine: [4297]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jan 10 18:25:12 dl380g5d pengine: [4297]: WARN: unpack_nodes: Blind faith: not fencing unseen nodes
Jan 10 18:25:12 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5c is online
Jan 10 18:25:12 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5d is online
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: unpack_rsc_op: Operation dummy:0_monitor_0 found resource dummy:0 active in master mode on dl380g5c
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: clone_print:  Master/Slave Set: stateful
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: short_print:      Masters: [ dl380g5c ]
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: short_print:      Stopped: [ dummy:1 ]
Jan 10 18:25:12 dl380g5d pengine: [4297]: info: master_color: Promoting dummy:0 (Master dl380g5c)
Jan 10 18:25:12 dl380g5d pengine: [4297]: info: master_color: stateful: Promoted 1 instances of a possible 1 to master
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (10s) for dummy:0 on dl380g5c
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (20s) for dummy:1 on dl380g5d
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (10s) for dummy:0 on dl380g5c
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (20s) for dummy:1 on dl380g5d
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: LogActions: Leave   resource dummy:0	(Master dl380g5c)
Jan 10 18:25:12 dl380g5d pengine: [4297]: notice: LogActions: Start   dummy:1	(dl380g5d)
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: unpack_graph: Unpacked transition 2: 5 actions in 5 synapses
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: do_te_invoke: Processing graph 2 (ref=pe_calc-dc-1357809912-24) derived from /var/lib/pengine/pe-input-2.bz2
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 9: monitor dummy:0_monitor_10000 on dl380g5c
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: te_pseudo_action: Pseudo action 12 fired and confirmed
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 10: start dummy:1_start_0 on dl380g5d (local)
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: do_lrm_rsc_op: Performing key=10:2:0:f619d469-9fad-4ee5-acd6-20cb09c4099c op=dummy:1_start_0 )
Jan 10 18:25:12 dl380g5d lrmd: [4288]: info: rsc:dummy:1 start[3] (pid 4503)
Jan 10 18:25:12 dl380g5d pengine: [4297]: info: process_pe_message: Transition 2: PEngine Input stored in: /var/lib/pengine/pe-input-2.bz2
Jan 10 18:25:12 dl380g5d crm_attribute: [4512]: info: Invoked: crm_attribute -N dl380g5d -n master-dummy:1 -l reboot -v 5 
Jan 10 18:25:12 dl380g5d attrd: [4290]: info: find_hash_entry: Creating hash entry for master-dummy:1
Jan 10 18:25:12 dl380g5d attrd: [4290]: info: attrd_trigger_update: Sending flush op to all hosts for: master-dummy:1 (5)
Jan 10 18:25:12 dl380g5d attrd: [4290]: info: attrd_perform_update: Sent update 16: master-dummy:1=5
Jan 10 18:25:12 dl380g5d lrmd: [4288]: info: operation start[3] on dummy:1 for client 4291: pid 4503 exited with return code 0
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: process_lrm_event: LRM operation dummy:1_start_0 (call=3, rc=0, cib-update=55, confirmed=true) ok
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: abort_transition_graph: te_update_diff:150 - Triggered transition abort (complete=0, tag=nvpair, id=status-3f6ec88d-ee47-4f63-bfeb-652b8dd96027-master-dummy:1, name=master-dummy:1, value=5, magic=NA, cib=0.7.9) : Transient attribute: update
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: update_abort_priority: Abort priority upgraded from 0 to 1000000
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: update_abort_priority: Abort action done superceeded by restart
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: match_graph_event: Action dummy:1_start_0 (10) confirmed on dl380g5d (rc=0)
Jan 10 18:25:12 dl380g5d crmd: [4291]: info: te_pseudo_action: Pseudo action 13 fired and confirmed
Jan 10 18:25:13 dl380g5d ccm: [4286]: info: client (pid=4536) removed from ccm
Jan 10 18:25:14 dl380g5d crmd: [4291]: info: match_graph_event: Action dummy:0_monitor_10000 (9) confirmed on dl380g5c (rc=0)
Jan 10 18:25:14 dl380g5d crmd: [4291]: info: run_graph: ====================================================
Jan 10 18:25:14 dl380g5d crmd: [4291]: notice: run_graph: Transition 2 (Complete=4, Pending=0, Fired=0, Skipped=1, Incomplete=0, Source=/var/lib/pengine/pe-input-2.bz2): Stopped
Jan 10 18:25:14 dl380g5d crmd: [4291]: info: te_graph_trigger: Transition 2 is now complete
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped!
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_pe_invoke: Query 56: Requesting the current CIB: S_POLICY_ENGINE
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_pe_invoke_callback: Invoking the PE: query=56, ref=pe_calc-dc-1357809916-27, seq=2, quorate=1
Jan 10 18:25:16 dl380g5d pengine: [4297]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jan 10 18:25:16 dl380g5d pengine: [4297]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jan 10 18:25:16 dl380g5d pengine: [4297]: WARN: unpack_nodes: Blind faith: not fencing unseen nodes
Jan 10 18:25:16 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5c is online
Jan 10 18:25:16 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5d is online
Jan 10 18:25:16 dl380g5d pengine: [4297]: notice: unpack_rsc_op: Operation dummy:0_monitor_0 found resource dummy:0 active in master mode on dl380g5c
Jan 10 18:25:16 dl380g5d pengine: [4297]: notice: clone_print:  Master/Slave Set: stateful
Jan 10 18:25:16 dl380g5d pengine: [4297]: notice: short_print:      Masters: [ dl380g5c ]
Jan 10 18:25:16 dl380g5d pengine: [4297]: notice: short_print:      Slaves: [ dl380g5d ]
Jan 10 18:25:16 dl380g5d pengine: [4297]: info: master_color: Promoting dummy:0 (Master dl380g5c)
Jan 10 18:25:16 dl380g5d pengine: [4297]: info: master_color: stateful: Promoted 1 instances of a possible 1 to master
Jan 10 18:25:16 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (20s) for dummy:1 on dl380g5d
Jan 10 18:25:16 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (20s) for dummy:1 on dl380g5d
Jan 10 18:25:16 dl380g5d pengine: [4297]: notice: LogActions: Leave   resource dummy:0	(Master dl380g5c)
Jan 10 18:25:16 dl380g5d pengine: [4297]: notice: LogActions: Leave   resource dummy:1	(Slave dl380g5d)
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: unpack_graph: Unpacked transition 3: 1 actions in 1 synapses
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_te_invoke: Processing graph 3 (ref=pe_calc-dc-1357809916-27) derived from /var/lib/pengine/pe-input-3.bz2
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 13: monitor dummy:1_monitor_20000 on dl380g5d (local)
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_lrm_rsc_op: Performing key=13:3:0:f619d469-9fad-4ee5-acd6-20cb09c4099c op=dummy:1_monitor_20000 )
Jan 10 18:25:16 dl380g5d lrmd: [4288]: info: rsc:dummy:1 monitor[4] (pid 4577)
Jan 10 18:25:16 dl380g5d pengine: [4297]: info: process_pe_message: Transition 3: PEngine Input stored in: /var/lib/pengine/pe-input-3.bz2
Jan 10 18:25:16 dl380g5d lrmd: [4288]: info: operation monitor[4] on dummy:1 for client 4291: pid 4577 exited with return code 0
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: process_lrm_event: LRM operation dummy:1_monitor_20000 (call=4, rc=0, cib-update=57, confirmed=false) ok
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: match_graph_event: Action dummy:1_monitor_20000 (13) confirmed on dl380g5d (rc=0)
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: run_graph: ====================================================
Jan 10 18:25:16 dl380g5d crmd: [4291]: notice: run_graph: Transition 3 (Complete=1, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pengine/pe-input-3.bz2): Complete
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: te_graph_trigger: Transition 3 is now complete
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: notify_crmd: Transition 3 status: done - <null>
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Jan 10 18:25:16 dl380g5d crmd: [4291]: info: do_state_transition: Starting PEngine Recheck Timer
Jan 10 18:25:17 dl380g5d ccm: [4286]: info: client (pid=4604) removed from ccm
Jan 10 18:26:14 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:26:15 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:26:18 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:26:18 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="7" num_updates="12" />
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="8" num_updates="1" >
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   <configuration >
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     <nodes >
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       <node id="21c624bd-c426-43dc-9665-bbfb92054bcd" >
Jan 10 18:26:56 dl380g5d crmd: [4291]: info: abort_transition_graph: need_abort:59 - Triggered transition abort (complete=1) : Non-status change
Jan 10 18:26:56 dl380g5d crmd: [4291]: info: need_abort: Aborting on change to admin_epoch
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         <instance_attributes id="nodes-21c624bd-c426-43dc-9665-bbfb92054bcd" __crm_diff_marker__="added:top" >
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +           <nvpair id="nodes-21c624bd-c426-43dc-9665-bbfb92054bcd-standby" name="standby" value="on" />
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +         </instance_attributes>
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +       </node>
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +     </nodes>
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: +   </configuration>
Jan 10 18:26:56 dl380g5d cib: [4287]: info: log_data_element: cib:diff: + </cib>
Jan 10 18:26:56 dl380g5d cib: [4287]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=dl380g5c/crm_attribute/5, version=0.8.1): ok (rc=0)
Jan 10 18:26:56 dl380g5d cib: [4724]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-5.raw
Jan 10 18:26:56 dl380g5d cib: [4724]: info: write_cib_contents: Wrote version 0.8.0 of the CIB to disk (digest: 8a7b0043b1514bc6c3d748a7903b5842)
Jan 10 18:26:56 dl380g5d cib: [4724]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.CLINT9 (digest: /var/lib/heartbeat/crm/cib.gMs2Qd)
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped!
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: do_pe_invoke: Query 58: Requesting the current CIB: S_POLICY_ENGINE
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: do_pe_invoke_callback: Invoking the PE: query=58, ref=pe_calc-dc-1357810018-31, seq=2, quorate=1
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jan 10 18:26:58 dl380g5d pengine: [4297]: WARN: unpack_nodes: Blind faith: not fencing unseen nodes
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: unpack_status: Node dl380g5c is in standby-mode
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5c is standby
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5d is online
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: unpack_rsc_op: Operation dummy:0_monitor_0 found resource dummy:0 active in master mode on dl380g5c
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: clone_print:  Master/Slave Set: stateful
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: short_print:      Masters: [ dl380g5c ]
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: short_print:      Slaves: [ dl380g5d ]
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: native_color: Resource dummy:0 cannot run anywhere
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: master_color: Promoting dummy:1 (Slave dl380g5d)
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: master_color: stateful: Promoted 1 instances of a possible 1 to master
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (10s) for dummy:1 on dl380g5d
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: RecurringOp: Cancelling action dummy:1_monitor_20000 (Slave vs. Master)
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (10s) for dummy:1 on dl380g5d
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: RecurringOp: Cancelling action dummy:1_monitor_20000 (Slave vs. Master)
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: LogActions: Demote  dummy:0	(Master -> Stopped dl380g5c)
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: LogActions: Stop    resource dummy:0	(dl380g5c)
Jan 10 18:26:58 dl380g5d pengine: [4297]: notice: LogActions: Promote dummy:1	(Slave -> Master dl380g5d)
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: unpack_graph: Unpacked transition 4: 12 actions in 12 synapses
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: do_te_invoke: Processing graph 4 (ref=pe_calc-dc-1357810018-31) derived from /var/lib/pengine/pe-input-4.bz2
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 2: cancel dummy:1_monitor_20000 on dl380g5d (local)
Jan 10 18:26:58 dl380g5d lrmd: [4288]: info: cancel_op: operation monitor[4] on dummy:1 for client 4291, its parameters: CRM_meta_clone=[1] CRM_meta_master_node_max=[1] CRM_meta_clone_node_max=[1] CRM_meta_clone_max=[2] CRM_meta_notify=[false] CRM_meta_master_max=[1] crm_feature_set=[3.0.1] CRM_meta_globally_unique=[false] CRM_meta_on_fail=[restart] CRM_meta_name=[monitor] CRM_meta_role=[Slave] CRM_meta_interval=[20000] CRM_meta_timeout=[100000]  cancelled
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: send_direct_ack: ACK'ing resource op dummy:1_monitor_20000 from 2:4:0:f619d469-9fad-4ee5-acd6-20cb09c4099c: lrm_invoke-lrmd-1357810018-33
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: process_te_message: Processing (N)ACK lrm_invoke-lrmd-1357810018-33 from dl380g5d
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: match_graph_event: Action dummy:1_monitor_20000 (2) confirmed on dl380g5d (rc=0)
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: te_pseudo_action: Pseudo action 20 fired and confirmed
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: process_lrm_event: LRM operation dummy:1_monitor_20000 (call=4, status=1, cib-update=0, confirmed=true) Cancelled
Jan 10 18:26:58 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 7: demote dummy:0_demote_0 on dl380g5c
Jan 10 18:26:58 dl380g5d pengine: [4297]: info: process_pe_message: Transition 4: PEngine Input stored in: /var/lib/pengine/pe-input-4.bz2
Jan 10 18:26:59 dl380g5d crmd: [4291]: WARN: status_from_rc: Action 7 (dummy:0_demote_0) on dl380g5c failed (target: 0 vs. rc: 1): Error
Jan 10 18:26:59 dl380g5d crmd: [4291]: WARN: update_failcount: Updating failcount for dummy:0 on dl380g5c after failed demote: rc=1 (update=value++, time=1357810019)
Jan 10 18:26:59 dl380g5d crmd: [4291]: info: abort_transition_graph: match_graph_event:299 - Triggered transition abort (complete=0, tag=lrm_rsc_op, id=dummy:0_demote_0, magic=0:1;7:4:0:f619d469-9fad-4ee5-acd6-20cb09c4099c, cib=0.8.3) : Event failed
Jan 10 18:26:59 dl380g5d crmd: [4291]: info: update_abort_priority: Abort priority upgraded from 0 to 1
Jan 10 18:26:59 dl380g5d crmd: [4291]: info: update_abort_priority: Abort action done superceeded by restart
Jan 10 18:26:59 dl380g5d crmd: [4291]: info: match_graph_event: Action dummy:0_demote_0 (7) confirmed on dl380g5c (rc=4)
Jan 10 18:26:59 dl380g5d crmd: [4291]: info: te_pseudo_action: Pseudo action 21 fired and confirmed
Jan 10 18:26:59 dl380g5d crmd: [4291]: info: run_graph: ====================================================
Jan 10 18:26:59 dl380g5d crmd: [4291]: notice: run_graph: Transition 4 (Complete=4, Pending=0, Fired=0, Skipped=6, Incomplete=2, Source=/var/lib/pengine/pe-input-4.bz2): Stopped
Jan 10 18:26:59 dl380g5d crmd: [4291]: info: te_graph_trigger: Transition 4 is now complete
Jan 10 18:27:00 dl380g5d ccm: [4286]: info: client (pid=4744) removed from ccm
Jan 10 18:27:01 dl380g5d attrd: [4290]: info: attrd_ha_callback: flush message from dl380g5c
Jan 10 18:27:01 dl380g5d attrd: [4290]: info: find_hash_entry: Creating hash entry for fail-count-dummy:0
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: abort_transition_graph: te_update_diff:150 - Triggered transition abort (complete=1, tag=nvpair, id=status-21c624bd-c426-43dc-9665-bbfb92054bcd-fail-count-dummy:0, name=fail-count-dummy:0, value=1, magic=NA, cib=0.8.4) : Transient attribute: update
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: abort_transition_graph: te_update_diff:150 - Triggered transition abort (complete=1, tag=nvpair, id=status-21c624bd-c426-43dc-9665-bbfb92054bcd-last-failure-dummy:0, name=last-failure-dummy:0, value=1357810019, magic=NA, cib=0.8.5) : Transient attribute: update
Jan 10 18:27:01 dl380g5d attrd: [4290]: info: attrd_ha_callback: flush message from dl380g5c
Jan 10 18:27:01 dl380g5d attrd: [4290]: info: find_hash_entry: Creating hash entry for last-failure-dummy:0
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped!
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: do_pe_invoke: Query 60: Requesting the current CIB: S_POLICY_ENGINE
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: do_pe_invoke_callback: Invoking the PE: query=60, ref=pe_calc-dc-1357810021-35, seq=2, quorate=1
Jan 10 18:27:01 dl380g5d pengine: [4297]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jan 10 18:27:01 dl380g5d pengine: [4297]: WARN: unpack_nodes: Blind faith: not fencing unseen nodes
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: unpack_status: Node dl380g5c is in standby-mode
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5c is standby
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5d is online
Jan 10 18:27:01 dl380g5d pengine: [4297]: notice: unpack_rsc_op: Operation dummy:0_monitor_0 found resource dummy:0 active in master mode on dl380g5c
Jan 10 18:27:01 dl380g5d pengine: [4297]: WARN: unpack_rsc_op: Processing failed op dummy:0_demote_0 on dl380g5c: unknown error (1)
Jan 10 18:27:01 dl380g5d pengine: [4297]: WARN: unpack_rsc_op: Forcing dummy:0 to stop after a failed demote action
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: native_add_running: resource dummy:0 isnt managed
Jan 10 18:27:01 dl380g5d pengine: [4297]: notice: clone_print:  Master/Slave Set: stateful
Jan 10 18:27:01 dl380g5d pengine: [4297]: notice: native_print:      dummy:0	(ocf::pacemaker:Stateful):	Slave dl380g5c (unmanaged) FAILED
Jan 10 18:27:01 dl380g5d pengine: [4297]: notice: short_print:      Slaves: [ dl380g5d ]
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: get_failcount: stateful has failed 1 times on dl380g5c
Jan 10 18:27:01 dl380g5d pengine: [4297]: WARN: common_apply_stickiness: Forcing stateful away from dl380g5c after 1 failures (max=1)
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: get_failcount: stateful has failed 1 times on dl380g5c
Jan 10 18:27:01 dl380g5d pengine: [4297]: WARN: common_apply_stickiness: Forcing stateful away from dl380g5c after 1 failures (max=1)
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: native_color: Unmanaged resource dummy:0 allocated to 'nowhere': failed
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: master_color: Promoting dummy:1 (Slave dl380g5d)
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: master_color: stateful: Promoted 1 instances of a possible 1 to master
Jan 10 18:27:01 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (10s) for dummy:1 on dl380g5d
Jan 10 18:27:01 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (10s) for dummy:1 on dl380g5d
Jan 10 18:27:01 dl380g5d pengine: [4297]: notice: LogActions: Leave   resource dummy:0	(Slave unmanaged)
Jan 10 18:27:01 dl380g5d pengine: [4297]: notice: LogActions: Promote dummy:1	(Slave -> Master dl380g5d)
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jan 10 18:27:01 dl380g5d crmd: [4291]: WARN: destroy_action: Cancelling timer for action 2 (src=77)
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: unpack_graph: Unpacked transition 5: 4 actions in 4 synapses
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: do_te_invoke: Processing graph 5 (ref=pe_calc-dc-1357810021-35) derived from /var/lib/pengine/pe-input-5.bz2
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: te_pseudo_action: Pseudo action 17 fired and confirmed
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 11: promote dummy:1_promote_0 on dl380g5d (local)
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: do_lrm_rsc_op: Performing key=11:5:0:f619d469-9fad-4ee5-acd6-20cb09c4099c op=dummy:1_promote_0 )
Jan 10 18:27:01 dl380g5d lrmd: [4288]: info: rsc:dummy:1 promote[5] (pid 4769)
Jan 10 18:27:01 dl380g5d pengine: [4297]: info: process_pe_message: Transition 5: PEngine Input stored in: /var/lib/pengine/pe-input-5.bz2
Jan 10 18:27:01 dl380g5d crm_attribute: [4779]: info: Invoked: crm_attribute -N dl380g5d -n master-dummy:1 -l reboot -v 10 
Jan 10 18:27:01 dl380g5d attrd: [4290]: info: attrd_trigger_update: Sending flush op to all hosts for: master-dummy:1 (10)
Jan 10 18:27:01 dl380g5d attrd: [4290]: info: attrd_perform_update: Sent update 20: master-dummy:1=10
Jan 10 18:27:01 dl380g5d lrmd: [4288]: info: operation promote[5] on dummy:1 for client 4291: pid 4769 exited with return code 0
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: process_lrm_event: LRM operation dummy:1_promote_0 (call=5, rc=0, cib-update=61, confirmed=true) ok
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: abort_transition_graph: te_update_diff:150 - Triggered transition abort (complete=0, tag=nvpair, id=status-3f6ec88d-ee47-4f63-bfeb-652b8dd96027-master-dummy:1, name=NA, value=10, magic=NA, cib=0.8.6) : Transient attribute: update
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: update_abort_priority: Abort priority upgraded from 0 to 1000000
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: update_abort_priority: Abort action done superceeded by restart
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: match_graph_event: Action dummy:1_promote_0 (11) confirmed on dl380g5d (rc=0)
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: te_pseudo_action: Pseudo action 18 fired and confirmed
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: run_graph: ====================================================
Jan 10 18:27:01 dl380g5d crmd: [4291]: notice: run_graph: Transition 5 (Complete=3, Pending=0, Fired=0, Skipped=1, Incomplete=0, Source=/var/lib/pengine/pe-input-5.bz2): Stopped
Jan 10 18:27:01 dl380g5d crmd: [4291]: info: te_graph_trigger: Transition 5 is now complete
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped!
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_pe_invoke: Query 62: Requesting the current CIB: S_POLICY_ENGINE
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_pe_invoke_callback: Invoking the PE: query=62, ref=pe_calc-dc-1357810023-37, seq=2, quorate=1
Jan 10 18:27:03 dl380g5d pengine: [4297]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jan 10 18:27:03 dl380g5d pengine: [4297]: WARN: unpack_nodes: Blind faith: not fencing unseen nodes
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: unpack_status: Node dl380g5c is in standby-mode
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5c is standby
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: determine_online_status: Node dl380g5d is online
Jan 10 18:27:03 dl380g5d pengine: [4297]: notice: unpack_rsc_op: Operation dummy:0_monitor_0 found resource dummy:0 active in master mode on dl380g5c
Jan 10 18:27:03 dl380g5d pengine: [4297]: WARN: unpack_rsc_op: Processing failed op dummy:0_demote_0 on dl380g5c: unknown error (1)
Jan 10 18:27:03 dl380g5d pengine: [4297]: WARN: unpack_rsc_op: Forcing dummy:0 to stop after a failed demote action
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: native_add_running: resource dummy:0 isnt managed
Jan 10 18:27:03 dl380g5d pengine: [4297]: notice: clone_print:  Master/Slave Set: stateful
Jan 10 18:27:03 dl380g5d pengine: [4297]: notice: native_print:      dummy:0	(ocf::pacemaker:Stateful):	Slave dl380g5c (unmanaged) FAILED
Jan 10 18:27:03 dl380g5d pengine: [4297]: notice: short_print:      Masters: [ dl380g5d ]
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: get_failcount: stateful has failed 1 times on dl380g5c
Jan 10 18:27:03 dl380g5d pengine: [4297]: WARN: common_apply_stickiness: Forcing stateful away from dl380g5c after 1 failures (max=1)
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: get_failcount: stateful has failed 1 times on dl380g5c
Jan 10 18:27:03 dl380g5d pengine: [4297]: WARN: common_apply_stickiness: Forcing stateful away from dl380g5c after 1 failures (max=1)
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: native_color: Unmanaged resource dummy:0 allocated to 'nowhere': failed
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: master_color: Promoting dummy:1 (Master dl380g5d)
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: master_color: stateful: Promoted 1 instances of a possible 1 to master
Jan 10 18:27:03 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (10s) for dummy:1 on dl380g5d
Jan 10 18:27:03 dl380g5d pengine: [4297]: notice: RecurringOp:  Start recurring monitor (10s) for dummy:1 on dl380g5d
Jan 10 18:27:03 dl380g5d pengine: [4297]: notice: LogActions: Leave   resource dummy:0	(Slave unmanaged)
Jan 10 18:27:03 dl380g5d pengine: [4297]: notice: LogActions: Leave   resource dummy:1	(Master dl380g5d)
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: unpack_graph: Unpacked transition 6: 1 actions in 1 synapses
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_te_invoke: Processing graph 6 (ref=pe_calc-dc-1357810023-37) derived from /var/lib/pengine/pe-input-6.bz2
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: te_rsc_command: Initiating action 12: monitor dummy:1_monitor_10000 on dl380g5d (local)
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_lrm_rsc_op: Performing key=12:6:8:f619d469-9fad-4ee5-acd6-20cb09c4099c op=dummy:1_monitor_10000 )
Jan 10 18:27:03 dl380g5d lrmd: [4288]: info: rsc:dummy:1 monitor[6] (pid 4780)
Jan 10 18:27:03 dl380g5d pengine: [4297]: info: process_pe_message: Transition 6: PEngine Input stored in: /var/lib/pengine/pe-input-6.bz2
Jan 10 18:27:03 dl380g5d lrmd: [4288]: info: operation monitor[6] on dummy:1 for client 4291: pid 4780 exited with return code 8
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: process_lrm_event: LRM operation dummy:1_monitor_10000 (call=6, rc=8, cib-update=63, confirmed=false) master
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: match_graph_event: Action dummy:1_monitor_10000 (12) confirmed on dl380g5d (rc=0)
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: run_graph: ====================================================
Jan 10 18:27:03 dl380g5d crmd: [4291]: notice: run_graph: Transition 6 (Complete=1, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pengine/pe-input-6.bz2): Complete
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: te_graph_trigger: Transition 6 is now complete
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: notify_crmd: Transition 6 status: done - <null>
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Jan 10 18:27:03 dl380g5d crmd: [4291]: info: do_state_transition: Starting PEngine Recheck Timer
Jan 10 18:27:55 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:27:57 dl380g5d root: Mark:HB_REPORT:1357810074
Jan 10 18:27:59 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:27:59 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:28:00 dl380g5d ccm: [4286]: info: client (pid=5047) removed from ccm
Jan 10 18:28:00 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:28:00 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:29:00 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:29:00 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:29:01 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:29:01 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:30:00 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:30:00 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:30:02 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:30:03 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:31:00 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:31:00 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:31:02 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:31:03 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:31:04 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:31:05 dl380g5d root: Mark:HB_REPORT:1357810264
Jan 10 18:31:06 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:31:07 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:31:07 dl380g5d ccm: [4286]: info: client (pid=7008) removed from ccm
Jan 10 18:31:07 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:31:08 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:32:09 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:32:09 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:32:10 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:32:11 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:32:52 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:32:53 dl380g5d root: Mark:HB_REPORT:1357810372
Jan 10 18:32:54 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:32:55 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:32:56 dl380g5d ccm: [4286]: info: client (pid=8850) removed from ccm
Jan 10 18:32:56 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:32:56 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:33:29 dl380g5d cib: [4287]: info: cib_stats: Processed 214 operations (1261.00us average, 0% utilization) in the last 10min
Jan 10 18:34:27 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:34:27 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:34:30 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:34:30 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
Jan 10 18:34:48 dl380g5d crmd: [4291]: info: handle_request: Current ping state: S_IDLE
