[ClusterLabs] Issue with Stonith Resource parameters
vija ar
vjav78 at gmail.com
Wed Mar 9 03:19:45 EST 2016
here is the config ....
<cib epoch="20" num_updates="0" admin_epoch="0"
validate-with="pacemaker-1.2" cib-last-written="Wed Mar 9 00:56:57 2016"
update-origin="server02" update-client="cibadmin" update-user="hacluster"
crm_feature_set="3.0.8" have-quorum="1" dc-uuid="server01">
<configuration>
<crm_config>
<cluster_property_set id="cib-bootstrap-options">
<nvpair name="stonith-enabled" value="true"
id="cib-bootstrap-options-stonith-enabled"/>
<nvpair name="no-quorum-policy" value="ignore"
id="cib-bootstrap-options-no-quorum-policy"/>
<nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
value="1.1.11-3ca8c3b"/>
<nvpair id="cib-bootstrap-options-cluster-infrastructure"
name="cluster-infrastructure" value="classic openais (with plugin)"/>
<nvpair id="cib-bootstrap-options-expected-quorum-votes"
name="expected-quorum-votes" value="2"/>
<nvpair name="stonith-action" value="reboot"
id="cib-bootstrap-options-stonith-action"/>
<nvpair name="stonith-timeout" value="150s"
id="cib-bootstrap-options-stonith-timeout"/>
</cluster_property_set>
</crm_config>
<nodes>
<node id="server02" uname="server02"/>
<node id="server01" uname="server01"/>
</nodes>
<resources>
<primitive id="STONITH-server01" class="stonith" type="external/ipmi">
<operations>
<op name="monitor" interval="0" timeout="60s"
id="STONITH-server01-monitor-0"/>
<op name="monitor" interval="300s" timeout="60s"
on-fail="restart" id="STONITH-server01-monitor-300s"/>
<op name="start" interval="0" timeout="60s" on-fail="restart"
id="STONITH-server01-start-0"/>
</operations>
<instance_attributes id="STONITH-server01-instance_attributes">
<nvpair name="hostname" value="server01"
id="STONITH-server01-instance_attributes-hostname"/>
<nvpair name="ipaddr" value="server01-ipmi"
id="STONITH-server01-instance_attributes-ipaddr"/>
<nvpair name="userid" value="administrator"
id="STONITH-server01-instance_attributes-userid"/>
<nvpair name="passwd" value="To12"
id="STONITH-server01-instance_attributes-passwd"/>
<nvpair name="interface" value="lanplus"
id="STONITH-server01-instance_attributes-interface"/>
</instance_attributes>
</primitive>
<primitive id="STONITH-server02" class="stonith" type="external/ipmi">
<operations>
<op name="monitor" interval="0" timeout="60s"
id="STONITH-server02-monitor-0"/>
<op name="monitor" interval="300s" timeout="60s"
on-fail="restart" id="STONITH-server02-monitor-300s"/>
<op name="start" interval="0" timeout="60s" on-fail="restart"
id="STONITH-server02-start-0"/>
</operations>
<instance_attributes id="STONITH-server02-instance_attributes">
<nvpair name="hostname" value="server02"
id="STONITH-server02-instance_attributes-hostname"/>
<nvpair name="ipaddr" value="server02-ipmi"
id="STONITH-server02-instance_attributes-ipaddr"/>
<nvpair name="userid" value="administrator"
id="STONITH-server02-instance_attributes-userid"/>
<nvpair name="passwd" value="To12"
id="STONITH-server02-instance_attributes-passwd"/>
<nvpair name="interface" value="lanplus"
id="STONITH-server02-instance_attributes-interface"/>
</instance_attributes>
</primitive>
<primitive id="VIRTUAL-IP" class="ocf" provider="heartbeat"
type="IPaddr2">
<instance_attributes id="VIRTUAL-IP-instance_attributes">
<nvpair name="ip" value="10.0.0.44"
id="VIRTUAL-IP-instance_attributes-ip"/>
</instance_attributes>
<operations>
<op name="monitor" timeout="20s" interval="10s"
id="VIRTUAL-IP-monitor-10s"/>
</operations>
<meta_attributes id="VIRTUAL-IP-meta_attributes">
<nvpair name="is-managed" value="true"
id="VIRTUAL-IP-meta_attributes-is-managed"/>
<nvpair name="target-role" value="Started"
id="VIRTUAL-IP-meta_attributes-target-role"/>
</meta_attributes>
</primitive>
</resources>
<constraints>
<rsc_location id="LOC_STONITH_server01" rsc="STONITH-server01"
score="INFINITY" node="server02"/>
<rsc_location id="LOC_STONITH_server02" rsc="STONITH-server02"
score="INFINITY" node="server01"/>
</constraints>
<rsc_defaults>
<meta_attributes id="rsc-options">
<nvpair name="migration-threshold" value="5000"
id="rsc-options-migration-threshold"/>
<nvpair name="resource-stickiness" value="1000"
id="rsc-options-resource-stickiness"/>
</meta_attributes>
</rsc_defaults>
<op_defaults>
<meta_attributes id="op-options">
<nvpair name="timeout" value="600" id="op-options-timeout"/>
<nvpair name="record-pending" value="false"
id="op-options-record-pending"/>
</meta_attributes>
</op_defaults>
</configuration>
</cib>
On Wed, Mar 9, 2016 at 1:25 PM, emmanuel segura <emi2fast at gmail.com> wrote:
> I think you should give the parameters to the stonith agent, anyway
> show your config.
>
> 2016-03-09 5:29 GMT+01:00 vija ar <vjav78 at gmail.com>:
> > I have configured SLEHA cluster on cisco ucs boxes with ipmi configured,
> i
> > have tested IPMI using impitool, however ipmitool to function neatly i
> have
> > to pass parameter -y i.e. <hex key> along with username and password,
> >
> > however to configure stonith there is no parameter in pacemaker to pass
> > <hex key>? and due to which stonith is failing
> >
> > can you please let me know if there is any way to add it or is this a
> bug?
> >
> > *******************
> >
> >
> >
> > Mar 9 00:26:28 server02 stonith: external_status: 'ipmi status' failed
> with
> > rc 1
> > Mar 9 00:26:28 server02 stonith: external/ipmi device not accessible.
> > Mar 9 00:26:28 server02 stonith-ng[99114]: notice: log_operation:
> > Operation 'monitor' [99200] for device 'STONITH-server02' returned: -201
> > (Generic Pacemaker error)
> > Mar 9 00:26:28 server02 stonith-ng[99114]: warning: log_operation:
> > STONITH-server02:99200 [ Performing: stonith -t external/ipmi -S ]
> > Mar 9 00:26:28 server02 stonith-ng[99114]: warning: log_operation:
> > STONITH-server02:99200 [ logd is not runningfailed: 1 ]
> > Mar 9 00:26:28 server02 crmd[99118]: error: process_lrm_event: LRM
> > operation STONITH-server02_start_0 (call=13, status=4, cib-update=13,
> > confirmed=true) Error
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update:
> Sent
> > update 35: fail-count-STONITH-server02=INFINITY
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server02
> > (1457463388)
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update:
> Sent
> > update 37: last-failure-STONITH-server02=1457463388
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update:
> Sent
> > update 39: fail-count-STONITH-server02=INFINITY
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server02
> > (1457463388)
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update:
> Sent
> > update 41: last-failure-STONITH-server02=1457463388
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update:
> Sent
> > update 43: fail-count-STONITH-server02=INFINITY
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server02
> > (1457463388)
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update:
> Sent
> > update 45: last-failure-STONITH-server02=1457463388
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update:
> Sent
> > update 47: fail-count-STONITH-server02=INFINITY
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server02
> > (1457463388)
> > Mar 9 00:26:28 server02 attrd[99116]: notice: attrd_perform_update:
> Sent
> > update 49: last-failure-STONITH-server02=1457463388
> > Mar 9 00:26:28 server02 crmd[99118]: notice: process_lrm_event: LRM
> > operation STONITH-server02_stop_0 (call=14, rc=0, cib-update=14,
> > confirmed=true) ok
> > Mar 9 00:26:28 server01 crmd[16809]: warning: status_from_rc: Action 9
> > (STONITH-server02_start_0) on server02 failed (target: 0 vs. rc: 1):
> Error
> > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount:
> Updating
> > failcount for STONITH-server02 on server02 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount:
> Updating
> > failcount for STONITH-server02 on server02 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar 9 00:26:28 server01 crmd[16809]: warning: status_from_rc: Action 9
> > (STONITH-server02_start_0) on server02 failed (target: 0 vs. rc: 1):
> Error
> > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount:
> Updating
> > failcount for STONITH-server02 on server02 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount:
> Updating
> > failcount for STONITH-server02 on server02 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar 9 00:26:28 server01 stonith: external_status: 'ipmi status' failed
> with
> > rc 1
> > Mar 9 00:26:28 server01 stonith: external/ipmi device not accessible.
> > Mar 9 00:26:28 server01 stonith-ng[16805]: notice: log_operation:
> > Operation 'monitor' [16891] for device 'STONITH-server01' returned: -201
> > (Generic Pacemaker error)
> > Mar 9 00:26:28 server01 stonith-ng[16805]: warning: log_operation:
> > STONITH-server01:16891 [ Performing: stonith -t external/ipmi -S ]
> > Mar 9 00:26:28 server01 stonith-ng[16805]: warning: log_operation:
> > STONITH-server01:16891 [ logd is not runningfailed: 1 ]
> > Mar 9 00:26:28 server01 crmd[16809]: error: process_lrm_event: LRM
> > operation STONITH-server01_start_0 (call=13, status=4, cib-update=49,
> > confirmed=true) Error
> > Mar 9 00:26:28 server01 crmd[16809]: warning: status_from_rc: Action 7
> > (STONITH-server01_start_0) on server01 failed (target: 0 vs. rc: 1):
> Error
> > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount:
> Updating
> > failcount for STONITH-server01 on server01 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY)
> > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount:
> Updating
> > failcount for STONITH-server01 on server01 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar 9 00:26:28 server01 crmd[16809]: warning: status_from_rc: Action 7
> > (STONITH-server01_start_0) on server01 failed (target: 0 vs. rc: 1):
> Error
> > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount:
> Updating
> > failcount for STONITH-server01 on server01 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar 9 00:26:28 server01 crmd[16809]: warning: update_failcount:
> Updating
> > failcount for STONITH-server01 on server01 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update:
> Sent
> > update 47: fail-count-STONITH-server01=INFINITY
> > Mar 9 00:26:28 server01 crmd[16809]: notice: run_graph: Transition 3
> > (Complete=5, Pending=0, Fired=0, Skipped=2, Incomplete=0,
> > Source=/var/lib/pacemaker/pengine/pe-input-70.bz2): Stopped
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server01
> > (1457463388)
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update:
> Sent
> > update 49: last-failure-STONITH-server01=1457463388
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: unpack_config: On
> loss of
> > CCM Quorum: Ignore
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server01: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server02: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server01 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server02 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server02 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Recover
> > STONITH-server01 (Started server01)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop
> > STONITH-server02 (server02)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: process_pe_message:
> > Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-71.bz2
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update:
> Sent
> > update 51: fail-count-STONITH-server01=INFINITY
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server01
> > (1457463388)
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update:
> Sent
> > update 53: last-failure-STONITH-server01=1457463388
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: unpack_config: On
> loss of
> > CCM Quorum: Ignore
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server01: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server02: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server01 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server01 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server02 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server02 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop
> > STONITH-server01 (server01)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop
> > STONITH-server02 (server02)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: process_pe_message:
> > Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-72.bz2
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update:
> Sent
> > update 55: fail-count-STONITH-server01=INFINITY
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server01
> > (1457463388)
> > Mar 9 00:26:28 server01 attrd[16807]: notice: attrd_perform_update:
> Sent
> > update 57: last-failure-STONITH-server01=1457463388
> > Mar 9 00:26:28 server01 pengine[16808]: notice: unpack_config: On
> loss of
> > CCM Quorum: Ignore
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server01: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server02: unknown
> error
> > (1)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server01 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server01 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server02 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server02 after 1000000 failures
> (max=3)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop
> > STONITH-server01 (server01)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: LogActions: Stop
> > STONITH-server02 (server02)
> > Mar 9 00:26:28 server01 pengine[16808]: notice: process_pe_message:
> > Calculated Transition 6: /var/lib/pacemaker/pengine/pe-input-73.bz2
> > Mar 9 00:26:28 server01 crmd[16809]: notice: do_te_invoke: Processing
> > graph 6 (ref=pe_calc-dc-1457463388-32) derived from
> > /var/lib/pacemaker/pengine/pe-input-73.bz2
> > Mar 9 00:26:28 server01 crmd[16809]: notice: te_rsc_command:
> Initiating
> > action 1: stop STONITH-server01_stop_0 on server01 (local)
> > Mar 9 00:26:28 server01 crmd[16809]: notice: te_rsc_command:
> Initiating
> > action 2: stop STONITH-server02_stop_0 on server02
> > Mar 9 00:26:28 server01 crmd[16809]: notice: process_lrm_event: LRM
> > operation STONITH-server01_stop_0 (call=14, rc=0, cib-update=55,
> > confirmed=true) ok
> > Mar 9 00:26:28 server01 crmd[16809]: notice: run_graph: Transition 6
> > (Complete=3, Pending=0, Fired=0, Skipped=0, Incomplete=0,
> > Source=/var/lib/pacemaker/pengine/pe-input-73.bz2): Complete
> > Mar 9 00:26:28 server01 crmd[16809]: notice: do_state_transition:
> State
> > transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
> > cause=C_FSA_INTERNAL origin=notify_crmd ]
> >
> >
> >
> > _______________________________________________
> > Users mailing list: Users at clusterlabs.org
> > http://clusterlabs.org/mailman/listinfo/users
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> >
>
>
>
> --
> .~.
> /V\
> // \\
> /( )\
> ^`~'^
>
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20160309/14c15c92/attachment-0003.html>
More information about the Users
mailing list