[Pacemaker] STONITH is not performed after stonithd reboots
Andrew Beekhof
andrew at beekhof.net
Wed May 9 07:11:55 UTC 2012
On Mon, May 7, 2012 at 7:52 PM, Kazunori INOUE
<inouekazu at intellilink.co.jp> wrote:
> Hi,
>
> On the Pacemkaer-1.1 + Corosync stack, although stonithd reboots
> after an abnormal end, STONITH is not performed after that.
>
> I am using the newest devel.
> - pacemaker : db5e16736cc2682fbf37f81cd47be7d17d5a2364
> - corosync : 88dd3e1eeacd64701d665f10acbc40f3795dd32f
> - glue : 2686:66d5f0c135c9
>
>
> * 0. cluster's state.
>
> [root at vm1 ~]# crm_mon -r1
> ============
> Last updated: Wed May 2 16:07:29 2012
> Last change: Wed May 2 16:06:35 2012 via cibadmin on vm1
> Stack: corosync
> Current DC: vm1 (1) - partition WITHOUT quorum
> Version: 1.1.7-db5e167
> 2 Nodes configured, unknown expected votes
> 3 Resources configured.
> ============
>
> Online: [ vm1 vm2 ]
>
> Full list of resources:
>
> prmDummy (ocf::pacemaker:Dummy): Started vm2
> prmStonith1 (stonith:external/libvirt): Started vm2
> prmStonith2 (stonith:external/libvirt): Started vm1
>
> [root at vm1 ~]# crm configure show
> node $id="1" vm1
> node $id="2" vm2
> primitive prmDummy ocf:pacemaker:Dummy \
> op start interval="0s" timeout="60s" on-fail="restart" \
> op monitor interval="10s" timeout="60s" on-fail="fence" \
> op stop interval="0s" timeout="60s" on-fail="stop"
> primitive prmStonith1 stonith:external/libvirt \
> params hostlist="vm1" hypervisor_uri="qemu+ssh://f/system" \
> op start interval="0s" timeout="60s" \
> op monitor interval="3600s" timeout="60s" \
> op stop interval="0s" timeout="60s"
> primitive prmStonith2 stonith:external/libvirt \
> params hostlist="vm2" hypervisor_uri="qemu+ssh://g/system" \
> op start interval="0s" timeout="60s" \
> op monitor interval="3600s" timeout="60s" \
> op stop interval="0s" timeout="60s"
> location rsc_location-prmDummy prmDummy \
> rule $id="rsc_location-prmDummy-rule" 200: #uname eq vm2
> location rsc_location-prmStonith1 prmStonith1 \
> rule $id="rsc_location-prmStonith1-rule" 200: #uname eq vm2 \
> rule $id="rsc_location-prmStonith1-rule-0" -inf: #uname eq vm1
> location rsc_location-prmStonith2 prmStonith2 \
> rule $id="rsc_location-prmStonith2-rule" 200: #uname eq vm1 \
> rule $id="rsc_location-prmStonith2-rule-0" -inf: #uname eq vm2
> property $id="cib-bootstrap-options" \
> dc-version="1.1.7-db5e167" \
> cluster-infrastructure="corosync" \
> no-quorum-policy="ignore" \
> stonith-enabled="true" \
> startup-fencing="false" \
> stonith-timeout="120s"
> rsc_defaults $id="rsc-options" \
> resource-stickiness="INFINITY" \
> migration-threshold="1"
>
>
> * 1. terminate stonithd forcibly.
>
> [root at vm1 ~]# pkill -9 stonithd
>
>
> * 2. I cause STONITH, but stonithd says that a device is not found and
> does not STONITH.
>
> [root at vm1 ~]# ssh vm2 'rm /var/run/Dummy-prmDummy.state'
> [root at vm1 ~]# grep Found /var/log/ha-debug
> May 2 16:13:07 vm1 stonith-ng[15115]: debug: stonith_query: Found 0 matching devices for 'vm2'
> May 2 16:13:19 vm1 stonith-ng[15115]: debug: stonith_query: Found 0 matching devices for 'vm2'
> May 2 16:13:31 vm1 stonith-ng[15115]: debug: stonith_query: Found 0 matching devices for 'vm2'
> May 2 16:13:43 vm1 stonith-ng[15115]: debug: stonith_query: Found 0 matching devices for 'vm2'
> (snip)
> [root at vm1 ~]#
>
>
> After stonithd reboots, it seems that STONITH-resource or lrmd needs
> to be rebooted.. is this the designed behavior?
No, that sounds like a bug.
>
> # crm resource restart <STONITH resource (prmStonith2)>
> or
> # /usr/lib64/heartbeat/lrmd -r (on the node which stonithd rebooted)
>
> ----
> Best regards,
> Kazunori INOUE
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
More information about the Pacemaker
mailing list