[Pacemaker] question about stonith:external/libvirt
Matthew O'Connor
matt at ecsorl.com
Sun May 20 06:40:20 CEST 2012
After using the tutorial on the Hastexo site for setting up stonith via
libvirt, I believe I have it working correctly...but...some strange
things are happening. I have two nodes, with shared storage provided by
a dual-primary DRBD resource and OCFS2. Here is one of my stonith
primitives:
primitive p_fence-l2 stonith:external/libvirt \
params hostlist="l2:l2.sandbox"
hypervisor_uri="qemu+ssh://matt@hv01/system" stonith-timeout="30"
pcmk_host_check="none" \
op start interval="0" timeout="15" \
op stop interval="0" timeout="15" \
op monitor interval="60" \
meta target-role="Started"
This cluster has stonith-enabled="true" in the cluster options, plus the
necessary location statements in the cib.
To watch the DLM, I run dbench on the shared storage on the node I let
live. While it's running, I creatively nuke the other node. If I just
"killall pacemakerd" on l2 for instance, the DLM seems unaffected and
the fence takes place, rebooting the now "failed" node l2. No real
interruption of service on the surviving node, l3. Yet, if I "halt -f
-n" on l2, the fence still takes place but the surviving node's (l3's)
DLM hangs and won't come back until I bring the failed node back
online. Note that l2 and l3 can be interchanged - the results are the
same. Note that when the DLM is hung as in the latter case, eventually
kernel messages about hung tasks start populating the syslog.
I thought I had recently read some posts concerning this very topic, but
for the life of me I can't find them...
Any ideas on how I should proceed, or what I should look for next?
Thanks!
-- Matt
More information about the Pacemaker
mailing list