[Pacemaker] problem with pacemaker and stonith resources
Matteo
motogintom at gmail.com
Tue May 25 22:37:51 UTC 2010
Hi folks,
I'm running a 2 node cluster with pacemaker, DRBD dual primary, ocfs2.
Now I'm trying to setup stonith correctly, but my stonith resources don't
start. I did some research but I didn't find a solution to my problem.
This is my cib:
node server1
node server2
primitive DLM ocf:pacemaker:controld \
op monitor interval="120s"
primitive DRBD ocf:linbit:drbd \
params drbd_resource="r0" \
operations $id="DRBD-operations" \
op monitor interval="20" role="Master" timeout="20" \
op monitor interval="30" role="Slave" timeout="20"
primitive FS ocf:heartbeat:Filesystem \
params device="/dev/drbd1" directory="/drbd" fstype="ocfs2" \
op monitor interval="120s" \
meta target-role="Started"
primitive O2CB ocf:pacemaker:o2cb \
op monitor interval="120s"
primitive STONITH1 stonith:external/ipmi \
params hostname="server1" ipaddr="10.0.0.1" userid="user" passwd="user"
interface="lan" \
meta target-role="Started"
primitive STONITH2 stonith:external/ipmi \
params hostname="server2" ipaddr="10.0.0.2" userid="user" passwd="user"
interface="lan" \
meta target-role="Started"
ms ms-DRBD DRBD \
meta resource-stickines="100" notify="true" master-max="2" interleave="true"
target-role="Stopped"
clone cloneDLM DLM \
meta globally-unique="false" interleave="true" target-role="Started"
clone cloneFS FS \
meta interleave="true" ordered="true"
clone cloneO2CB O2CB \
meta globally-unique="false" interleave="true" target-role="Started"
location loc-stonith1 STONITH1 -inf: server1
location loc-stonith2 STONITH2 -inf: server2
colocation DLM-DRBD inf: cloneDLM ms-DRBD:Master
colocation FS-O2CB inf: cloneFS cloneO2CB
colocation O2CB-DLM inf: cloneO2CB cloneDLM
order DLM-before-O2CB inf: cloneDLM:start cloneO2CB:start
order DRBD-before-DLM inf: ms-DRBD:promote cloneDLM:start
order O2CB-before-FS inf: cloneO2CB:start cloneFS:start
property $id="cib-bootstrap-options" \
dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \
cluster-infrastructure="openais" \
expected-quorum-votes="2" \
no-quorum-policy="ignore" \
stonith-enabled="true" \
stonith-action="poweroff" \
default-resource-stickiness="1000"
Is there something wrong?
-----------------------------------
crm_mon -n:
Last updated: Wed May 26 00:04:53 2010
Stack: openais
Current DC: server1 - partition with quorum
Version: 1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd
2 Nodes configured, 2 expected votes
6 Resources configured.
============
Node server2: online
DLM:0 (ocf::pacemaker:controld) Started
O2CB:0 (ocf::pacemaker:o2cb) Started
FS:0 (ocf::heartbeat:Filesystem) Started
DRBD:0 (ocf::linbit:drbd) Master
Node server1: online
DRBD:1 (ocf::linbit:drbd) Master
DLM:1 (ocf::pacemaker:controld) Started
O2CB:1 (ocf::pacemaker:o2cb) Started
FS:1 (ocf::heartbeat:Filesystem) Started
Failed actions:
STONITH2_start_0 (node=server1, call=8, rc=1, status=complete): unknown
erro r
STONITH1_start_0 (node=server2, call=8, rc=1, status=complete): unknown
erro r
------------------------
crm_verify -L -V:
crm_verify[5695]: 2010/05/26_00:17:19 WARN: unpack_rsc_op: Processing failed
op STONITH2_start_0 on server1: unknown error(1)
crm_verify[5695]: 2010/05/26_00:17:19 WARN: unpack_rsc_op: Processing failed
op STONITH1_start_0 on server2: unknown error(1)
crm_verify[5695]: 2010/05/26_00:17:19 WARN: common_apply_stickiness: Forcing
STONITH1 away from server2 after 1000000 failures (max=1000000)
crm_verify[5695]: 2010/05/26_00:17:19 WARN: common_apply_stickiness: Forcing
STONITH2 away from server1 after 1000000 failures (max=1000000)
I hope someone can help me,
Thank you!
Matt
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20100526/a714ed9b/attachment-0001.html>
More information about the Pacemaker
mailing list