[Pacemaker] Pacemaker 0.7.3: can't create resources
Bruno Voigt
Bruno.Voigt at ic3s.de
Wed Sep 24 13:56:43 UTC 2008
Hi Andrew,
thanks a lot.
I just reread the PDF and using xsltproc I just got my test resource
hbtest1a configured.
But the CRM tells me that it could not be started on either of my both
Nodes (running Xen+DRBD).
(Under the debian heartbeat it worked as expected.)
It tells me:
Sep 24 15:20:55 xen20a crm_verify: [7184]: info: main: =#=#=#=#= Getting
XML =#=#=#=#=
Sep 24 15:20:55 xen20a crm_verify: [7184]: info: main: Reading XML from:
live cluster
Sep 24 15:20:55 xen20a crm_verify: [7184]: notice: main: Required
feature set: 2.0
Sep 24 15:20:55 xen20a crm_verify: [7184]: WARN: main: Your
configuration was internally updated to the latest version (pacemaker-1.0)
Sep 24 15:20:55 xen20a crm_verify: [7184]: WARN: unpack_resources: No
STONITH resources have been defined
Sep 24 15:20:55 xen20a crm_verify: [7184]: info:
determine_online_status: Node xen20b.test.mytld.com is online
Sep 24 15:20:55 xen20a crm_verify: [7184]: info: unpack_rsc_op:
Remapping hbtest1a_start_0 (rc=5) on xen20b.test.mytld.com to an ERROR
(expected 0)
Sep 24 15:20:55 xen20a crm_verify: [7184]: ERROR: unpack_rsc_op: Hard
error - hbtest1a_start_0 failed with rc=5: Preventing hbtest1a from
re-starting on xen20b.test.mytld.com
Sep 24 15:20:55 xen20a crm_verify: [7184]: WARN: unpack_rsc_op:
Processing failed op hbtest1a_start_0 on xen20b.test.mytld.com: Error
Sep 24 15:20:55 xen20a crm_verify: [7184]: WARN: unpack_rsc_op:
Compatability handling for failed op hbtest1a_start_0 on
xen20b.test.mytld.com
Sep 24 15:20:55 xen20a crm_verify: [7184]: info:
determine_online_status: Node xen20a.test.mytld.com is online
Sep 24 15:20:55 xen20a crm_verify: [7184]: info: unpack_rsc_op:
Remapping hbtest1a_start_0 (rc=5) on xen20a.test.mytld.com to an ERROR
(expected 0)
Sep 24 15:20:55 xen20a crm_verify: [7184]: ERROR: unpack_rsc_op: Hard
error - hbtest1a_start_0 failed with rc=5: Preventing hbtest1a from
re-starting on xen20a.test.mytld.com
Sep 24 15:20:55 xen20a crm_verify: [7184]: WARN: unpack_rsc_op:
Processing failed op hbtest1a_start_0 on xen20a.test.mytld.com: Error
Sep 24 15:20:55 xen20a crm_verify: [7184]: WARN: unpack_rsc_op:
Compatability handling for failed op hbtest1a_start_0 on
xen20a.test.mytld.com
Sep 24 15:20:55 xen20a crm_verify: [7184]: info: get_failcount: hbtest1a
has failed 1000000 times on xen20b.test.mytld.com
Sep 24 15:20:55 xen20a crm_verify: [7184]: info: get_failcount: hbtest1a
has failed 1000000 times on xen20a.test.mytld.com
Sep 24 15:20:55 xen20a crm_verify: [7184]: WARN: native_color: Resource
hbtest1a cannot run anywhere
Perhaps you could me give me a hint how to find out,
why the resource could not be started.
Thanks a lot,
Bruno
cibadmin -Q yields:
<cib epoch="36" num_updates="7" admin_epoch="0"
validate-with="pacemaker-1.0" have-quorum="1" crm_feature_set="3.0"
dc-uuid="278bebc6-2a59-4fa9-be2f-f6e262ce8936">
<configuration>
<crm_config>
<cluster_property_set id="cib-bootstrap-options">
<nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
value="1.0.0-rc1-node: 9b1e9d2785edf8eadd60e1e89f0ecacf67b2bc8c"/>
<nvpair id="cib-bootstrap-options-last-lrm-refresh"
name="last-lrm-refresh" value="1222261375"/>
</cluster_property_set>
</crm_config>
<nodes>
<node id="278bebc6-2a59-4fa9-be2f-f6e262ce8936"
uname="xen20b.test.mytld.com" type="normal"/>
<node id="d06bdc02-b705-4bf9-9cd4-6f827ffdfe2e"
uname="xen20a.test.mytld.com" type="normal"/>
</nodes>
<resources>
<primitive id="hbtest1a" class="ocf" type="Xen" provider="heartbeat">
<operations>
<op name="start" interval="0" id="hbtest1a-op01" timeout="60s"
start-delay="0"/>
<op name="stop" interval="0" id="hbtest1a-op02" timeout="300s"/>
<op name="monitor" interval="30s" id="hbtest1a-op03"
timeout="60s" start-delay="300s" requires="nothing">
<instance_attributes id="instance_attributes.hbtest1a.op03">
<nvpair id="hbtest1a-op03-inst-01" name="check_xyz"
value="something"/>
</instance_attributes>
</op>
</operations>
<instance_attributes id="instance_attributes.hbtest1a">
<nvpair id="hbtest1a.ia03" name="monitor_scripts"
value="/opt/gucky/monitor/monitor_ssh"/>
</instance_attributes>
<meta_attributes id="meta_attributes.hbtest1a">
<nvpair id="hbtest1a.me01" name="is-managed" value="true"/>
<nvpair id="hbtest1a.me02" name="allow-migrate" value="true"/>
<nvpair id="hbtest1a.me03" name="target-role" value="started"/>
</meta_attributes>
<instance_attributes id="hbtest1a-instance_attributes"/>
</primitive>
</resources>
<constraints/>
</configuration>
<status>
<node_state id="278bebc6-2a59-4fa9-be2f-f6e262ce8936"
uname="xen20b.test.mytld.com" ha="active" in_ccm="true" crmd="online"
join="member" expected="member" crm-debug-origin="do_update_resource"
shutdown="0">
<lrm id="278bebc6-2a59-4fa9-be2f-f6e262ce8936">
<lrm_resources>
<lrm_resource id="hbtest1a" type="Xen" class="ocf"
provider="heartbeat">
<lrm_rsc_op id="hbtest1a_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0"
transition-key="4:38:7:41183756-ab7d-45c5-9114-f6af4d0ab52f"
transition-magic="0:7;4:38:7:41183756-ab7d-45c5-9114-f6af4d0ab52f"
call-id="13" rc-code="7" op-status="0" interval="0"
last-run="1222261375" last-rc-change="1222261375" exec-time="400"
queue-time="0" op-digest="49a8ad7917b43502c13560299454cb0c"/>
<lrm_rsc_op id="hbtest1a_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0"
transition-key="6:38:0:41183756-ab7d-45c5-9114-f6af4d0ab52f"
transition-magic="0:5;6:38:0:41183756-ab7d-45c5-9114-f6af4d0ab52f"
call-id="14" rc-code="5" op-status="0" interval="0"
last-run="1222261375" last-rc-change="1222261375" exec-time="410"
queue-time="0" op-digest="61f0950751fea1b5f7a5c22a0923c546"/>
<lrm_rsc_op id="hbtest1a_stop_0" operation="stop"
crm-debug-origin="do_update_resource" crm_feature_set="3.0"
transition-key="1:39:0:41183756-ab7d-45c5-9114-f6af4d0ab52f"
transition-magic="0:0;1:39:0:41183756-ab7d-45c5-9114-f6af4d0ab52f"
call-id="15" rc-code="0" op-status="0" interval="0"
last-run="1222261376" last-rc-change="1222261376" exec-time="460"
queue-time="0" op-digest="b5aefbd8964a51dea7e69fac50cd2db7"/>
</lrm_resource>
</lrm_resources>
</lrm>
<transient_attributes id="278bebc6-2a59-4fa9-be2f-f6e262ce8936">
<instance_attributes
id="status-278bebc6-2a59-4fa9-be2f-f6e262ce8936">
<nvpair
id="status-278bebc6-2a59-4fa9-be2f-f6e262ce8936-probe_complete"
name="probe_complete" value="true"/>
<nvpair
id="status-278bebc6-2a59-4fa9-be2f-f6e262ce8936-last-failure-hbtest1a"
name="last-failure-hbtest1a" value="1222261376"/>
<nvpair
id="status-278bebc6-2a59-4fa9-be2f-f6e262ce8936-fail-count-hbtest1a"
name="fail-count-hbtest1a" value="INFINITY"/>
</instance_attributes>
</transient_attributes>
</node_state>
<node_state id="d06bdc02-b705-4bf9-9cd4-6f827ffdfe2e"
uname="xen20a.test.mytld.com" ha="active" in_ccm="true" crmd="online"
join="member" expected="member" crm-debug-origin="do_update_resource"
shutdown="0">
<lrm id="d06bdc02-b705-4bf9-9cd4-6f827ffdfe2e">
<lrm_resources>
<lrm_resource id="hbtest1a" type="Xen" class="ocf"
provider="heartbeat">
<lrm_rsc_op id="hbtest1a_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0"
transition-key="5:35:7:41183756-ab7d-45c5-9114-f6af4d0ab52f"
transition-magic="0:7;5:35:7:41183756-ab7d-45c5-9114-f6af4d0ab52f"
call-id="13" rc-code="7" op-status="0" interval="0"
last-run="1222261368" last-rc-change="1222261368" exec-time="410"
queue-time="0" op-digest="49a8ad7917b43502c13560299454cb0c"/>
<lrm_rsc_op id="hbtest1a_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0"
transition-key="5:36:0:41183756-ab7d-45c5-9114-f6af4d0ab52f"
transition-magic="0:5;5:36:0:41183756-ab7d-45c5-9114-f6af4d0ab52f"
call-id="14" rc-code="5" op-status="0" interval="0"
last-run="1222261369" last-rc-change="1222261369" exec-time="410"
queue-time="0" op-digest="61f0950751fea1b5f7a5c22a0923c546"/>
<lrm_rsc_op id="hbtest1a_stop_0" operation="stop"
crm-debug-origin="do_update_resource" crm_feature_set="3.0"
transition-key="1:37:0:41183756-ab7d-45c5-9114-f6af4d0ab52f"
transition-magic="0:0;1:37:0:41183756-ab7d-45c5-9114-f6af4d0ab52f"
call-id="15" rc-code="0" op-status="0" interval="0"
last-run="1222261371" last-rc-change="1222261371" exec-time="490"
queue-time="0" op-digest="b5aefbd8964a51dea7e69fac50cd2db7"/>
</lrm_resource>
</lrm_resources>
</lrm>
<transient_attributes id="d06bdc02-b705-4bf9-9cd4-6f827ffdfe2e">
<instance_attributes
id="status-d06bdc02-b705-4bf9-9cd4-6f827ffdfe2e">
<nvpair
id="status-d06bdc02-b705-4bf9-9cd4-6f827ffdfe2e-probe_complete"
name="probe_complete" value="true"/>
<nvpair
id="status-d06bdc02-b705-4bf9-9cd4-6f827ffdfe2e-last-failure-hbtest1a"
name="last-failure-hbtest1a" value="1222261371"/>
<nvpair
id="status-d06bdc02-b705-4bf9-9cd4-6f827ffdfe2e-fail-count-hbtest1a"
name="fail-count-hbtest1a" value="INFINITY"/>
</instance_attributes>
</transient_attributes>
</node_state>
</status>
</cib>
Andrew Beekhof wrote:
>
> On Sep 24, 2008, at 12:09 PM, Bruno Voigt wrote:
>
>> Hi,
>> I've just installed pacemaker unstable from
>> http://download.opensuse.org/repositories/server:/ha-clustering:/UNSTABLE/Debian_Etch/amd64/
>>
>> replacing a previous configuration of debian heartbeat.
>>
>> The CIB contains only the node configuration.
>>
>> If I try to add a resource definition I get the following error
>> and was not able yet to figure out, why it is rejected.
>>
>> root# cibadmin -V -C -o resources -x ./test-resource.xml
>> cibadmin[25909]: 2008/09/24_12:06:40 info: main: Starting mainloop
>> cibadmin[25909]: 2008/09/24_12:06:41 WARN: cibadmin_op_callback: Call
>> cib_create failed (-47): Update does not conform to the configured
>> schema/DTD
>> Call cib_create failed (-47): Update does not conform to the configured
>> schema/DTD
>> <null>
>>
>>
>> test-resource.xml contains:
>>
>> <resources>
>> <primitive id="hbtest1a" class="ocf" type="Xen" provider="heartbeat">
>> <operations>
>> <op id="hbtest1a-op01" name="start" timeout="60s" start_delay="0"/>
>
> start_delay -> start-delay
>
>>
>> <op id="hbtest1a-op02" name="stop" timeout="300s"/>
>> <op id="hbtest1a-op03" name="monitor" interval="30s" timeout="60s"
>> start_delay="300s" prereq="nothing"/>
>> </operations>
>
>
> prereq -> requires
>
>>
>>
>> <instance_attributes id="hbtest1a">
>> <attributes>
>
> the attributes scaffolding is no longer used
>
>>
>> <nvpair id="hbtest1a-attr01" name="xmfile"
>> value="/etc/xen/hbtest1a.cfg"/>
>> </attributes>
>> </instance_attributes>
>>
>> <meta_attributes id="hbtest1a-meta01">
>> <attributes>
>> <nvpair id="hbtest1a-meta-attr01" name="is_managed" value="true"/>
>> <nvpair id="hbtest1a-meta-attr02" name="allow_migrate" value="true"/>
>> <nvpair id="hbtest1a-meta-attr03" name="target_role" value="stopped"/>
>
> _ -> - for all options
>
>>
>> </attributes>
>> </meta_attributes>
>>
>> </primitive>
>> </resources>
>>
>> How must a similar resource definition look like for pacemaker?
>
> http://clusterlabs.org/mw/Image:Configuration_Explained_1.0.pdf
>
> specifically, you might want to read the bit on upgrading an old
> configuration
>
>>
>>
>> TIA for any hints,
>> Bruno
>> --
>> bruno.voigt at ic3s.de
>>
>>
>> _______________________________________________
>> Pacemaker mailing list
>> Pacemaker at clusterlabs.org
>> http://list.clusterlabs.org/mailman/listinfo/pacemaker
>
>
> _______________________________________________
> Pacemaker mailing list
> Pacemaker at clusterlabs.org
> http://list.clusterlabs.org/mailman/listinfo/pacemaker
More information about the Pacemaker
mailing list