[Pacemaker] Pacemaker 0.7.3: How to use pingd clone resource and constraints
Bruno Voigt
Bruno.Voigt at ic3s.de
Wed Sep 24 20:04:10 UTC 2008
I defined two ping clone resources,
to be used independently by different resources:
<clone id="clone-pingd-internal">
<primitive id="pingd-internal" provider="pacemaker" class="ocf"
type="pingd">
<instance_attributes id="pingd-internal-ia">
<nvpair id="pingd-internal-ia01" name="name"
value="pingd-internal"/>
<nvpair id="pingd-internal-ia02" name="dampen" value="5s"/>
<nvpair id="pingd-internal-ia03" name="multiplier"
value="1000"/>
<nvpair id="pingd-internal-ia04" name="host_list"
value="172.17.32.23 192.168.132.23"/>
</instance_attributes>
</primitive>
</clone>
<clone id="clone-pingd-external">
<primitive id="pingd-external" provider="pacemaker" class="ocf"
type="pingd">
<instance_attributes id="pingd-external-ia">
<nvpair id="pingd-external-ia01" name="name"
value="pingd-external"/>
<nvpair id="pingd-external-ia02" name="dampen" value="5s"/>
<nvpair id="pingd-external-ia03" name="multiplier"
value="1000"/>
<nvpair id="pingd-external-ia04" name="host_list"
value="195.244.97.241"/>
</instance_attributes>
</primitive>
</clone>
I defined a constraint for a resource so that it depends on pingd-internal
<constraints>
<rsc_location id="hbtest1b-connectivity" rsc="hbtest1b">
<rule id="hbtest1b-connectivity-exclude-rule" score="-INFINITY" >
<expression id="hbtest1b-connectivity-exclude"
attribute="pingd-internal" operation="not_defined"/>
</rule>
</rsc_location>
</constraints>
But this causes the resource to be unrunnable on either of my both nodes,
There are as expected 2 pingd daemons running:
root 6132 1 0 21:07 ? 00:00:00 /usr/lib/heartbeat/pingd
-D -p /var/run/heartbeat/rsctmp/pingd-pingd-internal:0 -a pingd-internal
-d 5s -m 1000 -h 172.17.32.23 -h 192.168.132.23
root 13142 1 0 21:47 ? 00:00:00 /usr/lib/heartbeat/pingd
-D -p /var/run/heartbeat/rsctmp/pingd-pingd-external:0 -a pingd-external
-d 5s -m 1000 -h 195.244.97.241
The problem is, I can't see in the cibadmin -Q output that the pingd
daemons have
have stored their results anywhere..
In the log I see the following output:
Sep 24 22:01:46 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo (0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:47 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo (0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:48 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo (0):
8, code=0, seq=261, id=0, check=22762
Sep 24 22:01:48 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
8, code=0, seq=263, id=0, check=22250
Sep 24 22:01:50 xen20a pingd: [13142]: info: stand_alone_ping: Node
195.244.97.241 is alive (1)
Sep 24 22:01:50 xen20a pingd: [13142]: info: send_update: 1 active ping
nodes
Sep 24 22:01:51 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:52 xen20a pingd: [6132]: info: stand_alone_ping: Node
172.17.32.23 is alive (3)
Sep 24 22:01:53 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:54 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
3, code=1, seq=0, id=0, check=25823
Sep 24 22:01:55 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:57 xen20a pingd: [6132]: info: stand_alone_ping: Node
192.168.132.23 is alive (2)
Sep 24 22:01:57 xen20a pingd: [6132]: info: send_update: 2 active ping nodes
Where should the current pingd status be located in the cib ?
What is wrong with my setup ?
TIA,
Bruno
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 197 bytes
Desc: OpenPGP digital signature
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20080924/4658119a/attachment-0001.sig>
More information about the Pacemaker
mailing list