I defined two ping clone resources,
to be used independently by different resources:
<clone id="clone-pingd-internal">
<primitive id="pingd-internal" provider="pacemaker" class="ocf"
type="pingd">
<instance_attributes id="pingd-internal-ia">
<nvpair id="pingd-internal-ia01" name="name"
value="pingd-internal"/>
<nvpair id="pingd-internal-ia02" name="dampen" value="5s"/>
<nvpair id="pingd-internal-ia03" name="multiplier"
value="1000"/>
<nvpair id="pingd-internal-ia04" name="host_list"
value="172.17.32.23 192.168.132.23"/>
</instance_attributes>
</primitive>
</clone>
<clone id="clone-pingd-external">
<primitive id="pingd-external" provider="pacemaker" class="ocf"
type="pingd">
<instance_attributes id="pingd-external-ia">
<nvpair id="pingd-external-ia01" name="name"
value="pingd-external"/>
<nvpair id="pingd-external-ia02" name="dampen" value="5s"/>
<nvpair id="pingd-external-ia03" name="multiplier"
value="1000"/>
<nvpair id="pingd-external-ia04" name="host_list"
value="195.244.97.241"/>
</instance_attributes>
</primitive>
</clone>
I defined a constraint for a resource so that it depends on pingd-
internal
<constraints>
<rsc_location id="hbtest1b-connectivity" rsc="hbtest1b">
<rule id="hbtest1b-connectivity-exclude-rule" score="-INFINITY" >
<expression id="hbtest1b-connectivity-exclude"
attribute="pingd-internal" operation="not_defined"/>
</rule>
</rsc_location>
</constraints>
But this causes the resource to be unrunnable on either of my both
nodes,
There are as expected 2 pingd daemons running:
root 6132 1 0 21:07 ? 00:00:00 /usr/lib/heartbeat/
pingd
-D -p /var/run/heartbeat/rsctmp/pingd-pingd-internal:0 -a pingd-
internal
-d 5s -m 1000 -h 172.17.32.23 -h 192.168.132.23
root 13142 1 0 21:47 ? 00:00:00 /usr/lib/heartbeat/
pingd
-D -p /var/run/heartbeat/rsctmp/pingd-pingd-external:0 -a pingd-
external
-d 5s -m 1000 -h 195.244.97.241
The problem is, I can't see in the cibadmin -Q output that the pingd
daemons have
have stored their results anywhere..
In the log I see the following output:
Sep 24 22:01:46 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo
(0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:47 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo
(0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:48 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo
(0):
8, code=0, seq=261, id=0, check=22762
Sep 24 22:01:48 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo
(0):
8, code=0, seq=263, id=0, check=22250
Sep 24 22:01:50 xen20a pingd: [13142]: info: stand_alone_ping: Node
195.244.97.241 is alive (1)
Sep 24 22:01:50 xen20a pingd: [13142]: info: send_update: 1 active
ping
nodes
Sep 24 22:01:51 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo
(0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:52 xen20a pingd: [6132]: info: stand_alone_ping: Node
172.17.32.23 is alive (3)
Sep 24 22:01:53 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo
(0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:54 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo
(0):
3, code=1, seq=0, id=0, check=25823
Sep 24 22:01:55 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo
(0):
3, code=1, seq=0, id=0, check=24031
Sep 24 22:01:57 xen20a pingd: [6132]: info: stand_alone_ping: Node
192.168.132.23 is alive (2)
Sep 24 22:01:57 xen20a pingd: [6132]: info: send_update: 2 active
ping nodes
Where should the current pingd status be located in the cib ?
What is wrong with my setup ?
TIA,
Bruno
_______________________________________________
Pacemaker mailing list
[email protected]
http://list.clusterlabs.org/mailman/listinfo/pacemaker