[Pacemaker] Pacemaker 0.7.3: How to use pingd clone resource and constraints

Bruno Voigt Bruno.Voigt at ic3s.de
Thu Sep 25 15:19:37 EDT 2008


Hi Andrew,

is pingd doing alive tests differently compared to the normal ping command?
normal & flood ping of these hosts show 0% packet lost from my both nodes.

In the log below pingd - besides the warnings -
states that the node is alive and that it had sent an update,
but it does not show up in the cib.

WR,
Bruno

Andrew Beekhof wrote:
>
> On Sep 24, 2008, at 10:36 PM, Serge Dubrouski wrote:
>
>> There is a problem with attrd that affects  pingd in Pacemeaker
>> 0.7.3/Heartbeat 2.99. I've already created a Bugzilla ticket for it.
>> You can add your information there:
>>
>> http://developerbugs.linux-foundation.org/show_bug.cgi?id=1969
>
> I'm not so sure this is the same thing.
> Those "Bad echo" messages look suspicious
>
>>> Sep 24 22:01:46 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo
>>> (0):
>>> 3, code=1, seq=0, id=0, check=24031
>
> In fact, type=3 is ICMP_DEST_UNREACH - so pingd really is having
> trouble contacting that host.
>
>
>>
>>
>> On Wed, Sep 24, 2008 at 2:04 PM, Bruno Voigt <Bruno.Voigt at ic3s.de>
>> wrote:
>>> I defined two ping clone resources,
>>> to be used independently by different resources:
>>>
>>>     <clone id="clone-pingd-internal">
>>>       <primitive id="pingd-internal" provider="pacemaker" class="ocf"
>>> type="pingd">
>>>         <instance_attributes id="pingd-internal-ia">
>>>           <nvpair id="pingd-internal-ia01" name="name"
>>> value="pingd-internal"/>
>>>           <nvpair id="pingd-internal-ia02" name="dampen" value="5s"/>
>>>           <nvpair id="pingd-internal-ia03" name="multiplier"
>>> value="1000"/>
>>>           <nvpair id="pingd-internal-ia04" name="host_list"
>>> value="172.17.32.23 192.168.132.23"/>
>>>         </instance_attributes>
>>>       </primitive>
>>>     </clone>
>>>
>>>     <clone id="clone-pingd-external">
>>>       <primitive id="pingd-external" provider="pacemaker" class="ocf"
>>> type="pingd">
>>>         <instance_attributes id="pingd-external-ia">
>>>           <nvpair id="pingd-external-ia01" name="name"
>>> value="pingd-external"/>
>>>           <nvpair id="pingd-external-ia02" name="dampen" value="5s"/>
>>>           <nvpair id="pingd-external-ia03" name="multiplier"
>>> value="1000"/>
>>>           <nvpair id="pingd-external-ia04" name="host_list"
>>> value="195.244.97.241"/>
>>>         </instance_attributes>
>>>       </primitive>
>>>     </clone>
>>>
>>> I defined a constraint for a resource so that it depends on
>>> pingd-internal
>>>
>>> <constraints>
>>> <rsc_location id="hbtest1b-connectivity" rsc="hbtest1b">
>>>   <rule id="hbtest1b-connectivity-exclude-rule" score="-INFINITY" >
>>>     <expression id="hbtest1b-connectivity-exclude"
>>> attribute="pingd-internal" operation="not_defined"/>
>>>   </rule>
>>> </rsc_location>
>>> </constraints>
>>>
>>> But this causes the resource to be unrunnable on either of my both
>>> nodes,
>>>
>>>
>>> There are as expected 2 pingd daemons running:
>>>
>>> root      6132     1  0 21:07 ?        00:00:00
>>> /usr/lib/heartbeat/pingd
>>> -D -p /var/run/heartbeat/rsctmp/pingd-pingd-internal:0 -a
>>> pingd-internal
>>> -d 5s -m 1000 -h 172.17.32.23 -h 192.168.132.23
>>> root     13142     1  0 21:47 ?        00:00:00
>>> /usr/lib/heartbeat/pingd
>>> -D -p /var/run/heartbeat/rsctmp/pingd-pingd-external:0 -a
>>> pingd-external
>>> -d 5s -m 1000 -h 195.244.97.241
>>>
>>> The problem is, I can't see in the cibadmin -Q output that the pingd
>>> daemons have
>>> have stored  their results anywhere..
>>>
>>> In the log I see the following output:
>>>
>>> Sep 24 22:01:46 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo
>>> (0):
>>> 3, code=1, seq=0, id=0, check=24031
>>> Sep 24 22:01:47 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo
>>> (0):
>>> 3, code=1, seq=0, id=0, check=24031
>>> Sep 24 22:01:48 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo
>>> (0):
>>> 8, code=0, seq=261, id=0, check=22762
>>> Sep 24 22:01:48 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
>>> 8, code=0, seq=263, id=0, check=22250
>>> Sep 24 22:01:50 xen20a pingd: [13142]: info: stand_alone_ping: Node
>>> 195.244.97.241 is alive (1)
>>> Sep 24 22:01:50 xen20a pingd: [13142]: info: send_update: 1 active ping
>>> nodes
>>> Sep 24 22:01:51 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
>>> 3, code=1, seq=0, id=0, check=24031
>>> Sep 24 22:01:52 xen20a pingd: [6132]: info: stand_alone_ping: Node
>>> 172.17.32.23 is alive (3)
>>> Sep 24 22:01:53 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
>>> 3, code=1, seq=0, id=0, check=24031
>>> Sep 24 22:01:54 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
>>> 3, code=1, seq=0, id=0, check=25823
>>> Sep 24 22:01:55 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0):
>>> 3, code=1, seq=0, id=0, check=24031
>>> Sep 24 22:01:57 xen20a pingd: [6132]: info: stand_alone_ping: Node
>>> 192.168.132.23 is alive (2)
>>> Sep 24 22:01:57 xen20a pingd: [6132]: info: send_update: 2 active
>>> ping nodes
>>>
>>> Where should the current pingd status be located in the cib ?
>>> What is wrong with my setup ?
>>>
>>> TIA,
>>> Bruno
>>>
>>
>> -- 
>> Serge Dubrouski.
>

-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 197 bytes
Desc: OpenPGP digital signature
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20080925/8885619a/attachment-0001.sig>


More information about the Pacemaker mailing list