[Pacemaker] how to test network access and fail over accordingly?

Craig Hurley lists at thehurley.com
Wed Oct 6 20:21:36 UTC 2010


I tried using ping instead of pingd and I added "number" to the
evaluation, I get the same results :/

primitive p_ping ocf:pacemaker:ping params host_list=172.20.0.254
clone c_ping p_ping meta globally-unique=false
location loc_ping g_cluster_services rule -inf: not_defined p_ping or
p_ping number:lte 0

Regards,
Craig.


On 6 October 2010 20:43, Jayakrishnan <jayakrishnanlll at gmail.com> wrote:
>
> Hello,
>
> Guess the change:--
> location loc_pingd g_cluster_services rule -inf: not_defined pingd or pingd
> number:lte 0
>
> should work
>
>
> --
> Regards,
>
> Jayakrishnan. L
>
> Visit:
> www.foralllinux.blogspot.com
> www.jayakrishnan.bravehost.com
>
>
> On Wed, Oct 6, 2010 at 11:56 AM, Claus Denk <denk at us.es> wrote:
>>
>> I am having a similar problem, so let's wait for the experts, But in the
>> meanwhile, try changing
>>
>>
>> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd
>> or p_pingd lte 0
>>
>> to
>>
>> location loc_pingd g_cluster_services rule -inf: not_defined pingd
>> or pingd number:lte 0
>>
>> and see what happens. As far as I have read, it is also more recommended
>> to use the "ping"
>> resource instead of "pingd"...
>>
>> kind regards, Claus
>>
>>
>>
>>
>>
>>
>> On 10/06/2010 05:45 AM, Craig Hurley wrote:
>>>
>>> Hello,
>>>
>>> I have a 2 node cluster, running DRBD, heartbeat and pacemaker in
>>> active/passive mode.  On both nodes, eth0 is connected to the main
>>> network, eth1 is used to connect the nodes directly to each other.
>>> The nodes share a virtual IP address on eth0.  Pacemaker is also
>>> controlling a custom service with an LSB compliant script in
>>> /etc/init.d/.  All of this is working fine and I'm happy with it.
>>>
>>> I'd like to configure the nodes so that they fail over if eth0 goes
>>> down (or if they cannot access a particular gateway), so I tried
>>> adding the following (as per
>>> http://www.clusterlabs.org/wiki/Example_configurations#Set_up_pingd)
>>>
>>> primitive p_pingd ocf:pacemaker:pingd params host_list=172.20.0.254 op
>>> monitor interval=15s timeout=5s
>>> clone c_pingd p_pingd meta globally-unique=false
>>> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd
>>> or p_pingd lte 0
>>>
>>> ... but when I do add that, all resource are stopped and they don't
>>> come back up on either node.  Am I making a basic mistake or do you
>>> need more info from me?
>>>
>>> All help is appreciated,
>>> Craig.
>>>
>>>
>>> pacemaker
>>> Version: 1.0.8+hg15494-2ubuntu2
>>>
>>> heartbeat
>>> Version: 1:3.0.3-1ubuntu1
>>>
>>> drbd8-utils
>>> Version: 2:8.3.7-1ubuntu2.1
>>>
>>>
>>> rp at rpalpha:~$ sudo crm configure show
>>> node $id="32482293-7b0f-466e-b405-c64bcfa2747d" rpalpha
>>> node $id="3f2aac12-05aa-4ac7-b91f-c47fa28efb44" rpbravo
>>> primitive p_drbd_data ocf:linbit:drbd \
>>>         params drbd_resource="data" \
>>>         op monitor interval="30s"
>>> primitive p_fs_data ocf:heartbeat:Filesystem \
>>>         params device="/dev/drbd/by-res/data" directory="/mnt/data"
>>> fstype="ext4"
>>> primitive p_ip ocf:heartbeat:IPaddr2 \
>>>         params ip="172.20.50.3" cidr_netmask="255.255.0.0" nic="eth0" \
>>>         op monitor interval="30s"
>>> primitive p_rp lsb:rp \
>>>         op monitor interval="30s" \
>>>         meta target-role="Started"
>>> group g_cluster_services p_ip p_fs_data p_rp
>>> ms ms_drbd p_drbd_data \
>>>         meta master-max="1" master-node-max="1" clone-max="2"
>>> clone-node-max="1" notify="true"
>>> location loc_preferred_master g_cluster_services inf: rpalpha
>>> colocation colo_mnt_on_master inf: g_cluster_services ms_drbd:Master
>>> order ord_mount_after_drbd inf: ms_drbd:promote g_cluster_services:start
>>> property $id="cib-bootstrap-options" \
>>>         dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \
>>>         cluster-infrastructure="Heartbeat" \
>>>         no-quorum-policy="ignore" \
>>>         stonith-enabled="false" \
>>>         expected-quorum-votes="2" \
>>>
>>>
>>> rp at rpalpha:~$ sudo cat /etc/ha.d/ha.cf
>>> node rpalpha
>>> node rpbravo
>>>
>>> keepalive 2
>>> warntime 5
>>> deadtime 15
>>> initdead 60
>>>
>>> mcast eth0 239.0.0.43 694 1 0
>>> bcast eth1
>>>
>>> use_logd yes
>>> autojoin none
>>> crm respawn
>>>
>>>
>>> rp at rpalpha:~$ sudo cat /etc/drbd.conf
>>> global {
>>>         usage-count no;
>>> }
>>> common {
>>>         protocol C;
>>>
>>>         handlers {}
>>>
>>>         startup {}
>>>
>>>         disk {}
>>>
>>>         net {
>>>                 cram-hmac-alg sha1;
>>>                 shared-secret "foobar";
>>>         }
>>>
>>>         syncer {
>>>                 verify-alg sha1;
>>>                 rate 100M;
>>>         }
>>> }
>>> resource data {
>>>         device /dev/drbd0;
>>>         meta-disk internal;
>>>         on rpalpha {
>>>                 disk /dev/mapper/rpalpha-data;
>>>                 address 192.168.1.1:7789;
>>>         }
>>>         on rpbravo {
>>>                 disk /dev/mapper/rpbravo-data;
>>>                 address 192.168.1.2:7789;
>>>         }
>>> }
>>>




More information about the Pacemaker mailing list