[Pacemaker] how to test network access and fail over accordingly?

Vadym Chepkov vchepkov at gmail.com
Wed Oct 6 16:29:21 EDT 2010


On Oct 6, 2010, at 4:21 PM, Craig Hurley wrote:

> I tried using ping instead of pingd and I added "number" to the
> evaluation, I get the same results :/
> 
> primitive p_ping ocf:pacemaker:ping params host_list=172.20.0.254
> clone c_ping p_ping meta globally-unique=false
> location loc_ping g_cluster_services rule -inf: not_defined p_ping or
> p_ping number:lte 0
> 


try this

primitive p_ping ocf:pacemaker:ping params name="p_ping" host_list="172.20.0.254" op monitor interval="10s"

Vadym

> Regards,
> Craig.
> 
> 
> On 6 October 2010 20:43, Jayakrishnan <jayakrishnanlll at gmail.com> wrote:
>> 
>> Hello,
>> 
>> Guess the change:--
>> location loc_pingd g_cluster_services rule -inf: not_defined pingd or pingd
>> number:lte 0
>> 
>> should work
>> 
>> 
>> --
>> Regards,
>> 
>> Jayakrishnan. L
>> 
>> Visit:
>> www.foralllinux.blogspot.com
>> www.jayakrishnan.bravehost.com
>> 
>> 
>> On Wed, Oct 6, 2010 at 11:56 AM, Claus Denk <denk at us.es> wrote:
>>> 
>>> I am having a similar problem, so let's wait for the experts, But in the
>>> meanwhile, try changing
>>> 
>>> 
>>> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd
>>> or p_pingd lte 0
>>> 
>>> to
>>> 
>>> location loc_pingd g_cluster_services rule -inf: not_defined pingd
>>> or pingd number:lte 0
>>> 
>>> and see what happens. As far as I have read, it is also more recommended
>>> to use the "ping"
>>> resource instead of "pingd"...
>>> 
>>> kind regards, Claus
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>>> On 10/06/2010 05:45 AM, Craig Hurley wrote:
>>>> 
>>>> Hello,
>>>> 
>>>> I have a 2 node cluster, running DRBD, heartbeat and pacemaker in
>>>> active/passive mode.  On both nodes, eth0 is connected to the main
>>>> network, eth1 is used to connect the nodes directly to each other.
>>>> The nodes share a virtual IP address on eth0.  Pacemaker is also
>>>> controlling a custom service with an LSB compliant script in
>>>> /etc/init.d/.  All of this is working fine and I'm happy with it.
>>>> 
>>>> I'd like to configure the nodes so that they fail over if eth0 goes
>>>> down (or if they cannot access a particular gateway), so I tried
>>>> adding the following (as per
>>>> http://www.clusterlabs.org/wiki/Example_configurations#Set_up_pingd)
>>>> 
>>>> primitive p_pingd ocf:pacemaker:pingd params host_list=172.20.0.254 op
>>>> monitor interval=15s timeout=5s
>>>> clone c_pingd p_pingd meta globally-unique=false
>>>> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd
>>>> or p_pingd lte 0
>>>> 
>>>> ... but when I do add that, all resource are stopped and they don't
>>>> come back up on either node.  Am I making a basic mistake or do you
>>>> need more info from me?
>>>> 
>>>> All help is appreciated,
>>>> Craig.
>>>> 
>>>> 
>>>> pacemaker
>>>> Version: 1.0.8+hg15494-2ubuntu2
>>>> 
>>>> heartbeat
>>>> Version: 1:3.0.3-1ubuntu1
>>>> 
>>>> drbd8-utils
>>>> Version: 2:8.3.7-1ubuntu2.1
>>>> 
>>>> 
>>>> rp at rpalpha:~$ sudo crm configure show
>>>> node $id="32482293-7b0f-466e-b405-c64bcfa2747d" rpalpha
>>>> node $id="3f2aac12-05aa-4ac7-b91f-c47fa28efb44" rpbravo
>>>> primitive p_drbd_data ocf:linbit:drbd \
>>>>         params drbd_resource="data" \
>>>>         op monitor interval="30s"
>>>> primitive p_fs_data ocf:heartbeat:Filesystem \
>>>>         params device="/dev/drbd/by-res/data" directory="/mnt/data"
>>>> fstype="ext4"
>>>> primitive p_ip ocf:heartbeat:IPaddr2 \
>>>>         params ip="172.20.50.3" cidr_netmask="255.255.0.0" nic="eth0" \
>>>>         op monitor interval="30s"
>>>> primitive p_rp lsb:rp \
>>>>         op monitor interval="30s" \
>>>>         meta target-role="Started"
>>>> group g_cluster_services p_ip p_fs_data p_rp
>>>> ms ms_drbd p_drbd_data \
>>>>         meta master-max="1" master-node-max="1" clone-max="2"
>>>> clone-node-max="1" notify="true"
>>>> location loc_preferred_master g_cluster_services inf: rpalpha
>>>> colocation colo_mnt_on_master inf: g_cluster_services ms_drbd:Master
>>>> order ord_mount_after_drbd inf: ms_drbd:promote g_cluster_services:start
>>>> property $id="cib-bootstrap-options" \
>>>>         dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \
>>>>         cluster-infrastructure="Heartbeat" \
>>>>         no-quorum-policy="ignore" \
>>>>         stonith-enabled="false" \
>>>>         expected-quorum-votes="2" \
>>>> 
>>>> 
>>>> rp at rpalpha:~$ sudo cat /etc/ha.d/ha.cf
>>>> node rpalpha
>>>> node rpbravo
>>>> 
>>>> keepalive 2
>>>> warntime 5
>>>> deadtime 15
>>>> initdead 60
>>>> 
>>>> mcast eth0 239.0.0.43 694 1 0
>>>> bcast eth1
>>>> 
>>>> use_logd yes
>>>> autojoin none
>>>> crm respawn
>>>> 
>>>> 
>>>> rp at rpalpha:~$ sudo cat /etc/drbd.conf
>>>> global {
>>>>         usage-count no;
>>>> }
>>>> common {
>>>>         protocol C;
>>>> 
>>>>         handlers {}
>>>> 
>>>>         startup {}
>>>> 
>>>>         disk {}
>>>> 
>>>>         net {
>>>>                 cram-hmac-alg sha1;
>>>>                 shared-secret "foobar";
>>>>         }
>>>> 
>>>>         syncer {
>>>>                 verify-alg sha1;
>>>>                 rate 100M;
>>>>         }
>>>> }
>>>> resource data {
>>>>         device /dev/drbd0;
>>>>         meta-disk internal;
>>>>         on rpalpha {
>>>>                 disk /dev/mapper/rpalpha-data;
>>>>                 address 192.168.1.1:7789;
>>>>         }
>>>>         on rpbravo {
>>>>                 disk /dev/mapper/rpbravo-data;
>>>>                 address 192.168.1.2:7789;
>>>>         }
>>>> }
>>>> 
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker





More information about the Pacemaker mailing list