[Pacemaker] how to test network access and fail over accordingly?

Vadym Chepkov vchepkov at gmail.com
Wed Oct 6 12:19:30 UTC 2010


On Oct 6, 2010, at 3:43 AM, Jayakrishnan wrote:

> 
> Hello,
>  
> Guess the change:--  
> location loc_pingd g_cluster_services rule -inf: not_defined pingd or pingd number:lte 0
> 
> should work
>  
> 
> 


ocf:pacemaker:ping is recommended as a replacement for pingd RA

Both RA define node attribute "pingd" by default, I think this question arises a lot, since crm ra meta for both agents is misleading:

#  crm ra meta ocf:pacemaker:ping

name (string, [undef]): Attribute name
    The name of the attributes to set.  This is the name to be used in the constraints.

I think it should say "pingd" instead of "undef"

Obviously, you can redefine any name you like and use it instead, but, unfortunately, "pingd" is the only attribute name that crm_mon -f would display, the name is hardcoded in crm_mon.c:

    if(safe_str_eq("pingd", g_hash_table_lookup(rsc->meta, "type"))) {

this is inconvenience for multi-homed clusters where you need to define separate ping clones for each network, so maybe crm_mon should display attributes starting with "ping". Just a thought.


Vadym


> -- 
> Regards,
> 
> Jayakrishnan. L
> 
> Visit: 
> www.foralllinux.blogspot.com
> www.jayakrishnan.bravehost.com
>  
>  
> On Wed, Oct 6, 2010 at 11:56 AM, Claus Denk <denk at us.es> wrote:
> I am having a similar problem, so let's wait for the experts, But in the meanwhile, try changing
> 
> 
> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd
> or p_pingd lte 0
> 
> to
> 
> location loc_pingd g_cluster_services rule -inf: not_defined pingd
> or pingd number:lte 0
> 
> and see what happens. As far as I have read, it is also more recommended to use the "ping"
> resource instead of "pingd"...
> 
> kind regards, Claus
> 
> 
> 
> 
> 
> 
> On 10/06/2010 05:45 AM, Craig Hurley wrote:
> Hello,
> 
> I have a 2 node cluster, running DRBD, heartbeat and pacemaker in
> active/passive mode.  On both nodes, eth0 is connected to the main
> network, eth1 is used to connect the nodes directly to each other.
> The nodes share a virtual IP address on eth0.  Pacemaker is also
> controlling a custom service with an LSB compliant script in
> /etc/init.d/.  All of this is working fine and I'm happy with it.
> 
> I'd like to configure the nodes so that they fail over if eth0 goes
> down (or if they cannot access a particular gateway), so I tried
> adding the following (as per
> http://www.clusterlabs.org/wiki/Example_configurations#Set_up_pingd)
> 
> primitive p_pingd ocf:pacemaker:pingd params host_list=172.20.0.254 op
> monitor interval=15s timeout=5s
> clone c_pingd p_pingd meta globally-unique=false
> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd
> or p_pingd lte 0
> 
> ... but when I do add that, all resource are stopped and they don't
> come back up on either node.  Am I making a basic mistake or do you
> need more info from me?
> 
> All help is appreciated,
> Craig.
> 
> 
> pacemaker
> Version: 1.0.8+hg15494-2ubuntu2
> 
> heartbeat
> Version: 1:3.0.3-1ubuntu1
> 
> drbd8-utils
> Version: 2:8.3.7-1ubuntu2.1
> 
> 
> rp at rpalpha:~$ sudo crm configure show
> node $id="32482293-7b0f-466e-b405-c64bcfa2747d" rpalpha
> node $id="3f2aac12-05aa-4ac7-b91f-c47fa28efb44" rpbravo
> primitive p_drbd_data ocf:linbit:drbd \
>         params drbd_resource="data" \
>         op monitor interval="30s"
> primitive p_fs_data ocf:heartbeat:Filesystem \
>         params device="/dev/drbd/by-res/data" directory="/mnt/data"
> fstype="ext4"
> primitive p_ip ocf:heartbeat:IPaddr2 \
>         params ip="172.20.50.3" cidr_netmask="255.255.0.0" nic="eth0" \
>         op monitor interval="30s"
> primitive p_rp lsb:rp \
>         op monitor interval="30s" \
>         meta target-role="Started"
> group g_cluster_services p_ip p_fs_data p_rp
> ms ms_drbd p_drbd_data \
>         meta master-max="1" master-node-max="1" clone-max="2"
> clone-node-max="1" notify="true"
> location loc_preferred_master g_cluster_services inf: rpalpha
> colocation colo_mnt_on_master inf: g_cluster_services ms_drbd:Master
> order ord_mount_after_drbd inf: ms_drbd:promote g_cluster_services:start
> property $id="cib-bootstrap-options" \
>         dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \
>         cluster-infrastructure="Heartbeat" \
>         no-quorum-policy="ignore" \
>         stonith-enabled="false" \
>         expected-quorum-votes="2" \
> 
> 
> rp at rpalpha:~$ sudo cat /etc/ha.d/ha.cf
> node rpalpha
> node rpbravo
> 
> keepalive 2
> warntime 5
> deadtime 15
> initdead 60
> 
> mcast eth0 239.0.0.43 694 1 0
> bcast eth1
> 
> use_logd yes
> autojoin none
> crm respawn
> 
> 
> rp at rpalpha:~$ sudo cat /etc/drbd.conf
> global {
>         usage-count no;
> }
> common {
>         protocol C;
> 
>         handlers {}
> 
>         startup {}
> 
>         disk {}
> 
>         net {
>                 cram-hmac-alg sha1;
>                 shared-secret "foobar";
>         }
> 
>         syncer {
>                 verify-alg sha1;
>                 rate 100M;
>         }
> }
> resource data {
>         device /dev/drbd0;
>         meta-disk internal;
>         on rpalpha {
>                 disk /dev/mapper/rpalpha-data;
>                 address 192.168.1.1:7789;
>         }
>         on rpbravo {
>                 disk /dev/mapper/rpbravo-data;
>                 address 192.168.1.2:7789;
>         }
> }
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
>   
> 
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
> 
> 
>  
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20101006/10296e41/attachment-0002.htm>


More information about the Pacemaker mailing list