[Pacemaker] Infinite fail-count and migration-threshold after node fail-back

Pavlos Parissis pavlos.parissis at gmail.com
Thu Nov 11 08:37:01 EST 2010

On 11 November 2010 13:04, Dan Frincu <dfrincu at streamwide.ro> wrote:
> Hi,
> Andrew Beekhof wrote:
> On Mon, Oct 11, 2010 at 9:40 AM, Dan Frincu <dfrincu at streamwide.ro> wrote:
> Hi all,
> I've managed to make this setup work, basically the issue with a
> symmetric-cluster="false" and specifying the resources' location manually
> means that the resources will always obey the location constraint, and (as
> far as I could see) disregard the rsc_defaults resource-stickiness values.
> This definitely should not be the case.
> Possibly your stickiness setting is being eclipsed by the combination
> of the location constraint scores.
> Try INFINITY instead.
> I understand your point and I believe also this to be the case, however I've
> noticed that by specifying symmetric-cluster="false" for each resource I
> need to add 2 location constraints, which overcrowds the config, and if I
> want (and hope) to go to a config with multiple servers and resources, each
> with specific rules, then also adding location constraints for each resource
> is an overhead which I'd rather not include, if possible.

>From the documentation [1]
6.2.2. Asymmetrical "Opt-In" Clusters
To create an opt-in cluster, start by preventing resources from
running anywhere by default
crm_attribute --attr-name symmetric-cluster --attr-value false
Then start enabling nodes. The following fragment says that the web
server prefers sles-1, the database prefers sles-2 and both can
failover to sles-3 if their most preferred node fails.

    <rsc_location id="loc-1" rsc="Webserver" node="sles-1" score="200"/>
    <rsc_location id="loc-2" rsc="Webserver" node="sles-3" score="0"/>
    <rsc_location id="loc-3" rsc="Database" node="sles-2" score="200"/>
    <rsc_location id="loc-4" rsc="Database" node="sles-3" score="0"/>
Example 6.1. Example set of opt-in location constraints

At the moment you have symmetric-cluster=false, you need to add
location constraints in order to get your resources running.
Below is my conf and it works as expected, pbx_service_01 starts on
node-01 and never fails back, in case failed over to node-03 and
node-01 is back on line, due to resource-stickiness="1000", but take a
look at the score in location constraint, very low scores compared to
1000 - I could  have also set it to inf
location PrimaryNode-drbd_01 ms-drbd_01 100: node-01
location PrimaryNode-drbd_02 ms-drbd_02 100: node-02
location PrimaryNode-pbx_service_01 pbx_service_01 200: node-01
location PrimaryNode-pbx_service_02 pbx_service_02 200: node-02
location SecondaryNode-drbd_01 ms-drbd_01 0: node-03
location SecondaryNode-drbd_02 ms-drbd_02 0: node-03
location SecondaryNode-pbx_service_01 pbx_service_01 10: node-03
location SecondaryNode-pbx_service_02 pbx_service_02 10: node-03
location fencing-on-node-01 pdu 1: node-01
location fencing-on-node-02 pdu 1: node-02
location fencing-on-node-03 pdu 1: node-03
rsc_defaults $id="rsc-options"   resource-stickiness="1000"


More information about the Pacemaker mailing list