[Pacemaker] Problem understanding resource stickyness

David Vossel dvossel at redhat.com
Wed Apr 25 17:31:58 UTC 2012


----- Original Message -----
> From: "Bernhard Schmidt" <berni at birkenwald.de>
> To: pacemaker at clusterlabs.org
> Sent: Tuesday, April 24, 2012 4:30:55 AM
> Subject: [Pacemaker] Problem understanding resource stickyness
> 
> Hi everyone,
> 
> I have a small problem with a simple two-node virtualization cluster
> that runs a pair of firewall VMs. They run on DRBD devices and may
> both
> run on one host, but they should be distributed when two nodes are
> available. Also, if a VM has to be migrated, it should be the second
> one
> (the VMs have internal HA and the first one is usually active, so
> that
> should stay where it is).
> 
> I'm using the following configuration
> 
> primitive drbd-greatfw1 ocf:linbit:drbd \
>         params drbd_resource="greatfw1" \
>         op monitor interval="15s"
> primitive drbd-greatfw2 ocf:linbit:drbd \
>         params drbd_resource="greatfw2" \
>         op monitor interval="15s"
> 
> primitive kvm-greatfw1 heartbeat:kvm \
>         params 1="greatfw1" \
>         meta resource-stickiness="1000" target-role="Started"
> primitive kvm-greatfw2 heartbeat:kvm \
>         params 1="greatfw2"
> 
> ms ms-drbd-greatfw1 drbd-greatfw1 \
>         meta master-max="1" master-node-max="1" clone-max="2"
> clone-node-max="1" notify="true"
> ms ms-drbd-greatfw2 drbd-greatfw2 \
>         meta master-max="1" master-node-max="1" clone-max="2"
> clone-node-max="1" notify="true"
> 
> colocation vm-greatfw1 inf: kvm-greatfw1 ms-drbd-greatfw1:Master
> colocation vm-greatfw2 inf: kvm-greatfw2 ms-drbd-greatfw2:Master
> colocation col-greatfw1-greatfw2 -2000: kvm-greatfw1 kvm-greatfw2

Switch the order on the above colocation constraint to look like this.

colocation col-greatfw1-greatfw2 -2000: kvm-greatfw2 kvm-greatfw1

Where kvm-greatfw1 is is the with-rsc argument... Basically just reverse the two resources.  This fixed it for me, but I'm not convinced this isn't a bug.  I'm looking further into this.

-- Vossel


> 
> order vm-greatfw1-order inf: ms-drbd-greatfw1:promote
> kvm-greatfw1:start
> order vm-greatfw2-order inf: ms-drbd-greatfw2:promote
> kvm-greatfw2:start
> 
> property $id="cib-bootstrap-options" \
>         dc-version="1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b" \
>         cluster-infrastructure="Heartbeat" \
>         stonith-enabled="false" \
>         no-quorum-policy="ignore" \
>         default-resource-stickiness="200" \
>         last-lrm-refresh="1332228434" \
>         start-failure-is-fatal="false"
> 
> I figure, greatfw1 is configured exactly as greatfw2, but has a
> higher
> resource stickyness, so it should stay where it was. But when I
> standby
> one host (migrating both to the same host) and then take it back
> online,
> greatfw1 is migrated to the other side. 
> Debian Stable, pacemaker 1.0.9.1+hg15626-1 with heartbeat 1:3.0.3-2.
> 
> Best Regards,
> Bernhard
> 
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started:
> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
> 




More information about the Pacemaker mailing list