[Pacemaker] help building 2 node config

Alex Samad - Yieldbroker Alex.Samad at yieldbroker.com
Tue Mar 18 02:36:54 UTC 2014


Hi



> -----Original Message-----
> From: Andrew Beekhof [mailto:andrew at beekhof.net]
> Sent: Tuesday, 18 March 2014 11:51 AM
> To: The Pacemaker cluster resource manager
> Subject: Re: [Pacemaker] help building 2 node config
> 
> 
> On 13 Mar 2014, at 4:13 pm, Alex Samad - Yieldbroker
> <Alex.Samad at yieldbroker.com> wrote:
> 
[snip]
> 
> 
> ^^^ you only need the -clone versions of these constraints.
> other than that its fine
> 
> >
> >
> > Have I done anything silly ?
> >
> > Also as I don't have the application actually running on my nodes, I notice
> fails occur very fast, more than 1  sec, where its that configured and how do I
> configure it such that after 2 or 3,4 or 5 attempts it fails over to the other
> node.
> 
> Look for migration-threshold in the docs

okay

> 
> > I also want then resources to move back to the original nodes when
> > they come back
> 
> resource-stickiness=0

Added as a meta 

> 
> >
[snip]


I have made some slight changes, firstly if I don't use the IP load balancing it works fine, how I expect, it's the floating IP and its has that's the problem

pcs resource create ybrpip ocf:heartbeat:IPaddr2 params ip=10.32.21.20 cidr_netmask=24 nic=eth0 clusterip_hash=sourceip-sourceport \
    op start interval="0s" timeout="60s" \
    op monitor interval="5s" timeout="20s" \
    op stop interval="0s" timeout="60s" \
	
pcs resource meta ybrpip stickiness=0


pcs resource create ybrpstat ocf:yb:proxy op \
    op start interval="10s" timeout="60s" \
    op monitor interval="5s" timeout="20s" \
    op stop interval="10s" timeout="60s" \
	
pcs resource meta ybrpstat stickiness=0


# This runs on both boxes at the same time no issue
pcs resource clone ybrpstat globally-unique=false clone-max=2 clone-node-max=1
# has to be unique (works on ip hash)
pcs resource clone ybrpip globally-unique=true clone-max=2 clone-node-max=2


# So this I hope states that IP can't be started unless ybrpstat is okay
pcs constraint colocation add ybrpip-clone ybrpstat-clone INFINITY

# this states stat must be started before ip 
pcs constraint order ybrpstat-clone then ybrpip-clone

# not sure about this
pcs constraint location ybrpip-clone prefers dc1devrp01

I think the last line say that the clone of ybrpip prefers dc1devrp01, does that mean the other resource prefers the other node

pcs status
pcs status
Cluster name: ybrp
Last updated: Tue Mar 18 13:31:29 2014
Last change: Tue Mar 18 13:26:51 2014 via cibadmin on alcdevrp01
Stack: cman
Current DC: dc1devrp01 - partition with quorum
Version: 1.1.10-14.el6-368c726
2 Nodes configured
4 Resources configured


Online: [ alcdevrp01 dc1devrp01 ]

Full list of resources:

 Clone Set: ybrpstat-clone [ybrpstat]
     Started: [ alcdevrp01 dc1devrp01 ]
 Clone Set: ybrpip-clone [ybrpip] (unique)
     ybrpip:0   (ocf::heartbeat:IPaddr2):       Started dc1devrp01 
     ybrpip:1   (ocf::heartbeat:IPaddr2):       Started dc1devrp01

this is after I rebooted the alcdevrp01, the resources that were on it moved over to the dc1devrp01, but I want one of the ybrpip resources to move back to it. I guess another way of saying this is, can I place a constraint on an instance of 1 of the cloned resources ...


So this is the result after reboot dc1devrp01
pcs status
Cluster name: ybrp
Last updated: Tue Mar 18 13:32:08 2014
Last change: Tue Mar 18 13:26:51 2014 via cibadmin on alcdevrp01
Stack: cman
Current DC: alcdevrp01 - partition with quorum
Version: 1.1.10-14.el6-368c726
2 Nodes configured
4 Resources configured


Online: [ alcdevrp01 dc1devrp01 ]

Full list of resources:

 Clone Set: ybrpstat-clone [ybrpstat]
     Started: [ alcdevrp01 dc1devrp01 ]
 Clone Set: ybrpip-clone [ybrpip] (unique)
     ybrpip:0   (ocf::heartbeat:IPaddr2):       Started alcdevrp01 
     ybrpip:1   (ocf::heartbeat:IPaddr2):       Started dc1devrp01 


the ybrpip resource rebalanced themselves, I am guessing because I have the constraint for location

pcs config
Cluster Name: ybrp
Corosync Nodes:
 
Pacemaker Nodes:
 alcdevrp01 dc1devrp01 

Resources: 
 Clone: ybrpstat-clone
  Meta Attrs: globally-unique=false clone-max=2 clone-node-max=1 
  Resource: ybrpstat (class=ocf provider=yb type=proxy)
   Meta Attrs: stickiness=0 
   Operations: start interval=10s timeout=60s (ybrpstat-start-interval-10s)
               monitor interval=5s timeout=20s (ybrpstat-monitor-interval-5s)
               stop interval=10s timeout=60s (ybrpstat-stop-interval-10s)
 Clone: ybrpip-clone
  Meta Attrs: globally-unique=true clone-max=2 clone-node-max=2 
  Resource: ybrpip (class=ocf provider=heartbeat type=IPaddr2)
   Attributes: ip=10.32.21.20 cidr_netmask=24 nic=eth0 clusterip_hash=sourceip-sourceport 
   Meta Attrs: stickiness=0 
   Operations: start interval=0s timeout=60s (ybrpip-start-interval-0s)
               monitor interval=5s timeout=20s (ybrpip-monitor-interval-5s)
               stop interval=0s timeout=60s (ybrpip-stop-interval-0s)

Stonith Devices: 
Fencing Levels: 

Location Constraints:
  Resource: ybrpip-clone
    Enabled on: dc1devrp01 (score:INFINITY) (id:location-ybrpip-clone-dc1devrp01-INFINITY)
Ordering Constraints:
  start ybrpstat-clone then start ybrpip-clone (Mandatory) (id:order-ybrpstat-clone-ybrpip-clone-mandatory)
Colocation Constraints:
  ybrpip-clone with ybrpstat-clone (INFINITY) (id:colocation-ybrpip-clone-ybrpstat-clone-INFINITY)

Cluster Properties:
 cluster-infrastructure: cman
 dc-version: 1.1.10-14.el6-368c726
 no-quorum-policy: ignore
 stonith-enabled: false



If I have these constraints on location how do I manually move a resource from one node to another.... do I just move the node into standby mode ?

A













More information about the Pacemaker mailing list