[ClusterLabs] Default resource stickiness issue with colocation constraint

Strahil Nikolov hunter86_bg at yahoo.com
Sun Mar 8 14:11:09 EDT 2020

Hello All,

can someone help me figure something out.

I have a test cluster with 2 resource groups:

[root at node3 cluster]# pcs status
Cluster name: HACLUSTER16
Stack: corosync
Current DC: node3.localdomain (version 1.1.20-5.el7_7.2-3c4c782f70) - partition with quorum
Last updated: Sun Mar  8 20:00:48 2020
Last change: Sun Mar  8 20:00:04 2020 by root via cibadmin on node3.localdomain

3 nodes configured
14 resources configured

Node node2.localdomain: standby
Node node3.localdomain: standby
Online: [ node1.localdomain ]

Full list of resources:

 RHEVM  (stonith:fence_rhevm):  Started node1.localdomain
 MPATH  (stonith:fence_mpath):  Started node1.localdomain
 Resource Group: NFS
     NFS_LVM    (ocf::heartbeat:LVM):   Started node1.localdomain
     NFS_infodir        (ocf::heartbeat:Filesystem):    Started node1.localdomain
     NFS_data   (ocf::heartbeat:Filesystem):    Started node1.localdomain
     NFS_IP     (ocf::heartbeat:IPaddr2):       Started node1.localdomain
     NFS_SRV    (ocf::heartbeat:nfsserver):     Started node1.localdomain
     NFS_XPRT1  (ocf::heartbeat:exportfs):      Started node1.localdomain
     NFS_NTFY   (ocf::heartbeat:nfsnotify):     Started node1.localdomain
 Resource Group: APACHE
     APACHE_LVM (ocf::heartbeat:LVM):   Started node1.localdomain
     APACHE_cfg (ocf::heartbeat:Filesystem):    Started node1.localdomain
     APACHE_data        (ocf::heartbeat:Filesystem):    Started node1.localdomain
     APACHE_IP  (ocf::heartbeat:IPaddr2):       Started node1.localdomain
     APACHE_SRV (ocf::heartbeat:apache):        Started node1.localdomain

The constraints I have put are:

[root at node3 cluster]# pcs constraint
Location Constraints:
  Resource: APACHE
    Enabled on: node1.localdomain (score:3000)
    Enabled on: node2.localdomain (score:2000)
    Enabled on: node3.localdomain (score:1000)
  Resource: NFS
    Enabled on: node1.localdomain (score:1000)
    Enabled on: node2.localdomain (score:2000)
    Enabled on: node3.localdomain (score:3000)
Ordering Constraints:
Colocation Constraints:
  APACHE with NFS (score:-1000)
Ticket Constraints:

[root at node3 cluster]# pcs resource defaults

As you can see the default stickiness is 1000 per resource or 7000 for the APACHE group.
The colocation rule score is just -1000 and as per my understanding it should be ignored when the 2 nodes are removed from standby.

Can someone clarify why the APACHE group is moved , when the resource stickiness score is higher than the colocation score.

I have attached a file with the crm_simulate output (the output is correct, when the standby is removed - the group is moved).

Best Regards,
Strahil Nikolov
-------------- next part --------------
An embedded and charset-unspecified text was scrubbed...
Name: resource-stickiness_vs_colocation.txt
URL: <http://lists.clusterlabs.org/pipermail/users/attachments/20200308/4cb10fbd/attachment.txt>

More information about the Users mailing list