[ClusterLabs] Help with service banning on a node
    Leon Botes 
    leon at trusc.net
       
    Mon May 16 06:36:10 UTC 2016
    
    
  
Hi List.
I have the following configuration:
pcs -f ha_config property set symmetric-cluster="true"
pcs -f ha_config property set no-quorum-policy="stop"
pcs -f ha_config property set stonith-enabled="false"
pcs -f ha_config resource defaults resource-stickiness="200"
pcs -f ha_config resource create drbd ocf:linbit:drbd drbd_resource=r0 
op monitor interval=60s
pcs -f ha_config resource master drbd master-max=1 master-node-max=1 
clone-max=2 clone-node-max=1 notify=true
pcs -f ha_config resource create vip-blue ocf:heartbeat:IPaddr2 
ip=192.168.101.100 cidr_netmask=32 nic=blue op monitor interval=20s
pcs -f ha_config resource create vip-green ocf:heartbeat:IPaddr2 
ip=192.168.102.100 cidr_netmask=32 nic=blue op monitor interval=20s
pcs -f ha_config constraint colocation add vip-blue drbd-master INFINITY 
with-rsc-role=Master
pcs -f ha_config constraint colocation add vip-green drbd-master 
INFINITY with-rsc-role=Master
pcs -f ha_config constraint location drbd-master prefers stor-san1=50
pcs -f ha_config constraint location drbd-master avoids stor-node1=INFINITY
pcs -f ha_config constraint location vip-blue prefers stor-san1=50
pcs -f ha_config constraint location vip-blue avoids stor-node1=INFINITY
pcs -f ha_config constraint location vip-green prefers stor-san1=50
pcs -f ha_config constraint location vip-green avoids stor-node1=INFINITY
pcs -f ha_config constraint order promote drbd-master then start vip-blue
pcs -f ha_config constraint order start vip-blue then start vip-green
Which results in:
[root at san1 ~]# pcs status
Cluster name: ha_cluster
Last updated: Mon May 16 08:21:28 2016          Last change: Mon May 16 
08:21:25 2016 by root via crm_resource on iscsiA-san1
Stack: corosync
Current DC: iscsiA-node1 (version 1.1.13-10.el7_2.2-44eb2dd) - partition 
with quorum
3 nodes and 4 resources configured
Online: [ iscsiA-node1 iscsiA-san1 iscsiA-san2 ]
Full list of resources:
  Master/Slave Set: drbd-master [drbd]
      drbd       (ocf::linbit:drbd):     FAILED iscsiA-node1 (unmanaged)
      Masters: [ iscsiA-san1 ]
      Stopped: [ iscsiA-san2 ]
  vip-blue       (ocf::heartbeat:IPaddr2):       Started iscsiA-san1
  vip-green      (ocf::heartbeat:IPaddr2):       Started iscsiA-san1
Failed Actions:
* drbd_stop_0 on iscsiA-node1 'not installed' (5): call=18, 
status=complete, exitreason='none',
     last-rc-change='Mon May 16 08:20:16 2016', queued=0ms, exec=45ms
PCSD Status:
   iscsiA-san1: Online
   iscsiA-san2: Online
   iscsiA-node1: Online
Daemon Status:
   corosync: active/disabled
   pacemaker: active/disabled
   pcsd: active/enabled
Is there any way in the configuration to make the drbd sections 
completely be ignored on  iscsiA-node1  to avoide this:
  drbd (ocf::linbit:drbd): FAILED iscsiA-node1 (unmanaged)
and
Failed Actions:
* drbd_stop_0 on iscsiA-node1 'not installed' (5): call=18, 
status=complete, exitreason='none',
last-rc-change='Mon May 16 08:20:16 2016', queued=0ms, exec=45ms
Tried the ban statements but that seesm to have the same result.
Also is there any better way to write the configuration so that the drbd 
starts first then the vip's and colocate together. Also ensure that they 
run on only san1 or san2. Tried grouping but that seems to fail with 
Master / Slave resourcess.
-- 
Regards
Leon
    
    
More information about the Users
mailing list