[Pacemaker] standby attribute and same resources running at the	same time
    Leon Fauster 
    leonfauster at googlemail.com
       
    Mon Mar  4 12:20:41 EST 2013
    
    
  
Dear list,
just to excuse the triviality - i started to deploy a ha environment 
in a test lab and therefore i do not have much experience. 
i started to setup a 2-node cluster 
  corosync-1.4.1-15.el6.x86_64
  pacemaker-1.1.8-7.el6.x86_64
  cman-3.0.12.1-49.el6.x86_64
with rhel6.3 and then switched to rhel6.4. 
This update brings some differences. The crm shell is gone and pcs is added.
Anyway i found some equivalent commands to setup/configure resources. 
So far all good. I am doing some stress test now and noticed that rebooting
one node (n2), that node (n2) will be marked as standby in the cib (shown on the 
other node (n1)).
After rebooting the node (n2) crm_mon on that node shows that the other node (n1) 
is offline and begins to start the ressources. While the other node (n1) that wasn't
rebooted still shows n2 as standby. At that point both nodes are runnnig the "same" 
resources. After a couple of minutes that situation is noticed and both nodes 
renegotiate the current state. Then one node take over the responsibility to provide
the resources. On both nodes the previously rebooted node is still listed as standby.
  cat /var/log/messages |grep error
  Mar  4 17:32:33 cn1 pengine[1378]:    error: native_create_actions: Resource resIP (ocf::IPaddr2) is active on 2 nodes attempting recovery
  Mar  4 17:32:33 cn1 pengine[1378]:    error: native_create_actions: Resource resApache (ocf::apache) is active on 2 nodes attempting recovery
  Mar  4 17:32:33 cn1 pengine[1378]:    error: process_pe_message: Calculated Transition 1: /var/lib/pacemaker/pengine/pe-error-6.bz2
  Mar  4 17:32:48 cn1 crmd[1379]:   notice: run_graph: Transition 1 (Complete=9, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-error-6.bz2): Complete
  crm_mon -1
  Last updated: Mon Mar  4 17:49:08 2013
  Last change: Mon Mar  4 10:22:53 2013 via crm_resource on cn1.localdomain
  Stack: cman
  Current DC: cn1.localdomain - partition with quorum
  Version: 1.1.8-7.el6-394e906
  2 Nodes configured, 2 expected votes
  2 Resources configured.
  Node cn2.localdomain: standby
  Online: [ cn1.localdomain ]
  resIP	(ocf::heartbeat:IPaddr2):	Started cn1.localdomain
  resApache	(ocf::heartbeat:apache):	Started cn1.localdomain
i checked the init scripts and found that the standby "behavior" comes
from a function that is called on "service pacemaker stop" (added in rhel6.4).
cman_pre_stop() 
{
    cname=`crm_node --name`
    crm_attribute -N $cname -n standby -v true -l reboot
    echo -n "Waiting for shutdown of managed resources"
...
i could not delete the standby attribute with
crm_attribute -G --node=cn2.localdomain -n standby
okay - recap: 
1st. i have this delay where the two nodes dont see each 
other (after rebooting) and the result are resources running on both 
nodes while they should only run on one node - this will be corrected 
by the cluster it self but this situation should not happen.
2nd. the standby attribute (and there must be a reason why redhat 
added this) will prevent to migrate resources to that node. How 
do i delete this attribute?
i appreciate any comments
--
Leon
A. $ cat /etc/cluster/cluster.conf
<?xml version="1.0"?>
 <cluster name="HA" config_version="5">
   <logging debug="off"/>
   <clusternodes>
     <clusternode name="cn1.localdomain" votes="1" nodeid="1">
       <fence>
         <method name="pcmk-redirect">
           <device name="pcmk" port="cn1.localdomain"/>
         </method>
       </fence>
     </clusternode>
     <clusternode name="cn2.localdomain" votes="1" nodeid="2">
       <fence>
         <method name="pcmk-redirect">
           <device name="pcmk" port="cn2.localdomain"/>
         </method>
       </fence>
     </clusternode>
   </clusternodes>
   <fencedevices>
     <fencedevice name="pcmk" agent="fence_pcmk"/>
   </fencedevices>
   <rm>
     <failoverdomains/>
     <resources/>
   </rm>
 </cluster>
B. $ pcs config
Corosync Nodes:
 
Pacemaker Nodes:
 cn1.localdomain cn2.localdomain 
Resources: 
 Resource: resIP (provider=heartbeat type=IPaddr2 class=ocf)
  Attributes: ip=192.168.201.220 nic=eth0 cidr_netmask=24 
  Operations: monitor interval=30s
 Resource: resApache (provider=heartbeat type=apache class=ocf)
  Attributes: httpd=/usr/sbin/httpd configfile=/etc/httpd/conf/httpd.conf 
  Operations: monitor interval=1min
Location Constraints:
Ordering Constraints:
  start resApache then start resIP
Colocation Constraints:
  resIP with resApache
Cluster Properties:
 dc-version: 1.1.8-7.el6-394e906
 cluster-infrastructure: cman
 expected-quorum-votes: 2
 stonith-enabled: false
 no-quorum-policy: ignore
    
    
More information about the Pacemaker
mailing list