[Pacemaker] fail-count is not updated

David Vossel dvossel at redhat.com
Tue Apr 3 16:45:14 EDT 2012


----- Original Message -----
> From: "Kazunori INOUE" <inouekazu at intellilink.co.jp>
> To: "pacemaker at oss" <pacemaker at oss.clusterlabs.org>
> Cc: tanakakza at intellilink.co.jp
> Sent: Monday, April 2, 2012 12:40:20 AM
> Subject: [Pacemaker]  fail-count is not updated
> 
> Hi, Andrew
> 
> When combined with pacemaker-1.1.7 and corosync-1.99.9,
> fail-count is not updated at the time of monitor failure.
> 
> I am using the newest devel.
> - pacemaker : 7172b7323bb72c51999ce11c6fa5d3ff0a0a4b4f
> - corosync  : 4b2cfc3f6beabe517b28ea31c5340bf3b0a6b455
> - glue      : 041b464f74c8
> - libqb     : 7b13d09afbb684f9ee59def23b155b38a21987df
> 
> # crm_mon -f1
> ============
> Last updated: Mon Apr  2 14:03:03 2012
> Last change: Mon Apr  2 14:02:33 2012 via cibadmin on vm1
> Stack: corosync
> Current DC: vm1 (224766144) - partition with quorum
> Version: 1.1.7-7172b73
> 2 Nodes configured, unknown expected votes
> 1 Resources configured.
> ============
> 
> Online: [ vm1 vm2 ]
> 
>  prmDummy1      (ocf::pacemaker:Dummy): Started vm1
> 
> Migration summary:
> * Node vm1:
> * Node vm2:
> 
> Failed actions:
>     prmDummy1_monitor_10000 (node=vm1, call=4, rc=7,
>     status=complete): not running
> #
> 
> I think this is because corosync's nodeID and hostname are
> intermingled in the
> value which identifies a cluster node.
> I added the debugging code. (l.769)
> 
> # vi pacemaker/tools/attrd.c
> <snip>
> 752 void
> 753 attrd_local_callback(xmlNode * msg)
> 754 {
> <snip>
> 768
> 769 crm_info("DEBUG:
> [%s,%s,%s,%s,%s],[%s]\n",from,op,attr,value,host,attrd_uname);
> 770     if (host != NULL && safe_str_neq(host, attrd_uname)) {
> 771         send_cluster_message(host, crm_msg_attrd, msg, FALSE);
> 772         return;
> 773     }
> 774
> 775     crm_debug("%s message from %s: %s=%s", op, from, attr,
> crm_str(value));
> 
> [root at vm1 ~]# grep DEBUG /var/log/ha-debug
> <snip>
> Apr  2 14:02:34 vm1 Dummy(prmDummy1)[21140]: DEBUG: prmDummy1 monitor
> : 7
> Apr  2 14:02:34 vm1 attrd[21077]:     info: attrd_local_callback:
> DEBUG: [crmd,update,probe_complete,true,(null)],[vm1]
> Apr  2 14:02:34 vm1 Dummy(prmDummy1)[21151]: DEBUG: prmDummy1 start :
> 0
> Apr  2 14:02:34 vm1 Dummy(prmDummy1)[21159]: DEBUG: prmDummy1 monitor
> : 0
> Apr  2 14:02:44 vm1 Dummy(prmDummy1)[21166]: DEBUG: prmDummy1 monitor
> : 0
> Apr  2 14:02:54 vm1 Dummy(prmDummy1)[21175]: DEBUG: prmDummy1 monitor
> : 7
> Apr  2 14:02:54 vm1 attrd[21077]:     info: attrd_local_callback:
> DEBUG: [crmd,update,fail-count-prmDummy1,value++,224766144],[vm1]
> Apr  2 14:02:54 vm1 attrd[21077]:     info: attrd_local_callback:
> DEBUG:
> [crmd,update,last-failure-prmDummy1,1333342974,224766144],[vm1]
> Apr  2 14:02:54 vm1 Dummy(prmDummy1)[21182]: DEBUG: prmDummy1 stop :
> 0
> Apr  2 14:02:54 vm1 Dummy(prmDummy1)[21189]: DEBUG: prmDummy1 start :
> 0
> Apr  2 14:02:54 vm1 Dummy(prmDummy1)[21201]: DEBUG: prmDummy1 monitor
> : 0
> 
> Corosync's nodeID was stored in variable 'host', and hostname was
> stored in
> variable 'attrd_uname'..
> 
> [root at vm1 ~]# corosync-cfgtool -s | grep node
> Local node ID 224766144
> [root at vm1 ~]#
> 
> Regards,
> Kazunori INOUE

Yep, I am seeing this as well.  I created a bug report for the issue here, http://bugs.clusterlabs.org/show_bug.cgi?id=5053.  I'll take a shot at fixing this soon.

-- Vossel




More information about the Pacemaker mailing list