<div dir="ltr"><div>Update : I tried to spot the problem by running 2 Wheezy virtual machines configured with debian pinning like this :<br><br># cat /etc/apt/preferences<br>Package: *<br>Pin: release a=wheezy<br>Pin-Priority: 900<br>
<br>Package: *<br>Pin: release a=squeeze<br>Pin-Priority: 800<br><br># aptitude install corosync<br># aptitude install pacemaker/squeeze<br><br></div><div>so :<br><br>root@pcmk2:/etc/corosync# dpkg -l | grep pacem<br>ii pacemaker 1.0.9.1+hg15626-1 amd64 HA cluster resource manager<br>
root@pcmk2:/etc/corosync# dpkg -l | grep corosync<br>ii corosync 1.4.2-3 amd64 Standards-based cluster framework (daemon and modules)<br>ii libcorosync4 1.4.2-3 all Standards-based cluster framework (transitional package)<br>
<br><br></div>and the problem did not occur :<br><br>root@pcmk1:~/pacemaker# crm_mon -1<br>============<br>Last updated: Thu Aug 29 05:53:50 2013<br>Stack: openais<br>Current DC: pcmk1 - partition with quorum<br>Version: 1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b<br>
2 Nodes configured, 2 expected votes<br>2 Resources configured.<br>============<br><br>Online: [ pcmk2 pcmk1 ]<br><br> ip (ocf::heartbeat:IPaddr2): Started pcmk1<br> Clone Set: mysql-mm (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started pcmk2 (unmanaged)<br>
mysql:1 (ocf::heartbeat:mysql): Started pcmk1 (unmanaged)<br><br><div><div><div>root@pcmk2:/etc/corosync# /etc/init.d/mysql stop<br>[ ok ] Stopping MySQL database server: mysqld.<br><br>root@pcmk1:~/pacemaker# crm_mon -1<br>
============<br>Last updated: Thu Aug 29 05:55:39 2013<br>Stack: openais<br>Current DC: pcmk1 - partition with quorum<br>Version: 1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b<br>2 Nodes configured, 2 expected votes<br>2 Resources configured.<br>
============<br><br>Online: [ pcmk2 pcmk1 ]<br><br> ip (ocf::heartbeat:IPaddr2): Started pcmk1<br> Clone Set: mysql-mm (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started pcmk2 (unmanaged) FAILED<br> mysql:1 (ocf::heartbeat:mysql): Started pcmk1 (unmanaged)<br>
<br>Failed actions:<br> mysql:0_monitor_15000 (node=pcmk2, call=5, rc=7, status=complete): not running<br><br><br>root@pcmk2:/etc/corosync# /etc/init.d/mysql start<br>[ ok ] Starting MySQL database server: mysqld ..<br>
[info] Checking for tables which need an upgrade, are corrupt or were <br>not closed cleanly..<br><br>root@pcmk1:~/pacemaker# crm_mon -1<br>============<br>Last updated: Thu Aug 29 05:56:34 2013<br>Stack: openais<br>Current DC: pcmk1 - partition with quorum<br>
Version: 1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b<br>2 Nodes configured, 2 expected votes<br>2 Resources configured.<br>============<br><br>Online: [ pcmk2 pcmk1 ]<br><br> ip (ocf::heartbeat:IPaddr2): Started pcmk1<br>
Clone Set: mysql-mm (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started pcmk2 (unmanaged)<br> mysql:1 (ocf::heartbeat:mysql): Started pcmk1 (unmanaged)<br><br><br><br>-----<br><br></div><div>What I noticed :<br>
<br></div><div>with pacemaker 1.1.7, crm see 3 resources configured when in 1.0.9 it sees 2 resources (for the exact same configuration)<br><br></div></div></div><div class="gmail_extra"><br><br><div class="gmail_quote">2013/8/27 tetsuo shima <span dir="ltr"><<a href="mailto:tetsuo.41.shima@gmail.com" target="_blank">tetsuo.41.shima@gmail.com</a>></span><br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div dir="ltr"><div><div><div><div><div><div>Hi list !<br><br></div>I'm having an issue with corosync, here is the scenario :<br>
<br># crm_mon -1<br>============<br>Last updated: Tue Aug 27 09:50:13 2013<br>Last change: Mon Aug 26 16:06:01 2013 via cibadmin on node2<br>
Stack: openais<br>Current DC: node1 - partition with quorum<br>Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff<br>2 Nodes configured, 2 expected votes<br>3 Resources configured.<br>============<br><br>Online: [ node2 node1 ]<br>
<br> ip (ocf::heartbeat:IPaddr2): Started node1<br> Clone Set: mysql-mm [mysql] (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started node1 (unmanaged)<br> mysql:1 (ocf::heartbeat:mysql): Started node2 (unmanaged)<br>
<br># /etc/init.d/mysql stop<br>[ ok ] Stopping MySQL database server: mysqld.<br><br># crm_mon -1<br>============<br>Last updated: Tue Aug 27 09:50:30 2013<br>Last change: Mon Aug 26 16:06:01 2013 via cibadmin on node2<br>
Stack: openais<br>Current DC: node1 - partition with quorum<br>Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff<br>2 Nodes configured, 2 expected votes<br>3 Resources configured.<br>============<br><br>Online: [ node2 node1 ]<br>
<br> ip (ocf::heartbeat:IPaddr2): Started node1<br> Clone Set: mysql-mm [mysql] (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started node1 (unmanaged)<br> mysql:1 (ocf::heartbeat:mysql): Started node2 (unmanaged) FAILED<br>
<br>Failed actions:<br> mysql:0_monitor_15000 (node=node2, call=27, rc=7, status=complete): not running<br><br># /etc/init.d/mysql start<br>[ ok ] Starting MySQL database server: mysqld ..<br>[info] Checking for tables which need an upgrade, are corrupt or were <br>
not closed cleanly..<br><br># sleep 60 && crm_mon -1<br>============<br>Last updated: Tue Aug 27 09:51:54 2013<br>Last change: Mon Aug 26 16:06:01 2013 via cibadmin on node2<br>Stack: openais<br>Current DC: node1 - partition with quorum<br>
Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff<br>2 Nodes configured, 2 expected votes<br>3 Resources configured.<br>============<br><br>Online: [ node2 node1 ]<br><br> ip (ocf::heartbeat:IPaddr2): Started node1<br>
Clone Set: mysql-mm [mysql] (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started node1 (unmanaged)<br> mysql:1 (ocf::heartbeat:mysql): Started node2 (unmanaged) FAILED<br><br>Failed actions:<br> mysql:0_monitor_15000 (node=node2, call=27, rc=7, status=complete): not running<br>
<br></div>As you can see, every time I stop Mysql (which is unmanaged), the resource is marked as failed :<br><br>crmd: [1828]: info: process_lrm_event: LRM operation mysql:0_monitor_15000 (call=4, rc=7, cib-update=10, confirmed=false) not running<br>
<br></div><div>When I restart the resource :<br><br>crmd: [1828]: info: process_lrm_event: LRM operation mysql:0_monitor_15000 (call=4, rc=0, cib-update=11, confirmed=false) ok<br><br></div><div>The resource is still in failed state and does not recover until I manually clean up the resource.<br>
<br># crm_mon --one-shot --operations<br>============<br>Last updated: Tue Aug 27 10:17:30 2013<br>Last change: Mon Aug 26 16:06:01 2013 via cibadmin on node2<br>Stack: openais<br>Current DC: node1 - partition with quorum<br>
Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff<br>2 Nodes configured, 2 expected votes<br>3 Resources configured.<br>============<br><br>Online: [ node2 node1 ]<br><br> ip (ocf::heartbeat:IPaddr2): Started node1<br>
Clone Set: mysql-mm [mysql] (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started node1 (unmanaged)<br> mysql:1 (ocf::heartbeat:mysql): Started node2 (unmanaged) FAILED<br><br>Operations:<br>* Node node1: <br>
ip: migration-threshold=1<br> + (57) probe: rc=0 (ok)<br> mysql:0: migration-threshold=1 fail-count=1<br> + (58) probe: rc=0 (ok)<br> + (59) monitor: interval=15000ms rc=0 (ok)<br>* Node node2: <br> mysql:0: migration-threshold=1 fail-count=3<br>
+ (27) monitor: interval=15000ms rc=7 (not running)<br> + (27) monitor: interval=15000ms rc=0 (ok)<br><br>Failed actions:<br> mysql:0_monitor_15000 (node=node2, call=27, rc=7, status=complete): not running<br><br>
</div><div>---<br><br></div>Here is some details about my configuration :<br><br># cat /etc/debian_version <br>7.1<br><br># dpk# dpkg -l | grep corosync<br>ii corosync 1.4.2-3 amd64 Standards-based cluster framework <br>
<br># dpkg -l | grep pacem <br>ii pacemaker 1.1.7-1 amd64 HA cluster resource manager<br><br># crm configure show<br>node node2 \<br> attributes standby="off"<br>
node node1<br>primitive ip ocf:heartbeat:IPaddr2 \<br> params ip="192.168.0.20" cidr_netmask="255.255.0.0" nic="eth2.2755" iflabel="mysql" \<br> meta is-managed="true" target-role="Started" \<br>
meta resource-stickiness="100"<br>primitive mysql ocf:heartbeat:mysql \<br> op monitor interval="15" timeout="30"<br>clone mysql-mm mysql \<br> meta is-managed="false"<br>
location cli-prefer-ip ip 50: node1<br>colocation ip-on-mysql-mm 200: ip mysql-mm<br>property $id="cib-bootstrap-options" \<br> dc-version="1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff" \<br> cluster-infrastructure="openais" \<br>
expected-quorum-votes="2" \<br> stonith-enabled="false" \<br> no-quorum-policy="ignore" \<br> last-lrm-refresh="1377513557" \<br> start-failure-is-fatal="false"<br>
rsc_defaults $id="rsc-options" \<br> resource-stickiness="1" \<br> migration-threshold="1"<br><br>---<br><br></div>Does anyone know what is wrong with my configuration ?<br><br></div>Thanks for the help,<br>
<br></div>Best regards.<br><div><div><div><div><div><br> <br></div></div></div></div></div></div>
</blockquote></div><br></div></div>