<div dir="ltr"><div><div><div><div><div><div>Hi list !<br><br></div>I'm having an issue with corosync, here is the scenario :<br><br># crm_mon -1<br>============<br>Last updated: Tue Aug 27 09:50:13 2013<br>Last change: Mon Aug 26 16:06:01 2013 via cibadmin on node2<br>
Stack: openais<br>Current DC: node1 - partition with quorum<br>Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff<br>2 Nodes configured, 2 expected votes<br>3 Resources configured.<br>============<br><br>Online: [ node2 node1 ]<br>
<br> ip (ocf::heartbeat:IPaddr2): Started node1<br> Clone Set: mysql-mm [mysql] (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started node1 (unmanaged)<br> mysql:1 (ocf::heartbeat:mysql): Started node2 (unmanaged)<br>
<br># /etc/init.d/mysql stop<br>[ ok ] Stopping MySQL database server: mysqld.<br><br># crm_mon -1<br>============<br>Last updated: Tue Aug 27 09:50:30 2013<br>Last change: Mon Aug 26 16:06:01 2013 via cibadmin on node2<br>
Stack: openais<br>Current DC: node1 - partition with quorum<br>Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff<br>2 Nodes configured, 2 expected votes<br>3 Resources configured.<br>============<br><br>Online: [ node2 node1 ]<br>
<br> ip (ocf::heartbeat:IPaddr2): Started node1<br> Clone Set: mysql-mm [mysql] (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started node1 (unmanaged)<br> mysql:1 (ocf::heartbeat:mysql): Started node2 (unmanaged) FAILED<br>
<br>Failed actions:<br> mysql:0_monitor_15000 (node=node2, call=27, rc=7, status=complete): not running<br><br># /etc/init.d/mysql start<br>[ ok ] Starting MySQL database server: mysqld ..<br>[info] Checking for tables which need an upgrade, are corrupt or were <br>
not closed cleanly..<br><br># sleep 60 && crm_mon -1<br>============<br>Last updated: Tue Aug 27 09:51:54 2013<br>Last change: Mon Aug 26 16:06:01 2013 via cibadmin on node2<br>Stack: openais<br>Current DC: node1 - partition with quorum<br>
Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff<br>2 Nodes configured, 2 expected votes<br>3 Resources configured.<br>============<br><br>Online: [ node2 node1 ]<br><br> ip (ocf::heartbeat:IPaddr2): Started node1<br>
Clone Set: mysql-mm [mysql] (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started node1 (unmanaged)<br> mysql:1 (ocf::heartbeat:mysql): Started node2 (unmanaged) FAILED<br><br>Failed actions:<br> mysql:0_monitor_15000 (node=node2, call=27, rc=7, status=complete): not running<br>
<br></div>As you can see, every time I stop Mysql (which is unmanaged), the resource is marked as failed :<br><br>crmd: [1828]: info: process_lrm_event: LRM operation mysql:0_monitor_15000 (call=4, rc=7, cib-update=10, confirmed=false) not running<br>
<br></div><div>When I restart the resource :<br><br>crmd: [1828]: info: process_lrm_event: LRM operation mysql:0_monitor_15000 (call=4, rc=0, cib-update=11, confirmed=false) ok<br><br></div><div>The resource is still in failed state and does not recover until I manually clean up the resource.<br>
<br># crm_mon --one-shot --operations<br>============<br>Last updated: Tue Aug 27 10:17:30 2013<br>Last change: Mon Aug 26 16:06:01 2013 via cibadmin on node2<br>Stack: openais<br>Current DC: node1 - partition with quorum<br>
Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff<br>2 Nodes configured, 2 expected votes<br>3 Resources configured.<br>============<br><br>Online: [ node2 node1 ]<br><br> ip (ocf::heartbeat:IPaddr2): Started node1<br>
Clone Set: mysql-mm [mysql] (unmanaged)<br> mysql:0 (ocf::heartbeat:mysql): Started node1 (unmanaged)<br> mysql:1 (ocf::heartbeat:mysql): Started node2 (unmanaged) FAILED<br><br>Operations:<br>* Node node1: <br>
ip: migration-threshold=1<br> + (57) probe: rc=0 (ok)<br> mysql:0: migration-threshold=1 fail-count=1<br> + (58) probe: rc=0 (ok)<br> + (59) monitor: interval=15000ms rc=0 (ok)<br>* Node node2: <br> mysql:0: migration-threshold=1 fail-count=3<br>
+ (27) monitor: interval=15000ms rc=7 (not running)<br> + (27) monitor: interval=15000ms rc=0 (ok)<br><br>Failed actions:<br> mysql:0_monitor_15000 (node=node2, call=27, rc=7, status=complete): not running<br><br>
</div><div>---<br><br></div>Here is some details about my configuration :<br><br># cat /etc/debian_version <br>7.1<br><br># dpk# dpkg -l | grep corosync<br>ii corosync 1.4.2-3 amd64 Standards-based cluster framework <br>
<br># dpkg -l | grep pacem <br>ii pacemaker 1.1.7-1 amd64 HA cluster resource manager<br><br># crm configure show<br>node node2 \<br> attributes standby="off"<br>
node node1<br>primitive ip ocf:heartbeat:IPaddr2 \<br> params ip="192.168.0.20" cidr_netmask="255.255.0.0" nic="eth2.2755" iflabel="mysql" \<br> meta is-managed="true" target-role="Started" \<br>
meta resource-stickiness="100"<br>primitive mysql ocf:heartbeat:mysql \<br> op monitor interval="15" timeout="30"<br>clone mysql-mm mysql \<br> meta is-managed="false"<br>
location cli-prefer-ip ip 50: node1<br>colocation ip-on-mysql-mm 200: ip mysql-mm<br>property $id="cib-bootstrap-options" \<br> dc-version="1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff" \<br> cluster-infrastructure="openais" \<br>
expected-quorum-votes="2" \<br> stonith-enabled="false" \<br> no-quorum-policy="ignore" \<br> last-lrm-refresh="1377513557" \<br> start-failure-is-fatal="false"<br>
rsc_defaults $id="rsc-options" \<br> resource-stickiness="1" \<br> migration-threshold="1"<br><br>---<br><br></div>Does anyone know what is wrong with my configuration ?<br><br></div>Thanks for the help,<br>
<br></div>Best regards.<br><div><div><div><div><div><br> <br></div></div></div></div></div></div>