<div dir="ltr"><div><div>Hi, Andrew:<br></div>If I cut down the network connection of the running
node by:<br>service network stop, <br>"crm status" will show me the node is put into "OFFLINE" status.
The affected resource can also be failed over to another online node
correctly. But the issue is that, when I re-connect the network by:<br>service network start.<br>to put
the "OFFLINE" node to be "Online" again, all the resource is firstly
stopped , then some resource are restarted again on the original online
node and some other resource are going back to the newly "Online" node.
This behavior seems not related to the resource-stickiness
configuration.<br>
</div><div>I'm just curious if it's the default behavior.<br></div><div>And if I tried to reboot the OFFLINE node, when it's online again, the resource won't be stopped.<br></div><div>Is this expected that "service network start" triggers Pacemaker to reassign resource?<br>
</div>Thanks.<div class=""><div id=":2p" class="" tabindex="0"><img class="" src="https://mail.google.com/mail/u/0/images/cleardot.gif"></div></div></div><div class="gmail_extra"><br><br><div class="gmail_quote">On Fri, Jun 14, 2013 at 10:06 AM, Andrew Beekhof <span dir="ltr"><<a href="mailto:andrew@beekhof.net" target="_blank">andrew@beekhof.net</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div class="im"><br>
On 13/06/2013, at 5:15 PM, Xiaomin Zhang <<a href="mailto:zhangxiaomin@gmail.com">zhangxiaomin@gmail.com</a>> wrote:<br>
<br>
> Thanks Andrew.<br>
> Yes, the fs_ssn service (ocf:FileSystem) is still running when the machine loses network. I configure it as primitive:<br>
> primitive fs_ssn ocf:heartbeat:Filesystem \<br>
> op monitor interval="15s" \<br>
> params device="/dev/drbd0" directory="/drbd" fstype="ext3" \<br>
> meta target-role="Started"<br>
> As I assume this resource can only be started on 1 node, I think it should be stopped automatically when pacemaker detects it's not in a HA cluster.<br>
> Is this incorrect assumption?<br>
<br>
</div>No. But I'd need to see logs from all the nodes (please use attachments) to be able to comment further.<br>
<div class="HOEnZb"><div class="h5"><br>
> Thanks.<br>
><br>
><br>
><br>
> On Thu, Jun 13, 2013 at 1:50 PM, Andrew Beekhof <<a href="mailto:andrew@beekhof.net">andrew@beekhof.net</a>> wrote:<br>
><br>
> On 13/06/2013, at 2:43 PM, Xiaomin Zhang <<a href="mailto:zhangxiaomin@gmail.com">zhangxiaomin@gmail.com</a>> wrote:<br>
><br>
> > Andrew Beekhof <andrew@...> writes:<br>
> ><br>
> >><br>
> >> Try increasing your stickiness as it is being exceeded by the location<br>
> > constraints.<br>
> >> For the biggest stick, try 'infinity' which means - never move unless the<br>
> > node dies.<br>
> >><br>
> ><br>
> > Thanks, Andrew, I applied infinity resource stickiness. However, the sst<br>
> > resource is still switched to the node which is online back from failure.<br>
> > And I found sth in the log:<br>
> ><br>
> > Jun 13 11:46:29 node3 pengine[27813]: warning: unpack_rsc_op: Processing<br>
> > failed op monitor for ip_ssn on node2: not running (7)<br>
> > Jun 13 11:46:29 node3 pengine[27813]: error: native_create_actions:<br>
> > Resource fs_ssn (ocf::Filesystem) is active on 2 nodes attempting recovery<br>
> > Jun 13 11:46:29 node3 pengine[27813]: warning: native_create_actions: See<br>
> > <a href="http://clusterlabs.org/wiki/FAQ#Resource_is_Too_Active" target="_blank">http://clusterlabs.org/wiki/FAQ#Resource_is_Too_Active</a> for more information.<br>
> ><br>
> > Is this log showing that pacemaker tries to restart all the resource when<br>
> > the failed node is back again?<br>
><br>
> No, thats a log showing the services were already running there when pacemaker started.<br>
><br>
> ><br>
> ><br>
> >>> Thanks.<br>
> >>><br>
> >>> Below is my configure:<br>
> >>> ------------------CONFIG START--------------------------------------<br>
> >>> node node3 \<br>
> >>> attributes standby="on"<br>
> >>> node node1<br>
> >>> node node2<br>
> >>> primitive drbd_ssn ocf:linbit:drbd \<br>
> >>> params drbd_resource="r0" \<br>
> >>> op monitor interval="15s"<br>
> >>> primitive fs_ssn ocf:heartbeat:Filesystem \<br>
> >>> op monitor interval="15s" \<br>
> >>> params device="/dev/drbd0" directory="/drbd" fstype="ext3" \<br>
> >>> meta target-role="Started"<br>
> >>> primitive ip_ssn ocf:heartbeat:IPaddr2 \<br>
> >>> params ip="192.168.241.1" cidr_netmask="32" \<br>
> >>> op monitor interval="15s" \<br>
> >>> meta target-role="Started"<br>
> >>> primitive ip_sst ocf:heartbeat:IPaddr2 \<br>
> >>> params ip="192.168.241.2" cidr_netmask="32" \<br>
> >>> op monitor interval="15s" \<br>
> >>> meta target-role="Started"<br>
> >>> primitive sst lsb:sst \<br>
> >>> op monitor interval="15s" \<br>
> >>> meta target-role="stopped"<br>
> >>> primitive ssn lsb:ssn \<br>
> >>> op monitor interval="15s" \<br>
> >>> meta target-role="stopped"<br>
> >>> ms ms_drbd_ssn drbd_ssn \<br>
> >>> meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1"<br>
> >>> notify="true" target-role="Started"<br>
> >>> location sst_ip_prefer ip_sst 50: node1<br>
> >>> location drbd_ssn_prefer ms_drbd_ssn 50: node1<br>
> >>> colocation fs_ssn_coloc inf: ip_ssn fs_ssn<br>
> >>> colocation fs_on_drbd_coloc inf: fs_ssn ms_drbd_ssn:Master<br>
> >>> colocation sst_ip_coloc inf: sst ip_sst<br>
> >>> colocation ssn_ip_coloc inf: ssn ip_ssn<br>
> >>> order ssn_after_drbd inf: ms_drbd_ssn:promote fs_ssn:start<br>
> >>> order ip_after_fs inf: fs_ssn:start ip_ssn:start<br>
> >>> order sst_after_ip inf: ip_sst:start sst:start<br>
> >>> order sst_after_ssn inf: ssn:start sst:start<br>
> >>> order ssn_after_ip inf: ip_ssn:start ssn:start<br>
> >>> property $id="cib-bootstrap-options" \<br>
> >>> dc-version="1.1.8-7.el6-394e906" \<br>
> >>> cluster-infrastructure="classic openais (with plugin)" \<br>
> >>> expected-quorum-votes="3" \<br>
> >>> stonith-enabled="false"<br>
> >>> rsc_defaults $id="rsc-options" \<br>
> >>> resource-stickiness="100"<br>
> >>><br>
> >>> -------------------CONFIG END----------------------------------------<br>
> >>><br>
> > Best Regards.<br>
> > Xiaomin<br>
> ><br>
> ><br>
> ><br>
> ><br>
> ><br>
> ><br>
> ><br>
> > _______________________________________________<br>
> > Pacemaker mailing list: <a href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>
> > <a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker" target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>
> ><br>
> > Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>
> > Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>
> > Bugs: <a href="http://bugs.clusterlabs.org" target="_blank">http://bugs.clusterlabs.org</a><br>
><br>
><br>
> _______________________________________________<br>
> Pacemaker mailing list: <a href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>
> <a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker" target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>
><br>
> Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>
> Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>
> Bugs: <a href="http://bugs.clusterlabs.org" target="_blank">http://bugs.clusterlabs.org</a><br>
><br>
> _______________________________________________<br>
> Pacemaker mailing list: <a href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>
> <a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker" target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>
><br>
> Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>
> Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>
> Bugs: <a href="http://bugs.clusterlabs.org" target="_blank">http://bugs.clusterlabs.org</a><br>
<br>
<br>
_______________________________________________<br>
Pacemaker mailing list: <a href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>
<a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker" target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>
<br>
Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>
Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>
Bugs: <a href="http://bugs.clusterlabs.org" target="_blank">http://bugs.clusterlabs.org</a><br>
</div></div></blockquote></div><br></div>