<div dir="ltr">Thanks Andrew.<br>Yes, the fs_ssn service (ocf:FileSystem) is still running when the machine loses network. I configure it as primitive:<br>primitive fs_ssn ocf:heartbeat:Filesystem \<br> op monitor interval="15s" \<br>
params device="/dev/drbd0" directory="/drbd" fstype="ext3" \<br> meta target-role="Started"<br>As I assume this resource can only be started on 1 node, I think it should be stopped automatically when pacemaker detects it's not in a HA cluster.<br>
Is this incorrect assumption?<br>Thanks.<br><br><div class="gmail_extra"><br><br><div class="gmail_quote">On Thu, Jun 13, 2013 at 1:50 PM, Andrew Beekhof <span dir="ltr"><<a href="mailto:andrew@beekhof.net" target="_blank">andrew@beekhof.net</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div class="im"><br>
On 13/06/2013, at 2:43 PM, Xiaomin Zhang <<a href="mailto:zhangxiaomin@gmail.com">zhangxiaomin@gmail.com</a>> wrote:<br>
<br>
> Andrew Beekhof <andrew@...> writes:<br>
><br>
>><br>
>> Try increasing your stickiness as it is being exceeded by the location<br>
> constraints.<br>
>> For the biggest stick, try 'infinity' which means - never move unless the<br>
> node dies.<br>
>><br>
><br>
> Thanks, Andrew, I applied infinity resource stickiness. However, the sst<br>
> resource is still switched to the node which is online back from failure.<br>
> And I found sth in the log:<br>
><br>
> Jun 13 11:46:29 node3 pengine[27813]: warning: unpack_rsc_op: Processing<br>
> failed op monitor for ip_ssn on node2: not running (7)<br>
> Jun 13 11:46:29 node3 pengine[27813]: error: native_create_actions:<br>
> Resource fs_ssn (ocf::Filesystem) is active on 2 nodes attempting recovery<br>
> Jun 13 11:46:29 node3 pengine[27813]: warning: native_create_actions: See<br>
> <a href="http://clusterlabs.org/wiki/FAQ#Resource_is_Too_Active" target="_blank">http://clusterlabs.org/wiki/FAQ#Resource_is_Too_Active</a> for more information.<br>
><br>
> Is this log showing that pacemaker tries to restart all the resource when<br>
> the failed node is back again?<br>
<br>
</div>No, thats a log showing the services were already running there when pacemaker started.<br>
<div class="HOEnZb"><div class="h5"><br>
><br>
><br>
>>> Thanks.<br>
>>><br>
>>> Below is my configure:<br>
>>> ------------------CONFIG START--------------------------------------<br>
>>> node node3 \<br>
>>> attributes standby="on"<br>
>>> node node1<br>
>>> node node2<br>
>>> primitive drbd_ssn ocf:linbit:drbd \<br>
>>> params drbd_resource="r0" \<br>
>>> op monitor interval="15s"<br>
>>> primitive fs_ssn ocf:heartbeat:Filesystem \<br>
>>> op monitor interval="15s" \<br>
>>> params device="/dev/drbd0" directory="/drbd" fstype="ext3" \<br>
>>> meta target-role="Started"<br>
>>> primitive ip_ssn ocf:heartbeat:IPaddr2 \<br>
>>> params ip="192.168.241.1" cidr_netmask="32" \<br>
>>> op monitor interval="15s" \<br>
>>> meta target-role="Started"<br>
>>> primitive ip_sst ocf:heartbeat:IPaddr2 \<br>
>>> params ip="192.168.241.2" cidr_netmask="32" \<br>
>>> op monitor interval="15s" \<br>
>>> meta target-role="Started"<br>
>>> primitive sst lsb:sst \<br>
>>> op monitor interval="15s" \<br>
>>> meta target-role="stopped"<br>
>>> primitive ssn lsb:ssn \<br>
>>> op monitor interval="15s" \<br>
>>> meta target-role="stopped"<br>
>>> ms ms_drbd_ssn drbd_ssn \<br>
>>> meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1"<br>
>>> notify="true" target-role="Started"<br>
>>> location sst_ip_prefer ip_sst 50: node1<br>
>>> location drbd_ssn_prefer ms_drbd_ssn 50: node1<br>
>>> colocation fs_ssn_coloc inf: ip_ssn fs_ssn<br>
>>> colocation fs_on_drbd_coloc inf: fs_ssn ms_drbd_ssn:Master<br>
>>> colocation sst_ip_coloc inf: sst ip_sst<br>
>>> colocation ssn_ip_coloc inf: ssn ip_ssn<br>
>>> order ssn_after_drbd inf: ms_drbd_ssn:promote fs_ssn:start<br>
>>> order ip_after_fs inf: fs_ssn:start ip_ssn:start<br>
>>> order sst_after_ip inf: ip_sst:start sst:start<br>
>>> order sst_after_ssn inf: ssn:start sst:start<br>
>>> order ssn_after_ip inf: ip_ssn:start ssn:start<br>
>>> property $id="cib-bootstrap-options" \<br>
>>> dc-version="1.1.8-7.el6-394e906" \<br>
>>> cluster-infrastructure="classic openais (with plugin)" \<br>
>>> expected-quorum-votes="3" \<br>
>>> stonith-enabled="false"<br>
>>> rsc_defaults $id="rsc-options" \<br>
>>> resource-stickiness="100"<br>
>>><br>
>>> -------------------CONFIG END----------------------------------------<br>
>>><br>
> Best Regards.<br>
> Xiaomin<br>
><br>
><br>
><br>
><br>
><br>
><br>
><br>
> _______________________________________________<br>
> Pacemaker mailing list: <a href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>
> <a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker" target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>
><br>
> Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>
> Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>
> Bugs: <a href="http://bugs.clusterlabs.org" target="_blank">http://bugs.clusterlabs.org</a><br>
<br>
<br>
_______________________________________________<br>
Pacemaker mailing list: <a href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>
<a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker" target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>
<br>
Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>
Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>
Bugs: <a href="http://bugs.clusterlabs.org" target="_blank">http://bugs.clusterlabs.org</a><br>
</div></div></blockquote></div><br></div></div>