<html><head><style type="text/css"><!-- DIV {margin:0px;} --></style></head><body><div style="font-family:'times new roman', 'new york', times, serif;font-size:12pt"><div>Andrew,</div><div><br></div><div>Thanks for responding. Comments inline with <Bob></div><div style="font-family:times new roman, new york, times, serif;font-size:12pt"><br><div style="font-family:arial, helvetica, sans-serif;font-size:13px"><font size="2" face="Tahoma"><hr size="1"><b><span style="font-weight: bold;">From:</span></b> Andrew Beekhof <andrew@beekhof.net><br><b><span style="font-weight: bold;">To:</span></b> The Pacemaker cluster resource manager <pacemaker@oss.clusterlabs.org><br><b><span style="font-weight: bold;">Cc:</span></b> Bob Schatz <bschatz@yahoo.com><br><b><span style="font-weight: bold;">Sent:</span></b> Tue, April 12, 2011 11:23:14 PM<br><b><span style="font-weight: bold;">Subject:</span></b> Re: [Pacemaker] Question regarding
starting of master/slave resources and ELECTIONs<br></font><br>On Wed, Apr 13, 2011 at 4:54 AM, Bob Schatz <<a ymailto="mailto:bschatz@yahoo.com" href="mailto:bschatz@yahoo.com">bschatz@yahoo.com</a>> wrote:<br>> Hi,<br>> I am running Pacemaker 1.0.9 with Heartbeat 3.0.3.<br>> I create 5 master/slave resources in /etc/ha.d/resource.d/startstop during<br>> post-start.<br><br>I had no idea this was possible. Why would you do this?</div><div style="font-family:arial, helvetica, sans-serif;font-size:13px"><br></div><div style="font-family:arial, helvetica, sans-serif;font-size:13px"><Bob> We and I know of a couple of other companies, bundle LinuxHA/Pacemaker into an appliance. For me, when the appliance boots, it creates HA resources based on the hardware it discovers. I assumed that once POST-START was called in the startstop script and we have a DC then the cluster is up and running. I then use "crm"
commands to create the configuration, etc. I further assumed that since we have one DC in the cluster then all "crm" commands which modify the configuration would be ordered even if the DC fails over to a different node. Is this incorrect?<br><br>> I noticed that 4 of the master/slave resources will start right away but the<br>> 5 master/slave resource seems to take a minute or so and I am only running<br>> with one node.<br>> Is this expected?<br><br>Probably, if the other 4 take around a minute each to start.<br>There is an lrmd config variable that controls how much parallelism it<br>allows (but i forget the name).</div><div style="font-family:arial, helvetica, sans-serif;font-size:13px"><br></div><div style="font-family:arial, helvetica, sans-serif;font-size:13px"><Bob> It's max-children and I set it to 40 for this test to see if it would change the behavior. (/sbin/lrmadmin -p max-children 40)<br><br>> My
configuration is below and I have also attached ha-debug.<br>> Also, what triggers a crmd election?<br><br>Node up/down events and whenever someone replaces the cib (which the<br>shell used to do a lot).</div><div style="font-family:arial, helvetica, sans-serif;font-size:13px"><br></div><div style="font-family:arial, helvetica, sans-serif;font-size:13px"><Bob> For my test, I only started one node so that I could avoid node up/down events. I believe the log shows the cib being replaced. Since I am using crm then I assume it must be due to crm. Do the crm_resource, etc commands also replace the cib? Would that avoid elections as a result of cibs being replaced?</div><div style="font-family:arial, helvetica, sans-serif;font-size:13px"><br></div><div style="font-family:arial, helvetica, sans-serif;font-size:13px"><br></div><div style="font-family:arial, helvetica, sans-serif;font-size:13px">Thanks,</div><div
style="font-family:arial, helvetica, sans-serif;font-size:13px"><br></div><div style="font-family:arial, helvetica, sans-serif;font-size:13px">Bob<br><br>> I seemed to have a lot of elections in<br>> the attached log. I was assuming that on a single node I would only run the<br>> election once in the beginning and then there would not be another one until<br>> a new node joined.<br>><br>> Thanks,<br>> Bob<br>><br>> My configuration is:<br>> node $id="856c1f72-7cd1-4906-8183-8be87eef96f2" mgraid-s000030311-1<br>> primitive SSJ000030312 ocf:omneon:ss \<br>> params ss_resource="SSJ000030312"<br>> ssconf="/var/omneon/config/config.J000030312" \<br>> op monitor interval="3s" role="Master" timeout="7s" \<br>> op monitor interval="10s" role="Slave" timeout="7" \<br>> op
stop interval="0" timeout="20" \<br>> op start interval="0" timeout="300"<br>> primitive SSJ000030313 ocf:omneon:ss \<br>> params ss_resource="SSJ000030313"<br>> ssconf="/var/omneon/config/config.J000030313" \<br>> op monitor interval="3s" role="Master" timeout="7s" \<br>> op monitor interval="10s" role="Slave" timeout="7" \<br>> op stop interval="0" timeout="20" \<br>> op start interval="0" timeout="300"<br>> primitive SSJ000030314 ocf:omneon:ss \<br>> params ss_resource="SSJ000030314"<br>> ssconf="/var/omneon/config/config.J000030314" \<br>> op monitor interval="3s" role="Master" timeout="7s" \<br>> op monitor
interval="10s" role="Slave" timeout="7" \<br>> op stop interval="0" timeout="20" \<br>> op start interval="0" timeout="300"<br>> primitive SSJ000030315 ocf:omneon:ss \<br>> params ss_resource="SSJ000030315"<br>> ssconf="/var/omneon/config/config.J000030315" \<br>> op monitor interval="3s" role="Master" timeout="7s" \<br>> op monitor interval="10s" role="Slave" timeout="7" \<br>> op stop interval="0" timeout="20" \<br>> op start interval="0" timeout="300"<br>> primitive SSS000030311 ocf:omneon:ss \<br>> params ss_resource="SSS000030311"<br>> ssconf="/var/omneon/config/config.S000030311" \<br>> op monitor interval="3s"
role="Master" timeout="7s" \<br>> op monitor interval="10s" role="Slave" timeout="7" \<br>> op stop interval="0" timeout="20" \<br>> op start interval="0" timeout="300"<br>> primitive icms lsb:S53icms \<br>> op monitor interval="5s" timeout="7" \<br>> op start interval="0" timeout="5"<br>> primitive mgraid-stonith stonith:external/mgpstonith \<br>> params hostlist="mgraid-canister" \<br>> op monitor interval="0" timeout="20s"<br>> primitive omserver lsb:S49omserver \<br>> op monitor interval="5s" timeout="7" \<br>> op start interval="0" timeout="5"<br>> ms ms-SSJ000030312 SSJ000030312 \<br>>
meta clone-max="2" notify="true" globally-unique="false"<br>> target-role="Started"<br>> ms ms-SSJ000030313 SSJ000030313 \<br>> meta clone-max="2" notify="true" globally-unique="false"<br>> target-role="Started"<br>> ms ms-SSJ000030314 SSJ000030314 \<br>> meta clone-max="2" notify="true" globally-unique="false"<br>> target-role="Started"<br>> ms ms-SSJ000030315 SSJ000030315 \<br>> meta clone-max="2" notify="true" globally-unique="false"<br>> target-role="Started"<br>> ms ms-SSS000030311 SSS000030311 \<br>> meta clone-max="2" notify="true" globally-unique="false"<br>> target-role="Started"<br>> clone Fencing mgraid-stonith<br>> clone cloneIcms icms<br>> clone cloneOmserver omserver<br>> location ms-SSJ000030312-master-w1 ms-SSJ000030312 \<br>>
rule $id="ms-SSJ000030312-master-w1-rule" $role="master" 100: #uname<br>> eq mgraid-s000030311-0<br>> location ms-SSJ000030313-master-w1 ms-SSJ000030313 \<br>> rule $id="ms-SSJ000030313-master-w1-rule" $role="master" 100: #uname<br>> eq mgraid-s000030311-0<br>> location ms-SSJ000030314-master-w1 ms-SSJ000030314 \<br>> rule $id="ms-SSJ000030314-master-w1-rule" $role="master" 100: #uname<br>> eq mgraid-s000030311-0<br>> location ms-SSJ000030315-master-w1 ms-SSJ000030315 \<br>> rule $id="ms-SSJ000030315-master-w1-rule" $role="master" 100: #uname<br>> eq mgraid-s000030311-0<br>> location ms-SSS000030311-master-w1 ms-SSS000030311 \<br>> rule $id="ms-SSS000030311-master-w1-rule" $role="master" 100: #uname<br>> eq mgraid-s000030311-0<br>> order orderms-SSJ000030312 0:
cloneIcms ms-SSJ000030312<br>> order orderms-SSJ000030313 0: cloneIcms ms-SSJ000030313<br>> order orderms-SSJ000030314 0: cloneIcms ms-SSJ000030314<br>> order orderms-SSJ000030315 0: cloneIcms ms-SSJ000030315<br>> order orderms-SSS000030311 0: cloneIcms ms-SSS000030311<br>> property $id="cib-bootstrap-options" \<br>> dc-version="1.0.9-89bd754939df5150de7cd76835f98fe90851b677" \<br>> cluster-infrastructure="Heartbeat" \<br>> dc-deadtime="5s" \<br>> stonith-enabled="true"<br>><br>><br>> _______________________________________________<br>> Pacemaker mailing list: <a ymailto="mailto:Pacemaker@oss.clusterlabs.org" href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>> <a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker"
target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>><br>> Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>> Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>> Bugs:<br>> <a href="http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker" target="_blank">http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker</a><br>><br>><br></div></div><div style="position:fixed"></div>
</div><br>
</body></html>