<div dir="ltr">Hi Andrew,<div><br></div><div>Thanks for the responses! I set the cman quorum timeout to 0, and the first node came up.</div><div><br></div><div>However, I'm seeing newer issues now that I hadn't seen before, when creating a resource of an LSB compliant service. I will send another mail to the ML to keep that discussion separate.</div>
<div><br></div><div>Regards,</div><div>Vijay</div><div class="gmail_extra"><br><br><div class="gmail_quote">On Thu, Jun 26, 2014 at 5:45 PM, Andrew Beekhof <span dir="ltr"><<a href="mailto:andrew@beekhof.net" target="_blank">andrew@beekhof.net</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div class=""><br>
On 27 Jun 2014, at 10:22 am, Vijay B <<a href="mailto:os.vbvs@gmail.com">os.vbvs@gmail.com</a>> wrote:<br>
<br>
> Hi,<br>
><br>
> I'm trying to set up a three node cluster using pacemaker+corosync, and I installed the required packages on each node, checked for their network connectivity so they can see each other, added the required startup scripts and edited the cluster.conf file as well so it includes all three nodes.<br>
><br>
> Now, when on the first node, I attempt to start up cman using service cman start, it times out thus:<br>
><br>
> vagrant@precise64-pmk1:~$ sudo service cman start<br>
> Starting cluster:<br>
> Checking if cluster has been disabled at boot... [ OK ]<br>
> Checking Network Manager... [ OK ]<br>
> Global setup... [ OK ]<br>
> Loading kernel modules... [ OK ]<br>
> Mounting configfs... [ OK ]<br>
> Starting cman... [ OK ]<br>
> Waiting for quorum... Timed-out waiting for cluster<br>
> [FAILED]<br>
> vagrant@precise64-pmk1:~$<br>
><br>
> Why is this? Is it because I have three nodes to begin with in my /etc/cluster/cluster.conf, and so this node expects that the cluster quorum is 2, and so it should be able to talk to at least one other node?<br>
<br>
</div>Yes. CMAN refuses to start unless it can see enough other nodes to form quorum.<br>
Pacemaker doesn't need quorum to behave sanely though, so try:<br>
<br>
echo "CMAN_QUORUM_TIMEOUT=0" >> /etc/sysconfig/cman<br>
<div class=""><br>
> At this point, I haven't started the cman or pacemaker services on the other nodes.<br>
><br>
> If this is the case, what will happen when two nodes of the three die?<br>
<br>
</div>Unrelated. This only affects startup.<br>
What happens in this case depends on how you configured no-quorum-policy.<br>
<div class=""><br>
> In case cluster.conf changes accordingly to reflect the new cluster membership, what if all three nodes are simply powered off and one rebooted? The cluster will be down, won't it?<br>
><br>
> What is the best way to get around this? I don't want to set CMAN_QUORUM_TIMEOUT=0, since as I understand it, the node would then go ahead and start itself as a cluster without waiting for the other nodes, and if this causes my service to start up<br>
<br>
</div>Pacemaker won't start things without quorum - unless you specifically tell it to.<br>
<div class=""><br>
> and it is already started/running on another node, it could cause issues.<br>
><br>
> Now, I don't know how to configure quorum disks for pacemaker - is it possible to do this with pacemaker?<br>
<br>
</div>Pacemaker doesn't care. Thats a cman/quorum detail. Pacemaker only cares "do we have quorum".<br>
<div class=""><br>
> How does it work? What are the recommended ways to address the above problem? I infer that if this disk is configured, the node that grabs the disk first becomes the president of the pacemaker cluster. In this context, I have another question - does corosync have its own cluster membership state distributed across all cluster nodes? If so, I guess quorum is configured at the corosync level rather than at the pacemaker level?<br>
><br>
> Apologies in advance if my queries above are addressed in the documentation already - I felt it would be quicker and more accurate to ask the community for reliable info.<br>
><br>
> Thanks!<br>
> Regards,<br>
> Vijay<br>
><br>
><br>
</div>> _______________________________________________<br>
> Pacemaker mailing list: <a href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>
> <a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker" target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>
><br>
> Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>
> Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>
> Bugs: <a href="http://bugs.clusterlabs.org" target="_blank">http://bugs.clusterlabs.org</a><br>
<br>
<br>_______________________________________________<br>
Pacemaker mailing list: <a href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>
<a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker" target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>
<br>
Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>
Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>
Bugs: <a href="http://bugs.clusterlabs.org" target="_blank">http://bugs.clusterlabs.org</a><br>
<br></blockquote></div><br></div></div>