[ClusterLabs] [ClusterLab] : Corosync not initializing successfully
Nikhil Utane
nikhil.subscribed at gmail.com
Tue May 17 17:13:27 UTC 2016
Hi Honza,
Just checking if you have the official patch available for this issue.
As far as I am concerned barring couple of issues, everything seems to be
working fine on our big-endian system. Much relieved. :)
-Thanks
Nikhil
On Thu, May 5, 2016 at 3:24 PM, Nikhil Utane <nikhil.subscribed at gmail.com>
wrote:
> It worked for me. :)
> I'll wait for your formal patch but until then I am able to proceed
> further. (Don't know if I'll run into something else)
>
> However now encountering issue in pacemaker.
>
> May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit:
> The cib process (15224) can no longer be respawned, shutting the cluster
> down.
> May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit:
> The stonith-ng process (15225) can no longer be respawned, shutting the
> cluster down.
> May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit:
> The lrmd process (15226) can no longer be respawned, shutting the cluster
> down.
> May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit:
> The crmd process (15229) can no longer be respawned, shutting the cluster
> down.
> May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit:
> The pengine process (15228) can no longer be respawned, shutting the
> cluster down.
> May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit:
> The attrd process (15227) can no longer be respawned, shutting the cluster
> down.
>
> Looking into it.
>
> -Thanks
> Nikhil
>
> On Thu, May 5, 2016 at 2:58 PM, Jan Friesse <jfriesse at redhat.com> wrote:
>
>> Nikhil
>>
>> Found the root-cause.
>>> In file schedwrk.c, the function handle2void() uses a union which was not
>>> initialized.
>>> Because of that the handle value was computed incorrectly (lower half was
>>> garbage).
>>>
>>> 56 static hdb_handle_t
>>> 57 void2handle (const void *v) { union u u={}; u.v = v; return u.h; }
>>> 58 static const void *
>>> 59 handle2void (hdb_handle_t h) { union u u={}; u.h = h; return u.v; }
>>>
>>> After initializing (as highlighted), the corosync initialization seems to
>>> be going through fine. Will check other things.
>>>
>>
>> Your patch is incorrect and actually doesn't work. As I said (when
>> pointing you to schedwrk.c), I will send you proper patch, but fix that
>> issue correctly is not easy.
>>
>> Regards,
>> Honza
>>
>>
>>> -Regards
>>> Nikhil
>>>
>>> On Tue, May 3, 2016 at 7:04 PM, Nikhil Utane <
>>> nikhil.subscribed at gmail.com>
>>> wrote:
>>>
>>> Thanks for your response Dejan.
>>>>
>>>> I do not know yet whether this has anything to do with endianness.
>>>> FWIW, there could be something quirky with the system so keeping all
>>>> options open. :)
>>>>
>>>> I added some debug prints to understand what's happening under the hood.
>>>>
>>>> *Success case: (on x86 machine): *
>>>> [TOTEM ] entering OPERATIONAL state.
>>>> [TOTEM ] A new membership (10.206.1.7:137220) was formed. Members
>>>> joined:
>>>> 181272839
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=0,
>>>> my_high_delivered=0
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>>> my_high_delivered=0
>>>> [TOTEM ] Delivering 0 to 1
>>>> [TOTEM ] Delivering MCAST message with seq 1 to pending delivery queue
>>>> [SYNC ] Nikhil: Inside sync_deliver_fn. header->id=1
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=2,
>>>> my_high_delivered=1
>>>> [TOTEM ] Delivering 1 to 2
>>>> [TOTEM ] Delivering MCAST message with seq 2 to pending delivery queue
>>>> [SYNC ] Nikhil: Inside sync_deliver_fn. header->id=0
>>>> [SYNC ] Nikhil: Entering sync_barrier_handler
>>>> [SYNC ] Committing synchronization for corosync configuration map
>>>> access
>>>> .
>>>> [TOTEM ] Delivering 2 to 4
>>>> [TOTEM ] Delivering MCAST message with seq 3 to pending delivery queue
>>>> [TOTEM ] Delivering MCAST message with seq 4 to pending delivery queue
>>>> [CPG ] comparing: sender r(0) ip(10.206.1.7) ; members(old:0 left:0)
>>>> [CPG ] chosen downlist: sender r(0) ip(10.206.1.7) ; members(old:0
>>>> left:0)
>>>> [SYNC ] Committing synchronization for corosync cluster closed process
>>>> group service v1.01
>>>> *[MAIN ] Completed service synchronization, ready to provide service.*
>>>>
>>>>
>>>> *Failure case: (on ppc)*:
>>>>
>>>> [TOTEM ] entering OPERATIONAL state.
>>>> [TOTEM ] A new membership (10.207.24.101:16) was formed. Members
>>>> joined:
>>>> 181344357
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=0,
>>>> my_high_delivered=0
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>>> my_high_delivered=0
>>>> [TOTEM ] Delivering 0 to 1
>>>> [TOTEM ] Delivering MCAST message with seq 1 to pending delivery queue
>>>> [SYNC ] Nikhil: Inside sync_deliver_fn header->id=1
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>>> my_high_delivered=1
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>>> my_high_delivered=1
>>>> Above message repeats continuously.
>>>>
>>>> So it appears that in failure case I do not receive messages with
>>>> sequence
>>>> number 2-4.
>>>> If somebody can throw some ideas that'll help a lot.
>>>>
>>>> -Thanks
>>>> Nikhil
>>>>
>>>> On Tue, May 3, 2016 at 5:26 PM, Dejan Muhamedagic <dejanmm at fastmail.fm>
>>>> wrote:
>>>>
>>>> Hi,
>>>>>
>>>>> On Mon, May 02, 2016 at 08:54:09AM +0200, Jan Friesse wrote:
>>>>>
>>>>>> As your hardware is probably capable of running ppcle and if you have
>>>>>>>
>>>>>> an
>>>>>
>>>>>> environment
>>>>>>> at hand without too much effort it might pay off to try that.
>>>>>>> There are of course distributions out there support corosync on
>>>>>>> big-endian architectures
>>>>>>> but I don't know if there is an automatized regression for corosync
>>>>>>> on
>>>>>>> big-endian that
>>>>>>> would catch big-endian-issues right away with something as current as
>>>>>>> your 2.3.5.
>>>>>>>
>>>>>>
>>>>>> No we are not testing big-endian.
>>>>>>
>>>>>> So totally agree with Klaus. Give a try to ppcle. Also make sure all
>>>>>> nodes are little-endian. Corosync should work in mixed BE/LE
>>>>>> environment but because it's not tested, it may not work (and it's a
>>>>>> bug, so if ppcle works I will try to fix BE).
>>>>>>
>>>>>
>>>>> I tested a cluster consisting of big endian/little endian nodes
>>>>> (s390 and x86-64), but that was a while ago. IIRC, all relevant
>>>>> bugs in corosync got fixed at that time. Don't know what is the
>>>>> situation with the latest version.
>>>>>
>>>>> Thanks,
>>>>>
>>>>> Dejan
>>>>>
>>>>> Regards,
>>>>>> Honza
>>>>>>
>>>>>>
>>>>>>> Regards,
>>>>>>> Klaus
>>>>>>>
>>>>>>> On 05/02/2016 06:44 AM, Nikhil Utane wrote:
>>>>>>>
>>>>>>>> Re-sending as I don't see my post on the thread.
>>>>>>>>
>>>>>>>> On Sun, May 1, 2016 at 4:21 PM, Nikhil Utane
>>>>>>>> <nikhil.subscribed at gmail.com <mailto:nikhil.subscribed at gmail.com>>
>>>>>>>>
>>>>>>> wrote:
>>>>>
>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> Looking for some guidance here as we are completely blocked
>>>>>>>> otherwise :(.
>>>>>>>>
>>>>>>>> -Regards
>>>>>>>> Nikhil
>>>>>>>>
>>>>>>>> On Fri, Apr 29, 2016 at 6:11 PM, Sriram <sriram.ec at gmail.com
>>>>>>>> <mailto:sriram.ec at gmail.com>> wrote:
>>>>>>>>
>>>>>>>> Corrected the subject.
>>>>>>>>
>>>>>>>> We went ahead and captured corosync debug logs for our ppc
>>>>>>>>
>>>>>>> board.
>>>>>
>>>>>> After log analysis and comparison with the sucessful logs(
>>>>>>>> from x86 machine) ,
>>>>>>>> we didnt find *"[ MAIN ] Completed service
>>>>>>>> synchronization,
>>>>>>>> ready to provide service.*" in ppc logs.
>>>>>>>> So, looks like corosync is not in a position to accept
>>>>>>>> connection from Pacemaker.
>>>>>>>> Even I tried with the new corosync.conf with no success.
>>>>>>>>
>>>>>>>> Any hints on this issue would be really helpful.
>>>>>>>>
>>>>>>>> Attaching ppc_notworking.log, x86_working.log,
>>>>>>>> corosync.conf.
>>>>>>>>
>>>>>>>> Regards,
>>>>>>>> Sriram
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Fri, Apr 29, 2016 at 2:44 PM, Sriram <
>>>>>>>> sriram.ec at gmail.com
>>>>>>>> <mailto:sriram.ec at gmail.com>> wrote:
>>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> I went ahead and made some changes in file system(Like
>>>>>>>> I
>>>>>>>> brought in /etc/init.d/corosync and
>>>>>>>>
>>>>>>> /etc/init.d/pacemaker,
>>>>>
>>>>>> /etc/sysconfig ), After that I was able to run "pcs
>>>>>>>> cluster start".
>>>>>>>> But it failed with the following error
>>>>>>>> # pcs cluster start
>>>>>>>> Starting Cluster...
>>>>>>>> Starting Pacemaker Cluster Manager[FAILED]
>>>>>>>> Error: unable to start pacemaker
>>>>>>>>
>>>>>>>> And in the /var/log/pacemaker.log, I saw these errors
>>>>>>>> pacemakerd: info: mcp_read_config: cmap connection
>>>>>>>> setup failed: CS_ERR_TRY_AGAIN. Retrying in 4s
>>>>>>>> Apr 29 08:53:47 [15863] node_cu pacemakerd: info:
>>>>>>>> mcp_read_config: cmap connection setup failed:
>>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 5s
>>>>>>>> Apr 29 08:53:52 [15863] node_cu pacemakerd: warning:
>>>>>>>> mcp_read_config: Could not connect to Cluster
>>>>>>>> Configuration Database API, error 6
>>>>>>>> Apr 29 08:53:52 [15863] node_cu pacemakerd: notice:
>>>>>>>> main: Could not obtain corosync config data,
>>>>>>>> exiting
>>>>>>>> Apr 29 08:53:52 [15863] node_cu pacemakerd: info:
>>>>>>>> crm_xml_cleanup: Cleaning up memory from libxml2
>>>>>>>>
>>>>>>>>
>>>>>>>> And in the /var/log/Debuglog, I saw these errors coming
>>>>>>>> from corosync
>>>>>>>> 20160429 085347.487050 <tel:085347.487050> airv_cu
>>>>>>>> daemon.warn corosync[12857]: [QB ] Denied
>>>>>>>>
>>>>>>> connection,
>>>>>
>>>>>> is not ready (12857-15863-14)
>>>>>>>> 20160429 085347.487067 <tel:085347.487067> airv_cu
>>>>>>>> daemon.info <http://daemon.info> corosync[12857]:
>>>>>>>> [QB
>>>>>>>> ] Denied connection, is not ready (12857-15863-14)
>>>>>>>>
>>>>>>>>
>>>>>>>> I browsed the code of libqb to find that it is failing
>>>>>>>> in
>>>>>>>>
>>>>>>>>
>>>>>>>> https://github.com/ClusterLabs/libqb/blob/master/lib/ipc_setup.c
>>>>>
>>>>>>
>>>>>>>> Line 600 :
>>>>>>>> handle_new_connection function
>>>>>>>>
>>>>>>>> Line 637:
>>>>>>>> if (auth_result == 0 &&
>>>>>>>> c->service->serv_fns.connection_accept) {
>>>>>>>> res = c->service->serv_fns.connection_accept(c,
>>>>>>>> c->euid, c->egid);
>>>>>>>> }
>>>>>>>> if (res != 0) {
>>>>>>>> goto send_response;
>>>>>>>> }
>>>>>>>>
>>>>>>>> Any hints on this issue would be really helpful for me
>>>>>>>> to
>>>>>>>> go ahead.
>>>>>>>> Please let me know if any logs are required,
>>>>>>>>
>>>>>>>> Regards,
>>>>>>>> Sriram
>>>>>>>>
>>>>>>>> On Thu, Apr 28, 2016 at 2:42 PM, Sriram
>>>>>>>> <sriram.ec at gmail.com <mailto:sriram.ec at gmail.com>>
>>>>>>>>
>>>>>>> wrote:
>>>>>
>>>>>>
>>>>>>>> Thanks Ken and Emmanuel.
>>>>>>>> Its a big endian machine. I will try with running
>>>>>>>>
>>>>>>> "pcs
>>>>>
>>>>>> cluster setup" and "pcs cluster start"
>>>>>>>> Inside cluster.py, "service pacemaker start" and
>>>>>>>> "service corosync start" are executed to bring up
>>>>>>>> pacemaker and corosync.
>>>>>>>> Those service scripts and the infrastructure needed
>>>>>>>>
>>>>>>> to
>>>>>
>>>>>> bring up the processes in the above said manner
>>>>>>>> doesn't exist in my board.
>>>>>>>> As it is a embedded board with the limited memory,
>>>>>>>> full fledged linux is not installed.
>>>>>>>> Just curious to know, what could be reason the
>>>>>>>> pacemaker throws that error.
>>>>>>>>
>>>>>>>> /"cmap connection setup failed: CS_ERR_TRY_AGAIN.
>>>>>>>> Retrying in 1s"
>>>>>>>>
>>>>>>>> /
>>>>>>>> Thanks for response.
>>>>>>>>
>>>>>>>> Regards,
>>>>>>>> Sriram.
>>>>>>>>
>>>>>>>> On Thu, Apr 28, 2016 at 8:55 AM, Ken Gaillot
>>>>>>>> <kgaillot at redhat.com <mailto:kgaillot at redhat.com>>
>>>>>>>>
>>>>>>> wrote:
>>>>>
>>>>>>
>>>>>>>> On 04/27/2016 11:25 AM, emmanuel segura wrote:
>>>>>>>> > you need to use pcs to do everything, pcs
>>>>>>>> cluster setup and pcs
>>>>>>>> > cluster start, try to use the redhat docs for
>>>>>>>> more information.
>>>>>>>>
>>>>>>>> Agreed -- pcs cluster setup will create a
>>>>>>>> proper
>>>>>>>> corosync.conf for you.
>>>>>>>> Your corosync.conf below uses corosync 1
>>>>>>>> syntax,
>>>>>>>> and there were
>>>>>>>> significant changes in corosync 2. In
>>>>>>>> particular,
>>>>>>>> you don't need the
>>>>>>>> file created in step 4, because pacemaker is no
>>>>>>>> longer launched via a
>>>>>>>> corosync plugin.
>>>>>>>>
>>>>>>>> > 2016-04-27 17:28 GMT+02:00 Sriram
>>>>>>>> <sriram.ec at gmail.com <mailto:
>>>>>>>> sriram.ec at gmail.com
>>>>>>>>
>>>>>>> :
>>>>>>>
>>>>>>>> >> Dear All,
>>>>>>>> >>
>>>>>>>> >> I m trying to use pacemaker and corosync for
>>>>>>>> the clustering requirement that
>>>>>>>> >> came up recently.
>>>>>>>> >> We have cross compiled corosync, pacemaker
>>>>>>>> and
>>>>>>>> pcs(python) for ppc
>>>>>>>> >> environment (Target board where pacemaker
>>>>>>>> and
>>>>>>>> corosync are supposed to run)
>>>>>>>> >> I m having trouble bringing up pacemaker in
>>>>>>>> that environment, though I could
>>>>>>>> >> successfully bring up corosync.
>>>>>>>> >> Any help is welcome.
>>>>>>>> >>
>>>>>>>> >> I m using these versions of pacemaker and
>>>>>>>>
>>>>>>> corosync
>>>>>
>>>>>> >> [root at node_cu pacemaker]# corosync -v
>>>>>>>> >> Corosync Cluster Engine, version '2.3.5'
>>>>>>>> >> Copyright (c) 2006-2009 Red Hat, Inc.
>>>>>>>> >> [root at node_cu pacemaker]# pacemakerd -$
>>>>>>>> >> Pacemaker 1.1.14
>>>>>>>> >> Written by Andrew Beekhof
>>>>>>>> >>
>>>>>>>> >> For running corosync, I did the following.
>>>>>>>> >> 1. Created the following directories,
>>>>>>>> >> /var/lib/pacemaker
>>>>>>>> >> /var/lib/corosync
>>>>>>>> >> /var/lib/pacemaker
>>>>>>>> >> /var/lib/pacemaker/cores
>>>>>>>> >> /var/lib/pacemaker/pengine
>>>>>>>> >> /var/lib/pacemaker/blackbox
>>>>>>>> >> /var/lib/pacemaker/cib
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>> >> 2. Created a file called corosync.conf under
>>>>>>>> /etc/corosync folder with the
>>>>>>>> >> following contents
>>>>>>>> >>
>>>>>>>> >> totem {
>>>>>>>> >>
>>>>>>>> >> version: 2
>>>>>>>> >> token: 5000
>>>>>>>> >> token_retransmits_before_loss_const:
>>>>>>>>
>>>>>>> 20
>>>>>
>>>>>> >> join: 1000
>>>>>>>> >> consensus: 7500
>>>>>>>> >> vsftype: none
>>>>>>>> >> max_messages: 20
>>>>>>>> >> secauth: off
>>>>>>>> >> cluster_name: mycluster
>>>>>>>> >> transport: udpu
>>>>>>>> >> threads: 0
>>>>>>>> >> clear_node_high_bit: yes
>>>>>>>> >>
>>>>>>>> >> interface {
>>>>>>>> >> ringnumber: 0
>>>>>>>> >> # The following three values
>>>>>>>> need to be set based on your
>>>>>>>> >> environment
>>>>>>>> >> bindnetaddr: 10.x.x.x
>>>>>>>> >> mcastaddr: 226.94.1.1
>>>>>>>> >> mcastport: 5405
>>>>>>>> >> }
>>>>>>>> >> }
>>>>>>>> >>
>>>>>>>> >> logging {
>>>>>>>> >> fileline: off
>>>>>>>> >> to_syslog: yes
>>>>>>>> >> to_stderr: no
>>>>>>>> >> to_syslog: yes
>>>>>>>> >> logfile: /var/log/corosync.log
>>>>>>>> >> syslog_facility: daemon
>>>>>>>> >> debug: on
>>>>>>>> >> timestamp: on
>>>>>>>> >> }
>>>>>>>> >>
>>>>>>>> >> amf {
>>>>>>>> >> mode: disabled
>>>>>>>> >> }
>>>>>>>> >>
>>>>>>>> >> quorum {
>>>>>>>> >> provider: corosync_votequorum
>>>>>>>> >> }
>>>>>>>> >>
>>>>>>>> >> nodelist {
>>>>>>>> >> node {
>>>>>>>> >> ring0_addr: node_cu
>>>>>>>> >> nodeid: 1
>>>>>>>> >> }
>>>>>>>> >> }
>>>>>>>> >>
>>>>>>>> >> 3. Created authkey under /etc/corosync
>>>>>>>> >>
>>>>>>>> >> 4. Created a file called pcmk under
>>>>>>>> /etc/corosync/service.d and contents as
>>>>>>>> >> below,
>>>>>>>> >> cat pcmk
>>>>>>>> >> service {
>>>>>>>> >> # Load the Pacemaker Cluster
>>>>>>>> Resource
>>>>>>>> Manager
>>>>>>>> >> name: pacemaker
>>>>>>>> >> ver: 1
>>>>>>>> >> }
>>>>>>>> >>
>>>>>>>> >> 5. Added the node name "node_cu" in
>>>>>>>> /etc/hosts
>>>>>>>> with 10.X.X.X ip
>>>>>>>> >>
>>>>>>>> >> 6. ./corosync -f -p & --> this step started
>>>>>>>> corosync
>>>>>>>> >>
>>>>>>>> >> [root at node_cu pacemaker]# netstat -alpn |
>>>>>>>>
>>>>>>> grep
>>>>>
>>>>>> -i coros
>>>>>>>> >> udp 0 0 10.X.X.X:61841
>>>>>>>>
>>>>>>> 0.0.0.0:*
>>>>>
>>>>>> >> 9133/corosync
>>>>>>>> >> udp 0 0 10.X.X.X:5405
>>>>>>>>
>>>>>>> 0.0.0.0:*
>>>>>
>>>>>> >> 9133/corosync
>>>>>>>> >> unix 2 [ ACC ] STREAM
>>>>>>>> LISTENING
>>>>>>>> 148888 9133/corosync
>>>>>>>> >> @quorum
>>>>>>>> >> unix 2 [ ACC ] STREAM
>>>>>>>> LISTENING
>>>>>>>> 148884 9133/corosync
>>>>>>>> >> @cmap
>>>>>>>> >> unix 2 [ ACC ] STREAM
>>>>>>>> LISTENING
>>>>>>>> 148887 9133/corosync
>>>>>>>> >> @votequorum
>>>>>>>> >> unix 2 [ ACC ] STREAM
>>>>>>>> LISTENING
>>>>>>>> 148885 9133/corosync
>>>>>>>> >> @cfg
>>>>>>>> >> unix 2 [ ACC ] STREAM
>>>>>>>> LISTENING
>>>>>>>> 148886 9133/corosync
>>>>>>>> >> @cpg
>>>>>>>> >> unix 2 [ ] DGRAM
>>>>>>>> 148840 9133/corosync
>>>>>>>> >>
>>>>>>>> >> 7. ./pacemakerd -f & gives the following
>>>>>>>> error
>>>>>>>> and exits.
>>>>>>>> >> [root at node_cu pacemaker]# pacemakerd -f
>>>>>>>> >> cmap connection setup failed:
>>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 1s
>>>>>>>> >> cmap connection setup failed:
>>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 2s
>>>>>>>> >> cmap connection setup failed:
>>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 3s
>>>>>>>> >> cmap connection setup failed:
>>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 4s
>>>>>>>> >> cmap connection setup failed:
>>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 5s
>>>>>>>> >> Could not connect to Cluster Configuration
>>>>>>>> Database API, error 6
>>>>>>>> >>
>>>>>>>> >> Can you please point me, what is missing in
>>>>>>>> these steps ?
>>>>>>>> >>
>>>>>>>> >> Before trying these steps, I tried running
>>>>>>>>
>>>>>>> "pcs
>>>>>
>>>>>> cluster start", but that
>>>>>>>> >> command fails with "service" script not
>>>>>>>> found.
>>>>>>>> As the root filesystem
>>>>>>>> >> doesn't contain either /etc/init.d/ or
>>>>>>>> /sbin/service
>>>>>>>> >>
>>>>>>>> >> So, the plan is to bring up corosync and
>>>>>>>> pacemaker manually, later do the
>>>>>>>> >> cluster configuration using "pcs" commands.
>>>>>>>> >>
>>>>>>>> >> Regards,
>>>>>>>> >> Sriram
>>>>>>>> >>
>>>>>>>> >>
>>>>>>>>
>>>>>>> _______________________________________________
>>>>>
>>>>>> >> Users mailing list: Users at clusterlabs.org
>>>>>>>> <mailto:Users at clusterlabs.org>
>>>>>>>> >>
>>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>> >>
>>>>>>>> >> Project Home: http://www.clusterlabs.org
>>>>>>>> >> Getting started:
>>>>>>>>
>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>> >> Bugs: http://bugs.clusterlabs.org
>>>>>>>> >>
>>>>>>>> >
>>>>>>>> >
>>>>>>>> >
>>>>>>>>
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>>>> <mailto:Users at clusterlabs.org>
>>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>>
>>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>>> Getting started:
>>>>>>>>
>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>>>> <mailto:Users at clusterlabs.org>
>>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>>
>>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>>> Getting started:
>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>>
>>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>>> Getting started:
>>>>>>>>
>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>
>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>> Getting started:
>>>>>>>
>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>
>>>>>> Project Home: http://www.clusterlabs.org
>>>>>> Getting started:
>>>>>>
>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list: Users at clusterlabs.org
>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>
>>>>> Project Home: http://www.clusterlabs.org
>>>>> Getting started:
>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>>> _______________________________________________
>>> Users mailing list: Users at clusterlabs.org
>>> http://clusterlabs.org/mailman/listinfo/users
>>>
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>>>
>>>
>>
>> _______________________________________________
>> Users mailing list: Users at clusterlabs.org
>> http://clusterlabs.org/mailman/listinfo/users
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20160517/b3cbb81e/attachment.htm>
More information about the Users
mailing list