[ClusterLabs] [ClusterLab] : Corosync not initializing successfully

Nikhil Utane nikhil.subscribed at gmail.com
Tue May 17 13:13:27 EDT 2016


Hi Honza,

Just checking if you have the official patch available for this issue.

As far as I am concerned barring couple of issues, everything seems to be
working fine on our big-endian system. Much relieved. :)

-Thanks
Nikhil


On Thu, May 5, 2016 at 3:24 PM, Nikhil Utane <nikhil.subscribed at gmail.com>
wrote:

> It worked for me. :)
> I'll wait for your formal patch but until then I am able to proceed
> further. (Don't know if I'll run into something else)
>
> However now encountering issue in pacemaker.
>
> May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:
>  The cib process (15224) can no longer be respawned, shutting the cluster
> down.
> May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:
>  The stonith-ng process (15225) can no longer be respawned, shutting the
> cluster down.
> May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:
>  The lrmd process (15226) can no longer be respawned, shutting the cluster
> down.
> May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:
>  The crmd process (15229) can no longer be respawned, shutting the cluster
> down.
> May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:
>  The pengine process (15228) can no longer be respawned, shutting the
> cluster down.
> May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:
>  The attrd process (15227) can no longer be respawned, shutting the cluster
> down.
>
> Looking into it.
>
> -Thanks
> Nikhil
>
> On Thu, May 5, 2016 at 2:58 PM, Jan Friesse <jfriesse at redhat.com> wrote:
>
>> Nikhil
>>
>> Found the root-cause.
>>> In file schedwrk.c, the function handle2void() uses a union which was not
>>> initialized.
>>> Because of that the handle value was computed incorrectly (lower half was
>>> garbage).
>>>
>>>   56 static hdb_handle_t
>>>   57 void2handle (const void *v) { union u u={}; u.v = v; return u.h; }
>>>   58 static const void *
>>>   59 handle2void (hdb_handle_t h) { union u u={}; u.h = h; return u.v; }
>>>
>>> After initializing (as highlighted), the corosync initialization seems to
>>> be going through fine. Will check other things.
>>>
>>
>> Your patch is incorrect and actually doesn't work. As I said (when
>> pointing you to schedwrk.c), I will send you proper patch, but fix that
>> issue correctly is not easy.
>>
>> Regards,
>>   Honza
>>
>>
>>> -Regards
>>> Nikhil
>>>
>>> On Tue, May 3, 2016 at 7:04 PM, Nikhil Utane <
>>> nikhil.subscribed at gmail.com>
>>> wrote:
>>>
>>> Thanks for your response Dejan.
>>>>
>>>> I do not know yet whether this has anything to do with endianness.
>>>> FWIW, there could be something quirky with the system so keeping all
>>>> options open. :)
>>>>
>>>> I added some debug prints to understand what's happening under the hood.
>>>>
>>>> *Success case: (on x86 machine): *
>>>> [TOTEM ] entering OPERATIONAL state.
>>>> [TOTEM ] A new membership (10.206.1.7:137220) was formed. Members
>>>> joined:
>>>> 181272839
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=0,
>>>> my_high_delivered=0
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>>> my_high_delivered=0
>>>> [TOTEM ] Delivering 0 to 1
>>>> [TOTEM ] Delivering MCAST message with seq 1 to pending delivery queue
>>>> [SYNC  ] Nikhil: Inside sync_deliver_fn. header->id=1
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=2,
>>>> my_high_delivered=1
>>>> [TOTEM ] Delivering 1 to 2
>>>> [TOTEM ] Delivering MCAST message with seq 2 to pending delivery queue
>>>> [SYNC  ] Nikhil: Inside sync_deliver_fn. header->id=0
>>>> [SYNC  ] Nikhil: Entering sync_barrier_handler
>>>> [SYNC  ] Committing synchronization for corosync configuration map
>>>> access
>>>> .
>>>> [TOTEM ] Delivering 2 to 4
>>>> [TOTEM ] Delivering MCAST message with seq 3 to pending delivery queue
>>>> [TOTEM ] Delivering MCAST message with seq 4 to pending delivery queue
>>>> [CPG   ] comparing: sender r(0) ip(10.206.1.7) ; members(old:0 left:0)
>>>> [CPG   ] chosen downlist: sender r(0) ip(10.206.1.7) ; members(old:0
>>>> left:0)
>>>> [SYNC  ] Committing synchronization for corosync cluster closed process
>>>> group service v1.01
>>>> *[MAIN  ] Completed service synchronization, ready to provide service.*
>>>>
>>>>
>>>> *Failure case: (on ppc)*:
>>>>
>>>> [TOTEM ] entering OPERATIONAL state.
>>>> [TOTEM ] A new membership (10.207.24.101:16) was formed. Members
>>>> joined:
>>>> 181344357
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=0,
>>>> my_high_delivered=0
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>>> my_high_delivered=0
>>>> [TOTEM ] Delivering 0 to 1
>>>> [TOTEM ] Delivering MCAST message with seq 1 to pending delivery queue
>>>> [SYNC  ] Nikhil: Inside sync_deliver_fn header->id=1
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>>> my_high_delivered=1
>>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>>> my_high_delivered=1
>>>> Above message repeats continuously.
>>>>
>>>> So it appears that in failure case I do not receive messages with
>>>> sequence
>>>> number 2-4.
>>>> If somebody can throw some ideas that'll help a lot.
>>>>
>>>> -Thanks
>>>> Nikhil
>>>>
>>>> On Tue, May 3, 2016 at 5:26 PM, Dejan Muhamedagic <dejanmm at fastmail.fm>
>>>> wrote:
>>>>
>>>> Hi,
>>>>>
>>>>> On Mon, May 02, 2016 at 08:54:09AM +0200, Jan Friesse wrote:
>>>>>
>>>>>> As your hardware is probably capable of running ppcle and if you have
>>>>>>>
>>>>>> an
>>>>>
>>>>>> environment
>>>>>>> at hand without too much effort it might pay off to try that.
>>>>>>> There are of course distributions out there support corosync on
>>>>>>> big-endian architectures
>>>>>>> but I don't know if there is an automatized regression for corosync
>>>>>>> on
>>>>>>> big-endian that
>>>>>>> would catch big-endian-issues right away with something as current as
>>>>>>> your 2.3.5.
>>>>>>>
>>>>>>
>>>>>> No we are not testing big-endian.
>>>>>>
>>>>>> So totally agree with Klaus. Give a try to ppcle. Also make sure all
>>>>>> nodes are little-endian. Corosync should work in mixed BE/LE
>>>>>> environment but because it's not tested, it may not work (and it's a
>>>>>> bug, so if ppcle works I will try to fix BE).
>>>>>>
>>>>>
>>>>> I tested a cluster consisting of big endian/little endian nodes
>>>>> (s390 and x86-64), but that was a while ago. IIRC, all relevant
>>>>> bugs in corosync got fixed at that time. Don't know what is the
>>>>> situation with the latest version.
>>>>>
>>>>> Thanks,
>>>>>
>>>>> Dejan
>>>>>
>>>>> Regards,
>>>>>>    Honza
>>>>>>
>>>>>>
>>>>>>> Regards,
>>>>>>> Klaus
>>>>>>>
>>>>>>> On 05/02/2016 06:44 AM, Nikhil Utane wrote:
>>>>>>>
>>>>>>>> Re-sending as I don't see my post on the thread.
>>>>>>>>
>>>>>>>> On Sun, May 1, 2016 at 4:21 PM, Nikhil Utane
>>>>>>>> <nikhil.subscribed at gmail.com <mailto:nikhil.subscribed at gmail.com>>
>>>>>>>>
>>>>>>> wrote:
>>>>>
>>>>>>
>>>>>>>>      Hi,
>>>>>>>>
>>>>>>>>      Looking for some guidance here as we are completely blocked
>>>>>>>>      otherwise :(.
>>>>>>>>
>>>>>>>>      -Regards
>>>>>>>>      Nikhil
>>>>>>>>
>>>>>>>>      On Fri, Apr 29, 2016 at 6:11 PM, Sriram <sriram.ec at gmail.com
>>>>>>>>      <mailto:sriram.ec at gmail.com>> wrote:
>>>>>>>>
>>>>>>>>          Corrected the subject.
>>>>>>>>
>>>>>>>>          We went ahead and captured corosync debug logs for our ppc
>>>>>>>>
>>>>>>> board.
>>>>>
>>>>>>          After log analysis and comparison with the sucessful logs(
>>>>>>>>          from x86 machine) ,
>>>>>>>>          we didnt find *"[ MAIN  ] Completed service
>>>>>>>> synchronization,
>>>>>>>>          ready to provide service.*" in ppc logs.
>>>>>>>>          So, looks like corosync is not in a position to accept
>>>>>>>>          connection from Pacemaker.
>>>>>>>>          Even I tried with the new corosync.conf with no success.
>>>>>>>>
>>>>>>>>          Any hints on this issue would be really helpful.
>>>>>>>>
>>>>>>>>          Attaching ppc_notworking.log, x86_working.log,
>>>>>>>> corosync.conf.
>>>>>>>>
>>>>>>>>          Regards,
>>>>>>>>          Sriram
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>          On Fri, Apr 29, 2016 at 2:44 PM, Sriram <
>>>>>>>> sriram.ec at gmail.com
>>>>>>>>          <mailto:sriram.ec at gmail.com>> wrote:
>>>>>>>>
>>>>>>>>              Hi,
>>>>>>>>
>>>>>>>>              I went ahead and made some changes in file system(Like
>>>>>>>> I
>>>>>>>>              brought in /etc/init.d/corosync and
>>>>>>>>
>>>>>>> /etc/init.d/pacemaker,
>>>>>
>>>>>>              /etc/sysconfig ), After that I was able to run  "pcs
>>>>>>>>              cluster start".
>>>>>>>>              But it failed with the following error
>>>>>>>>               # pcs cluster start
>>>>>>>>              Starting Cluster...
>>>>>>>>              Starting Pacemaker Cluster Manager[FAILED]
>>>>>>>>              Error: unable to start pacemaker
>>>>>>>>
>>>>>>>>              And in the /var/log/pacemaker.log, I saw these errors
>>>>>>>>              pacemakerd:     info: mcp_read_config:  cmap connection
>>>>>>>>              setup failed: CS_ERR_TRY_AGAIN.  Retrying in 4s
>>>>>>>>              Apr 29 08:53:47 [15863] node_cu pacemakerd:     info:
>>>>>>>>              mcp_read_config:  cmap connection setup failed:
>>>>>>>>              CS_ERR_TRY_AGAIN.  Retrying in 5s
>>>>>>>>              Apr 29 08:53:52 [15863] node_cu pacemakerd:  warning:
>>>>>>>>              mcp_read_config:  Could not connect to Cluster
>>>>>>>>              Configuration Database API, error 6
>>>>>>>>              Apr 29 08:53:52 [15863] node_cu pacemakerd:   notice:
>>>>>>>>              main:     Could not obtain corosync config data,
>>>>>>>> exiting
>>>>>>>>              Apr 29 08:53:52 [15863] node_cu pacemakerd:     info:
>>>>>>>>              crm_xml_cleanup:  Cleaning up memory from libxml2
>>>>>>>>
>>>>>>>>
>>>>>>>>              And in the /var/log/Debuglog, I saw these errors coming
>>>>>>>>              from corosync
>>>>>>>>              20160429 085347.487050 <tel:085347.487050> airv_cu
>>>>>>>>              daemon.warn corosync[12857]:   [QB    ] Denied
>>>>>>>>
>>>>>>> connection,
>>>>>
>>>>>>              is not ready (12857-15863-14)
>>>>>>>>              20160429 085347.487067 <tel:085347.487067> airv_cu
>>>>>>>>              daemon.info <http://daemon.info> corosync[12857]:
>>>>>>>>  [QB
>>>>>>>>              ] Denied connection, is not ready (12857-15863-14)
>>>>>>>>
>>>>>>>>
>>>>>>>>              I browsed the code of libqb to find that it is failing
>>>>>>>> in
>>>>>>>>
>>>>>>>>
>>>>>>>> https://github.com/ClusterLabs/libqb/blob/master/lib/ipc_setup.c
>>>>>
>>>>>>
>>>>>>>>              Line 600 :
>>>>>>>>              handle_new_connection function
>>>>>>>>
>>>>>>>>              Line 637:
>>>>>>>>              if (auth_result == 0 &&
>>>>>>>>              c->service->serv_fns.connection_accept) {
>>>>>>>>                      res = c->service->serv_fns.connection_accept(c,
>>>>>>>>                                               c->euid, c->egid);
>>>>>>>>                  }
>>>>>>>>                  if (res != 0) {
>>>>>>>>                      goto send_response;
>>>>>>>>                  }
>>>>>>>>
>>>>>>>>              Any hints on this issue would be really helpful for me
>>>>>>>> to
>>>>>>>>              go ahead.
>>>>>>>>              Please let me know if any logs are required,
>>>>>>>>
>>>>>>>>              Regards,
>>>>>>>>              Sriram
>>>>>>>>
>>>>>>>>              On Thu, Apr 28, 2016 at 2:42 PM, Sriram
>>>>>>>>              <sriram.ec at gmail.com <mailto:sriram.ec at gmail.com>>
>>>>>>>>
>>>>>>> wrote:
>>>>>
>>>>>>
>>>>>>>>                  Thanks Ken and Emmanuel.
>>>>>>>>                  Its a big endian machine. I will try with running
>>>>>>>>
>>>>>>> "pcs
>>>>>
>>>>>>                  cluster setup" and "pcs cluster start"
>>>>>>>>                  Inside cluster.py, "service pacemaker start" and
>>>>>>>>                  "service corosync start" are executed to bring up
>>>>>>>>                  pacemaker and corosync.
>>>>>>>>                  Those service scripts and the infrastructure needed
>>>>>>>>
>>>>>>> to
>>>>>
>>>>>>                  bring up the processes in the above said manner
>>>>>>>>                  doesn't exist in my board.
>>>>>>>>                  As it is a embedded board with the limited memory,
>>>>>>>>                  full fledged linux is not installed.
>>>>>>>>                  Just curious to know, what could be reason the
>>>>>>>>                  pacemaker throws that error.
>>>>>>>>
>>>>>>>>                  /"cmap connection setup failed: CS_ERR_TRY_AGAIN.
>>>>>>>>                  Retrying in 1s"
>>>>>>>>
>>>>>>>>                  /
>>>>>>>>                  Thanks for response.
>>>>>>>>
>>>>>>>>                  Regards,
>>>>>>>>                  Sriram.
>>>>>>>>
>>>>>>>>                  On Thu, Apr 28, 2016 at 8:55 AM, Ken Gaillot
>>>>>>>>                  <kgaillot at redhat.com <mailto:kgaillot at redhat.com>>
>>>>>>>>
>>>>>>> wrote:
>>>>>
>>>>>>
>>>>>>>>                      On 04/27/2016 11:25 AM, emmanuel segura wrote:
>>>>>>>>                      > you need to use pcs to do everything, pcs
>>>>>>>>                      cluster setup and pcs
>>>>>>>>                      > cluster start, try to use the redhat docs for
>>>>>>>>                      more information.
>>>>>>>>
>>>>>>>>                      Agreed -- pcs cluster setup will create a
>>>>>>>> proper
>>>>>>>>                      corosync.conf for you.
>>>>>>>>                      Your corosync.conf below uses corosync 1
>>>>>>>> syntax,
>>>>>>>>                      and there were
>>>>>>>>                      significant changes in corosync 2. In
>>>>>>>> particular,
>>>>>>>>                      you don't need the
>>>>>>>>                      file created in step 4, because pacemaker is no
>>>>>>>>                      longer launched via a
>>>>>>>>                      corosync plugin.
>>>>>>>>
>>>>>>>>                      > 2016-04-27 17:28 GMT+02:00 Sriram
>>>>>>>>                      <sriram.ec at gmail.com <mailto:
>>>>>>>> sriram.ec at gmail.com
>>>>>>>>
>>>>>>> :
>>>>>>>
>>>>>>>>                      >> Dear All,
>>>>>>>>                      >>
>>>>>>>>                      >> I m trying to use pacemaker and corosync for
>>>>>>>>                      the clustering requirement that
>>>>>>>>                      >> came up recently.
>>>>>>>>                      >> We have cross compiled corosync, pacemaker
>>>>>>>> and
>>>>>>>>                      pcs(python) for ppc
>>>>>>>>                      >> environment (Target board where pacemaker
>>>>>>>> and
>>>>>>>>                      corosync are supposed to run)
>>>>>>>>                      >> I m having trouble bringing up pacemaker in
>>>>>>>>                      that environment, though I could
>>>>>>>>                      >> successfully bring up corosync.
>>>>>>>>                      >> Any help is welcome.
>>>>>>>>                      >>
>>>>>>>>                      >> I m using these versions of pacemaker and
>>>>>>>>
>>>>>>> corosync
>>>>>
>>>>>>                      >> [root at node_cu pacemaker]# corosync -v
>>>>>>>>                      >> Corosync Cluster Engine, version '2.3.5'
>>>>>>>>                      >> Copyright (c) 2006-2009 Red Hat, Inc.
>>>>>>>>                      >> [root at node_cu pacemaker]# pacemakerd -$
>>>>>>>>                      >> Pacemaker 1.1.14
>>>>>>>>                      >> Written by Andrew Beekhof
>>>>>>>>                      >>
>>>>>>>>                      >> For running corosync, I did the following.
>>>>>>>>                      >> 1. Created the following directories,
>>>>>>>>                      >>     /var/lib/pacemaker
>>>>>>>>                      >>     /var/lib/corosync
>>>>>>>>                      >>     /var/lib/pacemaker
>>>>>>>>                      >>     /var/lib/pacemaker/cores
>>>>>>>>                      >>     /var/lib/pacemaker/pengine
>>>>>>>>                      >>     /var/lib/pacemaker/blackbox
>>>>>>>>                      >>     /var/lib/pacemaker/cib
>>>>>>>>                      >>
>>>>>>>>                      >>
>>>>>>>>                      >> 2. Created a file called corosync.conf under
>>>>>>>>                      /etc/corosync folder with the
>>>>>>>>                      >> following contents
>>>>>>>>                      >>
>>>>>>>>                      >> totem {
>>>>>>>>                      >>
>>>>>>>>                      >>         version: 2
>>>>>>>>                      >>         token:          5000
>>>>>>>>                      >>         token_retransmits_before_loss_const:
>>>>>>>>
>>>>>>> 20
>>>>>
>>>>>>                      >>         join:           1000
>>>>>>>>                      >>         consensus:      7500
>>>>>>>>                      >>         vsftype:        none
>>>>>>>>                      >>         max_messages:   20
>>>>>>>>                      >>         secauth:        off
>>>>>>>>                      >>         cluster_name:   mycluster
>>>>>>>>                      >>         transport:      udpu
>>>>>>>>                      >>         threads:        0
>>>>>>>>                      >>         clear_node_high_bit: yes
>>>>>>>>                      >>
>>>>>>>>                      >>         interface {
>>>>>>>>                      >>                 ringnumber: 0
>>>>>>>>                      >>                 # The following three values
>>>>>>>>                      need to be set based on your
>>>>>>>>                      >> environment
>>>>>>>>                      >>                 bindnetaddr: 10.x.x.x
>>>>>>>>                      >>                 mcastaddr: 226.94.1.1
>>>>>>>>                      >>                 mcastport: 5405
>>>>>>>>                      >>         }
>>>>>>>>                      >>  }
>>>>>>>>                      >>
>>>>>>>>                      >>  logging {
>>>>>>>>                      >>         fileline: off
>>>>>>>>                      >>         to_syslog: yes
>>>>>>>>                      >>         to_stderr: no
>>>>>>>>                      >>         to_syslog: yes
>>>>>>>>                      >>         logfile: /var/log/corosync.log
>>>>>>>>                      >>         syslog_facility: daemon
>>>>>>>>                      >>         debug: on
>>>>>>>>                      >>         timestamp: on
>>>>>>>>                      >>  }
>>>>>>>>                      >>
>>>>>>>>                      >>  amf {
>>>>>>>>                      >>         mode: disabled
>>>>>>>>                      >>  }
>>>>>>>>                      >>
>>>>>>>>                      >>  quorum {
>>>>>>>>                      >>         provider: corosync_votequorum
>>>>>>>>                      >>  }
>>>>>>>>                      >>
>>>>>>>>                      >> nodelist {
>>>>>>>>                      >>   node {
>>>>>>>>                      >>         ring0_addr: node_cu
>>>>>>>>                      >>         nodeid: 1
>>>>>>>>                      >>        }
>>>>>>>>                      >> }
>>>>>>>>                      >>
>>>>>>>>                      >> 3.  Created authkey under /etc/corosync
>>>>>>>>                      >>
>>>>>>>>                      >> 4.  Created a file called pcmk under
>>>>>>>>                      /etc/corosync/service.d and contents as
>>>>>>>>                      >> below,
>>>>>>>>                      >>       cat pcmk
>>>>>>>>                      >>       service {
>>>>>>>>                      >>          # Load the Pacemaker Cluster
>>>>>>>> Resource
>>>>>>>>                      Manager
>>>>>>>>                      >>          name: pacemaker
>>>>>>>>                      >>          ver:  1
>>>>>>>>                      >>       }
>>>>>>>>                      >>
>>>>>>>>                      >> 5. Added the node name "node_cu" in
>>>>>>>> /etc/hosts
>>>>>>>>                      with 10.X.X.X ip
>>>>>>>>                      >>
>>>>>>>>                      >> 6. ./corosync -f -p & --> this step started
>>>>>>>>                      corosync
>>>>>>>>                      >>
>>>>>>>>                      >> [root at node_cu pacemaker]# netstat -alpn |
>>>>>>>>
>>>>>>> grep
>>>>>
>>>>>>                      -i coros
>>>>>>>>                      >> udp        0      0 10.X.X.X:61841
>>>>>>>>
>>>>>>>   0.0.0.0:*
>>>>>
>>>>>>                      >> 9133/corosync
>>>>>>>>                      >> udp        0      0 10.X.X.X:5405
>>>>>>>>
>>>>>>> 0.0.0.0:*
>>>>>
>>>>>>                      >> 9133/corosync
>>>>>>>>                      >> unix  2      [ ACC ]     STREAM
>>>>>>>>  LISTENING
>>>>>>>>                         148888 9133/corosync
>>>>>>>>                      >> @quorum
>>>>>>>>                      >> unix  2      [ ACC ]     STREAM
>>>>>>>>  LISTENING
>>>>>>>>                         148884 9133/corosync
>>>>>>>>                      >> @cmap
>>>>>>>>                      >> unix  2      [ ACC ]     STREAM
>>>>>>>>  LISTENING
>>>>>>>>                         148887 9133/corosync
>>>>>>>>                      >> @votequorum
>>>>>>>>                      >> unix  2      [ ACC ]     STREAM
>>>>>>>>  LISTENING
>>>>>>>>                         148885 9133/corosync
>>>>>>>>                      >> @cfg
>>>>>>>>                      >> unix  2      [ ACC ]     STREAM
>>>>>>>>  LISTENING
>>>>>>>>                         148886 9133/corosync
>>>>>>>>                      >> @cpg
>>>>>>>>                      >> unix  2      [ ]         DGRAM
>>>>>>>>                        148840 9133/corosync
>>>>>>>>                      >>
>>>>>>>>                      >> 7. ./pacemakerd -f & gives the following
>>>>>>>> error
>>>>>>>>                      and exits.
>>>>>>>>                      >> [root at node_cu pacemaker]# pacemakerd -f
>>>>>>>>                      >> cmap connection setup failed:
>>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 1s
>>>>>>>>                      >> cmap connection setup failed:
>>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 2s
>>>>>>>>                      >> cmap connection setup failed:
>>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 3s
>>>>>>>>                      >> cmap connection setup failed:
>>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 4s
>>>>>>>>                      >> cmap connection setup failed:
>>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 5s
>>>>>>>>                      >> Could not connect to Cluster Configuration
>>>>>>>>                      Database API, error 6
>>>>>>>>                      >>
>>>>>>>>                      >> Can you please point me, what is missing in
>>>>>>>>                      these steps ?
>>>>>>>>                      >>
>>>>>>>>                      >> Before trying these steps, I tried running
>>>>>>>>
>>>>>>> "pcs
>>>>>
>>>>>>                      cluster start", but that
>>>>>>>>                      >> command fails with "service" script not
>>>>>>>> found.
>>>>>>>>                      As the root filesystem
>>>>>>>>                      >> doesn't contain either /etc/init.d/ or
>>>>>>>>                      /sbin/service
>>>>>>>>                      >>
>>>>>>>>                      >> So, the plan is to bring up corosync and
>>>>>>>>                      pacemaker manually, later do the
>>>>>>>>                      >> cluster configuration using "pcs" commands.
>>>>>>>>                      >>
>>>>>>>>                      >> Regards,
>>>>>>>>                      >> Sriram
>>>>>>>>                      >>
>>>>>>>>                      >>
>>>>>>>>
>>>>>>> _______________________________________________
>>>>>
>>>>>>                      >> Users mailing list: Users at clusterlabs.org
>>>>>>>>                      <mailto:Users at clusterlabs.org>
>>>>>>>>                      >>
>>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>>                      >>
>>>>>>>>                      >> Project Home: http://www.clusterlabs.org
>>>>>>>>                      >> Getting started:
>>>>>>>>
>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>>                      >> Bugs: http://bugs.clusterlabs.org
>>>>>>>>                      >>
>>>>>>>>                      >
>>>>>>>>                      >
>>>>>>>>                      >
>>>>>>>>
>>>>>>>>
>>>>>>>>                      _______________________________________________
>>>>>>>>                      Users mailing list: Users at clusterlabs.org
>>>>>>>>                      <mailto:Users at clusterlabs.org>
>>>>>>>>                      http://clusterlabs.org/mailman/listinfo/users
>>>>>>>>
>>>>>>>>                      Project Home: http://www.clusterlabs.org
>>>>>>>>                      Getting started:
>>>>>>>>
>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>>                      Bugs: http://bugs.clusterlabs.org
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>          _______________________________________________
>>>>>>>>          Users mailing list: Users at clusterlabs.org
>>>>>>>>          <mailto:Users at clusterlabs.org>
>>>>>>>>          http://clusterlabs.org/mailman/listinfo/users
>>>>>>>>
>>>>>>>>          Project Home: http://www.clusterlabs.org
>>>>>>>>          Getting started:
>>>>>>>>          http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>          Bugs: http://bugs.clusterlabs.org
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>>
>>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>>> Getting started:
>>>>>>>>
>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>
>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>> Getting started:
>>>>>>>
>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>
>>>>>> Project Home: http://www.clusterlabs.org
>>>>>> Getting started:
>>>>>>
>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list: Users at clusterlabs.org
>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>
>>>>> Project Home: http://www.clusterlabs.org
>>>>> Getting started:
>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>>> _______________________________________________
>>> Users mailing list: Users at clusterlabs.org
>>> http://clusterlabs.org/mailman/listinfo/users
>>>
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>>>
>>>
>>
>> _______________________________________________
>> Users mailing list: Users at clusterlabs.org
>> http://clusterlabs.org/mailman/listinfo/users
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20160517/b3cbb81e/attachment-0003.html>


More information about the Users mailing list