[ClusterLabs] [ClusterLab] : Corosync not initializing successfully

Nikhil Utane nikhil.subscribed at gmail.com
Thu May 5 09:54:22 UTC 2016


It worked for me. :)
I'll wait for your formal patch but until then I am able to proceed
further. (Don't know if I'll run into something else)

However now encountering issue in pacemaker.

May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:  The
cib process (15224) can no longer be respawned, shutting the cluster down.
May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:  The
stonith-ng process (15225) can no longer be respawned, shutting the cluster
down.
May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:  The
lrmd process (15226) can no longer be respawned, shutting the cluster down.
May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:  The
crmd process (15229) can no longer be respawned, shutting the cluster down.
May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:  The
pengine process (15228) can no longer be respawned, shutting the cluster
down.
May 05 09:35:53 [15184] airv_cu pacemakerd:  warning: pcmk_child_exit:  The
attrd process (15227) can no longer be respawned, shutting the cluster down.

Looking into it.

-Thanks
Nikhil

On Thu, May 5, 2016 at 2:58 PM, Jan Friesse <jfriesse at redhat.com> wrote:

> Nikhil
>
> Found the root-cause.
>> In file schedwrk.c, the function handle2void() uses a union which was not
>> initialized.
>> Because of that the handle value was computed incorrectly (lower half was
>> garbage).
>>
>>   56 static hdb_handle_t
>>   57 void2handle (const void *v) { union u u={}; u.v = v; return u.h; }
>>   58 static const void *
>>   59 handle2void (hdb_handle_t h) { union u u={}; u.h = h; return u.v; }
>>
>> After initializing (as highlighted), the corosync initialization seems to
>> be going through fine. Will check other things.
>>
>
> Your patch is incorrect and actually doesn't work. As I said (when
> pointing you to schedwrk.c), I will send you proper patch, but fix that
> issue correctly is not easy.
>
> Regards,
>   Honza
>
>
>> -Regards
>> Nikhil
>>
>> On Tue, May 3, 2016 at 7:04 PM, Nikhil Utane <nikhil.subscribed at gmail.com
>> >
>> wrote:
>>
>> Thanks for your response Dejan.
>>>
>>> I do not know yet whether this has anything to do with endianness.
>>> FWIW, there could be something quirky with the system so keeping all
>>> options open. :)
>>>
>>> I added some debug prints to understand what's happening under the hood.
>>>
>>> *Success case: (on x86 machine): *
>>> [TOTEM ] entering OPERATIONAL state.
>>> [TOTEM ] A new membership (10.206.1.7:137220) was formed. Members
>>> joined:
>>> 181272839
>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=0,
>>> my_high_delivered=0
>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>> my_high_delivered=0
>>> [TOTEM ] Delivering 0 to 1
>>> [TOTEM ] Delivering MCAST message with seq 1 to pending delivery queue
>>> [SYNC  ] Nikhil: Inside sync_deliver_fn. header->id=1
>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=2,
>>> my_high_delivered=1
>>> [TOTEM ] Delivering 1 to 2
>>> [TOTEM ] Delivering MCAST message with seq 2 to pending delivery queue
>>> [SYNC  ] Nikhil: Inside sync_deliver_fn. header->id=0
>>> [SYNC  ] Nikhil: Entering sync_barrier_handler
>>> [SYNC  ] Committing synchronization for corosync configuration map access
>>> .
>>> [TOTEM ] Delivering 2 to 4
>>> [TOTEM ] Delivering MCAST message with seq 3 to pending delivery queue
>>> [TOTEM ] Delivering MCAST message with seq 4 to pending delivery queue
>>> [CPG   ] comparing: sender r(0) ip(10.206.1.7) ; members(old:0 left:0)
>>> [CPG   ] chosen downlist: sender r(0) ip(10.206.1.7) ; members(old:0
>>> left:0)
>>> [SYNC  ] Committing synchronization for corosync cluster closed process
>>> group service v1.01
>>> *[MAIN  ] Completed service synchronization, ready to provide service.*
>>>
>>>
>>> *Failure case: (on ppc)*:
>>>
>>> [TOTEM ] entering OPERATIONAL state.
>>> [TOTEM ] A new membership (10.207.24.101:16) was formed. Members joined:
>>> 181344357
>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=0,
>>> my_high_delivered=0
>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>> my_high_delivered=0
>>> [TOTEM ] Delivering 0 to 1
>>> [TOTEM ] Delivering MCAST message with seq 1 to pending delivery queue
>>> [SYNC  ] Nikhil: Inside sync_deliver_fn header->id=1
>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>> my_high_delivered=1
>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1,
>>> my_high_delivered=1
>>> Above message repeats continuously.
>>>
>>> So it appears that in failure case I do not receive messages with
>>> sequence
>>> number 2-4.
>>> If somebody can throw some ideas that'll help a lot.
>>>
>>> -Thanks
>>> Nikhil
>>>
>>> On Tue, May 3, 2016 at 5:26 PM, Dejan Muhamedagic <dejanmm at fastmail.fm>
>>> wrote:
>>>
>>> Hi,
>>>>
>>>> On Mon, May 02, 2016 at 08:54:09AM +0200, Jan Friesse wrote:
>>>>
>>>>> As your hardware is probably capable of running ppcle and if you have
>>>>>>
>>>>> an
>>>>
>>>>> environment
>>>>>> at hand without too much effort it might pay off to try that.
>>>>>> There are of course distributions out there support corosync on
>>>>>> big-endian architectures
>>>>>> but I don't know if there is an automatized regression for corosync on
>>>>>> big-endian that
>>>>>> would catch big-endian-issues right away with something as current as
>>>>>> your 2.3.5.
>>>>>>
>>>>>
>>>>> No we are not testing big-endian.
>>>>>
>>>>> So totally agree with Klaus. Give a try to ppcle. Also make sure all
>>>>> nodes are little-endian. Corosync should work in mixed BE/LE
>>>>> environment but because it's not tested, it may not work (and it's a
>>>>> bug, so if ppcle works I will try to fix BE).
>>>>>
>>>>
>>>> I tested a cluster consisting of big endian/little endian nodes
>>>> (s390 and x86-64), but that was a while ago. IIRC, all relevant
>>>> bugs in corosync got fixed at that time. Don't know what is the
>>>> situation with the latest version.
>>>>
>>>> Thanks,
>>>>
>>>> Dejan
>>>>
>>>> Regards,
>>>>>    Honza
>>>>>
>>>>>
>>>>>> Regards,
>>>>>> Klaus
>>>>>>
>>>>>> On 05/02/2016 06:44 AM, Nikhil Utane wrote:
>>>>>>
>>>>>>> Re-sending as I don't see my post on the thread.
>>>>>>>
>>>>>>> On Sun, May 1, 2016 at 4:21 PM, Nikhil Utane
>>>>>>> <nikhil.subscribed at gmail.com <mailto:nikhil.subscribed at gmail.com>>
>>>>>>>
>>>>>> wrote:
>>>>
>>>>>
>>>>>>>      Hi,
>>>>>>>
>>>>>>>      Looking for some guidance here as we are completely blocked
>>>>>>>      otherwise :(.
>>>>>>>
>>>>>>>      -Regards
>>>>>>>      Nikhil
>>>>>>>
>>>>>>>      On Fri, Apr 29, 2016 at 6:11 PM, Sriram <sriram.ec at gmail.com
>>>>>>>      <mailto:sriram.ec at gmail.com>> wrote:
>>>>>>>
>>>>>>>          Corrected the subject.
>>>>>>>
>>>>>>>          We went ahead and captured corosync debug logs for our ppc
>>>>>>>
>>>>>> board.
>>>>
>>>>>          After log analysis and comparison with the sucessful logs(
>>>>>>>          from x86 machine) ,
>>>>>>>          we didnt find *"[ MAIN  ] Completed service synchronization,
>>>>>>>          ready to provide service.*" in ppc logs.
>>>>>>>          So, looks like corosync is not in a position to accept
>>>>>>>          connection from Pacemaker.
>>>>>>>          Even I tried with the new corosync.conf with no success.
>>>>>>>
>>>>>>>          Any hints on this issue would be really helpful.
>>>>>>>
>>>>>>>          Attaching ppc_notworking.log, x86_working.log,
>>>>>>> corosync.conf.
>>>>>>>
>>>>>>>          Regards,
>>>>>>>          Sriram
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>          On Fri, Apr 29, 2016 at 2:44 PM, Sriram <
>>>>>>> sriram.ec at gmail.com
>>>>>>>          <mailto:sriram.ec at gmail.com>> wrote:
>>>>>>>
>>>>>>>              Hi,
>>>>>>>
>>>>>>>              I went ahead and made some changes in file system(Like I
>>>>>>>              brought in /etc/init.d/corosync and
>>>>>>>
>>>>>> /etc/init.d/pacemaker,
>>>>
>>>>>              /etc/sysconfig ), After that I was able to run  "pcs
>>>>>>>              cluster start".
>>>>>>>              But it failed with the following error
>>>>>>>               # pcs cluster start
>>>>>>>              Starting Cluster...
>>>>>>>              Starting Pacemaker Cluster Manager[FAILED]
>>>>>>>              Error: unable to start pacemaker
>>>>>>>
>>>>>>>              And in the /var/log/pacemaker.log, I saw these errors
>>>>>>>              pacemakerd:     info: mcp_read_config:  cmap connection
>>>>>>>              setup failed: CS_ERR_TRY_AGAIN.  Retrying in 4s
>>>>>>>              Apr 29 08:53:47 [15863] node_cu pacemakerd:     info:
>>>>>>>              mcp_read_config:  cmap connection setup failed:
>>>>>>>              CS_ERR_TRY_AGAIN.  Retrying in 5s
>>>>>>>              Apr 29 08:53:52 [15863] node_cu pacemakerd:  warning:
>>>>>>>              mcp_read_config:  Could not connect to Cluster
>>>>>>>              Configuration Database API, error 6
>>>>>>>              Apr 29 08:53:52 [15863] node_cu pacemakerd:   notice:
>>>>>>>              main:     Could not obtain corosync config data, exiting
>>>>>>>              Apr 29 08:53:52 [15863] node_cu pacemakerd:     info:
>>>>>>>              crm_xml_cleanup:  Cleaning up memory from libxml2
>>>>>>>
>>>>>>>
>>>>>>>              And in the /var/log/Debuglog, I saw these errors coming
>>>>>>>              from corosync
>>>>>>>              20160429 085347.487050 <tel:085347.487050> airv_cu
>>>>>>>              daemon.warn corosync[12857]:   [QB    ] Denied
>>>>>>>
>>>>>> connection,
>>>>
>>>>>              is not ready (12857-15863-14)
>>>>>>>              20160429 085347.487067 <tel:085347.487067> airv_cu
>>>>>>>              daemon.info <http://daemon.info> corosync[12857]:   [QB
>>>>>>>              ] Denied connection, is not ready (12857-15863-14)
>>>>>>>
>>>>>>>
>>>>>>>              I browsed the code of libqb to find that it is failing
>>>>>>> in
>>>>>>>
>>>>>>>
>>>>>>> https://github.com/ClusterLabs/libqb/blob/master/lib/ipc_setup.c
>>>>
>>>>>
>>>>>>>              Line 600 :
>>>>>>>              handle_new_connection function
>>>>>>>
>>>>>>>              Line 637:
>>>>>>>              if (auth_result == 0 &&
>>>>>>>              c->service->serv_fns.connection_accept) {
>>>>>>>                      res = c->service->serv_fns.connection_accept(c,
>>>>>>>                                               c->euid, c->egid);
>>>>>>>                  }
>>>>>>>                  if (res != 0) {
>>>>>>>                      goto send_response;
>>>>>>>                  }
>>>>>>>
>>>>>>>              Any hints on this issue would be really helpful for me
>>>>>>> to
>>>>>>>              go ahead.
>>>>>>>              Please let me know if any logs are required,
>>>>>>>
>>>>>>>              Regards,
>>>>>>>              Sriram
>>>>>>>
>>>>>>>              On Thu, Apr 28, 2016 at 2:42 PM, Sriram
>>>>>>>              <sriram.ec at gmail.com <mailto:sriram.ec at gmail.com>>
>>>>>>>
>>>>>> wrote:
>>>>
>>>>>
>>>>>>>                  Thanks Ken and Emmanuel.
>>>>>>>                  Its a big endian machine. I will try with running
>>>>>>>
>>>>>> "pcs
>>>>
>>>>>                  cluster setup" and "pcs cluster start"
>>>>>>>                  Inside cluster.py, "service pacemaker start" and
>>>>>>>                  "service corosync start" are executed to bring up
>>>>>>>                  pacemaker and corosync.
>>>>>>>                  Those service scripts and the infrastructure needed
>>>>>>>
>>>>>> to
>>>>
>>>>>                  bring up the processes in the above said manner
>>>>>>>                  doesn't exist in my board.
>>>>>>>                  As it is a embedded board with the limited memory,
>>>>>>>                  full fledged linux is not installed.
>>>>>>>                  Just curious to know, what could be reason the
>>>>>>>                  pacemaker throws that error.
>>>>>>>
>>>>>>>                  /"cmap connection setup failed: CS_ERR_TRY_AGAIN.
>>>>>>>                  Retrying in 1s"
>>>>>>>
>>>>>>>                  /
>>>>>>>                  Thanks for response.
>>>>>>>
>>>>>>>                  Regards,
>>>>>>>                  Sriram.
>>>>>>>
>>>>>>>                  On Thu, Apr 28, 2016 at 8:55 AM, Ken Gaillot
>>>>>>>                  <kgaillot at redhat.com <mailto:kgaillot at redhat.com>>
>>>>>>>
>>>>>> wrote:
>>>>
>>>>>
>>>>>>>                      On 04/27/2016 11:25 AM, emmanuel segura wrote:
>>>>>>>                      > you need to use pcs to do everything, pcs
>>>>>>>                      cluster setup and pcs
>>>>>>>                      > cluster start, try to use the redhat docs for
>>>>>>>                      more information.
>>>>>>>
>>>>>>>                      Agreed -- pcs cluster setup will create a proper
>>>>>>>                      corosync.conf for you.
>>>>>>>                      Your corosync.conf below uses corosync 1 syntax,
>>>>>>>                      and there were
>>>>>>>                      significant changes in corosync 2. In
>>>>>>> particular,
>>>>>>>                      you don't need the
>>>>>>>                      file created in step 4, because pacemaker is no
>>>>>>>                      longer launched via a
>>>>>>>                      corosync plugin.
>>>>>>>
>>>>>>>                      > 2016-04-27 17:28 GMT+02:00 Sriram
>>>>>>>                      <sriram.ec at gmail.com <mailto:
>>>>>>> sriram.ec at gmail.com
>>>>>>>
>>>>>> :
>>>>>>
>>>>>>>                      >> Dear All,
>>>>>>>                      >>
>>>>>>>                      >> I m trying to use pacemaker and corosync for
>>>>>>>                      the clustering requirement that
>>>>>>>                      >> came up recently.
>>>>>>>                      >> We have cross compiled corosync, pacemaker
>>>>>>> and
>>>>>>>                      pcs(python) for ppc
>>>>>>>                      >> environment (Target board where pacemaker and
>>>>>>>                      corosync are supposed to run)
>>>>>>>                      >> I m having trouble bringing up pacemaker in
>>>>>>>                      that environment, though I could
>>>>>>>                      >> successfully bring up corosync.
>>>>>>>                      >> Any help is welcome.
>>>>>>>                      >>
>>>>>>>                      >> I m using these versions of pacemaker and
>>>>>>>
>>>>>> corosync
>>>>
>>>>>                      >> [root at node_cu pacemaker]# corosync -v
>>>>>>>                      >> Corosync Cluster Engine, version '2.3.5'
>>>>>>>                      >> Copyright (c) 2006-2009 Red Hat, Inc.
>>>>>>>                      >> [root at node_cu pacemaker]# pacemakerd -$
>>>>>>>                      >> Pacemaker 1.1.14
>>>>>>>                      >> Written by Andrew Beekhof
>>>>>>>                      >>
>>>>>>>                      >> For running corosync, I did the following.
>>>>>>>                      >> 1. Created the following directories,
>>>>>>>                      >>     /var/lib/pacemaker
>>>>>>>                      >>     /var/lib/corosync
>>>>>>>                      >>     /var/lib/pacemaker
>>>>>>>                      >>     /var/lib/pacemaker/cores
>>>>>>>                      >>     /var/lib/pacemaker/pengine
>>>>>>>                      >>     /var/lib/pacemaker/blackbox
>>>>>>>                      >>     /var/lib/pacemaker/cib
>>>>>>>                      >>
>>>>>>>                      >>
>>>>>>>                      >> 2. Created a file called corosync.conf under
>>>>>>>                      /etc/corosync folder with the
>>>>>>>                      >> following contents
>>>>>>>                      >>
>>>>>>>                      >> totem {
>>>>>>>                      >>
>>>>>>>                      >>         version: 2
>>>>>>>                      >>         token:          5000
>>>>>>>                      >>         token_retransmits_before_loss_const:
>>>>>>>
>>>>>> 20
>>>>
>>>>>                      >>         join:           1000
>>>>>>>                      >>         consensus:      7500
>>>>>>>                      >>         vsftype:        none
>>>>>>>                      >>         max_messages:   20
>>>>>>>                      >>         secauth:        off
>>>>>>>                      >>         cluster_name:   mycluster
>>>>>>>                      >>         transport:      udpu
>>>>>>>                      >>         threads:        0
>>>>>>>                      >>         clear_node_high_bit: yes
>>>>>>>                      >>
>>>>>>>                      >>         interface {
>>>>>>>                      >>                 ringnumber: 0
>>>>>>>                      >>                 # The following three values
>>>>>>>                      need to be set based on your
>>>>>>>                      >> environment
>>>>>>>                      >>                 bindnetaddr: 10.x.x.x
>>>>>>>                      >>                 mcastaddr: 226.94.1.1
>>>>>>>                      >>                 mcastport: 5405
>>>>>>>                      >>         }
>>>>>>>                      >>  }
>>>>>>>                      >>
>>>>>>>                      >>  logging {
>>>>>>>                      >>         fileline: off
>>>>>>>                      >>         to_syslog: yes
>>>>>>>                      >>         to_stderr: no
>>>>>>>                      >>         to_syslog: yes
>>>>>>>                      >>         logfile: /var/log/corosync.log
>>>>>>>                      >>         syslog_facility: daemon
>>>>>>>                      >>         debug: on
>>>>>>>                      >>         timestamp: on
>>>>>>>                      >>  }
>>>>>>>                      >>
>>>>>>>                      >>  amf {
>>>>>>>                      >>         mode: disabled
>>>>>>>                      >>  }
>>>>>>>                      >>
>>>>>>>                      >>  quorum {
>>>>>>>                      >>         provider: corosync_votequorum
>>>>>>>                      >>  }
>>>>>>>                      >>
>>>>>>>                      >> nodelist {
>>>>>>>                      >>   node {
>>>>>>>                      >>         ring0_addr: node_cu
>>>>>>>                      >>         nodeid: 1
>>>>>>>                      >>        }
>>>>>>>                      >> }
>>>>>>>                      >>
>>>>>>>                      >> 3.  Created authkey under /etc/corosync
>>>>>>>                      >>
>>>>>>>                      >> 4.  Created a file called pcmk under
>>>>>>>                      /etc/corosync/service.d and contents as
>>>>>>>                      >> below,
>>>>>>>                      >>       cat pcmk
>>>>>>>                      >>       service {
>>>>>>>                      >>          # Load the Pacemaker Cluster
>>>>>>> Resource
>>>>>>>                      Manager
>>>>>>>                      >>          name: pacemaker
>>>>>>>                      >>          ver:  1
>>>>>>>                      >>       }
>>>>>>>                      >>
>>>>>>>                      >> 5. Added the node name "node_cu" in
>>>>>>> /etc/hosts
>>>>>>>                      with 10.X.X.X ip
>>>>>>>                      >>
>>>>>>>                      >> 6. ./corosync -f -p & --> this step started
>>>>>>>                      corosync
>>>>>>>                      >>
>>>>>>>                      >> [root at node_cu pacemaker]# netstat -alpn |
>>>>>>>
>>>>>> grep
>>>>
>>>>>                      -i coros
>>>>>>>                      >> udp        0      0 10.X.X.X:61841
>>>>>>>
>>>>>>   0.0.0.0:*
>>>>
>>>>>                      >> 9133/corosync
>>>>>>>                      >> udp        0      0 10.X.X.X:5405
>>>>>>>
>>>>>> 0.0.0.0:*
>>>>
>>>>>                      >> 9133/corosync
>>>>>>>                      >> unix  2      [ ACC ]     STREAM     LISTENING
>>>>>>>                         148888 9133/corosync
>>>>>>>                      >> @quorum
>>>>>>>                      >> unix  2      [ ACC ]     STREAM     LISTENING
>>>>>>>                         148884 9133/corosync
>>>>>>>                      >> @cmap
>>>>>>>                      >> unix  2      [ ACC ]     STREAM     LISTENING
>>>>>>>                         148887 9133/corosync
>>>>>>>                      >> @votequorum
>>>>>>>                      >> unix  2      [ ACC ]     STREAM     LISTENING
>>>>>>>                         148885 9133/corosync
>>>>>>>                      >> @cfg
>>>>>>>                      >> unix  2      [ ACC ]     STREAM     LISTENING
>>>>>>>                         148886 9133/corosync
>>>>>>>                      >> @cpg
>>>>>>>                      >> unix  2      [ ]         DGRAM
>>>>>>>                        148840 9133/corosync
>>>>>>>                      >>
>>>>>>>                      >> 7. ./pacemakerd -f & gives the following
>>>>>>> error
>>>>>>>                      and exits.
>>>>>>>                      >> [root at node_cu pacemaker]# pacemakerd -f
>>>>>>>                      >> cmap connection setup failed:
>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 1s
>>>>>>>                      >> cmap connection setup failed:
>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 2s
>>>>>>>                      >> cmap connection setup failed:
>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 3s
>>>>>>>                      >> cmap connection setup failed:
>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 4s
>>>>>>>                      >> cmap connection setup failed:
>>>>>>>                      CS_ERR_TRY_AGAIN.  Retrying in 5s
>>>>>>>                      >> Could not connect to Cluster Configuration
>>>>>>>                      Database API, error 6
>>>>>>>                      >>
>>>>>>>                      >> Can you please point me, what is missing in
>>>>>>>                      these steps ?
>>>>>>>                      >>
>>>>>>>                      >> Before trying these steps, I tried running
>>>>>>>
>>>>>> "pcs
>>>>
>>>>>                      cluster start", but that
>>>>>>>                      >> command fails with "service" script not
>>>>>>> found.
>>>>>>>                      As the root filesystem
>>>>>>>                      >> doesn't contain either /etc/init.d/ or
>>>>>>>                      /sbin/service
>>>>>>>                      >>
>>>>>>>                      >> So, the plan is to bring up corosync and
>>>>>>>                      pacemaker manually, later do the
>>>>>>>                      >> cluster configuration using "pcs" commands.
>>>>>>>                      >>
>>>>>>>                      >> Regards,
>>>>>>>                      >> Sriram
>>>>>>>                      >>
>>>>>>>                      >>
>>>>>>>
>>>>>> _______________________________________________
>>>>
>>>>>                      >> Users mailing list: Users at clusterlabs.org
>>>>>>>                      <mailto:Users at clusterlabs.org>
>>>>>>>                      >>
>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>                      >>
>>>>>>>                      >> Project Home: http://www.clusterlabs.org
>>>>>>>                      >> Getting started:
>>>>>>>
>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>
>>>>>                      >> Bugs: http://bugs.clusterlabs.org
>>>>>>>                      >>
>>>>>>>                      >
>>>>>>>                      >
>>>>>>>                      >
>>>>>>>
>>>>>>>
>>>>>>>                      _______________________________________________
>>>>>>>                      Users mailing list: Users at clusterlabs.org
>>>>>>>                      <mailto:Users at clusterlabs.org>
>>>>>>>                      http://clusterlabs.org/mailman/listinfo/users
>>>>>>>
>>>>>>>                      Project Home: http://www.clusterlabs.org
>>>>>>>                      Getting started:
>>>>>>>
>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>
>>>>>                      Bugs: http://bugs.clusterlabs.org
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>          _______________________________________________
>>>>>>>          Users mailing list: Users at clusterlabs.org
>>>>>>>          <mailto:Users at clusterlabs.org>
>>>>>>>          http://clusterlabs.org/mailman/listinfo/users
>>>>>>>
>>>>>>>          Project Home: http://www.clusterlabs.org
>>>>>>>          Getting started:
>>>>>>>          http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>          Bugs: http://bugs.clusterlabs.org
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>>
>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>> Getting started:
>>>>>>>
>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>
>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>>
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list: Users at clusterlabs.org
>>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>>
>>>>>> Project Home: http://www.clusterlabs.org
>>>>>> Getting started:
>>>>>>
>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>
>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list: Users at clusterlabs.org
>>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>>
>>>>> Project Home: http://www.clusterlabs.org
>>>>> Getting started:
>>>>>
>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>
>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list: Users at clusterlabs.org
>>>> http://clusterlabs.org/mailman/listinfo/users
>>>>
>>>> Project Home: http://www.clusterlabs.org
>>>> Getting started:
>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>> Bugs: http://bugs.clusterlabs.org
>>>>
>>>>
>>>
>>>
>>
>>
>> _______________________________________________
>> Users mailing list: Users at clusterlabs.org
>> http://clusterlabs.org/mailman/listinfo/users
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>>
>
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20160505/ab9e7255/attachment.htm>


More information about the Users mailing list