[Pacemaker] hangs pending

Andrew Beekhof andrew at beekhof.net
Wed Feb 19 02:42:48 UTC 2014


On 18 Feb 2014, at 9:29 pm, Andrey Groshev <greenx at yandex.ru> wrote:

> Hi, ALL and Andrew!
> 
> Today is a good day - I killed a lot, and a lot of shooting at me.
> In general - I am happy (almost like an elephant)   :) 
> Except resources on the node are important to me eight processes: corosync,pacemakerd,cib,stonithd,lrmd,attrd,pengine,crmd.
> I killed them with different signals (4,6,11 and even 9).
> Behavior does not depend of number signal - it's good.
> If STONITH send reboot to the node - it rebooted and rejoined the cluster - too it's good.
> But the behavior is different from killing various demons.
> 
> Turned four groups:
> 1. corosync,cib - STONITH work 100%.
> Kill via any signals - call STONITH and reboot.

excellent

> 3. stonithd,attrd,pengine - not need STONITH 
> This daemons simple restart, resources - stay running.

right

> 
> 2. lrmd,crmd - strange behavior STONITH.
> Sometimes called STONITH - and the corresponding reaction.
> Sometimes restart daemon

The daemon will always try to restart, the only variable is how long it takes the peer to notice and initiate fencing.
If the failure happens just before a they're due to receive totem token, the failure will be very quickly detected and the node fenced.
If the failure happens just after, then detection will take longer - giving the node longer to recover and not be fenced.

So fence/not fence is normal and to be expected.


> and restart resources with large delay MS:pgsql.
> One time after restart crmd - pgsql don't restart.

I would not expect pgsql to ever restart - if the RA does its job properly anyway.
In the case the node is not fenced, the crmd will respawn and the the PE will request that it re-detect the state of all resources.

If the agent reports "all good", then there is nothing more to do.
If the agent is not reporting "all good", you should really be asking why.

> 
> 4. pacemakerd - nothing happens.

On non-systemd based machines, correct.

On a systemd based machine pacemakerd is respawned and reattaches to the existing daemons.
Any subsequent daemon failure will be detected and the daemon respawned.

> And then I can kill any process of the third group. They do not restart.

Until they become needed.
Eg. if the DC goes to invoke the policy engine, that will fail causing the crmd to fail and the node to be fenced.

> Generaly don't touch corosync,cib and maybe lrmd,crmd.
> 
> What do you think about this?
> The main question of this topic - we decided. 
> But this varied behavior - another big problem.
> 
> 
> 
> 
> 17.02.2014, 08:52, "Andrey Groshev" <greenx at yandex.ru>:
>> 17.02.2014, 02:27, "Andrew Beekhof" <andrew at beekhof.net>:
>> 
>>>  With no quick follow-up, dare one hope that means the patch worked? :-)
>> 
>> Hi,
>> No, unfortunately the chief changed my plans on Friday and all day I was engaged in a parallel project.
>> I hope that today have time to carry out the necessary tests.
>> 
>>>  On 14 Feb 2014, at 3:37 pm, Andrey Groshev <greenx at yandex.ru> wrote:
>>>>   Yes, of course. Now beginning build world and test )
>>>> 
>>>>   14.02.2014, 04:41, "Andrew Beekhof" <andrew at beekhof.net>:
>>>>>   The previous patch wasn't quite right.
>>>>>   Could you try this new one?
>>>>> 
>>>>>      http://paste.fedoraproject.org/77123/13923376/
>>>>> 
>>>>>   [11:23 AM] beekhof at f19 ~/Development/sources/pacemaker/devel ☺ # git diff
>>>>>   diff --git a/crmd/callbacks.c b/crmd/callbacks.c
>>>>>   index ac4b905..d49525b 100644
>>>>>   --- a/crmd/callbacks.c
>>>>>   +++ b/crmd/callbacks.c
>>>>>   @@ -199,8 +199,7 @@ peer_update_callback(enum crm_status_type type, crm_node_t * node, const void *d
>>>>>                    stop_te_timer(down->timer);
>>>>> 
>>>>>                    flags |= node_update_join | node_update_expected;
>>>>>   -                crm_update_peer_join(__FUNCTION__, node, crm_join_none);
>>>>>   -                crm_update_peer_expected(__FUNCTION__, node, CRMD_JOINSTATE_DOWN);
>>>>>   +                crmd_peer_down(node, FALSE);
>>>>>                    check_join_state(fsa_state, __FUNCTION__);
>>>>> 
>>>>>                    update_graph(transition_graph, down);
>>>>>   diff --git a/crmd/crmd_utils.h b/crmd/crmd_utils.h
>>>>>   index bc472c2..1a2577a 100644
>>>>>   --- a/crmd/crmd_utils.h
>>>>>   +++ b/crmd/crmd_utils.h
>>>>>   @@ -100,6 +100,7 @@ void crmd_join_phase_log(int level);
>>>>>    const char *get_timer_desc(fsa_timer_t * timer);
>>>>>    gboolean too_many_st_failures(void);
>>>>>    void st_fail_count_reset(const char * target);
>>>>>   +void crmd_peer_down(crm_node_t *peer, bool full);
>>>>> 
>>>>>    #  define fsa_register_cib_callback(id, flag, data, fn) do {              \
>>>>>            fsa_cib_conn->cmds->register_callback(                          \
>>>>>   diff --git a/crmd/te_actions.c b/crmd/te_actions.c
>>>>>   index f31d4ec..3bfce59 100644
>>>>>   --- a/crmd/te_actions.c
>>>>>   +++ b/crmd/te_actions.c
>>>>>   @@ -80,11 +80,8 @@ send_stonith_update(crm_action_t * action, const char *target, const char *uuid)
>>>>>            crm_info("Recording uuid '%s' for node '%s'", uuid, target);
>>>>>            peer->uuid = strdup(uuid);
>>>>>        }
>>>>>   -    crm_update_peer_proc(__FUNCTION__, peer, crm_proc_none, NULL);
>>>>>   -    crm_update_peer_state(__FUNCTION__, peer, CRM_NODE_LOST, 0);
>>>>>   -    crm_update_peer_expected(__FUNCTION__, peer, CRMD_JOINSTATE_DOWN);
>>>>>   -    crm_update_peer_join(__FUNCTION__, peer, crm_join_none);
>>>>> 
>>>>>   +    crmd_peer_down(peer, TRUE);
>>>>>        node_state =
>>>>>            do_update_node_cib(peer,
>>>>>                               node_update_cluster | node_update_peer | node_update_join |
>>>>>   diff --git a/crmd/te_utils.c b/crmd/te_utils.c
>>>>>   index ad7e573..0c92e95 100644
>>>>>   --- a/crmd/te_utils.c
>>>>>   +++ b/crmd/te_utils.c
>>>>>   @@ -247,10 +247,7 @@ tengine_stonith_notify(stonith_t * st, stonith_event_t * st_event)
>>>>> 
>>>>>            }
>>>>> 
>>>>>   -        crm_update_peer_proc(__FUNCTION__, peer, crm_proc_none, NULL);
>>>>>   -        crm_update_peer_state(__FUNCTION__, peer, CRM_NODE_LOST, 0);
>>>>>   -        crm_update_peer_expected(__FUNCTION__, peer, CRMD_JOINSTATE_DOWN);
>>>>>   -        crm_update_peer_join(__FUNCTION__, peer, crm_join_none);
>>>>>   +        crmd_peer_down(peer, TRUE);
>>>>>         }
>>>>>    }
>>>>> 
>>>>>   diff --git a/crmd/utils.c b/crmd/utils.c
>>>>>   index 3988cfe..2df53ab 100644
>>>>>   --- a/crmd/utils.c
>>>>>   +++ b/crmd/utils.c
>>>>>   @@ -1077,3 +1077,13 @@ update_attrd_remote_node_removed(const char *host, const char *user_name)
>>>>>        crm_trace("telling attrd to clear attributes for remote host %s", host);
>>>>>        update_attrd_helper(host, NULL, NULL, user_name, TRUE, 'C');
>>>>>    }
>>>>>   +
>>>>>   +void crmd_peer_down(crm_node_t *peer, bool full)
>>>>>   +{
>>>>>   +    if(full && peer->state == NULL) {
>>>>>   +        crm_update_peer_state(__FUNCTION__, peer, CRM_NODE_LOST, 0);
>>>>>   +        crm_update_peer_proc(__FUNCTION__, peer, crm_proc_none, NULL);
>>>>>   +    }
>>>>>   +    crm_update_peer_join(__FUNCTION__, peer, crm_join_none);
>>>>>   +    crm_update_peer_expected(__FUNCTION__, peer, CRMD_JOINSTATE_DOWN);
>>>>>   +}
>>>>> 
>>>>>   On 16 Jan 2014, at 7:24 pm, Andrey Groshev <greenx at yandex.ru> wrote:
>>>>>>    16.01.2014, 01:30, "Andrew Beekhof" <andrew at beekhof.net>:
>>>>>>>    On 16 Jan 2014, at 12:41 am, Andrey Groshev <greenx at yandex.ru> wrote:
>>>>>>>>     15.01.2014, 02:53, "Andrew Beekhof" <andrew at beekhof.net>:
>>>>>>>>>     On 15 Jan 2014, at 12:15 am, Andrey Groshev <greenx at yandex.ru> wrote:
>>>>>>>>>>      14.01.2014, 10:00, "Andrey Groshev" <greenx at yandex.ru>:
>>>>>>>>>>>      14.01.2014, 07:47, "Andrew Beekhof" <andrew at beekhof.net>:
>>>>>>>>>>>>       Ok, here's what happens:
>>>>>>>>>>>> 
>>>>>>>>>>>>       1. node2 is lost
>>>>>>>>>>>>       2. fencing of node2 starts
>>>>>>>>>>>>       3. node2 reboots (and cluster starts)
>>>>>>>>>>>>       4. node2 returns to the membership
>>>>>>>>>>>>       5. node2 is marked as a cluster member
>>>>>>>>>>>>       6. DC tries to bring it into the cluster, but needs to cancel the active transition first.
>>>>>>>>>>>>          Which is a problem since the node2 fencing operation is part of that
>>>>>>>>>>>>       7. node2 is in a transition (pending) state until fencing passes or fails
>>>>>>>>>>>>       8a. fencing fails: transition completes and the node joins the cluster
>>>>>>>>>>>> 
>>>>>>>>>>>>       Thats in theory, except we automatically try again. Which isn't appropriate.
>>>>>>>>>>>>       This should be relatively easy to fix.
>>>>>>>>>>>> 
>>>>>>>>>>>>       8b. fencing passes: the node is incorrectly marked as offline
>>>>>>>>>>>> 
>>>>>>>>>>>>       This I have no idea how to fix yet.
>>>>>>>>>>>> 
>>>>>>>>>>>>       On another note, it doesn't look like this agent works at all.
>>>>>>>>>>>>       The node has been back online for a long time and the agent is still timing out after 10 minutes.
>>>>>>>>>>>>       So "Once the script makes sure that the victim will rebooted and again available via ssh - it exit with 0." does not seem true.
>>>>>>>>>>>      Damn. Looks like you're right. At some time I broke my agent and had not noticed it. Who will understand.
>>>>>>>>>>      I repaired my agent - after send reboot he is wait STDIN.
>>>>>>>>>>      Returned "normally" a behavior - hangs "pending", until manually send reboot. :)
>>>>>>>>>     Right. Now you're in case 8b.
>>>>>>>>> 
>>>>>>>>>     Can you try this patch:  http://paste.fedoraproject.org/68450/38973966
>>>>>>>>     Killed all day experiences.
>>>>>>>>     It turns out here that:
>>>>>>>>     1. Did cluster.
>>>>>>>>     2. On the node-2 send signal (-4) - killed corosink
>>>>>>>>     3. From node-1 (there DC) - stonith sent reboot
>>>>>>>>     4. Noda rebooted and resources start.
>>>>>>>>     5. Again. On the node-2 send signal (-4) - killed corosink
>>>>>>>>     6. Again. From node-1 (there DC) - stonith sent reboot
>>>>>>>>     7. Noda-2 rebooted and hangs in "pending"
>>>>>>>>     8. Waiting, waiting..... manually reboot.
>>>>>>>>     9. Noda-2 reboot and raised resources start.
>>>>>>>>     10. GOTO p.2
>>>>>>>    Logs?
>>>>>>    Yesterday I wrote an additional letter why not put the logs.
>>>>>>    Read it please, it contains a few more questions.
>>>>>>    Today again began to hang and continue along the same cycle.
>>>>>>    Logs here http://send2me.ru/crmrep2.tar.bz2
>>>>>>>>>>      New logs: http://send2me.ru/crmrep1.tar.bz2
>>>>>>>>>>>>       On 14 Jan 2014, at 1:19 pm, Andrew Beekhof <andrew at beekhof.net> wrote:
>>>>>>>>>>>>>        Apart from anything else, your timeout needs to be bigger:
>>>>>>>>>>>>> 
>>>>>>>>>>>>>        Jan 13 12:21:36 [17223] dev-cluster2-node1.unix.tensor.ru stonith-ng: (  commands.c:1321  )   error: log_operation: Operation 'reboot' [11331] (call 2 from crmd.17227) for host 'dev-cluster2-node2.unix.tensor.ru' with device 'st1' returned: -62 (Timer expired)
>>>>>>>>>>>>> 
>>>>>>>>>>>>>        On 14 Jan 2014, at 7:18 am, Andrew Beekhof <andrew at beekhof.net> wrote:
>>>>>>>>>>>>>>        On 13 Jan 2014, at 8:31 pm, Andrey Groshev <greenx at yandex.ru> wrote:
>>>>>>>>>>>>>>>        13.01.2014, 02:51, "Andrew Beekhof" <andrew at beekhof.net>:
>>>>>>>>>>>>>>>>        On 10 Jan 2014, at 9:55 pm, Andrey Groshev <greenx at yandex.ru> wrote:
>>>>>>>>>>>>>>>>>        10.01.2014, 14:31, "Andrey Groshev" <greenx at yandex.ru>:
>>>>>>>>>>>>>>>>>>        10.01.2014, 14:01, "Andrew Beekhof" <andrew at beekhof.net>:
>>>>>>>>>>>>>>>>>>>        On 10 Jan 2014, at 5:03 pm, Andrey Groshev <greenx at yandex.ru> wrote:
>>>>>>>>>>>>>>>>>>>>         10.01.2014, 05:29, "Andrew Beekhof" <andrew at beekhof.net>:
>>>>>>>>>>>>>>>>>>>>>          On 9 Jan 2014, at 11:11 pm, Andrey Groshev <greenx at yandex.ru> wrote:
>>>>>>>>>>>>>>>>>>>>>>           08.01.2014, 06:22, "Andrew Beekhof" <andrew at beekhof.net>:
>>>>>>>>>>>>>>>>>>>>>>>           On 29 Nov 2013, at 7:17 pm, Andrey Groshev <greenx at yandex.ru> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>            Hi, ALL.
>>>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>>>>            I'm still trying to cope with the fact that after the fence - node hangs in "pending".
>>>>>>>>>>>>>>>>>>>>>>>           Please define "pending".  Where did you see this?
>>>>>>>>>>>>>>>>>>>>>>           In crm_mon:
>>>>>>>>>>>>>>>>>>>>>>           ......
>>>>>>>>>>>>>>>>>>>>>>           Node dev-cluster2-node2 (172793105): pending
>>>>>>>>>>>>>>>>>>>>>>           ......
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>>           The experiment was like this:
>>>>>>>>>>>>>>>>>>>>>>           Four nodes in cluster.
>>>>>>>>>>>>>>>>>>>>>>           On one of them kill corosync or pacemakerd (signal 4 or 6 oк 11).
>>>>>>>>>>>>>>>>>>>>>>           Thereafter, the remaining start it constantly reboot, under various pretexts, "softly whistling", "fly low", "not a cluster member!" ...
>>>>>>>>>>>>>>>>>>>>>>           Then in the log fell out "Too many failures ...."
>>>>>>>>>>>>>>>>>>>>>>           All this time in the status in crm_mon is "pending".
>>>>>>>>>>>>>>>>>>>>>>           Depending on the wind direction changed to "UNCLEAN"
>>>>>>>>>>>>>>>>>>>>>>           Much time has passed and I can not accurately describe the behavior...
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>>           Now I am in the following state:
>>>>>>>>>>>>>>>>>>>>>>           I tried locate the problem. Came here with this.
>>>>>>>>>>>>>>>>>>>>>>           I set big value in property stonith-timeout="600s".
>>>>>>>>>>>>>>>>>>>>>>           And got the following behavior:
>>>>>>>>>>>>>>>>>>>>>>           1. pkill -4 corosync
>>>>>>>>>>>>>>>>>>>>>>           2. from node with DC call my fence agent "sshbykey"
>>>>>>>>>>>>>>>>>>>>>>           3. It sends reboot victim and waits until she comes to life again.
>>>>>>>>>>>>>>>>>>>>>          Hmmm.... what version of pacemaker?
>>>>>>>>>>>>>>>>>>>>>          This sounds like a timing issue that we fixed a while back
>>>>>>>>>>>>>>>>>>>>         Was a version 1.1.11 from December 3.
>>>>>>>>>>>>>>>>>>>>         Now try full update and retest.
>>>>>>>>>>>>>>>>>>>        That should be recent enough.  Can you create a crm_report the next time you reproduce?
>>>>>>>>>>>>>>>>>>        Of course yes. Little delay.... :)
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>        ......
>>>>>>>>>>>>>>>>>>        cc1: warnings being treated as errors
>>>>>>>>>>>>>>>>>>        upstart.c: In function ‘upstart_job_property’:
>>>>>>>>>>>>>>>>>>        upstart.c:264: error: implicit declaration of function ‘g_variant_lookup_value’
>>>>>>>>>>>>>>>>>>        upstart.c:264: error: nested extern declaration of ‘g_variant_lookup_value’
>>>>>>>>>>>>>>>>>>        upstart.c:264: error: assignment makes pointer from integer without a cast
>>>>>>>>>>>>>>>>>>        gmake[2]: *** [libcrmservice_la-upstart.lo] Error 1
>>>>>>>>>>>>>>>>>>        gmake[2]: Leaving directory `/root/ha/pacemaker/lib/services'
>>>>>>>>>>>>>>>>>>        make[1]: *** [all-recursive] Error 1
>>>>>>>>>>>>>>>>>>        make[1]: Leaving directory `/root/ha/pacemaker/lib'
>>>>>>>>>>>>>>>>>>        make: *** [core] Error 1
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>        I'm trying to solve this a problem.
>>>>>>>>>>>>>>>>>        Do not get solved quickly...
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>        https://developer.gnome.org/glib/2.28/glib-GVariant.html#g-variant-lookup-value
>>>>>>>>>>>>>>>>>        g_variant_lookup_value () Since 2.28
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>        # yum list installed glib2
>>>>>>>>>>>>>>>>>        Loaded plugins: fastestmirror, rhnplugin, security
>>>>>>>>>>>>>>>>>        This system is receiving updates from RHN Classic or Red Hat Satellite.
>>>>>>>>>>>>>>>>>        Loading mirror speeds from cached hostfile
>>>>>>>>>>>>>>>>>        Installed Packages
>>>>>>>>>>>>>>>>>        glib2.x86_64                                                              2.26.1-3.el6                                                               installed
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>        # cat /etc/issue
>>>>>>>>>>>>>>>>>        CentOS release 6.5 (Final)
>>>>>>>>>>>>>>>>>        Kernel \r on an \m
>>>>>>>>>>>>>>>>        Can you try this patch?
>>>>>>>>>>>>>>>>        Upstart jobs wont work, but the code will compile
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>        diff --git a/lib/services/upstart.c b/lib/services/upstart.c
>>>>>>>>>>>>>>>>        index 831e7cf..195c3a4 100644
>>>>>>>>>>>>>>>>        --- a/lib/services/upstart.c
>>>>>>>>>>>>>>>>        +++ b/lib/services/upstart.c
>>>>>>>>>>>>>>>>        @@ -231,12 +231,21 @@ upstart_job_exists(const char *name)
>>>>>>>>>>>>>>>>        static char *
>>>>>>>>>>>>>>>>        upstart_job_property(const char *obj, const gchar * iface, const char *name)
>>>>>>>>>>>>>>>>        {
>>>>>>>>>>>>>>>>        +    char *output = NULL;
>>>>>>>>>>>>>>>>        +
>>>>>>>>>>>>>>>>        +#if !GLIB_CHECK_VERSION(2,28,0)
>>>>>>>>>>>>>>>>        +    static bool err = TRUE;
>>>>>>>>>>>>>>>>        +
>>>>>>>>>>>>>>>>        +    if(err) {
>>>>>>>>>>>>>>>>        +        crm_err("This version of glib is too old to support upstart jobs");
>>>>>>>>>>>>>>>>        +        err = FALSE;
>>>>>>>>>>>>>>>>        +    }
>>>>>>>>>>>>>>>>        +#else
>>>>>>>>>>>>>>>>           GError *error = NULL;
>>>>>>>>>>>>>>>>           GDBusProxy *proxy;
>>>>>>>>>>>>>>>>           GVariant *asv = NULL;
>>>>>>>>>>>>>>>>           GVariant *value = NULL;
>>>>>>>>>>>>>>>>           GVariant *_ret = NULL;
>>>>>>>>>>>>>>>>        -    char *output = NULL;
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>           crm_info("Calling GetAll on %s", obj);
>>>>>>>>>>>>>>>>           proxy = get_proxy(obj, BUS_PROPERTY_IFACE);
>>>>>>>>>>>>>>>>        @@ -272,6 +281,7 @@ upstart_job_property(const char *obj, const gchar * iface, const char *name)
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>           g_object_unref(proxy);
>>>>>>>>>>>>>>>>           g_variant_unref(_ret);
>>>>>>>>>>>>>>>>        +#endif
>>>>>>>>>>>>>>>>           return output;
>>>>>>>>>>>>>>>>        }
>>>>>>>>>>>>>>>        Ok :) I patch source.
>>>>>>>>>>>>>>>        Type "make rc" - the same error.
>>>>>>>>>>>>>>        Because its not building your local changes
>>>>>>>>>>>>>>>        Make new copy via "fetch" - the same error.
>>>>>>>>>>>>>>>        It seems that if not exist ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz, then download it.
>>>>>>>>>>>>>>>        Otherwise use exist archive.
>>>>>>>>>>>>>>>        Cutted log .......
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>        # make rc
>>>>>>>>>>>>>>>        make TAG=Pacemaker-1.1.11-rc3 rpm
>>>>>>>>>>>>>>>        make[1]: Entering directory `/root/ha/pacemaker'
>>>>>>>>>>>>>>>        rm -f pacemaker-dirty.tar.* pacemaker-tip.tar.* pacemaker-HEAD.tar.*
>>>>>>>>>>>>>>>        if [ ! -f ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz ]; then                                             \
>>>>>>>>>>>>>>>                 rm -f pacemaker.tar.*;                                              \
>>>>>>>>>>>>>>>                 if [ Pacemaker-1.1.11-rc3 = dirty ]; then                                   \
>>>>>>>>>>>>>>>                     git commit -m "DO-NOT-PUSH" -a;                                 \
>>>>>>>>>>>>>>>                     git archive --prefix=ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3/ HEAD | gzip > ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz;       \
>>>>>>>>>>>>>>>                     git reset --mixed HEAD^;                                        \
>>>>>>>>>>>>>>>                 else                                                                \
>>>>>>>>>>>>>>>                     git archive --prefix=ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3/ Pacemaker-1.1.11-rc3 | gzip > ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz;     \
>>>>>>>>>>>>>>>                 fi;                                                                 \
>>>>>>>>>>>>>>>                 echo `date`: Rebuilt ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz;                                     \
>>>>>>>>>>>>>>>             else                                                                    \
>>>>>>>>>>>>>>>                 echo `date`: Using existing tarball: ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz;                     \
>>>>>>>>>>>>>>>             fi
>>>>>>>>>>>>>>>        Mon Jan 13 13:23:21 MSK 2014: Using existing tarball: ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz
>>>>>>>>>>>>>>>        .......
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>        Well, "make rpm" - build rpms and I create cluster.
>>>>>>>>>>>>>>>        I spent the same tests and confirmed the behavior.
>>>>>>>>>>>>>>>        crm_reoprt log here - http://send2me.ru/crmrep.tar.bz2
>>>>>>>>>>>>>>        Thanks!
>>>>>>>>>>>>       ,
>>>>>>>>>>>>       _______________________________________________
>>>>>>>>>>>>       Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>>>>>>>       http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>>>>> 
>>>>>>>>>>>>       Project Home: http://www.clusterlabs.org
>>>>>>>>>>>>       Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>>>>>       Bugs: http://bugs.clusterlabs.org
>>>>>>>>>>>      _______________________________________________
>>>>>>>>>>>      Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>>>>>>      http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>>>> 
>>>>>>>>>>>      Project Home: http://www.clusterlabs.org
>>>>>>>>>>>      Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>>>>      Bugs: http://bugs.clusterlabs.org
>>>>>>>>>>      _______________________________________________
>>>>>>>>>>      Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>>>>>      http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>>> 
>>>>>>>>>>      Project Home: http://www.clusterlabs.org
>>>>>>>>>>      Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>>>      Bugs: http://bugs.clusterlabs.org
>>>>>>>>>     ,
>>>>>>>>>     _______________________________________________
>>>>>>>>>     Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>>>>     http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>> 
>>>>>>>>>     Project Home: http://www.clusterlabs.org
>>>>>>>>>     Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>>     Bugs: http://bugs.clusterlabs.org
>>>>>>>>     _______________________________________________
>>>>>>>>     Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>>>     http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>> 
>>>>>>>>     Project Home: http://www.clusterlabs.org
>>>>>>>>     Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>     Bugs: http://bugs.clusterlabs.org
>>>>>>>    ,
>>>>>>>    _______________________________________________
>>>>>>>    Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>>    http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>> 
>>>>>>>    Project Home: http://www.clusterlabs.org
>>>>>>>    Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>    Bugs: http://bugs.clusterlabs.org
>>>>>>    _______________________________________________
>>>>>>    Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>    http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>> 
>>>>>>    Project Home: http://www.clusterlabs.org
>>>>>>    Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>    Bugs: http://bugs.clusterlabs.org
>>>>>   ,
>>>>>   _______________________________________________
>>>>>   Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>   http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>> 
>>>>>   Project Home: http://www.clusterlabs.org
>>>>>   Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>   Bugs: http://bugs.clusterlabs.org
>>>>   _______________________________________________
>>>>   Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>   http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>> 
>>>>   Project Home: http://www.clusterlabs.org
>>>>   Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>   Bugs: http://bugs.clusterlabs.org
>>>  ,
>>>  _______________________________________________
>>>  Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>  http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>> 
>>>  Project Home: http://www.clusterlabs.org
>>>  Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>  Bugs: http://bugs.clusterlabs.org
>> 
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> 
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org

-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 841 bytes
Desc: Message signed with OpenPGP using GPGMail
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20140219/66480f34/attachment-0004.sig>


More information about the Pacemaker mailing list