[Pacemaker] Question regarding starting of master/slave resources and ELECTIONs

Andrew Beekhof andrew at beekhof.net
Fri Apr 15 11:28:52 UTC 2011


On Fri, Apr 15, 2011 at 5:58 AM, Bob Schatz <bschatz at yahoo.com> wrote:
> Andrew,
> Thanks for the help
> Comments inline with <BS>
> ________________________________
> From: Andrew Beekhof <andrew at beekhof.net>
> To: Bob Schatz <bschatz at yahoo.com>
> Cc: The Pacemaker cluster resource manager <pacemaker at oss.clusterlabs.org>
> Sent: Thu, April 14, 2011 2:14:40 AM
> Subject: Re: [Pacemaker] Question regarding starting of master/slave
> resources and ELECTIONs
>
> On Thu, Apr 14, 2011 at 10:49 AM, Andrew Beekhof <andrew at beekhof.net> wrote:
>
>>>> I noticed that 4 of the master/slave resources will start right away but
>>>> the
>>>> 5 master/slave resource seems to take a minute or so and I am only
>>>> running
>>>> with one node.
>>>> Is this expected?
>>>
>>> Probably, if the other 4 take around a minute each to start.
>>> There is an lrmd config variable that controls how much parallelism it
>>> allows (but i forget the name).
>>> <Bob> It's max-children and I set it to 40 for this test to see if it
>>> would
>>> change the behavior.  (/sbin/lrmadmin -p max-children 40)
>>
>> Thats surprising.  I'll have a look at the logs.
>
> Looking at the logs, I see a couple of things:
>
>
> This is very bad:
> Apr 12 19:33:42 mgraid-S000030311-1 crmd: [17529]: WARN: get_uuid:
> Could not calculate UUID for mgraid-s000030311-0
> Apr 12 19:33:42 mgraid-S000030311-1 crmd: [17529]: WARN:
> populate_cib_nodes_ha: Node mgraid-s000030311-0: no uuid found
>
> For some reason pacemaker cant get the node's uuid from heartbeat.
>
> <BS> I create the uuid when the node comes up.

Heartbeat should have already created it before pacemaker even got
started though.

>
> So we start a few things:
>
> Apr 12 19:33:41 mgraid-S000030311-1 crmd: [17529]: info:
> do_lrm_rsc_op: Performing
> key=23:3:0:48aac631-8177-4cda-94ea-48dfa9b1a90f
> op=SSS000030311:0_start_0 )
> Apr 12 19:33:41 mgraid-S000030311-1 crmd: [17529]: info:
> do_lrm_rsc_op: Performing
> key=49:3:0:48aac631-8177-4cda-94ea-48dfa9b1a90f
> op=SSJ000030312:0_start_0 )
> Apr 12 19:33:41 mgraid-S000030311-1 crmd: [17529]: info:
> do_lrm_rsc_op: Performing
> key=75:3:0:48aac631-8177-4cda-94ea-48dfa9b1a90f
> op=SSJ000030313:0_start_0 )
> Apr 12 19:33:41 mgraid-S000030311-1 crmd: [17529]: info:
> do_lrm_rsc_op: Performing
> key=101:3:0:48aac631-8177-4cda-94ea-48dfa9b1a90f
> op=SSJ000030314:0_start_0 )
>
> But then another change comes in:
>
> Apr 12 19:33:41 mgraid-S000030311-1 crmd: [17529]: info:
> abort_transition_graph: need_abort:59 - Triggered transition abort
> (complete=0) : Non-status change
>
> Normally we'd recompute and keep going, but it was a(nother) replace
> operation, so:
>
> Apr 12 19:33:42 mgraid-S000030311-1 crmd: [17529]: info:
> do_state_transition: State transition S_TRANSITION_ENGINE ->
> S_ELECTION [ input=I_ELECTION cause=C_FSA_INTERNAL
> origin=do_cib_replaced ]
>
> All the time goes here:
>
> Apr 12 19:35:31 mgraid-S000030311-1 crmd: [17529]: WARN:
> action_timer_callback: Timer popped (timeout=20000,
> abort_level=1000000, complete=true)
> Apr 12 19:35:31 mgraid-S000030311-1 crmd: [17529]: WARN:
> action_timer_callback: Ignoring timeout while not in transition
> Apr 12 19:35:31 mgraid-S000030311-1 crmd: [17529]: WARN:
> action_timer_callback: Timer popped (timeout=20000,
> abort_level=1000000, complete=true)
> Apr 12 19:35:31 mgraid-S000030311-1 crmd: [17529]: WARN:
> action_timer_callback: Ignoring timeout while not in transition
> Apr 12 19:35:31 mgraid-S000030311-1 crmd: [17529]: WARN:
> action_timer_callback: Timer popped (timeout=20000,
> abort_level=1000000, complete=true)
> Apr 12 19:35:31 mgraid-S000030311-1 crmd: [17529]: WARN:
> action_timer_callback: Ignoring timeout while not in transition
> Apr 12 19:35:31 mgraid-S000030311-1 crmd: [17529]: WARN:
> action_timer_callback: Timer popped (timeout=20000,
> abort_level=1000000, complete=true)
> Apr 12 19:35:31 mgraid-S000030311-1 crmd: [17529]: WARN:
> action_timer_callback: Ignoring timeout while not in transition
> Apr 12 19:37:00 mgraid-S000030311-1 crmd: [17529]: ERROR:
> crm_timer_popped: Integration Timer (I_INTEGRATED) just popped!
>
> but its not at all clear to me why - although certainly avoiding the
> election would help.
> Is there any chance to load all the changes at once?
>
> <BS> Yes.  That worked.  I created the configuration in a file and then did
> a "crm configure load update <filename>" to avoid the election
> Possibly the delay related to the UUID issue above, possibly it might
> be related to one of these two patches that went in after 1.0.9
>
> andrew (stable-1.0)    High: crmd: Make sure we always poke the FSA after
> a transition to clear any TE_HALT actions CS: 9187c0506fd3 On:
> 2010-07-07
> andrew (stable-1.0)    High: crmd: Reschedule the PE_START action if its
> not already running when we try to use it CS: e44dfe49e448 On:
> 2010-11-11
>
> Could you try turning on debug and/or a more recent version?
>
> <BS>  I turned on debug and grabbed the logs, configuration and and
> /var/lib/pengine directory.   They are attached.
>      Unfortunately I cannot try a new version with this hardware at this
> time. :(

Based on the new logs, it looks like the problem is resolved by
loading everything once.
All start actions appear to occur within a second of Apr 14 20:35:43.




More information about the Pacemaker mailing list