[Pacemaker] Announce: Pacemaker 1.1.10 now available

Andrew Beekhof andrew at beekhof.net
Thu Aug 1 01:47:17 EDT 2013


On 01/08/2013, at 1:56 PM, Takatoshi MATSUO <matsuo.tak at gmail.com> wrote:

> Hi Andrew
> 
> This patch works fine.

Excellent. Thankyou for confirming!

Pushed:
	https://github.com/beekhof/pacemaker/commit/5896233

> 
> 2013/8/1 Andrew Beekhof <andrew at beekhof.net>:
>> 
>> On 01/08/2013, at 10:18 AM, Takatoshi MATSUO <matsuo.tak at gmail.com> wrote:
>> 
>>> Hi Andrew
>>> 
>>> I'm about to collect logs of crm_report,
>>> but Rainer already provides it.
>>> 
>>> Could you see his reports ?
>> 
>> I had just written:
>> 
>> "I can but they're insufficiently helpful."
>> 
>> when a thought struck me....
>> 
>> Can you try the following patch?
>> It would explain why I couldn't reproduce it locally earlier today.
>> 
>> diff --git a/crmd/lrm.c b/crmd/lrm.c
>> index d6b0dd0..4bce39a 100644
>> --- a/crmd/lrm.c
>> +++ b/crmd/lrm.c
>> @@ -1744,7 +1744,9 @@ do_lrm_rsc_op(lrm_state_t * lrm_state, lrmd_rsc_info_t * rsc, const char *operat
>>     CRM_CHECK(op != NULL, return);
>> 
>>     /* stop any previous monitor operations before changing the resource state */
>> -    if (op->interval == 0 && strcmp(operation, CRMD_ACTION_STATUS) != 0) {
>> +    if (op->interval == 0
>> +        && strcmp(operation, CRMD_ACTION_STATUS) != 0
>> +        && strcmp(operation, CRMD_ACTION_NOTIFY) != 0) {
>>         guint removed = 0;
>>         struct stop_recurring_action_s data;
>> 
>> 
>> 
>>> 
>>> Thanks,
>>> Takatoshi MATSUO
>>> 
>>> 
>>> 2013/8/1 Rainer Brestan <rainer.brestan at gmx.net>:
>>>> Base situation for the logs:
>>>> Pacemaker stop on int2node1 and int2node2
>>>> Master/slave resource msABC already configured.
>>>> Included in the crm_report is also per node a file "a", this is the one,
>>>> which the modified Stateful RA writes to log each action performed.
>>>> 
>>>> 1.) 19:22:25 start Pacemaker on int2node1
>>>> https://www.dropbox.com/s/ftbdl71ol2iyi42/step1.log.tar.bz2
>>>> monitor on master is called
>>>> 
>>>> 2.) 19:32:14 start Pacemaker on int2node2
>>>> https://www.dropbox.com/s/s3jnxqvod9mlyz1/step2.log.tar.bz2
>>>> monitor on master is not called any more
>>>> 
>>>> 3.) 19:37:14 stop Pacemaker on int2node2
>>>> https://www.dropbox.com/s/w75myab6fxh7mak/step3.log.tar.bz2
>>>> monitor on master is still not called any more
>>>> 
>>>> 4.) 19:42:14 start Pacemaker on in2node2
>>>> https://www.dropbox.com/s/p00wl9kx4vwhilh/step4.log.tar.bz2
>>>> monitor on master is called normally
>>>> 
>>>> Hope this gives a clearer picture which component has forgotten the monitor
>>>> action.
>>>> 
>>>> Rainer
>>>> Gesendet: Mittwoch, 31. Juli 2013 um 14:19 Uhr
>>>> 
>>>> Von: "Andrew Beekhof" <andrew at beekhof.net>
>>>> An: "The Pacemaker cluster resource manager" <pacemaker at oss.clusterlabs.org>
>>>> Betreff: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available
>>>> 
>>>> On 31/07/2013, at 5:17 PM, Rainer Brestan <rainer.brestan at gmx.net> wrote:
>>>> 
>>>>> Modified the RA to log each action call performed and from this log there
>>>>> is no call of monitor action.
>>>>> 
>>>>> From the logs i do not think it is the policy engine, it might be the LRM
>>>>> part of crmd (the is the only relevant change be seen after git diff between
>>>>> 1.1.10-rc7 and 1.1.10).
>>>> 
>>>> Ok. Can you still send me a crm_report though?
>>>> Even if the PE isn't at fault, it shows me what the cib looked like at the
>>>> time which can be surprisingly helpful.
>>>> And it would have all the logs...
>>>> 
>>>>> 
>>>>> Explanation of the below log:
>>>>> primitive resABC ocf:heartbeat:Stateful \
>>>>> op start interval="0s" timeout="60s" on-fail="restart" \
>>>>> op monitor interval="30s" timeout="60s" on-fail="restart" \
>>>>> op promote interval="0s" timeout="60s" on-fail="restart" \
>>>>> op demote interval="0" timeout="60s" on-fail="restart" \
>>>>> op stop interval="0" timeout="60s" on-fail="restart" \
>>>>> op monitor interval="20" role="Master" timeout="60"
>>>>> ms msABC resABC \
>>>>> meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1"
>>>>> notify="true"
>>>>> crm_mon at begin of log:
>>>>> Last updated: Wed Jul 31 08:30:57 2013
>>>>> Last change: Tue Jul 30 13:01:36 2013 via crmd on int2node1
>>>>> Stack: corosync
>>>>> Current DC: int2node1 (1743917066) - partition with quorum
>>>>> Version: 1.1.10-1.el6-368c726
>>>>> 2 Nodes configured
>>>>> 5 Resources configured
>>>>> Online: [ int2node1 int2node2 ]
>>>>> Master/Slave Set: msABC [resABC]
>>>>> Masters: [ int2node1 ]
>>>>> Slaves: [ int2node2 ]
>>>>> crm_mon at end of log:
>>>>> Last updated: Wed Jul 31 08:55:29 2013
>>>>> Last change: Tue Jul 30 13:01:36 2013 via crmd on int2node1
>>>>> Stack: corosync
>>>>> Current DC: int2node1 (1743917066) - partition with quorum
>>>>> Version: 1.1.10-1.el6-368c726
>>>>> 2 Nodes configured
>>>>> 5 Resources configured
>>>>> Online: [ int2node1 ]
>>>>> OFFLINE: [ int2node2 ]
>>>>> Master/Slave Set: msABC [resABC]
>>>>> Masters: [ int2node1 ]
>>>>> 
>>>>> int2node1 is running, int2node2 is started
>>>>> 2013-07-31T08:30:52.631+02:00 int2node1 pengine[16443] notice: notice:
>>>>> LogActions: Start resABC:1 (int2node2)
>>>>> 2013-07-31T08:30:52.638+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 9: monitor resABC:1_monitor_0 on int2node2
>>>>> 2013-07-31T08:30:52.638+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 54: notify resABC_pre_notify_start_0 on
>>>>> int2node1 (local)
>>>>> 2013-07-31T08:30:52.681+02:00 int2node1 crmd[16444] notice: notice:
>>>>> process_lrm_event: LRM operation resABC_notify_0 (call=64, rc=0,
>>>>> cib-update=0, confirmed=true) ok
>>>>> 2013-07-31T08:30:52.780+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 25: start resABC:1_start_0 on int2node2
>>>>> 2013-07-31T08:30:52.940+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 55: notify resABC_post_notify_start_0 on
>>>>> int2node1 (local)
>>>>> 2013-07-31T08:30:52.943+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 56: notify resABC:1_post_notify_start_0 on
>>>>> int2node2
>>>>> 2013-07-31T08:30:52.982+02:00 int2node1 crmd[16444] notice: notice:
>>>>> process_lrm_event: LRM operation resABC_notify_0 (call=67, rc=0,
>>>>> cib-update=0, confirmed=true) ok
>>>>> 2013-07-31T08:30:52.992+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 24: monitor resABC_monitor_20000 on
>>>>> int2node1 (local)
>>>>> 2013-07-31T08:30:52.996+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 26: monitor resABC:1_monitor_30000 on
>>>>> int2node2
>>>>> 2013-07-31T08:30:53.035+02:00 int2node1 crmd[16444] notice: notice:
>>>>> process_lrm_event: LRM operation resABC_monitor_20000 (call=70, rc=8,
>>>>> cib-update=149, confirmed=false) master
>>>>> 
>>>>> At this point int2node2 is stopped.
>>>>> 2013-07-31T08:37:51.457+02:00 int2node1 crmd[16444] notice: notice:
>>>>> do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [
>>>>> input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
>>>>> 2013-07-31T08:37:51.462+02:00 int2node1 pengine[16443] notice: notice:
>>>>> unpack_config: On loss of CCM Quorum: Ignore
>>>>> 2013-07-31T08:37:51.465+02:00 int2node1 pengine[16443] notice: notice:
>>>>> stage6: Scheduling Node int2node2 for shutdown
>>>>> 2013-07-31T08:37:51.466+02:00 int2node1 pengine[16443] notice: notice:
>>>>> LogActions: Stop resABC:1 (int2node2)
>>>>> 2013-07-31T08:37:51.469+02:00 int2node1 pengine[16443] notice: notice:
>>>>> process_pe_message: Calculated Transition 86:
>>>>> /var/lib/pacemaker/pengine/pe-input-125.bz2
>>>>> 2013-07-31T08:37:51.471+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 56: notify resABC_pre_notify_stop_0 on
>>>>> int2node1 (local)
>>>>> 2013-07-31T08:37:51.474+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 58: notify resABC_pre_notify_stop_0 on
>>>>> int2node2
>>>>> 2013-07-31T08:37:51.512+02:00 int2node1 crmd[16444] notice: notice:
>>>>> process_lrm_event: LRM operation resABC_notify_0 (call=74, rc=0,
>>>>> cib-update=0, confirmed=true) ok
>>>>> 2013-07-31T08:37:51.514+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 23: stop resABC_stop_0 on int2node2
>>>>> 2013-07-31T08:37:51.654+02:00 int2node1 crmd[16444] notice: notice:
>>>>> te_rsc_command: Initiating action 57: notify resABC_post_notify_stop_0 on
>>>>> int2node1 (local)
>>>>> 2013-07-31T08:37:51.699+02:00 int2node1 crmd[16444] notice: notice:
>>>>> process_lrm_event: LRM operation resABC_notify_0 (call=78, rc=0,
>>>>> cib-update=0, confirmed=true) ok
>>>>> 2013-07-31T08:37:51.699+02:00 int2node1 crmd[16444] notice: notice:
>>>>> run_graph: Transition 86 (Complete=13, Pending=0, Fired=0, Skipped=2,
>>>>> Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-125.bz2): Stopped
>>>>> 2013-07-31T08:37:51.705+02:00 int2node1 pengine[16443] notice: notice:
>>>>> unpack_config: On loss of CCM Quorum: Ignore
>>>>> 2013-07-31T08:37:51.705+02:00 int2node1 pengine[16443] notice: notice:
>>>>> stage6: Scheduling Node int2node2 for shutdown
>>>>> 2013-07-31T08:37:51.706+02:00 int2node1 pengine[16443] notice: notice:
>>>>> process_pe_message: Calculated Transition 87:
>>>>> /var/lib/pacemaker/pengine/pe-input-126.bz2
>>>>> 2013-07-31T08:37:51.707+02:00 int2node1 crmd[16444] notice: notice:
>>>>> run_graph: Transition 87 (Complete=1, Pending=0, Fired=0, Skipped=0,
>>>>> Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-126.bz2): Complete
>>>>> 2013-07-31T08:37:51.707+02:00 int2node1 crmd[16444] notice: notice:
>>>>> do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [
>>>>> input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
>>>>> 2013-07-31T08:37:51.720+02:00 int2node1 crmd[16444] notice: notice:
>>>>> peer_update_callback: do_shutdown of int2node2 (op 45) is complete
>>>>> 
>>>>> Output from RA on int2node1:
>>>>> Wed Jul 31 08:30:52 CEST 2013 resABC: operation notify, type pre,
>>>>> operation start
>>>>> Wed Jul 31 08:30:52 CEST 2013 resABC: operation notify, type post,
>>>>> operation start
>>>>> Wed Jul 31 08:30:53 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:31:13 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:31:33 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:31:53 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:32:13 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:32:33 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:32:53 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:33:13 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:33:33 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:33:53 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:34:13 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:34:33 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:34:53 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:35:13 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:35:33 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:35:53 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:36:13 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:36:33 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:36:53 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:37:13 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:37:33 CEST 2013 resABC: operation monitor, type , operation
>>>>> Wed Jul 31 08:37:51 CEST 2013 resABC: operation notify, type pre,
>>>>> operation stop
>>>>> Wed Jul 31 08:37:51 CEST 2013 resABC: operation notify, type post,
>>>>> operation stop
>>>>> 
>>>>> After 08:37:51 no log output from Pacemaker for resABC, nor any output
>>>>> from RA on int2node1.
>>>>> 
>>>>> Gesendet: Mittwoch, 31. Juli 2013 um 02:10 Uhr
>>>>> Von: "Andrew Beekhof" <andrew at beekhof.net>
>>>>> An: "The Pacemaker cluster resource manager"
>>>>> <pacemaker at oss.clusterlabs.org>
>>>>> Betreff: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available
>>>>> 
>>>>> On 30/07/2013, at 9:13 PM, Rainer Brestan <rainer.brestan at gmx.net> wrote:
>>>>> 
>>>>>> I can agree, Master monitor operation is broken in 1.1.10 release.
>>>>>> When the slave monitor action is started, the master monitor action is
>>>>>> not called any more.
>>>>> 
>>>>> Based on?
>>>>> 
>>>>>> 
>>>>>> I have created a setup with Stateful resource with two nodes.
>>>>>> Then the Pacemaker installation is changed to different versions without
>>>>>> changing the configuration part of the CIB.
>>>>>> 
>>>>>> Result:
>>>>>> 1.1.10-rc5, 1.1.10-rc6 and 1.1.10-rc7 does not have this error
>>>>>> 1.1.10-1 release has the error
>>>>>> 
>>>>>> Installation order (just that anybody know how it was done):
>>>>>> 1.1.10-1 -> error
>>>>>> 1.1.10-rc5 -> no error
>>>>>> 1.1.10-rc6 -> no error
>>>>>> 1.1.10-rc7 -> no error
>>>>>> 1.1.10-1 -> error
>>>>>> 
>>>>>> Rainer
>>>>>> Gesendet: Freitag, 26. Juli 2013 um 09:32 Uhr
>>>>>> Von: "Takatoshi MATSUO" <matsuo.tak at gmail.com>
>>>>>> An: "The Pacemaker cluster resource manager"
>>>>>> <pacemaker at oss.clusterlabs.org>
>>>>>> Betreff: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available
>>>>>> Hi
>>>>>> 
>>>>>> I used Stateful RA and caught a same issue.
>>>>>> 
>>>>>> 1. before starting slave
>>>>>> 
>>>>>> # crm_simulate -VVV -S -x /var/lib/pacemaker/pengine/pe-input-1543.bz2
>>>>>> | grep "Resource action"
>>>>>> * Resource action: stateful monitor=2000 on 16-sl6
>>>>>> 
>>>>>> 2. starting slave
>>>>>> # crm_simulate -VVV -S -x /var/lib/pacemaker/pengine/pe-input-1544.bz2
>>>>>> | grep "Resource action"
>>>>>> * Resource action: stateful monitor on 17-sl6
>>>>>> * Resource action: stateful notify on 16-sl6
>>>>>> * Resource action: stateful start on 17-sl6
>>>>>> * Resource action: stateful notify on 16-sl6
>>>>>> * Resource action: stateful notify on 17-sl6
>>>>>> * Resource action: stateful monitor=3000 on 17-sl6
>>>>>> 
>>>>>> 3. after
>>>>>> # crm_simulate -VVV -S -x /var/lib/pacemaker/pengine/pe-input-1545.bz2
>>>>>> | grep "Resource action"
>>>>>> * Resource action: stateful monitor=3000 on 17-sl6
>>>>>> 
>>>>>> Monitor=2000 is deleted.
>>>>>> Is this correct ?
>>>>>> 
>>>>>> 
>>>>>> My setting
>>>>>> --------
>>>>>> property \
>>>>>> no-quorum-policy="ignore" \
>>>>>> stonith-enabled="false"
>>>>>> 
>>>>>> rsc_defaults \
>>>>>> resource-stickiness="INFINITY" \
>>>>>> migration-threshold="1"
>>>>>> 
>>>>>> ms msStateful stateful \
>>>>>> meta \
>>>>>> master-max="1" \
>>>>>> master-node-max="1" \
>>>>>> clone-max="2" \
>>>>>> clone-node-max="1" \
>>>>>> notify="true"
>>>>>> 
>>>>>> primitive stateful ocf:heartbeat:Stateful \
>>>>>> op start timeout="60s" interval="0s" on-fail="restart" \
>>>>>> op monitor timeout="60s" interval="3s" on-fail="restart" \
>>>>>> op monitor timeout="60s" interval="2s" on-fail="restart" role="Master" \
>>>>>> op promote timeout="60s" interval="0s" on-fail="restart" \
>>>>>> op demote timeout="60s" interval="0s" on-fail="stop" \
>>>>>> op stop timeout="60s" interval="0s" on-fail="block"
>>>>>> --------
>>>>>> 
>>>>>> Regards,
>>>>>> Takatoshi MATSUO
>>>>>> 
>>>>>> 2013/7/26 Takatoshi MATSUO <matsuo.tak at gmail.com>:
>>>>>>> Hi
>>>>>>> 
>>>>>>> My report is late for 1.1.10 :(
>>>>>>> 
>>>>>>> I am using pacemaker 1.1.10-0.1.ab2e209.git.
>>>>>>> It seems that master's monitor is stopped when slave is started.
>>>>>>> 
>>>>>>> Does someone encounter same problem ?
>>>>>>> I attach a log and settings.
>>>>>>> 
>>>>>>> 
>>>>>>> Thanks,
>>>>>>> Takatoshi MATSUO
>>>>>>> 
>>>>>>> 2013/7/26 Digimer <lists at alteeve.ca>:
>>>>>>>> Congrats!! I know this was a long time in the making.
>>>>>>>> 
>>>>>>>> digimer
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On 25/07/13 20:43, Andrew Beekhof wrote:
>>>>>>>>> 
>>>>>>>>> Announcing the release of Pacemaker 1.1.10
>>>>>>>>> 
>>>>>>>>> https://github.com/ClusterLabs/pacemaker/releases/Pacemaker-1.1.10
>>>>>>>>> 
>>>>>>>>> There were three changes of note since rc7:
>>>>>>>>> 
>>>>>>>>> + Bug cl#5161 - crmd: Prevent memory leak in operation cache
>>>>>>>>> + cib: Correctly read back archived configurations if the primary is
>>>>>>>>> corrupted
>>>>>>>>> + cman: Do not pretend we know the state of nodes we've never seen
>>>>>>>>> 
>>>>>>>>> Along with assorted bug fixes, the major topics for this release
>>>>>>>>> were:
>>>>>>>>> 
>>>>>>>>> - stonithd fixes
>>>>>>>>> - fixing memory leaks, often caused by incorrect use of glib
>>>>>>>>> reference
>>>>>>>>> counting
>>>>>>>>> - supportability improvements (code cleanup and deduplication,
>>>>>>>>> standardized error codes)
>>>>>>>>> 
>>>>>>>>> Release candidates for the next Pacemaker release (1.1.11) can be
>>>>>>>>> expected some time around Novemeber.
>>>>>>>>> 
>>>>>>>>> A big thankyou to everyone that spent time testing the release
>>>>>>>>> candidates and/or contributed patches. However now that Pacemaker is
>>>>>>>>> perfect, anyone reporting bugs will be shot :-)
>>>>>>>>> 
>>>>>>>>> To build `rpm` packages:
>>>>>>>>> 
>>>>>>>>> 1. Clone the current sources:
>>>>>>>>> 
>>>>>>>>> # git clone --depth 0 git://github.com/ClusterLabs/pacemaker.git
>>>>>>>>> # cd pacemaker
>>>>>>>>> 
>>>>>>>>> 1. Install dependancies (if you haven't already)
>>>>>>>>> 
>>>>>>>>> [Fedora] # sudo yum install -y yum-utils
>>>>>>>>> [ALL] # make rpm-dep
>>>>>>>>> 
>>>>>>>>> 1. Build Pacemaker
>>>>>>>>> 
>>>>>>>>> # make release
>>>>>>>>> 
>>>>>>>>> 1. Copy and deploy as needed
>>>>>>>>> 
>>>>>>>>> ## Details - 1.1.10 - final
>>>>>>>>> 
>>>>>>>>> Changesets: 602
>>>>>>>>> Diff: 143 files changed, 8162 insertions(+), 5159 deletions(-)
>>>>>>>>> 
>>>>>>>>> ## Highlights
>>>>>>>>> 
>>>>>>>>> ### Features added since Pacemaker-1.1.9
>>>>>>>>> 
>>>>>>>>> + Core: Convert all exit codes to positive errno values
>>>>>>>>> + crm_error: Add the ability to list and print error symbols
>>>>>>>>> + crm_resource: Allow individual resources to be reprobed
>>>>>>>>> + crm_resource: Allow options to be set recursively
>>>>>>>>> + crm_resource: Implement --ban for moving resources away from nodes
>>>>>>>>> and --clear (replaces --unmove)
>>>>>>>>> + crm_resource: Support OCF tracing when using
>>>>>>>>> --force-(check|start|stop)
>>>>>>>>> + PE: Allow active nodes in our current membership to be fenced
>>>>>>>>> without
>>>>>>>>> quorum
>>>>>>>>> + PE: Suppress meaningless IDs when displaying anonymous clone
>>>>>>>>> status
>>>>>>>>> + Turn off auto-respawning of systemd services when the cluster
>>>>>>>>> starts
>>>>>>>>> them
>>>>>>>>> + Bug cl#5128 - pengine: Support maintenance mode for a single node
>>>>>>>>> 
>>>>>>>>> ### Changes since Pacemaker-1.1.9
>>>>>>>>> 
>>>>>>>>> + crmd: cib: stonithd: Memory leaks resolved and improved use of
>>>>>>>>> glib
>>>>>>>>> reference counting
>>>>>>>>> + attrd: Fixes deleted attributes during dc election
>>>>>>>>> + Bug cf#5153 - Correctly display clone failcounts in crm_mon
>>>>>>>>> + Bug cl#5133 - pengine: Correctly observe on-fail=block for failed
>>>>>>>>> demote operation
>>>>>>>>> + Bug cl#5148 - legacy: Correctly remove a node that used to have a
>>>>>>>>> different nodeid
>>>>>>>>> + Bug cl#5151 - Ensure node names are consistently compared without
>>>>>>>>> case
>>>>>>>>> + Bug cl#5152 - crmd: Correctly clean up fenced nodes during
>>>>>>>>> membership
>>>>>>>>> changes
>>>>>>>>> + Bug cl#5154 - Do not expire failures when on-fail=block is present
>>>>>>>>> + Bug cl#5155 - pengine: Block the stop of resources if any
>>>>>>>>> depending
>>>>>>>>> resource is unmanaged
>>>>>>>>> + Bug cl#5157 - Allow migration in the absence of some colocation
>>>>>>>>> constraints
>>>>>>>>> + Bug cl#5161 - crmd: Prevent memory leak in operation cache
>>>>>>>>> + Bug cl#5164 - crmd: Fixes crash when using pacemaker-remote
>>>>>>>>> + Bug cl#5164 - pengine: Fixes segfault when calculating transition
>>>>>>>>> with remote-nodes.
>>>>>>>>> + Bug cl#5167 - crm_mon: Only print "stopped" node list for
>>>>>>>>> incomplete
>>>>>>>>> clone sets
>>>>>>>>> + Bug cl#5168 - Prevent clones from being bounced around the cluster
>>>>>>>>> due to location constraints
>>>>>>>>> + Bug cl#5170 - Correctly support on-fail=block for clones
>>>>>>>>> + cib: Correctly read back archived configurations if the primary is
>>>>>>>>> corrupted
>>>>>>>>> + cib: The result is not valid when diffs fail to apply cleanly for
>>>>>>>>> CLI
>>>>>>>>> tools
>>>>>>>>> + cib: Restore the ability to embed comments in the configuration
>>>>>>>>> + cluster: Detect and warn about node names with capitals
>>>>>>>>> + cman: Do not pretend we know the state of nodes we've never seen
>>>>>>>>> + cman: Do not unconditionally start cman if it is already running
>>>>>>>>> + cman: Support non-blocking CPG calls
>>>>>>>>> + Core: Ensure the blackbox is saved on abnormal program termination
>>>>>>>>> + corosync: Detect the loss of members for which we only know the
>>>>>>>>> nodeid
>>>>>>>>> + corosync: Do not pretend we know the state of nodes we've never
>>>>>>>>> seen
>>>>>>>>> + corosync: Ensure removed peers are erased from all caches
>>>>>>>>> + corosync: Nodes that can persist in sending CPG messages must be
>>>>>>>>> alive afterall
>>>>>>>>> + crmd: Do not get stuck in S_POLICY_ENGINE if a node we couldn't
>>>>>>>>> fence
>>>>>>>>> returns
>>>>>>>>> + crmd: Do not update fail-count and last-failure for old failures
>>>>>>>>> + crmd: Ensure all membership operations can complete while trying
>>>>>>>>> to
>>>>>>>>> cancel a transition
>>>>>>>>> + crmd: Ensure operations for cleaned up resources don't block
>>>>>>>>> recovery
>>>>>>>>> + crmd: Ensure we return to a stable state if there have been too
>>>>>>>>> many
>>>>>>>>> fencing failures
>>>>>>>>> + crmd: Initiate node shutdown if another node claims to have
>>>>>>>>> successfully fenced us
>>>>>>>>> + crmd: Prevent messages for remote crmd clients from being relayed
>>>>>>>>> to
>>>>>>>>> wrong daemons
>>>>>>>>> + crmd: Properly handle recurring monitor operations for remote-node
>>>>>>>>> agent
>>>>>>>>> + crmd: Store last-run and last-rc-change for all operations
>>>>>>>>> + crm_mon: Ensure stale pid files are updated when a new process is
>>>>>>>>> started
>>>>>>>>> + crm_report: Correctly collect logs when 'uname -n' reports fully
>>>>>>>>> qualified names
>>>>>>>>> + fencing: Fail the operation once all peers have been exhausted
>>>>>>>>> + fencing: Restore the ability to manually confirm that fencing
>>>>>>>>> completed
>>>>>>>>> + ipc: Allow unpriviliged clients to clean up after server failures
>>>>>>>>> + ipc: Restore the ability for members of the haclient group to
>>>>>>>>> connect
>>>>>>>>> to the cluster
>>>>>>>>> + legacy: Support "crm_node --remove" with a node name for corosync
>>>>>>>>> plugin (bnc#805278)
>>>>>>>>> + lrmd: Default to the upstream location for resource agent scratch
>>>>>>>>> directory
>>>>>>>>> + lrmd: Pass errors from lsb metadata generation back to the caller
>>>>>>>>> + pengine: Correctly handle resources that recover before we operate
>>>>>>>>> on
>>>>>>>>> them
>>>>>>>>> + pengine: Delete the old resource state on every node whenever the
>>>>>>>>> resource type is changed
>>>>>>>>> + pengine: Detect constraints with inappropriate actions (ie.
>>>>>>>>> promote
>>>>>>>>> for a clone)
>>>>>>>>> + pengine: Ensure per-node resource parameters are used during
>>>>>>>>> probes
>>>>>>>>> + pengine: If fencing is unavailable or disabled, block further
>>>>>>>>> recovery for resources that fail to stop
>>>>>>>>> + pengine: Implement the rest of get_timet_now() and rename to
>>>>>>>>> get_effective_time
>>>>>>>>> + pengine: Re-initiate _active_ recurring monitors that previously
>>>>>>>>> failed but have timed out
>>>>>>>>> + remote: Workaround for inconsistent tls handshake behavior between
>>>>>>>>> gnutls versions
>>>>>>>>> + systemd: Ensure we get shut down correctly by systemd
>>>>>>>>> + systemd: Reload systemd after adding/removing override files for
>>>>>>>>> cluster services
>>>>>>>>> + xml: Check for and replace non-printing characters with their
>>>>>>>>> octal
>>>>>>>>> equivalent while exporting xml text
>>>>>>>>> + xml: Prevent lockups by setting a more reliable buffer allocation
>>>>>>>>> strategy
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> _______________________________________________
>>>>>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>> 
>>>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>>>> Getting started:
>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> --
>>>>>>>> Digimer
>>>>>>>> Papers and Projects: https://alteeve.ca/w/
>>>>>>>> What if the cure for cancer is trapped in the mind of a person
>>>>>>>> without
>>>>>>>> access to education?
>>>>>>>> 
>>>>>>>> 
>>>>>>>> _______________________________________________
>>>>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>> 
>>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>>> Getting started:
>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>> 
>>>>>> _______________________________________________
>>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>> 
>>>>>> Project Home: http://www.clusterlabs.org
>>>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>> _______________________________________________
>>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>> 
>>>>>> Project Home: http://www.clusterlabs.org
>>>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>> 
>>>>> Project Home: http://www.clusterlabs.org
>>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>> Bugs: http://bugs.clusterlabs.org
>>>>> _______________________________________________
>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>> 
>>>>> Project Home: http://www.clusterlabs.org
>>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>> Bugs: http://bugs.clusterlabs.org
>>>> 
>>>> 
>>>> _______________________________________________
>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>> 
>>>> Project Home: http://www.clusterlabs.org
>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>> Bugs: http://bugs.clusterlabs.org
>>>> 
>>>> _______________________________________________
>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>> 
>>>> Project Home: http://www.clusterlabs.org
>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>> Bugs: http://bugs.clusterlabs.org
>>>> 
>>> 
>>> _______________________________________________
>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>> 
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>> 
>> 
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> 
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org





More information about the Pacemaker mailing list