[ClusterLabs] Cluster failover failure with Unresolved dependency
Ken Gaillot
kgaillot at redhat.com
Mon Mar 14 13:44:29 EDT 2016
On 03/10/2016 09:49 AM, Lorand Kelemen wrote:
> Dear List,
>
> After the creation and testing of a simple 2 node active-passive
> drbd+postfix cluster nearly everything works flawlessly (standby, failure
> of a filesystem resource + failover, splitbrain + manual recovery) however
> when delibarately killing the postfix instance, after reaching the
> migration threshold failover does not occur and resources revert to the
> Stopped state (except the master-slave drbd resource, which works as
> expected).
>
> Ordering and colocation is configured, STONITH and quorum disabled, the
> goal is to always have one node running all the resources and at any sign
> of error it should fail over to the passive node, nothing fancy.
>
> Is my configuration wrong or am I hitting a bug?
>
> All software from centos 7 + elrepo repositories.
With these versions, you can set "two_node: 1" in
/etc/corosync/corosync.conf (which will be done automatically if you
used "pcs cluster setup" initially), and then you don't need to ignore
quorum in pacemaker.
> Regarding STONITH: the machines are running on free ESXi instances on
> separate machines, so the Vmware fencing agents won't work because in the
> free version the API is read-only.
> Still trying to figure out a way to go, until then manual recovery + huge
> arp cache times on the upstream firewall...
>
> Please find pe-input*.bz files attached, logs and config below. The
> situation: on node mail1 postfix was killed 3 times (migration threshold),
> it should have failed over to mail2.
> When killing a filesystem resource three times this happens flawlessly.
>
> Thanks for your input!
>
> Best regards,
> Lorand
>
>
> Cluster Name: mailcluster
> Corosync Nodes:
> mail1 mail2
> Pacemaker Nodes:
> mail1 mail2
>
> Resources:
> Group: network-services
> Resource: virtualip-1 (class=ocf provider=heartbeat type=IPaddr2)
> Attributes: ip=10.20.64.10 cidr_netmask=24 nic=lan0
> Operations: start interval=0s timeout=20s (virtualip-1-start-interval-0s)
> stop interval=0s timeout=20s (virtualip-1-stop-interval-0s)
> monitor interval=30s (virtualip-1-monitor-interval-30s)
> Master: spool-clone
> Meta Attrs: master-max=1 master-node-max=1 clone-max=2 clone-node-max=1
> notify=true
> Resource: spool (class=ocf provider=linbit type=drbd)
> Attributes: drbd_resource=spool
> Operations: start interval=0s timeout=240 (spool-start-interval-0s)
> promote interval=0s timeout=90 (spool-promote-interval-0s)
> demote interval=0s timeout=90 (spool-demote-interval-0s)
> stop interval=0s timeout=100 (spool-stop-interval-0s)
> monitor interval=10s (spool-monitor-interval-10s)
> Master: mail-clone
> Meta Attrs: master-max=1 master-node-max=1 clone-max=2 clone-node-max=1
> notify=true
> Resource: mail (class=ocf provider=linbit type=drbd)
> Attributes: drbd_resource=mail
> Operations: start interval=0s timeout=240 (mail-start-interval-0s)
> promote interval=0s timeout=90 (mail-promote-interval-0s)
> demote interval=0s timeout=90 (mail-demote-interval-0s)
> stop interval=0s timeout=100 (mail-stop-interval-0s)
> monitor interval=10s (mail-monitor-interval-10s)
> Group: fs-services
> Resource: fs-spool (class=ocf provider=heartbeat type=Filesystem)
> Attributes: device=/dev/drbd0 directory=/var/spool/postfix fstype=ext4
> options=nodev,nosuid,noexec
> Operations: start interval=0s timeout=60 (fs-spool-start-interval-0s)
> stop interval=0s timeout=60 (fs-spool-stop-interval-0s)
> monitor interval=20 timeout=40 (fs-spool-monitor-interval-20)
> Resource: fs-mail (class=ocf provider=heartbeat type=Filesystem)
> Attributes: device=/dev/drbd1 directory=/var/spool/mail fstype=ext4
> options=nodev,nosuid,noexec
> Operations: start interval=0s timeout=60 (fs-mail-start-interval-0s)
> stop interval=0s timeout=60 (fs-mail-stop-interval-0s)
> monitor interval=20 timeout=40 (fs-mail-monitor-interval-20)
> Group: mail-services
> Resource: postfix (class=ocf provider=heartbeat type=postfix)
> Operations: start interval=0s timeout=20s (postfix-start-interval-0s)
> stop interval=0s timeout=20s (postfix-stop-interval-0s)
> monitor interval=45s (postfix-monitor-interval-45s)
>
> Stonith Devices:
> Fencing Levels:
>
> Location Constraints:
> Ordering Constraints:
> start network-services then promote mail-clone (kind:Mandatory)
> (id:order-network-services-mail-clone-mandatory)
> promote mail-clone then promote spool-clone (kind:Mandatory)
> (id:order-mail-clone-spool-clone-mandatory)
> promote spool-clone then start fs-services (kind:Mandatory)
> (id:order-spool-clone-fs-services-mandatory)
> start fs-services then start mail-services (kind:Mandatory)
> (id:order-fs-services-mail-services-mandatory)
> Colocation Constraints:
> network-services with spool-clone (score:INFINITY) (rsc-role:Started)
> (with-rsc-role:Master) (id:colocation-network-services-spool-clone-INFINITY)
> network-services with mail-clone (score:INFINITY) (rsc-role:Started)
> (with-rsc-role:Master) (id:colocation-network-services-mail-clone-INFINITY)
> network-services with fs-services (score:INFINITY)
> (id:colocation-network-services-fs-services-INFINITY)
> network-services with mail-services (score:INFINITY)
> (id:colocation-network-services-mail-services-INFINITY)
I'm not sure whether it's causing your issue, but I would make the
constraints reflect the logical relationships better.
For example, network-services only needs to be colocated with
mail-services logically; it's mail-services that needs to be with
fs-services, and fs-services that needs to be with
spool-clone/mail-clone master. In other words, don't make the
highest-level resource depend on everything else, make each level depend
on the level below it.
Also, I would guess that the virtual IP only needs to be ordered before
mail-services, and mail-clone and spool-clone could both be ordered
before fs-services, rather than ordering mail-clone before spool-clone.
> Resources Defaults:
> migration-threshold: 3
> Operations Defaults:
> on-fail: restart
>
> Cluster Properties:
> cluster-infrastructure: corosync
> cluster-name: mailcluster
> cluster-recheck-interval: 5min
> dc-version: 1.1.13-10.el7_2.2-44eb2dd
> default-resource-stickiness: infinity
> have-watchdog: false
> last-lrm-refresh: 1457613674
> no-quorum-policy: ignore
> pe-error-series-max: 1024
> pe-input-series-max: 1024
> pe-warn-series-max: 1024
> stonith-enabled: false
>
>
>
>
>
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: --- 0.197.15 2
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: +++ 0.197.16 (null)
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: + /cib: @num_updates=16
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: +
> /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='postfix']/lrm_rsc_op[@id='postfix_last_failure_0']:
> @transition-key=4:1234:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @transition-magic=0:7;4:1234:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @call-id=1274, @last-rc-change=1457613440
> Mar 10 13:37:20 [7420] HWJ-626.domain.local crmd: info:
> abort_transition_graph: Transition aborted by postfix_monitor_45000
> 'modify' on mail1: Inactive graph
> (magic=0:7;4:1234:0:ae755a85-c250-498f-9c94-ddd8a7e2788a, cib=0.197.16,
> source=process_graph_event:598, 1)
> Mar 10 13:37:20 [7420] HWJ-626.domain.local crmd: info:
> update_failcount: Updating failcount for postfix on mail1 after failed
> monitor: rc=7 (update=value++, time=1457613440)
> Mar 10 13:37:20 [7418] HWJ-626.domain.local attrd: info:
> attrd_client_update: Expanded fail-count-postfix=value++ to 3
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_process_request: Completed cib_modify operation for section status: OK
> (rc=0, origin=mail1/crmd/196, version=0.197.16)
> Mar 10 13:37:20 [7418] HWJ-626.domain.local attrd: info:
> attrd_peer_update: Setting fail-count-postfix[mail1]: 2 -> 3 from mail2
> Mar 10 13:37:20 [7418] HWJ-626.domain.local attrd: info:
> write_attribute: Sent update 400 with 2 changes for
> fail-count-postfix, id=<n/a>, set=(null)
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_process_request: Forwarding cib_modify operation for section status to
> master (origin=local/attrd/400)
> Mar 10 13:37:20 [7420] HWJ-626.domain.local crmd: info:
> process_graph_event: Detected action (1234.4)
> postfix_monitor_45000.1274=not running: failed
> Mar 10 13:37:20 [7418] HWJ-626.domain.local attrd: info:
> attrd_peer_update: Setting last-failure-postfix[mail1]: 1457613347 ->
> 1457613440 from mail2
> Mar 10 13:37:20 [7420] HWJ-626.domain.local crmd: notice:
> do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [
> input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
> Mar 10 13:37:20 [7418] HWJ-626.domain.local attrd: info:
> write_attribute: Sent update 401 with 2 changes for
> last-failure-postfix, id=<n/a>, set=(null)
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: --- 0.197.16 2
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: +++ 0.197.17 (null)
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: + /cib: @num_updates=17
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: +
> /cib/status/node_state[@id='1']/transient_attributes[@id='1']/instance_attributes[@id='status-1']/nvpair[@id='status-1-fail-count-postfix']:
> @value=3
> Mar 10 13:37:20 [7415] HWJ-626.domain.local cib: info:
> cib_process_request: Completed cib_modify operation for section status: OK
> (rc=0, origin=mail2/attrd/400, version=0.197.17)
> Mar 10 13:37:20 [7420] HWJ-626.domain.local crmd: info:
> abort_transition_graph: Transition aborted by
> status-1-fail-count-postfix, fail-count-postfix=3: Transient attribute
> change (modify cib=0.197.17, source=abort_unless_down:319,
> path=/cib/status/node_state[@id='1']/transient_attributes[@id='1']/instance_attributes[@id='status-1']/nvpair[@id='status-1-fail-count-postfix'],
> 1)
> Mar 10 13:37:20 [7418] HWJ-626.domain.local attrd: info:
> attrd_cib_callback: Update 400 for fail-count-postfix: OK (0)
> Mar 10 13:37:20 [7418] HWJ-626.domain.local attrd: info:
> attrd_cib_callback: Update 400 for fail-count-postfix[mail1]=3: OK (0)
> Mar 10 13:37:20 [7418] HWJ-626.domain.local attrd: info:
> attrd_cib_callback: Update 400 for fail-count-postfix[mail2]=(null): OK
> (0)
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_process_request: Forwarding cib_modify operation for section status to
> master (origin=local/attrd/401)
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: --- 0.197.17 2
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: +++ 0.197.18 (null)
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: + /cib: @num_updates=18
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: +
> /cib/status/node_state[@id='1']/transient_attributes[@id='1']/instance_attributes[@id='status-1']/nvpair[@id='status-1-last-failure-postfix']:
> @value=1457613440
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_process_request: Completed cib_modify operation for section status: OK
> (rc=0, origin=mail2/attrd/401, version=0.197.18)
> Mar 10 13:37:21 [7418] HWJ-626.domain.local attrd: info:
> attrd_cib_callback: Update 401 for last-failure-postfix: OK (0)
> Mar 10 13:37:21 [7418] HWJ-626.domain.local attrd: info:
> attrd_cib_callback: Update 401 for
> last-failure-postfix[mail1]=1457613440: OK (0)
> Mar 10 13:37:21 [7418] HWJ-626.domain.local attrd: info:
> attrd_cib_callback: Update 401 for
> last-failure-postfix[mail2]=1457610376: OK (0)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: info:
> abort_transition_graph: Transition aborted by
> status-1-last-failure-postfix, last-failure-postfix=1457613440: Transient
> attribute change (modify cib=0.197.18, source=abort_unless_down:319,
> path=/cib/status/node_state[@id='1']/transient_attributes[@id='1']/instance_attributes[@id='status-1']/nvpair[@id='status-1-last-failure-postfix'],
> 1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> unpack_config: On loss of CCM Quorum: Ignore
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_online_status: Node mail1 is online
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_online_status: Node mail2 is online
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource mail:0 active in
> master mode on mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource spool:0 active in
> master mode on mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource fs-spool active on
> mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource fs-mail active on
> mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: warning:
> unpack_rsc_op_failure: Processing failed op monitor for postfix on
> mail1: not running (7)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource spool:1 active in
> master mode on mail2
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource mail:1 active in
> master mode on mail2
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> group_print: Resource Group: network-services
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_print: virtualip-1 (ocf::heartbeat:IPaddr2): Started
> mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> clone_print: Master/Slave Set: spool-clone [spool]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> short_print: Masters: [ mail1 ]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> short_print: Slaves: [ mail2 ]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> clone_print: Master/Slave Set: mail-clone [mail]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> short_print: Masters: [ mail1 ]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> short_print: Slaves: [ mail2 ]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> group_print: Resource Group: fs-services
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_print: fs-spool (ocf::heartbeat:Filesystem): Started mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_print: fs-mail (ocf::heartbeat:Filesystem): Started mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> group_print: Resource Group: mail-services
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_print: postfix (ocf::heartbeat:postfix): FAILED mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> get_failcount_full: postfix has failed 3 times on mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: warning:
> common_apply_stickiness: Forcing postfix away from mail1 after 3
> failures (max=3)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> master_color: Promoting mail:0 (Master mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> master_color: mail-clone: Promoted 1 instances of a possible 1 to master
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> master_color: Promoting spool:0 (Master mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> master_color: spool-clone: Promoted 1 instances of a possible 1 to master
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> rsc_merge_weights: postfix: Rolling back scores from virtualip-1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_color: Resource virtualip-1 cannot run anywhere
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> RecurringOp: Start recurring monitor (45s) for postfix on mail2
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> LogActions: Stop virtualip-1 (mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> LogActions: Leave spool:0 (Master mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> LogActions: Leave spool:1 (Slave mail2)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> LogActions: Leave mail:0 (Master mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> LogActions: Leave mail:1 (Slave mail2)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> LogActions: Stop fs-spool (Started mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> LogActions: Stop fs-mail (Started mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> LogActions: Stop postfix (Started mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> process_pe_message: Calculated Transition 1235:
> /var/lib/pacemaker/pengine/pe-input-302.bz2
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: info:
> handle_response: pe_calc calculation pe_calc-dc-1457613441-3756 is
> obsolete
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> unpack_config: On loss of CCM Quorum: Ignore
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_online_status: Node mail1 is online
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_online_status: Node mail2 is online
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource mail:0 active in
> master mode on mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource spool:0 active in
> master mode on mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource fs-spool active on
> mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource fs-mail active on
> mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: warning:
> unpack_rsc_op_failure: Processing failed op monitor for postfix on
> mail1: not running (7)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource spool:1 active in
> master mode on mail2
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> determine_op_status: Operation monitor found resource mail:1 active in
> master mode on mail2
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> group_print: Resource Group: network-services
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_print: virtualip-1 (ocf::heartbeat:IPaddr2): Started
> mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> clone_print: Master/Slave Set: spool-clone [spool]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> short_print: Masters: [ mail1 ]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> short_print: Slaves: [ mail2 ]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> clone_print: Master/Slave Set: mail-clone [mail]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> short_print: Masters: [ mail1 ]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> short_print: Slaves: [ mail2 ]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> group_print: Resource Group: fs-services
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_print: fs-spool (ocf::heartbeat:Filesystem): Started mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_print: fs-mail (ocf::heartbeat:Filesystem): Started mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> group_print: Resource Group: mail-services
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_print: postfix (ocf::heartbeat:postfix): FAILED mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> get_failcount_full: postfix has failed 3 times on mail1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: warning:
> common_apply_stickiness: Forcing postfix away from mail1 after 3
> failures (max=3)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> master_color: Promoting mail:0 (Master mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> master_color: mail-clone: Promoted 1 instances of a possible 1 to master
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> master_color: Promoting spool:0 (Master mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> master_color: spool-clone: Promoted 1 instances of a possible 1 to master
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> rsc_merge_weights: postfix: Rolling back scores from virtualip-1
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> native_color: Resource virtualip-1 cannot run anywhere
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> RecurringOp: Start recurring monitor (45s) for postfix on mail2
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> LogActions: Stop virtualip-1 (mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> LogActions: Leave spool:0 (Master mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> LogActions: Leave spool:1 (Slave mail2)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> LogActions: Leave mail:0 (Master mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: info:
> LogActions: Leave mail:1 (Slave mail2)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> LogActions: Stop fs-spool (Started mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> LogActions: Stop fs-mail (Started mail1)
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> LogActions: Stop postfix (Started mail1)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: info:
> do_state_transition: State transition S_POLICY_ENGINE ->
> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE
> origin=handle_response ]
> Mar 10 13:37:21 [7419] HWJ-626.domain.local pengine: notice:
> process_pe_message: Calculated Transition 1236:
> /var/lib/pacemaker/pengine/pe-input-303.bz2
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: info:
> do_te_invoke: Processing graph 1236 (ref=pe_calc-dc-1457613441-3757)
> derived from /var/lib/pacemaker/pengine/pe-input-303.bz2
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> te_rsc_command: Initiating action 12: stop virtualip-1_stop_0 on mail1
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> te_rsc_command: Initiating action 5: stop postfix_stop_0 on mail1
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: --- 0.197.18 2
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: +++ 0.197.19 (null)
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: + /cib: @num_updates=19
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: +
> /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='virtualip-1']/lrm_rsc_op[@id='virtualip-1_last_0']:
> @operation_key=virtualip-1_stop_0, @operation=stop,
> @transition-key=12:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @transition-magic=0:0;12:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @call-id=1276, @last-run=1457613441, @last-rc-change=1457613441,
> @exec-time=66
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_process_request: Completed cib_modify operation for section status: OK
> (rc=0, origin=mail1/crmd/197, version=0.197.19)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: info:
> match_graph_event: Action virtualip-1_stop_0 (12) confirmed on mail1
> (rc=0)
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: --- 0.197.19 2
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: +++ 0.197.20 (null)
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: + /cib: @num_updates=20
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: +
> /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='postfix']/lrm_rsc_op[@id='postfix_last_0']:
> @operation_key=postfix_stop_0, @operation=stop,
> @transition-key=5:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @transition-magic=0:0;5:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @call-id=1278, @last-run=1457613441, @last-rc-change=1457613441,
> @exec-time=476
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: info:
> match_graph_event: Action postfix_stop_0 (5) confirmed on mail1 (rc=0)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> te_rsc_command: Initiating action 79: stop fs-mail_stop_0 on mail1
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_process_request: Completed cib_modify operation for section status: OK
> (rc=0, origin=mail1/crmd/198, version=0.197.20)
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: --- 0.197.20 2
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: +++ 0.197.21 (null)
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: + /cib: @num_updates=21
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: +
> /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='fs-mail']/lrm_rsc_op[@id='fs-mail_last_0']:
> @operation_key=fs-mail_stop_0, @operation=stop,
> @transition-key=79:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @transition-magic=0:0;79:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @call-id=1280, @last-run=1457613441, @last-rc-change=1457613441,
> @exec-time=88, @queue-time=1
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_process_request: Completed cib_modify operation for section status: OK
> (rc=0, origin=mail1/crmd/199, version=0.197.21)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: info:
> match_graph_event: Action fs-mail_stop_0 (79) confirmed on mail1 (rc=0)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> te_rsc_command: Initiating action 77: stop fs-spool_stop_0 on mail1
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: --- 0.197.21 2
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: Diff: +++ 0.197.22 (null)
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: + /cib: @num_updates=22
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_perform_op: +
> /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='fs-spool']/lrm_rsc_op[@id='fs-spool_last_0']:
> @operation_key=fs-spool_stop_0, @operation=stop,
> @transition-key=77:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @transition-magic=0:0;77:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
> @call-id=1282, @last-run=1457613441, @last-rc-change=1457613441,
> @exec-time=86
> Mar 10 13:37:21 [7415] HWJ-626.domain.local cib: info:
> cib_process_request: Completed cib_modify operation for section status: OK
> (rc=0, origin=mail1/crmd/200, version=0.197.22)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: info:
> match_graph_event: Action fs-spool_stop_0 (77) confirmed on mail1 (rc=0)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: warning:
> run_graph: Transition 1236 (Complete=11, Pending=0, Fired=0, Skipped=0,
> Incomplete=1, Source=/var/lib/pacemaker/pengine/pe-input-303.bz2):
> Terminated
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: warning:
> te_graph_trigger: Transition failed: terminated
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_graph: Graph 1236 with 12 actions: batch-limit=12 jobs,
> network-delay=0ms
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 16]: Completed pseudo op
> network-services_stopped_0 on N/A (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 15]: Completed pseudo op
> network-services_stop_0 on N/A (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 12]: Completed rsc op virtualip-1_stop_0
> on mail1 (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 84]: Completed pseudo op
> fs-services_stopped_0 on N/A (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 83]: Completed pseudo op fs-services_stop_0
> on N/A (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 77]: Completed rsc op fs-spool_stop_0
> on mail1 (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 79]: Completed rsc op fs-mail_stop_0
> on mail1 (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 90]: Completed pseudo op
> mail-services_stopped_0 on N/A (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 89]: Completed pseudo op
> mail-services_stop_0 on N/A (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 86]: Pending rsc op postfix_monitor_45000
> on mail2 (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: * [Input 85]: Unresolved dependency rsc op
> postfix_start_0 on mail2
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 5]: Completed rsc op postfix_stop_0
> on mail1 (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> print_synapse: [Action 8]: Completed pseudo op all_stopped
> on N/A (priority: 0, waiting: none)
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: info: do_log:
> FSA: Input I_TE_SUCCESS from notify_crmd() received in state
> S_TRANSITION_ENGINE
> Mar 10 13:37:21 [7420] HWJ-626.domain.local crmd: notice:
> do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [
> input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
> Mar 10 13:37:26 [7415] HWJ-626.domain.local cib: info:
> cib_process_ping: Reporting our current digest to mail2:
> 3896ee29cdb6ba128330b0ef6e41bd79 for 0.197.22 (0x1544a30 0)
More information about the Users
mailing list