[ClusterLabs] Cluster failover failure with Unresolved dependency

Lorand Kelemen lorand.kelemen at gmail.com
Thu Mar 10 16:49:29 CET 2016


Dear List,

After the creation and testing of a simple 2 node active-passive
drbd+postfix cluster nearly everything works flawlessly (standby, failure
of a filesystem resource + failover, splitbrain + manual recovery) however
when delibarately killing the postfix instance, after reaching the
migration threshold failover does not occur and resources revert to the
Stopped state (except the master-slave drbd resource, which works as
expected).

Ordering and colocation is configured, STONITH and quorum disabled, the
goal is to always have one node running all the resources and at any sign
of error it should fail over to the passive node, nothing fancy.

Is my configuration wrong or am I hitting a bug?

All software from centos 7 + elrepo repositories.

Regarding STONITH: the machines are running on free ESXi instances on
separate machines, so the Vmware fencing agents won't work because in the
free version the API is read-only.
Still trying to figure out a way to go, until then manual recovery + huge
arp cache times on the upstream firewall...

Please find pe-input*.bz files attached, logs and config below. The
situation: on node mail1 postfix was killed 3 times (migration threshold),
it should have failed over to mail2.
When killing a filesystem resource three times this happens flawlessly.

Thanks for your input!

Best regards,
Lorand


Cluster Name: mailcluster
Corosync Nodes:
 mail1 mail2
Pacemaker Nodes:
 mail1 mail2

Resources:
 Group: network-services
  Resource: virtualip-1 (class=ocf provider=heartbeat type=IPaddr2)
   Attributes: ip=10.20.64.10 cidr_netmask=24 nic=lan0
   Operations: start interval=0s timeout=20s (virtualip-1-start-interval-0s)
               stop interval=0s timeout=20s (virtualip-1-stop-interval-0s)
               monitor interval=30s (virtualip-1-monitor-interval-30s)
 Master: spool-clone
  Meta Attrs: master-max=1 master-node-max=1 clone-max=2 clone-node-max=1
notify=true
  Resource: spool (class=ocf provider=linbit type=drbd)
   Attributes: drbd_resource=spool
   Operations: start interval=0s timeout=240 (spool-start-interval-0s)
               promote interval=0s timeout=90 (spool-promote-interval-0s)
               demote interval=0s timeout=90 (spool-demote-interval-0s)
               stop interval=0s timeout=100 (spool-stop-interval-0s)
               monitor interval=10s (spool-monitor-interval-10s)
 Master: mail-clone
  Meta Attrs: master-max=1 master-node-max=1 clone-max=2 clone-node-max=1
notify=true
  Resource: mail (class=ocf provider=linbit type=drbd)
   Attributes: drbd_resource=mail
   Operations: start interval=0s timeout=240 (mail-start-interval-0s)
               promote interval=0s timeout=90 (mail-promote-interval-0s)
               demote interval=0s timeout=90 (mail-demote-interval-0s)
               stop interval=0s timeout=100 (mail-stop-interval-0s)
               monitor interval=10s (mail-monitor-interval-10s)
 Group: fs-services
  Resource: fs-spool (class=ocf provider=heartbeat type=Filesystem)
   Attributes: device=/dev/drbd0 directory=/var/spool/postfix fstype=ext4
options=nodev,nosuid,noexec
   Operations: start interval=0s timeout=60 (fs-spool-start-interval-0s)
               stop interval=0s timeout=60 (fs-spool-stop-interval-0s)
               monitor interval=20 timeout=40 (fs-spool-monitor-interval-20)
  Resource: fs-mail (class=ocf provider=heartbeat type=Filesystem)
   Attributes: device=/dev/drbd1 directory=/var/spool/mail fstype=ext4
options=nodev,nosuid,noexec
   Operations: start interval=0s timeout=60 (fs-mail-start-interval-0s)
               stop interval=0s timeout=60 (fs-mail-stop-interval-0s)
               monitor interval=20 timeout=40 (fs-mail-monitor-interval-20)
 Group: mail-services
  Resource: postfix (class=ocf provider=heartbeat type=postfix)
   Operations: start interval=0s timeout=20s (postfix-start-interval-0s)
               stop interval=0s timeout=20s (postfix-stop-interval-0s)
               monitor interval=45s (postfix-monitor-interval-45s)

Stonith Devices:
Fencing Levels:

Location Constraints:
Ordering Constraints:
  start network-services then promote mail-clone (kind:Mandatory)
(id:order-network-services-mail-clone-mandatory)
  promote mail-clone then promote spool-clone (kind:Mandatory)
(id:order-mail-clone-spool-clone-mandatory)
  promote spool-clone then start fs-services (kind:Mandatory)
(id:order-spool-clone-fs-services-mandatory)
  start fs-services then start mail-services (kind:Mandatory)
(id:order-fs-services-mail-services-mandatory)
Colocation Constraints:
  network-services with spool-clone (score:INFINITY) (rsc-role:Started)
(with-rsc-role:Master) (id:colocation-network-services-spool-clone-INFINITY)
  network-services with mail-clone (score:INFINITY) (rsc-role:Started)
(with-rsc-role:Master) (id:colocation-network-services-mail-clone-INFINITY)
  network-services with fs-services (score:INFINITY)
(id:colocation-network-services-fs-services-INFINITY)
  network-services with mail-services (score:INFINITY)
(id:colocation-network-services-mail-services-INFINITY)

Resources Defaults:
 migration-threshold: 3
Operations Defaults:
 on-fail: restart

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: mailcluster
 cluster-recheck-interval: 5min
 dc-version: 1.1.13-10.el7_2.2-44eb2dd
 default-resource-stickiness: infinity
 have-watchdog: false
 last-lrm-refresh: 1457613674
 no-quorum-policy: ignore
 pe-error-series-max: 1024
 pe-input-series-max: 1024
 pe-warn-series-max: 1024
 stonith-enabled: false





Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: --- 0.197.15 2
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: +++ 0.197.16 (null)
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +  /cib:  @num_updates=16
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +
 /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='postfix']/lrm_rsc_op[@id='postfix_last_failure_0']:
 @transition-key=4:1234:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@transition-magic=0:7;4:1234:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@call-id=1274, @last-rc-change=1457613440
Mar 10 13:37:20 [7420] HWJ-626.domain.local       crmd:     info:
abort_transition_graph:       Transition aborted by postfix_monitor_45000
'modify' on mail1: Inactive graph
(magic=0:7;4:1234:0:ae755a85-c250-498f-9c94-ddd8a7e2788a, cib=0.197.16,
source=process_graph_event:598, 1)
Mar 10 13:37:20 [7420] HWJ-626.domain.local       crmd:     info:
update_failcount:     Updating failcount for postfix on mail1 after failed
monitor: rc=7 (update=value++, time=1457613440)
Mar 10 13:37:20 [7418] HWJ-626.domain.local      attrd:     info:
attrd_client_update:  Expanded fail-count-postfix=value++ to 3
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_process_request:  Completed cib_modify operation for section status: OK
(rc=0, origin=mail1/crmd/196, version=0.197.16)
Mar 10 13:37:20 [7418] HWJ-626.domain.local      attrd:     info:
attrd_peer_update:    Setting fail-count-postfix[mail1]: 2 -> 3 from mail2
Mar 10 13:37:20 [7418] HWJ-626.domain.local      attrd:     info:
write_attribute:      Sent update 400 with 2 changes for
fail-count-postfix, id=<n/a>, set=(null)
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_process_request:  Forwarding cib_modify operation for section status to
master (origin=local/attrd/400)
Mar 10 13:37:20 [7420] HWJ-626.domain.local       crmd:     info:
process_graph_event:  Detected action (1234.4)
postfix_monitor_45000.1274=not running: failed
Mar 10 13:37:20 [7418] HWJ-626.domain.local      attrd:     info:
attrd_peer_update:    Setting last-failure-postfix[mail1]: 1457613347 ->
1457613440 from mail2
Mar 10 13:37:20 [7420] HWJ-626.domain.local       crmd:   notice:
do_state_transition:  State transition S_IDLE -> S_POLICY_ENGINE [
input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
Mar 10 13:37:20 [7418] HWJ-626.domain.local      attrd:     info:
write_attribute:      Sent update 401 with 2 changes for
last-failure-postfix, id=<n/a>, set=(null)
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: --- 0.197.16 2
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: +++ 0.197.17 (null)
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +  /cib:  @num_updates=17
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +
 /cib/status/node_state[@id='1']/transient_attributes[@id='1']/instance_attributes[@id='status-1']/nvpair[@id='status-1-fail-count-postfix']:
 @value=3
Mar 10 13:37:20 [7415] HWJ-626.domain.local        cib:     info:
cib_process_request:  Completed cib_modify operation for section status: OK
(rc=0, origin=mail2/attrd/400, version=0.197.17)
Mar 10 13:37:20 [7420] HWJ-626.domain.local       crmd:     info:
abort_transition_graph:       Transition aborted by
status-1-fail-count-postfix, fail-count-postfix=3: Transient attribute
change (modify cib=0.197.17, source=abort_unless_down:319,
path=/cib/status/node_state[@id='1']/transient_attributes[@id='1']/instance_attributes[@id='status-1']/nvpair[@id='status-1-fail-count-postfix'],
1)
Mar 10 13:37:20 [7418] HWJ-626.domain.local      attrd:     info:
attrd_cib_callback:   Update 400 for fail-count-postfix: OK (0)
Mar 10 13:37:20 [7418] HWJ-626.domain.local      attrd:     info:
attrd_cib_callback:   Update 400 for fail-count-postfix[mail1]=3: OK (0)
Mar 10 13:37:20 [7418] HWJ-626.domain.local      attrd:     info:
attrd_cib_callback:   Update 400 for fail-count-postfix[mail2]=(null): OK
(0)
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_process_request:  Forwarding cib_modify operation for section status to
master (origin=local/attrd/401)
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: --- 0.197.17 2
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: +++ 0.197.18 (null)
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +  /cib:  @num_updates=18
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +
 /cib/status/node_state[@id='1']/transient_attributes[@id='1']/instance_attributes[@id='status-1']/nvpair[@id='status-1-last-failure-postfix']:
 @value=1457613440
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_process_request:  Completed cib_modify operation for section status: OK
(rc=0, origin=mail2/attrd/401, version=0.197.18)
Mar 10 13:37:21 [7418] HWJ-626.domain.local      attrd:     info:
attrd_cib_callback:   Update 401 for last-failure-postfix: OK (0)
Mar 10 13:37:21 [7418] HWJ-626.domain.local      attrd:     info:
attrd_cib_callback:   Update 401 for
last-failure-postfix[mail1]=1457613440: OK (0)
Mar 10 13:37:21 [7418] HWJ-626.domain.local      attrd:     info:
attrd_cib_callback:   Update 401 for
last-failure-postfix[mail2]=1457610376: OK (0)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:     info:
abort_transition_graph:       Transition aborted by
status-1-last-failure-postfix, last-failure-postfix=1457613440: Transient
attribute change (modify cib=0.197.18, source=abort_unless_down:319,
path=/cib/status/node_state[@id='1']/transient_attributes[@id='1']/instance_attributes[@id='status-1']/nvpair[@id='status-1-last-failure-postfix'],
1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
unpack_config:        On loss of CCM Quorum: Ignore
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_online_status:      Node mail1 is online
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_online_status:      Node mail2 is online
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource mail:0 active in
master mode on mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource spool:0 active in
master mode on mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource fs-spool active on
mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource fs-mail active on
mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:  warning:
unpack_rsc_op_failure:        Processing failed op monitor for postfix on
mail1: not running (7)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource spool:1 active in
master mode on mail2
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource mail:1 active in
master mode on mail2
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
group_print:   Resource Group: network-services
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_print:      virtualip-1        (ocf::heartbeat:IPaddr2):     Started
mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
clone_print:   Master/Slave Set: spool-clone [spool]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
short_print:       Masters: [ mail1 ]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
short_print:       Slaves: [ mail2 ]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
clone_print:   Master/Slave Set: mail-clone [mail]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
short_print:       Masters: [ mail1 ]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
short_print:       Slaves: [ mail2 ]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
group_print:   Resource Group: fs-services
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_print:      fs-spool   (ocf::heartbeat:Filesystem):    Started mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_print:      fs-mail    (ocf::heartbeat:Filesystem):    Started mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
group_print:   Resource Group: mail-services
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_print:      postfix    (ocf::heartbeat:postfix):       FAILED mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
get_failcount_full:   postfix has failed 3 times on mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:  warning:
common_apply_stickiness:      Forcing postfix away from mail1 after 3
failures (max=3)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
master_color: Promoting mail:0 (Master mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
master_color: mail-clone: Promoted 1 instances of a possible 1 to master
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
master_color: Promoting spool:0 (Master mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
master_color: spool-clone: Promoted 1 instances of a possible 1 to master
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
rsc_merge_weights:    postfix: Rolling back scores from virtualip-1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_color: Resource virtualip-1 cannot run anywhere
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
RecurringOp:   Start recurring monitor (45s) for postfix on mail2
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
LogActions:   Stop    virtualip-1     (mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
LogActions:   Leave   spool:0 (Master mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
LogActions:   Leave   spool:1 (Slave mail2)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
LogActions:   Leave   mail:0  (Master mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
LogActions:   Leave   mail:1  (Slave mail2)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
LogActions:   Stop    fs-spool        (Started mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
LogActions:   Stop    fs-mail (Started mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
LogActions:   Stop    postfix (Started mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
process_pe_message:   Calculated Transition 1235:
/var/lib/pacemaker/pengine/pe-input-302.bz2
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:     info:
handle_response:      pe_calc calculation pe_calc-dc-1457613441-3756 is
obsolete
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
unpack_config:        On loss of CCM Quorum: Ignore
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_online_status:      Node mail1 is online
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_online_status:      Node mail2 is online
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource mail:0 active in
master mode on mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource spool:0 active in
master mode on mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource fs-spool active on
mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource fs-mail active on
mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:  warning:
unpack_rsc_op_failure:        Processing failed op monitor for postfix on
mail1: not running (7)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource spool:1 active in
master mode on mail2
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
determine_op_status:  Operation monitor found resource mail:1 active in
master mode on mail2
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
group_print:   Resource Group: network-services
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_print:      virtualip-1        (ocf::heartbeat:IPaddr2):     Started
mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
clone_print:   Master/Slave Set: spool-clone [spool]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
short_print:       Masters: [ mail1 ]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
short_print:       Slaves: [ mail2 ]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
clone_print:   Master/Slave Set: mail-clone [mail]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
short_print:       Masters: [ mail1 ]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
short_print:       Slaves: [ mail2 ]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
group_print:   Resource Group: fs-services
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_print:      fs-spool   (ocf::heartbeat:Filesystem):    Started mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_print:      fs-mail    (ocf::heartbeat:Filesystem):    Started mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
group_print:   Resource Group: mail-services
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_print:      postfix    (ocf::heartbeat:postfix):       FAILED mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
get_failcount_full:   postfix has failed 3 times on mail1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:  warning:
common_apply_stickiness:      Forcing postfix away from mail1 after 3
failures (max=3)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
master_color: Promoting mail:0 (Master mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
master_color: mail-clone: Promoted 1 instances of a possible 1 to master
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
master_color: Promoting spool:0 (Master mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
master_color: spool-clone: Promoted 1 instances of a possible 1 to master
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
rsc_merge_weights:    postfix: Rolling back scores from virtualip-1
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
native_color: Resource virtualip-1 cannot run anywhere
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
RecurringOp:   Start recurring monitor (45s) for postfix on mail2
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
LogActions:   Stop    virtualip-1     (mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
LogActions:   Leave   spool:0 (Master mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
LogActions:   Leave   spool:1 (Slave mail2)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
LogActions:   Leave   mail:0  (Master mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:     info:
LogActions:   Leave   mail:1  (Slave mail2)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
LogActions:   Stop    fs-spool        (Started mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
LogActions:   Stop    fs-mail (Started mail1)
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
LogActions:   Stop    postfix (Started mail1)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:     info:
do_state_transition:  State transition S_POLICY_ENGINE ->
S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE
origin=handle_response ]
Mar 10 13:37:21 [7419] HWJ-626.domain.local    pengine:   notice:
process_pe_message:   Calculated Transition 1236:
/var/lib/pacemaker/pengine/pe-input-303.bz2
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:     info:
do_te_invoke: Processing graph 1236 (ref=pe_calc-dc-1457613441-3757)
derived from /var/lib/pacemaker/pengine/pe-input-303.bz2
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
te_rsc_command:       Initiating action 12: stop virtualip-1_stop_0 on mail1
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
te_rsc_command:       Initiating action 5: stop postfix_stop_0 on mail1
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: --- 0.197.18 2
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: +++ 0.197.19 (null)
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +  /cib:  @num_updates=19
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +
 /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='virtualip-1']/lrm_rsc_op[@id='virtualip-1_last_0']:
 @operation_key=virtualip-1_stop_0, @operation=stop,
@transition-key=12:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@transition-magic=0:0;12:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@call-id=1276, @last-run=1457613441, @last-rc-change=1457613441,
@exec-time=66
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_process_request:  Completed cib_modify operation for section status: OK
(rc=0, origin=mail1/crmd/197, version=0.197.19)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:     info:
match_graph_event:    Action virtualip-1_stop_0 (12) confirmed on mail1
(rc=0)
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: --- 0.197.19 2
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: +++ 0.197.20 (null)
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +  /cib:  @num_updates=20
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +
 /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='postfix']/lrm_rsc_op[@id='postfix_last_0']:
 @operation_key=postfix_stop_0, @operation=stop,
@transition-key=5:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@transition-magic=0:0;5:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@call-id=1278, @last-run=1457613441, @last-rc-change=1457613441,
@exec-time=476
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:     info:
match_graph_event:    Action postfix_stop_0 (5) confirmed on mail1 (rc=0)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
te_rsc_command:       Initiating action 79: stop fs-mail_stop_0 on mail1
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_process_request:  Completed cib_modify operation for section status: OK
(rc=0, origin=mail1/crmd/198, version=0.197.20)
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: --- 0.197.20 2
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: +++ 0.197.21 (null)
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +  /cib:  @num_updates=21
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +
 /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='fs-mail']/lrm_rsc_op[@id='fs-mail_last_0']:
 @operation_key=fs-mail_stop_0, @operation=stop,
@transition-key=79:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@transition-magic=0:0;79:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@call-id=1280, @last-run=1457613441, @last-rc-change=1457613441,
@exec-time=88, @queue-time=1
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_process_request:  Completed cib_modify operation for section status: OK
(rc=0, origin=mail1/crmd/199, version=0.197.21)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:     info:
match_graph_event:    Action fs-mail_stop_0 (79) confirmed on mail1 (rc=0)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
te_rsc_command:       Initiating action 77: stop fs-spool_stop_0 on mail1
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: --- 0.197.21 2
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       Diff: +++ 0.197.22 (null)
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +  /cib:  @num_updates=22
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_perform_op:       +
 /cib/status/node_state[@id='1']/lrm[@id='1']/lrm_resources/lrm_resource[@id='fs-spool']/lrm_rsc_op[@id='fs-spool_last_0']:
 @operation_key=fs-spool_stop_0, @operation=stop,
@transition-key=77:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@transition-magic=0:0;77:1236:0:ae755a85-c250-498f-9c94-ddd8a7e2788a,
@call-id=1282, @last-run=1457613441, @last-rc-change=1457613441,
@exec-time=86
Mar 10 13:37:21 [7415] HWJ-626.domain.local        cib:     info:
cib_process_request:  Completed cib_modify operation for section status: OK
(rc=0, origin=mail1/crmd/200, version=0.197.22)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:     info:
match_graph_event:    Action fs-spool_stop_0 (77) confirmed on mail1 (rc=0)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:  warning:
run_graph:    Transition 1236 (Complete=11, Pending=0, Fired=0, Skipped=0,
Incomplete=1, Source=/var/lib/pacemaker/pengine/pe-input-303.bz2):
Terminated
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:  warning:
te_graph_trigger:     Transition failed: terminated
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_graph:  Graph 1236 with 12 actions: batch-limit=12 jobs,
network-delay=0ms
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   16]: Completed pseudo op
network-services_stopped_0     on N/A (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   15]: Completed pseudo op
network-services_stop_0        on N/A (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   12]: Completed rsc op virtualip-1_stop_0
             on mail1 (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   84]: Completed pseudo op
fs-services_stopped_0          on N/A (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   83]: Completed pseudo op fs-services_stop_0
            on N/A (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   77]: Completed rsc op fs-spool_stop_0
            on mail1 (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   79]: Completed rsc op fs-mail_stop_0
             on mail1 (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   90]: Completed pseudo op
mail-services_stopped_0        on N/A (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   89]: Completed pseudo op
mail-services_stop_0           on N/A (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action   86]: Pending rsc op postfix_monitor_45000
            on mail2 (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:         * [Input 85]: Unresolved dependency rsc op
postfix_start_0 on mail2
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action    5]: Completed rsc op postfix_stop_0
             on mail1 (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
print_synapse:        [Action    8]: Completed pseudo op all_stopped
             on N/A (priority: 0, waiting: none)
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:     info: do_log:
    FSA: Input I_TE_SUCCESS from notify_crmd() received in state
S_TRANSITION_ENGINE
Mar 10 13:37:21 [7420] HWJ-626.domain.local       crmd:   notice:
do_state_transition:  State transition S_TRANSITION_ENGINE -> S_IDLE [
input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Mar 10 13:37:26 [7415] HWJ-626.domain.local        cib:     info:
cib_process_ping:     Reporting our current digest to mail2:
3896ee29cdb6ba128330b0ef6e41bd79 for 0.197.22 (0x1544a30 0)
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://clusterlabs.org/pipermail/users/attachments/20160310/22941f39/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: pe-input-302.bz2
Type: application/x-bzip2
Size: 3462 bytes
Desc: not available
URL: <http://clusterlabs.org/pipermail/users/attachments/20160310/22941f39/attachment-0002.bz2>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: pe-input-303.bz2
Type: application/x-bzip2
Size: 3460 bytes
Desc: not available
URL: <http://clusterlabs.org/pipermail/users/attachments/20160310/22941f39/attachment-0003.bz2>


More information about the Users mailing list