[Pacemaker] Problems with migration of kvm on primary/primary cluster

Patrik Plank p.plank at st-georgen-gusen.ooe.gv.at
Tue Jun 7 08:26:31 UTC 2011


Hallo,

I have a "little" Problem with the migration of kvm vm. 
I build a cluster with drbd,ocfs2 and pacemaker on two nodes (primary/primary).
If I do the migration manually with "virt-manager" or with "crm resource" all works fine, but if I reboot or shutdown the node, 
pacemaker STOP the vm completely and start it again on the other node.

I don't know what to do next.
I cant find the error in the config? 


name of the nodes: virtualserver03 and virtualserver05
resource kvm vm: unbuntu


cib:

node virtualserver03 \
        attributes standby="off"
node virtualserver05 \
        attributes standby="off"
primitive resDLM ocf:pacemaker:controld \
        op monitor interval="120s"
primitive resDRBD ocf:linbit:drbd \
        params drbd_resource="disk0" \
        operations $id="resDRBD-operations" \
        op monitor interval="20" role="Master" timeout="20" \
        op monitor interval="30" role="Slave" timeout="20"
primitive resFS ocf:heartbeat:Filesystem \
        params device="/dev/drbd/by-res/disk0" directory="/opt" fstype="ocfs2" \
        op monitor interval="120s"
primitive resO2CB ocf:pacemaker:o2cb \
        op monitor interval="120s"
primitive ubuntu ocf:heartbeat:VirtualDomain \
        params config="/etc/libvirt/qemu/ubuntu.xml" hypervisor="qemu:///system" migration_transport="ssh" force_stop="0" \
        op start interval="0" timeout="90" \
        op stop interval="0" timeout="300" \
        op monitor interval="10" timeout="30" depth="0" \
        op migrate_from interval="0" timeout="240" \
        op migrate_to interval="0" timeout="240" \
        meta allow-migrate="true" target-role="started" is-managed="true"
ms msDRBD resDRBD \
        meta resource-stickines="100" notify="true" master-max="2" interleave="true"
clone cloneDLM resDLM \
        meta globally-unique="false" interleave="true"
clone cloneFS resFS \
        meta interleave="true" ordered="true"
clone cloneO2CB resO2CB \
        meta globally-unique="false" interleave="true"
colocation colDLMDRBD inf: cloneDLM msDRBD:Master
colocation colFSO2CB inf: cloneFS cloneO2CB
colocation colO2CBDLM inf: cloneO2CB cloneDLM
order ordDLMO2CB 0: cloneDLM cloneO2CB
order ordDRBDDLM 0: msDRBD:promote cloneDLM
order ordO2CBFS 0: cloneO2CB cloneFS
property $id="cib-bootstrap-options" \
        dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \
        cluster-infrastructure="openais" \
        stonith-enabled="false" \
        no-quorum-policy="ignore" \
        expected-quorum-votes="2" \
        last-lrm-refresh="1307433605"


The logfile when virtualserver03 go into standby


Jun  7 09:09:23 virtualserver03 crmd: [903]: notice: run_graph: Transition 26 (Complete=2, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pengine/pe-input-52.bz2): Complete
Jun  7 09:09:23 virtualserver03 crmd: [903]: info: te_graph_trigger: Transition 26 is now complete
Jun  7 09:09:23 virtualserver03 crmd: [903]: info: notify_crmd: Transition 26 status: done - <null>
Jun  7 09:09:23 virtualserver03 crmd: [903]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Jun  7 09:09:23 virtualserver03 crmd: [903]: info: do_state_transition: Starting PEngine Recheck Timer
Jun  7 09:09:25 virtualserver03 cibadmin: [14056]: info: Invoked: /usr/sbin/cibadmin -Ql 
Jun  7 09:09:32 virtualserver03 cib: [898]: WARN: send_ipc_message: IPC Channel to 1403 is not connected
Jun  7 09:09:34 virtualserver03 kernel: [ 2920.781261] vnet0: no IPv6 routers present
Jun  7 09:09:38 virtualserver03 kernel: [ 2924.891266] br0: port 2(vnet0) entering forwarding state
Jun  7 09:09:43 virtualserver03 cib: [898]: WARN: send_ipc_message: IPC Channel to 1403 is not connected
Jun  7 09:10:15 virtualserver03 cib: last message repeated 2 times
Jun  7 09:10:15 virtualserver03 cibadmin: [15019]: info: Invoked: cibadmin -Ql -o nodes 
Jun  7 09:10:15 virtualserver03 crm_attribute: [15024]: info: Invoked: crm_attribute -N virtualserver03 -n standby -v on -l forever 
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="138" num_updates="3" >
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: -   <configuration >
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: -     <nodes >
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: -       <node id="virtualserver03" >
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: -         <instance_attributes id="nodes-virtualserver03" >
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: -           <nvpair value="off" id="nodes-virtualserver03-standby" />
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: -         </instance_attributes>
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: -       </node>
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: -     </nodes>
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: -   </configuration>
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: - </cib>
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: abort_transition_graph: need_abort:59 - Triggered transition abort (complete=1) : Non-status change
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="139" num_updates="1" >
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: need_abort: Aborting on change to admin_epoch
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: +   <configuration >
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: +     <nodes >
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: +       <node id="virtualserver03" >
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: +         <instance_attributes id="nodes-virtualserver03" >
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: +           <nvpair value="on" id="nodes-virtualserver03-standby" />
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_pe_invoke: Query 160: Requesting the current CIB: S_POLICY_ENGINE
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: +         </instance_attributes>
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: +       </node>
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: +     </nodes>
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: +   </configuration>
Jun  7 09:10:15 virtualserver03 cib: [898]: info: log_data_element: cib:diff: + </cib>
Jun  7 09:10:15 virtualserver03 cib: [898]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crm_attribute/4, version=0.139.1): ok (rc=0)
Jun  7 09:10:15 virtualserver03 cib: [15025]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-25.raw
Jun  7 09:10:15 virtualserver03 cib: [15025]: info: write_cib_contents: Wrote version 0.139.0 of the CIB to disk (digest: 8ac9dbbc5de14a957fbfa8d372ff03c9)
Jun  7 09:10:15 virtualserver03 cib: [15025]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.gNfdj7 (digest: /var/lib/heartbeat/crm/cib.3YnDJX)
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_pe_invoke_callback: Invoking the PE: query=160, ref=pe_calc-dc-1307430615-172, seq=132, quorate=1
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: unpack_status: Node virtualserver03 is in standby-mode
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: determine_online_status: Node virtualserver03 is standby
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: determine_online_status: Node virtualserver05 is online
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_print:  Master/Slave Set: msDRBD
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: short_print:      Masters: [ virtualserver03 virtualserver05 ]
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_print:  Clone Set: cloneDLM
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: short_print:      Started: [ virtualserver05 virtualserver03 ]
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_print:  Clone Set: cloneO2CB
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: short_print:      Started: [ virtualserver05 virtualserver03 ]
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_print:  Clone Set: cloneFS
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: short_print:      Started: [ virtualserver05 virtualserver03 ]
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: native_print: ubuntu#011(ocf::heartbeat:VirtualDomain):#011Started virtualserver03
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneDLM
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneO2CB
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: native_color: Resource resDRBD:0 cannot run anywhere
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneDLM
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneO2CB
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: master_color: Promoting resDRBD:1 (Master virtualserver05)
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: master_color: msDRBD: Promoted 1 instances of a possible 2 to master
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: native_merge_weights: cloneDLM: Rolling back scores from cloneO2CB
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: master_color: Promoting resDRBD:1 (Master virtualserver05)
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: master_color: msDRBD: Promoted 1 instances of a possible 2 to master
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: master_color: Promoting resDRBD:1 (Master virtualserver05)
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: master_color: msDRBD: Promoted 1 instances of a possible 2 to master
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: native_color: Resource resDLM:1 cannot run anywhere
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: native_color: Resource resO2CB:1 cannot run anywhere
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: native_color: Resource resFS:1 cannot run anywhere
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: RecurringOp:  Start recurring monitor (10s) for ubuntu on virtualserver05
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resO2CB:0 with resDLM:0 on virtualserver05
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDLM:0 with resO2CB:0
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDLM:0 with resO2CB:0 on virtualserver05
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resO2CB:0 with resDLM:0
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDLM:1 with resO2CB:1 on virtualserver03
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resO2CB:1 with resDLM:1
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDLM:0 with resDRBD:1 on virtualserver05
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDRBD:1 with resDLM:0
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDRBD:0 with resDLM:1 on virtualserver03
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDLM:1 with resDRBD:0
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDRBD:1 with resDLM:0 on virtualserver05
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDLM:0 with resDRBD:1
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resFS:0 with resO2CB:0 on virtualserver05
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resO2CB:0 with resFS:0
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resO2CB:0 with resFS:0 on virtualserver05
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resFS:0 with resO2CB:0
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resO2CB:1 with resFS:1 on virtualserver03
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resFS:1 with resO2CB:1
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Demote resDRBD:0#011(Master -> Stopped virtualserver03)
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Stop resource resDRBD:0#011(virtualserver03)
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resDRBD:1#011(Master virtualserver05)
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resDLM:0#011(Started virtualserver05)
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Stop resource resDLM:1#011(virtualserver03)
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resO2CB:0#011(Started virtualserver05)
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Stop resource resO2CB:1#011(virtualserver03)
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resFS:0#011(Started virtualserver05)
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Stop resource resFS:1#011(virtualserver03)
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: complex_migrate_reload: Migrating ubuntu from virtualserver03 to virtualserver05
Jun  7 09:10:15 virtualserver03 pengine: [902]: notice: LogActions: Migrate resource ubuntu#011(Started virtualserver03 -> virtualserver05)
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: unpack_graph: Unpacked transition 27: 34 actions in 34 synapses
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_te_invoke: Processing graph 27 (ref=pe_calc-dc-1307430615-172) derived from /var/lib/pengine/pe-input-53.bz2
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 40 fired and confirmed
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 49 fired and confirmed
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 63 fired and confirmed
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 65: migrate_to ubuntu_migrate_to_0 on virtualserver03 (local)
Jun  7 09:10:15 virtualserver03 pengine: [902]: info: process_pe_message: Transition 27: PEngine Input stored in: /var/lib/pengine/pe-input-53.bz2
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: cancel_op: operation monitor[56] on ocf::VirtualDomain::ubuntu for client 903, its parameters: hypervisor=[qemu:///system] CRM_meta_depth=[0] CRM_meta_interval=[10000] force_stop=[0] config=[/etc/libvirt/qemu/ubuntu.xml] depth=[0] crm_feature_set=[3.0.1] CRM_meta_name=[monitor] migration_transport=[ssh] CRM_meta_timeout=[30000]  cancelled
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_lrm_rsc_op: Performing key=65:27:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2 op=ubuntu_migrate_to_0 )
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: rsc:ubuntu:57: migrate_to
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation ubuntu_monitor_10000 (call=56, status=1, cib-update=0, confirmed=true) Cancelled
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 84: notify resDRBD:0_pre_notify_demote_0 on virtualserver03 (local)
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_lrm_rsc_op: Performing key=84:27:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2 op=resDRBD:0_notify_0 )
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: rsc:resDRBD:0:58: notify
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 86: notify resDRBD:1_pre_notify_demote_0 on virtualserver05
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 60: stop resFS:1_stop_0 on virtualserver03 (local)
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: cancel_op: operation monitor[45] on ocf::Filesystem::resFS:1 for client 903, its parameters: CRM_meta_clone=[1] fstype=[ocfs2] device=[/dev/drbd/by-res/disk0] CRM_meta_clone_max=[2] crm_feature_set=[3.0.1] CRM_meta_globally_unique=[false] directory=[/opt] CRM_meta_name=[monitor] CRM_meta_timeout=[20000] CRM_meta_interval=[120000]  cancelled
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_lrm_rsc_op: Performing key=60:27:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2 op=resFS:1_stop_0 )
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: rsc:resFS:1:59: stop
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resFS:1_monitor_120000 (call=45, status=1, cib-update=0, confirmed=true) Cancelled
Jun  7 09:10:15 virtualserver03 Filesystem[15028]: INFO: Running stop for /dev/drbd/by-res/disk0 on /opt
Jun  7 09:10:15 virtualserver03 VirtualDomain[15026]: INFO: ubuntu: Starting live migration to virtualserver05 (using remote hypervisor URI qemu+ssh://virtualserver05/system).
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: match_graph_event: Action resDRBD:1_pre_notify_demote_0 (86) confirmed on virtualserver05 (rc=0)
Jun  7 09:10:15 virtualserver03 Filesystem[15028]: INFO: Trying to unmount /opt
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resDRBD:0_notify_0 (call=58, rc=0, cib-update=161, confirmed=true) ok
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resFS:1:stop:stderr) umount: /opt: device is busy.#012        (In some cases useful info about processes that use#012         the device is found by lsof(8) or fuser(1))
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: match_graph_event: Action resDRBD:0_pre_notify_demote_0 (84) confirmed on virtualserver03 (rc=0)
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 41 fired and confirmed
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 38 fired and confirmed
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 14: demote resDRBD:0_demote_0 on virtualserver03 (local)
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: cancel_op: operation monitor[40] on ocf::drbd::resDRBD:0 for client 903, its parameters: CRM_meta_clone=[0] CRM_meta_timeout=[20000] CRM_meta_notify_slave_resource=[ ] CRM_meta_notify_active_resource=[ ] CRM_meta_notify_demote_uname=[ ] drbd_resource=[disk0] CRM_meta_notify_stop_resource=[ ] CRM_meta_notify_master_resource=[resDRBD:1 ] CRM_meta_clone_max=[2] CRM_meta_notify_demote_resource=[ ] CRM_meta_notify_start_resource=[resDRBD:0 ] CRM_meta_notify_master_uname=[virtualserver05 ] CRM_meta_master_max=[2] CRM_meta_ cancelled
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: do_lrm_rsc_op: Performing key=14:27:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2 op=resDRBD:0_demote_0 )
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: rsc:resDRBD:0:60: demote
Jun  7 09:10:15 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resDRBD:0_monitor_20000 (call=40, status=1, cib-update=0, confirmed=true) Cancelled
Jun  7 09:10:15 virtualserver03 Filesystem[15028]: ERROR: Couldn't unmount /opt; trying cleanup with SIGTERM
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resFS:1:stop:stderr) /opt:
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resFS:1:stop:stderr)  
Jun  7 09:10:15 virtualserver03 lrmd: last message repeated 14 times
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resFS:1:stop:stdout)  14027
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resFS:1:stop:stderr) 
Jun  7 09:10:15 virtualserver03 Filesystem[15028]: INFO: Some processes on /opt were signalled
Jun  7 09:10:15 virtualserver03 kernel: [ 2961.696583] block drbd0: State change failed: Device is held open by someone
Jun  7 09:10:15 virtualserver03 kernel: [ 2961.703532] br0: port 2(vnet0) entering disabled state
Jun  7 09:10:15 virtualserver03 kernel: [ 2961.706567] block drbd0:   state = { cs:Connected ro:Primary/Primary ds:UpToDate/UpToDate r--- }
Jun  7 09:10:15 virtualserver03 kernel: [ 2961.706567] block drbd0:  wanted = { cs:Connected ro:Secondary/Primary ds:UpToDate/UpToDate r--- }
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stderr) 0: State change failed: (-12) Device is held open by someone
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stderr) Command 'drbdsetup 0 secondary' terminated with exit code 11
Jun  7 09:10:15 virtualserver03 drbd[15127]: ERROR: disk0: Called drbdadm -c /etc/drbd.conf secondary disk0
Jun  7 09:10:15 virtualserver03 drbd[15127]: ERROR: disk0: Exit code 11
Jun  7 09:10:15 virtualserver03 drbd[15127]: ERROR: disk0: Command output: 
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stdout) 
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stderr) 0: State change failed: (-12) Device is held open by someone
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stderr) Command 'drbdsetup 0 secondary' terminated with exit code 11
Jun  7 09:10:15 virtualserver03 drbd[15127]: ERROR: disk0: Called drbdadm -c /etc/drbd.conf secondary disk0
Jun  7 09:10:15 virtualserver03 drbd[15127]: ERROR: disk0: Exit code 11
Jun  7 09:10:15 virtualserver03 drbd[15127]: ERROR: disk0: Command output: 
Jun  7 09:10:15 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stdout) 
Jun  7 09:10:15 virtualserver03 kernel: [ 2961.791246] device vnet0 left promiscuous mode
Jun  7 09:10:15 virtualserver03 kernel: [ 2961.791741] block drbd0: State change failed: Device is held open by someone
Jun  7 09:10:15 virtualserver03 kernel: [ 2961.791748] block drbd0:   state = { cs:Connected ro:Primary/Primary ds:UpToDate/UpToDate r--- }
Jun  7 09:10:15 virtualserver03 kernel: [ 2961.791753] block drbd0:  wanted = { cs:Connected ro:Secondary/Primary ds:UpToDate/UpToDate r--- }
Jun  7 09:10:15 virtualserver03 kernel: [ 2961.866998] br0: port 2(vnet0) entering disabled state
Jun  7 09:10:16 virtualserver03 kernel: [ 2962.249486] type=1505 audit(1307430616.014:25):  operation="profile_remove" pid=15205 name="libvirt-147ae184-c43d-80d3-9ce1-67c3801154ae" namespace="root"
Jun  7 09:10:16 virtualserver03 cib: [898]: WARN: send_ipc_message: IPC Channel to 1403 is not connected
Jun  7 09:10:16 virtualserver03 kernel: [ 2962.814480] block drbd0: State change failed: Device is held open by someone
Jun  7 09:10:16 virtualserver03 kernel: [ 2962.824463] block drbd0:   state = { cs:Connected ro:Primary/Primary ds:UpToDate/UpToDate r--- }
Jun  7 09:10:16 virtualserver03 kernel: [ 2962.824463] block drbd0:  wanted = { cs:Connected ro:Secondary/Primary ds:UpToDate/UpToDate r--- }
Jun  7 09:10:16 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stderr) 0: State change failed: (-12) Device is held open by someone
Jun  7 09:10:16 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stderr) Command 'drbdsetup 0 secondary' terminated with exit code 11
Jun  7 09:10:16 virtualserver03 drbd[15127]: ERROR: disk0: Called drbdadm -c /etc/drbd.conf secondary disk0
Jun  7 09:10:16 virtualserver03 drbd[15127]: ERROR: disk0: Exit code 11
Jun  7 09:10:16 virtualserver03 drbd[15127]: ERROR: disk0: Command output: 
Jun  7 09:10:16 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stdout) 
Jun  7 09:10:16 virtualserver03 libvirtd: 09:10:16.668: error : qemudDomainMigratePerform:7615 : Requested operation is not valid: domain is not running
Jun  7 09:10:16 virtualserver03 kernel: [ 2962.913569] ocfs2: Unmounting device (147,0) on (node 1728096)
Jun  7 09:10:16 virtualserver03 Filesystem[15028]: INFO: unmounted /opt successfully
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resFS:1_stop_0 (call=59, rc=0, cib-update=162, confirmed=true) ok
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: match_graph_event: Action resFS:1_stop_0 (60) confirmed on virtualserver03 (rc=0)
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 64 fired and confirmed
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 56 fired and confirmed
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 53: stop resO2CB:1_stop_0 on virtualserver03 (local)
Jun  7 09:10:16 virtualserver03 lrmd: [900]: info: cancel_op: operation monitor[43] on ocf::o2cb::resO2CB:1 for client 903, its parameters: CRM_meta_clone_max=[2] CRM_meta_name=[monitor] crm_feature_set=[3.0.1] CRM_meta_timeout=[20000] CRM_meta_globally_unique=[false] CRM_meta_interval=[120000] CRM_meta_clone=[1]  cancelled
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: do_lrm_rsc_op: Performing key=53:27:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2 op=resO2CB:1_stop_0 )
Jun  7 09:10:16 virtualserver03 lrmd: [900]: info: rsc:resO2CB:1:61: stop
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resO2CB:1_monitor_120000 (call=43, status=1, cib-update=0, confirmed=true) Cancelled
Jun  7 09:10:16 virtualserver03 o2cb[15259]: INFO: Stopping resO2CB:1
Jun  7 09:10:16 virtualserver03 o2cb[15259]: INFO: Stopping ocfs2_controld.pcmk
Jun  7 09:10:16 virtualserver03 ocfs2_controld[6156]: Caught signal 15, exiting
Jun  7 09:10:16 virtualserver03 lrmd: [900]: info: RA output: (ubuntu:migrate_to:stderr) error: Unknown failure
Jun  7 09:10:16 virtualserver03 VirtualDomain[15026]: ERROR: ubuntu: live migration to qemu+ssh://virtualserver05/system failed: 1
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation ubuntu_migrate_to_0 (call=57, rc=1, cib-update=163, confirmed=true) unknown error
Jun  7 09:10:16 virtualserver03 crmd: [903]: WARN: status_from_rc: Action 65 (ubuntu_migrate_to_0) on virtualserver03 failed (target: 0 vs. rc: 1): Error
Jun  7 09:10:16 virtualserver03 crmd: [903]: WARN: update_failcount: Updating failcount for ubuntu on virtualserver03 after failed stop: rc=1 (update=INFINITY, time=1307430616)
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: abort_transition_graph: match_graph_event:272 - Triggered transition abort (complete=0, tag=lrm_rsc_op, id=ubuntu_stop_0, magic=0:1;65:27:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2, cib=0.139.5) : Event failed
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: update_abort_priority: Abort priority upgraded from 0 to 1
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: update_abort_priority: Abort action done superceeded by restart
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: match_graph_event: Action ubuntu_stop_0 (65) confirmed on virtualserver03 (rc=4)
Jun  7 09:10:16 virtualserver03 attrd: [901]: info: attrd_trigger_update: Sending flush op to all hosts for: fail-count-ubuntu (INFINITY)
Jun  7 09:10:16 virtualserver03 attrd: [901]: info: attrd_perform_update: Sent update 79: fail-count-ubuntu=INFINITY
Jun  7 09:10:16 virtualserver03 attrd: [901]: info: attrd_trigger_update: Sending flush op to all hosts for: last-failure-ubuntu (1307430616)
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: abort_transition_graph: te_update_diff:146 - Triggered transition abort (complete=0, tag=transient_attributes, id=virtualserver03, magic=NA, cib=0.139.6) : Transient attribute: update
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: update_abort_priority: Abort priority upgraded from 1 to 1000000
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: update_abort_priority: 'Event failed' abort superceeded
Jun  7 09:10:16 virtualserver03 attrd: [901]: info: attrd_perform_update: Sent update 81: last-failure-ubuntu=1307430616
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: abort_transition_graph: te_update_diff:146 - Triggered transition abort (complete=0, tag=transient_attributes, id=virtualserver03, magic=NA, cib=0.139.7) : Transient attribute: update
Jun  7 09:10:16 virtualserver03 kernel: [ 2963.161914] ocfs2: Unregistered cluster interface user
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resO2CB:1_stop_0 (call=61, rc=0, cib-update=164, confirmed=true) ok
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: match_graph_event: Action resO2CB:1_stop_0 (53) confirmed on virtualserver03 (rc=0)
Jun  7 09:10:16 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 57 fired and confirmed
Jun  7 09:10:17 virtualserver03 cibadmin: [15342]: info: Invoked: /usr/sbin/cibadmin -Ql 
Jun  7 09:10:17 virtualserver03 kernel: [ 2963.898376] block drbd0: role( Primary -> Secondary ) 
Jun  7 09:10:17 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:demote:stdout) 
Jun  7 09:10:18 virtualserver03 cibadmin: [15385]: info: Invoked: /usr/sbin/cibadmin -Ql 
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resDRBD:0_demote_0 (call=60, rc=0, cib-update=165, confirmed=true) ok
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: match_graph_event: Action resDRBD:0_demote_0 (14) confirmed on virtualserver03 (rc=0)
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 39 fired and confirmed
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 42 fired and confirmed
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 85: notify resDRBD:0_post_notify_demote_0 on virtualserver03 (local)
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_lrm_rsc_op: Performing key=85:27:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2 op=resDRBD:0_notify_0 )
Jun  7 09:10:18 virtualserver03 lrmd: [900]: info: rsc:resDRBD:0:62: notify
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 87: notify resDRBD:1_post_notify_demote_0 on virtualserver05
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: match_graph_event: Action resDRBD:1_post_notify_demote_0 (87) confirmed on virtualserver05 (rc=0)
Jun  7 09:10:18 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:notify:stdout) 
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resDRBD:0_notify_0 (call=62, rc=0, cib-update=166, confirmed=true) ok
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: match_graph_event: Action resDRBD:0_post_notify_demote_0 (85) confirmed on virtualserver03 (rc=0)
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 43 fired and confirmed
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: run_graph: ====================================================
Jun  7 09:10:18 virtualserver03 crmd: [903]: notice: run_graph: Transition 27 (Complete=19, Pending=0, Fired=0, Skipped=10, Incomplete=5, Source=/var/lib/pengine/pe-input-53.bz2): Stopped
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_graph_trigger: Transition 27 is now complete
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=notify_crmd ]
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_pe_invoke: Query 167: Requesting the current CIB: S_POLICY_ENGINE
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_pe_invoke_callback: Invoking the PE: query=167, ref=pe_calc-dc-1307430618-181, seq=132, quorate=1
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: unpack_status: Node virtualserver03 is in standby-mode
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: determine_online_status: Node virtualserver03 is standby
Jun  7 09:10:18 virtualserver03 pengine: [902]: WARN: unpack_rsc_op: Processing failed op ubuntu_stop_0 on virtualserver03: unknown error (1)
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_add_running: resource ubuntu isnt managed
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: determine_online_status: Node virtualserver05 is online
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_print:  Master/Slave Set: msDRBD
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: short_print:      Masters: [ virtualserver05 ]
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: short_print:      Slaves: [ virtualserver03 ]
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_print:  Clone Set: cloneDLM
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: short_print:      Started: [ virtualserver05 virtualserver03 ]
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_print:  Clone Set: cloneO2CB
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: short_print:      Started: [ virtualserver05 ]
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: short_print:      Stopped: [ resO2CB:1 ]
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_print:  Clone Set: cloneFS
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: short_print:      Started: [ virtualserver05 ]
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: short_print:      Stopped: [ resFS:1 ]
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: native_print: ubuntu#011(ocf::heartbeat:VirtualDomain):#011Started virtualserver03 (unmanaged) FAILED
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: get_failcount: ubuntu has failed 1000000 times on virtualserver03
Jun  7 09:10:18 virtualserver03 pengine: [902]: WARN: common_apply_stickiness: Forcing ubuntu away from virtualserver03 after 1000000 failures (max=1000000)
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneDLM
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneO2CB
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_color: Resource resDRBD:0 cannot run anywhere
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneDLM
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneO2CB
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: master_color: Promoting resDRBD:1 (Master virtualserver05)
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: master_color: msDRBD: Promoted 1 instances of a possible 2 to master
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_merge_weights: cloneDLM: Rolling back scores from cloneO2CB
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: master_color: Promoting resDRBD:1 (Master virtualserver05)
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: master_color: msDRBD: Promoted 1 instances of a possible 2 to master
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: master_color: Promoting resDRBD:1 (Master virtualserver05)
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: master_color: msDRBD: Promoted 1 instances of a possible 2 to master
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_color: Resource resDLM:1 cannot run anywhere
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_color: Resource resO2CB:1 cannot run anywhere
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_color: Resource resFS:1 cannot run anywhere
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: native_color: Unmanaged resource ubuntu allocated to 'nowhere': failed
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resO2CB:0 with resDLM:0 on virtualserver05
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDLM:0 with resO2CB:0
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDLM:0 with resO2CB:0 on virtualserver05
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resO2CB:0 with resDLM:0
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDLM:0 with resDRBD:1 on virtualserver05
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDRBD:1 with resDLM:0
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDRBD:0 with resDLM:1 on virtualserver03
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDLM:1 with resDRBD:0
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDRBD:1 with resDLM:0 on virtualserver05
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDLM:0 with resDRBD:1
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resFS:0 with resO2CB:0 on virtualserver05
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resO2CB:0 with resFS:0
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resO2CB:0 with resFS:0 on virtualserver05
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resFS:0 with resO2CB:0
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: LogActions: Stop resource resDRBD:0#011(virtualserver03)
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resDRBD:1#011(Master virtualserver05)
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resDLM:0#011(Started virtualserver05)
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: LogActions: Stop resource resDLM:1#011(virtualserver03)
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resO2CB:0#011(Started virtualserver05)
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resO2CB:1#011(Stopped)
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resFS:0#011(Started virtualserver05)
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resFS:1#011(Stopped)
Jun  7 09:10:18 virtualserver03 pengine: [902]: notice: LogActions: Leave resource ubuntu#011(Started unmanaged)
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: unpack_graph: Unpacked transition 28: 14 actions in 14 synapses
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_te_invoke: Processing graph 28 (ref=pe_calc-dc-1307430618-181) derived from /var/lib/pengine/pe-input-54.bz2
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 24 fired and confirmed
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 45 fired and confirmed
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 76: notify resDRBD:0_pre_notify_stop_0 on virtualserver03 (local)
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_lrm_rsc_op: Performing key=76:28:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2 op=resDRBD:0_notify_0 )
Jun  7 09:10:18 virtualserver03 lrmd: [900]: info: rsc:resDRBD:0:63: notify
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 77: notify resDRBD:1_pre_notify_stop_0 on virtualserver05
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 42: stop resDLM:1_stop_0 on virtualserver03 (local)
Jun  7 09:10:18 virtualserver03 lrmd: [900]: info: cancel_op: operation monitor[41] on ocf::controld::resDLM:1 for client 903, its parameters: CRM_meta_clone_max=[2] CRM_meta_name=[monitor] crm_feature_set=[3.0.1] CRM_meta_timeout=[20000] CRM_meta_globally_unique=[false] CRM_meta_interval=[120000] CRM_meta_clone=[1]  cancelled
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_lrm_rsc_op: Performing key=42:28:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2 op=resDLM:1_stop_0 )
Jun  7 09:10:18 virtualserver03 lrmd: [900]: info: rsc:resDLM:1:64: stop
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resDLM:1_monitor_120000 (call=41, status=1, cib-update=0, confirmed=true) Cancelled
Jun  7 09:10:18 virtualserver03 pengine: [902]: info: process_pe_message: Transition 28: PEngine Input stored in: /var/lib/pengine/pe-input-54.bz2
Jun  7 09:10:18 virtualserver03 kernel: [ 2965.048034] dlm: closing connection to node 1761650880
Jun  7 09:10:18 virtualserver03 kernel: [ 2965.059725] dlm: closing connection to node 1728096448
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: match_graph_event: Action resDRBD:1_pre_notify_stop_0 (77) confirmed on virtualserver05 (rc=0)
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resDRBD:0_notify_0 (call=63, rc=0, cib-update=168, confirmed=true) ok
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: match_graph_event: Action resDRBD:0_pre_notify_stop_0 (76) confirmed on virtualserver03 (rc=0)
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 25 fired and confirmed
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 22 fired and confirmed
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 11: stop resDRBD:0_stop_0 on virtualserver03 (local)
Jun  7 09:10:18 virtualserver03 crmd: [903]: info: do_lrm_rsc_op: Performing key=11:28:0:f9c426e5-68d1-49f0-ae7f-0eff420e35a2 op=resDRBD:0_stop_0 )
Jun  7 09:10:18 virtualserver03 lrmd: [900]: info: rsc:resDRBD:0:65: stop
Jun  7 09:10:18 virtualserver03 kernel: [ 2965.141353] block drbd0: peer( Primary -> Unknown ) conn( Connected -> Disconnecting ) pdsk( UpToDate -> DUnknown ) 
Jun  7 09:10:18 virtualserver03 kernel: [ 2965.150786] block drbd0: meta connection shut down by peer.
Jun  7 09:10:18 virtualserver03 kernel: [ 2965.176483] block drbd0: asender terminated
Jun  7 09:10:18 virtualserver03 kernel: [ 2965.188282] block drbd0: Terminating asender thread
Jun  7 09:10:18 virtualserver03 kernel: [ 2965.199854] block drbd0: Connection closed
Jun  7 09:10:18 virtualserver03 kernel: [ 2965.211136] block drbd0: conn( Disconnecting -> StandAlone ) 
Jun  7 09:10:19 virtualserver03 kernel: [ 2965.222421] block drbd0: receiver terminated
Jun  7 09:10:19 virtualserver03 kernel: [ 2965.222468] block drbd0: disk( UpToDate -> Diskless ) 
Jun  7 09:10:19 virtualserver03 kernel: [ 2965.244230] block drbd0: Terminating receiver thread
Jun  7 09:10:19 virtualserver03 kernel: [ 2965.252864] block drbd0: State change failed: Disk state is lower than outdated
Jun  7 09:10:19 virtualserver03 kernel: [ 2965.252870] block drbd0:   state = { cs:StandAlone ro:Secondary/Unknown ds:Diskless/DUnknown r--- }
Jun  7 09:10:19 virtualserver03 kernel: [ 2965.252874] block drbd0:  wanted = { cs:StandAlone ro:Secondary/Unknown ds:Outdated/DUnknown r--- }
Jun  7 09:10:19 virtualserver03 kernel: [ 2965.319324] block drbd0: drbd_bm_resize called with capacity == 0
Jun  7 09:10:19 virtualserver03 kernel: [ 2965.330489] block drbd0: worker terminated
Jun  7 09:10:19 virtualserver03 kernel: [ 2965.341395] block drbd0: Terminating worker thread
Jun  7 09:10:19 virtualserver03 attrd: [901]: info: attrd_trigger_update: Sending flush op to all hosts for: master-resDRBD:0 (-INFINITY)
Jun  7 09:10:19 virtualserver03 attrd: [901]: info: attrd_perform_update: Sent update 83: master-resDRBD:0=-INFINITY
Jun  7 09:10:19 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:stop:stdout) #012
Jun  7 09:10:19 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:stop:stdout) 
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: abort_transition_graph: te_update_diff:146 - Triggered transition abort (complete=0, tag=transient_attributes, id=virtualserver03, magic=NA, cib=0.139.14) : Transient attribute: update
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: update_abort_priority: Abort priority upgraded from 0 to 1000000
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: update_abort_priority: Abort action done superceeded by restart
Jun  7 09:10:19 virtualserver03 crm_attribute: [15500]: info: Invoked: crm_attribute -N virtualserver03 -n master-resDRBD:0 -l reboot -D 
Jun  7 09:10:19 virtualserver03 attrd: [901]: info: attrd_trigger_update: Sending flush op to all hosts for: master-resDRBD:0 (<null>)
Jun  7 09:10:19 virtualserver03 attrd: [901]: info: attrd_perform_update: Sent delete 85: node=virtualserver03, attr=master-resDRBD:0, id=<n/a>, set=(null), section=status
Jun  7 09:10:19 virtualserver03 lrmd: [900]: info: RA output: (resDRBD:0:stop:stdout) 
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resDRBD:0_stop_0 (call=65, rc=0, cib-update=169, confirmed=true) ok
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: abort_transition_graph: te_update_diff:157 - Triggered transition abort (complete=0, tag=transient_attributes, id=virtualserver03, magic=NA, cib=0.139.15) : Transient attribute: removal
Jun  7 09:10:19 virtualserver03 attrd: [901]: info: attrd_perform_update: Sent delete -22: node=virtualserver03, attr=master-resDRBD:0, id=<n/a>, set=(null), section=status
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: match_graph_event: Action resDRBD:0_stop_0 (11) confirmed on virtualserver03 (rc=0)
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 23 fired and confirmed
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 26 fired and confirmed
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: te_rsc_command: Initiating action 78: notify resDRBD:1_post_notify_stop_0 on virtualserver05
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: match_graph_event: Action resDRBD:1_post_notify_stop_0 (78) confirmed on virtualserver05 (rc=0)
Jun  7 09:10:19 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 27 fired and confirmed
Jun  7 09:10:19 virtualserver03 lrmd: [900]: info: RA output: (resDLM:1:stop:stderr) dlm_controld.pcmk: no process found
Jun  7 09:10:20 virtualserver03 cibadmin: [15532]: info: Invoked: /usr/sbin/cibadmin -Ql 
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: process_lrm_event: LRM operation resDLM:1_stop_0 (call=64, rc=0, cib-update=170, confirmed=true) ok
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: match_graph_event: Action resDLM:1_stop_0 (42) confirmed on virtualserver03 (rc=0)
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: te_pseudo_action: Pseudo action 46 fired and confirmed
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: run_graph: ====================================================
Jun  7 09:10:20 virtualserver03 crmd: [903]: notice: run_graph: Transition 28 (Complete=13, Pending=0, Fired=0, Skipped=1, Incomplete=0, Source=/var/lib/pengine/pe-input-54.bz2): Stopped
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: te_graph_trigger: Transition 28 is now complete
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=notify_crmd ]
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: do_pe_invoke: Query 171: Requesting the current CIB: S_POLICY_ENGINE
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: do_pe_invoke_callback: Invoking the PE: query=171, ref=pe_calc-dc-1307430620-187, seq=132, quorate=1
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: unpack_status: Node virtualserver03 is in standby-mode
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: determine_online_status: Node virtualserver03 is standby
Jun  7 09:10:20 virtualserver03 pengine: [902]: WARN: unpack_rsc_op: Processing failed op ubuntu_stop_0 on virtualserver03: unknown error (1)
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_add_running: resource ubuntu isnt managed
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: determine_online_status: Node virtualserver05 is online
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_print:  Master/Slave Set: msDRBD
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: short_print:      Masters: [ virtualserver05 ]
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: short_print:      Stopped: [ resDRBD:0 ]
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_print:  Clone Set: cloneDLM
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: short_print:      Started: [ virtualserver05 ]
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: short_print:      Stopped: [ resDLM:1 ]
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_print:  Clone Set: cloneO2CB
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: short_print:      Started: [ virtualserver05 ]
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: short_print:      Stopped: [ resO2CB:1 ]
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_print:  Clone Set: cloneFS
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: short_print:      Started: [ virtualserver05 ]
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: short_print:      Stopped: [ resFS:1 ]
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: native_print: ubuntu#011(ocf::heartbeat:VirtualDomain):#011Started virtualserver03 (unmanaged) FAILED
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: get_failcount: ubuntu has failed 1000000 times on virtualserver03
Jun  7 09:10:20 virtualserver03 pengine: [902]: WARN: common_apply_stickiness: Forcing ubuntu away from virtualserver03 after 1000000 failures (max=1000000)
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneDLM
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneO2CB
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_color: Resource resDRBD:0 cannot run anywhere
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneDLM
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_merge_weights: msDRBD: Rolling back scores from cloneO2CB
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: master_color: Promoting resDRBD:1 (Master virtualserver05)
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: master_color: msDRBD: Promoted 1 instances of a possible 2 to master
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_merge_weights: cloneDLM: Rolling back scores from cloneO2CB
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: master_color: Promoting resDRBD:1 (Master virtualserver05)
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: master_color: msDRBD: Promoted 1 instances of a possible 2 to master
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: master_color: Promoting resDRBD:1 (Master virtualserver05)
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: master_color: msDRBD: Promoted 1 instances of a possible 2 to master
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_color: Resource resDLM:1 cannot run anywhere
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_color: Resource resO2CB:1 cannot run anywhere
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_color: Resource resFS:1 cannot run anywhere
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: native_color: Unmanaged resource ubuntu allocated to 'nowhere': failed
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resO2CB:0 with resDLM:0 on virtualserver05
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDLM:0 with resO2CB:0
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDLM:0 with resO2CB:0 on virtualserver05
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resO2CB:0 with resDLM:0
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDLM:0 with resDRBD:1 on virtualserver05
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDRBD:1 with resDLM:0
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resDRBD:1 with resDLM:0 on virtualserver05
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resDLM:0 with resDRBD:1
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resFS:0 with resO2CB:0 on virtualserver05
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resO2CB:0 with resFS:0
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: find_compatible_child: Colocating resO2CB:0 with resFS:0 on virtualserver05
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: clone_rsc_order_lh: Interleaving resFS:0 with resO2CB:0
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resDRBD:0#011(Stopped)
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resDRBD:1#011(Master virtualserver05)
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resDLM:0#011(Started virtualserver05)
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resDLM:1#011(Stopped)
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resO2CB:0#011(Started virtualserver05)
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resO2CB:1#011(Stopped)
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resFS:0#011(Started virtualserver05)
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: LogActions: Leave resource resFS:1#011(Stopped)
Jun  7 09:10:20 virtualserver03 pengine: [902]: notice: LogActions: Leave resource ubuntu#011(Started unmanaged)
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: unpack_graph: Unpacked transition 29: 0 actions in 0 synapses
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: do_te_invoke: Processing graph 29 (ref=pe_calc-dc-1307430620-187) derived from /var/lib/pengine/pe-input-55.bz2
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: run_graph: ====================================================
Jun  7 09:10:20 virtualserver03 crmd: [903]: notice: run_graph: Transition 29 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pengine/pe-input-55.bz2): Complete
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: te_graph_trigger: Transition 29 is now complete
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: notify_crmd: Transition 29 status: done - <null>
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Jun  7 09:10:20 virtualserver03 crmd: [903]: info: do_state_transition: Starting PEngine Recheck Timer
Jun  7 09:10:20 virtualserver03 pengine: [902]: info: process_pe_message: Transition 29: PEngine Input stored in: /var/lib/pengine/pe-input-55.bz2
Jun  7 09:10:22 virtualserver03 cibadmin: [15535]: info: Invoked: /usr/sbin/cibadmin -Ql 


crm_mon -n

Last updated: Tue Jun  7 10:00:46 2011
Stack: openais
Current DC: virtualserver05 - partition with quorum
Version: 1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd
2 Nodes configured, 2 expected votes
5 Resources configured.
============

Node virtualserver03: online
        resDRBD:0       (ocf::linbit:drbd) Master
        resDLM:1        (ocf::pacemaker:controld) Started
        resO2CB:1       (ocf::pacemaker:o2cb) Started
        resFS:1 (ocf::heartbeat:Filesystem) Started
        ubuntu  (ocf::heartbeat:VirtualDomain) Started
Node virtualserver05: online
        resDLM:0        (ocf::pacemaker:controld) Started
        resDRBD:1       (ocf::linbit:drbd) Master
        resO2CB:0       (ocf::pacemaker:o2cb) Started
        resFS:0 (ocf::heartbeat:Filesystem) Started

I hope somebody can help me?
sorry for my english :)

king regards
Patrik
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20110607/2118ab05/attachment-0003.html>


More information about the Pacemaker mailing list