Dec 18 14:10:20 bl460g1n7 corosync[6909]:   [MAIN  ] main.c:main:1171 Corosync Cluster Engine ('2.3.2.8-7014f'): started and ready to provide service.
Dec 18 14:10:20 bl460g1n7 corosync[6909]:   [MAIN  ] main.c:main:1172 Corosync built-in features: watchdog upstart snmp pie relro bindnow
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [TOTEM ] totemnet.c:totemnet_instance_initialize:242 Initializing transport (UDP/IP Multicast).
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [TOTEM ] totemcrypto.c:init_nss:579 Initializing transmit/receive security (NSS) crypto: aes256 hash: sha1
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [TOTEM ] totemnet.c:totemnet_instance_initialize:242 Initializing transport (UDP/IP Multicast).
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [TOTEM ] totemcrypto.c:init_nss:579 Initializing transmit/receive security (NSS) crypto: aes256 hash: sha1
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [TOTEM ] totemudp.c:timer_function_netif_check_timeout:670 The network interface [192.168.101.217] is now up.
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync configuration map access [0]
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: cmap
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync configuration service [1]
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: cfg
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync cluster closed process group service v1.01 [2]
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: cpg
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync profile loading service [4]
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [WD    ] wd.c:setup_watchdog:651 Watchdog is now been tickled by corosync.
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [WD    ] wd.c:wd_scan_resources:580 no resources configured.
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync watchdog service [7]
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [QUORUM] vsf_quorum.c:quorum_exec_init_fn:274 Using quorum provider corosync_votequorum
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync vote quorum service v1.0 [5]
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: votequorum
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync cluster quorum service v0.1 [3]
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: quorum
Dec 18 14:10:20 bl460g1n7 corosync[6910]:   [TOTEM ] totemudp.c:timer_function_netif_check_timeout:670 The network interface [192.168.102.217] is now up.
Dec 18 14:10:21 bl460g1n7 corosync[6910]:   [TOTEM ] totemsrp.c:memb_state_operational_enter:2016 A new membership (192.168.101.217:4) was formed. Members joined: -1062705703
Dec 18 14:10:21 bl460g1n7 corosync[6910]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[1]: -1062705703
Dec 18 14:10:21 bl460g1n7 corosync[6910]:   [MAIN  ] main.c:corosync_sync_completed:276 Completed service synchronization, ready to provide service.
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: crm_ipc_connect: Could not establish pacemakerd connection: Connection refused (111)
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: get_cluster_type: Detected an active 'corosync' cluster
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: mcp_read_config: Reading configure for stack: corosync
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:   notice: mcp_read_config: Configured corosync to accept connections from group 189: OK (1)
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:   notice: main: Starting Pacemaker 1.1.11-0.4.ce5d77c.git.el6 (Build: ce5d77c):  generated-manpages agent-manpages ascii-docs ncurses libqb-logging libqb-ipc lha-fencing nagios  corosync-native snmp
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: main: Maximum core file size is: 18446744073709551615
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: qb_ipcs_us_publish: server name: pacemakerd
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: crm_get_peer: Created entry 2f16ab34-eb06-4b4d-bac3-35fff7a78648/0x1b788c0 for node (null)/3232261593 (1 total)
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:   notice: cluster_connect_quorum: Quorum lost
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Using uid=189 and group=189 for process cib
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Forked child 6923 for process cib
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Forked child 6924 for process stonith-ng
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Forked child 6925 for process lrmd
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Using uid=189 and group=189 for process attrd
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Forked child 6926 for process attrd
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Using uid=189 and group=189 for process pengine
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Forked child 6927 for process pengine
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Using uid=189 and group=189 for process crmd
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: start_child: Forked child 6928 for process crmd
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: main: Starting mainloop
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:     info: pcmk_quorum_notification: Membership 4: quorum still lost (1)
Dec 18 14:10:23 bl460g1n7 pacemakerd[6919]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n7[3232261593] - state is now member (was (null))
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Dec 18 14:10:23 bl460g1n7 cib[6923]:   notice: main: Using new config location: /var/lib/pacemaker/cib
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.xml (digest: /var/lib/pacemaker/cib/cib.xml.sig)
Dec 18 14:10:23 bl460g1n7 cib[6923]:  warning: retrieveCib: Cluster configuration not found: /var/lib/pacemaker/cib/cib.xml
Dec 18 14:10:23 bl460g1n7 cib[6923]:  warning: readCibXmlFile: Primary configuration corrupt or unusable, trying backups in /var/lib/pacemaker/cib
Dec 18 14:10:23 bl460g1n7 cib[6923]:  warning: readCibXmlFile: Continuing with an empty configuration.
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: validate_with_relaxng: Creating RNG parser context
Dec 18 14:10:23 bl460g1n7 lrmd[6925]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Dec 18 14:10:23 bl460g1n7 lrmd[6925]:     info: qb_ipcs_us_publish: server name: lrmd
Dec 18 14:10:23 bl460g1n7 lrmd[6925]:     info: main: Starting
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Dec 18 14:10:23 bl460g1n7 attrd[6926]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Dec 18 14:10:23 bl460g1n7 pengine[6927]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Dec 18 14:10:23 bl460g1n7 pengine[6927]:     info: qb_ipcs_us_publish: server name: pengine
Dec 18 14:10:23 bl460g1n7 pengine[6927]:     info: main: Starting pengine
Dec 18 14:10:23 bl460g1n7 crmd[6928]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Dec 18 14:10:23 bl460g1n7 crmd[6928]:   notice: main: CRM Git Version: ce5d77c
Dec 18 14:10:23 bl460g1n7 crmd[6928]:     info: do_log: FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
Dec 18 14:10:23 bl460g1n7 crmd[6928]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Dec 18 14:10:23 bl460g1n7 crmd[6928]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Dec 18 14:10:23 bl460g1n7 crmd[6928]:     info: crm_ipc_connect: Could not establish cib_shm connection: Connection refused (111)
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: startCib: CIB Initialization completed successfully
Dec 18 14:10:23 bl460g1n7 cib[6923]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:     info: crm_get_peer: Created entry 24dab3b8-e797-429b-9389-e18fce5074d5/0x19642b0 for node (null)/3232261593 (1 total)
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: crm_get_peer: Created entry 868dbf8a-e243-40b7-b7ec-b64bbbca436a/0x25dad00 for node (null)/3232261593 (1 total)
Dec 18 14:10:23 bl460g1n7 attrd[6926]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 attrd[6926]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Dec 18 14:10:23 bl460g1n7 attrd[6926]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261593] - state is now member (was (null))
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:     info: init_cs_connection_once: Connection to 'corosync': established
Dec 18 14:10:23 bl460g1n7 cib[6923]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 cib[6923]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: init_cs_connection_once: Connection to 'corosync': established
Dec 18 14:10:23 bl460g1n7 attrd[6926]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 attrd[6926]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Dec 18 14:10:23 bl460g1n7 stonith-ng[6924]:     info: crm_ipc_connect: Could not establish cib_rw connection: Connection refused (111)
Dec 18 14:10:23 bl460g1n7 cib[6923]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:23 bl460g1n7 cib[6923]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: qb_ipcs_us_publish: server name: cib_ro
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: qb_ipcs_us_publish: server name: cib_rw
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: qb_ipcs_us_publish: server name: cib_shm
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: cib_init: Starting cib mainloop
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: pcmk_cpg_membership: Joined[0.0] cib.3232261593 
Dec 18 14:10:23 bl460g1n7 cib[6923]:     info: pcmk_cpg_membership: Member[0.0] cib.3232261593 
Dec 18 14:10:23 bl460g1n7 cib[6929]:     info: write_cib_contents: Wrote version 0.0.0 of the CIB to disk (digest: 97b40f67f7fe5fa1d0a8198f10e4d67a)
Dec 18 14:10:23 bl460g1n7 cib[6929]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.EccUi9 (digest: /var/lib/pacemaker/cib/cib.zDnLmG)
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x25ddd70 for uid=189 gid=189 pid=6928 id=43f851ae-ddcd-4289-bd76-645757de4211
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: do_cib_control: CIB connection established
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.0.0)
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: crm_get_peer: Created entry b7d2ea99-c8c6-44c6-a035-d05655a5d293/0xc6fa80 for node (null)/3232261593 (1 total)
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: init_cs_connection_once: Connection to 'corosync': established
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: peer_update_callback: bl460g1n7 is now (null)
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: cluster_connect_quorum: Quorum lost
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: do_ha_control: Connected to the cluster
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.0.0)
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: lrmd_ipc_connect: Connecting to lrmd
Dec 18 14:10:24 bl460g1n7 lrmd[6925]:     info: crm_client_new: Connecting 0xdd8990 for uid=189 gid=189 pid=6928 id=b9124534-93b7-4090-ae9a-928dff54ae3c
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: do_lrm_control: LRM connection established
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: do_started: Delaying start, no membership data (0000000000100000)
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.0.0)
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: pcmk_quorum_notification: Membership 4: quorum still lost (1)
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n7[3232261593] - state is now member (was (null))
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: peer_update_callback: bl460g1n7 is now member (was (null))
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x242cbf0 for uid=189 gid=189 pid=6926 id=4a32daeb-e3b4-4752-b592-778d51779f32
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x265edc0 for uid=0 gid=0 pid=6924 id=9f25b010-e2ce-4250-b324-36bf6d598b65
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: do_started: Delaying start, Config not read (0000000000000040)
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: qb_ipcs_us_publish: server name: crmd
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: do_started: The local CRM is operational
Dec 18 14:10:24 bl460g1n7 crmd[6928]:     info: do_log: FSA: Input I_PENDING from do_started() received in state S_STARTING
Dec 18 14:10:24 bl460g1n7 crmd[6928]:   notice: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/5, version=0.0.0)
Dec 18 14:10:24 bl460g1n7 stonith-ng[6924]:   notice: setup_cib: Watching for stonith topology changes
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.0.0)
Dec 18 14:10:24 bl460g1n7 stonith-ng[6924]:     info: qb_ipcs_us_publish: server name: stonith-ng
Dec 18 14:10:24 bl460g1n7 stonith-ng[6924]:     info: main: Starting stonith-ng mainloop
Dec 18 14:10:24 bl460g1n7 stonith-ng[6924]:     info: init_cib_cache_cb: Updating device list from the cib: init
Dec 18 14:10:24 bl460g1n7 stonith-ng[6924]:     info: unpack_nodes: Creating a fake local node
Dec 18 14:10:24 bl460g1n7 stonith-ng[6924]:     info: pcmk_cpg_membership: Joined[0.0] stonith-ng.3232261593 
Dec 18 14:10:24 bl460g1n7 stonith-ng[6924]:     info: pcmk_cpg_membership: Member[0.0] stonith-ng.3232261593 
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x265f320 for uid=0 gid=0 pid=6712 id=cca887e2-4bbf-4b5e-96b7-64372c33ad7d
Dec 18 14:10:24 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_mon/2, version=0.0.0)
Dec 18 14:10:25 bl460g1n7 crmd[6928]:     info: pcmk_cpg_membership: Joined[0.0] crmd.3232261593 
Dec 18 14:10:25 bl460g1n7 crmd[6928]:     info: pcmk_cpg_membership: Member[0.0] crmd.3232261593 
Dec 18 14:10:26 bl460g1n7 stonith-ng[6924]:     info: crm_client_new: Connecting 0x196af60 for uid=189 gid=189 pid=6928 id=dc49a74a-3037-40ab-9703-ad49346a3b2b
Dec 18 14:10:26 bl460g1n7 stonith-ng[6924]:     info: stonith_command: Processed register from crmd.6928: OK (0)
Dec 18 14:10:26 bl460g1n7 stonith-ng[6924]:     info: stonith_command: Processed st_notify from crmd.6928: OK (0)
Dec 18 14:10:26 bl460g1n7 stonith-ng[6924]:     info: stonith_command: Processed st_notify from crmd.6928: OK (0)
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: crm_timer_popped: Election Trigger (I_DC_TIMEOUT) just popped (20000ms)
Dec 18 14:10:45 bl460g1n7 crmd[6928]:  warning: do_log: FSA: Input I_DC_TIMEOUT from crm_timer_popped() received in state S_PENDING
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_PENDING -> S_ELECTION [ input=I_DC_TIMEOUT cause=C_TIMER_POPPED origin=crm_timer_popped ]
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: election_complete: Election election-0 complete
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: election_timeout_popped: Election failed: Declaring ourselves the winner
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_log: FSA: Input I_ELECTION_DC from election_timeout_popped() received in state S_ELECTION
Dec 18 14:10:45 bl460g1n7 crmd[6928]:   notice: do_state_transition: State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC cause=C_TIMER_POPPED origin=election_timeout_popped ]
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_te_control: Registering TE UUID: ddf348fe-fbad-4abb-9a12-8250f71b075a
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: set_graph_functions: Setting custom graph functions
Dec 18 14:10:45 bl460g1n7 pengine[6927]:     info: crm_client_new: Connecting 0xbbb4c0 for uid=189 gid=189 pid=6928 id=64eadb9c-aa18-4f66-95ab-524eb5a3a018
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_dc_takeover: Taking over DC status for this partition
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_readwrite: We are now in R/W mode
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_master operation for section 'all': OK (rc=0, origin=local/crmd/6, version=0.0.0)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section cib: OK (rc=0, origin=local/crmd/7, version=0.0.1)
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/crm_config//cluster_property_set//nvpair[@name='dc-version']: No such device or address (rc=-6, origin=local/crmd/8, version=0.0.1)
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: log_cib_diff: cib:diff: Local-only Change: 0.1.1
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="0" num_updates="1"/>
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <cluster_property_set id="cib-bootstrap-options">
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.1.11-0.4.ce5d77c.git.el6-ce5d77c"/>
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </cluster_property_set>
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section crm_config: OK (rc=0, origin=local/crmd/9, version=0.1.1)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/crm_config//cluster_property_set//nvpair[@name='cluster-infrastructure']: No such device or address (rc=-6, origin=local/crmd/10, version=0.1.1)
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: join_make_offer: Making join offers based on membership 4
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: join_make_offer: join-1: Sending offer to bl460g1n7
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n7[3232261593] - join-1 phase 0 -> 1
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_dc_join_offer_all: join-1: Waiting on 1 outstanding join acks
Dec 18 14:10:45 bl460g1n7 crmd[6928]:  warning: do_log: FSA: Input I_ELECTION_DC from do_election_check() received in state S_INTEGRATION
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: initialize_join: Node bl460g1n7[3232261593] - join-2 phase 1 -> 0
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: join_make_offer: join-2: Sending offer to bl460g1n7
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n7[3232261593] - join-2 phase 0 -> 1
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_dc_join_offer_all: join-2: Waiting on 1 outstanding join acks
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: update_dc: Set DC to bl460g1n7 (3.0.8)
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: log_cib_diff: cib:diff: Local-only Change: 0.2.1
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="1" num_updates="1"/>
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="corosync"/>
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section crm_config: OK (rc=0, origin=local/crmd/11, version=0.2.1)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/12, version=0.2.1)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/13, version=0.2.1)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/14, version=0.2.1)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/15, version=0.2.1)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/16, version=0.2.1)
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n7[3232261593] - join-2 phase 1 -> 2
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: crm_update_peer_expected: do_dc_join_filter_offer: Node bl460g1n7[3232261593] - expected state is now member (was (null))
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: crmd_join_phase_log: join-2: bl460g1n7=integrated
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_dc_join_finalize: join-2: Syncing our CIB to the rest of the cluster
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_sync operation for section 'all': OK (rc=0, origin=local/crmd/17, version=0.2.1)
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: finalize_join_for: Node bl460g1n7[3232261593] - join-2 phase 2 -> 3
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: log_cib_diff: cib:diff: Local-only Change: 0.3.1
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="2" num_updates="1"/>
Dec 18 14:10:45 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <node id="3232261593" uname="bl460g1n7"/>
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/transient_attributes
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/18, version=0.3.1)
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: update_attrd_helper: Connecting to attrd... 5 retries remaining
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n7']/transient_attributes: OK (rc=0, origin=local/crmd/19, version=0.3.1)
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_ack: Node bl460g1n7[3232261593] - join-2 phase 3 -> 4
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_dc_join_ack: join-2: Updating node state to member for bl460g1n7
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/lrm
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=local/crmd/20, version=0.3.1)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/21, version=0.3.2)
Dec 18 14:10:45 bl460g1n7 attrd[6926]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:45 bl460g1n7 attrd[6926]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: abort_transition_graph: do_te_invoke:151 - Triggered transition abort (complete=1) : Peer Cancelled
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/22, version=0.3.2)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/23, version=0.3.2)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section cib: OK (rc=0, origin=local/crmd/24, version=0.3.3)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/25, version=0.3.3)
Dec 18 14:10:45 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 2 with 1 changes for terminate, id=<n/a>, set=(null)
Dec 18 14:10:45 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 3 with 1 changes for shutdown, id=<n/a>, set=(null)
Dec 18 14:10:45 bl460g1n7 pengine[6927]:  warning: cluster_status: We do not have quorum - fencing and resource management disabled
Dec 18 14:10:45 bl460g1n7 pengine[6927]:    error: unpack_resources: Resource start-up disabled since no STONITH resources have been defined
Dec 18 14:10:45 bl460g1n7 pengine[6927]:    error: unpack_resources: Either configure some or disable STONITH with the stonith-enabled option
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/2, version=0.3.4)
Dec 18 14:10:45 bl460g1n7 pengine[6927]:    error: unpack_resources: NOTE: Clusters with shared data need STONITH to ensure data integrity
Dec 18 14:10:45 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n7 is active
Dec 18 14:10:45 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n7 is online
Dec 18 14:10:45 bl460g1n7 pengine[6927]:   notice: stage6: Delaying fencing operations until there are resources to manage
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/3, version=0.3.4)
Dec 18 14:10:45 bl460g1n7 pengine[6927]:   notice: process_pe_message: Calculated Transition 0: /var/lib/pacemaker/pengine/pe-input-0.bz2
Dec 18 14:10:45 bl460g1n7 pengine[6927]:   notice: process_pe_message: Configuration ERRORs found during PE processing.  Please run "crm_verify -L" to identify issues.
Dec 18 14:10:45 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 2 for terminate[bl460g1n7]=(null): OK (0)
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_te_invoke: Processing graph 0 (ref=pe_calc-dc-1387343445-9) derived from /var/lib/pacemaker/pengine/pe-input-0.bz2
Dec 18 14:10:45 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 2: probe_complete probe_complete on bl460g1n7 (local) - no waiting
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: te_rsc_command: Action 2 confirmed - no wait
Dec 18 14:10:45 bl460g1n7 crmd[6928]:   notice: run_graph: Transition 0 (Complete=1, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-0.bz2): Complete
Dec 18 14:10:45 bl460g1n7 crmd[6928]:     info: do_log: FSA: Input I_TE_SUCCESS from notify_crmd() received in state S_TRANSITION_ENGINE
Dec 18 14:10:45 bl460g1n7 crmd[6928]:   notice: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Dec 18 14:10:45 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 3 for shutdown[bl460g1n7]=(null): OK (0)
Dec 18 14:10:45 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 4 with 1 changes for probe_complete, id=<n/a>, set=(null)
Dec 18 14:10:45 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/4, version=0.3.5)
Dec 18 14:10:45 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 4 for probe_complete[bl460g1n7]=true: OK (0)
Dec 18 14:10:45 bl460g1n7 cib[6937]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-0.raw
Dec 18 14:10:45 bl460g1n7 cib[6937]:     info: write_cib_contents: Wrote version 0.1.0 of the CIB to disk (digest: 37cd07e611cf33396df3755a7423f762)
Dec 18 14:10:45 bl460g1n7 cib[6937]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.fgk7rd (digest: /var/lib/pacemaker/cib/cib.cM8YHI)
Dec 18 14:10:45 bl460g1n7 cib[6938]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-1.raw
Dec 18 14:10:45 bl460g1n7 cib[6938]:     info: write_cib_contents: Wrote version 0.3.0 of the CIB to disk (digest: e32ba67134080b5f95287a40b192a41a)
Dec 18 14:10:45 bl460g1n7 cib[6938]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.R4l3Aj (digest: /var/lib/pacemaker/cib/cib.x077bP)
Dec 18 14:10:48 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:10:50 bl460g1n7 corosync[6910]:   [TOTEM ] totemsrp.c:memb_state_operational_enter:2016 A new membership (192.168.101.216:8) was formed. Members joined: -1062705704 -1062705702
Dec 18 14:10:50 bl460g1n7 corosync[6910]:   [QUORUM] vsf_quorum.c:quorum_api_set_quorum:148 This node is within the primary component and will provide service.
Dec 18 14:10:50 bl460g1n7 corosync[6910]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[1]: -1062705703
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: pcmk_quorum_notification: Membership 4: quorum acquired (1)
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: pcmk_quorum_notification: Membership 4: quorum acquired (1)
Dec 18 14:10:50 bl460g1n7 crmd[6928]:     info: crm_cs_flush: Sent 0 CPG messages  (1 remaining, last=9): Try again (6)
Dec 18 14:10:50 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section cib: OK (rc=0, origin=local/crmd/26, version=0.3.6)
Dec 18 14:10:50 bl460g1n7 cib[6923]:     info: crm_cs_flush: Sent 0 CPG messages  (1 remaining, last=9): Try again (6)
Dec 18 14:10:50 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/27, version=0.3.6)
Dec 18 14:10:50 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/28, version=0.3.6)
Dec 18 14:10:50 bl460g1n7 corosync[6910]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[3]: -1062705704 -1062705703 -1062705702
Dec 18 14:10:50 bl460g1n7 crmd[6928]:     info: pcmk_quorum_notification: Membership 8: quorum retained (3)
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:     info: pcmk_quorum_notification: Membership 8: quorum retained (3)
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:     info: crm_get_peer: Created entry 4917b316-f43e-4027-9f1e-a0bc91edfb67/0x1b818a0 for node (null)/3232261592 (2 total)
Dec 18 14:10:50 bl460g1n7 crmd[6928]:     info: crm_get_peer: Created entry 1e17f5f3-74bd-492b-8122-4dfcbacaf4e9/0xe46c80 for node (null)/3232261592 (2 total)
Dec 18 14:10:50 bl460g1n7 corosync[6910]:   [MAIN  ] main.c:corosync_sync_completed:276 Completed service synchronization, ready to provide service.
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 crmd[6928]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Dec 18 14:10:50 bl460g1n7 crmd[6928]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261592
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261592
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261592] - state is now member (was (null))
Dec 18 14:10:50 bl460g1n7 crmd[6928]:     info: crm_get_peer: Created entry 68c1f948-a7dd-4ea0-96d1-caedda502422/0xe452f0 for node (null)/3232261594 (3 total)
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261592] - state is now member (was (null))
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:     info: crm_get_peer: Created entry ee9bb930-82d2-4361-9564-985766ed8290/0x1b7ff50 for node (null)/3232261594 (3 total)
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 crmd[6928]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Dec 18 14:10:50 bl460g1n7 crmd[6928]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261594
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261594
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 crmd[6928]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261594] - state is now member (was (null))
Dec 18 14:10:50 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/29, version=0.3.6)
Dec 18 14:10:50 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/30, version=0.3.7)
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:50 bl460g1n7 pacemakerd[6919]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261594] - state is now member (was (null))
Dec 18 14:10:50 bl460g1n7 crmd[6928]:     info: crm_cs_flush: Sent 2 CPG messages  (0 remaining, last=11): OK (1)
Dec 18 14:10:50 bl460g1n7 cib[6923]:     info: crm_cs_flush: Sent 2 CPG messages  (0 remaining, last=11): OK (1)
Dec 18 14:10:52 bl460g1n7 pacemakerd[6919]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Dec 18 14:10:52 bl460g1n7 pacemakerd[6919]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: pcmk_cpg_membership: Joined[1.0] stonith-ng.3232261592 
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: crm_get_peer: Created entry dc82c9cc-440a-4f8c-8686-379167bc9664/0x196a850 for node (null)/3232261592 (2 total)
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: pcmk_cpg_membership: Member[1.0] stonith-ng.3232261592 
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: pcmk_cpg_membership: Member[1.1] stonith-ng.3232261593 
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: pcmk_cpg_membership: Joined[1.0] cib.3232261592 
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: crm_get_peer: Created entry a874702b-51cb-4ce0-863f-0f646c3f3e65/0x2765340 for node (null)/3232261592 (2 total)
Dec 18 14:10:52 bl460g1n7 attrd[6926]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 attrd[6926]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 attrd[6926]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261592] - state is now member (was (null))
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: pcmk_cpg_membership: Joined[2.0] stonith-ng.3232261594 
Dec 18 14:10:52 bl460g1n7 cib[6923]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 cib[6923]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: pcmk_cpg_membership: Member[1.0] cib.3232261592 
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: pcmk_cpg_membership: Member[1.1] cib.3232261593 
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: pcmk_cpg_membership: Joined[2.0] cib.3232261594 
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: pcmk_cpg_membership: Member[2.0] stonith-ng.3232261592 
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: pcmk_cpg_membership: Member[2.1] stonith-ng.3232261593 
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: crm_get_peer: Created entry e341cf95-f046-4278-9a45-fa4013e2aa74/0x1966270 for node (null)/3232261594 (3 total)
Dec 18 14:10:52 bl460g1n7 attrd[6926]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 attrd[6926]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 cib[6923]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: pcmk_cpg_membership: Member[2.2] stonith-ng.3232261594 
Dec 18 14:10:52 bl460g1n7 stonith-ng[6924]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Dec 18 14:10:52 bl460g1n7 cib[6923]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: pcmk_cpg_membership: Member[2.0] cib.3232261592 
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: pcmk_cpg_membership: Member[2.1] cib.3232261593 
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: crm_get_peer: Created entry 76fdef63-c364-4caa-a319-f807600dae82/0x27653b0 for node (null)/3232261594 (3 total)
Dec 18 14:10:52 bl460g1n7 attrd[6926]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:52 bl460g1n7 attrd[6926]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:52 bl460g1n7 attrd[6926]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261594] - state is now member (was (null))
Dec 18 14:10:52 bl460g1n7 cib[6923]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:52 bl460g1n7 cib[6923]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: pcmk_cpg_membership: Member[2.2] cib.3232261594 
Dec 18 14:10:52 bl460g1n7 cib[6923]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: pcmk_cpg_membership: Joined[1.0] crmd.3232261592 
Dec 18 14:10:53 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:53 bl460g1n7 crmd[6928]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: pcmk_cpg_membership: Member[1.0] crmd.3232261592 
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: pcmk_cpg_membership: Member[1.1] crmd.3232261593 
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: pcmk_cpg_membership: Joined[2.0] crmd.3232261594 
Dec 18 14:10:53 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:53 bl460g1n7 crmd[6928]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: pcmk_cpg_membership: Member[2.0] crmd.3232261592 
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: pcmk_cpg_membership: Member[2.1] crmd.3232261593 
Dec 18 14:10:53 bl460g1n7 crmd[6928]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:53 bl460g1n7 crmd[6928]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: pcmk_cpg_membership: Member[2.2] crmd.3232261594 
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Dec 18 14:10:53 bl460g1n7 stonith-ng[6924]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Dec 18 14:10:53 bl460g1n7 stonith-ng[6924]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: peer_update_callback: bl460g1n8 is now member
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Dec 18 14:10:53 bl460g1n7 crmd[6928]:     info: peer_update_callback: bl460g1n6 is now member
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: throttle_send_command: Updated throttle state to 0000
Dec 18 14:10:54 bl460g1n7 crmd[6928]:   notice: do_state_transition: State transition S_IDLE -> S_INTEGRATION [ input=I_NODE_JOIN cause=C_HA_MESSAGE origin=route_message ]
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: do_dc_join_offer_one: join-2: Processing join_announce request from bl460g1n8 in state S_INTEGRATION
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: join_make_offer: Making join offers based on membership 8
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: join_make_offer: join-2: Sending offer to bl460g1n8
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n8[3232261594] - join-2 phase 0 -> 1
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n7[3232261593] - join-2 phase 4 -> 0
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: join_make_offer: join-2: Sending offer to bl460g1n7
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n7[3232261593] - join-2 phase 0 -> 1
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: abort_transition_graph: do_dc_join_offer_one:236 - Triggered transition abort (complete=1) : Node join
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: abort_transition_graph: do_te_invoke:158 - Triggered transition abort (complete=1) : Peer Halt
Dec 18 14:10:54 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/31, version=0.3.7)
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n7[3232261593] - join-2 phase 1 -> 2
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: do_dc_join_offer_one: join-2: Processing join_announce request from bl460g1n6 in state S_INTEGRATION
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: join_make_offer: join-2: Sending offer to bl460g1n6
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n6[3232261592] - join-2 phase 0 -> 1
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n7[3232261593] - join-2 phase 2 -> 0
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: join_make_offer: join-2: Sending offer to bl460g1n7
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n7[3232261593] - join-2 phase 0 -> 1
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: abort_transition_graph: do_dc_join_offer_one:236 - Triggered transition abort (complete=1) : Node join
Dec 18 14:10:54 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/32, version=0.3.7)
Dec 18 14:10:54 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n7[3232261593] - join-2 phase 1 -> 2
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n8[3232261594] - join-2 phase 1 -> 2
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_expected: do_dc_join_filter_offer: Node bl460g1n8[3232261594] - expected state is now member (was (null))
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n6[3232261592] - join-2 phase 1 -> 2
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_expected: do_dc_join_filter_offer: Node bl460g1n6[3232261592] - expected state is now member (was (null))
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crmd_join_phase_log: join-2: bl460g1n7=integrated
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crmd_join_phase_log: join-2: bl460g1n6=integrated
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crmd_join_phase_log: join-2: bl460g1n8=integrated
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: do_dc_join_finalize: join-2: Syncing our CIB to the rest of the cluster
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_sync operation for section 'all': OK (rc=0, origin=local/crmd/33, version=0.3.7)
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: finalize_join_for: Node bl460g1n7[3232261593] - join-2 phase 2 -> 3
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: finalize_join_for: Node bl460g1n6[3232261592] - join-2 phase 2 -> 3
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: finalize_join_for: Node bl460g1n8[3232261594] - join-2 phase 2 -> 3
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/34, version=0.3.7)
Dec 18 14:10:55 bl460g1n7 cib[6923]:   notice: cib:diff: Diff: --- 0.3.7
Dec 18 14:10:55 bl460g1n7 cib[6923]:   notice: cib:diff: Diff: +++ 0.4.1 e7d74f3ee87880fb031556177c1864a4
Dec 18 14:10:55 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="3" num_updates="7"/>
Dec 18 14:10:55 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <node id="3232261592" uname="bl460g1n6"/>
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/35, version=0.4.1)
Dec 18 14:10:55 bl460g1n7 cib[6923]:   notice: log_cib_diff: cib:diff: Local-only Change: 0.5.1
Dec 18 14:10:55 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="4" num_updates="1"/>
Dec 18 14:10:55 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <node id="3232261594" uname="bl460g1n8"/>
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/36, version=0.5.1)
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_ack: Node bl460g1n7[3232261593] - join-2 phase 3 -> 4
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: do_dc_join_ack: join-2: Updating node state to member for bl460g1n7
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/lrm
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=local/crmd/37, version=0.5.2)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n8']/transient_attributes: OK (rc=0, origin=bl460g1n8/crmd/8, version=0.5.2)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 5 with 2 changes for terminate, id=<n/a>, set=(null)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/38, version=0.5.3)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 6 with 2 changes for shutdown, id=<n/a>, set=(null)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/5, version=0.5.4)
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_ack: Node bl460g1n8[3232261594] - join-2 phase 3 -> 4
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: do_dc_join_ack: join-2: Updating node state to member for bl460g1n8
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n8']/lrm
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/6, version=0.5.4)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 6 for shutdown[bl460g1n7]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 6 for shutdown[bl460g1n8]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n8']/lrm: OK (rc=0, origin=local/crmd/39, version=0.5.4)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/40, version=0.5.5)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 5 for terminate[bl460g1n7]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 5 for terminate[bl460g1n8]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_ack: Node bl460g1n6[3232261592] - join-2 phase 3 -> 4
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: do_dc_join_ack: join-2: Updating node state to member for bl460g1n6
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 7 with 3 changes for terminate, id=<n/a>, set=(null)
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n6']/lrm
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/7, version=0.5.6)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n6']/lrm: OK (rc=0, origin=local/crmd/41, version=0.5.6)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 8 with 3 changes for shutdown, id=<n/a>, set=(null)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/42, version=0.5.7)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/8, version=0.5.7)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 8 for shutdown[bl460g1n6]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 8 for shutdown[bl460g1n7]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 8 for shutdown[bl460g1n8]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 7 for terminate[bl460g1n6]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 7 for terminate[bl460g1n7]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 7 for terminate[bl460g1n8]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n6']/transient_attributes: OK (rc=0, origin=bl460g1n6/crmd/8, version=0.5.8)
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: abort_transition_graph: do_te_invoke:151 - Triggered transition abort (complete=1) : Peer Cancelled
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/43, version=0.5.8)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/44, version=0.5.8)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section cib: OK (rc=0, origin=local/crmd/45, version=0.5.8)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/46, version=0.5.8)
Dec 18 14:10:55 bl460g1n7 pengine[6927]:    error: unpack_resources: Resource start-up disabled since no STONITH resources have been defined
Dec 18 14:10:55 bl460g1n7 pengine[6927]:    error: unpack_resources: Either configure some or disable STONITH with the stonith-enabled option
Dec 18 14:10:55 bl460g1n7 pengine[6927]:    error: unpack_resources: NOTE: Clusters with shared data need STONITH to ensure data integrity
Dec 18 14:10:55 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n7 is active
Dec 18 14:10:55 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n7 is online
Dec 18 14:10:55 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n6 is active
Dec 18 14:10:55 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n6 is online
Dec 18 14:10:55 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n8 is active
Dec 18 14:10:55 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n8 is online
Dec 18 14:10:55 bl460g1n7 pengine[6927]:   notice: stage6: Delaying fencing operations until there are resources to manage
Dec 18 14:10:55 bl460g1n7 pengine[6927]:   notice: process_pe_message: Calculated Transition 1: /var/lib/pacemaker/pengine/pe-input-1.bz2
Dec 18 14:10:55 bl460g1n7 pengine[6927]:   notice: process_pe_message: Configuration ERRORs found during PE processing.  Please run "crm_verify -L" to identify issues.
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: do_te_invoke: Processing graph 1 (ref=pe_calc-dc-1387343455-24) derived from /var/lib/pacemaker/pengine/pe-input-1.bz2
Dec 18 14:10:55 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 4: probe_complete probe_complete on bl460g1n8 - no waiting
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: te_rsc_command: Action 4 confirmed - no wait
Dec 18 14:10:55 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 2: probe_complete probe_complete on bl460g1n6 - no waiting
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: te_rsc_command: Action 2 confirmed - no wait
Dec 18 14:10:55 bl460g1n7 crmd[6928]:   notice: run_graph: Transition 1 (Complete=2, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-1.bz2): Complete
Dec 18 14:10:55 bl460g1n7 crmd[6928]:     info: do_log: FSA: Input I_TE_SUCCESS from notify_crmd() received in state S_TRANSITION_ENGINE
Dec 18 14:10:55 bl460g1n7 crmd[6928]:   notice: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 9 with 2 changes for probe_complete, id=<n/a>, set=(null)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/9, version=0.5.9)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 9 for probe_complete[bl460g1n6]=true: OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 9 for probe_complete[bl460g1n7]=true: OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 9 for probe_complete[bl460g1n8]=(null): OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 10 with 3 changes for probe_complete, id=<n/a>, set=(null)
Dec 18 14:10:55 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/10, version=0.5.10)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 10 for probe_complete[bl460g1n6]=true: OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 10 for probe_complete[bl460g1n7]=true: OK (0)
Dec 18 14:10:55 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 10 for probe_complete[bl460g1n8]=true: OK (0)
Dec 18 14:10:55 bl460g1n7 cib[6945]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-2.raw
Dec 18 14:10:55 bl460g1n7 cib[6945]:     info: write_cib_contents: Wrote version 0.5.0 of the CIB to disk (digest: 911c877da49466ea19e31eeb70a3aeba)
Dec 18 14:10:55 bl460g1n7 cib[6945]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.Xmv6Z4 (digest: /var/lib/pacemaker/cib/cib.Se3Ly2)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_replace_notify: Replaced: 0.5.10 -> 0.6.1 from bl460g1n6
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: Diff: --- 0.5.10
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: Diff: +++ 0.6.1 951cd61c854d70a1576a4c4eca5f5237
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: --         <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.1.11-0.4.ce5d77c.git.el6-ce5d77c"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: --         <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="corosync"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair name="no-quorum-policy" value="freeze" id="cib-bootstrap-options-no-quorum-policy"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair name="stonith-enabled" value="true" id="cib-bootstrap-options-stonith-enabled"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair name="startup-fencing" value="false" id="cib-bootstrap-options-startup-fencing"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair name="crmd-transition-delay" value="2s" id="cib-bootstrap-options-crmd-transition-delay"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <primitive id="prmDummy" class="ocf" provider="pacemaker" type="Dummy"/>
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_replaced_cb: Updating all attributes after cib_refresh_notify event
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <primitive id="prmVM2" class="ocf" provider="heartbeat" type="VirtualDomain">
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 11 with 3 changes for shutdown, id=<n/a>, set=(null)
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <meta_attributes id="prmVM2-meta_attributes">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <nvpair name="allow-migrate" value="true" id="prmVM2-meta_attributes-allow-migrate"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </meta_attributes>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <instance_attributes id="prmVM2-instance_attributes">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <nvpair name="config" value="/migrate_test/config/vm2.xml" id="prmVM2-instance_attributes-config"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <nvpair name="hypervisor" value="qemu:///system" id="prmVM2-instance_attributes-hypervisor"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <nvpair name="migration_transport" value="ssh" id="prmVM2-instance_attributes-migration_transport"/>
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 12 with 3 changes for terminate, id=<n/a>, set=(null)
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </instance_attributes>
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 13 with 3 changes for probe_complete, id=<n/a>, set=(null)
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <operations>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <op name="start" interval="0s" timeout="120s" on-fail="restart" id="prmVM2-start-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <op name="monitor" interval="10s" timeout="30s" on-fail="restart" id="prmVM2-monitor-10s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <op name="stop" interval="0s" timeout="120s" on-fail="fence" id="prmVM2-stop-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <op name="migrate_to" interval="0s" timeout="120s" on-fail="restart" id="prmVM2-migrate_to-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <op name="migrate_from" interval="0s" timeout="120s" on-fail="restart" id="prmVM2-migrate_from-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </operations>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </primitive>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <group id="grpStonith6">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <primitive id="prmStonith6" class="stonith" type="external/ipmi">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <instance_attributes id="prmStonith6-instance_attributes">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="pcmk_reboot_retries" value="1" id="prmStonith6-instance_attributes-pcmk_reboot_retries"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="pcmk_reboot_timeout" value="60s" id="prmStonith6-instance_attributes-pcmk_reboot_timeout"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="hostname" value="bl460g1n6" id="prmStonith6-instance_attributes-hostname"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="ipaddr" value="192.168.133.236" id="prmStonith6-instance_attributes-ipaddr"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="userid" value="USERID" id="prmStonith6-instance_attributes-userid"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="passwd" value="PASSW0RD" id="prmStonith6-instance_attributes-passwd"/>
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: abort_transition_graph: te_update_diff:126 - Triggered transition abort (complete=1, node=, tag=diff, id=(null), magic=NA, cib=0.6.1) : Non-status change
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:     info: stonith_level_remove: Node bl460g1n6 not found (0 active entries)
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:     info: stonith_level_register: Node bl460g1n6 has 1 active fencing levels
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:     info: stonith_level_remove: Node bl460g1n7 not found (1 active entries)
Dec 18 14:11:01 bl460g1n7 crmd[6928]:   notice: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="interface" value="lanplus" id="prmStonith6-instance_attributes-interface"/>
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:     info: stonith_level_register: Node bl460g1n7 has 1 active fencing levels
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_POLICY_ENGINE -> S_ELECTION [ input=I_ELECTION cause=C_FSA_INTERNAL origin=do_cib_replaced ]
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: update_dc: Unset DC. Was bl460g1n7
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           </instance_attributes>
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:     info: stonith_level_remove: Node bl460g1n8 not found (2 active entries)
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <operations>
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:     info: stonith_level_register: Node bl460g1n8 has 1 active fencing levels
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:  warning: handle_startup_fencing: Blind faith: not fencing unseen nodes
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:     info: cib_device_update: Device prmStonith6 is allowed on bl460g1n7: score=0
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="start" interval="0s" timeout="60s" on-fail="restart" id="prmStonith6-start-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="monitor" interval="3600s" timeout="60s" on-fail="restart" id="prmStonith6-monitor-3600s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="stop" interval="0s" timeout="60s" on-fail="ignore" id="prmStonith6-stop-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           </operations>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </primitive>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </group>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <group id="grpStonith7">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <primitive id="prmStonith7" class="stonith" type="external/ipmi">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <instance_attributes id="prmStonith7-instance_attributes">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="pcmk_reboot_retries" value="1" id="prmStonith7-instance_attributes-pcmk_reboot_retries"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="pcmk_reboot_timeout" value="60s" id="prmStonith7-instance_attributes-pcmk_reboot_timeout"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="hostname" value="bl460g1n7" id="prmStonith7-instance_attributes-hostname"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="ipaddr" value="192.168.133.237" id="prmStonith7-instance_attributes-ipaddr"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="userid" value="USERID" id="prmStonith7-instance_attributes-userid"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="passwd" value="PASSW0RD" id="prmStonith7-instance_attributes-passwd"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="interface" value="lanplus" id="prmStonith7-instance_attributes-interface"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           </instance_attributes>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <operations>
Dec 18 14:11:01 bl460g1n7 stonith-ng[6924]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="start" interval="0s" timeout="60s" on-fail="restart" id="prmStonith7-start-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="monitor" interval="3600s" timeout="60s" on-fail="restart" id="prmStonith7-monitor-3600s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="stop" interval="0s" timeout="60s" on-fail="ignore" id="prmStonith7-stop-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           </operations>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </primitive>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </group>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <group id="grpStonith8">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <primitive id="prmStonith8" class="stonith" type="external/ipmi">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <instance_attributes id="prmStonith8-instance_attributes">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="pcmk_reboot_retries" value="1" id="prmStonith8-instance_attributes-pcmk_reboot_retries"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="pcmk_reboot_timeout" value="60s" id="prmStonith8-instance_attributes-pcmk_reboot_timeout"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="hostname" value="bl460g1n8" id="prmStonith8-instance_attributes-hostname"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="ipaddr" value="192.168.133.238" id="prmStonith8-instance_attributes-ipaddr"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="userid" value="USERID" id="prmStonith8-instance_attributes-userid"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="passwd" value="PASSW0RD" id="prmStonith8-instance_attributes-passwd"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="interface" value="lanplus" id="prmStonith8-instance_attributes-interface"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           </instance_attributes>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <operations>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="start" interval="0s" timeout="60s" on-fail="restart" id="prmStonith8-start-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="monitor" interval="3600s" timeout="60s" on-fail="restart" id="prmStonith8-monitor-3600s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="stop" interval="0s" timeout="60s" on-fail="ignore" id="prmStonith8-stop-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           </operations>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </primitive>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </group>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <clone id="clnPing">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <primitive id="prmPing" class="ocf" provider="pacemaker" type="ping">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <instance_attributes id="prmPing-instance_attributes">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="name" value="default_ping_set" id="prmPing-instance_attributes-name"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="host_list" value="192.168.201.254" id="prmPing-instance_attributes-host_list"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="multiplier" value="100" id="prmPing-instance_attributes-multiplier"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="attempts" value="2" id="prmPing-instance_attributes-attempts"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <nvpair name="timeout" value="2" id="prmPing-instance_attributes-timeout"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           </instance_attributes>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <operations>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="start" interval="0s" timeout="60s" on-fail="restart" id="prmPing-start-0s"/>
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: election_complete: Election election-0 complete
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="monitor" interval="10s" timeout="60s" on-fail="restart" id="prmPing-monitor-10s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++             <op name="stop" interval="0s" timeout="60s" on-fail="ignore" id="prmPing-stop-0s"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           </operations>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </primitive>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </clone>
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: election_timeout_popped: Election failed: Declaring ourselves the winner
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <rsc_location id="l1" rsc="prmDummy">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="200" id="l1-rule">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="#uname" operation="eq" value="bl460g1n6" id="l1-expression"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_log: FSA: Input I_ELECTION_DC from election_timeout_popped() received in state S_ELECTION
Dec 18 14:11:01 bl460g1n7 crmd[6928]:   notice: do_state_transition: State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC cause=C_TIMER_POPPED origin=election_timeout_popped ]
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_dc_takeover: Taking over DC status for this partition
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="300" id="l1-rule-0">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="#uname" operation="eq" value="bl460g1n7" id="l1-expression-0"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="200" id="l1-rule-1">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="#uname" operation="eq" value="bl460g1n8" id="l1-expression-1"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </rsc_location>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <rsc_location id="l2" rsc="prmVM2">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="200" id="l2-rule">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="#uname" operation="eq" value="bl460g1n6" id="l2-expression"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="300" id="l2-rule-0">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="#uname" operation="eq" value="bl460g1n7" id="l2-expression-0"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="200" id="l2-rule-1">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="#uname" operation="eq" value="bl460g1n8" id="l2-expression-1"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="-INFINITY" boolean-op="or" id="l2-rule-2">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression operation="not_defined" attribute="default_ping_set" id="l2-expression-2"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="default_ping_set" operation="lt" value="100" id="l2-expression-3"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </rsc_location>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <rsc_location id="l3" rsc="grpStonith6">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="-INFINITY" id="l3-rule">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="#uname" operation="eq" value="bl460g1n6" id="l3-expression"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </rsc_location>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <rsc_location id="l4" rsc="grpStonith7">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="-INFINITY" id="l4-rule">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="#uname" operation="eq" value="bl460g1n7" id="l4-expression"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </rsc_location>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <rsc_location id="l5" rsc="grpStonith8">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <rule score="-INFINITY" id="l5-rule">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <expression attribute="#uname" operation="eq" value="bl460g1n8" id="l5-expression"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </rule>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </rsc_location>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <rsc_colocation id="c4" score="INFINITY" rsc="prmVM2" with-rsc="clnPing"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <rsc_order id="o4" score="0" first="clnPing" then="prmVM2" symmetrical="false"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++     <fencing-topology>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <fencing-level target="bl460g1n6" devices="prmStonith6" index="1" id="fencing"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <fencing-level target="bl460g1n7" devices="prmStonith7" index="1" id="fencing-0"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <fencing-level target="bl460g1n8" devices="prmStonith8" index="1" id="fencing-1"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++     </fencing-topology>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++     <rsc_defaults>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <meta_attributes id="rsc-options">
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair name="resource-stickiness" value="INFINITY" id="rsc-options-resource-stickiness"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair name="migration-threshold" value="1" id="rsc-options-migration-threshold"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++       </meta_attributes>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++     </rsc_defaults>
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_replace operation for section 'all': OK (rc=0, origin=bl460g1n6/cibadmin/2, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/47, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/48, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/49, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/50, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_master operation for section 'all': OK (rc=0, origin=local/crmd/51, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/11, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 11 for shutdown[bl460g1n6]=(null): OK (0)
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 11 for shutdown[bl460g1n7]=(null): OK (0)
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 11 for shutdown[bl460g1n8]=(null): OK (0)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/12, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 12 for terminate[bl460g1n6]=(null): OK (0)
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 12 for terminate[bl460g1n7]=(null): OK (0)
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 12 for terminate[bl460g1n8]=(null): OK (0)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/13, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 13 for probe_complete[bl460g1n6]=true: OK (0)
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 13 for probe_complete[bl460g1n7]=true: OK (0)
Dec 18 14:11:01 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 13 for probe_complete[bl460g1n8]=true: OK (0)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section cib: OK (rc=0, origin=local/crmd/52, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/crm_config//cluster_property_set//nvpair[@name='dc-version']: No such device or address (rc=-6, origin=local/crmd/53, version=0.6.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: log_cib_diff: cib:diff: Local-only Change: 0.7.1
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="6" num_updates="1"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.1.11-0.4.ce5d77c.git.el6-ce5d77c"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section crm_config: OK (rc=0, origin=local/crmd/54, version=0.7.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/crm_config//cluster_property_set//nvpair[@name='cluster-infrastructure']: No such device or address (rc=-6, origin=local/crmd/55, version=0.7.1)
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: initialize_join: Node bl460g1n7[3232261593] - join-3 phase 4 -> 0
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: initialize_join: Node bl460g1n6[3232261592] - join-3 phase 4 -> 0
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: initialize_join: Node bl460g1n8[3232261594] - join-3 phase 4 -> 0
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: join_make_offer: join-3: Sending offer to bl460g1n7
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n7[3232261593] - join-3 phase 0 -> 1
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: join_make_offer: join-3: Sending offer to bl460g1n6
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n6[3232261592] - join-3 phase 0 -> 1
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: join_make_offer: join-3: Sending offer to bl460g1n8
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n8[3232261594] - join-3 phase 0 -> 1
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_dc_join_offer_all: join-3: Waiting on 3 outstanding join acks
Dec 18 14:11:01 bl460g1n7 crmd[6928]:  warning: do_log: FSA: Input I_ELECTION_DC from do_election_check() received in state S_INTEGRATION
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: initialize_join: Node bl460g1n7[3232261593] - join-4 phase 1 -> 0
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: initialize_join: Node bl460g1n6[3232261592] - join-4 phase 1 -> 0
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: initialize_join: Node bl460g1n8[3232261594] - join-4 phase 1 -> 0
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: join_make_offer: join-4: Sending offer to bl460g1n7
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n7[3232261593] - join-4 phase 0 -> 1
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: join_make_offer: join-4: Sending offer to bl460g1n6
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n6[3232261592] - join-4 phase 0 -> 1
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: join_make_offer: join-4: Sending offer to bl460g1n8
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n8[3232261594] - join-4 phase 0 -> 1
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_dc_join_offer_all: join-4: Waiting on 3 outstanding join acks
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: log_cib_diff: cib:diff: Local-only Change: 0.8.1
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="7" num_updates="1"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="corosync"/>
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section crm_config: OK (rc=0, origin=local/crmd/56, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: update_dc: Set DC to bl460g1n7 (3.0.8)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/57, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/58, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/59, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/60, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/61, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n7[3232261593] - join-4 phase 1 -> 2
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n8[3232261594] - join-4 phase 1 -> 2
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n6[3232261592] - join-4 phase 1 -> 2
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crmd_join_phase_log: join-4: bl460g1n7=integrated
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crmd_join_phase_log: join-4: bl460g1n6=integrated
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crmd_join_phase_log: join-4: bl460g1n8=integrated
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_dc_join_finalize: join-4: Syncing our CIB to the rest of the cluster
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_sync operation for section 'all': OK (rc=0, origin=local/crmd/62, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: finalize_join_for: Node bl460g1n7[3232261593] - join-4 phase 2 -> 3
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: finalize_join_for: Node bl460g1n6[3232261592] - join-4 phase 2 -> 3
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: finalize_join_for: Node bl460g1n8[3232261594] - join-4 phase 2 -> 3
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/63, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/64, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/65, version=0.8.1)
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_ack: Node bl460g1n7[3232261593] - join-4 phase 3 -> 4
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_dc_join_ack: join-4: Updating node state to member for bl460g1n7
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/lrm
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=local/crmd/66, version=0.8.2)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/67, version=0.8.3)
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_ack: Node bl460g1n6[3232261592] - join-4 phase 3 -> 4
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_dc_join_ack: join-4: Updating node state to member for bl460g1n6
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n6']/lrm
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n6']/lrm: OK (rc=0, origin=local/crmd/68, version=0.8.4)
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: crm_update_peer_join: do_dc_join_ack: Node bl460g1n8[3232261594] - join-4 phase 3 -> 4
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_dc_join_ack: join-4: Updating node state to member for bl460g1n8
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n8']/lrm
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/69, version=0.8.5)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n8']/lrm: OK (rc=0, origin=local/crmd/70, version=0.8.6)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/71, version=0.8.7)
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 14:11:01 bl460g1n7 crmd[6928]:     info: abort_transition_graph: do_te_invoke:151 - Triggered transition abort (complete=1) : Peer Cancelled
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/72, version=0.8.7)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/73, version=0.8.7)
Dec 18 14:11:01 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section cib: OK (rc=0, origin=local/crmd/74, version=0.8.7)
Dec 18 14:11:01 bl460g1n7 cib[6953]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-3.raw
Dec 18 14:11:01 bl460g1n7 cib[6953]:     info: write_cib_contents: Wrote version 0.6.0 of the CIB to disk (digest: fd2d05bae499007bfa64d61b5c3ed77a)
Dec 18 14:11:01 bl460g1n7 cib[6953]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.xXK5kj (digest: /var/lib/pacemaker/cib/cib.oPaWex)
Dec 18 14:11:01 bl460g1n7 cib[6954]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-4.raw
Dec 18 14:11:01 bl460g1n7 cib[6954]:     info: write_cib_contents: Wrote version 0.8.0 of the CIB to disk (digest: 28a0078bb392734599b693ae67bffcfd)
Dec 18 14:11:01 bl460g1n7 cib[6954]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.VbQ9kt (digest: /var/lib/pacemaker/cib/cib.QiFwyH)
Dec 18 14:11:02 bl460g1n7 stonith-ng[6924]:   notice: stonith_device_register: Added 'prmStonith6' to the device list (1 active devices)
Dec 18 14:11:02 bl460g1n7 stonith-ng[6924]:     info: cib_device_update: Device prmStonith7 has been disabled on bl460g1n7: score=-INFINITY
Dec 18 14:11:02 bl460g1n7 stonith-ng[6924]:     info: cib_device_update: Device prmStonith8 is allowed on bl460g1n7: score=0
Dec 18 14:11:02 bl460g1n7 stonith-ng[6924]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:03 bl460g1n7 stonith-ng[6924]:   notice: stonith_device_register: Added 'prmStonith8' to the device list (2 active devices)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped (2000ms)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/75, version=0.8.7)
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n7 is active
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n7 is online
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n6 is active
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n6 is online
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n8 is active
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n8 is online
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: native_print: prmDummy	(ocf::pacemaker:Dummy):	Stopped 
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: native_print: prmVM2	(ocf::heartbeat:VirtualDomain):	Stopped 
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith6
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith6	(stonith:external/ipmi):	Stopped 
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith7
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith7	(stonith:external/ipmi):	Stopped 
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith8
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith8	(stonith:external/ipmi):	Stopped 
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: clone_print:  Clone Set: clnPing [prmPing]
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: short_print:      Stopped: [ bl460g1n6 bl460g1n7 bl460g1n8 ]
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: native_color: Resource prmVM2 cannot run anywhere
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (3600s) for prmStonith6 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (3600s) for prmStonith7 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (3600s) for prmStonith8 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (10s) for prmPing:0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (10s) for prmPing:1 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (10s) for prmPing:2 on bl460g1n7
Dec 18 14:11:03 bl460g1n7 pengine[6927]:   notice: LogActions: Start   prmDummy	(bl460g1n7)
Dec 18 14:11:03 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmVM2	(Stopped)
Dec 18 14:11:03 bl460g1n7 pengine[6927]:   notice: LogActions: Start   prmStonith6	(bl460g1n8)
Dec 18 14:11:03 bl460g1n7 pengine[6927]:   notice: LogActions: Start   prmStonith7	(bl460g1n6)
Dec 18 14:11:03 bl460g1n7 pengine[6927]:   notice: LogActions: Start   prmStonith8	(bl460g1n6)
Dec 18 14:11:03 bl460g1n7 pengine[6927]:   notice: LogActions: Start   prmPing:0	(bl460g1n6)
Dec 18 14:11:03 bl460g1n7 pengine[6927]:   notice: LogActions: Start   prmPing:1	(bl460g1n8)
Dec 18 14:11:03 bl460g1n7 pengine[6927]:   notice: LogActions: Start   prmPing:2	(bl460g1n7)
Dec 18 14:11:03 bl460g1n7 pengine[6927]:   notice: process_pe_message: Calculated Transition 2: /var/lib/pacemaker/pengine/pe-input-2.bz2
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_te_invoke: Processing graph 2 (ref=pe_calc-dc-1387343463-40) derived from /var/lib/pacemaker/pengine/pe-input-2.bz2
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 18: monitor prmDummy_monitor_0 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 11: monitor prmDummy_monitor_0 on bl460g1n7 (local)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_get_rsc_info: Resource 'prmDummy' not found (0 active resources)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_rsc_register: Added 'prmDummy' to the rsc list (1 active resources)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=11:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmDummy_monitor_0
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 4: monitor prmDummy_monitor_0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 19: monitor prmVM2_monitor_0 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 12: monitor prmVM2_monitor_0 on bl460g1n7 (local)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_get_rsc_info: Resource 'prmVM2' not found (1 active resources)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_rsc_register: Added 'prmVM2' to the rsc list (2 active resources)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=12:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmVM2_monitor_0
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 5: monitor prmVM2_monitor_0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 20: monitor prmStonith6_monitor_0 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 13: monitor prmStonith6_monitor_0 on bl460g1n7 (local)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith6' not found (2 active resources)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_rsc_register: Added 'prmStonith6' to the rsc list (3 active resources)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=13:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith6_monitor_0
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 6: monitor prmStonith6_monitor_0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 21: monitor prmStonith7_monitor_0 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 14: monitor prmStonith7_monitor_0 on bl460g1n7 (local)
Dec 18 14:11:03 bl460g1n7 stonith-ng[6924]:     info: crm_client_new: Connecting 0x1af6200 for uid=0 gid=0 pid=6925 id=ea4ca969-959d-4849-97b4-5ffa6ca0370a
Dec 18 14:11:03 bl460g1n7 stonith-ng[6924]:     info: stonith_command: Processed register from lrmd.6925: OK (0)
Dec 18 14:11:03 bl460g1n7 stonith-ng[6924]:     info: stonith_command: Processed st_notify from lrmd.6925: OK (0)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith7' not found (3 active resources)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_rsc_register: Added 'prmStonith7' to the rsc list (4 active resources)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=14:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith7_monitor_0
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 7: monitor prmStonith7_monitor_0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 22: monitor prmStonith8_monitor_0 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 15: monitor prmStonith8_monitor_0 on bl460g1n7 (local)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith8' not found (4 active resources)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_rsc_register: Added 'prmStonith8' to the rsc list (5 active resources)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=15:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith8_monitor_0
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 8: monitor prmStonith8_monitor_0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 9: monitor prmPing:0_monitor_0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 23: monitor prmPing:1_monitor_0 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 16: monitor prmPing:2_monitor_0 on bl460g1n7 (local)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_get_rsc_info: Resource 'prmPing' not found (5 active resources)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_get_rsc_info: Resource 'prmPing:2' not found (5 active resources)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: process_lrmd_rsc_register: Added 'prmPing' to the rsc list (6 active resources)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=16:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmPing_monitor_0
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: process_lrm_event: LRM operation prmStonith6_monitor_0 (call=13, rc=7, cib-update=76, confirmed=true) not running
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: process_lrm_event: LRM operation prmStonith7_monitor_0 (call=17, rc=7, cib-update=77, confirmed=true) not running
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: process_lrm_event: LRM operation prmStonith8_monitor_0 (call=21, rc=7, cib-update=78, confirmed=true) not running
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/76, version=0.8.8)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/77, version=0.8.9)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/78, version=0.8.10)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n8/crmd/15, version=0.8.11)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n8/crmd/16, version=0.8.12)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: services_os_action_execute: Managed Dummy_meta-data_0 process 6983 exited with rc=0
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: process_lrm_event: LRM operation prmDummy_monitor_0 (call=5, rc=7, cib-update=79, confirmed=true) not running
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith6_monitor_0 (13) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith7_monitor_0 (14) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith8_monitor_0 (15) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith6_monitor_0 (20) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith7_monitor_0 (21) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/79, version=0.8.13)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n8/crmd/17, version=0.8.14)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n8/crmd/18, version=0.8.15)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: services_os_action_execute: Managed ping_meta-data_0 process 6987 exited with rc=0
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: process_lrm_event: LRM operation prmPing_monitor_0 (call=26, rc=7, cib-update=80, confirmed=true) not running
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmDummy_monitor_0 (11) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmDummy_monitor_0 (18) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith8_monitor_0 (22) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmPing_monitor_0 (16) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/80, version=0.8.16)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith6_monitor_0 (6) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:03 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n6/crmd/15, version=0.8.17)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith7_monitor_0 (7) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmDummy_monitor_0 (4) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith8_monitor_0 (8) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmPing_monitor_0 (23) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmVM2_monitor_0 (19) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 17: probe_complete probe_complete on bl460g1n8 - no waiting
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: te_rsc_command: Action 17 confirmed - no wait
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmPing_monitor_0 (9) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:   notice: operation_finished: prmVM2_monitor_0:6958:stderr [ error: failed to get domain 'vm2' ]
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:   notice: operation_finished: prmVM2_monitor_0:6958:stderr [ error: Domain not found: no domain with matching name 'vm2' ]
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:   notice: operation_finished: prmVM2_monitor_0:6958:stderr [ error: failed to get domain 'vm2' ]
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:   notice: operation_finished: prmVM2_monitor_0:6958:stderr [ error: Domain not found: no domain with matching name 'vm2' ]
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: services_os_action_execute: Managed VirtualDomain_meta-data_0 process 7019 exited with rc=0
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: process_lrm_event: LRM operation prmVM2_monitor_0 (call=9, rc=7, cib-update=81, confirmed=true) not running
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmVM2_monitor_0 (5) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 3: probe_complete probe_complete on bl460g1n6 - no waiting
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: te_rsc_command: Action 3 confirmed - no wait
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmVM2_monitor_0 (12) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 10: probe_complete probe_complete on bl460g1n7 (local) - no waiting
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: te_rsc_command: Action 10 confirmed - no wait
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 24: start prmDummy_start_0 on bl460g1n7 (local)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=24:2:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmDummy_start_0
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: log_execute: executing - rsc:prmDummy action:start call_id:27
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 25: start prmStonith6_start_0 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 31: start prmStonith7_start_0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 37: start prmStonith8_start_0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 43: start prmPing:0_start_0 on bl460g1n6
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 45: start prmPing:1_start_0 on bl460g1n8
Dec 18 14:11:03 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 47: start prmPing:2_start_0 on bl460g1n7 (local)
Dec 18 14:11:03 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=47:2:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmPing_start_0
Dec 18 14:11:03 bl460g1n7 lrmd[6925]:     info: log_execute: executing - rsc:prmPing action:start call_id:28
Dec 18 14:11:04 bl460g1n7 lrmd[6925]:     info: log_finished: finished - rsc:prmDummy action:start call_id:27 pid:7025 exit-code:0 exec-time:13ms queue-time:0ms
Dec 18 14:11:04 bl460g1n7 crmd[6928]:   notice: process_lrm_event: LRM operation prmDummy_start_0 (call=27, rc=0, cib-update=82, confirmed=true) ok
Dec 18 14:11:04 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmDummy_start_0 (24) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:05 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 14 with 1 changes for default_ping_set, id=<n/a>, set=(null)
Dec 18 14:11:05 bl460g1n7 lrmd[6925]:     info: log_finished: finished - rsc:prmPing action:start call_id:28 pid:7026 exit-code:0 exec-time:1054ms queue-time:0ms
Dec 18 14:11:05 bl460g1n7 crmd[6928]:   notice: process_lrm_event: LRM operation prmPing_start_0 (call=28, rc=0, cib-update=83, confirmed=true) ok
Dec 18 14:11:05 bl460g1n7 crmd[6928]:     info: abort_transition_graph: te_update_diff:172 - Triggered transition abort (complete=0, node=bl460g1n7, tag=nvpair, id=status-3232261593-default_ping_set, name=default_ping_set, value=100, magic=NA, cib=0.8.27) : Transient attribute: update
Dec 18 14:11:05 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmPing_start_0 (47) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:05 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 14 for default_ping_set[bl460g1n7]=100: OK (0)
Dec 18 14:11:06 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 15 with 2 changes for default_ping_set, id=<n/a>, set=(null)
Dec 18 14:11:06 bl460g1n7 crmd[6928]:     info: abort_transition_graph: te_update_diff:172 - Triggered transition abort (complete=0, node=bl460g1n8, tag=nvpair, id=status-3232261594-default_ping_set, name=default_ping_set, value=100, magic=NA, cib=0.8.29) : Transient attribute: update
Dec 18 14:11:06 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmPing_start_0 (45) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:06 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 15 for default_ping_set[bl460g1n7]=100: OK (0)
Dec 18 14:11:06 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 15 for default_ping_set[bl460g1n8]=100: OK (0)
Dec 18 14:11:06 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith6_start_0 (25) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:07 bl460g1n7 attrd[6926]:   notice: write_attribute: Sent update 16 with 3 changes for default_ping_set, id=<n/a>, set=(null)
Dec 18 14:11:07 bl460g1n7 crmd[6928]:     info: abort_transition_graph: te_update_diff:172 - Triggered transition abort (complete=0, node=bl460g1n6, tag=nvpair, id=status-3232261592-default_ping_set, name=default_ping_set, value=100, magic=NA, cib=0.8.32) : Transient attribute: update
Dec 18 14:11:07 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/16, version=0.8.32)
Dec 18 14:11:07 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmPing_start_0 (43) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:07 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n6/crmd/21, version=0.8.33)
Dec 18 14:11:07 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 16 for default_ping_set[bl460g1n6]=100: OK (0)
Dec 18 14:11:07 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 16 for default_ping_set[bl460g1n7]=100: OK (0)
Dec 18 14:11:07 bl460g1n7 attrd[6926]:   notice: attrd_cib_callback: Update 16 for default_ping_set[bl460g1n8]=100: OK (0)
Dec 18 14:11:07 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith8_start_0 (37) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:07 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n6/crmd/22, version=0.8.34)
Dec 18 14:11:07 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith7_start_0 (31) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:07 bl460g1n7 crmd[6928]:   notice: run_graph: Transition 2 (Complete=34, Pending=0, Fired=0, Skipped=9, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-2.bz2): Stopped
Dec 18 14:11:07 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n6/crmd/23, version=0.8.35)
Dec 18 14:11:09 bl460g1n7 crmd[6928]:     info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped (2000ms)
Dec 18 14:11:09 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Dec 18 14:11:09 bl460g1n7 crmd[6928]:     info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/84, version=0.8.35)
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n7 is active
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n7 is online
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n6 is active
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n6 is online
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n8 is active
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n8 is online
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: native_print: prmDummy	(ocf::pacemaker:Dummy):	Started bl460g1n7 
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: native_print: prmVM2	(ocf::heartbeat:VirtualDomain):	Stopped 
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith6
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith6	(stonith:external/ipmi):	Started bl460g1n8 
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith7
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith7	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith8
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith8	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: clone_print:  Clone Set: clnPing [prmPing]
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: short_print:      Started: [ bl460g1n6 bl460g1n7 bl460g1n8 ]
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (10s) for prmVM2 on bl460g1n7
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (3600s) for prmStonith6 on bl460g1n8
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (3600s) for prmStonith7 on bl460g1n6
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (3600s) for prmStonith8 on bl460g1n6
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (10s) for prmPing:0 on bl460g1n7
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (10s) for prmPing:1 on bl460g1n6
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (10s) for prmPing:2 on bl460g1n8
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmDummy	(Started bl460g1n7)
Dec 18 14:11:09 bl460g1n7 pengine[6927]:   notice: LogActions: Start   prmVM2	(bl460g1n7)
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmStonith6	(Started bl460g1n8)
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmStonith7	(Started bl460g1n6)
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmStonith8	(Started bl460g1n6)
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmPing:0	(Started bl460g1n7)
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmPing:1	(Started bl460g1n6)
Dec 18 14:11:09 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmPing:2	(Started bl460g1n8)
Dec 18 14:11:09 bl460g1n7 pengine[6927]:   notice: process_pe_message: Calculated Transition 3: /var/lib/pacemaker/pengine/pe-input-3.bz2
Dec 18 14:11:09 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Dec 18 14:11:09 bl460g1n7 crmd[6928]:     info: do_te_invoke: Processing graph 3 (ref=pe_calc-dc-1387343469-69) derived from /var/lib/pacemaker/pengine/pe-input-3.bz2
Dec 18 14:11:09 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 8: start prmVM2_start_0 on bl460g1n7 (local)
Dec 18 14:11:09 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=8:3:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmVM2_start_0
Dec 18 14:11:09 bl460g1n7 lrmd[6925]:     info: log_execute: executing - rsc:prmVM2 action:start call_id:29
Dec 18 14:11:09 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 12: monitor prmStonith6_monitor_3600000 on bl460g1n8
Dec 18 14:11:09 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 19: monitor prmStonith7_monitor_3600000 on bl460g1n6
Dec 18 14:11:09 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 26: monitor prmStonith8_monitor_3600000 on bl460g1n6
Dec 18 14:11:09 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 33: monitor prmPing_monitor_10000 on bl460g1n7 (local)
Dec 18 14:11:09 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=33:3:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmPing_monitor_10000
Dec 18 14:11:09 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 36: monitor prmPing_monitor_10000 on bl460g1n6
Dec 18 14:11:09 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 39: monitor prmPing_monitor_10000 on bl460g1n8
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x2769300 for uid=0 gid=0 pid=7197 id=437e2811-8a13-42c9-998e-c7ca51ac4cd3
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.35)
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x2769300 for uid=0 gid=0 pid=7199 id=7e9c02c1-d09c-4143-98dd-4d6d282fd809
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.35)
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM2"]/utilization//nvpair[@name="cpu"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.8.35)
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.8.35)
Dec 18 14:11:09 bl460g1n7 crmd[6928]:     info: abort_transition_graph: te_update_diff:126 - Triggered transition abort (complete=0, node=, tag=diff, id=(null), magic=NA, cib=0.9.1) : Non-status change
Dec 18 14:11:09 bl460g1n7 cib[6923]:   notice: cib:diff: Diff: --- 0.8.35
Dec 18 14:11:09 bl460g1n7 cib[6923]:   notice: cib:diff: Diff: +++ 0.9.1 bab8a310cb4cee0f66f25156ae3a1c3f
Dec 18 14:11:09 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="8" num_updates="35"/>
Dec 18 14:11:09 bl460g1n7 cib[6923]:   notice: cib:diff: ++         <utilization id="prmVM2-utilization">
Dec 18 14:11:09 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <nvpair id="prmVM2-utilization-cpu" name="cpu" value="1"/>
Dec 18 14:11:09 bl460g1n7 cib[6923]:   notice: cib:diff: ++         </utilization>
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section resources: OK (rc=0, origin=local/crm_resource/5, version=0.9.1)
Dec 18 14:11:09 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:10 bl460g1n7 cib[7200]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-5.raw
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x2769300 for uid=0 gid=0 pid=7206 id=adab7ca0-763c-473a-b115-c9af64f7494a
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.9.1)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x2769300 for uid=0 gid=0 pid=7208 id=adc01dbc-e9a7-4812-86f3-d4ae53409d61
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.9.1)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM2"]/utilization//nvpair[@name="hv_memory"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.9.1)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.9.1)
Dec 18 14:11:10 bl460g1n7 crmd[6928]:     info: abort_transition_graph: te_update_diff:126 - Triggered transition abort (complete=0, node=, tag=diff, id=(null), magic=NA, cib=0.10.1) : Non-status change
Dec 18 14:11:10 bl460g1n7 cib[6923]:   notice: log_cib_diff: cib:diff: Local-only Change: 0.10.1
Dec 18 14:11:10 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="9" num_updates="1"/>
Dec 18 14:11:10 bl460g1n7 cib[6923]:   notice: cib:diff: ++           <nvpair id="prmVM2-utilization-hv_memory" name="hv_memory" value="2048"/>
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section resources: OK (rc=0, origin=local/crm_resource/5, version=0.10.1)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:10 bl460g1n7 lrmd[6925]:     info: log_finished: finished - rsc:prmVM2 action:start call_id:29 pid:7048 exit-code:0 exec-time:914ms queue-time:0ms
Dec 18 14:11:10 bl460g1n7 crmd[6928]:   notice: process_lrm_event: LRM operation prmVM2_start_0 (call=29, rc=0, cib-update=85, confirmed=true) ok
Dec 18 14:11:10 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmVM2_start_0 (8) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/85, version=0.10.2)
Dec 18 14:11:10 bl460g1n7 cib[7200]:     info: write_cib_contents: Wrote version 0.9.0 of the CIB to disk (digest: 089b299028d92f831b25092e9d16d596)
Dec 18 14:11:10 bl460g1n7 cib[7200]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.jgjnoK (digest: /var/lib/pacemaker/cib/cib.p1wcJk)
Dec 18 14:11:10 bl460g1n7 cib[7209]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-6.raw
Dec 18 14:11:10 bl460g1n7 crmd[6928]:   notice: process_lrm_event: LRM operation prmPing_monitor_10000 (call=30, rc=0, cib-update=86, confirmed=false) ok
Dec 18 14:11:10 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmPing_monitor_10000 (33) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/86, version=0.10.3)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n8/crmd/23, version=0.10.4)
Dec 18 14:11:10 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmPing_monitor_10000 (39) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:10 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmPing_monitor_10000 (36) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n6/crmd/24, version=0.10.5)
Dec 18 14:11:10 bl460g1n7 cib[7209]:     info: write_cib_contents: Wrote version 0.10.0 of the CIB to disk (digest: 14865fa78a03d2bdea5eb0d59a80267c)
Dec 18 14:11:10 bl460g1n7 cib[7209]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.okf6tU (digest: /var/lib/pacemaker/cib/cib.E9oDbv)
Dec 18 14:11:10 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith6_monitor_3600000 (12) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n8/crmd/24, version=0.10.6)
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n6/crmd/25, version=0.10.7)
Dec 18 14:11:10 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith7_monitor_3600000 (19) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:10 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmStonith8_monitor_3600000 (26) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:10 bl460g1n7 crmd[6928]:   notice: run_graph: Transition 3 (Complete=7, Pending=0, Fired=0, Skipped=1, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-3.bz2): Stopped
Dec 18 14:11:10 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n6/crmd/26, version=0.10.8)
Dec 18 14:11:12 bl460g1n7 crmd[6928]:     info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped (2000ms)
Dec 18 14:11:12 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Dec 18 14:11:12 bl460g1n7 crmd[6928]:     info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Dec 18 14:11:12 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/87, version=0.10.8)
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n7 is active
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n7 is online
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n6 is active
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n6 is online
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n8 is active
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n8 is online
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: native_print: prmDummy	(ocf::pacemaker:Dummy):	Started bl460g1n7 
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: native_print: prmVM2	(ocf::heartbeat:VirtualDomain):	Started bl460g1n7 
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith6
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith6	(stonith:external/ipmi):	Started bl460g1n8 
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith7
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith7	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith8
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith8	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: clone_print:  Clone Set: clnPing [prmPing]
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: short_print:      Started: [ bl460g1n6 bl460g1n7 bl460g1n8 ]
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (10s) for prmVM2 on bl460g1n7
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmDummy	(Started bl460g1n7)
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmVM2	(Started bl460g1n7)
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmStonith6	(Started bl460g1n8)
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmStonith7	(Started bl460g1n6)
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmStonith8	(Started bl460g1n6)
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmPing:0	(Started bl460g1n7)
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmPing:1	(Started bl460g1n6)
Dec 18 14:11:12 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmPing:2	(Started bl460g1n8)
Dec 18 14:11:12 bl460g1n7 pengine[6927]:   notice: process_pe_message: Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-4.bz2
Dec 18 14:11:12 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Dec 18 14:11:12 bl460g1n7 crmd[6928]:     info: do_te_invoke: Processing graph 4 (ref=pe_calc-dc-1387343472-77) derived from /var/lib/pacemaker/pengine/pe-input-4.bz2
Dec 18 14:11:12 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 16: monitor prmVM2_monitor_10000 on bl460g1n7 (local)
Dec 18 14:11:12 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=16:4:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmVM2_monitor_10000
Dec 18 14:11:12 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x27cc2f0 for uid=0 gid=0 pid=7247 id=c50581b3-b779-4967-a7c3-07eb5e2462a8
Dec 18 14:11:12 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.10.8)
Dec 18 14:11:12 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:12 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x27cc2f0 for uid=0 gid=0 pid=7253 id=b7f41f7d-e6f6-46e4-90d3-3f0c5b29ad2c
Dec 18 14:11:12 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.10.8)
Dec 18 14:11:12 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:12 bl460g1n7 crmd[6928]:   notice: process_lrm_event: LRM operation prmVM2_monitor_10000 (call=31, rc=0, cib-update=88, confirmed=false) ok
Dec 18 14:11:12 bl460g1n7 crmd[6928]:     info: match_graph_event: Action prmVM2_monitor_10000 (16) confirmed on bl460g1n7 (rc=0)
Dec 18 14:11:12 bl460g1n7 crmd[6928]:   notice: run_graph: Transition 4 (Complete=1, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-4.bz2): Complete
Dec 18 14:11:12 bl460g1n7 crmd[6928]:     info: do_log: FSA: Input I_TE_SUCCESS from notify_crmd() received in state S_TRANSITION_ENGINE
Dec 18 14:11:12 bl460g1n7 crmd[6928]:   notice: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Dec 18 14:11:12 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/88, version=0.10.9)
Dec 18 14:11:22 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x2769300 for uid=0 gid=0 pid=7319 id=04c09a50-de28-4207-aaf9-e0e1c287bf6e
Dec 18 14:11:22 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.10.9)
Dec 18 14:11:22 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:22 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x2769300 for uid=0 gid=0 pid=7325 id=0339a7fb-2432-457c-84c2-64e16cdd0b39
Dec 18 14:11:22 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.10.9)
Dec 18 14:11:22 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:32 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x2769300 for uid=0 gid=0 pid=7377 id=0e7969dc-532d-41f4-b9b3-f8491707c1dc
Dec 18 14:11:32 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.10.9)
Dec 18 14:11:32 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:32 bl460g1n7 cib[6923]:     info: crm_client_new: Connecting 0x2769300 for uid=0 gid=0 pid=7383 id=ea2e1cbc-01bb-49ed-9045-97516b7224c6
Dec 18 14:11:32 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.10.9)
Dec 18 14:11:32 bl460g1n7 cib[6923]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:34 bl460g1n7 crmd[6928]:     info: abort_transition_graph: te_update_diff:126 - Triggered transition abort (complete=1, node=, tag=diff, id=(null), magic=NA, cib=0.11.1) : Non-status change
Dec 18 14:11:34 bl460g1n7 cib[6923]:   notice: cib:diff: Diff: --- 0.10.9
Dec 18 14:11:34 bl460g1n7 cib[6923]:   notice: cib:diff: Diff: +++ 0.11.1 e30c22d35c3d338660160834fdf936e5
Dec 18 14:11:34 bl460g1n7 cib[6923]:   notice: cib:diff: -- <cib admin_epoch="0" epoch="10" num_updates="9"/>
Dec 18 14:11:34 bl460g1n7 cib[6923]:   notice: cib:diff: ++       <rsc_location id="cli-ban-prmVM2-on-bl460g1n7" rsc="prmVM2" role="Started" node="bl460g1n7" score="-INFINITY"/>
Dec 18 14:11:34 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_modify operation for section constraints: OK (rc=0, origin=bl460g1n6/crm_resource/3, version=0.11.1)
Dec 18 14:11:34 bl460g1n7 stonith-ng[6924]:     info: update_cib_stonith_devices: Updating device list from the cib: new location constraint
Dec 18 14:11:34 bl460g1n7 stonith-ng[6924]:     info: stonith_device_remove: Removed 'prmStonith6' from the device list (1 active devices)
Dec 18 14:11:34 bl460g1n7 stonith-ng[6924]:     info: cib_device_update: Device prmStonith6 is allowed on bl460g1n7: score=0
Dec 18 14:11:34 bl460g1n7 stonith-ng[6924]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:34 bl460g1n7 cib[7384]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-7.raw
Dec 18 14:11:34 bl460g1n7 cib[7384]:     info: write_cib_contents: Wrote version 0.11.0 of the CIB to disk (digest: 81971b854de85601a20f4f47275deb7a)
Dec 18 14:11:34 bl460g1n7 cib[7384]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.4jnSSj (digest: /var/lib/pacemaker/cib/cib.9qeITY)
Dec 18 14:11:35 bl460g1n7 stonith-ng[6924]:   notice: stonith_device_register: Added 'prmStonith6' to the device list (2 active devices)
Dec 18 14:11:35 bl460g1n7 stonith-ng[6924]:     info: cib_device_update: Device prmStonith7 has been disabled on bl460g1n7: score=-INFINITY
Dec 18 14:11:35 bl460g1n7 stonith-ng[6924]:     info: stonith_device_remove: Removed 'prmStonith8' from the device list (1 active devices)
Dec 18 14:11:35 bl460g1n7 stonith-ng[6924]:     info: cib_device_update: Device prmStonith8 is allowed on bl460g1n7: score=0
Dec 18 14:11:35 bl460g1n7 stonith-ng[6924]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:36 bl460g1n7 crmd[6928]:     info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped (2000ms)
Dec 18 14:11:36 bl460g1n7 crmd[6928]:   notice: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Dec 18 14:11:36 bl460g1n7 crmd[6928]:     info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Dec 18 14:11:36 bl460g1n7 cib[6923]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/89, version=0.11.1)
Dec 18 14:11:36 bl460g1n7 stonith-ng[6924]:   notice: stonith_device_register: Added 'prmStonith8' to the device list (2 active devices)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n7 is active
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n7 is online
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n6 is active
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n6 is online
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: determine_online_status_fencing: Node bl460g1n8 is active
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: determine_online_status: Node bl460g1n8 is online
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: native_print: prmDummy	(ocf::pacemaker:Dummy):	Started bl460g1n7 
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: native_print: prmVM2	(ocf::heartbeat:VirtualDomain):	Started bl460g1n7 
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith6
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith6	(stonith:external/ipmi):	Started bl460g1n8 
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith7
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith7	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: group_print:  Resource Group: grpStonith8
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: native_print:      prmStonith8	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: clone_print:  Clone Set: clnPing [prmPing]
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: short_print:      Started: [ bl460g1n6 bl460g1n7 bl460g1n8 ]
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: RecurringOp:  Start recurring monitor (10s) for prmVM2 on bl460g1n6
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmDummy	(Started bl460g1n7)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: MigrateRsc: Migrating prmVM2 from bl460g1n7 to bl460g1n6
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: MigrateRsc: Repairing c4: prmVM2 == clnPing (1000000)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:   notice: LogActions: Migrate prmVM2	(Started bl460g1n7 -> bl460g1n6)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmStonith6	(Started bl460g1n8)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmStonith7	(Started bl460g1n6)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmStonith8	(Started bl460g1n6)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmPing:0	(Started bl460g1n7)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmPing:1	(Started bl460g1n6)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:     info: LogActions: Leave   prmPing:2	(Started bl460g1n8)
Dec 18 14:11:36 bl460g1n7 pengine[6927]:   notice: process_pe_message: Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-5.bz2
Dec 18 14:11:36 bl460g1n7 crmd[6928]:     info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Dec 18 14:11:36 bl460g1n7 crmd[6928]:     info: do_te_invoke: Processing graph 5 (ref=pe_calc-dc-1387343496-79) derived from /var/lib/pacemaker/pengine/pe-input-5.bz2
Dec 18 14:11:36 bl460g1n7 crmd[6928]:   notice: te_rsc_command: Initiating action 47: migrate_to prmVM2_migrate_to_0 on bl460g1n7 (local)
Dec 18 14:11:36 bl460g1n7 lrmd[6925]:     info: cancel_recurring_action: Cancelling operation prmVM2_monitor_10000
Dec 18 14:11:36 bl460g1n7 crmd[6928]:     info: do_lrm_rsc_op: Performing key=47:5:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmVM2_migrate_to_0
Dec 18 14:11:36 bl460g1n7 lrmd[6925]:     info: log_execute: executing - rsc:prmVM2 action:migrate_to call_id:33
Dec 18 14:11:36 bl460g1n7 crmd[6928]:     info: process_lrm_event: LRM operation prmVM2_monitor_10000 (call=31, status=1, cib-update=0, confirmed=true) Cancelled
Dec 18 14:11:36 bl460g1n7 VirtualDomain(prmVM2)[7387]: INFO: vm2: Starting live migration to bl460g1n6 (using remote hypervisor URI qemu+ssh://bl460g1n6/system ).
Dec 18 14:11:42 bl460g1n7 VirtualDomain(prmVM2)[7387]: INFO: vm2: live migration to bl460g1n6 succeeded.
Dec 18 14:13:57 bl460g1n7 root: Mark:pcmk:1387343637
