Dec 18 14:10:49 bl460g1n6 corosync[17800]:   [MAIN  ] main.c:main:1171 Corosync Cluster Engine ('2.3.2.8-7014f'): started and ready to provide service.
Dec 18 14:10:49 bl460g1n6 corosync[17800]:   [MAIN  ] main.c:main:1172 Corosync built-in features: watchdog upstart snmp pie relro bindnow
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [TOTEM ] totemnet.c:totemnet_instance_initialize:242 Initializing transport (UDP/IP Multicast).
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [TOTEM ] totemcrypto.c:init_nss:579 Initializing transmit/receive security (NSS) crypto: aes256 hash: sha1
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [TOTEM ] totemnet.c:totemnet_instance_initialize:242 Initializing transport (UDP/IP Multicast).
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [TOTEM ] totemcrypto.c:init_nss:579 Initializing transmit/receive security (NSS) crypto: aes256 hash: sha1
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [TOTEM ] totemudp.c:timer_function_netif_check_timeout:670 The network interface [192.168.101.216] is now up.
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync configuration map access [0]
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: cmap
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync configuration service [1]
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: cfg
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync cluster closed process group service v1.01 [2]
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: cpg
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync profile loading service [4]
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [WD    ] wd.c:setup_watchdog:651 Watchdog is now been tickled by corosync.
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [WD    ] wd.c:wd_scan_resources:580 no resources configured.
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync watchdog service [7]
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [QUORUM] vsf_quorum.c:quorum_exec_init_fn:274 Using quorum provider corosync_votequorum
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync vote quorum service v1.0 [5]
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: votequorum
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync cluster quorum service v0.1 [3]
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:377 server name: quorum
Dec 18 14:10:49 bl460g1n6 corosync[17802]:   [TOTEM ] totemudp.c:timer_function_netif_check_timeout:670 The network interface [192.168.102.216] is now up.
Dec 18 14:10:50 bl460g1n6 corosync[17802]:   [TOTEM ] totemsrp.c:memb_state_operational_enter:2016 A new membership (192.168.101.216:4) was formed. Members joined: -1062705704
Dec 18 14:10:50 bl460g1n6 corosync[17802]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[1]: -1062705704
Dec 18 14:10:50 bl460g1n6 corosync[17802]:   [MAIN  ] main.c:corosync_sync_completed:276 Completed service synchronization, ready to provide service.
Dec 18 14:10:50 bl460g1n6 corosync[17802]:   [TOTEM ] totemsrp.c:memb_state_operational_enter:2016 A new membership (192.168.101.216:8) was formed. Members joined: -1062705703 -1062705702
Dec 18 14:10:50 bl460g1n6 corosync[17802]:   [QUORUM] vsf_quorum.c:quorum_api_set_quorum:148 This node is within the primary component and will provide service.
Dec 18 14:10:50 bl460g1n6 corosync[17802]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[1]: -1062705704
Dec 18 14:10:50 bl460g1n6 corosync[17802]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[3]: -1062705704 -1062705703 -1062705702
Dec 18 14:10:50 bl460g1n6 corosync[17802]:   [MAIN  ] main.c:corosync_sync_completed:276 Completed service synchronization, ready to provide service.
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_ipc_connect: Could not establish pacemakerd connection: Connection refused (111)
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: get_cluster_type: Detected an active 'corosync' cluster
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: mcp_read_config: Reading configure for stack: corosync
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: mcp_read_config: Configured corosync to accept connections from group 189: OK (1)
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: main: Starting Pacemaker 1.1.11-0.4.ce5d77c.git.el6 (Build: ce5d77c):  generated-manpages agent-manpages ascii-docs ncurses libqb-logging libqb-ipc lha-fencing nagios  corosync-native snmp
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: main: Maximum core file size is: 18446744073709551615
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: qb_ipcs_us_publish: server name: pacemakerd
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_get_peer: Created entry 35f15fae-6e46-4995-978c-f6b55c7ef23d/0x11bb8d0 for node (null)/3232261592 (1 total)
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261592] - corosync-cpg is now online
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: cluster_connect_quorum: Quorum acquired
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Using uid=189 and group=189 for process cib
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Forked child 17816 for process cib
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Forked child 17817 for process stonith-ng
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Forked child 17818 for process lrmd
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Using uid=189 and group=189 for process attrd
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Forked child 17819 for process attrd
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Using uid=189 and group=189 for process pengine
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Forked child 17820 for process pengine
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Using uid=189 and group=189 for process crmd
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: start_child: Forked child 17821 for process crmd
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: main: Starting mainloop
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: pcmk_quorum_notification: Membership 8: quorum retained (3)
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n6[3232261592] - state is now member (was (null))
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_get_peer: Created entry 266da416-63da-4bb7-847a-65c0393bba76/0x11bcf30 for node (null)/3232261593 (2 total)
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: main: Using new config location: /var/lib/pacemaker/cib
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.xml (digest: /var/lib/pacemaker/cib/cib.xml.sig)
Dec 18 14:10:52 bl460g1n6 cib[17816]:  warning: retrieveCib: Cluster configuration not found: /var/lib/pacemaker/cib/cib.xml
Dec 18 14:10:52 bl460g1n6 cib[17816]:  warning: readCibXmlFile: Primary configuration corrupt or unusable, trying backups in /var/lib/pacemaker/cib
Dec 18 14:10:52 bl460g1n6 cib[17816]:  warning: readCibXmlFile: Continuing with an empty configuration.
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: validate_with_relaxng: Creating RNG parser context
Dec 18 14:10:52 bl460g1n6 lrmd[17818]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Dec 18 14:10:52 bl460g1n6 lrmd[17818]:     info: qb_ipcs_us_publish: server name: lrmd
Dec 18 14:10:52 bl460g1n6 lrmd[17818]:     info: main: Starting
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Dec 18 14:10:52 bl460g1n6 attrd[17819]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Dec 18 14:10:52 bl460g1n6 pengine[17820]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Dec 18 14:10:52 bl460g1n6 pengine[17820]:     info: qb_ipcs_us_publish: server name: pengine
Dec 18 14:10:52 bl460g1n6 pengine[17820]:     info: main: Starting pengine
Dec 18 14:10:52 bl460g1n6 crmd[17821]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Dec 18 14:10:52 bl460g1n6 crmd[17821]:   notice: main: CRM Git Version: ce5d77c
Dec 18 14:10:52 bl460g1n6 crmd[17821]:     info: do_log: FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
Dec 18 14:10:52 bl460g1n6 crmd[17821]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Dec 18 14:10:52 bl460g1n6 crmd[17821]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Dec 18 14:10:52 bl460g1n6 crmd[17821]:     info: crm_ipc_connect: Could not establish cib_shm connection: Connection refused (111)
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261593
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: startCib: CIB Initialization completed successfully
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:     info: crm_get_peer: Created entry 7b3862f1-ca45-42d1-92d2-020cf1562f75/0x23772b0 for node (null)/3232261592 (1 total)
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_get_peer: Created entry ec2e986f-5451-40c3-8874-68b6be0a307b/0x25c60f0 for node (null)/3232261592 (1 total)
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Dec 18 14:10:52 bl460g1n6 attrd[17819]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261592] - corosync-cpg is now online
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:     info: init_cs_connection_once: Connection to 'corosync': established
Dec 18 14:10:52 bl460g1n6 attrd[17819]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 attrd[17819]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261592] - state is now member (was (null))
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261592] - corosync-cpg is now online
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: init_cs_connection_once: Connection to 'corosync': established
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261593] - state is now member (was (null))
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_get_peer: Created entry a01330a9-4d4a-4f42-ab88-62bdd5736df9/0x11bd4e0 for node (null)/3232261594 (3 total)
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Dec 18 14:10:52 bl460g1n6 stonith-ng[17817]:     info: crm_ipc_connect: Could not establish cib_rw connection: Connection refused (111)
Dec 18 14:10:52 bl460g1n6 attrd[17819]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 attrd[17819]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: qb_ipcs_us_publish: server name: cib_ro
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: qb_ipcs_us_publish: server name: cib_rw
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: qb_ipcs_us_publish: server name: cib_shm
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: cib_init: Starting cib mainloop
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Joined[0.0] cib.3232261592 
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Member[0.0] cib.3232261592 
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_get_peer: Created entry 3307ff21-ed49-4dd2-98da-7dec1db5a472/0x25c8800 for node (null)/3232261593 (2 total)
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261594
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Member[0.1] cib.3232261593 
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261593] - corosync-cpg is now online
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Joined[1.0] cib.3232261594 
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Member[1.0] cib.3232261592 
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Member[1.1] cib.3232261593 
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_get_peer: Created entry e31263a0-d5da-440c-9da1-c3c55923860a/0x25c8870 for node (null)/3232261594 (3 total)
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261594] - state is now member (was (null))
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Dec 18 14:10:52 bl460g1n6 pacemakerd[17812]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:52 bl460g1n6 cib[17816]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Member[1.2] cib.3232261594 
Dec 18 14:10:52 bl460g1n6 cib[17816]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Dec 18 14:10:52 bl460g1n6 cib[17822]:     info: write_cib_contents: Wrote version 0.0.0 of the CIB to disk (digest: e25219628aafa5da816bfed98f68664c)
Dec 18 14:10:52 bl460g1n6 cib[17822]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.8uFdVD (digest: /var/lib/pacemaker/cib/cib.15IhyF)
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x25c9070 for uid=189 gid=189 pid=17821 id=7f3e659f-bd3a-42db-a289-d843220237de
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: do_cib_control: CIB connection established
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.0.0)
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: crm_get_peer: Created entry f5b0f1e1-f750-4181-89b7-986d88bb59d4/0x2848a60 for node (null)/3232261592 (1 total)
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261592
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261592] - corosync-cpg is now online
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: init_cs_connection_once: Connection to 'corosync': established
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: peer_update_callback: bl460g1n6 is now (null)
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: cluster_connect_quorum: Quorum acquired
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.0.0)
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: do_ha_control: Connected to the cluster
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: lrmd_ipc_connect: Connecting to lrmd
Dec 18 14:10:53 bl460g1n6 lrmd[17818]:     info: crm_client_new: Connecting 0x22c5990 for uid=189 gid=189 pid=17821 id=45739341-e33f-4f2a-b972-5ddbd43b3990
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: do_lrm_control: LRM connection established
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: do_started: Delaying start, no membership data (0000000000100000)
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.0.0)
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: pcmk_quorum_notification: Membership 8: quorum retained (3)
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n6[3232261592] - state is now member (was (null))
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: peer_update_callback: bl460g1n6 is now member (was (null))
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: crm_get_peer: Created entry d1fff44d-2052-4c4b-a508-10f75f90e43f/0x298fd60 for node (null)/3232261593 (2 total)
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261593
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261593] - state is now member (was (null))
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: crm_get_peer: Created entry 541a083b-6c6f-4774-a8ba-65fcd7c0cd2e/0x298d8d0 for node (null)/3232261594 (3 total)
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x2417bf0 for uid=0 gid=0 pid=17817 id=7c9bf734-5305-456c-8461-93faa07d4b95
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x2649ff0 for uid=189 gid=189 pid=17819 id=772c2573-abe9-450c-b0c6-1941da554825
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261594
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:   notice: setup_cib: Watching for stonith topology changes
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: qb_ipcs_us_publish: server name: stonith-ng
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.0.0)
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: main: Starting stonith-ng mainloop
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: init_cib_cache_cb: Updating device list from the cib: init
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: unpack_nodes: Creating a fake local node
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Joined[0.0] stonith-ng.3232261592 
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Member[0.0] stonith-ng.3232261592 
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: crm_get_peer: Created entry fdd65038-8fe5-4c85-9418-e7f6f4696d22/0x237ffa0 for node (null)/3232261593 (2 total)
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Member[0.1] stonith-ng.3232261593 
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261593] - corosync-cpg is now online
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261593] - state is now member (was (null))
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Joined[1.0] stonith-ng.3232261594 
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Member[1.0] stonith-ng.3232261592 
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Member[1.1] stonith-ng.3232261593 
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: crm_get_peer: Created entry b176d09e-5ba7-40c5-942d-04686fabdcd7/0x2382100 for node (null)/3232261594 (3 total)
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261594] - state is now member (was (null))
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Member[1.2] stonith-ng.3232261594 
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Dec 18 14:10:53 bl460g1n6 stonith-ng[17817]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: do_started: Delaying start, Config not read (0000000000000040)
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: qb_ipcs_us_publish: server name: crmd
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: do_started: The local CRM is operational
Dec 18 14:10:53 bl460g1n6 crmd[17821]:     info: do_log: FSA: Input I_PENDING from do_started() received in state S_STARTING
Dec 18 14:10:53 bl460g1n6 crmd[17821]:   notice: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/5, version=0.0.0)
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261594] - state is now member (was (null))
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: attrd_peer_message: Recorded attribute writer: bl460g1n7
Dec 18 14:10:53 bl460g1n6 attrd[17819]:   notice: attrd_peer_message: Processing sync-response from bl460g1n7
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x264a550 for uid=0 gid=0 pid=3390 id=f542c84f-2cbf-4540-a3a1-d0355b3cc210
Dec 18 14:10:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_mon/7, version=0.0.0)
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Joined[0.0] crmd.3232261592 
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Member[0.0] crmd.3232261592 
Dec 18 14:10:54 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:54 bl460g1n6 crmd[17821]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Member[0.1] crmd.3232261593 
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261593] - corosync-cpg is now online
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Joined[1.0] crmd.3232261594 
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Member[1.0] crmd.3232261592 
Dec 18 14:10:54 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Dec 18 14:10:54 bl460g1n6 crmd[17821]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261593
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Member[1.1] crmd.3232261593 
Dec 18 14:10:54 bl460g1n6 crmd[17821]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Dec 18 14:10:54 bl460g1n6 crmd[17821]:   notice: get_node_name: Could not obtain a node name for corosync nodeid 3232261594
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Member[1.2] crmd.3232261594 
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: peer_update_callback: bl460g1n8 is now member
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: peer_update_callback: bl460g1n7 is now member
Dec 18 14:10:54 bl460g1n6 crmd[17821]:     info: update_dc: Set DC to bl460g1n7 (3.0.8)
Dec 18 14:10:54 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/6, version=0.0.0)
Dec 18 14:10:55 bl460g1n6 stonith-ng[17817]:     info: crm_client_new: Connecting 0x237c2d0 for uid=189 gid=189 pid=17821 id=c3f4e9a6-1580-4e8d-92f9-8ae4508b2b49
Dec 18 14:10:55 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed register from crmd.17821: OK (0)
Dec 18 14:10:55 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_notify from crmd.17821: OK (0)
Dec 18 14:10:55 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_notify from crmd.17821: OK (0)
Dec 18 14:10:55 bl460g1n6 cib[17816]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Dec 18 14:10:55 bl460g1n6 cib[17816]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_replace: Digest matched on replace from bl460g1n7: 19544907710d300373a234bd343458e0
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_replace: Replaced 0.0.0 with 0.3.7 from bl460g1n7
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_replace_notify: Replaced: 0.0.0 -> 0.3.7 from bl460g1n7
Dec 18 14:10:55 bl460g1n6 attrd[17819]:   notice: attrd_cib_replaced_cb: Updating all attributes after cib_refresh_notify event
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_replace operation for section 'all': OK (rc=0, origin=bl460g1n7/crmd/33, version=0.3.7)
Dec 18 14:10:55 bl460g1n6 crmd[17821]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n6']/transient_attributes
Dec 18 14:10:55 bl460g1n6 crmd[17821]:     info: update_attrd_helper: Connecting to attrd... 5 retries remaining
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section nodes: OK (rc=0, origin=bl460g1n7/crmd/35, version=0.4.1)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section nodes: OK (rc=0, origin=bl460g1n7/crmd/36, version=0.5.1)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Dec 18 14:10:55 bl460g1n6 crmd[17821]:     info: do_log: FSA: Input I_NOT_DC from do_cl_join_finalize_respond() received in state S_PENDING
Dec 18 14:10:55 bl460g1n6 crmd[17821]:   notice: do_state_transition: State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=bl460g1n7/crmd/37, version=0.5.2)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/38, version=0.5.3)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/attrd/5, version=0.5.4)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/40, version=0.5.5)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/7, version=0.5.5)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_delete operation for section //node_state[@uname='bl460g1n6']/transient_attributes to master (origin=local/crmd/8)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/attrd/7, version=0.5.6)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/42, version=0.5.7)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n6']/transient_attributes: OK (rc=0, origin=bl460g1n7/crmd/8, version=0.5.8)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/attrd/9, version=0.5.9)
Dec 18 14:10:55 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/attrd/10, version=0.5.10)
Dec 18 14:10:55 bl460g1n6 cib[17825]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-0.raw
Dec 18 14:10:55 bl460g1n6 cib[17825]:     info: write_cib_contents: Wrote version 0.5.0 of the CIB to disk (digest: 911c877da49466ea19e31eeb70a3aeba)
Dec 18 14:10:55 bl460g1n6 cib[17825]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.aR3eEU (digest: /var/lib/pacemaker/cib/cib.2c9yG4)
Dec 18 14:11:00 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x27d0670 for uid=0 gid=0 pid=17830 id=34687eab-a09a-45ff-aa63-03a6fbb204a8
Dec 18 14:11:00 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/cibadmin/2, version=0.5.10)
Dec 18 14:11:00 bl460g1n6 cib[17816]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x27d0670 for uid=0 gid=0 pid=17831 id=c0ca0362-ba63-4b75-ba27-09174005c1f3
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/cibadmin/2, version=0.5.10)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x27d0670 for uid=0 gid=0 pid=17869 id=c7e438c2-33c8-4435-ab0d-810a51664cb3
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_replace operation for section 'all' to master (origin=local/cibadmin/2)
Dec 18 14:11:01 bl460g1n6 crmd[17821]:     info: election_count_vote: Election 3 (owner: 3232261593) lost: vote from bl460g1n7 (Uptime)
Dec 18 14:11:01 bl460g1n6 crmd[17821]:     info: update_dc: Unset DC. Was bl460g1n7
Dec 18 14:11:01 bl460g1n6 crmd[17821]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_NOT_DC
Dec 18 14:11:01 bl460g1n6 crmd[17821]:   notice: do_state_transition: State transition S_NOT_DC -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_election_count_vote ]
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/9, version=0.5.10)
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: stonith_level_remove: Node bl460g1n6 not found (0 active entries)
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: stonith_level_register: Node bl460g1n6 has 1 active fencing levels
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: stonith_level_remove: Node bl460g1n7 not found (1 active entries)
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: stonith_level_register: Node bl460g1n7 has 1 active fencing levels
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: stonith_level_remove: Node bl460g1n8 not found (2 active entries)
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: stonith_level_register: Node bl460g1n8 has 1 active fencing levels
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section 'all': OK (rc=0, origin=bl460g1n7/cibadmin/2, version=0.6.1)
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:  warning: handle_startup_fencing: Blind faith: not fencing unseen nodes
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/10, version=0.6.1)
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: cib_device_update: Device prmStonith6 has been disabled on bl460g1n6: score=-INFINITY
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: cib_device_update: Device prmStonith7 is allowed on bl460g1n6: score=0
Dec 18 14:11:01 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:01 bl460g1n6 crmd[17821]:     info: update_dc: Set DC to bl460g1n7 (3.0.8)
Dec 18 14:11:01 bl460g1n6 crmd[17821]:     info: election_count_vote: Election 4 (owner: 3232261593) lost: vote from bl460g1n7 (Uptime)
Dec 18 14:11:01 bl460g1n6 crmd[17821]:     info: update_dc: Unset DC. Was bl460g1n7
Dec 18 14:11:01 bl460g1n6 crmd[17821]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_PENDING
Dec 18 14:11:01 bl460g1n6 crmd[17821]:     info: update_dc: Set DC to bl460g1n7 (3.0.8)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n7/crmd/54, version=0.7.1)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n7/crmd/56, version=0.8.1)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/11, version=0.8.1)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/12, version=0.8.1)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/13, version=0.8.1)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/14, version=0.8.1)
Dec 18 14:11:01 bl460g1n6 crmd[17821]:     info: do_log: FSA: Input I_NOT_DC from do_cl_join_finalize_respond() received in state S_PENDING
Dec 18 14:11:01 bl460g1n6 crmd[17821]:   notice: do_state_transition: State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_replace: Digest matched on replace from bl460g1n7: 2d6c960d5289b5475275e8aa894291b1
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_replace: Replaced 0.8.1 with 0.8.1 from bl460g1n7
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_replace operation for section 'all': OK (rc=0, origin=bl460g1n7/crmd/62, version=0.8.1)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x264aab0 for uid=0 gid=0 pid=17871 id=ac5e854b-c202-42dd-8f63-efab8de623b9
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=bl460g1n7/crmd/66, version=0.8.2)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/67, version=0.8.3)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n6']/lrm: OK (rc=0, origin=bl460g1n7/crmd/68, version=0.8.4)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/69, version=0.8.5)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n8']/lrm: OK (rc=0, origin=bl460g1n7/crmd/70, version=0.8.6)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/71, version=0.8.7)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/cibadmin/2, version=0.8.7)
Dec 18 14:11:01 bl460g1n6 cib[17816]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:01 bl460g1n6 cib[17872]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-1.raw
Dec 18 14:11:01 bl460g1n6 cib[17872]:     info: write_cib_contents: Wrote version 0.8.0 of the CIB to disk (digest: 28a0078bb392734599b693ae67bffcfd)
Dec 18 14:11:01 bl460g1n6 cib[17872]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.ZYqbP4 (digest: /var/lib/pacemaker/cib/cib.9aJ45u)
Dec 18 14:11:01 bl460g1n6 cib[17873]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-2.raw
Dec 18 14:11:01 bl460g1n6 cib[17873]:     info: write_cib_contents: Wrote version 0.8.0 of the CIB to disk (digest: 28a0078bb392734599b693ae67bffcfd)
Dec 18 14:11:01 bl460g1n6 cib[17873]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.DS7LUe (digest: /var/lib/pacemaker/cib/cib.D01hxF)
Dec 18 14:11:02 bl460g1n6 stonith-ng[17817]:   notice: stonith_device_register: Added 'prmStonith7' to the device list (1 active devices)
Dec 18 14:11:02 bl460g1n6 stonith-ng[17817]:     info: cib_device_update: Device prmStonith8 is allowed on bl460g1n6: score=0
Dec 18 14:11:02 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:03 bl460g1n6 stonith-ng[17817]:   notice: stonith_device_register: Added 'prmStonith8' to the device list (2 active devices)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_get_rsc_info: Resource 'prmDummy' not found (0 active resources)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_rsc_register: Added 'prmDummy' to the rsc list (1 active resources)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=4:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmDummy_monitor_0
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_get_rsc_info: Resource 'prmVM2' not found (1 active resources)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_rsc_register: Added 'prmVM2' to the rsc list (2 active resources)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=5:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmVM2_monitor_0
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith6' not found (2 active resources)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_rsc_register: Added 'prmStonith6' to the rsc list (3 active resources)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=6:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith6_monitor_0
Dec 18 14:11:03 bl460g1n6 stonith-ng[17817]:     info: crm_client_new: Connecting 0x250bef0 for uid=0 gid=0 pid=17818 id=76729c4d-d55d-4314-a2ce-df7789da49e1
Dec 18 14:11:03 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed register from lrmd.17818: OK (0)
Dec 18 14:11:03 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_notify from lrmd.17818: OK (0)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith7' not found (3 active resources)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_rsc_register: Added 'prmStonith7' to the rsc list (4 active resources)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=7:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith7_monitor_0
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/76, version=0.8.8)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/77, version=0.8.9)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: process_lrm_event: LRM operation prmStonith6_monitor_0 (call=13, rc=7, cib-update=15, confirmed=true) not running
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: process_lrm_event: LRM operation prmStonith7_monitor_0 (call=17, rc=7, cib-update=16, confirmed=true) not running
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/78, version=0.8.10)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/15, version=0.8.11)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/16, version=0.8.12)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/79, version=0.8.13)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/17, version=0.8.14)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: services_os_action_execute: Managed Dummy_meta-data_0 process 17896 exited with rc=0
Dec 18 14:11:03 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmDummy_monitor_0 (call=5, rc=7, cib-update=17, confirmed=true) not running
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/18, version=0.8.15)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith8' not found (4 active resources)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_rsc_register: Added 'prmStonith8' to the rsc list (5 active resources)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=8:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith8_monitor_0
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: process_lrm_event: LRM operation prmStonith8_monitor_0 (call=21, rc=7, cib-update=18, confirmed=true) not running
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_get_rsc_info: Resource 'prmPing' not found (5 active resources)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/80, version=0.8.16)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_get_rsc_info: Resource 'prmPing:0' not found (5 active resources)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: process_lrmd_rsc_register: Added 'prmPing' to the rsc list (6 active resources)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/15)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=9:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmPing_monitor_0
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/16)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/17)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/18)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/15, version=0.8.17)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/16, version=0.8.18)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/17, version=0.8.19)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/18, version=0.8.20)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/19, version=0.8.21)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: services_os_action_execute: Managed ping_meta-data_0 process 17922 exited with rc=0
Dec 18 14:11:03 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmPing_monitor_0 (call=26, rc=7, cib-update=19, confirmed=true) not running
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/20, version=0.8.22)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/19)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/19, version=0.8.23)
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:   notice: operation_finished: prmVM2_monitor_0:17876:stderr [ error: failed to get domain 'vm2' ]
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:   notice: operation_finished: prmVM2_monitor_0:17876:stderr [ error: Domain not found: no domain with matching name 'vm2' ]
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:   notice: operation_finished: prmVM2_monitor_0:17876:stderr [ error: failed to get domain 'vm2' ]
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:   notice: operation_finished: prmVM2_monitor_0:17876:stderr [ error: Domain not found: no domain with matching name 'vm2' ]
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: services_os_action_execute: Managed VirtualDomain_meta-data_0 process 17937 exited with rc=0
Dec 18 14:11:03 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmVM2_monitor_0 (call=9, rc=7, cib-update=20, confirmed=true) not running
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/20)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/20, version=0.8.24)
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=31:2:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith7_start_0
Dec 18 14:11:03 bl460g1n6 lrmd[17818]:     info: log_execute: executing - rsc:prmStonith7 action:start call_id:27
Dec 18 14:11:03 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=37:2:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith8_start_0
Dec 18 14:11:03 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/81, version=0.8.25)
Dec 18 14:11:03 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/82, version=0.8.26)
Dec 18 14:11:04 bl460g1n6 stonith-ng[17817]:   notice: stonith_device_register: Device 'prmStonith7' already existed in device list (2 active devices)
Dec 18 14:11:04 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_device_register from lrmd.17818: OK (0)
Dec 18 14:11:04 bl460g1n6 lrmd[17818]:     info: log_execute: executing - rsc:prmStonith8 action:start call_id:28
Dec 18 14:11:04 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=43:2:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmPing_start_0
Dec 18 14:11:04 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_execute from lrmd.17818: Operation now in progress (-115)
Dec 18 14:11:04 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:04 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/attrd/14, version=0.8.27)
Dec 18 14:11:04 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/83, version=0.8.28)
Dec 18 14:11:05 bl460g1n6 stonith-ng[17817]:   notice: stonith_device_register: Device 'prmStonith8' already existed in device list (2 active devices)
Dec 18 14:11:05 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_device_register from lrmd.17818: OK (0)
Dec 18 14:11:05 bl460g1n6 lrmd[17818]:     info: log_execute: executing - rsc:prmPing action:start call_id:29
Dec 18 14:11:05 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Dec 18 14:11:05 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_execute from lrmd.17818: Operation now in progress (-115)
Dec 18 14:11:05 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Dec 18 14:11:05 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/attrd/15, version=0.8.29)
Dec 18 14:11:05 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/21, version=0.8.30)
Dec 18 14:11:05 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/22, version=0.8.31)
Dec 18 14:11:06 bl460g1n6 lrmd[17818]:     info: log_finished: finished - rsc:prmPing action:start call_id:29 pid:17945 exit-code:0 exec-time:1045ms queue-time:0ms
Dec 18 14:11:06 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmPing_start_0 (call=29, rc=0, cib-update=21, confirmed=true) ok
Dec 18 14:11:06 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/21)
Dec 18 14:11:06 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/attrd/16, version=0.8.32)
Dec 18 14:11:06 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/21, version=0.8.33)
Dec 18 14:11:06 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith8:17949 [ Performing: stonith -t external/ipmi -S ]
Dec 18 14:11:06 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith8:17949 [ success:  0 ]
Dec 18 14:11:06 bl460g1n6 lrmd[17818]:     info: log_finished: finished - rsc:prmStonith8 action:start call_id:28  exit-code:0 exec-time:2123ms queue-time:0ms
Dec 18 14:11:06 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith7:17946 [ Performing: stonith -t external/ipmi -S ]
Dec 18 14:11:06 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith7:17946 [ success:  0 ]
Dec 18 14:11:06 bl460g1n6 lrmd[17818]:     info: log_finished: finished - rsc:prmStonith7 action:start call_id:27  exit-code:0 exec-time:3126ms queue-time:0ms
Dec 18 14:11:06 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmStonith8_start_0 (call=28, rc=0, cib-update=22, confirmed=true) ok
Dec 18 14:11:06 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/22)
Dec 18 14:11:06 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmStonith7_start_0 (call=27, rc=0, cib-update=23, confirmed=true) ok
Dec 18 14:11:06 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/23)
Dec 18 14:11:06 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/22, version=0.8.34)
Dec 18 14:11:06 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/23, version=0.8.35)
Dec 18 14:11:08 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=19:3:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith7_monitor_3600000
Dec 18 14:11:08 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=26:3:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmStonith8_monitor_3600000
Dec 18 14:11:08 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_execute from lrmd.17818: Operation now in progress (-115)
Dec 18 14:11:08 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Dec 18 14:11:08 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=36:3:0:ddf348fe-fbad-4abb-9a12-8250f71b075a op=prmPing_monitor_10000
Dec 18 14:11:08 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_execute from lrmd.17818: Operation now in progress (-115)
Dec 18 14:11:08 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n7/crm_resource/5, version=0.9.1)
Dec 18 14:11:09 bl460g1n6 cib[18035]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-3.raw
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n7/crm_resource/5, version=0.10.1)
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/85, version=0.10.2)
Dec 18 14:11:09 bl460g1n6 cib[18035]:     info: write_cib_contents: Wrote version 0.9.0 of the CIB to disk (digest: b072d452ad6fa080c4549c6ac0d606c0)
Dec 18 14:11:09 bl460g1n6 cib[18035]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.b8v5QH (digest: /var/lib/pacemaker/cib/cib.Z26vAu)
Dec 18 14:11:09 bl460g1n6 cib[18036]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-4.raw
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/86, version=0.10.3)
Dec 18 14:11:09 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmPing_monitor_10000 (call=32, rc=0, cib-update=24, confirmed=false) ok
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/23, version=0.10.4)
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/24)
Dec 18 14:11:09 bl460g1n6 cib[18036]:     info: write_cib_contents: Wrote version 0.10.0 of the CIB to disk (digest: 67b333859e4181541b5f3c35baa20285)
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/24, version=0.10.5)
Dec 18 14:11:09 bl460g1n6 cib[18036]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.Z0kmXR (digest: /var/lib/pacemaker/cib/cib.Z1sa2E)
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/24, version=0.10.6)
Dec 18 14:11:09 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith7:17992 [ Performing: stonith -t external/ipmi -S ]
Dec 18 14:11:09 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith7:17992 [ success:  0 ]
Dec 18 14:11:09 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmStonith7_monitor_3600000 (call=30, rc=0, cib-update=25, confirmed=false) ok
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/25)
Dec 18 14:11:09 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith8:17994 [ Performing: stonith -t external/ipmi -S ]
Dec 18 14:11:09 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith8:17994 [ success:  0 ]
Dec 18 14:11:09 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmStonith8_monitor_3600000 (call=31, rc=0, cib-update=26, confirmed=false) ok
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/26)
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/25, version=0.10.7)
Dec 18 14:11:09 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/26, version=0.10.8)
Dec 18 14:11:12 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n7/crmd/88, version=0.10.9)
Dec 18 14:11:23 bl460g1n6 crmd[17821]:     info: throttle_send_command: Updated throttle state to 0000
Dec 18 14:11:34 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x27d0670 for uid=0 gid=0 pid=18085 id=f01b0ed5-97b2-4882-ac0c-23d98a6e0b56
Dec 18 14:11:34 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.10.9)
Dec 18 14:11:34 bl460g1n6 cib[17816]:     info: cib_process_request: Forwarding cib_modify operation for section constraints to master (origin=local/crm_resource/3)
Dec 18 14:11:34 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_apply_diff operation for section constraints: OK (rc=0, origin=bl460g1n7/crm_resource/3, version=0.11.1)
Dec 18 14:11:34 bl460g1n6 stonith-ng[17817]:     info: update_cib_stonith_devices: Updating device list from the cib: new location constraint
Dec 18 14:11:34 bl460g1n6 stonith-ng[17817]:     info: cib_device_update: Device prmStonith6 has been disabled on bl460g1n6: score=-INFINITY
Dec 18 14:11:34 bl460g1n6 stonith-ng[17817]:     info: cib_device_update: Device prmStonith7 is allowed on bl460g1n6: score=0
Dec 18 14:11:34 bl460g1n6 cib[17816]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:11:34 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:34 bl460g1n6 stonith-ng[17817]:   notice: stonith_device_register: Device 'prmStonith7' already existed in device list (2 active devices)
Dec 18 14:11:34 bl460g1n6 stonith-ng[17817]:     info: cib_device_update: Device prmStonith8 is allowed on bl460g1n6: score=0
Dec 18 14:11:34 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Dec 18 14:11:34 bl460g1n6 cib[18086]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-5.raw
Dec 18 14:11:34 bl460g1n6 stonith-ng[17817]:   notice: stonith_device_register: Device 'prmStonith8' already existed in device list (2 active devices)
Dec 18 14:11:34 bl460g1n6 cib[18086]:     info: write_cib_contents: Wrote version 0.11.0 of the CIB to disk (digest: 8aa561ad5ad78095666c39e698c55659)
Dec 18 14:11:34 bl460g1n6 cib[18086]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.yrY136 (digest: /var/lib/pacemaker/cib/cib.1x53pY)
Dec 18 14:11:43 bl460g1n6 corosync[17802]:   [TOTEM ] totemsrp.c:timer_function_orf_token_timeout:1689 A processor failed, forming new configuration.
Dec 18 14:11:44 bl460g1n6 corosync[17802]:   [TOTEM ] totemsrp.c:memb_state_operational_enter:2016 A new membership (192.168.101.216:12) was formed. Members left: -1062705703
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Left[2.0] cib.3232261593 
Dec 18 14:11:44 bl460g1n6 attrd[17819]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node bl460g1n7[3232261593] - state is now lost (was member)
Dec 18 14:11:44 bl460g1n6 attrd[17819]:   notice: attrd_peer_remove: Removing all bl460g1n7 attributes for attrd_peer_change_cb
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node bl460g1n7[3232261593] - corosync-cpg is now offline
Dec 18 14:11:44 bl460g1n6 attrd[17819]:   notice: attrd_peer_change_cb: Lost attribute writer bl460g1n7
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Member[2.0] cib.3232261592 
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: pcmk_cpg_membership: Member[2.1] cib.3232261594 
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Left[2.0] crmd.3232261593 
Dec 18 14:11:44 bl460g1n6 pacemakerd[17812]:     info: crm_cs_flush: Sent 0 CPG messages  (1 remaining, last=8): Try again (6)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node bl460g1n7[3232261593] - corosync-cpg is now offline
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: peer_update_callback: Client bl460g1n7/peer now has status [offline] (DC=bl460g1n7)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:   notice: peer_update_callback: Our peer on the DC (bl460g1n7) is dead
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Member[2.0] crmd.3232261592 
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: pcmk_cpg_membership: Member[2.1] crmd.3232261594 
Dec 18 14:11:44 bl460g1n6 crmd[17821]:   notice: do_state_transition: State transition S_NOT_DC -> S_ELECTION [ input=I_ELECTION cause=C_CRMD_STATUS_CALLBACK origin=peer_update_callback ]
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: update_dc: Unset DC. Was bl460g1n7
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_cs_flush: Sent 0 CPG messages  (1 remaining, last=9): Try again (6)
Dec 18 14:11:44 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Left[2.0] stonith-ng.3232261593 
Dec 18 14:11:44 bl460g1n6 stonith-ng[17817]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node bl460g1n7[3232261593] - corosync-cpg is now offline
Dec 18 14:11:44 bl460g1n6 stonith-ng[17817]:     info: crm_cs_flush: Sent 0 CPG messages  (1 remaining, last=4): Try again (6)
Dec 18 14:11:44 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Member[2.0] stonith-ng.3232261592 
Dec 18 14:11:44 bl460g1n6 stonith-ng[17817]:     info: pcmk_cpg_membership: Member[2.1] stonith-ng.3232261594 
Dec 18 14:11:44 bl460g1n6 corosync[17802]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[2]: -1062705704 -1062705702
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: pcmk_quorum_notification: Membership 12: quorum retained (2)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n7[3232261593] - state is now lost (was member)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: peer_update_callback: bl460g1n7 is now lost (was member)
Dec 18 14:11:44 bl460g1n6 pacemakerd[17812]:     info: pcmk_quorum_notification: Membership 12: quorum retained (2)
Dec 18 14:11:44 bl460g1n6 pacemakerd[17812]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n7[3232261593] - state is now lost (was member)
Dec 18 14:11:44 bl460g1n6 corosync[17802]:   [MAIN  ] main.c:corosync_sync_completed:276 Completed service synchronization, ready to provide service.
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_cs_flush: Sent 2 CPG messages  (0 remaining, last=11): OK (1)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: election_count_vote: Election 1 (owner: 3232261594) pass: vote from bl460g1n8 (Uptime)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: election_complete: Election election-0 complete
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: election_timeout_popped: Election failed: Declaring ourselves the winner
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_log: FSA: Input I_ELECTION_DC from election_timeout_popped() received in state S_ELECTION
Dec 18 14:11:44 bl460g1n6 crmd[17821]:   notice: do_state_transition: State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC cause=C_TIMER_POPPED origin=election_timeout_popped ]
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_te_control: Registering TE UUID: eb4921e1-885a-48fd-9b1b-c46cee527da2
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: set_graph_functions: Setting custom graph functions
Dec 18 14:11:44 bl460g1n6 pengine[17820]:     info: crm_client_new: Connecting 0x1e434c0 for uid=189 gid=189 pid=17821 id=a3576366-b312-47ba-b72e-2355a1e898a8
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_dc_takeover: Taking over DC status for this partition
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_readwrite: We are now in R/W mode
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_master operation for section 'all': OK (rc=0, origin=local/crmd/27, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section cib: OK (rc=0, origin=local/crmd/28, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/crm_config//cluster_property_set//nvpair[@name='dc-version']: OK (rc=0, origin=local/crmd/29, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section crm_config: OK (rc=0, origin=local/crmd/30, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/crm_config//cluster_property_set//nvpair[@name='cluster-infrastructure']: OK (rc=0, origin=local/crmd/31, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: join_make_offer: Not making an offer to bl460g1n7: not active (lost)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_expected: join_make_offer: Node bl460g1n7[3232261593] - expected state is now down (was (null))
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: join_make_offer: Making join offers based on membership 12
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: join_make_offer: join-1: Sending offer to bl460g1n8
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n8[3232261594] - join-1 phase 0 -> 1
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: join_make_offer: join-1: Sending offer to bl460g1n6
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n6[3232261592] - join-1 phase 0 -> 1
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_dc_join_offer_all: join-1: Waiting on 2 outstanding join acks
Dec 18 14:11:44 bl460g1n6 crmd[17821]:  warning: do_log: FSA: Input I_ELECTION_DC from do_election_check() received in state S_INTEGRATION
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: initialize_join: Node bl460g1n8[3232261594] - join-2 phase 1 -> 0
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: initialize_join: Node bl460g1n6[3232261592] - join-2 phase 1 -> 0
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: join_make_offer: Not making an offer to bl460g1n7: not active (lost)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: join_make_offer: join-2: Sending offer to bl460g1n8
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n8[3232261594] - join-2 phase 0 -> 1
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: join_make_offer: join-2: Sending offer to bl460g1n6
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: join_make_offer: Node bl460g1n6[3232261592] - join-2 phase 0 -> 1
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_dc_join_offer_all: join-2: Waiting on 2 outstanding join acks
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: update_dc: Set DC to bl460g1n6 (3.0.8)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section crm_config: OK (rc=0, origin=local/crmd/32, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/33, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/34, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/35, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n8[3232261594] - join-2 phase 1 -> 2
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_expected: do_dc_join_filter_offer: Node bl460g1n8[3232261594] - expected state is now member (was (null))
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: do_dc_join_filter_offer: Node bl460g1n6[3232261592] - join-2 phase 1 -> 2
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_expected: do_dc_join_filter_offer: Node bl460g1n6[3232261592] - expected state is now member (was (null))
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_state_transition: State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crmd_join_phase_log: join-2: bl460g1n7=none
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crmd_join_phase_log: join-2: bl460g1n8=integrated
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crmd_join_phase_log: join-2: bl460g1n6=integrated
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_dc_join_finalize: join-2: Syncing our CIB to the rest of the cluster
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_sync operation for section 'all': OK (rc=0, origin=local/crmd/36, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: finalize_join_for: Node bl460g1n8[3232261594] - join-2 phase 2 -> 3
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: finalize_join_for: Node bl460g1n6[3232261592] - join-2 phase 2 -> 3
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/37, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/38, version=0.11.1)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: services_os_action_execute: Managed ping_meta-data_0 process 18230 exited with rc=0
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: do_dc_join_ack: Node bl460g1n6[3232261592] - join-2 phase 3 -> 4
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_dc_join_ack: join-2: Updating node state to member for bl460g1n6
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n6']/lrm
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: crm_update_peer_join: do_dc_join_ack: Node bl460g1n8[3232261594] - join-2 phase 3 -> 4
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_dc_join_ack: join-2: Updating node state to member for bl460g1n8
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n8']/lrm
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n6']/lrm: OK (rc=0, origin=local/crmd/39, version=0.11.2)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/40, version=0.11.3)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n8']/lrm: OK (rc=0, origin=local/crmd/41, version=0.11.4)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/42, version=0.11.5)
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: do_state_transition: State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED cause=C_FSA_INTERNAL origin=check_join_state ]
Dec 18 14:11:44 bl460g1n6 crmd[17821]:     info: abort_transition_graph: do_te_invoke:151 - Triggered transition abort (complete=1) : Peer Cancelled
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/43, version=0.11.5)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/44, version=0.11.6)
Dec 18 14:11:44 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section cib: OK (rc=0, origin=local/crmd/45, version=0.11.7)
Dec 18 14:11:46 bl460g1n6 crmd[17821]:     info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped (2000ms)
Dec 18 14:11:46 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/46, version=0.11.7)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:  warning: handle_startup_fencing: Blind faith: not fencing unseen nodes
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: determine_online_status_fencing: Node bl460g1n6 is active
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: determine_online_status: Node bl460g1n6 is online
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: determine_online_status_fencing: Node bl460g1n8 is active
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: determine_online_status: Node bl460g1n8 is online
Dec 18 14:11:46 bl460g1n6 pengine[17820]:  warning: pe_fence_node: Node bl460g1n7 will be fenced because prmDummy is thought to be active there
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: native_print: prmDummy	(ocf::pacemaker:Dummy):	Started bl460g1n7 
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: native_print: prmVM2	(ocf::heartbeat:VirtualDomain):	Started bl460g1n7 
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: group_print:  Resource Group: grpStonith6
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: native_print:      prmStonith6	(stonith:external/ipmi):	Started bl460g1n8 
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: group_print:  Resource Group: grpStonith7
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: native_print:      prmStonith7	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: group_print:  Resource Group: grpStonith8
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: native_print:      prmStonith8	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: clone_print:  Clone Set: clnPing [prmPing]
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: short_print:      Started: [ bl460g1n6 bl460g1n7 bl460g1n8 ]
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: native_color: Resource prmPing:0 cannot run anywhere
Dec 18 14:11:46 bl460g1n6 pengine[17820]:  warning: custom_action: Action prmDummy_stop_0 on bl460g1n7 is unrunnable (offline)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:  warning: custom_action: Action prmVM2_stop_0 on bl460g1n7 is unrunnable (offline)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: RecurringOp:  Start recurring monitor (10s) for prmVM2 on bl460g1n8
Dec 18 14:11:46 bl460g1n6 pengine[17820]:  warning: custom_action: Action prmPing:0_stop_0 on bl460g1n7 is unrunnable (offline)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:  warning: custom_action: Action prmPing:0_stop_0 on bl460g1n7 is unrunnable (offline)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:  warning: stage6: Scheduling Node bl460g1n7 for STONITH
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: native_stop_constraints: prmDummy_stop_0 is implicit after bl460g1n7 is fenced
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: native_stop_constraints: prmVM2_stop_0 is implicit after bl460g1n7 is fenced
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: native_stop_constraints: prmPing:0_stop_0 is implicit after bl460g1n7 is fenced
Dec 18 14:11:46 bl460g1n6 pengine[17820]:   notice: LogActions: Move    prmDummy	(Started bl460g1n7 -> bl460g1n6)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:   notice: LogActions: Move    prmVM2	(Started bl460g1n7 -> bl460g1n8)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmStonith6	(Started bl460g1n8)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmStonith7	(Started bl460g1n6)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmStonith8	(Started bl460g1n6)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:   notice: LogActions: Stop    prmPing:0	(bl460g1n7)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmPing:1	(Started bl460g1n6)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmPing:2	(Started bl460g1n8)
Dec 18 14:11:46 bl460g1n6 pengine[17820]:  warning: process_pe_message: Calculated Transition 0: /var/lib/pacemaker/pengine/pe-warn-0.bz2
Dec 18 14:11:46 bl460g1n6 crmd[17821]:     info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Dec 18 14:11:46 bl460g1n6 crmd[17821]:     info: do_te_invoke: Processing graph 0 (ref=pe_calc-dc-1387343506-21) derived from /var/lib/pacemaker/pengine/pe-warn-0.bz2
Dec 18 14:11:46 bl460g1n6 crmd[17821]:   notice: te_fence_node: Executing reboot fencing operation (43) on bl460g1n7 (timeout=60000)
Dec 18 14:11:46 bl460g1n6 stonith-ng[17817]:   notice: handle_request: Client crmd.17821.c3f4e9a6 wants to fence (reboot) 'bl460g1n7' with device '(any)'
Dec 18 14:11:46 bl460g1n6 stonith-ng[17817]:   notice: initiate_remote_stonith_op: Initiating remote operation reboot for bl460g1n7: e7460b1d-89f8-4bf3-9015-2e068bee302a (0)
Dec 18 14:11:46 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_fence from crmd.17821: Operation now in progress (-115)
Dec 18 14:11:46 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_query from bl460g1n6: OK (0)
Dec 18 14:11:46 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action list for agent fence_legacy (target=(null))
Dec 18 14:11:46 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action list for agent fence_legacy (target=(null))
Dec 18 14:11:46 bl460g1n6 stonith-ng[17817]:     info: process_remote_stonith_query: Query result 1 of 2 from bl460g1n8 (1 devices)
Dec 18 14:11:46 bl460g1n6 stonith-ng[17817]:     info: call_remote_stonith: Total remote op timeout set to 60 for fencing of node bl460g1n7 for crmd.17821.e7460b1d
Dec 18 14:11:47 bl460g1n6 stonith-ng[17817]:     info: call_remote_stonith: Waiting for additional peers capable of terminating bl460g1n7 with prmStonith7 for crmd.17821.e7460b1d
Dec 18 14:11:47 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_query reply from bl460g1n8: OK (0)
Dec 18 14:11:47 bl460g1n6 stonith-ng[17817]:     info: dynamic_list_search_cb: Refreshing port list for prmStonith7
Dec 18 14:11:47 bl460g1n6 stonith-ng[17817]:     info: dynamic_list_search_cb: Refreshing port list for prmStonith8
Dec 18 14:11:47 bl460g1n6 stonith-ng[17817]:     info: process_remote_stonith_query: Query result 2 of 2 from bl460g1n6 (1 devices)
Dec 18 14:11:47 bl460g1n6 stonith-ng[17817]:     info: call_remote_stonith: Requesting that bl460g1n6 perform op reboot bl460g1n7 with prmStonith7 for crmd.17821 (72s)
Dec 18 14:11:47 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_query reply from bl460g1n6: OK (0)
Dec 18 14:11:47 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_fence from bl460g1n6: Operation now in progress (-115)
Dec 18 14:11:47 bl460g1n6 stonith-ng[17817]:     info: stonith_action_create: Initiating action reboot for agent fence_legacy (target=bl460g1n7)
Dec 18 14:11:48 bl460g1n6 stonith-ng[17817]:   notice: log_operation: Operation 'reboot' [18264] (call 2 from crmd.17821) for host 'bl460g1n7' with device 'prmStonith7' returned: 0 (OK)
Dec 18 14:11:48 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith7:18264 [ Performing: stonith -t external/ipmi -T reset bl460g1n7 ]
Dec 18 14:11:48 bl460g1n6 stonith-ng[17817]:     info: log_operation: prmStonith7:18264 [ success: bl460g1n7 0 ]
Dec 18 14:11:48 bl460g1n6 stonith-ng[17817]:   notice: process_remote_stonith_exec: Call to prmStonith7 for bl460g1n7 on behalf of crmd.17821@bl460g1n6: OK (0)
Dec 18 14:11:48 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_fence reply from bl460g1n6: OK (0)
Dec 18 14:11:48 bl460g1n6 stonith-ng[17817]:   notice: remote_op_done: Operation reboot of bl460g1n7 by bl460g1n6 for crmd.17821@bl460g1n6.e7460b1d: OK
Dec 18 14:11:48 bl460g1n6 stonith-ng[17817]:     info: stonith_command: Processed st_notify reply from bl460g1n6: OK (0)
Dec 18 14:11:48 bl460g1n6 crmd[17821]:   notice: tengine_stonith_callback: Stonith operation 2/43:0:0:eb4921e1-885a-48fd-9b1b-c46cee527da2: OK (0)
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: crm_update_peer_proc: send_stonith_update: Node bl460g1n7[3232261593] - all processes are now offline
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: peer_update_callback: Client bl460g1n7/peer now has status [offline] (DC=true)
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/lrm
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/transient_attributes
Dec 18 14:11:48 bl460g1n6 crmd[17821]:   notice: tengine_stonith_notify: Peer bl460g1n7 was terminated (reboot) by bl460g1n6 for bl460g1n6: OK (ref=e7460b1d-89f8-4bf3-9015-2e068bee302a) by client crmd.17821
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/lrm
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/transient_attributes
Dec 18 14:11:48 bl460g1n6 crmd[17821]:   notice: te_rsc_command: Initiating action 11: start prmDummy_start_0 on bl460g1n6 (local)
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: do_lrm_rsc_op: Performing key=11:0:0:eb4921e1-885a-48fd-9b1b-c46cee527da2 op=prmDummy_start_0
Dec 18 14:11:48 bl460g1n6 lrmd[17818]:     info: log_execute: executing - rsc:prmDummy action:start call_id:33
Dec 18 14:11:48 bl460g1n6 crmd[17821]:   notice: te_rsc_command: Initiating action 13: start prmVM2_start_0 on bl460g1n8
Dec 18 14:11:48 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/47, version=0.11.8)
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: abort_transition_graph: te_update_diff:258 - Triggered transition abort (complete=0, node=bl460g1n7, tag=lrm_rsc_op, id=prmStonith6_last_0, magic=0:7;13:2:7:ddf348fe-fbad-4abb-9a12-8250f71b075a, cib=0.11.9) : Resource op removal
Dec 18 14:11:48 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=local/crmd/48, version=0.11.9)
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: abort_transition_graph: te_update_diff:188 - Triggered transition abort (complete=0, node=bl460g1n7, tag=transient_attributes, id=3232261593, magic=NA, cib=0.11.10) : Transient attribute: removal
Dec 18 14:11:48 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n7']/transient_attributes: OK (rc=0, origin=local/crmd/49, version=0.11.10)
Dec 18 14:11:48 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/50, version=0.11.10)
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: cib_fencing_updated: Fencing update 50 for bl460g1n7: complete
Dec 18 14:11:48 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=local/crmd/51, version=0.11.10)
Dec 18 14:11:48 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_delete operation for section //node_state[@uname='bl460g1n7']/transient_attributes: OK (rc=0, origin=local/crmd/52, version=0.11.10)
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: cib_fencing_updated: Fencing update 47 for bl460g1n7: complete
Dec 18 14:11:48 bl460g1n6 lrmd[17818]:     info: log_finished: finished - rsc:prmDummy action:start call_id:33 pid:18280 exit-code:0 exec-time:21ms queue-time:0ms
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: services_os_action_execute: Managed Dummy_meta-data_0 process 18285 exited with rc=0
Dec 18 14:11:48 bl460g1n6 crmd[17821]:   notice: process_lrm_event: LRM operation prmDummy_start_0 (call=33, rc=0, cib-update=53, confirmed=true) ok
Dec 18 14:11:48 bl460g1n6 crmd[17821]:     info: match_graph_event: Action prmDummy_start_0 (11) confirmed on bl460g1n6 (rc=0)
Dec 18 14:11:48 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/crmd/53, version=0.11.11)
Dec 18 14:11:49 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n8/crmd/28, version=0.11.12)
Dec 18 14:11:49 bl460g1n6 crmd[17821]:     info: match_graph_event: Action prmVM2_start_0 (13) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:49 bl460g1n6 crmd[17821]:   notice: run_graph: Transition 0 (Complete=10, Pending=0, Fired=0, Skipped=1, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-warn-0.bz2): Stopped
Dec 18 14:11:51 bl460g1n6 crmd[17821]:     info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped (2000ms)
Dec 18 14:11:51 bl460g1n6 crmd[17821]:     info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Dec 18 14:11:51 bl460g1n6 crmd[17821]:     info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Dec 18 14:11:51 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/54, version=0.11.12)
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: determine_online_status_fencing: Node bl460g1n6 is active
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: determine_online_status: Node bl460g1n6 is online
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: determine_online_status_fencing: Node bl460g1n8 is active
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: determine_online_status: Node bl460g1n8 is online
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: native_print: prmDummy	(ocf::pacemaker:Dummy):	Started bl460g1n6 
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: native_print: prmVM2	(ocf::heartbeat:VirtualDomain):	Started bl460g1n8 
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: group_print:  Resource Group: grpStonith6
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: native_print:      prmStonith6	(stonith:external/ipmi):	Started bl460g1n8 
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: group_print:  Resource Group: grpStonith7
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: native_print:      prmStonith7	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: group_print:  Resource Group: grpStonith8
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: native_print:      prmStonith8	(stonith:external/ipmi):	Started bl460g1n6 
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: clone_print:  Clone Set: clnPing [prmPing]
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: short_print:      Started: [ bl460g1n6 bl460g1n8 ]
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: short_print:      Stopped: [ bl460g1n7 ]
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: native_color: Resource prmPing:2 cannot run anywhere
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: RecurringOp:  Start recurring monitor (10s) for prmVM2 on bl460g1n8
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmDummy	(Started bl460g1n6)
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmVM2	(Started bl460g1n8)
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmStonith6	(Started bl460g1n8)
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmStonith7	(Started bl460g1n6)
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmStonith8	(Started bl460g1n6)
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmPing:0	(Started bl460g1n6)
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmPing:1	(Started bl460g1n8)
Dec 18 14:11:51 bl460g1n6 pengine[17820]:     info: LogActions: Leave   prmPing:2	(Stopped)
Dec 18 14:11:51 bl460g1n6 pengine[17820]:   notice: process_pe_message: Calculated Transition 1: /var/lib/pacemaker/pengine/pe-input-0.bz2
Dec 18 14:11:51 bl460g1n6 crmd[17821]:     info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Dec 18 14:11:51 bl460g1n6 crmd[17821]:     info: do_te_invoke: Processing graph 1 (ref=pe_calc-dc-1387343511-24) derived from /var/lib/pacemaker/pengine/pe-input-0.bz2
Dec 18 14:11:51 bl460g1n6 crmd[17821]:   notice: te_rsc_command: Initiating action 14: monitor prmVM2_monitor_10000 on bl460g1n8
Dec 18 14:11:51 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=bl460g1n8/crmd/29, version=0.11.13)
Dec 18 14:11:51 bl460g1n6 crmd[17821]:     info: match_graph_event: Action prmVM2_monitor_10000 (14) confirmed on bl460g1n8 (rc=0)
Dec 18 14:11:51 bl460g1n6 crmd[17821]:   notice: run_graph: Transition 1 (Complete=1, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-0.bz2): Complete
Dec 18 14:11:51 bl460g1n6 crmd[17821]:     info: do_log: FSA: Input I_TE_SUCCESS from notify_crmd() received in state S_TRANSITION_ENGINE
Dec 18 14:11:51 bl460g1n6 crmd[17821]:   notice: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: write_attribute: Sent update 2 with 2 changes for shutdown, id=<n/a>, set=(null)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: write_attribute: Sent update 3 with 2 changes for default_ping_set, id=<n/a>, set=(null)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: write_attribute: Sent update 4 with 2 changes for terminate, id=<n/a>, set=(null)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: write_attribute: Sent update 5 with 2 changes for probe_complete, id=<n/a>, set=(null)
Dec 18 14:11:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/2, version=0.11.13)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: attrd_cib_callback: Update 2 for shutdown[bl460g1n6]=(null): OK (0)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: attrd_cib_callback: Update 2 for shutdown[bl460g1n8]=(null): OK (0)
Dec 18 14:11:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/3, version=0.11.13)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: attrd_cib_callback: Update 3 for default_ping_set[bl460g1n6]=100: OK (0)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: attrd_cib_callback: Update 3 for default_ping_set[bl460g1n8]=100: OK (0)
Dec 18 14:11:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/4, version=0.11.13)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: attrd_cib_callback: Update 4 for terminate[bl460g1n6]=(null): OK (0)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: attrd_cib_callback: Update 4 for terminate[bl460g1n8]=(null): OK (0)
Dec 18 14:11:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_modify operation for section status: OK (rc=0, origin=local/attrd/5, version=0.11.13)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: attrd_cib_callback: Update 5 for probe_complete[bl460g1n6]=true: OK (0)
Dec 18 14:11:53 bl460g1n6 attrd[17819]:   notice: attrd_cib_callback: Update 5 for probe_complete[bl460g1n8]=true: OK (0)
Dec 18 14:13:46 bl460g1n6 root: Mark:pcmk:1387343626
Dec 18 14:13:53 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x27d0670 for uid=0 gid=0 pid=20475 id=44e81b7f-fae7-4aad-bdfe-55f1a04a9d76
Dec 18 14:13:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_mon/2, version=0.11.13)
Dec 18 14:13:53 bl460g1n6 cib[17816]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:13:53 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x27d0670 for uid=0 gid=0 pid=20477 id=6c9173ca-cebf-422f-b9f5-80fb7e96b6d9
Dec 18 14:13:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/cibadmin/2, version=0.11.13)
Dec 18 14:13:53 bl460g1n6 cib[17816]:     info: crm_client_destroy: Destroying 0 events
Dec 18 14:13:53 bl460g1n6 cib[17816]:     info: crm_client_new: Connecting 0x27d0670 for uid=0 gid=0 pid=20554 id=81b9ba66-8d28-4a68-a2b6-8c9ed409c656
Dec 18 14:13:53 bl460g1n6 cib[17816]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_mon/2, version=0.11.13)
Dec 18 14:13:53 bl460g1n6 cib[17816]:     info: crm_client_destroy: Destroying 0 events
