Oct 21 11:19:18 bl460g1n8 corosync[1580]:   [MAIN  ] main.c:main:1171 Corosync Cluster Engine ('2.3.2.4-805b3'): started and ready to provide service.
Oct 21 11:19:18 bl460g1n8 corosync[1580]:   [MAIN  ] main.c:main:1172 Corosync built-in features: watchdog upstart snmp pie relro bindnow
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [TOTEM ] totemnet.c:totemnet_instance_initialize:242 Initializing transport (UDP/IP Multicast).
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [TOTEM ] totemcrypto.c:init_nss:579 Initializing transmit/receive security (NSS) crypto: aes256 hash: sha1
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [TOTEM ] totemnet.c:totemnet_instance_initialize:242 Initializing transport (UDP/IP Multicast).
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [TOTEM ] totemcrypto.c:init_nss:579 Initializing transmit/receive security (NSS) crypto: aes256 hash: sha1
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [TOTEM ] totemudp.c:timer_function_netif_check_timeout:670 The network interface [192.168.101.218] is now up.
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync configuration map access [0]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on cmap [0]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: cmap
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync configuration service [1]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on cfg [1]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: cfg
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync cluster closed process group service v1.01 [2]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on cpg [2]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: cpg
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync profile loading service [4]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:851 NOT Initializing IPC on pload [4]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [WD    ] wd.c:setup_watchdog:651 Watchdog is now been tickled by corosync.
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [WD    ] wd.c:setup_watchdog:652 HP iLO2+ HW Watchdog Timer
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [WD    ] wd.c:wd_scan_resources:580 no resources configured.
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync watchdog service [7]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:851 NOT Initializing IPC on wd [7]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:quorum_exec_init_fn:274 Using quorum provider corosync_votequorum
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:votequorum_readconfig:967 Reading configuration (runtime: 0)
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:votequorum_read_nodelist_configuration:886 No nodelist defined or our node is not in the nodelist
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=1, expected_votes=3
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: Yes Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync vote quorum service v1.0 [5]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on votequorum [5]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: votequorum
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync cluster quorum service v0.1 [3]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on quorum [3]
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: quorum
Oct 21 11:19:18 bl460g1n8 corosync[1581]:   [TOTEM ] totemudp.c:timer_function_netif_check_timeout:670 The network interface [192.168.102.218] is now up.
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [MAIN  ] main.c:member_object_joined:333 Member joined: r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) 
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [TOTEM ] totemsrp.c:memb_state_operational_enter:1966 A new membership (192.168.101.218:4) was formed. Members joined: -1062705702
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261594
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261594]: votes: 1, expected: 3 flags: 8
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: Yes Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=1, expected_votes=3
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync configuration map access
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_sync_activate:386 Single node sync -> no action
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:downlist_log:776 comparing: sender r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ; members(old:0 left:0)
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:downlist_log:776 chosen downlist: sender r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ; members(old:0 left:0)
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync cluster closed process group service v1.01
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: Yes Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261594
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261594]: votes: 1, expected: 3 flags: 8
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: Yes Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=1, expected_votes=3
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261594
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[0]: votes: 0, expected: 0 flags: 0
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync vote quorum service v1.0
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=1, expected_votes=3
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[1]: -1062705702
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:send_library_notification:359 sending quorum notification to (nil), length = 52
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [MAIN  ] main.c:corosync_sync_completed:276 Completed service synchronization, ready to provide service.
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1586]
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1586-26)
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1586-26) state:2
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cfg-response-1581-1586-26-header
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cfg-event-1581-1586-26-header
Oct 21 11:19:19 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cfg-request-1581-1586-26-header
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] main.c:member_object_joined:333 Member joined: r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) 
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] main.c:member_object_joined:333 Member joined: r(0) ip(192.168.101.217) r(1) ip(192.168.102.217) 
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [TOTEM ] totemsrp.c:memb_state_operational_enter:1966 A new membership (192.168.101.216:16) was formed. Members joined: -1062705704 -1062705703
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync configuration map access
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_sync_activate:400 My config version is 0 -> no action
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:downlist_log:776 comparing: sender r(0) ip(192.168.101.217) r(1) ip(192.168.102.217) ; members(old:1 left:0)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:downlist_log:776 comparing: sender r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) ; members(old:1 left:0)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:downlist_log:776 comparing: sender r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ; members(old:1 left:0)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:downlist_log:776 chosen downlist: sender r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) ; members(old:1 left:0)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync cluster closed process group service v1.01
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261593
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261593]: votes: 1, expected: 3 flags: 0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=2, expected_votes=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:get_lowest_node_id:527 lowest node id: -1062705703 us: -1062705702
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:are_we_quorate:777 quorum regained, resuming activity
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261593
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[0]: votes: 0, expected: 0 flags: 0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261594
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261594]: votes: 1, expected: 3 flags: 0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=2, expected_votes=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:get_lowest_node_id:527 lowest node id: -1062705703 us: -1062705702
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261594
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[0]: votes: 0, expected: 0 flags: 0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261592
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261592]: votes: 1, expected: 3 flags: 0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=3, expected_votes=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261592 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:get_lowest_node_id:527 lowest node id: -1062705704 us: -1062705702
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261592
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[0]: votes: 0, expected: 0 flags: 0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync vote quorum service v1.0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=3, expected_votes=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261592 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [VOTEQ ] votequorum.c:get_lowest_node_id:527 lowest node id: -1062705704 us: -1062705702
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:quorum_api_set_quorum:148 This node is within the primary component and will provide service.
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[3]: -1062705704 -1062705703 -1062705702
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:send_library_notification:359 sending quorum notification to (nil), length = 60
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] main.c:corosync_sync_completed:276 Completed service synchronization, ready to provide service.
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705704 (r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) ) for pid 7676
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_ipc_connect: Could not establish pacemakerd connection: Connection refused (111)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1591]
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85b9f23d30
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1591]
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85b9e215e0
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: get_cluster_type: Detected an active 'corosync' cluster
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: mcp_read_config: Reading configure for stack: corosync
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1591-27)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1591-27) state:2
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85b9e215e0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1591-27-header
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1591-27-header
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1591-27-header
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: mcp_read_config: Configured corosync to accept connections from group 189: OK (1)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1591-26)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1591-26) state:2
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85b9f23d30
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1591-26-header
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: main: Starting Pacemaker 1.1.11-0.302.b6d42ed.git.el6 (Build: b6d42ed):  generated-manpages agent-manpages ascii-docs ncurses libqb-logging libqb-ipc lha-fencing nagios  corosync-native snmp
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1591-26-header
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: main: Maximum core file size is: 18446744073709551615
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1591-26-header
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: qb_ipcs_us_publish: server name: pacemakerd
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1591]
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1591]
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:cpg_lib_init_fn:1459 lib_init_fn: conn=0x7f85ba0241d0, cpd=0x7f85ba024864
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_get_peer: Created entry 3287844f-99cc-4967-ba64-c5d45bb11050/0x254f0b0 for node (null)/3232261594 (1 total)
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1591]
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:quorum_lib_init_fn:316 lib_init_fn: conn=0x7f85b9e229a0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_gettype:471 got quorum_type request on 0x7f85b9e229a0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_getquorate:395 got quorate request on 0x7f85b9e229a0
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: cluster_connect_quorum: Quorum acquired
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:412 got trackstart request on 0x7f85b9e229a0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:420 sending initial status to 0x7f85b9e229a0
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:send_library_notification:359 sending quorum notification to 0x7f85b9e229a0, length = 60
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1591]
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85ba228d60
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705702 (r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ) for pid 1591
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1591-29)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1591-29) state:2
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85ba228d60
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1591-29-header
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1591-29-header
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1591-29-header
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1591]
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Using uid=189 and group=189 for process cib
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Forked child 1595 for process cib
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Forked child 1596 for process stonith-ng
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Forked child 1597 for process lrmd
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Using uid=189 and group=189 for process attrd
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Forked child 1598 for process attrd
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Using uid=189 and group=189 for process pengine
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Forked child 1599 for process pengine
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Using uid=189 and group=189 for process crmd
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: start_child: Forked child 1600 for process crmd
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: main: Starting mainloop
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: pcmk_quorum_notification: Membership 16: quorum retained (3)
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_get_peer: Created entry 5cadcbca-14cd-4fab-a7a8-c2b617a7cdcd/0x26515c0 for node (null)/3232261592 (2 total)
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261592
Oct 21 11:19:21 bl460g1n8 cib[1595]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n8 cib[1595]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:19:21 bl460g1n8 cib[1595]:   notice: main: Using new config location: /var/lib/pacemaker/cib
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:19:21 bl460g1n8 lrmd[1597]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n8 lrmd[1597]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.xml (digest: /var/lib/pacemaker/cib/cib.xml.sig)
Oct 21 11:19:21 bl460g1n8 attrd[1598]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n8 attrd[1598]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:21 bl460g1n8 cib[1595]:  warning: retrieveCib: Cluster configuration not found: /var/lib/pacemaker/cib/cib.xml
Oct 21 11:19:21 bl460g1n8 cib[1595]:  warning: readCibXmlFile: Primary configuration corrupt or unusable, trying backups in /var/lib/pacemaker/cib
Oct 21 11:19:21 bl460g1n8 cib[1595]:  warning: readCibXmlFile: Continuing with an empty configuration.
Oct 21 11:19:21 bl460g1n8 lrmd[1597]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: validate_with_relaxng: Creating RNG parser context
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:21 bl460g1n8 lrmd[1597]:     info: qb_ipcs_us_publish: server name: lrmd
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:19:21 bl460g1n8 lrmd[1597]:     info: main: Starting
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: main: Starting up
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:19:21 bl460g1n8 attrd[1598]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:19:21 bl460g1n8 pengine[1599]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n8 pengine[1599]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:21 bl460g1n8 pengine[1599]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:19:21 bl460g1n8 pengine[1599]:     info: qb_ipcs_us_publish: server name: pengine
Oct 21 11:19:21 bl460g1n8 pengine[1599]:     info: main: Starting pengine
Oct 21 11:19:21 bl460g1n8 crmd[1600]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n8 crmd[1600]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:21 bl460g1n8 crmd[1600]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:19:21 bl460g1n8 crmd[1600]:   notice: main: CRM Git Version: b6d42ed
Oct 21 11:19:21 bl460g1n8 crmd[1600]:     info: do_log: FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
Oct 21 11:19:21 bl460g1n8 crmd[1600]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:19:21 bl460g1n8 crmd[1600]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:19:21 bl460g1n8 crmd[1600]:     info: crm_ipc_connect: Could not establish cib_shm connection: Connection refused (111)
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261592] - state is now member (was (null))
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_get_peer: Created entry af171a7f-f802-48af-85e0-780ea7b6d9b4/0x2650dd0 for node (null)/3232261593 (3 total)
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261593
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: startCib: CIB Initialization completed successfully
Oct 21 11:19:21 bl460g1n8 cib[1595]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: crm_get_peer: Created entry a5b410d6-a7a9-4ebe-83e4-719435526264/0xaa2120 for node (null)/3232261594 (1 total)
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:19:21 bl460g1n8 attrd[1598]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261594] - state is now member (was (null))
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:     info: crm_get_peer: Created entry cb9e6fce-c866-410b-9af4-c1eda0b441a5/0x1eb3660 for node (null)/3232261594 (1 total)
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261593] - state is now member (was (null))
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n8[3232261594] - state is now member (was (null))
Oct 21 11:19:21 bl460g1n8 pacemakerd[1591]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:21 bl460g1n8 attrd[1598]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:21 bl460g1n8 attrd[1598]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: main: Cluster connection active
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: qb_ipcs_us_publish: server name: attrd
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: main: Accepting attribute updates
Oct 21 11:19:21 bl460g1n8 attrd[1598]:     info: crm_ipc_connect: Could not establish cib_rw connection: Connection refused (111)
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: crm_get_peer: Created entry 69aef4b0-2a66-4f25-bcbc-9293678733f5/0x135ffd0 for node (null)/3232261594 (1 total)
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:21 bl460g1n8 stonith-ng[1596]:     info: crm_ipc_connect: Could not establish cib_rw connection: Connection refused (111)
Oct 21 11:19:21 bl460g1n8 cib[1595]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:21 bl460g1n8 cib[1595]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: qb_ipcs_us_publish: server name: cib_ro
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: qb_ipcs_us_publish: server name: cib_rw
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: qb_ipcs_us_publish: server name: cib_shm
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: cib_init: Starting cib mainloop
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: pcmk_cpg_membership: Joined[0.0] cib.3232261594 
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: crm_get_peer: Created entry f27c2fe7-8d75-4bb8-9c5c-4a5d1b8c0dac/0x13628e0 for node (null)/3232261592 (2 total)
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: pcmk_cpg_membership: Member[0.0] cib.3232261592 
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:19:21 bl460g1n8 cib[1595]:     info: pcmk_cpg_membership: Member[0.1] cib.3232261594 
Oct 21 11:19:21 bl460g1n8 cib[1601]:     info: write_cib_contents: Wrote version 0.0.0 of the CIB to disk (digest: 38e2a365180d27f2831950bf8df46420)
Oct 21 11:19:21 bl460g1n8 cib[1601]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.mBen3y (digest: /var/lib/pacemaker/cib/cib.P9XeYu)
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1363000 for uid=189 gid=189 pid=1600 id=55fbdd1c-678e-4263-8ad6-e414e025627d
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: do_cib_control: CIB connection established
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.0.0)
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: crm_get_peer: Created entry 4281a303-62fe-4103-85b7-62f235d691e6/0x1dbeec0 for node (null)/3232261594 (1 total)
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: peer_update_callback: bl460g1n8 is now (null)
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: cluster_connect_quorum: Quorum acquired
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1387470 for uid=189 gid=189 pid=1598 id=059f7fa5-71e7-40bb-abd1-e38a76c69b21
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: attrd_cib_connect: Connected to the CIB after 2 attempts
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: main: CIB connection active
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: pcmk_cpg_membership: Joined[0.0] attrd.3232261594 
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: crm_get_peer: Created entry 14851a28-c0c0-4cee-9ffd-4b23da03e465/0xaa8070 for node (null)/3232261592 (2 total)
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: pcmk_cpg_membership: Member[0.0] attrd.3232261592 
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n8 attrd[1598]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261592] - state is now member (was (null))
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: pcmk_cpg_membership: Member[0.1] attrd.3232261594 
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: do_ha_control: Connected to the cluster
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: lrmd_ipc_connect: Connecting to lrmd
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.0.0)
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x11b1be0 for uid=0 gid=0 pid=1596 id=8558224a-d620-49ac-97e3-b33569b45729
Oct 21 11:19:22 bl460g1n8 lrmd[1597]:     info: crm_client_new: Connecting 0x1e09d10 for uid=189 gid=189 pid=1600 id=64932f2b-865d-49b5-a4f4-9afcb9540d6f
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:   notice: setup_cib: Watching for stonith topology changes
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.0.0)
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: qb_ipcs_us_publish: server name: stonith-ng
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: main: Starting stonith-ng mainloop
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: do_lrm_control: LRM connection established
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: do_started: Delaying start, no membership data (0000000000100000)
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: init_cib_cache_cb: Updating device list from the cib: init
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: pcmk_quorum_notification: Membership 16: quorum retained (3)
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.0.0)
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: crm_get_peer: Created entry a6061a87-c6d2-42fe-8e26-2161de73d28a/0x1f04aa0 for node (null)/3232261592 (2 total)
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261592
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: unpack_nodes: Creating a fake local node
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: pcmk_cpg_membership: Joined[0.0] stonith-ng.3232261594 
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: crm_get_peer: Created entry b5d64c00-fd6d-47d4-9e2c-708d3e1386ae/0x1ebbbf0 for node (null)/3232261592 (2 total)
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: pcmk_cpg_membership: Member[0.0] stonith-ng.3232261592 
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261592] - state is now member (was (null))
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: crm_get_peer: Created entry 8d388136-9411-40ea-841e-c03aea01e501/0x1f04990 for node (null)/3232261593 (3 total)
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261593
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: pcmk_cpg_membership: Member[0.1] stonith-ng.3232261594 
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261593] - state is now member (was (null))
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n8[3232261594] - state is now member (was (null))
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: peer_update_callback: bl460g1n8 is now member (was (null))
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: do_started: Delaying start, Config not read (0000000000000040)
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: qb_ipcs_us_publish: server name: crmd
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: do_started: The local CRM is operational
Oct 21 11:19:22 bl460g1n8 crmd[1600]:     info: do_log: FSA: Input I_PENDING from do_started() received in state S_STARTING
Oct 21 11:19:22 bl460g1n8 crmd[1600]:   notice: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/5, version=0.0.0)
Oct 21 11:19:22 bl460g1n8 pacemakerd[1591]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: pcmk_cpg_membership: Joined[1.0] attrd.3232261593 
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: pcmk_cpg_membership: Joined[1.0] cib.3232261593 
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: pcmk_cpg_membership: Member[1.0] attrd.3232261592 
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: pcmk_cpg_membership: Member[1.0] cib.3232261592 
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: pcmk_cpg_membership: Joined[1.0] stonith-ng.3232261593 
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: crm_get_peer: Created entry 80ec14f1-1a14-470c-8ea9-cdf4b9908ae5/0xaa80e0 for node (null)/3232261593 (3 total)
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: pcmk_cpg_membership: Member[1.0] stonith-ng.3232261592 
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: crm_get_peer: Created entry baedea8b-30b6-4c66-a009-06c1f87d10f5/0x13ca870 for node (null)/3232261593 (3 total)
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: crm_get_peer: Created entry ace4ad1a-2d2a-47e4-98d4-a23df10b3eb8/0x1eba750 for node (null)/3232261593 (3 total)
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: pcmk_cpg_membership: Member[1.1] attrd.3232261593 
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: pcmk_cpg_membership: Member[1.1] cib.3232261593 
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: pcmk_cpg_membership: Member[1.1] stonith-ng.3232261593 
Oct 21 11:19:22 bl460g1n8 cib[1595]:     info: pcmk_cpg_membership: Member[1.2] cib.3232261594 
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n8 attrd[1598]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261593] - state is now member (was (null))
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: pcmk_cpg_membership: Member[1.2] stonith-ng.3232261594 
Oct 21 11:19:22 bl460g1n8 attrd[1598]:     info: pcmk_cpg_membership: Member[1.2] attrd.3232261594 
Oct 21 11:19:22 bl460g1n8 stonith-ng[1596]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:23 bl460g1n8 stonith-ng[1596]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: pcmk_cpg_membership: Joined[0.0] crmd.3232261594 
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: pcmk_cpg_membership: Member[0.0] crmd.3232261594 
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: pcmk_cpg_membership: Joined[1.0] crmd.3232261592 
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: pcmk_cpg_membership: Member[1.0] crmd.3232261592 
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: pcmk_cpg_membership: Member[1.1] crmd.3232261594 
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: peer_update_callback: bl460g1n6 is now member
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: pcmk_cpg_membership: Joined[2.0] crmd.3232261593 
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: pcmk_cpg_membership: Member[2.0] crmd.3232261592 
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: pcmk_cpg_membership: Member[2.1] crmd.3232261593 
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: pcmk_cpg_membership: Member[2.2] crmd.3232261594 
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:23 bl460g1n8 crmd[1600]:     info: peer_update_callback: bl460g1n7 is now member
Oct 21 11:19:24 bl460g1n8 stonith-ng[1596]:     info: crm_client_new: Connecting 0x1eb8610 for uid=189 gid=189 pid=1600 id=3ddb5881-bfe1-43cf-bedb-ac31c7dad58e
Oct 21 11:19:24 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed register from crmd.1600: OK (0)
Oct 21 11:19:24 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_notify from crmd.1600: OK (0)
Oct 21 11:19:24 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_notify from crmd.1600: OK (0)
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: election_count_vote: Election 1 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_PENDING
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1595]
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85ba2380b0
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1595-34)
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1595-34) state:2
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:43 bl460g1n8 cib[1595]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:43 bl460g1n8 cib[1595]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85ba2380b0
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1595-34-header
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1595-34-header
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1595-34-header
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section cib: OK (rc=0, origin=bl460g1n6/crmd/7, version=0.0.1)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n6/crmd/9, version=0.1.1)
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: election_count_vote: Election 2 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: update_dc: Unset DC. Was bl460g1n6
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_PENDING
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/6, version=0.1.1)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/7, version=0.1.1)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/8, version=0.1.1)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n6/crmd/11, version=0.2.1)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/9, version=0.2.1)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_replace: Digest matched on replace from bl460g1n6: 2a3b572276d3a9904ee84d7bd4781b52
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_replace: Replaced 0.2.1 with 0.2.1 from bl460g1n6
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_replace operation for section 'all': OK (rc=0, origin=bl460g1n6/crmd/16, version=0.2.1)
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n8']/transient_attributes
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: update_attrd_helper: Connecting to attrd... 5 retries remaining
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_delete operation for section //node_state[@uname='bl460g1n8']/transient_attributes to master (origin=local/crmd/10)
Oct 21 11:19:43 bl460g1n8 attrd[1598]:     info: crm_client_new: Connecting 0xaa5460 for uid=189 gid=189 pid=1600 id=400ed215-7a97-41ed-9ad8-488efd123e21
Oct 21 11:19:43 bl460g1n8 cib[1610]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-0.raw
Oct 21 11:19:43 bl460g1n8 crmd[1600]:     info: do_log: FSA: Input I_NOT_DC from do_cl_join_finalize_respond() received in state S_PENDING
Oct 21 11:19:43 bl460g1n8 crmd[1600]:   notice: do_state_transition: State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
Oct 21 11:19:43 bl460g1n8 attrd[1598]:     info: attrd_client_message: Starting an election to determine the writer
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1598]
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85ba2380b0
Oct 21 11:19:43 bl460g1n8 attrd[1598]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:43 bl460g1n8 attrd[1598]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section nodes: OK (rc=0, origin=bl460g1n6/crmd/17, version=0.3.1)
Oct 21 11:19:43 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting terminate[bl460g1n8] = (null)
Oct 21 11:19:43 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting shutdown[bl460g1n8] = (null)
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1598-34)
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1598-34) state:2
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85ba2380b0
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1598-34-header
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1598-34-header
Oct 21 11:19:43 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1598-34-header
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section nodes: OK (rc=0, origin=bl460g1n6/crmd/18, version=0.4.1)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section nodes: OK (rc=0, origin=bl460g1n6/crmd/19, version=0.5.1)
Oct 21 11:19:43 bl460g1n8 attrd[1598]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:43 bl460g1n8 attrd[1598]:     info: election_count_vote: Election 1 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/22, version=0.5.2)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:43 bl460g1n8 attrd[1598]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:43 bl460g1n8 attrd[1598]:     info: election_count_vote: Election 2 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/24, version=0.5.3)
Oct 21 11:19:43 bl460g1n8 attrd[1598]:   notice: attrd_peer_message: Processing sync-response from bl460g1n6
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/26, version=0.5.4)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/2, version=0.5.5)
Oct 21 11:19:43 bl460g1n8 cib[1610]:     info: write_cib_contents: Wrote version 0.1.0 of the CIB to disk (digest: 2b5d37bb9a92bb6b3485bb82e926f2e5)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section cib: OK (rc=0, origin=bl460g1n6/crmd/29, version=0.5.6)
Oct 21 11:19:43 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting probe_complete[bl460g1n8] = true
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/4, version=0.5.7)
Oct 21 11:19:43 bl460g1n8 cib[1610]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.sJbMrB (digest: /var/lib/pacemaker/cib/cib.C97Pxv)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/5, version=0.5.8)
Oct 21 11:19:43 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/6, version=0.5.9)
Oct 21 11:19:43 bl460g1n8 cib[1611]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-1.raw
Oct 21 11:19:43 bl460g1n8 cib[1611]:     info: write_cib_contents: Wrote version 0.5.0 of the CIB to disk (digest: fd1c17f0935512546ae5c2e678df50ae)
Oct 21 11:19:43 bl460g1n8 cib[1611]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.vlXV3I (digest: /var/lib/pacemaker/cib/cib.b96GtD)
Oct 21 11:20:18 bl460g1n8 crmd[1600]:     info: election_count_vote: Election 3 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:20:18 bl460g1n8 crmd[1600]:     info: update_dc: Unset DC. Was bl460g1n6
Oct 21 11:20:18 bl460g1n8 crmd[1600]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_NOT_DC
Oct 21 11:20:18 bl460g1n8 crmd[1600]:   notice: do_state_transition: State transition S_NOT_DC -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_election_count_vote ]
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/11, version=0.5.9)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section 'all': OK (rc=0, origin=bl460g1n6/cibadmin/2, version=0.6.1)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/12, version=0.6.1)
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_level_remove: Node bl460g1n6 not found (0 active entries)
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_level_register: Node bl460g1n6 has 1 active fencing levels
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_level_register: Node bl460g1n6 has 2 active fencing levels
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_level_remove: Node bl460g1n7 not found (1 active entries)
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_level_register: Node bl460g1n7 has 1 active fencing levels
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_level_register: Node bl460g1n7 has 2 active fencing levels
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_level_remove: Node bl460g1n8 not found (2 active entries)
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_level_register: Node bl460g1n8 has 1 active fencing levels
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_level_register: Node bl460g1n8 has 2 active fencing levels
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:  warning: handle_startup_fencing: Blind faith: not fencing unseen nodes
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:18 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:18 bl460g1n8 crmd[1600]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:20:18 bl460g1n8 crmd[1600]:     info: election_count_vote: Election 4 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:20:18 bl460g1n8 crmd[1600]:     info: update_dc: Unset DC. Was bl460g1n6
Oct 21 11:20:18 bl460g1n8 crmd[1600]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_PENDING
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n6/crmd/38, version=0.7.1)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/13, version=0.7.1)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/14, version=0.7.1)
Oct 21 11:20:18 bl460g1n8 crmd[1600]:  warning: join_query_callback: No DC for join-3
Oct 21 11:20:18 bl460g1n8 crmd[1600]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/15, version=0.7.1)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n6/crmd/40, version=0.8.1)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/16, version=0.8.1)
Oct 21 11:20:18 bl460g1n8 cib[1632]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-2.raw
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_replace: Digest matched on replace from bl460g1n6: 46121f46af7d94446d41f49a1119615a
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_replace: Replaced 0.8.1 with 0.8.1 from bl460g1n6
Oct 21 11:20:18 bl460g1n8 crmd[1600]:     info: do_log: FSA: Input I_NOT_DC from do_cl_join_finalize_respond() received in state S_PENDING
Oct 21 11:20:18 bl460g1n8 crmd[1600]:   notice: do_state_transition: State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_replace operation for section 'all': OK (rc=0, origin=bl460g1n6/crmd/46, version=0.8.1)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=bl460g1n6/crmd/50, version=0.8.2)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/51, version=0.8.3)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n8']/lrm: OK (rc=0, origin=bl460g1n6/crmd/52, version=0.8.4)
Oct 21 11:20:18 bl460g1n8 cib[1632]:     info: write_cib_contents: Wrote version 0.6.0 of the CIB to disk (digest: 790911eef6fe5f321a7dd0f980e93d42)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/53, version=0.8.5)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n6']/lrm: OK (rc=0, origin=bl460g1n6/crmd/54, version=0.8.6)
Oct 21 11:20:18 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/55, version=0.8.7)
Oct 21 11:20:18 bl460g1n8 cib[1632]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.0ECxTs (digest: /var/lib/pacemaker/cib/cib.FBeQeV)
Oct 21 11:20:18 bl460g1n8 cib[1636]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-3.raw
Oct 21 11:20:18 bl460g1n8 cib[1636]:     info: write_cib_contents: Wrote version 0.8.0 of the CIB to disk (digest: 691cb7e31283f65808b970435c0d8e48)
Oct 21 11:20:18 bl460g1n8 cib[1636]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.tjfLnD (digest: /var/lib/pacemaker/cib/cib.5hEC15)
Oct 21 11:20:19 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (1 active devices)
Oct 21 11:20:19 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:19 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:20 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (2 active devices)
Oct 21 11:20:20 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:20 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:20 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmVM1' not found (0 active resources)
Oct 21 11:20:20 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmVM1' to the rsc list (1 active resources)
Oct 21 11:20:20 bl460g1n8 crmd[1600]:     info: do_lrm_rsc_op: Performing key=26:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM1_monitor_0
Oct 21 11:20:20 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmVM2' not found (1 active resources)
Oct 21 11:20:20 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmVM2' to the rsc list (2 active resources)
Oct 21 11:20:20 bl460g1n8 crmd[1600]:     info: do_lrm_rsc_op: Performing key=27:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM2_monitor_0
Oct 21 11:20:20 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmVM3' not found (2 active resources)
Oct 21 11:20:20 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmVM3' to the rsc list (3 active resources)
Oct 21 11:20:20 bl460g1n8 crmd[1600]:     info: do_lrm_rsc_op: Performing key=28:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM3_monitor_0
Oct 21 11:20:20 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith6-1' not found (3 active resources)
Oct 21 11:20:20 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmStonith6-1' to the rsc list (4 active resources)
Oct 21 11:20:20 bl460g1n8 crmd[1600]:     info: do_lrm_rsc_op: Performing key=29:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-1_monitor_0
Oct 21 11:20:20 bl460g1n8 VirtualDomain(prmVM1)[1643]: DEBUG: Virtual domain vm1 is currently shut off.
Oct 21 11:20:20 bl460g1n8 VirtualDomain(prmVM2)[1646]: DEBUG: Virtual domain vm2 is currently shut off.
Oct 21 11:20:20 bl460g1n8 VirtualDomain(prmVM3)[1648]: DEBUG: Virtual domain vm3 is currently shut off.
Oct 21 11:20:20 bl460g1n8 crm_resource[1747]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n8 crm_resource[1747]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1389c40 for uid=0 gid=0 pid=1747 id=6698b20e-a005-42a2-b534-072ccde73617
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n8 crm_resource[1749]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n8 crm_resource[1749]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1389c40 for uid=0 gid=0 pid=1749 id=1c3656d2-aec8-498f-b51c-4b23d93035d9
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 crm_resource[1751]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n8 crm_resource[1751]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x13d19e0 for uid=0 gid=0 pid=1751 id=79c7db3f-9fcd-47bc-b017-1bce44a58c41
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n8 crm_resource[1753]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n8 crm_resource[1753]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1389c40 for uid=0 gid=0 pid=1753 id=2ee0dc22-5acc-4e7c-8c86-81175217501c
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n8 crm_resource[1755]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n8 crm_resource[1755]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM1"]/utilization//nvpair[@name="cpu"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x13d19e0 for uid=0 gid=0 pid=1755 id=464dc553-d37b-4e18-8ad2-d85e986fe21e
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM2"]/utilization//nvpair[@name="cpu"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:20 bl460g1n8 crm_resource[1757]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n8 crm_resource[1757]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x13e53b0 for uid=0 gid=0 pid=1757 id=d4abd1d8-e5fd-47f9-9358-d0ecf5c3c493
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM3"]/utilization//nvpair[@name="cpu"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.8.7)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.9.1)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.10.1)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.11.1)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n8 cib[1758]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-4.raw
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n8 crm_resource[1769]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n8 crm_resource[1769]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x13d19e0 for uid=0 gid=0 pid=1769 id=3426a56c-ede3-4685-9e22-6cec384899a8
Oct 21 11:20:20 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n8 crm_resource[1774]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n8 crm_resource[1774]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n8 cib[1758]:     info: write_cib_contents: Wrote version 0.9.0 of the CIB to disk (digest: 6706479431a247ca0e49bb8b54057d59)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x13d19e0 for uid=0 gid=0 pid=1774 id=c5225c1c-bc36-4694-b900-3c892ef3c4b4
Oct 21 11:20:21 bl460g1n8 crm_resource[1776]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n8 crm_resource[1776]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x14358b0 for uid=0 gid=0 pid=1776 id=43d4dc52-ad49-4c01-be18-3122d7e8d3a0
Oct 21 11:20:21 bl460g1n8 crm_resource[1778]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n8 crm_resource[1778]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1460960 for uid=0 gid=0 pid=1778 id=12c763fb-3610-46e6-b381-8fc12d98dbc6
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM1"]/utilization//nvpair[@name="hv_memory"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1758]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.iD6u0O (digest: /var/lib/pacemaker/cib/cib.BgtNpn)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:21 bl460g1n8 crm_resource[1780]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n8 crm_resource[1780]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x13d19e0 for uid=0 gid=0 pid=1780 id=c08b21d6-e866-4698-b6c2-48ddb501adbd
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 crm_resource[1782]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n8 crm_resource[1782]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x14358b0 for uid=0 gid=0 pid=1782 id=4c281d08-3e4d-446a-a373-10964c821ac3
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM2"]/utilization//nvpair[@name="hv_memory"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM3"]/utilization//nvpair[@name="hv_memory"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.11.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.12.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.13.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.14.1)
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n8 cib[1783]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-5.raw
Oct 21 11:20:21 bl460g1n8 cib[1783]:     info: write_cib_contents: Wrote version 0.12.0 of the CIB to disk (digest: f9164a9a0ff267cf10ecd24bb92b6cec)
Oct 21 11:20:21 bl460g1n8 cib[1783]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.qLz0iP (digest: /var/lib/pacemaker/cib/cib.C00a0n)
Oct 21 11:20:21 bl460g1n8 cib[1784]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-6.raw
Oct 21 11:20:21 bl460g1n8 cib[1784]:     info: write_cib_contents: Wrote version 0.14.0 of the CIB to disk (digest: 0fa312cc44a6846316c5d9be6b00ad4b)
Oct 21 11:20:21 bl460g1n8 cib[1784]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.MYjtoV (digest: /var/lib/pacemaker/cib/cib.xdOSpu)
Oct 21 11:20:21 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-1' to the device list (3 active devices)
Oct 21 11:20:21 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:21 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:22 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (4 active devices)
Oct 21 11:20:22 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-1 has been disabled on bl460g1n8: score=-INFINITY
Oct 21 11:20:22 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:22 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:23 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:23 bl460g1n8 stonith-ng[1596]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:23 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:23 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:23 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:24 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:24 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:24 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:24 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:25 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:25 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-1' from the device list (4 active devices)
Oct 21 11:20:25 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:25 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:26 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-1' to the device list (5 active devices)
Oct 21 11:20:26 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:26 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:26 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:27 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:27 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-1 has been disabled on bl460g1n8: score=-INFINITY
Oct 21 11:20:27 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:27 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:27 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:28 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:28 bl460g1n8 stonith-ng[1596]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:28 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:28 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:28 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:29 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:29 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:29 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:29 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:30 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:30 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-1' from the device list (4 active devices)
Oct 21 11:20:30 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:30 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:31 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-1' to the device list (5 active devices)
Oct 21 11:20:31 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:31 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:31 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:32 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:32 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-1 has been disabled on bl460g1n8: score=-INFINITY
Oct 21 11:20:32 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:32 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:32 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:33 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:33 bl460g1n8 stonith-ng[1596]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:33 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:33 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:33 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:34 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:34 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:34 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:34 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:35 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:35 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-1' from the device list (4 active devices)
Oct 21 11:20:35 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:35 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:36 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-1' to the device list (5 active devices)
Oct 21 11:20:36 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:36 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:36 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:37 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:37 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-1 has been disabled on bl460g1n8: score=-INFINITY
Oct 21 11:20:37 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:37 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:37 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:38 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:38 bl460g1n8 stonith-ng[1596]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:38 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:38 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:38 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:39 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:39 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:39 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:39 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:40 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:40 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-1' from the device list (4 active devices)
Oct 21 11:20:40 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:40 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:41 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-1' to the device list (5 active devices)
Oct 21 11:20:41 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:41 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:41 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:42 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:42 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-1 has been disabled on bl460g1n8: score=-INFINITY
Oct 21 11:20:42 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:42 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:42 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:43 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:43 bl460g1n8 stonith-ng[1596]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:43 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:43 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:43 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:44 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:44 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:44 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:44 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:45 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:45 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-1' from the device list (4 active devices)
Oct 21 11:20:45 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:45 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:46 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-1' to the device list (5 active devices)
Oct 21 11:20:46 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:46 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:46 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:47 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:47 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-1 has been disabled on bl460g1n8: score=-INFINITY
Oct 21 11:20:47 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:47 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:47 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:48 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:48 bl460g1n8 stonith-ng[1596]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:48 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:48 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:48 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:49 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:49 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:49 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:49 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:50 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:50 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-1' from the device list (4 active devices)
Oct 21 11:20:50 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-1 is allowed on bl460g1n8: score=0
Oct 21 11:20:50 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:51 bl460g1n8 crmd[1600]:  warning: crm_ipc_send: Request 44 to lrmd (0x1ec1dd0) failed: Resource temporarily unavailable (-11)
Oct 21 11:20:51 bl460g1n8 crmd[1600]:    error: lrmd_send_command: Couldn't perform lrmd_rsc_info operation (timeout=30000): -11: Connection timed out (110)
Oct 21 11:20:51 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-1' to the device list (5 active devices)
Oct 21 11:20:51 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:51 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:51 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: crm_ipc_send: Sending to lrmd (0x1ec1dd0) is disabled until pending reply is received
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: lrmd_send_command: Couldn't perform lrmd_rsc_register operation (timeout=0): -114: Connection timed out (110)
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: crm_ipc_send: Sending to lrmd (0x1ec1dd0) is disabled until pending reply is received
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: lrmd_send_command: Couldn't perform lrmd_rsc_info operation (timeout=30000): -114: Connection timed out (110)
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: get_lrm_resource: Could not add resource prmStonith6-2 to LRM
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: register_fsa_error_adv: Resetting the current action list
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: do_lrm_invoke: Invalid resource definition
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_lrm_invoke: bad input   <create_request_adv origin="te_rsc_command" t="crmd" version="3.0.7" subt="request" reference="lrm_invoke-tengine-1382322020-45" crm_task="lrm_invoke" crm_sys_to="lrmd" crm_sys_from="tengine" crm_host_to="bl460g1n8" src="bl460g1n6">
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_lrm_invoke: bad input     <crm_xml>
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_lrm_invoke: bad input       <rsc_op id="30" operation="monitor" operation_key="prmStonith6-2_monitor_0" on_node="bl460g1n8" on_node_uuid="3232261594" transition-key="30:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5">
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_lrm_invoke: bad input         <primitive id="prmStonith6-2" class="stonith" type="external/ssh"/>
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_lrm_invoke: bad input         <attributes CRM_meta_op_target_rc="7" CRM_meta_timeout="60000" crm_feature_set="3.0.7" hostlist="bl460g1n6" pcmk_reboot_retries="1" pcmk_reboot_timeout="40s"/>
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_lrm_invoke: bad input       </rsc_op>
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_lrm_invoke: bad input     </crm_xml>
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_lrm_invoke: bad input   </create_request_adv>
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_log: FSA: Input I_FAIL from get_lrm_resource() received in state S_NOT_DC
Oct 21 11:20:52 bl460g1n8 crmd[1600]:   notice: do_state_transition: State transition S_NOT_DC -> S_RECOVERY [ input=I_FAIL cause=C_FSA_INTERNAL origin=get_lrm_resource ]
Oct 21 11:20:52 bl460g1n8 crmd[1600]:  warning: do_recover: Fast-tracking shutdown in response to errors
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: do_log: FSA: Input I_TERMINATE from do_recover() received in state S_RECOVERY
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: do_state_transition: State transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE cause=C_FSA_INTERNAL origin=do_recover ]
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: do_shutdown: Disconnecting STONITH...
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: tengine_stonith_connection_destroy: Fencing daemon disconnected
Oct 21 11:20:52 bl460g1n8 crmd[1600]:   notice: lrm_state_verify_stopped: Stopped 0 recurring operations at shutdown (4 ops remaining)
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: lrm_state_verify_stopped: 4 pending LRM operations at shutdown
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: lrm_state_verify_stopped: Pending action: prmVM3:13 (prmVM3_monitor_0)
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: lrm_state_verify_stopped: Pending action: prmVM2:9 (prmVM2_monitor_0)
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: lrm_state_verify_stopped: Pending action: prmVM1:5 (prmVM1_monitor_0)
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: lrm_state_verify_stopped: Pending action: prmStonith6-1:17 (prmStonith6-1_monitor_0)
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: do_lrm_control: Disconnecting from the LRM
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: lrmd_api_disconnect: Disconnecting from lrmd service
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: lrmd_ipc_connection_destroy: IPC connection destroyed
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: lrm_connection_destroy: LRM Connection disconnected
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: lrmd_api_disconnect: Disconnecting from lrmd service
Oct 21 11:20:52 bl460g1n8 crmd[1600]:   notice: do_lrm_control: Disconnected from the LRM
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: crm_cluster_disconnect: Disconnecting from cluster infrastructure: corosync
Oct 21 11:20:52 bl460g1n8 crmd[1600]:   notice: terminate_cs_connection: Disconnecting from Corosync
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:message_handler_req_lib_cpg_leave:1544 got leave request on 0x7f85ba025530
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:message_handler_req_lib_cpg_finalize:1582 cpg finalize for conn=0x7f85ba025530
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1600-29)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1600-29) state:2
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:cpg_lib_exit_fn:1043 exit_fn for conn=0x7f85ba025530
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cpg-response-1581-1600-29-header
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: crm_cluster_disconnect: Disconnected from corosync
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: do_ha_control: Disconnected from the cluster
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: do_cib_control: Disconnecting CIB
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cpg-event-1581-1600-29-header
Oct 21 11:20:52 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/17, version=0.14.1)
Oct 21 11:20:52 bl460g1n8 cib[1595]:  warning: qb_ipcs_event_sendv: new_event_notification (1595-1600-10): Broken pipe (32)
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: crmd_cib_connection_destroy: Connection to the CIB terminated...
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: qb_ipcs_us_withdraw: withdrawing server sockets
Oct 21 11:20:52 bl460g1n8 cib[1595]:  warning: do_local_notify: A-Sync reply to crmd failed: No message of desired type
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: do_exit: Performing A_EXIT_0 - gracefully exiting the CRMd
Oct 21 11:20:52 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 1 events
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: do_exit: [crmd] stopped (0)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cpg-request-1581-1600-29-header
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: crmd_exit: Dropping I_TERMINATE: [ state=S_TERMINATE cause=C_FSA_INTERNAL origin=do_stop ]
Oct 21 11:20:52 bl460g1n8 attrd[1598]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: crmd_quorum_destroy: connection closed
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: crmd_cs_destroy: connection closed
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: crmd_init: 1600 stopped: OK (0)
Oct 21 11:20:52 bl460g1n8 crmd[1600]:    error: crmd_fast_exit: Could not recover from internal error
Oct 21 11:20:52 bl460g1n8 crmd[1600]:     info: crm_xml_cleanup: Cleaning up memory from libxml2
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1600-33)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1600-33) state:2
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:quorum_lib_exit_fn:328 lib_exit_fn: conn=0x7f85ba329d10
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-quorum-response-1581-1600-33-header
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-quorum-event-1581-1600-33-header
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-quorum-request-1581-1600-33-header
Oct 21 11:20:52 bl460g1n8 pacemakerd[1591]:    error: pcmk_child_exit: Child process crmd (1600) exited: Generic Pacemaker error (201)
Oct 21 11:20:52 bl460g1n8 pacemakerd[1591]:   notice: pcmk_process_exit: Respawning failed child process: crmd
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procleave:1276 got procleave message from cluster node -1062705702
Oct 21 11:20:52 bl460g1n8 pacemakerd[1591]:     info: start_child: Using uid=189 and group=189 for process crmd
Oct 21 11:20:52 bl460g1n8 pacemakerd[1591]:     info: start_child: Forked child 1841 for process crmd
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procleave:1276 got procleave message from cluster node -1062705703
Oct 21 11:20:52 bl460g1n8 crmd[1841]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:52 bl460g1n8 crmd[1841]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:20:52 bl460g1n8 crmd[1841]:   notice: main: CRM Git Version: b6d42ed
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: do_log: FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:20:52 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x14358b0 for uid=189 gid=189 pid=1841 id=eabcdde9-93a3-44d6-963e-33e9a294cc3a
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: do_cib_control: CIB connection established
Oct 21 11:20:52 bl460g1n8 crmd[1841]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1841]
Oct 21 11:20:52 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.14.1)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:cpg_lib_init_fn:1459 lib_init_fn: conn=0x7f85ba025530, cpd=0x7f85ba236a54
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: crm_get_peer: Created entry 24e01579-4483-4018-9096-c5866515490c/0x1c064a0 for node (null)/3232261594 (1 total)
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1841]
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85ba329ad0
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705702 (r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ) for pid 1841
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1841-33)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1841-33) state:2
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n8 crmd[1841]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n8 crmd[1841]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85ba329ad0
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1841-33-header
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: peer_update_callback: bl460g1n8 is now (null)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1841-33-header
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1841-33-header
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1841]
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:quorum_lib_init_fn:316 lib_init_fn: conn=0x7f85ba329ad0
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705703 (r(0) ip(192.168.101.217) r(1) ip(192.168.102.217) ) for pid 12703
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_gettype:471 got quorum_type request on 0x7f85ba329ad0
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_getquorate:395 got quorate request on 0x7f85ba329ad0
Oct 21 11:20:52 bl460g1n8 crmd[1841]:   notice: cluster_connect_quorum: Quorum acquired
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:412 got trackstart request on 0x7f85ba329ad0
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:420 sending initial status to 0x7f85ba329ad0
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QUORUM] vsf_quorum.c:send_library_notification:359 sending quorum notification to 0x7f85ba329ad0, length = 60
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1841]
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85ba237320
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1841-34)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1841-34) state:2
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85ba237320
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1841-34-header
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1841-34-header
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1841-34-header
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1841]
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85ba2371d0
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1841-34)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1841-34) state:2
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85ba2371d0
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1841-34-header
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1841-34-header
Oct 21 11:20:52 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1841-34-header
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: do_ha_control: Connected to the cluster
Oct 21 11:20:52 bl460g1n8 crmd[1841]:     info: lrmd_ipc_connect: Connecting to lrmd
Oct 21 11:20:52 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.14.1)
Oct 21 11:20:52 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:52 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-1 has been disabled on bl460g1n8: score=-INFINITY
Oct 21 11:20:52 bl460g1n8 stonith-ng[1596]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:52 bl460g1n8 stonith-ng[1596]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n8: score=0
Oct 21 11:20:52 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:53 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:53 bl460g1n8 stonith-ng[1596]:     info: crm_client_new: Connecting 0x1fbdac0 for uid=0 gid=0 pid=1597 id=9b03f481-23b8-4296-83f7-cabb2f1164af
Oct 21 11:20:53 bl460g1n8 stonith-ng[1596]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:53 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed register from lrmd.1597: OK (0)
Oct 21 11:20:53 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_notify from lrmd.1597: OK (0)
Oct 21 11:20:53 bl460g1n8 lrmd[1597]:  warning: qb_ipcs_event_sendv: new_event_notification (1597-1600-6): Broken pipe (32)
Oct 21 11:20:53 bl460g1n8 lrmd[1597]:  warning: send_client_notify: Notification of client crmd/64932f2b-865d-49b5-a4f4-9afcb9540d6f failed
Oct 21 11:20:53 bl460g1n8 lrmd[1597]:  warning: send_client_notify: Notification of client crmd/64932f2b-865d-49b5-a4f4-9afcb9540d6f failed
Oct 21 11:20:53 bl460g1n8 lrmd[1597]:  warning: send_client_notify: Notification of client crmd/64932f2b-865d-49b5-a4f4-9afcb9540d6f failed
Oct 21 11:20:53 bl460g1n8 lrmd[1597]:  warning: send_client_notify: Notification of client crmd/64932f2b-865d-49b5-a4f4-9afcb9540d6f failed
Oct 21 11:20:53 bl460g1n8 lrmd[1597]:     info: crm_client_new: Connecting 0x1e13ee0 for uid=189 gid=189 pid=1841 id=558c5cd3-d57c-4791-9fa2-ae7593aff936
Oct 21 11:20:53 bl460g1n8 lrmd[1597]:     info: crm_client_destroy: Destroying 4 events
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: do_lrm_control: LRM connection established
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: do_started: Delaying start, no membership data (0000000000100000)
Oct 21 11:20:53 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.14.1)
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: pcmk_quorum_notification: Membership 16: quorum retained (3)
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: crm_get_peer: Created entry eae9e1d8-14b3-40dc-8518-fff4e423b0bd/0x1d4f890 for node (null)/3232261592 (2 total)
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261592
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1841]
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85ba2371a0
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1841-34)
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1841-34) state:2
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:53 bl460g1n8 crmd[1841]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85ba2371a0
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:53 bl460g1n8 crmd[1841]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261592] - state is now member (was (null))
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1841-34-header
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: crm_get_peer: Created entry 823b93e1-2ec4-4e08-9154-ddbb3e69c002/0x1d49ff0 for node (null)/3232261593 (3 total)
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261593
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1841-34-header
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1841-34-header
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1841]
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85ba2371a0
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1841-34)
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1841-34) state:2
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:53 bl460g1n8 crmd[1841]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:53 bl460g1n8 crmd[1841]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261593] - state is now member (was (null))
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85ba2371a0
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:53 bl460g1n8 crmd[1841]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n8[3232261594] - state is now member (was (null))
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1841-34-header
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: peer_update_callback: bl460g1n8 is now member (was (null))
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1841-34-header
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1841-34-header
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [1841]
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f85ba2371a0
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (1581-1841-34)
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(1581-1841-34) state:2
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:53 bl460g1n8 crmd[1841]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:53 bl460g1n8 crmd[1841]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f85ba2371a0
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-1581-1841-34-header
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: qb_ipcs_us_publish: server name: crmd
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-1581-1841-34-header
Oct 21 11:20:53 bl460g1n8 crmd[1841]:   notice: do_started: The local CRM is operational
Oct 21 11:20:53 bl460g1n8 crmd[1841]:     info: do_log: FSA: Input I_PENDING from do_started() received in state S_STARTING
Oct 21 11:20:53 bl460g1n8 crmd[1841]:   notice: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
Oct 21 11:20:53 bl460g1n8 corosync[1581]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-1581-1841-34-header
Oct 21 11:20:53 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/5, version=0.14.1)
Oct 21 11:20:53 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/60, version=0.14.2)
Oct 21 11:20:53 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/61, version=0.14.3)
Oct 21 11:20:53 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/62, version=0.14.4)
Oct 21 11:20:53 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/63, version=0.14.5)
Oct 21 11:20:53 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/64, version=0.14.6)
Oct 21 11:20:53 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/65, version=0.14.7)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/66, version=0.14.8)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/67, version=0.14.9)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/68, version=0.14.10)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/69, version=0.14.11)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/70, version=0.14.12)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/71, version=0.14.13)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/72, version=0.14.14)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/73, version=0.14.15)
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: pcmk_cpg_membership: Joined[0.0] crmd.3232261594 
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: pcmk_cpg_membership: Member[0.0] crmd.3232261592 
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: pcmk_cpg_membership: Member[0.1] crmd.3232261594 
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: pcmk_cpg_membership: Joined[1.0] crmd.3232261593 
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: pcmk_cpg_membership: Member[1.0] crmd.3232261592 
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: pcmk_cpg_membership: Member[1.1] crmd.3232261593 
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: pcmk_cpg_membership: Member[1.2] crmd.3232261594 
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: peer_update_callback: bl460g1n6 is now member
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith7-1' not found (4 active resources)
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmStonith7-1' to the rsc list (5 active resources)
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=31:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-1_monitor_0
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmStonith7-1 in state S_PENDING
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith7-2' not found (5 active resources)
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmStonith7-2' to the rsc list (6 active resources)
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=32:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-2_monitor_0
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmStonith7-2 in state S_PENDING
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith8-1' not found (6 active resources)
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmStonith8-1' to the rsc list (7 active resources)
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=33:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-1_monitor_0
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmStonith8-1 in state S_PENDING
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith8-2' not found (7 active resources)
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmStonith8-2' to the rsc list (8 active resources)
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=34:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-2_monitor_0
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmStonith8-2 in state S_PENDING
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmPing' not found (8 active resources)
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmPing:2' not found (8 active resources)
Oct 21 11:20:54 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmPing' to the rsc list (9 active resources)
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=35:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmPing_monitor_0
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmPing in state S_PENDING
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: services_os_action_execute: Managed ping_meta-data_0 process 1866 exited with rc=0
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: peer_update_callback: bl460g1n7 is now member
Oct 21 11:20:54 bl460g1n8 crmd[1841]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/6, version=0.14.15)
Oct 21 11:20:54 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/7, version=0.14.15)
Oct 21 11:20:55 bl460g1n8 stonith-ng[1596]:     info: crm_client_new: Connecting 0x1eb8610 for uid=189 gid=189 pid=1841 id=201c189b-f5db-4e5f-a4d7-521be5d30d59
Oct 21 11:20:55 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed register from crmd.1841: OK (0)
Oct 21 11:20:55 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_notify from crmd.1841: OK (0)
Oct 21 11:20:55 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_notify from crmd.1841: OK (0)
Oct 21 11:20:56 bl460g1n8 crmd[1841]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n8']/transient_attributes
Oct 21 11:20:56 bl460g1n8 crmd[1841]:     info: update_attrd_helper: Connecting to attrd... 5 retries remaining
Oct 21 11:20:56 bl460g1n8 attrd[1598]:     info: crm_client_new: Connecting 0xaa5460 for uid=189 gid=189 pid=1841 id=280880bf-ee68-4d9a-8d87-f5115a6b6fd4
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_replace: Digest matched on replace from bl460g1n6: 9e817daca05ff13f2c868ff292f7703a
Oct 21 11:20:56 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting terminate[bl460g1n8] = (null)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_replace: Replaced 0.14.15 with 0.14.15 from bl460g1n6
Oct 21 11:20:56 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting shutdown[bl460g1n8] = (null)
Oct 21 11:20:56 bl460g1n8 crmd[1841]:     info: do_log: FSA: Input I_NOT_DC from do_cl_join_finalize_respond() received in state S_PENDING
Oct 21 11:20:56 bl460g1n8 crmd[1841]:   notice: do_state_transition: State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_replace operation for section 'all': OK (rc=0, origin=bl460g1n6/crmd/76, version=0.14.15)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_delete operation for section //node_state[@uname='bl460g1n8']/transient_attributes to master (origin=local/crmd/8)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n6']/lrm: OK (rc=0, origin=bl460g1n6/crmd/80, version=0.14.16)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n7']/transient_attributes: OK (rc=0, origin=bl460g1n6/crmd/8, version=0.14.17)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n8']/transient_attributes: OK (rc=0, origin=bl460g1n6/crmd/8, version=0.14.18)
Oct 21 11:20:56 bl460g1n8 cib[1874]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-7.raw
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/81, version=0.14.19)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=bl460g1n6/crmd/82, version=0.14.20)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/83, version=0.14.21)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n8']/lrm: OK (rc=0, origin=bl460g1n6/crmd/84, version=0.14.22)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/85, version=0.14.23)
Oct 21 11:20:56 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/87, version=0.14.24)
Oct 21 11:20:56 bl460g1n8 cib[1874]:     info: write_cib_contents: Wrote version 0.14.0 of the CIB to disk (digest: 7ed85a3c13981ed3a8cf49c95bde5769)
Oct 21 11:20:56 bl460g1n8 cib[1874]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.CZwHwk (digest: /var/lib/pacemaker/cib/cib.niIoGq)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=16:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM1_monitor_0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=17:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM2_monitor_0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=18:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM3_monitor_0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=19:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-1_monitor_0
Oct 21 11:20:58 bl460g1n8 lrmd[1597]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith6-2' not found (9 active resources)
Oct 21 11:20:58 bl460g1n8 lrmd[1597]:     info: process_lrmd_rsc_register: Added 'prmStonith6-2' to the rsc list (10 active resources)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=20:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-2_monitor_0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=21:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-1_monitor_0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: process_lrm_event: LRM operation prmStonith6-1_monitor_0 (call=42, rc=7, cib-update=9, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/9)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: process_lrm_event: LRM operation prmStonith6-2_monitor_0 (call=46, rc=7, cib-update=10, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/10)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: process_lrm_event: LRM operation prmStonith7-1_monitor_0 (call=47, rc=7, cib-update=11, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/11)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=22:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-2_monitor_0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: process_lrm_event: LRM operation prmStonith7-2_monitor_0 (call=48, rc=7, cib-update=12, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/12)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=23:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-1_monitor_0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: process_lrm_event: LRM operation prmStonith8-1_monitor_0 (call=49, rc=7, cib-update=13, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/13)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=24:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-2_monitor_0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: process_lrm_event: LRM operation prmStonith8-2_monitor_0 (call=50, rc=7, cib-update=14, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/14)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=25:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmPing_monitor_0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmPing_monitor_0 (call=51, rc=7, cib-update=15, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/9, version=0.14.25)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/15)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/10, version=0.14.26)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/11, version=0.14.27)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/12, version=0.14.28)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/9, version=0.14.29)
Oct 21 11:20:58 bl460g1n8 VirtualDomain(prmVM1)[1875]: DEBUG: Virtual domain vm1 is currently shut off.
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/10, version=0.14.30)
Oct 21 11:20:58 bl460g1n8 VirtualDomain(prmVM2)[1878]: DEBUG: Virtual domain vm2 is currently shut off.
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/11, version=0.14.31)
Oct 21 11:20:58 bl460g1n8 VirtualDomain(prmVM3)[1881]: DEBUG: Virtual domain vm3 is currently shut off.
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/12, version=0.14.32)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/13, version=0.14.33)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/14, version=0.14.34)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/13, version=0.14.35)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/14, version=0.14.36)
Oct 21 11:20:58 bl460g1n8 crm_resource[1981]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n8 crm_resource[1981]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/15, version=0.14.37)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1383010 for uid=0 gid=0 pid=1981 id=6b8be496-1d5a-493a-8f18-98f4ed5f04b1
Oct 21 11:20:58 bl460g1n8 crm_resource[1988]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n8 crm_resource[1988]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1389c40 for uid=0 gid=0 pid=1988 id=7b737b9b-5769-4780-a122-8ccffefd044c
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.37)
Oct 21 11:20:58 bl460g1n8 crm_resource[1990]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n8 crm_resource[1990]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/15, version=0.14.38)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x13d19e0 for uid=0 gid=0 pid=1990 id=ba2456af-fbb3-4501-b9e3-7c8f79f4d3f5
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n8 crm_resource[2005]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n8 crm_resource[2005]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1383010 for uid=0 gid=0 pid=2005 id=79f93e9c-ae53-4c11-a056-b1d8acf7c6fa
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n8 crm_resource[2007]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n8 crm_resource[2007]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1389c40 for uid=0 gid=0 pid=2007 id=4a025ede-60c0-47b8-9596-648134f32226
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n8 crm_resource[2009]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n8 crm_resource[2009]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_new: Connecting 0x1383010 for uid=0 gid=0 pid=2009 id=74d1642d-fa93-44d0-99d6-3704efe1a1e8
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: services_os_action_execute: Managed VirtualDomain_meta-data_0 process 2010 exited with rc=0
Oct 21 11:20:58 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmVM1_monitor_0 (call=36, rc=7, cib-update=16, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/16)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmVM2_monitor_0 (call=38, rc=7, cib-update=17, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/17)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmVM3_monitor_0 (call=40, rc=7, cib-update=18, confirmed=true) not running
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/18)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/16, version=0.14.39)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/17, version=0.14.40)
Oct 21 11:20:58 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting probe_complete[bl460g1n8] = true
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/18, version=0.14.41)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/16, version=0.14.42)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/17, version=0.14.43)
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=43:2:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-1_start_0
Oct 21 11:20:58 bl460g1n8 lrmd[1597]:     info: log_execute: executing - rsc:prmStonith7-1 action:start call_id:52
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/18, version=0.14.44)
Oct 21 11:20:58 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:58 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=63:2:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmPing_start_0
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/90, version=0.14.45)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/91, version=0.14.46)
Oct 21 11:20:58 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/92, version=0.14.47)
Oct 21 11:20:59 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Device 'prmStonith7-1' already existed in device list (5 active devices)
Oct 21 11:20:59 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_device_register from lrmd.1597: OK (0)
Oct 21 11:20:59 bl460g1n8 lrmd[1597]:     info: log_execute: executing - rsc:prmPing action:start call_id:53
Oct 21 11:20:59 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_execute from lrmd.1597: Operation now in progress (-115)
Oct 21 11:20:59 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:20:59 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/10, version=0.14.48)
Oct 21 11:20:59 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/93, version=0.14.49)
Oct 21 11:21:00 bl460g1n8 attrd_updater[2049]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:21:00 bl460g1n8 attrd_updater[2049]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:21:00 bl460g1n8 attrd[1598]:     info: crm_client_new: Connecting 0xaab940 for uid=0 gid=0 pid=2049 id=6095bf31-191d-418e-a12b-f249e77672e4
Oct 21 11:21:00 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting default_ping_set[bl460g1n8] = 100
Oct 21 11:21:00 bl460g1n8 attrd[1598]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:21:00 bl460g1n8 lrmd[1597]:     info: log_finished: finished - rsc:prmPing action:start call_id:53 pid:2017 exit-code:0 exec-time:1035ms queue-time:0ms
Oct 21 11:21:00 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmPing_start_0 (call=53, rc=0, cib-update=19, confirmed=true) ok
Oct 21 11:21:00 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/19)
Oct 21 11:21:00 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/11, version=0.14.50)
Oct 21 11:21:00 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/19, version=0.14.51)
Oct 21 11:21:00 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-1:2020 [ Performing: stonith -t external/ipmi -S ]
Oct 21 11:21:00 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-1:2020 [ success:  0 ]
Oct 21 11:21:00 bl460g1n8 lrmd[1597]:     info: log_finished: finished - rsc:prmStonith7-1 action:start call_id:52  exit-code:0 exec-time:2100ms queue-time:0ms
Oct 21 11:21:00 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmStonith7-1_start_0 (call=52, rc=0, cib-update=20, confirmed=true) ok
Oct 21 11:21:00 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/20)
Oct 21 11:21:00 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/20, version=0.14.52)
Oct 21 11:21:01 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/12, version=0.14.53)
Oct 21 11:21:01 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/19, version=0.14.54)
Oct 21 11:21:01 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/20, version=0.14.55)
Oct 21 11:21:01 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/21, version=0.14.56)
Oct 21 11:21:03 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=26:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-1_monitor_360000
Oct 21 11:21:03 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=27:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-2_start_0
Oct 21 11:21:03 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_execute from lrmd.1597: Operation now in progress (-115)
Oct 21 11:21:03 bl460g1n8 lrmd[1597]:     info: log_execute: executing - rsc:prmStonith7-2 action:start call_id:55
Oct 21 11:21:03 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:03 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:21:03 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=44:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmPing_monitor_10000
Oct 21 11:21:03 bl460g1n8 stonith-ng[1596]:   notice: stonith_device_register: Device 'prmStonith7-2' already existed in device list (5 active devices)
Oct 21 11:21:03 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_device_register from lrmd.1597: OK (0)
Oct 21 11:21:03 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_execute from lrmd.1597: Operation now in progress (-115)
Oct 21 11:21:03 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:03 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting probe_complete[bl460g1n8] = true
Oct 21 11:21:04 bl460g1n8 attrd_updater[2113]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:21:04 bl460g1n8 attrd_updater[2113]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:21:04 bl460g1n8 attrd[1598]:     info: crm_client_new: Connecting 0xaab940 for uid=0 gid=0 pid=2113 id=660b18c8-9d6c-4def-8be8-708869811ef8
Oct 21 11:21:04 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting default_ping_set[bl460g1n8] = 100
Oct 21 11:21:04 bl460g1n8 attrd[1598]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:21:04 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmPing_monitor_10000 (call=56, rc=0, cib-update=21, confirmed=false) ok
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/21)
Oct 21 11:21:04 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-2:2057 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:04 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-2:2057 [ success:  0 ]
Oct 21 11:21:04 bl460g1n8 lrmd[1597]:     info: log_finished: finished - rsc:prmStonith7-2 action:start call_id:55  exit-code:0 exec-time:1065ms queue-time:0ms
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/95, version=0.14.57)
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/21, version=0.14.58)
Oct 21 11:21:04 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmStonith7-2_start_0 (call=55, rc=0, cib-update=22, confirmed=true) ok
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/22)
Oct 21 11:21:04 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-1:2052 [ Performing: stonith -t external/ipmi -S ]
Oct 21 11:21:04 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-1:2052 [ success:  0 ]
Oct 21 11:21:04 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmStonith7-1_monitor_360000 (call=54, rc=0, cib-update=23, confirmed=false) ok
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/23)
Oct 21 11:21:04 bl460g1n8 crmd[1841]:     info: do_lrm_rsc_op: Performing key=28:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-2_monitor_10000
Oct 21 11:21:04 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_execute from lrmd.1597: Operation now in progress (-115)
Oct 21 11:21:04 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/22, version=0.14.59)
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/23, version=0.14.60)
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/22, version=0.14.61)
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/23, version=0.14.62)
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/24, version=0.14.63)
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/25, version=0.14.64)
Oct 21 11:21:04 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/26, version=0.14.65)
Oct 21 11:21:05 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-2:2127 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:05 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-2:2127 [ success:  0 ]
Oct 21 11:21:05 bl460g1n8 crmd[1841]:   notice: process_lrm_event: LRM operation prmStonith7-2_monitor_10000 (call=57, rc=0, cib-update=24, confirmed=false) ok
Oct 21 11:21:05 bl460g1n8 cib[1595]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/24)
Oct 21 11:21:05 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/24, version=0.14.66)
Oct 21 11:21:05 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/27, version=0.14.67)
Oct 21 11:21:05 bl460g1n8 cib[1595]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/28, version=0.14.68)
Oct 21 11:21:15 bl460g1n8 attrd_updater[2158]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:21:15 bl460g1n8 attrd_updater[2158]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:21:15 bl460g1n8 attrd[1598]:     info: crm_client_new: Connecting 0xaab940 for uid=0 gid=0 pid=2158 id=90c6f5fd-e6ed-4905-81e4-f3960e5cd6ca
Oct 21 11:21:15 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting default_ping_set[bl460g1n8] = 100
Oct 21 11:21:15 bl460g1n8 attrd[1598]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:21:15 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_execute from lrmd.1597: Operation now in progress (-115)
Oct 21 11:21:15 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:16 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-2:2159 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:16 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-2:2159 [ success:  0 ]
Oct 21 11:21:26 bl460g1n8 attrd_updater[2192]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:21:26 bl460g1n8 attrd_updater[2192]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:21:26 bl460g1n8 attrd[1598]:     info: crm_client_new: Connecting 0xaab940 for uid=0 gid=0 pid=2192 id=a61d48d1-589b-4e6e-ade9-61c08d137e61
Oct 21 11:21:26 bl460g1n8 attrd[1598]:     info: attrd_client_message: Broadcasting default_ping_set[bl460g1n8] = 100
Oct 21 11:21:26 bl460g1n8 attrd[1598]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:21:26 bl460g1n8 stonith-ng[1596]:     info: stonith_command: Processed st_execute from lrmd.1597: Operation now in progress (-115)
Oct 21 11:21:26 bl460g1n8 stonith-ng[1596]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:27 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-2:2193 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:27 bl460g1n8 stonith-ng[1596]:     info: log_operation: prmStonith7-2:2193 [ success:  0 ]
