Oct 21 11:19:18 bl460g1n7 corosync[12467]:   [MAIN  ] main.c:main:1171 Corosync Cluster Engine ('2.3.2.4-805b3'): started and ready to provide service.
Oct 21 11:19:18 bl460g1n7 corosync[12467]:   [MAIN  ] main.c:main:1172 Corosync built-in features: watchdog upstart snmp pie relro bindnow
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [TOTEM ] totemnet.c:totemnet_instance_initialize:242 Initializing transport (UDP/IP Multicast).
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [TOTEM ] totemcrypto.c:init_nss:579 Initializing transmit/receive security (NSS) crypto: aes256 hash: sha1
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [TOTEM ] totemnet.c:totemnet_instance_initialize:242 Initializing transport (UDP/IP Multicast).
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [TOTEM ] totemcrypto.c:init_nss:579 Initializing transmit/receive security (NSS) crypto: aes256 hash: sha1
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [TOTEM ] totemudp.c:timer_function_netif_check_timeout:670 The network interface [192.168.101.217] is now up.
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync configuration map access [0]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on cmap [0]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: cmap
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync configuration service [1]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on cfg [1]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: cfg
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync cluster closed process group service v1.01 [2]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on cpg [2]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: cpg
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync profile loading service [4]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:851 NOT Initializing IPC on pload [4]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [WD    ] wd.c:setup_watchdog:651 Watchdog is now been tickled by corosync.
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [WD    ] wd.c:setup_watchdog:652 HP iLO2+ HW Watchdog Timer
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [WD    ] wd.c:wd_scan_resources:580 no resources configured.
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync watchdog service [7]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:851 NOT Initializing IPC on wd [7]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:quorum_exec_init_fn:274 Using quorum provider corosync_votequorum
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:votequorum_readconfig:967 Reading configuration (runtime: 0)
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:votequorum_read_nodelist_configuration:886 No nodelist defined or our node is not in the nodelist
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=1, expected_votes=3
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: Yes Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync vote quorum service v1.0 [5]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on votequorum [5]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: votequorum
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [SERV  ] service.c:corosync_service_link_and_init:174 Service engine loaded: corosync cluster quorum service v0.1 [3]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_service_init:865 Initializing IPC on quorum [3]
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_get_ipc_type:811 No configured qb.ipc_type. Using native ipc
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [QB    ] ipc_setup.c:qb_ipcs_us_publish:374 server name: quorum
Oct 21 11:19:18 bl460g1n7 corosync[12468]:   [TOTEM ] totemudp.c:timer_function_netif_check_timeout:670 The network interface [192.168.102.217] is now up.
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [MAIN  ] main.c:member_object_joined:333 Member joined: r(0) ip(192.168.101.217) r(1) ip(192.168.102.217) 
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [TOTEM ] totemsrp.c:memb_state_operational_enter:1966 A new membership (192.168.101.217:4) was formed. Members joined: -1062705703
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261593
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261593]: votes: 1, expected: 3 flags: 8
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: Yes Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=1, expected_votes=3
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync configuration map access
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_sync_activate:386 Single node sync -> no action
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:downlist_log:776 comparing: sender r(0) ip(192.168.101.217) r(1) ip(192.168.102.217) ; members(old:0 left:0)
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:downlist_log:776 chosen downlist: sender r(0) ip(192.168.101.217) r(1) ip(192.168.102.217) ; members(old:0 left:0)
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync cluster closed process group service v1.01
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: Yes Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261593
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261593]: votes: 1, expected: 3 flags: 8
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: Yes Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=1, expected_votes=3
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261593
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[0]: votes: 0, expected: 0 flags: 0
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync vote quorum service v1.0
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=1, expected_votes=3
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[1]: -1062705703
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:send_library_notification:359 sending quorum notification to (nil), length = 52
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [MAIN  ] main.c:corosync_sync_completed:276 Completed service synchronization, ready to provide service.
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12473]
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12473-26)
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12473-26) state:2
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cfg-response-12468-12473-26-header
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cfg-event-12468-12473-26-header
Oct 21 11:19:19 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cfg-request-12468-12473-26-header
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] main.c:member_object_joined:333 Member joined: r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) 
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] main.c:member_object_joined:333 Member joined: r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) 
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [TOTEM ] totemsrp.c:memb_state_operational_enter:1966 A new membership (192.168.101.216:16) was formed. Members joined: -1062705704 -1062705702
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:     info: crm_ipc_connect: Could not establish pacemakerd connection: Connection refused (111)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12477]
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecde1ee0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12477]
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecee48a0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12477-27)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12477-27) state:2
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:     info: get_cluster_type: Detected an active 'corosync' cluster
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecee48a0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:     info: mcp_read_config: Reading configure for stack: corosync
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12477-27-header
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12477-27-header
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12477-27-header
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:   notice: mcp_read_config: Configured corosync to accept connections from group 189: OK (1)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12477-26)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12477-26) state:2
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecde1ee0
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12477-26-header
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12477-26-header
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:   notice: main: Starting Pacemaker 1.1.11-0.302.b6d42ed.git.el6 (Build: b6d42ed):  generated-manpages agent-manpages ascii-docs ncurses libqb-logging libqb-ipc lha-fencing nagios  corosync-native snmp
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:     info: main: Maximum core file size is: 18446744073709551615
Oct 21 11:19:21 bl460g1n7 pacemakerd[12477]:     info: qb_ipcs_us_publish: server name: pacemakerd
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12477-26-header
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12477]
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12477]
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:cpg_lib_init_fn:1459 lib_init_fn: conn=0x7f79ecee3700, cpd=0x7f79ecee3e04
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync configuration map access
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_sync_activate:400 My config version is 0 -> no action
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:downlist_log:776 comparing: sender r(0) ip(192.168.101.217) r(1) ip(192.168.102.217) ; members(old:1 left:0)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:downlist_log:776 comparing: sender r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) ; members(old:1 left:0)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:downlist_log:776 comparing: sender r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ; members(old:1 left:0)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:downlist_log:776 chosen downlist: sender r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) ; members(old:1 left:0)
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync cluster closed process group service v1.01
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261593
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261593]: votes: 1, expected: 3 flags: 0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=1, expected_votes=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261593
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[0]: votes: 0, expected: 0 flags: 0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261594
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261594]: votes: 1, expected: 3 flags: 0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=2, expected_votes=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:get_lowest_node_id:527 lowest node id: -1062705703 us: -1062705703
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:are_we_quorate:777 quorum regained, resuming activity
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261594
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[0]: votes: 0, expected: 0 flags: 0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261592
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[3232261592]: votes: 1, expected: 3 flags: 0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:decode_flags:587 flags: quorate: No Leaving: No WFA Status: No First: No Qdevice: No QdeviceAlive: No QdeviceCastVote: No QdeviceMasterWins: No
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=3, expected_votes=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261592 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:get_lowest_node_id:527 lowest node id: -1062705704 us: -1062705703
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1604 got nodeinfo message from cluster node 3232261592
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:message_handler_req_exec_votequorum_nodeinfo:1609 nodeinfo message[0]: votes: 0, expected: 0 flags: 0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [SYNC  ] sync.c:sync_barrier_handler:232 Committing synchronization for corosync vote quorum service v1.0
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:recalculate_quorum:851 total_votes=3, expected_votes=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261592 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261593 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:calculate_quorum:670 node 3232261594 state=1, votes=1, expected=3
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [VOTEQ ] votequorum.c:get_lowest_node_id:527 lowest node id: -1062705704 us: -1062705703
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:quorum_api_set_quorum:148 This node is within the primary component and will provide service.
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:log_view_list:132 Members[3]: -1062705704 -1062705703 -1062705702
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:send_library_notification:359 sending quorum notification to (nil), length = 60
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [MAIN  ] main.c:corosync_sync_completed:276 Completed service synchronization, ready to provide service.
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705704 (r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) ) for pid 7676
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705702 (r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ) for pid 1591
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705704 (r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) ) for pid 7683
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705704 (r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) ) for pid 7681
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705704 (r(0) ip(192.168.101.216) r(1) ip(192.168.102.216) ) for pid 7680
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705702 (r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ) for pid 1598
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705702 (r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ) for pid 1596
Oct 21 11:19:21 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705702 (r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ) for pid 1595
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_get_peer: Created entry 58303714-8164-4f9b-a562-86f401a9644a/0x18e8200 for node (null)/3232261593 (1 total)
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705703 (r(0) ip(192.168.101.217) r(1) ip(192.168.102.217) ) for pid 12477
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12477]
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:quorum_lib_init_fn:316 lib_init_fn: conn=0x7f79eceed230
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_gettype:471 got quorum_type request on 0x7f79eceed230
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_getquorate:395 got quorate request on 0x7f79eceed230
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:   notice: cluster_connect_quorum: Quorum acquired
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:412 got trackstart request on 0x7f79eceed230
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:420 sending initial status to 0x7f79eceed230
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:send_library_notification:359 sending quorum notification to 0x7f79eceed230, length = 60
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12477]
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79eceeed30
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12477-29)
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12477-29) state:2
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79eceeed30
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:22 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12477-29-header
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Using uid=189 and group=189 for process cib
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Forked child 12482 for process cib
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Forked child 12483 for process stonith-ng
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Forked child 12484 for process lrmd
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Using uid=189 and group=189 for process attrd
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Forked child 12485 for process attrd
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Using uid=189 and group=189 for process pengine
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Forked child 12486 for process pengine
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Using uid=189 and group=189 for process crmd
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: start_child: Forked child 12487 for process crmd
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: main: Starting mainloop
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: pcmk_quorum_notification: Membership 16: quorum retained (3)
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_get_peer: Created entry ffc98b22-f979-421c-9215-1c7644559bfb/0x19ea460 for node (null)/3232261592 (2 total)
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261592
Oct 21 11:19:22 bl460g1n7 cib[12482]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:22 bl460g1n7 cib[12482]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:19:22 bl460g1n7 cib[12482]:   notice: main: Using new config location: /var/lib/pacemaker/cib
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.xml (digest: /var/lib/pacemaker/cib/cib.xml.sig)
Oct 21 11:19:22 bl460g1n7 cib[12482]:  warning: retrieveCib: Cluster configuration not found: /var/lib/pacemaker/cib/cib.xml
Oct 21 11:19:22 bl460g1n7 lrmd[12484]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:22 bl460g1n7 cib[12482]:  warning: readCibXmlFile: Primary configuration corrupt or unusable, trying backups in /var/lib/pacemaker/cib
Oct 21 11:19:22 bl460g1n7 lrmd[12484]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:22 bl460g1n7 cib[12482]:  warning: readCibXmlFile: Continuing with an empty configuration.
Oct 21 11:19:22 bl460g1n7 attrd[12485]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: validate_with_relaxng: Creating RNG parser context
Oct 21 11:19:22 bl460g1n7 attrd[12485]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:22 bl460g1n7 lrmd[12484]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Oct 21 11:19:22 bl460g1n7 lrmd[12484]:     info: qb_ipcs_us_publish: server name: lrmd
Oct 21 11:19:22 bl460g1n7 lrmd[12484]:     info: main: Starting
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: main: Starting up
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:19:22 bl460g1n7 attrd[12485]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/root
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:19:22 bl460g1n7 pengine[12486]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:22 bl460g1n7 pengine[12486]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:22 bl460g1n7 pengine[12486]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:19:22 bl460g1n7 pengine[12486]:     info: qb_ipcs_us_publish: server name: pengine
Oct 21 11:19:22 bl460g1n7 pengine[12486]:     info: main: Starting pengine
Oct 21 11:19:22 bl460g1n7 crmd[12487]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:19:22 bl460g1n7 crmd[12487]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:19:22 bl460g1n7 crmd[12487]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:19:22 bl460g1n7 crmd[12487]:   notice: main: CRM Git Version: b6d42ed
Oct 21 11:19:22 bl460g1n7 crmd[12487]:     info: do_log: FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
Oct 21 11:19:22 bl460g1n7 crmd[12487]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:19:22 bl460g1n7 crmd[12487]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:19:22 bl460g1n7 crmd[12487]:     info: crm_ipc_connect: Could not establish cib_shm connection: Connection refused (111)
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261592] - state is now member (was (null))
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n7[3232261593] - state is now member (was (null))
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_get_peer: Created entry 0ca63de8-6d24-4809-8d5a-42cb629ff669/0x19e9da0 for node (null)/3232261594 (3 total)
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261594
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: startCib: CIB Initialization completed successfully
Oct 21 11:19:22 bl460g1n7 cib[12482]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: crm_get_peer: Created entry 7a406073-1528-4134-bab2-454f7f6e2d6b/0x1e21120 for node (null)/3232261593 (1 total)
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n7 attrd[12485]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261593] - state is now member (was (null))
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:     info: crm_get_peer: Created entry 629c45b3-c73e-49fb-9a8d-ca26179bf71a/0x1d46660 for node (null)/3232261593 (1 total)
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261594] - state is now member (was (null))
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_get_peer: Created entry 3a759773-4349-4a9d-83b8-6f8b1347086f/0xf153f0 for node (null)/3232261593 (1 total)
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:22 bl460g1n7 pacemakerd[12477]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:22 bl460g1n7 attrd[12485]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:22 bl460g1n7 attrd[12485]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: main: Cluster connection active
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: qb_ipcs_us_publish: server name: attrd
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: main: Accepting attribute updates
Oct 21 11:19:22 bl460g1n7 attrd[12485]:     info: crm_ipc_connect: Could not establish cib_rw connection: Connection refused (111)
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:22 bl460g1n7 stonith-ng[12483]:     info: crm_ipc_connect: Could not establish cib_rw connection: Connection refused (111)
Oct 21 11:19:22 bl460g1n7 cib[12482]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:22 bl460g1n7 cib[12482]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: qb_ipcs_us_publish: server name: cib_ro
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: qb_ipcs_us_publish: server name: cib_rw
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: qb_ipcs_us_publish: server name: cib_shm
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: cib_init: Starting cib mainloop
Oct 21 11:19:22 bl460g1n7 cib[12488]:     info: write_cib_contents: Wrote version 0.0.0 of the CIB to disk (digest: f10d1085fbaf76b0407b2ee23bf15b3f)
Oct 21 11:19:22 bl460g1n7 cib[12488]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.uANZdu (digest: /var/lib/pacemaker/cib/cib.nPu8Oq)
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: pcmk_cpg_membership: Joined[0.0] cib.3232261593 
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_get_peer: Created entry cfb2d8d2-0a46-4625-84b2-884db697727f/0xf15dc0 for node (null)/3232261592 (2 total)
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: pcmk_cpg_membership: Member[0.0] cib.3232261592 
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: pcmk_cpg_membership: Member[0.1] cib.3232261593 
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_get_peer: Created entry 9003f5dd-9102-4584-a214-06dda7e600eb/0xf15e30 for node (null)/3232261594 (3 total)
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: pcmk_cpg_membership: Member[0.2] cib.3232261594 
Oct 21 11:19:22 bl460g1n7 cib[12482]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:19:23 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf15fb0 for uid=189 gid=189 pid=12487 id=3ede0c58-7b89-4f9a-aa70-620393fd956e
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: do_cib_control: CIB connection established
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:19:23 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.0.0)
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: crm_get_peer: Created entry 3dc9e61e-b42b-4fb5-881d-045fb6bac96b/0x17eceb0 for node (null)/3232261593 (1 total)
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: peer_update_callback: bl460g1n7 is now (null)
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: cluster_connect_quorum: Quorum acquired
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: do_ha_control: Connected to the cluster
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: lrmd_ipc_connect: Connecting to lrmd
Oct 21 11:19:23 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.0.0)
Oct 21 11:19:23 bl460g1n7 lrmd[12484]:     info: crm_client_new: Connecting 0xeeed10 for uid=189 gid=189 pid=12487 id=eaecbe18-3d6e-4619-bc5a-af0e1d4415fe
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: do_lrm_control: LRM connection established
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: do_started: Delaying start, no membership data (0000000000100000)
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: pcmk_quorum_notification: Membership 16: quorum retained (3)
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: crm_get_peer: Created entry 6ffd0b4c-cee5-4cd0-b557-f795a809c53a/0x1930a30 for node (null)/3232261592 (2 total)
Oct 21 11:19:23 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.0.0)
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261592
Oct 21 11:19:23 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xd64be0 for uid=189 gid=189 pid=12485 id=6473688a-8f08-496a-8ecf-c9eb079caced
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: attrd_cib_connect: Connected to the CIB after 2 attempts
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: main: CIB connection active
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: pcmk_cpg_membership: Joined[0.0] attrd.3232261593 
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: crm_get_peer: Created entry 8d77373f-51e6-4ba8-9ea2-dd4fda504a2c/0x1e27070 for node (null)/3232261592 (2 total)
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: pcmk_cpg_membership: Member[0.0] attrd.3232261592 
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:19:23 bl460g1n7 attrd[12485]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261592] - state is now member (was (null))
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: pcmk_cpg_membership: Member[0.1] attrd.3232261593 
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: crm_get_peer: Created entry 7836c990-12ca-4312-be8b-9110f7cd3255/0x1e270e0 for node (null)/3232261594 (3 total)
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: pcmk_cpg_membership: Member[0.2] attrd.3232261594 
Oct 21 11:19:23 bl460g1n7 attrd[12485]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:19:23 bl460g1n7 attrd[12485]:   notice: crm_update_peer_state: attrd_peer_change_cb: Node (null)[3232261594] - state is now member (was (null))
Oct 21 11:19:23 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf37040 for uid=0 gid=0 pid=12483 id=afc526be-a88f-45f3-8f41-cce36597a275
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:   notice: setup_cib: Watching for stonith topology changes
Oct 21 11:19:23 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.0.0)
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: qb_ipcs_us_publish: server name: stonith-ng
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: main: Starting stonith-ng mainloop
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: init_cib_cache_cb: Updating device list from the cib: init
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261592] - state is now member (was (null))
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: unpack_nodes: Creating a fake local node
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n7[3232261593] - state is now member (was (null))
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: peer_update_callback: bl460g1n7 is now member (was (null))
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: crm_get_peer: Created entry a37b2111-21ec-4a7f-8ec2-990a7b5b8651/0x1932a90 for node (null)/3232261594 (3 total)
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: pcmk_cpg_membership: Joined[0.0] stonith-ng.3232261593 
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: crm_get_peer: Created entry 8ac67e93-ea68-40fa-8a4a-658f6006a69c/0x1d4f130 for node (null)/3232261592 (2 total)
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261594
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: pcmk_cpg_membership: Member[0.0] stonith-ng.3232261592 
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261594] - state is now member (was (null))
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: pcmk_cpg_membership: Member[0.1] stonith-ng.3232261593 
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: crm_get_peer: Created entry 85db216a-87eb-48e7-994a-14f9ef2756bb/0x1d4cd80 for node (null)/3232261594 (3 total)
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: pcmk_cpg_membership: Member[0.2] stonith-ng.3232261594 
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:23 bl460g1n7 stonith-ng[12483]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: do_started: Delaying start, Config not read (0000000000000040)
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: qb_ipcs_us_publish: server name: crmd
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: do_started: The local CRM is operational
Oct 21 11:19:23 bl460g1n7 crmd[12487]:     info: do_log: FSA: Input I_PENDING from do_started() received in state S_STARTING
Oct 21 11:19:23 bl460g1n7 crmd[12487]:   notice: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
Oct 21 11:19:23 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/5, version=0.0.0)
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: pcmk_cpg_membership: Joined[0.0] crmd.3232261593 
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: pcmk_cpg_membership: Member[0.0] crmd.3232261592 
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: pcmk_cpg_membership: Member[0.1] crmd.3232261593 
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: pcmk_cpg_membership: Member[0.2] crmd.3232261594 
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: peer_update_callback: bl460g1n6 is now member
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:24 bl460g1n7 crmd[12487]:     info: peer_update_callback: bl460g1n8 is now member
Oct 21 11:19:25 bl460g1n7 stonith-ng[12483]:     info: crm_client_new: Connecting 0x1d4b610 for uid=189 gid=189 pid=12487 id=cf0d96c0-b312-4a19-9970-b811533ee7b0
Oct 21 11:19:25 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed register from crmd.12487: OK (0)
Oct 21 11:19:25 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_notify from crmd.12487: OK (0)
Oct 21 11:19:25 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_notify from crmd.12487: OK (0)
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: election_count_vote: Election 1 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_PENDING
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12482]
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecef0a20
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12482-34)
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12482-34) state:2
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:43 bl460g1n7 cib[12482]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:43 bl460g1n7 cib[12482]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecef0a20
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12482-34-header
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12482-34-header
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12482-34-header
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section cib: OK (rc=0, origin=bl460g1n6/crmd/7, version=0.0.1)
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n6/crmd/9, version=0.1.1)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/6, version=0.1.1)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/7, version=0.1.1)
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: election_count_vote: Election 2 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: update_dc: Unset DC. Was bl460g1n6
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_PENDING
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/8, version=0.1.1)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n6/crmd/11, version=0.2.1)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/9, version=0.2.1)
Oct 21 11:19:43 bl460g1n7 cib[12496]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-0.raw
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_replace: Digest matched on replace from bl460g1n6: 2a3b572276d3a9904ee84d7bd4781b52
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12485]
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_replace: Replaced 0.2.1 with 0.2.1 from bl460g1n6
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_replace operation for section 'all': OK (rc=0, origin=bl460g1n6/crmd/16, version=0.2.1)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/transient_attributes
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: update_attrd_helper: Connecting to attrd... 5 retries remaining
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section nodes: OK (rc=0, origin=bl460g1n6/crmd/17, version=0.3.1)
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section nodes: OK (rc=0, origin=bl460g1n6/crmd/18, version=0.4.1)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section nodes: OK (rc=0, origin=bl460g1n6/crmd/19, version=0.5.1)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_delete operation for section //node_state[@uname='bl460g1n7']/transient_attributes to master (origin=local/crmd/10)
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecef0a20
Oct 21 11:19:43 bl460g1n7 attrd[12485]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:19:43 bl460g1n7 attrd[12485]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:19:43 bl460g1n7 attrd[12485]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:19:43 bl460g1n7 attrd[12485]:     info: election_count_vote: Election 1 (owner: 3232261594) lost: vote from bl460g1n8 (Uptime)
Oct 21 11:19:43 bl460g1n7 attrd[12485]:     info: crm_client_new: Connecting 0x1e28be0 for uid=189 gid=189 pid=12487 id=1144e652-2497-4eb0-b828-207f1aae1499
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12485-34)
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12485-34) state:2
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecef0a20
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12485-34-header
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12485-34-header
Oct 21 11:19:43 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting terminate[bl460g1n7] = (null)
Oct 21 11:19:43 bl460g1n7 crmd[12487]:     info: do_log: FSA: Input I_NOT_DC from do_cl_join_finalize_respond() received in state S_PENDING
Oct 21 11:19:43 bl460g1n7 crmd[12487]:   notice: do_state_transition: State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
Oct 21 11:19:43 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting shutdown[bl460g1n7] = (null)
Oct 21 11:19:43 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12485-34-header
Oct 21 11:19:43 bl460g1n7 cib[12496]:     info: write_cib_contents: Wrote version 0.1.0 of the CIB to disk (digest: 2b5d37bb9a92bb6b3485bb82e926f2e5)
Oct 21 11:19:43 bl460g1n7 attrd[12485]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:19:43 bl460g1n7 attrd[12485]:     info: election_count_vote: Election 1 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/22, version=0.5.2)
Oct 21 11:19:43 bl460g1n7 attrd[12485]:     info: election_count_vote: Election 2 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/24, version=0.5.3)
Oct 21 11:19:43 bl460g1n7 attrd[12485]:   notice: attrd_peer_message: Processing sync-response from bl460g1n6
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/26, version=0.5.4)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/2, version=0.5.5)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section cib: OK (rc=0, origin=bl460g1n6/crmd/29, version=0.5.6)
Oct 21 11:19:43 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting probe_complete[bl460g1n7] = true
Oct 21 11:19:43 bl460g1n7 cib[12496]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.F0aSjA (digest: /var/lib/pacemaker/cib/cib.LhJ1ws)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/4, version=0.5.7)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/5, version=0.5.8)
Oct 21 11:19:43 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/6, version=0.5.9)
Oct 21 11:19:43 bl460g1n7 cib[12497]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-1.raw
Oct 21 11:19:43 bl460g1n7 cib[12497]:     info: write_cib_contents: Wrote version 0.5.0 of the CIB to disk (digest: fd1c17f0935512546ae5c2e678df50ae)
Oct 21 11:19:43 bl460g1n7 cib[12497]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.Lc3X9G (digest: /var/lib/pacemaker/cib/cib.20SKIz)
Oct 21 11:20:18 bl460g1n7 crmd[12487]:     info: election_count_vote: Election 3 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:20:18 bl460g1n7 crmd[12487]:     info: update_dc: Unset DC. Was bl460g1n6
Oct 21 11:20:18 bl460g1n7 crmd[12487]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_NOT_DC
Oct 21 11:20:18 bl460g1n7 crmd[12487]:   notice: do_state_transition: State transition S_NOT_DC -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_election_count_vote ]
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/11, version=0.5.9)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section 'all': OK (rc=0, origin=bl460g1n6/cibadmin/2, version=0.6.1)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/12, version=0.6.1)
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_level_remove: Node bl460g1n6 not found (0 active entries)
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_level_register: Node bl460g1n6 has 1 active fencing levels
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_level_register: Node bl460g1n6 has 2 active fencing levels
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_level_remove: Node bl460g1n7 not found (1 active entries)
Oct 21 11:20:18 bl460g1n7 crmd[12487]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_level_register: Node bl460g1n7 has 1 active fencing levels
Oct 21 11:20:18 bl460g1n7 crmd[12487]:     info: election_count_vote: Election 4 (owner: 3232261592) lost: vote from bl460g1n6 (Uptime)
Oct 21 11:20:18 bl460g1n7 crmd[12487]:     info: update_dc: Unset DC. Was bl460g1n6
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_level_register: Node bl460g1n7 has 2 active fencing levels
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n6/crmd/38, version=0.7.1)
Oct 21 11:20:18 bl460g1n7 crmd[12487]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_level_remove: Node bl460g1n8 not found (2 active entries)
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_level_register: Node bl460g1n8 has 1 active fencing levels
Oct 21 11:20:18 bl460g1n7 crmd[12487]:     info: do_log: FSA: Input I_PENDING from do_election_count_vote() received in state S_PENDING
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_level_register: Node bl460g1n8 has 2 active fencing levels
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:  warning: handle_startup_fencing: Blind faith: not fencing unseen nodes
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:18 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section crm_config: OK (rc=0, origin=bl460g1n6/crmd/40, version=0.8.1)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/13, version=0.8.1)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/14, version=0.8.1)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/15, version=0.8.1)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/16, version=0.8.1)
Oct 21 11:20:18 bl460g1n7 cib[12518]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-2.raw
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_replace: Digest matched on replace from bl460g1n6: 46121f46af7d94446d41f49a1119615a
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_replace: Replaced 0.8.1 with 0.8.1 from bl460g1n6
Oct 21 11:20:18 bl460g1n7 crmd[12487]:     info: do_log: FSA: Input I_NOT_DC from do_cl_join_finalize_respond() received in state S_PENDING
Oct 21 11:20:18 bl460g1n7 crmd[12487]:   notice: do_state_transition: State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_replace operation for section 'all': OK (rc=0, origin=bl460g1n6/crmd/46, version=0.8.1)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=bl460g1n6/crmd/50, version=0.8.2)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/51, version=0.8.3)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n8']/lrm: OK (rc=0, origin=bl460g1n6/crmd/52, version=0.8.4)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/53, version=0.8.5)
Oct 21 11:20:18 bl460g1n7 cib[12518]:     info: write_cib_contents: Wrote version 0.6.0 of the CIB to disk (digest: 790911eef6fe5f321a7dd0f980e93d42)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n6']/lrm: OK (rc=0, origin=bl460g1n6/crmd/54, version=0.8.6)
Oct 21 11:20:18 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/55, version=0.8.7)
Oct 21 11:20:18 bl460g1n7 cib[12518]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.bIVtUq (digest: /var/lib/pacemaker/cib/cib.FcjvrR)
Oct 21 11:20:18 bl460g1n7 cib[12520]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-3.raw
Oct 21 11:20:18 bl460g1n7 cib[12520]:     info: write_cib_contents: Wrote version 0.8.0 of the CIB to disk (digest: 691cb7e31283f65808b970435c0d8e48)
Oct 21 11:20:18 bl460g1n7 cib[12520]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.CpaINA (digest: /var/lib/pacemaker/cib/cib.kwymF1)
Oct 21 11:20:19 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (1 active devices)
Oct 21 11:20:19 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:19 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:20 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (2 active devices)
Oct 21 11:20:20 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-1 has been disabled on bl460g1n7: score=-INFINITY
Oct 21 11:20:20 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:20 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:20 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmVM1' not found (0 active resources)
Oct 21 11:20:20 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmVM1' to the rsc list (1 active resources)
Oct 21 11:20:20 bl460g1n7 crmd[12487]:     info: do_lrm_rsc_op: Performing key=15:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM1_monitor_0
Oct 21 11:20:20 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmVM2' not found (1 active resources)
Oct 21 11:20:20 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmVM2' to the rsc list (2 active resources)
Oct 21 11:20:20 bl460g1n7 crmd[12487]:     info: do_lrm_rsc_op: Performing key=16:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM2_monitor_0
Oct 21 11:20:20 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmVM3' not found (2 active resources)
Oct 21 11:20:20 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmVM3' to the rsc list (3 active resources)
Oct 21 11:20:20 bl460g1n7 crmd[12487]:     info: do_lrm_rsc_op: Performing key=17:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM3_monitor_0
Oct 21 11:20:20 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith6-1' not found (3 active resources)
Oct 21 11:20:20 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmStonith6-1' to the rsc list (4 active resources)
Oct 21 11:20:20 bl460g1n7 crmd[12487]:     info: do_lrm_rsc_op: Performing key=18:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-1_monitor_0
Oct 21 11:20:20 bl460g1n7 VirtualDomain(prmVM1)[12523]: DEBUG: Virtual domain vm1 is currently shut off.
Oct 21 11:20:20 bl460g1n7 VirtualDomain(prmVM2)[12525]: DEBUG: Virtual domain vm2 is currently shut off.
Oct 21 11:20:20 bl460g1n7 VirtualDomain(prmVM3)[12527]: DEBUG: Virtual domain vm3 is currently shut off.
Oct 21 11:20:20 bl460g1n7 crm_resource[12617]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n7 crm_resource[12617]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9caf0 for uid=0 gid=0 pid=12617 id=7c9e61d4-655b-4947-838e-dca2dcdc0b76
Oct 21 11:20:20 bl460g1n7 crm_resource[12619]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n7 crm_resource[12619]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xfe2520 for uid=0 gid=0 pid=12619 id=cf9fec80-fdb5-467f-b7ba-e4ff0206ff47
Oct 21 11:20:20 bl460g1n7 crm_resource[12621]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n7 crm_resource[12621]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9caf0 for uid=0 gid=0 pid=12621 id=89fd9711-deef-4c26-886c-caec9cbff03a
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n7 crm_resource[12623]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n7 crm_resource[12623]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9caf0 for uid=0 gid=0 pid=12623 id=b6d8f366-2d26-4b00-8fd3-0b8ddff22ef4
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 crm_resource[12625]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n7 crm_resource[12625]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xfe2520 for uid=0 gid=0 pid=12625 id=5d02d421-02ad-49cd-b159-40367cec2095
Oct 21 11:20:20 bl460g1n7 crm_resource[12627]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n7 crm_resource[12627]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9d170 for uid=0 gid=0 pid=12627 id=577e1d6e-a72c-4984-8fd9-fb8b5f61042d
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM1"]/utilization//nvpair[@name="cpu"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM2"]/utilization//nvpair[@name="cpu"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM3"]/utilization//nvpair[@name="cpu"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.8.7)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.9.1)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.10.1)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.11.1)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:20 bl460g1n7 cib[12628]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-4.raw
Oct 21 11:20:20 bl460g1n7 crm_resource[12638]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:20 bl460g1n7 crm_resource[12638]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0x1013950 for uid=0 gid=0 pid=12638 id=036d6450-bcf9-4620-bb56-309af479a76c
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:20 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n7 crm_resource[12644]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n7 crm_resource[12644]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0x1013950 for uid=0 gid=0 pid=12644 id=72c408a8-d5d0-40b7-b3c9-2efa3baaa3ce
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 crm_resource[12646]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n7 crm_resource[12646]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9d170 for uid=0 gid=0 pid=12646 id=164b877d-fbe8-4049-b271-4c48cc809bc4
Oct 21 11:20:21 bl460g1n7 cib[12628]:     info: write_cib_contents: Wrote version 0.9.0 of the CIB to disk (digest: 6706479431a247ca0e49bb8b54057d59)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 crm_resource[12648]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n7 crm_resource[12648]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0x1013950 for uid=0 gid=0 pid=12648 id=481e1569-ce8c-4b9a-b1ad-016dc5bdbe29
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 crm_resource[12650]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n7 crm_resource[12652]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:21 bl460g1n7 crm_resource[12652]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n7 crm_resource[12650]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM1"]/utilization//nvpair[@name="hv_memory"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9d170 for uid=0 gid=0 pid=12650 id=231896b0-0257-4460-b579-8daac32aec0f
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xfe2520 for uid=0 gid=0 pid=12652 id=d5c20d67-9fd3-4c14-baf9-94ad75adf536
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 cib[12628]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.Rczi7M (digest: /var/lib/pacemaker/cib/cib.tvNdGj)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM2"]/utilization//nvpair[@name="hv_memory"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section //cib/configuration/resources//*[@id="prmVM3"]/utilization//nvpair[@name="hv_memory"]: No such device or address (rc=-6, origin=local/crm_resource/3, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section /cib: OK (rc=0, origin=local/crm_resource/4, version=0.11.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section resources to master (origin=local/crm_resource/5)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.12.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.13.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section resources: OK (rc=0, origin=bl460g1n6/crm_resource/5, version=0.14.1)
Oct 21 11:20:21 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:21 bl460g1n7 cib[12653]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-5.raw
Oct 21 11:20:21 bl460g1n7 cib[12653]:     info: write_cib_contents: Wrote version 0.12.0 of the CIB to disk (digest: f9164a9a0ff267cf10ecd24bb92b6cec)
Oct 21 11:20:21 bl460g1n7 cib[12653]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.SIxYxM (digest: /var/lib/pacemaker/cib/cib.t29vsj)
Oct 21 11:20:21 bl460g1n7 cib[12654]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-6.raw
Oct 21 11:20:21 bl460g1n7 cib[12654]:     info: write_cib_contents: Wrote version 0.14.0 of the CIB to disk (digest: 0fa312cc44a6846316c5d9be6b00ad4b)
Oct 21 11:20:21 bl460g1n7 cib[12654]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.9mLzyW (digest: /var/lib/pacemaker/cib/cib.v0OtPt)
Oct 21 11:20:21 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (3 active devices)
Oct 21 11:20:21 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:21 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:22 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-1' to the device list (4 active devices)
Oct 21 11:20:22 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:22 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:23 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:23 bl460g1n7 stonith-ng[12483]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:23 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:23 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:23 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:24 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:24 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:24 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:24 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:25 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:25 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-1 has been disabled on bl460g1n7: score=-INFINITY
Oct 21 11:20:25 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:25 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:25 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:26 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:26 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-1' from the device list (4 active devices)
Oct 21 11:20:26 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:26 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:27 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-1' to the device list (5 active devices)
Oct 21 11:20:27 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:27 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:27 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:28 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:28 bl460g1n7 stonith-ng[12483]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:28 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:28 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:28 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:29 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:29 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:29 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:29 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:30 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:30 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-1 has been disabled on bl460g1n7: score=-INFINITY
Oct 21 11:20:30 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:30 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:30 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:31 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:31 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-1' from the device list (4 active devices)
Oct 21 11:20:31 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:31 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:32 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-1' to the device list (5 active devices)
Oct 21 11:20:32 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:32 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:32 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:33 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:33 bl460g1n7 stonith-ng[12483]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:33 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:33 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:33 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:34 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:34 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:34 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:34 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:35 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:35 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-1 has been disabled on bl460g1n7: score=-INFINITY
Oct 21 11:20:35 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:35 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:35 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:36 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:36 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-1' from the device list (4 active devices)
Oct 21 11:20:36 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:36 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:37 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-1' to the device list (5 active devices)
Oct 21 11:20:37 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:37 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:37 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:38 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:38 bl460g1n7 stonith-ng[12483]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:38 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:38 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:38 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:39 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:39 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:39 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:39 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:40 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:40 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-1 has been disabled on bl460g1n7: score=-INFINITY
Oct 21 11:20:40 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:40 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:40 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:41 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:41 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-1' from the device list (4 active devices)
Oct 21 11:20:41 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:41 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:42 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-1' to the device list (5 active devices)
Oct 21 11:20:42 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:42 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:42 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:43 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:43 bl460g1n7 stonith-ng[12483]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:43 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:43 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:43 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:44 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:44 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:44 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:44 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:45 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:45 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-1 has been disabled on bl460g1n7: score=-INFINITY
Oct 21 11:20:45 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:45 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:45 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:46 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:46 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-1' from the device list (4 active devices)
Oct 21 11:20:46 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:46 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:47 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-1' to the device list (5 active devices)
Oct 21 11:20:47 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:47 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:47 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:48 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:48 bl460g1n7 stonith-ng[12483]:     info: update_cib_stonith_devices: Updating device list from the cib: new resource
Oct 21 11:20:48 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-1' from the device list (4 active devices)
Oct 21 11:20:48 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:48 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:49 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-1' to the device list (5 active devices)
Oct 21 11:20:49 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith6-2' from the device list (4 active devices)
Oct 21 11:20:49 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith6-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:49 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:50 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith6-2' to the device list (5 active devices)
Oct 21 11:20:50 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-1 has been disabled on bl460g1n7: score=-INFINITY
Oct 21 11:20:50 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith7-2' from the device list (4 active devices)
Oct 21 11:20:50 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith7-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:50 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:51 bl460g1n7 crmd[12487]:  warning: crm_ipc_send: Request 44 to lrmd (0x18efdc0) failed: Resource temporarily unavailable (-11)
Oct 21 11:20:51 bl460g1n7 crmd[12487]:    error: lrmd_send_command: Couldn't perform lrmd_rsc_info operation (timeout=30000): -11: Connection timed out (110)
Oct 21 11:20:51 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith7-2' to the device list (5 active devices)
Oct 21 11:20:51 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-1' from the device list (4 active devices)
Oct 21 11:20:51 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-1 is allowed on bl460g1n7: score=0
Oct 21 11:20:51 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: crm_ipc_send: Sending to lrmd (0x18efdc0) is disabled until pending reply is received
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: lrmd_send_command: Couldn't perform lrmd_rsc_register operation (timeout=0): -114: Connection timed out (110)
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: crm_ipc_send: Sending to lrmd (0x18efdc0) is disabled until pending reply is received
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: lrmd_send_command: Couldn't perform lrmd_rsc_info operation (timeout=30000): -114: Connection timed out (110)
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: get_lrm_resource: Could not add resource prmStonith6-2 to LRM
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: register_fsa_error_adv: Resetting the current action list
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: do_lrm_invoke: Invalid resource definition
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_lrm_invoke: bad input   <create_request_adv origin="te_rsc_command" t="crmd" version="3.0.7" subt="request" reference="lrm_invoke-tengine-1382322020-46" crm_task="lrm_invoke" crm_sys_to="lrmd" crm_sys_from="tengine" crm_host_to="bl460g1n7" src="bl460g1n6">
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_lrm_invoke: bad input     <crm_xml>
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_lrm_invoke: bad input       <rsc_op id="19" operation="monitor" operation_key="prmStonith6-2_monitor_0" on_node="bl460g1n7" on_node_uuid="3232261593" transition-key="19:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5">
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_lrm_invoke: bad input         <primitive id="prmStonith6-2" class="stonith" type="external/ssh"/>
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_lrm_invoke: bad input         <attributes CRM_meta_op_target_rc="7" CRM_meta_timeout="60000" crm_feature_set="3.0.7" hostlist="bl460g1n6" pcmk_reboot_retries="1" pcmk_reboot_timeout="40s"/>
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_lrm_invoke: bad input       </rsc_op>
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_lrm_invoke: bad input     </crm_xml>
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_lrm_invoke: bad input   </create_request_adv>
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_log: FSA: Input I_FAIL from get_lrm_resource() received in state S_NOT_DC
Oct 21 11:20:52 bl460g1n7 crmd[12487]:   notice: do_state_transition: State transition S_NOT_DC -> S_RECOVERY [ input=I_FAIL cause=C_FSA_INTERNAL origin=get_lrm_resource ]
Oct 21 11:20:52 bl460g1n7 crmd[12487]:  warning: do_recover: Fast-tracking shutdown in response to errors
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: do_log: FSA: Input I_TERMINATE from do_recover() received in state S_RECOVERY
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: do_state_transition: State transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE cause=C_FSA_INTERNAL origin=do_recover ]
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: do_shutdown: Disconnecting STONITH...
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: tengine_stonith_connection_destroy: Fencing daemon disconnected
Oct 21 11:20:52 bl460g1n7 crmd[12487]:   notice: lrm_state_verify_stopped: Stopped 0 recurring operations at shutdown (4 ops remaining)
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: lrm_state_verify_stopped: 4 pending LRM operations at shutdown
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: lrm_state_verify_stopped: Pending action: prmVM3:13 (prmVM3_monitor_0)
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: lrm_state_verify_stopped: Pending action: prmVM2:9 (prmVM2_monitor_0)
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: lrm_state_verify_stopped: Pending action: prmVM1:5 (prmVM1_monitor_0)
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: lrm_state_verify_stopped: Pending action: prmStonith6-1:17 (prmStonith6-1_monitor_0)
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: do_lrm_control: Disconnecting from the LRM
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: lrmd_api_disconnect: Disconnecting from lrmd service
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: lrmd_ipc_connection_destroy: IPC connection destroyed
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: lrm_connection_destroy: LRM Connection disconnected
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: lrmd_api_disconnect: Disconnecting from lrmd service
Oct 21 11:20:52 bl460g1n7 crmd[12487]:   notice: do_lrm_control: Disconnected from the LRM
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: crm_cluster_disconnect: Disconnecting from cluster infrastructure: corosync
Oct 21 11:20:52 bl460g1n7 crmd[12487]:   notice: terminate_cs_connection: Disconnecting from Corosync
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_lib_cpg_leave:1544 got leave request on 0x7f79ecef6b40
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_lib_cpg_finalize:1582 cpg finalize for conn=0x7f79ecef6b40
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12487-32)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12487-32) state:2
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:cpg_lib_exit_fn:1043 exit_fn for conn=0x7f79ecef6b40
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cpg-response-12468-12487-32-header
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: crm_cluster_disconnect: Disconnected from corosync
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: do_ha_control: Disconnected from the cluster
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: do_cib_control: Disconnecting CIB
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cpg-event-12468-12487-32-header
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: crmd_cib_connection_destroy: Connection to the CIB terminated...
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: qb_ipcs_us_withdraw: withdrawing server sockets
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: do_exit: Performing A_EXIT_0 - gracefully exiting the CRMd
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: do_exit: [crmd] stopped (0)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cpg-request-12468-12487-32-header
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: crmd_exit: Dropping I_TERMINATE: [ state=S_TERMINATE cause=C_FSA_INTERNAL origin=do_stop ]
Oct 21 11:20:52 bl460g1n7 attrd[12485]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: crmd_quorum_destroy: connection closed
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: crmd_cs_destroy: connection closed
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: crmd_init: 12487 stopped: OK (0)
Oct 21 11:20:52 bl460g1n7 crmd[12487]:    error: crmd_fast_exit: Could not recover from internal error
Oct 21 11:20:52 bl460g1n7 crmd[12487]:     info: crm_xml_cleanup: Cleaning up memory from libxml2
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12487-33)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12487-33) state:2
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:quorum_lib_exit_fn:328 lib_exit_fn: conn=0x7f79ecefa720
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-quorum-response-12468-12487-33-header
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-quorum-event-12468-12487-33-header
Oct 21 11:20:52 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-quorum-request-12468-12487-33-header
Oct 21 11:20:52 bl460g1n7 pacemakerd[12477]:    error: pcmk_child_exit: Child process crmd (12487) exited: Generic Pacemaker error (201)
Oct 21 11:20:52 bl460g1n7 pacemakerd[12477]:   notice: pcmk_process_exit: Respawning failed child process: crmd
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procleave:1276 got procleave message from cluster node -1062705702
Oct 21 11:20:52 bl460g1n7 pacemakerd[12477]:     info: start_child: Using uid=189 and group=189 for process crmd
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procleave:1276 got procleave message from cluster node -1062705703
Oct 21 11:20:52 bl460g1n7 pacemakerd[12477]:     info: start_child: Forked child 12703 for process crmd
Oct 21 11:20:52 bl460g1n7 crmd[12703]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:52 bl460g1n7 crmd[12703]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: crm_log_init: Changed active directory to /var/lib/heartbeat/cores/hacluster
Oct 21 11:20:52 bl460g1n7 crmd[12703]:   notice: main: CRM Git Version: b6d42ed
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: do_log: FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: get_cluster_type: Verifying cluster type: 'corosync'
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: get_cluster_type: Assuming an active 'corosync' cluster
Oct 21 11:20:52 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0x1013950 for uid=189 gid=189 pid=12703 id=91b45748-e104-464b-bb4a-c2b1086081d7
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: do_cib_control: CIB connection established
Oct 21 11:20:52 bl460g1n7 crmd[12703]:   notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12703]
Oct 21 11:20:52 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.14.1)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:cpg_lib_init_fn:1459 lib_init_fn: conn=0x7f79ecef6b40, cpd=0x7f79ecef75e4
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: crm_get_peer: Created entry 5e56141d-5d93-4896-a51f-dc47edfebd3c/0x135d490 for node (null)/3232261593 (1 total)
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: crm_get_peer: Node 3232261593 has uuid 3232261593
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: crm_update_peer_proc: cluster_connect_cpg: Node (null)[3232261593] - corosync-cpg is now online
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: init_cs_connection_once: Connection to 'corosync': established
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12703]
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecef8c30
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705702 (r(0) ip(192.168.101.218) r(1) ip(192.168.102.218) ) for pid 1841
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CPG   ] cpg.c:message_handler_req_exec_cpg_procjoin:1260 got procjoin message from cluster node -1062705703 (r(0) ip(192.168.101.217) r(1) ip(192.168.102.217) ) for pid 12703
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12703-33)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12703-33) state:2
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n7 crmd[12703]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n7 crmd[12703]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecef8c30
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: crm_get_peer: Node 3232261593 is now known as bl460g1n7
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12703-33-header
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: peer_update_callback: bl460g1n7 is now (null)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12703-33-header
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12703-33-header
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12703]
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:quorum_lib_init_fn:316 lib_init_fn: conn=0x7f79ecef8c30
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_gettype:471 got quorum_type request on 0x7f79ecef8c30
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_getquorate:395 got quorate request on 0x7f79ecef8c30
Oct 21 11:20:52 bl460g1n7 crmd[12703]:   notice: cluster_connect_quorum: Quorum acquired
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:412 got trackstart request on 0x7f79ecef8c30
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:message_handler_req_lib_quorum_trackstart:420 sending initial status to 0x7f79ecef8c30
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QUORUM] vsf_quorum.c:send_library_notification:359 sending quorum notification to 0x7f79ecef8c30, length = 60
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12703]
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecef9bd0
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12703-34)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12703-34) state:2
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecef9bd0
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12703-34-header
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12703-34-header
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12703-34-header
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12703]
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecef99b0
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12703-34)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12703-34) state:2
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecef99b0
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12703-34-header
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12703-34-header
Oct 21 11:20:52 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12703-34-header
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: do_ha_control: Connected to the cluster
Oct 21 11:20:52 bl460g1n7 crmd[12703]:     info: lrmd_ipc_connect: Connecting to lrmd
Oct 21 11:20:52 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.14.1)
Oct 21 11:20:52 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-1' to the device list (5 active devices)
Oct 21 11:20:52 bl460g1n7 stonith-ng[12483]:     info: stonith_device_remove: Removed 'prmStonith8-2' from the device list (4 active devices)
Oct 21 11:20:52 bl460g1n7 stonith-ng[12483]:     info: cib_device_update: Device prmStonith8-2 is allowed on bl460g1n7: score=0
Oct 21 11:20:52 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:53 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Added 'prmStonith8-2' to the device list (5 active devices)
Oct 21 11:20:53 bl460g1n7 stonith-ng[12483]:     info: crm_client_new: Connecting 0x1e614a0 for uid=0 gid=0 pid=12484 id=fc68ded4-0ee7-4f68-9f10-1b9b05109295
Oct 21 11:20:53 bl460g1n7 stonith-ng[12483]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:53 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed register from lrmd.12484: OK (0)
Oct 21 11:20:53 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_notify from lrmd.12484: OK (0)
Oct 21 11:20:53 bl460g1n7 lrmd[12484]:  warning: qb_ipcs_event_sendv: new_event_notification (12484-12487-6): Broken pipe (32)
Oct 21 11:20:53 bl460g1n7 lrmd[12484]:  warning: send_client_notify: Notification of client crmd/eaecbe18-3d6e-4619-bc5a-af0e1d4415fe failed
Oct 21 11:20:53 bl460g1n7 lrmd[12484]:  warning: send_client_notify: Notification of client crmd/eaecbe18-3d6e-4619-bc5a-af0e1d4415fe failed
Oct 21 11:20:53 bl460g1n7 lrmd[12484]:  warning: send_client_notify: Notification of client crmd/eaecbe18-3d6e-4619-bc5a-af0e1d4415fe failed
Oct 21 11:20:53 bl460g1n7 lrmd[12484]:  warning: send_client_notify: Notification of client crmd/eaecbe18-3d6e-4619-bc5a-af0e1d4415fe failed
Oct 21 11:20:53 bl460g1n7 lrmd[12484]:     info: crm_client_new: Connecting 0xef8350 for uid=189 gid=189 pid=12703 id=346e0c4c-6f69-4c4b-93fb-ba62c8870072
Oct 21 11:20:53 bl460g1n7 lrmd[12484]:     info: crm_client_destroy: Destroying 4 events
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: do_lrm_control: LRM connection established
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: do_started: Delaying start, no membership data (0000000000100000)
Oct 21 11:20:53 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.14.1)
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: pcmk_quorum_notification: Membership 16: quorum retained (3)
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: crm_get_peer: Created entry 7cda00ef-9c87-4924-8a9d-67bcf779cef9/0x14a1110 for node (null)/3232261592 (2 total)
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: crm_get_peer: Node 3232261592 has uuid 3232261592
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261592
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12703]
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecef99b0
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12703-34)
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12703-34) state:2
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecef99b0
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12703-34-header
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12703-34-header
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12703-34-header
Oct 21 11:20:53 bl460g1n7 crmd[12703]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261592
Oct 21 11:20:53 bl460g1n7 crmd[12703]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261592] - state is now member (was (null))
Oct 21 11:20:53 bl460g1n7 crmd[12703]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node bl460g1n7[3232261593] - state is now member (was (null))
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: peer_update_callback: bl460g1n7 is now member (was (null))
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: crm_get_peer: Created entry ac3fc444-4769-4f0d-ae17-0e5d1aa8b9cb/0x14a6880 for node (null)/3232261594 (3 total)
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: crm_get_peer: Node 3232261594 has uuid 3232261594
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: pcmk_quorum_notification: Obtaining name for new node 3232261594
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12703]
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecef99b0
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12703-34)
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12703-34) state:2
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecef99b0
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12703-34-header
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12703-34-header
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12703-34-header
Oct 21 11:20:53 bl460g1n7 crmd[12703]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261594
Oct 21 11:20:53 bl460g1n7 crmd[12703]:   notice: crm_update_peer_state: pcmk_quorum_notification: Node (null)[3232261594] - state is now member (was (null))
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [12703]
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_open_2:236 shm size:1048589; real_size:1052672; rb->word_size:263168
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_created:272 connection created
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_init_fn:306 lib_init_fn: conn=0x7f79ecef99b0
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_dispatch_connection_request:757 HUP conn (12468-12703-34)
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ipcs.c:qb_ipcs_disconnect:605 qb_ipcs_disconnect(12468-12703-34) state:2
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] loop_poll_epoll.c:_del:117 epoll_ctl(del): Bad file descriptor (9)
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_closed:417 cs_ipcs_connection_closed() 
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [CMAP  ] cmap.c:cmap_lib_exit_fn:325 exit_fn for conn=0x7f79ecef99b0
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [MAIN  ] ipc_glue.c:cs_ipcs_connection_destroyed:390 cs_ipcs_connection_destroyed() 
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-response-12468-12703-34-header
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-event-12468-12703-34-header
Oct 21 11:20:53 bl460g1n7 corosync[12468]:   [QB    ] ringbuffer.c:qb_rb_close:299 Free'ing ringbuffer: /dev/shm/qb-cmap-request-12468-12703-34-header
Oct 21 11:20:53 bl460g1n7 crmd[12703]:   notice: corosync_node_name: Unable to get node name for nodeid 3232261593
Oct 21 11:20:53 bl460g1n7 crmd[12703]:   notice: get_node_name: Defaulting to uname -n for the local corosync node name
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: qb_ipcs_us_publish: server name: crmd
Oct 21 11:20:53 bl460g1n7 crmd[12703]:   notice: do_started: The local CRM is operational
Oct 21 11:20:53 bl460g1n7 crmd[12703]:     info: do_log: FSA: Input I_PENDING from do_started() received in state S_STARTING
Oct 21 11:20:53 bl460g1n7 crmd[12703]:   notice: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
Oct 21 11:20:53 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_slave operation for section 'all': OK (rc=0, origin=local/crmd/5, version=0.14.1)
Oct 21 11:20:53 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/60, version=0.14.2)
Oct 21 11:20:53 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/61, version=0.14.3)
Oct 21 11:20:53 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/62, version=0.14.4)
Oct 21 11:20:53 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/63, version=0.14.5)
Oct 21 11:20:53 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/64, version=0.14.6)
Oct 21 11:20:53 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/65, version=0.14.7)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/66, version=0.14.8)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/67, version=0.14.9)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/68, version=0.14.10)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/69, version=0.14.11)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/70, version=0.14.12)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/71, version=0.14.13)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/72, version=0.14.14)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/73, version=0.14.15)
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: pcmk_cpg_membership: Joined[0.0] crmd.3232261593 
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: pcmk_cpg_membership: Member[0.0] crmd.3232261592 
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261592] - corosync-cpg is now online
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: pcmk_cpg_membership: Member[0.1] crmd.3232261593 
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: pcmk_cpg_membership: Member[0.2] crmd.3232261594 
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: crm_update_peer_proc: pcmk_cpg_membership: Node (null)[3232261594] - corosync-cpg is now online
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: crm_get_peer: Node 3232261592 is now known as bl460g1n6
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: peer_update_callback: bl460g1n6 is now member
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith7-1' not found (4 active resources)
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmStonith7-1' to the rsc list (5 active resources)
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=20:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-1_monitor_0
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmStonith7-1 in state S_PENDING
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith7-2' not found (5 active resources)
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmStonith7-2' to the rsc list (6 active resources)
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=21:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-2_monitor_0
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmStonith7-2 in state S_PENDING
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith8-1' not found (6 active resources)
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmStonith8-1' to the rsc list (7 active resources)
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=22:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-1_monitor_0
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmStonith8-1 in state S_PENDING
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith8-2' not found (7 active resources)
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmStonith8-2' to the rsc list (8 active resources)
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=23:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-2_monitor_0
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmStonith8-2 in state S_PENDING
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: crm_get_peer: Node 3232261594 is now known as bl460g1n8
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: peer_update_callback: bl460g1n8 is now member
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmPing' not found (8 active resources)
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmPing:1' not found (8 active resources)
Oct 21 11:20:54 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmPing' to the rsc list (9 active resources)
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=24:1:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmPing_monitor_0
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Discarding attempt to perform action monitor on prmPing in state S_PENDING
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: services_os_action_execute: Managed ping_meta-data_0 process 12729 exited with rc=0
Oct 21 11:20:54 bl460g1n7 crmd[12703]:     info: update_dc: Set DC to bl460g1n6 (3.0.7)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/6, version=0.14.15)
Oct 21 11:20:54 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/7, version=0.14.15)
Oct 21 11:20:55 bl460g1n7 stonith-ng[12483]:     info: crm_client_new: Connecting 0x1d4b610 for uid=189 gid=189 pid=12703 id=aa19df49-48bc-4998-b540-503d9a3e9781
Oct 21 11:20:55 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed register from crmd.12703: OK (0)
Oct 21 11:20:55 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_notify from crmd.12703: OK (0)
Oct 21 11:20:55 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_notify from crmd.12703: OK (0)
Oct 21 11:20:56 bl460g1n7 crmd[12703]:     info: erase_status_tag: Deleting xpath: //node_state[@uname='bl460g1n7']/transient_attributes
Oct 21 11:20:56 bl460g1n7 crmd[12703]:     info: update_attrd_helper: Connecting to attrd... 5 retries remaining
Oct 21 11:20:56 bl460g1n7 attrd[12485]:     info: crm_client_new: Connecting 0x1e28be0 for uid=189 gid=189 pid=12703 id=eb884e99-d2fd-43c4-b840-5db1a3c30423
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_replace: Digest matched on replace from bl460g1n6: 9e817daca05ff13f2c868ff292f7703a
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_replace: Replaced 0.14.15 with 0.14.15 from bl460g1n6
Oct 21 11:20:56 bl460g1n7 crmd[12703]:     info: do_log: FSA: Input I_NOT_DC from do_cl_join_finalize_respond() received in state S_PENDING
Oct 21 11:20:56 bl460g1n7 crmd[12703]:   notice: do_state_transition: State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
Oct 21 11:20:56 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting terminate[bl460g1n7] = (null)
Oct 21 11:20:56 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting shutdown[bl460g1n7] = (null)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_replace operation for section 'all': OK (rc=0, origin=bl460g1n6/crmd/76, version=0.14.15)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_delete operation for section //node_state[@uname='bl460g1n7']/transient_attributes to master (origin=local/crmd/8)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n6']/lrm: OK (rc=0, origin=bl460g1n6/crmd/80, version=0.14.16)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n7']/transient_attributes: OK (rc=0, origin=bl460g1n6/crmd/8, version=0.14.17)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n8']/transient_attributes: OK (rc=0, origin=bl460g1n6/crmd/8, version=0.14.18)
Oct 21 11:20:56 bl460g1n7 cib[12736]:     info: write_cib_contents: Archived previous version as /var/lib/pacemaker/cib/cib-7.raw
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/81, version=0.14.19)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n7']/lrm: OK (rc=0, origin=bl460g1n6/crmd/82, version=0.14.20)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/83, version=0.14.21)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section //node_state[@uname='bl460g1n8']/lrm: OK (rc=0, origin=bl460g1n6/crmd/84, version=0.14.22)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/85, version=0.14.23)
Oct 21 11:20:56 bl460g1n7 cib[12736]:     info: write_cib_contents: Wrote version 0.14.0 of the CIB to disk (digest: 7ed85a3c13981ed3a8cf49c95bde5769)
Oct 21 11:20:56 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/87, version=0.14.24)
Oct 21 11:20:56 bl460g1n7 cib[12736]:     info: retrieveCib: Reading cluster configuration from: /var/lib/pacemaker/cib/cib.1asUFh (digest: /var/lib/pacemaker/cib/cib.6q3tYl)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=5:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM1_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=6:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM2_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=7:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmVM3_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=8:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-1_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: process_lrm_event: LRM operation prmStonith6-1_monitor_0 (call=42, rc=7, cib-update=9, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 lrmd[12484]:     info: process_lrmd_get_rsc_info: Resource 'prmStonith6-2' not found (9 active resources)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/9)
Oct 21 11:20:58 bl460g1n7 lrmd[12484]:     info: process_lrmd_rsc_register: Added 'prmStonith6-2' to the rsc list (10 active resources)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=9:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-2_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=10:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-1_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: process_lrm_event: LRM operation prmStonith6-2_monitor_0 (call=46, rc=7, cib-update=10, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/10)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: process_lrm_event: LRM operation prmStonith7-1_monitor_0 (call=47, rc=7, cib-update=11, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/11)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=11:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith7-2_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: process_lrm_event: LRM operation prmStonith7-2_monitor_0 (call=48, rc=7, cib-update=12, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/12)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=12:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-1_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: process_lrm_event: LRM operation prmStonith8-1_monitor_0 (call=49, rc=7, cib-update=13, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/13)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=13:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-2_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=14:2:7:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmPing_monitor_0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: process_lrm_event: LRM operation prmStonith8-2_monitor_0 (call=50, rc=7, cib-update=14, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/14)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmPing_monitor_0 (call=51, rc=7, cib-update=15, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/9, version=0.14.25)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/15)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/10, version=0.14.26)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/11, version=0.14.27)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/12, version=0.14.28)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/9, version=0.14.29)
Oct 21 11:20:58 bl460g1n7 VirtualDomain(prmVM1)[12738]: DEBUG: Virtual domain vm1 is currently shut off.
Oct 21 11:20:58 bl460g1n7 VirtualDomain(prmVM2)[12741]: DEBUG: Virtual domain vm2 is currently shut off.
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/10, version=0.14.30)
Oct 21 11:20:58 bl460g1n7 VirtualDomain(prmVM3)[12744]: DEBUG: Virtual domain vm3 is currently shut off.
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/11, version=0.14.31)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/12, version=0.14.32)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/13, version=0.14.33)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/14, version=0.14.34)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/13, version=0.14.35)
Oct 21 11:20:58 bl460g1n7 crm_resource[12838]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n7 crm_resource[12838]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/14, version=0.14.36)
Oct 21 11:20:58 bl460g1n7 crm_resource[12840]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n7 crm_resource[12840]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/15, version=0.14.37)
Oct 21 11:20:58 bl460g1n7 crm_resource[12842]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n7 crm_resource[12842]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/15, version=0.14.38)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9ca10 for uid=0 gid=0 pid=12838 id=2641a13b-84d2-46bb-bd3e-b5cb6449beef
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9d170 for uid=0 gid=0 pid=12840 id=3d8d1691-fc2f-4a83-b87b-9419702624d0
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0x10294f0 for uid=0 gid=0 pid=12842 id=f8a2dbdd-ef1d-4c52-8c51-97c36c8f5efc
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n7 crm_resource[12856]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n7 crm_resource[12856]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9ca10 for uid=0 gid=0 pid=12856 id=f726491b-fe18-4267-a9f6-66f6b52eeb04
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n7 crm_resource[12858]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n7 crm_resource[12858]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0xf9d170 for uid=0 gid=0 pid=12858 id=11e4c385-5a4d-4e7f-9776-9d1caf147367
Oct 21 11:20:58 bl460g1n7 crm_resource[12860]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:20:58 bl460g1n7 crm_resource[12860]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_new: Connecting 0x10294f0 for uid=0 gid=0 pid=12860 id=9bb235b4-870b-4014-86aa-e4ff07a2313b
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_query operation for section 'all': OK (rc=0, origin=local/crm_resource/2, version=0.14.38)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: services_os_action_execute: Managed VirtualDomain_meta-data_0 process 12861 exited with rc=0
Oct 21 11:20:58 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmVM2_monitor_0 (call=38, rc=7, cib-update=16, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/16)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmVM1_monitor_0 (call=36, rc=7, cib-update=17, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/17)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmVM3_monitor_0 (call=40, rc=7, cib-update=18, confirmed=true) not running
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/18)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/16, version=0.14.39)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/17, version=0.14.40)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/18, version=0.14.41)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/16, version=0.14.42)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/17, version=0.14.43)
Oct 21 11:20:58 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting probe_complete[bl460g1n7] = true
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=35:2:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-1_start_0
Oct 21 11:20:58 bl460g1n7 lrmd[12484]:     info: log_execute: executing - rsc:prmStonith6-1 action:start call_id:52
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/18, version=0.14.44)
Oct 21 11:20:58 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=51:2:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-1_start_0
Oct 21 11:20:58 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/90, version=0.14.45)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/91, version=0.14.46)
Oct 21 11:20:58 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/92, version=0.14.47)
Oct 21 11:20:59 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Device 'prmStonith6-1' already existed in device list (5 active devices)
Oct 21 11:20:59 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_device_register from lrmd.12484: OK (0)
Oct 21 11:20:59 bl460g1n7 lrmd[12484]:     info: log_execute: executing - rsc:prmStonith8-1 action:start call_id:53
Oct 21 11:20:59 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=61:2:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmPing_start_0
Oct 21 11:20:59 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:20:59 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:20:59 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/10, version=0.14.48)
Oct 21 11:20:59 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/93, version=0.14.49)
Oct 21 11:21:00 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Device 'prmStonith8-1' already existed in device list (5 active devices)
Oct 21 11:21:00 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_device_register from lrmd.12484: OK (0)
Oct 21 11:21:00 bl460g1n7 lrmd[12484]:     info: log_execute: executing - rsc:prmPing action:start call_id:54
Oct 21 11:21:00 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:00 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:00 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:00 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/11, version=0.14.50)
Oct 21 11:21:00 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/19, version=0.14.51)
Oct 21 11:21:00 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/20, version=0.14.52)
Oct 21 11:21:01 bl460g1n7 attrd_updater[12914]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:21:01 bl460g1n7 attrd_updater[12914]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:21:01 bl460g1n7 attrd[12485]:     info: crm_client_new: Connecting 0x1e27e90 for uid=0 gid=0 pid=12914 id=821b1407-95e2-4ceb-a80c-be729f64c9a6
Oct 21 11:21:01 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting default_ping_set[bl460g1n7] = 100
Oct 21 11:21:01 bl460g1n7 attrd[12485]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:21:01 bl460g1n7 lrmd[12484]:     info: log_finished: finished - rsc:prmPing action:start call_id:54 pid:12869 exit-code:0 exec-time:1037ms queue-time:0ms
Oct 21 11:21:01 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmPing_start_0 (call=54, rc=0, cib-update=19, confirmed=true) ok
Oct 21 11:21:01 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/19)
Oct 21 11:21:01 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/attrd/12, version=0.14.53)
Oct 21 11:21:01 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/19, version=0.14.54)
Oct 21 11:21:01 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-1:12873 [ Performing: stonith -t external/ipmi -S ]
Oct 21 11:21:01 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-1:12873 [ success:  0 ]
Oct 21 11:21:01 bl460g1n7 lrmd[12484]:     info: log_finished: finished - rsc:prmStonith8-1 action:start call_id:53  exit-code:0 exec-time:2103ms queue-time:0ms
Oct 21 11:21:01 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmStonith8-1_start_0 (call=53, rc=0, cib-update=20, confirmed=true) ok
Oct 21 11:21:01 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-1:12872 [ Performing: stonith -t external/ipmi -S ]
Oct 21 11:21:01 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-1:12872 [ success:  0 ]
Oct 21 11:21:01 bl460g1n7 lrmd[12484]:     info: log_finished: finished - rsc:prmStonith6-1 action:start call_id:52  exit-code:0 exec-time:3107ms queue-time:0ms
Oct 21 11:21:01 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmStonith6-1_start_0 (call=52, rc=0, cib-update=21, confirmed=true) ok
Oct 21 11:21:01 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/20)
Oct 21 11:21:01 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/21)
Oct 21 11:21:01 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/20, version=0.14.55)
Oct 21 11:21:01 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/21, version=0.14.56)
Oct 21 11:21:03 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=17:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-1_monitor_360000
Oct 21 11:21:03 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=18:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-2_start_0
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:03 bl460g1n7 lrmd[12484]:     info: log_execute: executing - rsc:prmStonith6-2 action:start call_id:56
Oct 21 11:21:03 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=35:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-1_monitor_360000
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Device 'prmStonith6-2' already existed in device list (5 active devices)
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_device_register from lrmd.12484: OK (0)
Oct 21 11:21:03 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=36:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-2_start_0
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:03 bl460g1n7 lrmd[12484]:     info: log_execute: executing - rsc:prmStonith8-2 action:start call_id:58
Oct 21 11:21:03 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=50:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmPing_monitor_10000
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action metadata for agent fence_legacy (target=(null))
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:   notice: stonith_device_register: Device 'prmStonith8-2' already existed in device list (5 active devices)
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_device_register from lrmd.12484: OK (0)
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:03 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting probe_complete[bl460g1n7] = true
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:03 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/95, version=0.14.57)
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-2:12920 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-2:12920 [ success:  0 ]
Oct 21 11:21:04 bl460g1n7 lrmd[12484]:     info: log_finished: finished - rsc:prmStonith6-2 action:start call_id:56  exit-code:0 exec-time:1064ms queue-time:0ms
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/21, version=0.14.58)
Oct 21 11:21:04 bl460g1n7 attrd_updater[13008]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:21:04 bl460g1n7 attrd_updater[13008]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:21:04 bl460g1n7 attrd[12485]:     info: crm_client_new: Connecting 0x1e27e90 for uid=0 gid=0 pid=13008 id=1175168e-bef5-4537-ac2b-d0e134d7be01
Oct 21 11:21:04 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting default_ping_set[bl460g1n7] = 100
Oct 21 11:21:04 bl460g1n7 attrd[12485]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:21:04 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmStonith6-2_start_0 (call=56, rc=0, cib-update=22, confirmed=true) ok
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/22)
Oct 21 11:21:04 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmPing_monitor_10000 (call=59, rc=0, cib-update=23, confirmed=false) ok
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/23)
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-1:12916 [ Performing: stonith -t external/ipmi -S ]
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-1:12916 [ success:  0 ]
Oct 21 11:21:04 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmStonith6-1_monitor_360000 (call=55, rc=0, cib-update=24, confirmed=false) ok
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-2:12933 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-2:12933 [ success:  0 ]
Oct 21 11:21:04 bl460g1n7 lrmd[12484]:     info: log_finished: finished - rsc:prmStonith8-2 action:start call_id:58  exit-code:0 exec-time:1066ms queue-time:0ms
Oct 21 11:21:04 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmStonith8-2_start_0 (call=58, rc=0, cib-update=25, confirmed=true) ok
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/22, version=0.14.59)
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/24)
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/25)
Oct 21 11:21:04 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=19:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith6-2_monitor_10000
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/23, version=0.14.60)
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/22, version=0.14.61)
Oct 21 11:21:04 bl460g1n7 crmd[12703]:     info: do_lrm_rsc_op: Performing key=37:3:0:38db9a68-056c-4e65-8658-75f0c3cc91e5 op=prmStonith8-2_monitor_10000
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/23, version=0.14.62)
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/24, version=0.14.63)
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/25, version=0.14.64)
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-1:12932 [ Performing: stonith -t external/ipmi -S ]
Oct 21 11:21:04 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-1:12932 [ success:  0 ]
Oct 21 11:21:04 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmStonith8-1_monitor_360000 (call=57, rc=0, cib-update=26, confirmed=false) ok
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/26)
Oct 21 11:21:04 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/26, version=0.14.65)
Oct 21 11:21:05 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-2:13017 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:05 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-2:13017 [ success:  0 ]
Oct 21 11:21:05 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmStonith6-2_monitor_10000 (call=60, rc=0, cib-update=27, confirmed=false) ok
Oct 21 11:21:05 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/24, version=0.14.66)
Oct 21 11:21:05 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/27)
Oct 21 11:21:05 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-2:13018 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:05 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-2:13018 [ success:  0 ]
Oct 21 11:21:05 bl460g1n7 crmd[12703]:   notice: process_lrm_event: LRM operation prmStonith8-2_monitor_10000 (call=61, rc=0, cib-update=28, confirmed=false) ok
Oct 21 11:21:05 bl460g1n7 cib[12482]:     info: cib_process_request: Forwarding cib_modify operation for section status to master (origin=local/crmd/28)
Oct 21 11:21:05 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/27, version=0.14.67)
Oct 21 11:21:05 bl460g1n7 cib[12482]:     info: cib_process_request: Completed cib_apply_diff operation for section status: OK (rc=0, origin=bl460g1n6/crmd/28, version=0.14.68)
Oct 21 11:21:15 bl460g1n7 attrd_updater[13062]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:21:15 bl460g1n7 attrd_updater[13062]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:21:15 bl460g1n7 attrd[12485]:     info: crm_client_new: Connecting 0x1e25780 for uid=0 gid=0 pid=13062 id=2300260d-3b9b-488c-9ac4-f5975d03f49b
Oct 21 11:21:15 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting default_ping_set[bl460g1n7] = 100
Oct 21 11:21:15 bl460g1n7 attrd[12485]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:21:15 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:15 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:15 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:15 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:16 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-2:13063 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:16 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-2:13063 [ success:  0 ]
Oct 21 11:21:16 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-2:13064 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:16 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-2:13064 [ success:  0 ]
Oct 21 11:21:26 bl460g1n7 attrd_updater[13105]:   notice: crm_add_logfile: Additional logging available in /var/log/ha-debug
Oct 21 11:21:26 bl460g1n7 attrd_updater[13105]:    debug: crm_update_callsites: Enabling callsites based on priority=7, files=(null), functions=(null), formats=(null), tags=(null)
Oct 21 11:21:26 bl460g1n7 attrd[12485]:     info: crm_client_new: Connecting 0x1e25780 for uid=0 gid=0 pid=13105 id=517bf940-525e-4008-b2c1-f2104e00b585
Oct 21 11:21:26 bl460g1n7 attrd[12485]:     info: attrd_client_message: Broadcasting default_ping_set[bl460g1n7] = 100
Oct 21 11:21:26 bl460g1n7 attrd[12485]:     info: crm_client_destroy: Destroying 0 events
Oct 21 11:21:26 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:26 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:26 bl460g1n7 stonith-ng[12483]:     info: stonith_command: Processed st_execute from lrmd.12484: Operation now in progress (-115)
Oct 21 11:21:26 bl460g1n7 stonith-ng[12483]:     info: stonith_action_create: Initiating action monitor for agent fence_legacy (target=(null))
Oct 21 11:21:27 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-2:13106 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:27 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith6-2:13106 [ success:  0 ]
Oct 21 11:21:27 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-2:13107 [ Performing: stonith -t external/ssh -S ]
Oct 21 11:21:27 bl460g1n7 stonith-ng[12483]:     info: log_operation: prmStonith8-2:13107 [ success:  0 ]
Oct 21 11:21:35 bl460g1n7 corosync[12468]:   [QB    ] ipc_shm.c:qb_ipcs_shm_connect:295 connecting to client [13173]
