Jul 24 11:08:05 node02 heartbeat: [2350]: WARN: Logging daemon is disabled --enabling logging daemon is recommended
Jul 24 11:08:05 node02 heartbeat: [2350]: info: **************************
Jul 24 11:08:05 node02 heartbeat: [2350]: info: Configuration validated. Starting heartbeat 3.0.5
Jul 24 11:08:05 node02 heartbeat: [2351]: info: heartbeat: version 3.0.5
Jul 24 11:08:05 node02 heartbeat: [2351]: info: Heartbeat generation: 1341987739
Jul 24 11:08:05 node02 heartbeat: [2351]: info: glib: UDP Broadcast heartbeat started on port 694 (694) interface eth2
Jul 24 11:08:05 node02 heartbeat: [2351]: info: glib: UDP Broadcast heartbeat closed on port 694 interface eth2 - Status: 1
Jul 24 11:08:05 node02 heartbeat: [2351]: notice: Using watchdog device: /dev/watchdog
Jul 24 11:08:05 node02 heartbeat: [2351]: info: Local status now set to: 'up'
Jul 24 11:08:05 node02 heartbeat: [2351]: info: Link node02:eth2 up.
Jul 24 11:08:05 node02 heartbeat: [2351]: info: Link node01:eth2 up.
Jul 24 11:08:06 node02 heartbeat: [2351]: info: Status update for node node01: status active
Jul 24 11:08:06 node02 heartbeat: [2351]: info: Comm_now_up(): updating status to active
Jul 24 11:08:06 node02 heartbeat: [2351]: info: Local status now set to: 'active'
Jul 24 11:08:06 node02 heartbeat: [2351]: info: Starting child client "/usr/lib64/heartbeat/ccm" (495,489)
Jul 24 11:08:06 node02 heartbeat: [2351]: info: Starting child client "/usr/lib64/heartbeat/cib" (495,489)
Jul 24 11:08:06 node02 heartbeat: [2351]: info: Starting child client "/usr/lib64/heartbeat/lrmd -r" (0,0)
Jul 24 11:08:06 node02 heartbeat: [2351]: info: Starting child client "/usr/lib64/heartbeat/stonithd" (0,0)
Jul 24 11:08:06 node02 heartbeat: [2351]: info: Starting child client "/usr/lib64/heartbeat/attrd" (495,489)
Jul 24 11:08:06 node02 heartbeat: [2351]: info: Starting child client "/usr/lib64/heartbeat/crmd" (495,489)
Jul 24 11:08:06 node02 heartbeat: [2364]: info: Starting "/usr/lib64/heartbeat/crmd" as uid 495  gid 489 (pid 2364)
Jul 24 11:08:06 node02 heartbeat: [2363]: info: Starting "/usr/lib64/heartbeat/attrd" as uid 495  gid 489 (pid 2363)
Jul 24 11:08:06 node02 heartbeat: [2361]: info: Starting "/usr/lib64/heartbeat/lrmd -r" as uid 0  gid 0 (pid 2361)
Jul 24 11:08:06 node02 heartbeat: [2360]: info: Starting "/usr/lib64/heartbeat/cib" as uid 495  gid 489 (pid 2360)
Jul 24 11:08:06 node02 heartbeat: [2362]: info: Starting "/usr/lib64/heartbeat/stonithd" as uid 0  gid 0 (pid 2362)
Jul 24 11:08:06 node02 heartbeat: [2359]: info: Starting "/usr/lib64/heartbeat/ccm" as uid 495  gid 489 (pid 2359)
Jul 24 11:08:06 node02 lrmd: [2361]: info: enabling coredumps
Jul 24 11:08:06 node02 lrmd: [2361]: info: Started.
Jul 24 11:08:06 node02 cib: [2360]: info: Invoked: /usr/lib64/heartbeat/cib 
Jul 24 11:08:06 node02 cib: [2360]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.xml (digest: /var/lib/heartbeat/crm/cib.xml.sig)
Jul 24 11:08:06 node02 attrd: [2363]: info: Invoked: /usr/lib64/heartbeat/attrd 
Jul 24 11:08:06 node02 attrd: [2363]: info: main: Starting up
Jul 24 11:08:06 node02 ccm: [2359]: info: Hostname: node02
Jul 24 11:08:06 node02 crmd: [2364]: info: Invoked: /usr/lib64/heartbeat/crmd 
Jul 24 11:08:06 node02 crmd: [2364]: info: main: CRM Hg Version: 066152e
Jul 24 11:08:06 node02 crmd: [2364]: info: crmd_init: Starting crmd
Jul 24 11:08:06 node02 cib: [2360]: WARN: retrieveCib: Cluster configuration not found: /var/lib/heartbeat/crm/cib.xml
Jul 24 11:08:06 node02 cib: [2360]: WARN: readCibXmlFile: Primary configuration corrupt or unusable, trying backup...
Jul 24 11:08:06 node02 cib: [2360]: WARN: readCibXmlFile: Continuing with an empty configuration.
Jul 24 11:08:06 node02 heartbeat: [2351]: info: the send queue length from heartbeat to client ccm is set to 1024
Jul 24 11:08:06 node02 attrd: [2363]: info: register_heartbeat_conn: Hostname: node02
Jul 24 11:08:06 node02 attrd: [2363]: info: register_heartbeat_conn: UUID: 554581e7-6f22-45a9-a637-9a6606234a2d
Jul 24 11:08:06 node02 attrd: [2363]: info: crm_cluster_connect: Connecting to Heartbeat
Jul 24 11:08:06 node02 attrd: [2363]: info: main: Cluster connection active
Jul 24 11:08:06 node02 attrd: [2363]: info: main: Accepting attribute updates
Jul 24 11:08:06 node02 attrd: [2363]: info: main: Starting mainloop...
Jul 24 11:08:06 node02 cib: [2360]: info: startCib: CIB Initialization completed successfully
Jul 24 11:08:06 node02 heartbeat: [2351]: info: the send queue length from heartbeat to client attrd is set to 1024
Jul 24 11:08:06 node02 stonithd: [2362]: info: register_heartbeat_conn: Hostname: node02
Jul 24 11:08:06 node02 stonithd: [2362]: info: register_heartbeat_conn: UUID: 554581e7-6f22-45a9-a637-9a6606234a2d
Jul 24 11:08:06 node02 stonithd: [2362]: info: crm_cluster_connect: Connecting to Heartbeat
Jul 24 11:08:06 node02 heartbeat: [2351]: info: the send queue length from heartbeat to client stonithd is set to 1024
Jul 24 11:08:06 node02 stonithd: [2362]: notice: /usr/lib64/heartbeat/stonithd start up successfully.
Jul 24 11:08:06 node02 cib: [2360]: info: register_heartbeat_conn: Hostname: node02
Jul 24 11:08:06 node02 cib: [2360]: info: register_heartbeat_conn: UUID: 554581e7-6f22-45a9-a637-9a6606234a2d
Jul 24 11:08:06 node02 cib: [2360]: info: crm_cluster_connect: Connecting to Heartbeat
Jul 24 11:08:06 node02 cib: [2360]: info: ccm_connect: Registering with CCM...
Jul 24 11:08:06 node02 cib: [2360]: WARN: ccm_connect: CCM Activation failed
Jul 24 11:08:06 node02 cib: [2360]: WARN: ccm_connect: CCM Connection failed 1 times (30 max)
Jul 24 11:08:07 node02 heartbeat: [2351]: info: the send queue length from heartbeat to client cib is set to 1024
Jul 24 11:08:07 node02 crmd: [2364]: info: do_cib_control: Could not connect to the CIB service: connection failed
Jul 24 11:08:07 node02 crmd: [2364]: WARN: do_cib_control: Couldn't complete CIB registration 1 times... pause and retry
Jul 24 11:08:07 node02 crmd: [2364]: info: crmd_init: Starting crmd's mainloop
Jul 24 11:08:09 node02 crmd: [2364]: info: crm_timer_popped: Wait Timer (I_NULL) just popped!
Jul 24 11:08:09 node02 cib: [2360]: info: ccm_connect: Registering with CCM...
Jul 24 11:08:09 node02 cib: [2360]: WARN: ccm_connect: CCM Activation failed
Jul 24 11:08:09 node02 cib: [2360]: WARN: ccm_connect: CCM Connection failed 2 times (30 max)
Jul 24 11:08:10 node02 crmd: [2364]: info: do_cib_control: Could not connect to the CIB service: connection failed
Jul 24 11:08:10 node02 crmd: [2364]: WARN: do_cib_control: Couldn't complete CIB registration 2 times... pause and retry
Jul 24 11:08:12 node02 crmd: [2364]: info: crm_timer_popped: Wait Timer (I_NULL) just popped!
Jul 24 11:08:12 node02 cib: [2360]: info: ccm_connect: Registering with CCM...
Jul 24 11:08:12 node02 cib: [2360]: info: cib_init: Requesting the list of configured nodes
Jul 24 11:08:12 node02 cib: [2360]: info: cib_init: Starting cib mainloop
Jul 24 11:08:12 node02 cib: [2360]: info: cib_client_status_callback: Status update: Client node02/cib now has status [join]
Jul 24 11:08:12 node02 cib: [2360]: info: crm_new_peer: Node 0 is now known as node02
Jul 24 11:08:12 node02 cib: [2360]: info: crm_update_peer_proc: node02.cib is now online
Jul 24 11:08:12 node02 cib: [2368]: info: write_cib_contents: Wrote version 0.0.0 of the CIB to disk (digest: fe034d007b1a32fe64dda5b238bfefef)
Jul 24 11:08:12 node02 cib: [2368]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.3D6Nlf (digest: /var/lib/heartbeat/crm/cib.BxSaLx)
Jul 24 11:08:13 node02 ccm: [2359]: info: Break tie for 2 nodes cluster
Jul 24 11:08:13 node02 cib: [2360]: info: cib_client_status_callback: Status update: Client node02/cib now has status [online]
Jul 24 11:08:13 node02 cib: [2360]: info: mem_handle_event: Got an event OC_EV_MS_NEW_MEMBERSHIP from ccm
Jul 24 11:08:13 node02 cib: [2360]: info: mem_handle_event: instance=1, nodes=1, new=1, lost=0, n_idx=0, new_idx=0, old_idx=3
Jul 24 11:08:13 node02 cib: [2360]: info: cib_ccm_msg_callback: Processing CCM event=NEW MEMBERSHIP (id=1)
Jul 24 11:08:13 node02 cib: [2360]: info: crm_get_peer: Node node02 now has id: 1
Jul 24 11:08:13 node02 cib: [2360]: info: crm_update_peer: Node node02: id=1 state=member (new) addr=(null) votes=-1 born=1 seen=1 proc=00000000000000000000000000000100
Jul 24 11:08:13 node02 cib: [2360]: info: crm_update_peer_proc: node02.ais is now online
Jul 24 11:08:13 node02 cib: [2360]: info: crm_update_peer_proc: node02.crmd is now online
Jul 24 11:08:13 node02 crmd: [2364]: info: do_cib_control: CIB connection established
Jul 24 11:08:13 node02 heartbeat: [2351]: WARN: 1 lost packet(s) for [node01] [17:19]
Jul 24 11:08:13 node02 heartbeat: [2351]: info: No pkts missing from node01!
Jul 24 11:08:13 node02 crmd: [2364]: info: register_heartbeat_conn: Hostname: node02
Jul 24 11:08:13 node02 crmd: [2364]: info: register_heartbeat_conn: UUID: 554581e7-6f22-45a9-a637-9a6606234a2d
Jul 24 11:08:13 node02 crmd: [2364]: info: crm_cluster_connect: Connecting to Heartbeat
Jul 24 11:08:13 node02 cib: [2360]: info: cib_client_status_callback: Status update: Client node01/cib now has status [online]
Jul 24 11:08:14 node02 heartbeat: [2351]: info: the send queue length from heartbeat to client crmd is set to 1024
Jul 24 11:08:14 node02 cib: [2360]: info: crm_new_peer: Node 0 is now known as node01
Jul 24 11:08:14 node02 cib: [2360]: info: crm_update_peer_proc: node01.cib is now online
Jul 24 11:08:14 node02 cib: [2360]: info: mem_handle_event: Got an event OC_EV_MS_INVALID from ccm
Jul 24 11:08:14 node02 cib: [2360]: info: mem_handle_event: no mbr_track info
Jul 24 11:08:14 node02 cib: [2360]: info: mem_handle_event: Got an event OC_EV_MS_NEW_MEMBERSHIP from ccm
Jul 24 11:08:14 node02 cib: [2360]: info: mem_handle_event: instance=2, nodes=2, new=1, lost=0, n_idx=0, new_idx=2, old_idx=4
Jul 24 11:08:14 node02 cib: [2360]: info: cib_ccm_msg_callback: Processing CCM event=NEW MEMBERSHIP (id=2)
Jul 24 11:08:14 node02 cib: [2360]: info: crm_update_peer: Node node01: id=0 state=member (new) addr=(null) votes=-1 born=2 seen=2 proc=00000000000000000000000000000100
Jul 24 11:08:14 node02 cib: [2360]: info: crm_update_peer_proc: node01.ais is now online
Jul 24 11:08:14 node02 cib: [2360]: info: crm_update_peer_proc: node01.crmd is now online
Jul 24 11:08:14 node02 crmd: [2364]: info: do_ha_control: Connected to the cluster
Jul 24 11:08:14 node02 crmd: [2364]: info: do_ccm_control: CCM connection established... waiting for first callback
Jul 24 11:08:14 node02 crmd: [2364]: info: do_started: Delaying start, CCM (0000000000100000) not connected
Jul 24 11:08:14 node02 crmd: [2364]: info: config_query_callback: Checking for expired actions every 900000ms
Jul 24 11:08:14 node02 crmd: [2364]: notice: crmd_client_status_callback: Status update: Client node02/crmd now has status [online] (DC=false)
Jul 24 11:08:14 node02 crmd: [2364]: info: crm_new_peer: Node 0 is now known as node02
Jul 24 11:08:14 node02 crmd: [2364]: info: crm_update_peer_proc: node02.crmd is now online
Jul 24 11:08:14 node02 crmd: [2364]: info: crmd_client_status_callback: Not the DC
Jul 24 11:08:14 node02 crmd: [2364]: notice: crmd_client_status_callback: Status update: Client node01/crmd now has status [online] (DC=false)
Jul 24 11:08:15 node02 heartbeat: [2351]: WARN: 1 lost packet(s) for [node01] [23:25]
Jul 24 11:08:15 node02 heartbeat: [2351]: info: No pkts missing from node01!
Jul 24 11:08:15 node02 crmd: [2364]: info: crm_new_peer: Node 0 is now known as node01
Jul 24 11:08:15 node02 crmd: [2364]: info: crm_update_peer_proc: node01.crmd is now online
Jul 24 11:08:15 node02 crmd: [2364]: info: crmd_client_status_callback: Not the DC
Jul 24 11:08:15 node02 crmd: [2364]: notice: crmd_client_status_callback: Status update: Client node02/crmd now has status [online] (DC=false)
Jul 24 11:08:15 node02 crmd: [2364]: info: crmd_client_status_callback: Not the DC
Jul 24 11:08:15 node02 crmd: [2364]: notice: crmd_client_status_callback: Status update: Client node01/crmd now has status [online] (DC=false)
Jul 24 11:08:15 node02 crmd: [2364]: info: crmd_client_status_callback: Not the DC
Jul 24 11:08:15 node02 crmd: [2364]: info: mem_handle_event: Got an event OC_EV_MS_NEW_MEMBERSHIP from ccm
Jul 24 11:08:15 node02 crmd: [2364]: info: mem_handle_event: instance=2, nodes=2, new=2, lost=0, n_idx=0, new_idx=0, old_idx=4
Jul 24 11:08:15 node02 crmd: [2364]: info: crmd_ccm_msg_callback: Quorum (re)attained after event=NEW MEMBERSHIP (id=2)
Jul 24 11:08:15 node02 crmd: [2364]: info: ccm_event_detail: NEW MEMBERSHIP: trans=2, nodes=2, new=2, lost=0 n_idx=0, new_idx=0, old_idx=4
Jul 24 11:08:15 node02 crmd: [2364]: info: ccm_event_detail: 	CURRENT: node02 [nodeid=1, born=1]
Jul 24 11:08:15 node02 crmd: [2364]: info: ccm_event_detail: 	CURRENT: node01 [nodeid=0, born=2]
Jul 24 11:08:15 node02 crmd: [2364]: info: ccm_event_detail: 	NEW:     node02 [nodeid=1, born=1]
Jul 24 11:08:15 node02 crmd: [2364]: info: ccm_event_detail: 	NEW:     node01 [nodeid=0, born=2]
Jul 24 11:08:15 node02 crmd: [2364]: info: crm_get_peer: Node node02 now has id: 1
Jul 24 11:08:15 node02 crmd: [2364]: info: crm_update_peer: Node node02: id=1 state=member (new) addr=(null) votes=-1 born=1 seen=2 proc=00000000000000000000000000000200
Jul 24 11:08:15 node02 crmd: [2364]: info: crm_update_peer_proc: node02.ais is now online
Jul 24 11:08:15 node02 crmd: [2364]: info: crm_update_peer: Node node01: id=0 state=member (new) addr=(null) votes=-1 born=2 seen=2 proc=00000000000000000000000000000200
Jul 24 11:08:15 node02 crmd: [2364]: info: crm_update_peer_proc: node01.ais is now online
Jul 24 11:08:15 node02 crmd: [2364]: info: do_started: The local CRM is operational
Jul 24 11:08:15 node02 crmd: [2364]: info: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
Jul 24 11:08:16 node02 attrd: [2363]: info: cib_connect: Connected to the CIB after 2 signon attempts
Jul 24 11:08:16 node02 attrd: [2363]: info: cib_connect: Sending full refresh
Jul 24 11:08:36 node02 crmd: [2364]: info: crm_timer_popped: Election Trigger (I_DC_TIMEOUT) just popped!
Jul 24 11:08:36 node02 crmd: [2364]: WARN: do_log: FSA: Input I_DC_TIMEOUT from crm_timer_popped() received in state S_PENDING
Jul 24 11:08:36 node02 crmd: [2364]: info: do_state_transition: State transition S_PENDING -> S_ELECTION [ input=I_DC_TIMEOUT cause=C_TIMER_POPPED origin=crm_timer_popped ]
Jul 24 11:08:37 node02 crmd: [2364]: info: do_election_count_vote: Election 2 (owner: d6a57ba8-8f59-4ce8-8ae7-7b757c721f7c) pass: vote from node01 (Age)
Jul 24 11:08:38 node02 crmd: [2364]: info: do_state_transition: State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC cause=C_FSA_INTERNAL origin=do_election_check ]
Jul 24 11:08:38 node02 crmd: [2364]: info: do_te_control: Registering TE UUID: b5b7e743-bc71-49ef-bf53-c1fa81be29b5
Jul 24 11:08:38 node02 crmd: [2364]: info: set_graph_functions: Setting custom graph functions
Jul 24 11:08:38 node02 crmd: [2364]: info: unpack_graph: Unpacked transition -1: 0 actions in 0 synapses
Jul 24 11:08:38 node02 crmd: [2364]: info: start_subsystem: Starting sub-system "pengine"
Jul 24 11:08:38 node02 pengine: [2369]: info: Invoked: /usr/lib64/heartbeat/pengine 
Jul 24 11:08:38 node02 pengine: [2369]: info: main: Starting pengine
Jul 24 11:08:42 node02 crmd: [2364]: info: do_dc_takeover: Taking over DC status for this partition
Jul 24 11:08:42 node02 cib: [2360]: info: cib_process_readwrite: We are now in R/W mode
Jul 24 11:08:42 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_master for section 'all' (origin=local/crmd/5, version=0.0.0): ok (rc=0)
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="0" num_updates="0" />
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: + <cib crm_feature_set="3.0.1" admin_epoch="0" epoch="1" num_updates="1" />
Jul 24 11:08:42 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section cib (origin=local/crmd/6, version=0.1.1): ok (rc=0)
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="1" num_updates="1" />
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="2" num_updates="1" >
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +   <configuration >
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +     <crm_config >
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +       <cluster_property_set id="cib-bootstrap-options" __crm_diff_marker__="added:top" >
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.0.12-066152e" />
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +       </cluster_property_set>
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +     </crm_config>
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +   </configuration>
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: + </cib>
Jul 24 11:08:42 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section crm_config (origin=local/crmd/9, version=0.2.1): ok (rc=0)
Jul 24 11:08:42 node02 crmd: [2364]: info: join_make_offer: Making join offers based on membership 2
Jul 24 11:08:42 node02 crmd: [2364]: info: do_dc_join_offer_all: join-1: Waiting on 2 outstanding join acks
Jul 24 11:08:42 node02 crmd: [2364]: info: te_connect_stonith: Attempting connection to fencing daemon...
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="2" num_updates="1" />
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="3" num_updates="1" >
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +   <configuration >
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +     <crm_config >
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +       <cluster_property_set id="cib-bootstrap-options" >
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="Heartbeat" __crm_diff_marker__="added:top" />
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +       </cluster_property_set>
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +     </crm_config>
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: +   </configuration>
Jul 24 11:08:42 node02 cib: [2360]: info: log_data_element: cib:diff: + </cib>
Jul 24 11:08:42 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section crm_config (origin=local/crmd/12, version=0.3.1): ok (rc=0)
Jul 24 11:08:42 node02 cib: [2370]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-0.raw
Jul 24 11:08:42 node02 cib: [2370]: info: write_cib_contents: Wrote version 0.3.0 of the CIB to disk (digest: bcd749be236199e0f718a9588b835079)
Jul 24 11:08:42 node02 cib: [2370]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.j2CjRC (digest: /var/lib/heartbeat/crm/cib.FvA1jg)
Jul 24 11:08:43 node02 crmd: [2364]: info: te_connect_stonith: Connected
Jul 24 11:08:43 node02 crmd: [2364]: info: config_query_callback: Checking for expired actions every 900000ms
Jul 24 11:08:43 node02 crmd: [2364]: info: config_query_callback: Checking for expired actions every 900000ms
Jul 24 11:08:43 node02 crmd: [2364]: info: update_dc: Set DC to node02 (3.0.1)
Jul 24 11:08:43 node02 crmd: [2364]: info: config_query_callback: Checking for expired actions every 900000ms
Jul 24 11:08:44 node02 crmd: [2364]: info: do_state_transition: State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED cause=C_FSA_INTERNAL origin=check_join_state ]
Jul 24 11:08:44 node02 crmd: [2364]: info: do_state_transition: All 2 cluster nodes responded to the join offer.
Jul 24 11:08:44 node02 crmd: [2364]: info: do_dc_join_finalize: join-1: Syncing the CIB from node02 to the rest of the cluster
Jul 24 11:08:44 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_sync for section 'all' (origin=local/crmd/17, version=0.3.1): ok (rc=0)
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="3" num_updates="1" />
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="4" num_updates="1" >
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +   <configuration >
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +     <nodes >
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +       <node id="d6a57ba8-8f59-4ce8-8ae7-7b757c721f7c" uname="node01" type="normal" __crm_diff_marker__="added:top" />
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +     </nodes>
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +   </configuration>
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: + </cib>
Jul 24 11:08:44 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/18, version=0.4.1): ok (rc=0)
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="4" num_updates="1" />
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="5" num_updates="1" >
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +   <configuration >
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +     <nodes >
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +       <node id="554581e7-6f22-45a9-a637-9a6606234a2d" uname="node02" type="normal" __crm_diff_marker__="added:top" />
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +     </nodes>
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: +   </configuration>
Jul 24 11:08:44 node02 cib: [2360]: info: log_data_element: cib:diff: + </cib>
Jul 24 11:08:44 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/19, version=0.5.1): ok (rc=0)
Jul 24 11:08:44 node02 cib: [2371]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-1.raw
Jul 24 11:08:44 node02 cib: [2371]: info: write_cib_contents: Wrote version 0.5.0 of the CIB to disk (digest: 1d4d92a73653653a7a08770f2b187669)
Jul 24 11:08:44 node02 cib: [2371]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.zQnzX3 (digest: /var/lib/heartbeat/crm/cib.CkYYvL)
Jul 24 11:08:44 node02 crmd: [2364]: info: update_attrd: Connecting to attrd...
Jul 24 11:08:44 node02 attrd: [2363]: info: find_hash_entry: Creating hash entry for terminate
Jul 24 11:08:44 node02 attrd: [2363]: info: find_hash_entry: Creating hash entry for shutdown
Jul 24 11:08:44 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='node02']/transient_attributes (origin=local/crmd/20, version=0.5.1): ok (rc=0)
Jul 24 11:08:44 node02 crmd: [2364]: info: erase_xpath_callback: Deletion of "//node_state[@uname='node02']/transient_attributes": ok (rc=0)
Jul 24 11:08:45 node02 crmd: [2364]: info: do_dc_join_ack: join-1: Updating node state to member for node02
Jul 24 11:08:45 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='node02']/lrm (origin=local/crmd/21, version=0.5.1): ok (rc=0)
Jul 24 11:08:45 node02 crmd: [2364]: info: erase_xpath_callback: Deletion of "//node_state[@uname='node02']/lrm": ok (rc=0)
Jul 24 11:08:45 node02 crmd: [2364]: info: do_dc_join_ack: join-1: Updating node state to member for node01
Jul 24 11:08:45 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='node01']/transient_attributes (origin=node01/crmd/9, version=0.5.2): ok (rc=0)
Jul 24 11:08:45 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='node01']/lrm (origin=local/crmd/23, version=0.5.2): ok (rc=0)
Jul 24 11:08:45 node02 crmd: [2364]: info: erase_xpath_callback: Deletion of "//node_state[@uname='node01']/lrm": ok (rc=0)
Jul 24 11:08:45 node02 crmd: [2364]: info: do_state_transition: State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED cause=C_FSA_INTERNAL origin=check_join_state ]
Jul 24 11:08:45 node02 crmd: [2364]: info: populate_cib_nodes_ha: Requesting the list of configured nodes
Jul 24 11:08:47 node02 crmd: [2364]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jul 24 11:08:47 node02 crmd: [2364]: info: do_dc_join_final: Ensuring DC, quorum and node attributes are up-to-date
Jul 24 11:08:47 node02 crmd: [2364]: info: crm_update_quorum: Updating quorum status to true (call=27)
Jul 24 11:08:47 node02 crmd: [2364]: info: abort_transition_graph: do_te_invoke:185 - Triggered transition abort (complete=1) : Peer Cancelled
Jul 24 11:08:47 node02 crmd: [2364]: info: do_pe_invoke: Query 28: Requesting the current CIB: S_POLICY_ENGINE
Jul 24 11:08:47 node02 attrd: [2363]: info: attrd_local_callback: Sending full refresh (origin=crmd)
Jul 24 11:08:47 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/25, version=0.5.3): ok (rc=0)
Jul 24 11:08:47 node02 crmd: [2364]: info: abort_transition_graph: need_abort:59 - Triggered transition abort (complete=1) : Non-status change
Jul 24 11:08:47 node02 crmd: [2364]: info: need_abort: Aborting on change to have-quorum
Jul 24 11:08:47 node02 crmd: [2364]: info: do_pe_invoke: Query 29: Requesting the current CIB: S_POLICY_ENGINE
Jul 24 11:08:47 node02 cib: [2360]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="5" num_updates="3" />
Jul 24 11:08:47 node02 cib: [2360]: info: log_data_element: cib:diff: + <cib have-quorum="1" dc-uuid="554581e7-6f22-45a9-a637-9a6606234a2d" admin_epoch="0" epoch="6" num_updates="1" />
Jul 24 11:08:47 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section cib (origin=local/crmd/27, version=0.6.1): ok (rc=0)
Jul 24 11:08:47 node02 crmd: [2364]: info: do_pe_invoke_callback: Invoking the PE: query=29, ref=pe_calc-dc-1343095727-10, seq=2, quorate=1
Jul 24 11:08:47 node02 pengine: [2369]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jul 24 11:08:47 node02 pengine: [2369]: ERROR: unpack_resources: Resource start-up disabled since no STONITH resources have been defined
Jul 24 11:08:47 node02 pengine: [2369]: ERROR: unpack_resources: Either configure some or disable STONITH with the stonith-enabled option
Jul 24 11:08:47 node02 pengine: [2369]: ERROR: unpack_resources: NOTE: Clusters with shared data need STONITH to ensure data integrity
Jul 24 11:08:47 node02 pengine: [2369]: info: determine_online_status: Node node02 is online
Jul 24 11:08:47 node02 pengine: [2369]: info: determine_online_status: Node node01 is online
Jul 24 11:08:47 node02 pengine: [2369]: info: stage6: Delaying fencing operations until there are resources to manage
Jul 24 11:08:47 node02 crmd: [2364]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jul 24 11:08:47 node02 crmd: [2364]: info: unpack_graph: Unpacked transition 0: 2 actions in 2 synapses
Jul 24 11:08:47 node02 crmd: [2364]: info: do_te_invoke: Processing graph 0 (ref=pe_calc-dc-1343095727-10) derived from /var/lib/pengine/pe-input-0.bz2
Jul 24 11:08:47 node02 crmd: [2364]: info: te_rsc_command: Initiating action 2: probe_complete probe_complete on node01 - no waiting
Jul 24 11:08:47 node02 crmd: [2364]: info: te_rsc_command: Initiating action 3: probe_complete probe_complete on node02 (local) - no waiting
Jul 24 11:08:47 node02 attrd: [2363]: info: find_hash_entry: Creating hash entry for probe_complete
Jul 24 11:08:47 node02 attrd: [2363]: info: attrd_trigger_update: Sending flush op to all hosts for: probe_complete (true)
Jul 24 11:08:47 node02 crmd: [2364]: info: run_graph: ====================================================
Jul 24 11:08:47 node02 crmd: [2364]: notice: run_graph: Transition 0 (Complete=2, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pengine/pe-input-0.bz2): Complete
Jul 24 11:08:47 node02 crmd: [2364]: info: te_graph_trigger: Transition 0 is now complete
Jul 24 11:08:47 node02 crmd: [2364]: info: notify_crmd: Transition 0 status: done - <null>
Jul 24 11:08:47 node02 crmd: [2364]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Jul 24 11:08:47 node02 crmd: [2364]: info: do_state_transition: Starting PEngine Recheck Timer
Jul 24 11:08:47 node02 attrd: [2363]: info: attrd_perform_update: Sent update 4: probe_complete=true
Jul 24 11:08:47 node02 pengine: [2369]: info: process_pe_message: Transition 0: PEngine Input stored in: /var/lib/pengine/pe-input-0.bz2
Jul 24 11:08:47 node02 pengine: [2369]: info: process_pe_message: Configuration ERRORs found during PE processing.  Please run "crm_verify -L" to identify issues.
Jul 24 11:08:47 node02 cib: [2372]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-2.raw
Jul 24 11:08:47 node02 cib: [2372]: info: write_cib_contents: Wrote version 0.6.0 of the CIB to disk (digest: 41dd33f3cd30cb841ff18306f848bcdc)
Jul 24 11:08:47 node02 cib: [2372]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.TL2Eng (digest: /var/lib/heartbeat/crm/cib.AqAti6)
Jul 24 11:08:48 node02 attrd: [2363]: info: attrd_ha_callback: flush message from node01
Jul 24 11:08:52 node02 crmd: [2364]: info: abort_transition_graph: need_abort:59 - Triggered transition abort (complete=1) : Non-status change
Jul 24 11:08:52 node02 crmd: [2364]: info: need_abort: Aborting on change to admin_epoch
Jul 24 11:08:52 node02 crmd: [2364]: info: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
Jul 24 11:08:52 node02 crmd: [2364]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jul 24 11:08:52 node02 crmd: [2364]: info: do_pe_invoke: Query 30: Requesting the current CIB: S_POLICY_ENGINE
Jul 24 11:08:52 node02 cib: [2360]: info: cib_replace_notify: Replaced: 0.6.3 -> 0.7.1 from node01
Jul 24 11:08:52 node02 crmd: [2364]: info: populate_cib_nodes_ha: Requesting the list of configured nodes
Jul 24 11:08:52 node02 attrd: [2363]: info: do_cib_replaced: Sending full refresh
Jul 24 11:08:52 node02 attrd: [2363]: info: attrd_trigger_update: Sending flush op to all hosts for: probe_complete (true)
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: - <cib admin_epoch="0" epoch="6" num_updates="3" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: + <cib admin_epoch="0" epoch="7" num_updates="1" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +   <configuration >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +     <crm_config >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +       <cluster_property_set id="cib-bootstrap-options" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-no-quorum-policy" name="no-quorum-policy" value="ignore" __crm_diff_marker__="added:top" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="false" __crm_diff_marker__="added:top" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-startup-fencing" name="startup-fencing" value="false" __crm_diff_marker__="added:top" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         <nvpair id="cib-bootstrap-options-crmd-transition-delay" name="crmd-transition-delay" value="2s" __crm_diff_marker__="added:top" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +       </cluster_property_set>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +     </crm_config>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +     <resources >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +       <clone id="clnPingd01" __crm_diff_marker__="added:top" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         <primitive class="ocf" id="pingd01" provider="pacemaker" type="pingd" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +           <instance_attributes id="pingd01-instance_attributes" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <nvpair id="pingd01-instance_attributes-name" name="name" value="default_ping_set01" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <nvpair id="pingd01-instance_attributes-host_list" name="host_list" value="192.168.201.254" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <nvpair id="pingd01-instance_attributes-multiplier" name="multiplier" value="100" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +           </instance_attributes>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +           <operations >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <op id="pingd01-start-0s" interval="0s" name="start" on-fail="restart" timeout="100s" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <op id="pingd01-monitor-2s" interval="2s" name="monitor" on-fail="restart" timeout="100s" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <op id="pingd01-stop-0s" interval="0s" name="stop" on-fail="ignore" timeout="100s" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +           </operations>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         </primitive>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +       </clone>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +       <clone id="clnPingd02" __crm_diff_marker__="added:top" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         <primitive class="ocf" id="pingd02" provider="pacemaker" type="pingd" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +           <instance_attributes id="pingd02-instance_attributes" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <nvpair id="pingd02-instance_attributes-name" name="name" value="default_ping_set02" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <nvpair id="pingd02-instance_attributes-host_list" name="host_list" value="192.168.133.11" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <nvpair id="pingd02-instance_attributes-multiplier" name="multiplier" value="100" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +           </instance_attributes>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +           <operations >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <op id="pingd02-start-0s" interval="0s" name="start" on-fail="restart" timeout="100s" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <op id="pingd02-monitor-2s" interval="2s" name="monitor" on-fail="restart" timeout="100s" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +             <op id="pingd02-stop-0s" interval="0s" name="stop" on-fail="ignore" timeout="100s" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +           </operations>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         </primitive>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +       </clone>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +     </resources>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +     <rsc_defaults __crm_diff_marker__="added:top" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +       <meta_attributes id="rsc-options" >
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         <nvpair id="rsc-options-resource-stickiness" name="resource-stickiness" value="INFINITY" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +         <nvpair id="rsc-options-migration-threshold" name="migration-threshold" value="1" />
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +       </meta_attributes>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +     </rsc_defaults>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: +   </configuration>
Jul 24 11:08:52 node02 cib: [2360]: info: log_data_element: cib:diff: + </cib>
Jul 24 11:08:52 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_replace for section 'all' (origin=node01/cibadmin/2, version=0.7.1): ok (rc=0)
Jul 24 11:08:52 node02 cib: [2373]: info: write_cib_contents: Archived previous version as /var/lib/heartbeat/crm/cib-3.raw
Jul 24 11:08:52 node02 cib: [2373]: info: write_cib_contents: Wrote version 0.7.0 of the CIB to disk (digest: 0d90d060ebd8776722c9d65fa23801ce)
Jul 24 11:08:52 node02 cib: [2373]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.rviubt (digest: /var/lib/heartbeat/crm/cib.6UUFUw)
Jul 24 11:08:53 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/32, version=0.7.1): ok (rc=0)
Jul 24 11:08:54 node02 crmd: [2364]: info: config_query_callback: Checking for expired actions every 900000ms
Jul 24 11:08:54 node02 crmd: [2364]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_ELECTION [ input=I_ELECTION cause=C_FSA_INTERNAL origin=do_cib_replaced ]
Jul 24 11:08:54 node02 crmd: [2364]: info: update_dc: Unset DC node02
Jul 24 11:08:54 node02 crmd: [2364]: info: do_pe_invoke: Query 34: Requesting the current CIB: S_ELECTION
Jul 24 11:08:54 node02 crmd: [2364]: info: do_state_transition: State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC cause=C_FSA_INTERNAL origin=do_election_check ]
Jul 24 11:08:54 node02 crmd: [2364]: info: start_subsystem: Starting sub-system "pengine"
Jul 24 11:08:54 node02 crmd: [2364]: WARN: start_subsystem: Client pengine already running as pid 2369
Jul 24 11:08:58 node02 crmd: [2364]: info: do_dc_takeover: Taking over DC status for this partition
Jul 24 11:08:58 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_master for section 'all' (origin=local/crmd/35, version=0.7.1): ok (rc=0)
Jul 24 11:08:58 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section cib (origin=local/crmd/36, version=0.7.1): ok (rc=0)
Jul 24 11:08:58 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section crm_config (origin=local/crmd/38, version=0.7.1): ok (rc=0)
Jul 24 11:08:58 node02 crmd: [2364]: info: do_dc_join_offer_all: join-2: Waiting on 2 outstanding join acks
Jul 24 11:08:58 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section crm_config (origin=local/crmd/40, version=0.7.1): ok (rc=0)
Jul 24 11:08:58 node02 crmd: [2364]: info: config_query_callback: Checking for expired actions every 900000ms
Jul 24 11:08:59 node02 crmd: [2364]: info: update_dc: Set DC to node02 (3.0.1)
Jul 24 11:09:00 node02 crmd: [2364]: info: do_state_transition: State transition S_INTEGRATION -> S_FINALIZE_JOIN [ input=I_INTEGRATED cause=C_FSA_INTERNAL origin=check_join_state ]
Jul 24 11:09:00 node02 crmd: [2364]: info: do_state_transition: All 2 cluster nodes responded to the join offer.
Jul 24 11:09:00 node02 crmd: [2364]: info: do_dc_join_finalize: join-2: Syncing the CIB from node02 to the rest of the cluster
Jul 24 11:09:00 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_sync for section 'all' (origin=local/crmd/43, version=0.7.1): ok (rc=0)
Jul 24 11:09:00 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/44, version=0.7.1): ok (rc=0)
Jul 24 11:09:00 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/45, version=0.7.1): ok (rc=0)
Jul 24 11:09:01 node02 crmd: [2364]: info: do_dc_join_ack: join-2: Updating node state to member for node02
Jul 24 11:09:01 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='node02']/lrm (origin=local/crmd/46, version=0.7.2): ok (rc=0)
Jul 24 11:09:01 node02 crmd: [2364]: info: erase_xpath_callback: Deletion of "//node_state[@uname='node02']/lrm": ok (rc=0)
Jul 24 11:09:01 node02 attrd: [2363]: info: attrd_ha_callback: flush message from node01
Jul 24 11:09:01 node02 crmd: [2364]: info: do_dc_join_ack: join-2: Updating node state to member for node01
Jul 24 11:09:01 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_delete for section //node_state[@uname='node01']/lrm (origin=local/crmd/48, version=0.7.4): ok (rc=0)
Jul 24 11:09:01 node02 crmd: [2364]: info: erase_xpath_callback: Deletion of "//node_state[@uname='node01']/lrm": ok (rc=0)
Jul 24 11:09:01 node02 crmd: [2364]: info: do_state_transition: State transition S_FINALIZE_JOIN -> S_POLICY_ENGINE [ input=I_FINALIZED cause=C_FSA_INTERNAL origin=check_join_state ]
Jul 24 11:09:01 node02 crmd: [2364]: info: populate_cib_nodes_ha: Requesting the list of configured nodes
Jul 24 11:09:03 node02 crmd: [2364]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jul 24 11:09:03 node02 crmd: [2364]: info: do_dc_join_final: Ensuring DC, quorum and node attributes are up-to-date
Jul 24 11:09:03 node02 crmd: [2364]: info: crm_update_quorum: Updating quorum status to true (call=52)
Jul 24 11:09:03 node02 crmd: [2364]: info: abort_transition_graph: do_te_invoke:185 - Triggered transition abort (complete=1) : Peer Cancelled
Jul 24 11:09:03 node02 attrd: [2363]: info: attrd_local_callback: Sending full refresh (origin=crmd)
Jul 24 11:09:03 node02 attrd: [2363]: info: attrd_trigger_update: Sending flush op to all hosts for: probe_complete (true)
Jul 24 11:09:03 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section nodes (origin=local/crmd/50, version=0.7.5): ok (rc=0)
Jul 24 11:09:03 node02 cib: [2360]: info: cib_process_request: Operation complete: op cib_modify for section cib (origin=local/crmd/52, version=0.7.5): ok (rc=0)
Jul 24 11:09:05 node02 crmd: [2364]: info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped!
Jul 24 11:09:05 node02 crmd: [2364]: info: do_pe_invoke: Query 53: Requesting the current CIB: S_POLICY_ENGINE
Jul 24 11:09:05 node02 crmd: [2364]: info: do_pe_invoke_callback: Invoking the PE: query=53, ref=pe_calc-dc-1343095745-20, seq=2, quorate=1
Jul 24 11:09:05 node02 pengine: [2369]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jul 24 11:09:05 node02 pengine: [2369]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jul 24 11:09:05 node02 pengine: [2369]: WARN: unpack_nodes: Blind faith: not fencing unseen nodes
Jul 24 11:09:05 node02 pengine: [2369]: info: determine_online_status: Node node02 is online
Jul 24 11:09:05 node02 pengine: [2369]: info: determine_online_status: Node node01 is online
Jul 24 11:09:05 node02 pengine: [2369]: notice: clone_print:  Clone Set: clnPingd01
Jul 24 11:09:05 node02 pengine: [2369]: notice: short_print:      Stopped: [ pingd01:0 pingd01:1 ]
Jul 24 11:09:05 node02 pengine: [2369]: notice: clone_print:  Clone Set: clnPingd02
Jul 24 11:09:05 node02 pengine: [2369]: notice: short_print:      Stopped: [ pingd02:0 pingd02:1 ]
Jul 24 11:09:05 node02 pengine: [2369]: notice: RecurringOp:  Start recurring monitor (2s) for pingd01:0 on node01
Jul 24 11:09:05 node02 pengine: [2369]: notice: RecurringOp:  Start recurring monitor (2s) for pingd01:1 on node02
Jul 24 11:09:05 node02 pengine: [2369]: notice: RecurringOp:  Start recurring monitor (2s) for pingd02:0 on node01
Jul 24 11:09:05 node02 pengine: [2369]: notice: RecurringOp:  Start recurring monitor (2s) for pingd02:1 on node02
Jul 24 11:09:05 node02 pengine: [2369]: notice: LogActions: Start   pingd01:0	(node01)
Jul 24 11:09:05 node02 pengine: [2369]: notice: LogActions: Start   pingd01:1	(node02)
Jul 24 11:09:05 node02 pengine: [2369]: notice: LogActions: Start   pingd02:0	(node01)
Jul 24 11:09:05 node02 pengine: [2369]: notice: LogActions: Start   pingd02:1	(node02)
Jul 24 11:09:05 node02 crmd: [2364]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jul 24 11:09:05 node02 crmd: [2364]: info: unpack_graph: Unpacked transition 1: 21 actions in 21 synapses
Jul 24 11:09:05 node02 crmd: [2364]: info: do_te_invoke: Processing graph 1 (ref=pe_calc-dc-1343095745-20) derived from /var/lib/pengine/pe-input-1.bz2
Jul 24 11:09:05 node02 crmd: [2364]: info: te_rsc_command: Initiating action 4: monitor pingd01:0_monitor_0 on node01
Jul 24 11:09:05 node02 crmd: [2364]: info: te_rsc_command: Initiating action 7: monitor pingd01:1_monitor_0 on node02 (local)
Jul 24 11:09:05 node02 crmd: [2364]: info: do_lrm_rsc_op: Performing key=7:1:7:b5b7e743-bc71-49ef-bf53-c1fa81be29b5 op=pingd01:1_monitor_0 )
Jul 24 11:09:05 node02 pengine: [2369]: info: process_pe_message: Transition 1: PEngine Input stored in: /var/lib/pengine/pe-input-1.bz2
Jul 24 11:09:05 node02 lrmd: [2361]: info: rsc:pingd01:1 probe[2] (pid 2374)
Jul 24 11:09:05 node02 crmd: [2364]: info: te_rsc_command: Initiating action 5: monitor pingd02:0_monitor_0 on node01
Jul 24 11:09:05 node02 crmd: [2364]: info: te_rsc_command: Initiating action 8: monitor pingd02:1_monitor_0 on node02 (local)
Jul 24 11:09:05 node02 crmd: [2364]: info: do_lrm_rsc_op: Performing key=8:1:7:b5b7e743-bc71-49ef-bf53-c1fa81be29b5 op=pingd02:1_monitor_0 )
Jul 24 11:09:05 node02 lrmd: [2361]: info: rsc:pingd02:1 probe[3] (pid 2375)
Jul 24 11:09:05 node02 lrmd: [2361]: info: operation monitor[2] on pingd01:1 for client 2364: pid 2374 exited with return code 7
Jul 24 11:09:05 node02 crmd: [2364]: info: process_lrm_event: LRM operation pingd01:1_monitor_0 (call=2, rc=7, cib-update=54, confirmed=true) not running
Jul 24 11:09:05 node02 crmd: [2364]: info: match_graph_event: Action pingd01:1_monitor_0 (7) confirmed on node02 (rc=0)
Jul 24 11:09:05 node02 lrmd: [2361]: info: operation monitor[3] on pingd02:1 for client 2364: pid 2375 exited with return code 7
Jul 24 11:09:05 node02 crmd: [2364]: info: process_lrm_event: LRM operation pingd02:1_monitor_0 (call=3, rc=7, cib-update=55, confirmed=true) not running
Jul 24 11:09:05 node02 crmd: [2364]: info: match_graph_event: Action pingd02:1_monitor_0 (8) confirmed on node02 (rc=0)
Jul 24 11:09:05 node02 crmd: [2364]: info: te_rsc_command: Initiating action 6: probe_complete probe_complete on node02 (local) - no waiting
Jul 24 11:09:06 node02 crmd: [2364]: info: match_graph_event: Action pingd02:0_monitor_0 (5) confirmed on node01 (rc=0)
Jul 24 11:09:06 node02 crmd: [2364]: info: match_graph_event: Action pingd01:0_monitor_0 (4) confirmed on node01 (rc=0)
Jul 24 11:09:06 node02 crmd: [2364]: info: te_rsc_command: Initiating action 3: probe_complete probe_complete on node01 - no waiting
Jul 24 11:09:06 node02 crmd: [2364]: info: te_pseudo_action: Pseudo action 2 fired and confirmed
Jul 24 11:09:06 node02 crmd: [2364]: info: te_pseudo_action: Pseudo action 15 fired and confirmed
Jul 24 11:09:06 node02 crmd: [2364]: info: te_pseudo_action: Pseudo action 23 fired and confirmed
Jul 24 11:09:06 node02 crmd: [2364]: info: te_pseudo_action: Pseudo action 13 fired and confirmed
Jul 24 11:09:06 node02 crmd: [2364]: info: te_pseudo_action: Pseudo action 21 fired and confirmed
Jul 24 11:09:06 node02 crmd: [2364]: info: te_rsc_command: Initiating action 9: start pingd01:0_start_0 on node01
Jul 24 11:09:06 node02 crmd: [2364]: info: te_rsc_command: Initiating action 11: start pingd01:1_start_0 on node02 (local)
Jul 24 11:09:06 node02 crmd: [2364]: info: do_lrm_rsc_op: Performing key=11:1:0:b5b7e743-bc71-49ef-bf53-c1fa81be29b5 op=pingd01:1_start_0 )
Jul 24 11:09:06 node02 lrmd: [2361]: info: rsc:pingd01:1 start[4] (pid 2380)
Jul 24 11:09:06 node02 crmd: [2364]: info: te_rsc_command: Initiating action 17: start pingd02:0_start_0 on node01
Jul 24 11:09:06 node02 crmd: [2364]: info: te_rsc_command: Initiating action 19: start pingd02:1_start_0 on node02 (local)
Jul 24 11:09:06 node02 crmd: [2364]: info: do_lrm_rsc_op: Performing key=19:1:0:b5b7e743-bc71-49ef-bf53-c1fa81be29b5 op=pingd02:1_start_0 )
Jul 24 11:09:06 node02 lrmd: [2361]: info: rsc:pingd02:1 start[5] (pid 2381)
Jul 24 11:09:06 node02 pingd: [2386]: info: Invoked: /usr/lib64/heartbeat/pingd -D -p /var/run//pingd-default_ping_set02 -a default_ping_set02 -d 5s -m 100 -i 1 -h 192.168.133.11 
Jul 24 11:09:06 node02 lrmd: [2361]: info: operation start[5] on pingd02:1 for client 2364: pid 2381 exited with return code 0
Jul 24 11:09:06 node02 pingd: [2387]: info: Invoked: /usr/lib64/heartbeat/pingd -D -p /var/run//pingd-default_ping_set01 -a default_ping_set01 -d 5s -m 100 -i 1 -h 192.168.201.254 
Jul 24 11:09:06 node02 pingd: [2388]: info: main: Starting pingd
Jul 24 11:09:06 node02 pingd: [2388]: info: attrd_lazy_update: Connecting to cluster... 5 retries remaining
Jul 24 11:09:06 node02 attrd: [2363]: info: find_hash_entry: Creating hash entry for default_ping_set02
Jul 24 11:09:06 node02 pingd: [2390]: info: main: Starting pingd
Jul 24 11:09:06 node02 pingd: [2390]: info: attrd_lazy_update: Connecting to cluster... 5 retries remaining
Jul 24 11:09:06 node02 attrd: [2363]: info: find_hash_entry: Creating hash entry for default_ping_set01
Jul 24 11:09:06 node02 lrmd: [2361]: info: operation start[4] on pingd01:1 for client 2364: pid 2380 exited with return code 0
Jul 24 11:09:06 node02 crmd: [2364]: info: process_lrm_event: LRM operation pingd02:1_start_0 (call=5, rc=0, cib-update=56, confirmed=true) ok
Jul 24 11:09:06 node02 crmd: [2364]: info: process_lrm_event: LRM operation pingd01:1_start_0 (call=4, rc=0, cib-update=57, confirmed=true) ok
Jul 24 11:09:06 node02 crmd: [2364]: info: match_graph_event: Action pingd02:1_start_0 (19) confirmed on node02 (rc=0)
Jul 24 11:09:06 node02 crmd: [2364]: info: te_rsc_command: Initiating action 20: monitor pingd02:1_monitor_2000 on node02 (local)
Jul 24 11:09:06 node02 crmd: [2364]: info: do_lrm_rsc_op: Performing key=20:1:0:b5b7e743-bc71-49ef-bf53-c1fa81be29b5 op=pingd02:1_monitor_2000 )
Jul 24 11:09:06 node02 lrmd: [2361]: info: rsc:pingd02:1 monitor[6] (pid 2394)
Jul 24 11:09:06 node02 crmd: [2364]: info: match_graph_event: Action pingd01:1_start_0 (11) confirmed on node02 (rc=0)
Jul 24 11:09:06 node02 crmd: [2364]: info: te_rsc_command: Initiating action 12: monitor pingd01:1_monitor_2000 on node02 (local)
Jul 24 11:09:06 node02 crmd: [2364]: info: do_lrm_rsc_op: Performing key=12:1:0:b5b7e743-bc71-49ef-bf53-c1fa81be29b5 op=pingd01:1_monitor_2000 )
Jul 24 11:09:06 node02 lrmd: [2361]: info: rsc:pingd01:1 monitor[7] (pid 2395)
Jul 24 11:09:06 node02 lrmd: [2361]: info: operation monitor[6] on pingd02:1 for client 2364: pid 2394 exited with return code 0
Jul 24 11:09:06 node02 crmd: [2364]: info: process_lrm_event: LRM operation pingd02:1_monitor_2000 (call=6, rc=0, cib-update=58, confirmed=false) ok
Jul 24 11:09:06 node02 crmd: [2364]: info: match_graph_event: Action pingd02:1_monitor_2000 (20) confirmed on node02 (rc=0)
Jul 24 11:09:06 node02 lrmd: [2361]: info: operation monitor[7] on pingd01:1 for client 2364: pid 2395 exited with return code 0
Jul 24 11:09:06 node02 crmd: [2364]: info: process_lrm_event: LRM operation pingd01:1_monitor_2000 (call=7, rc=0, cib-update=59, confirmed=false) ok
Jul 24 11:09:06 node02 crmd: [2364]: info: match_graph_event: Action pingd01:1_monitor_2000 (12) confirmed on node02 (rc=0)
Jul 24 11:09:08 node02 crmd: [2364]: info: match_graph_event: Action pingd02:0_start_0 (17) confirmed on node01 (rc=0)
Jul 24 11:09:08 node02 crmd: [2364]: info: te_rsc_command: Initiating action 18: monitor pingd02:0_monitor_2000 on node01
Jul 24 11:09:08 node02 crmd: [2364]: info: te_pseudo_action: Pseudo action 22 fired and confirmed
Jul 24 11:09:08 node02 crmd: [2364]: info: match_graph_event: Action pingd01:0_start_0 (9) confirmed on node01 (rc=0)
Jul 24 11:09:08 node02 crmd: [2364]: info: te_rsc_command: Initiating action 10: monitor pingd01:0_monitor_2000 on node01
Jul 24 11:09:08 node02 crmd: [2364]: info: te_pseudo_action: Pseudo action 14 fired and confirmed
Jul 24 11:09:09 node02 crmd: [2364]: info: match_graph_event: Action pingd02:0_monitor_2000 (18) confirmed on node01 (rc=0)
Jul 24 11:09:09 node02 crmd: [2364]: info: match_graph_event: Action pingd01:0_monitor_2000 (10) confirmed on node01 (rc=0)
Jul 24 11:09:09 node02 crmd: [2364]: info: run_graph: ====================================================
Jul 24 11:09:09 node02 crmd: [2364]: notice: run_graph: Transition 1 (Complete=21, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pengine/pe-input-1.bz2): Complete
Jul 24 11:09:09 node02 crmd: [2364]: info: te_graph_trigger: Transition 1 is now complete
Jul 24 11:09:09 node02 crmd: [2364]: info: notify_crmd: Transition 1 status: done - <null>
Jul 24 11:09:09 node02 crmd: [2364]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Jul 24 11:09:09 node02 crmd: [2364]: info: do_state_transition: Starting PEngine Recheck Timer
Jul 24 11:09:11 node02 attrd: [2363]: info: attrd_trigger_update: Sending flush op to all hosts for: default_ping_set02 (100)
Jul 24 11:09:11 node02 attrd: [2363]: info: attrd_trigger_update: Sending flush op to all hosts for: default_ping_set01 (100)
Jul 24 11:09:12 node02 attrd: [2363]: info: attrd_ha_callback: flush message from node02
Jul 24 11:09:12 node02 attrd: [2363]: info: attrd_perform_update: Sent update 15: default_ping_set02=100
Jul 24 11:09:12 node02 attrd: [2363]: info: attrd_ha_callback: flush message from node02
Jul 24 11:09:12 node02 crmd: [2364]: info: abort_transition_graph: te_update_diff:150 - Triggered transition abort (complete=1, tag=nvpair, id=status-554581e7-6f22-45a9-a637-9a6606234a2d-default_ping_set02, name=default_ping_set02, value=100, magic=NA, cib=0.7.18) : Transient attribute: update
Jul 24 11:09:12 node02 attrd: [2363]: info: attrd_perform_update: Sent update 18: default_ping_set01=100
Jul 24 11:09:12 node02 crmd: [2364]: info: abort_transition_graph: te_update_diff:150 - Triggered transition abort (complete=1, tag=nvpair, id=status-554581e7-6f22-45a9-a637-9a6606234a2d-default_ping_set01, name=default_ping_set01, value=100, magic=NA, cib=0.7.19) : Transient attribute: update
Jul 24 11:09:12 node02 crmd: [2364]: info: abort_transition_graph: te_update_diff:150 - Triggered transition abort (complete=1, tag=nvpair, id=status-d6a57ba8-8f59-4ce8-8ae7-7b757c721f7c-default_ping_set02, name=default_ping_set02, value=100, magic=NA, cib=0.7.20) : Transient attribute: update
Jul 24 11:09:12 node02 crmd: [2364]: info: abort_transition_graph: te_update_diff:150 - Triggered transition abort (complete=1, tag=nvpair, id=status-d6a57ba8-8f59-4ce8-8ae7-7b757c721f7c-default_ping_set01, name=default_ping_set01, value=100, magic=NA, cib=0.7.21) : Transient attribute: update
Jul 24 11:09:13 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:14 node02 crmd: [2364]: info: crm_timer_popped: New Transition Timer (I_PE_CALC) just popped!
Jul 24 11:09:14 node02 crmd: [2364]: info: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED origin=crm_timer_popped ]
Jul 24 11:09:14 node02 crmd: [2364]: info: do_state_transition: Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Jul 24 11:09:14 node02 crmd: [2364]: info: do_state_transition: All 2 cluster nodes are eligible to run resources.
Jul 24 11:09:14 node02 crmd: [2364]: info: do_pe_invoke: Query 60: Requesting the current CIB: S_POLICY_ENGINE
Jul 24 11:09:14 node02 pengine: [2369]: notice: unpack_config: On loss of CCM Quorum: Ignore
Jul 24 11:09:14 node02 pengine: [2369]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jul 24 11:09:14 node02 pengine: [2369]: WARN: unpack_nodes: Blind faith: not fencing unseen nodes
Jul 24 11:09:14 node02 pengine: [2369]: info: determine_online_status: Node node02 is online
Jul 24 11:09:14 node02 crmd: [2364]: info: do_pe_invoke_callback: Invoking the PE: query=60, ref=pe_calc-dc-1343095754-35, seq=2, quorate=1
Jul 24 11:09:14 node02 pengine: [2369]: info: determine_online_status: Node node01 is online
Jul 24 11:09:14 node02 pengine: [2369]: notice: clone_print:  Clone Set: clnPingd01
Jul 24 11:09:14 node02 pengine: [2369]: notice: short_print:      Started: [ node01 node02 ]
Jul 24 11:09:14 node02 pengine: [2369]: notice: clone_print:  Clone Set: clnPingd02
Jul 24 11:09:14 node02 pengine: [2369]: notice: short_print:      Started: [ node01 node02 ]
Jul 24 11:09:14 node02 pengine: [2369]: notice: LogActions: Leave   resource pingd01:0	(Started node01)
Jul 24 11:09:14 node02 pengine: [2369]: notice: LogActions: Leave   resource pingd01:1	(Started node02)
Jul 24 11:09:14 node02 pengine: [2369]: notice: LogActions: Leave   resource pingd02:0	(Started node01)
Jul 24 11:09:14 node02 pengine: [2369]: notice: LogActions: Leave   resource pingd02:1	(Started node02)
Jul 24 11:09:14 node02 crmd: [2364]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
Jul 24 11:09:14 node02 crmd: [2364]: info: unpack_graph: Unpacked transition 2: 0 actions in 0 synapses
Jul 24 11:09:14 node02 crmd: [2364]: info: do_te_invoke: Processing graph 2 (ref=pe_calc-dc-1343095754-35) derived from /var/lib/pengine/pe-input-2.bz2
Jul 24 11:09:14 node02 crmd: [2364]: info: run_graph: ====================================================
Jul 24 11:09:14 node02 crmd: [2364]: notice: run_graph: Transition 2 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pengine/pe-input-2.bz2): Complete
Jul 24 11:09:14 node02 crmd: [2364]: info: te_graph_trigger: Transition 2 is now complete
Jul 24 11:09:14 node02 crmd: [2364]: info: notify_crmd: Transition 2 status: done - <null>
Jul 24 11:09:14 node02 crmd: [2364]: info: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
Jul 24 11:09:14 node02 crmd: [2364]: info: do_state_transition: Starting PEngine Recheck Timer
Jul 24 11:09:14 node02 pengine: [2369]: info: process_pe_message: Transition 2: PEngine Input stored in: /var/lib/pengine/pe-input-2.bz2
Jul 24 11:09:16 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:18 node02 attrd: [2363]: info: attrd_ha_callback: flush message from node01
Jul 24 11:09:18 node02 attrd: [2363]: info: attrd_ha_callback: flush message from node01
Jul 24 11:09:23 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:30 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:31 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:38 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:39 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:49 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:50 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:51 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:09:53 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:01 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:13 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:15 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:20 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:21 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:27 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:33 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:38 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:39 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:41 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:49 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:50 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:51 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:52 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:53 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:54 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:55 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:56 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:57 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:58 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:10:59 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:00 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:01 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:02 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:03 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:04 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:05 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:06 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:08 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:09 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:10 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:11 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:12 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:13 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:14 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:15 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:16 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:17 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:18 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:19 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:20 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:21 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:23 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:11:41 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:04 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:14 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:40 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:41 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:43 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:46 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:47 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:49 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:54 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:12:55 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:03 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:19 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:20 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:21 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:22 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:23 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:24 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:25 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:26 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:27 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:28 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:29 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:30 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:31 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:32 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:33 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:34 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:35 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:36 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:37 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:38 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:39 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:40 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:41 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:42 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:43 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:44 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:45 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:46 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:47 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:48 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:50 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:51 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:52 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:53 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:54 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:55 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:56 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:57 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:58 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:13:59 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:00 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:01 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:02 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:03 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:10 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:15 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:21 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:33 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:43 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:44 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:51 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:53 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:54 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:56 node02 pingd: [2390]: info: ping_read: Retrying...
Jul 24 11:14:58 node02 pingd: [2390]: info: ping_read: Retrying...
