[ClusterLabs] node is always offline

Eric Ren zren at suse.com
Tue Aug 16 07:55:19 UTC 2016


Hi,

On 08/16/2016 01:29 PM, wslium at 126.com wrote:
> thank you for your reply.
> The network is well,  firewall has been closed,  and selinux has been disabled.
> /var/log/pacemaker.log as follows:
To be honest, this piece of logs is useless. A whole log file should be attached, which logs 
messages from
the first time when things go bad until now;-)

Also,  lots of info is missing, like OS distributions, packages' version and cluster 
configuration, etc.
A "hb_report" (man hb_report) would be great if package "cluster-glue" is intalled.

Eric
> Aug 16 13:11:12 [8980] node0       crmd:     info: peer_update_callback:        node0 is now (null)
> Aug 16 13:11:12 [8980] node0       crmd:     info: crm_get_peer:        Node 1 has uuid 1
> Aug 16 13:11:12 [8980] node0       crmd:     info: crm_update_peer_proc:        cluster_connect_cpg: Node node0[1] - corosync-cpg is now online
> Aug 16 13:11:12 [8980] node0       crmd:     info: peer_update_callback:        Client node0/peer now has status [online] (DC=<null>)
> Aug 16 13:11:12 [8980] node0       crmd:     info: init_cs_connection_once:     Connection to 'corosync': established
> Aug 16 13:11:12 [8980] node0       crmd:   notice: cluster_connect_quorum:      Quorum acquired
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_ha_control:       Connected to the cluster
> Aug 16 13:11:12 [8980] node0       crmd:     info: lrmd_ipc_connect:    Connecting to lrmd
> Aug 16 13:11:12 [8977] node0       lrmd:     info: crm_client_new:      Connecting 0x16025a0 for uid=189 gid=189 pid=8980 id=e317bf62-fb55-46a1-af17-9284234917b8
> Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.1.0)
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_lrm_control:      LRM connection established
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_started:  Delaying start, no membership data (0000000000100000)
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_started:  Delaying start, no membership data (0000000000100000)
> Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.1.0)
> Aug 16 13:11:12 [8980] node0       crmd:     info: pcmk_quorum_notification:    Membership 8571476093872377196: quorum retained (1)
> Aug 16 13:11:12 [8980] node0       crmd:   notice: crm_update_peer_state:       pcmk_quorum_notification: Node node0[1] - state is now member (was (null))
> Aug 16 13:11:12 [8980] node0       crmd:     info: peer_update_callback:        node0 is now member (was (null))
> Aug 16 13:11:12 [8980] node0       crmd:   notice: crm_update_peer_state:       pcmk_quorum_notification: Node node0[1] - state is now lost (was member)
> Aug 16 13:11:12 [8980] node0       crmd:     info: peer_update_callback:        node0 is now lost (was member)
> Aug 16 13:11:12 [8980] node0       crmd:    error: reap_dead_nodes:     We're not part of the cluster anymore
> Aug 16 13:11:12 [8975] node0        cib:     info: crm_client_new:      Connecting 0x1cbcd80 for uid=0 gid=0 pid=8976 id=9afb683d-8390-408a-a690-fb5447fefd37
> Aug 16 13:11:12 [8976] node0 stonith-ng:   notice: setup_cib:   Watching for stonith topology changes
> Aug 16 13:11:12 [8976] node0 stonith-ng:     info: qb_ipcs_us_publish:  server name: stonith-ng
> Aug 16 13:11:12 [8976] node0 stonith-ng:     info: main:        Starting stonith-ng mainloop
> Aug 16 13:11:12 [8976] node0 stonith-ng:     info: pcmk_cpg_membership:         Joined[0.0] stonith-ng.1
> Aug 16 13:11:12 [8976] node0 stonith-ng:     info: pcmk_cpg_membership:         Member[0.0] stonith-ng.1
> Aug 16 13:11:12 [8980] node0       crmd:    error: do_log:      FSA: Input I_ERROR from reap_dead_nodes() received in state S_STARTING
> Aug 16 13:11:12 [8980] node0       crmd:   notice: do_state_transition:         State transition S_STARTING -> S_RECOVERY [ input=I_ERROR cause=C_FSA_INTERNAL origin=reap_dead_nodes ]
> Aug 16 13:11:12 [8980] node0       crmd:  warning: do_recover:  Fast-tracking shutdown in response to errors
> Aug 16 13:11:12 [8980] node0       crmd:    error: do_started:  Start cancelled... S_RECOVERY
> Aug 16 13:11:12 [8980] node0       crmd:    error: do_log:      FSA: Input I_TERMINATE from do_recover() received in state S_RECOVERY
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_state_transition:         State transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE cause=C_FSA_INTERNAL origin=do_recover ]
> Aug 16 13:11:12 [8980] node0       crmd:   notice: lrm_state_verify_stopped:    Stopped 0 recurring operations at shutdown (0 ops remaining)
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_lrm_control:      Disconnecting from the LRM
> Aug 16 13:11:12 [8980] node0       crmd:     info: lrmd_api_disconnect:         Disconnecting from lrmd service
> Aug 16 13:11:12 [8980] node0       crmd:     info: lrmd_ipc_connection_destroy:         IPC connection destroyed
> Aug 16 13:11:12 [8980] node0       crmd:     info: lrm_connection_destroy:      LRM Connection disconnected
> Aug 16 13:11:12 [8980] node0       crmd:     info: lrmd_api_disconnect:         Disconnecting from lrmd service
> Aug 16 13:11:12 [8980] node0       crmd:   notice: do_lrm_control:      Disconnected from the LRM
> Aug 16 13:11:12 [8980] node0       crmd:     info: crm_cluster_disconnect:      Disconnecting from cluster infrastructure: corosync
> Aug 16 13:11:12 [8980] node0       crmd:   notice: terminate_cs_connection:     Disconnecting from Corosync
> Aug 16 13:11:12 [8980] node0       crmd:     info: crm_cluster_disconnect:      Disconnected from corosync
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_ha_control:       Disconnected from the cluster
> Aug 16 13:11:12 [8977] node0       lrmd:     info: crm_client_destroy:  Destroying 0 events
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_cib_control:      Disconnecting CIB
> Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.1.0)
> Aug 16 13:11:12 [8976] node0 stonith-ng:     info: init_cib_cache_cb:   Updating device list from the cib: init
> Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_cib_connection_destroy:         Connection to the CIB terminated...
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_exit:     Performing A_EXIT_0 - gracefully exiting the CRMd
> Aug 16 13:11:12 [8980] node0       crmd:     info: do_exit:     [crmd] stopped (0)
> Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_exit:   Dropping I_TERMINATE: [ state=S_TERMINATE cause=C_FSA_INTERNAL origin=do_stop ]
> Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_quorum_destroy:         connection closed
> Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_cs_destroy:     connection closed
> Aug 16 13:11:12 [8980] node0       crmd:     info: crmd_init:   8980 stopped: OK (0)
> Aug 16 13:11:12 [8980] node0       crmd:    error: crmd_fast_exit:      Could not recover from internal error
> Aug 16 13:11:12 [8980] node0       crmd:     info: crm_xml_cleanup:     Cleaning up memory from libxml2
> Aug 16 13:11:12 [8975] node0        cib:     info: crm_client_destroy:  Destroying 0 events
> Aug 16 13:11:12 [8974] node0 pacemakerd:    error: pcmk_child_exit:     Child process crmd (8980) exited: Generic Pacemaker error (201)
> Aug 16 13:11:12 [8974] node0 pacemakerd:    error: pcmk_child_exit:     Child process crmd (8980) exited: Generic Pacemaker error (201)
> Aug 16 13:11:12 [8974] node0 pacemakerd:   notice: pcmk_process_exit:   Respawning failed child process: crmd
> Aug 16 13:11:12 [8974] node0 pacemakerd:   notice: pcmk_process_exit:   Respawning failed child process: crmd
> Aug 16 13:11:12 [8974] node0 pacemakerd:     info: start_child:         Using uid=189 and group=189 for process crmd
> Aug 16 13:11:12 [8974] node0 pacemakerd:     info: start_child:         Using uid=189 and group=189 for process crmd
> Aug 16 13:11:12 [8974] node0 pacemakerd:     info: start_child:         Forked child 8982 for process crmd
> Aug 16 13:11:12 [8974] node0 pacemakerd:     info: start_child:         Forked child 8982 for process crmd
> Aug 16 13:11:12 [8982] node0       crmd:     info: crm_log_init:        Changed active directory to /var/lib/pacemaker/cores/hacluster
> Aug 16 13:11:12 [8982] node0       crmd:   notice: main:        CRM Git Version: 368c726
> Aug 16 13:11:12 [8982] node0       crmd:     info: do_log:      FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
> Aug 16 13:11:12 [8982] node0       crmd:     info: get_cluster_type:    Verifying cluster type: 'corosync'
> Aug 16 13:11:12 [8982] node0       crmd:     info: get_cluster_type:    Assuming an active 'corosync' cluster
> Aug 16 13:11:12 [8975] node0        cib:     info: crm_client_new:      Connecting 0x1a8dd80 for uid=189 gid=189 pid=8982 id=5e050be4-f112-4f32-831b-1704ce1872dd
> Aug 16 13:11:12 [8982] node0       crmd:     info: do_cib_control:      CIB connection established
> Aug 16 13:11:12 [8982] node0       crmd:   notice: crm_cluster_connect:         Connecting to cluster infrastructure: corosync
> Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         Completed cib_query operation for section 'all': OK (rc=0, origin=local/crmd/2, version=0.1.0)
> Aug 16 13:11:12 [8982] node0       crmd:     info: crm_get_peer:        Created entry c01a1bd2-9f44-4061-9ea3-95c1df3de86d/0xea3c20 for node (null)/1 (1 total)
> Aug 16 13:11:12 [8982] node0       crmd:     info: crm_get_peer:        Node 1 is now known as node0
> Aug 16 13:11:12 [8982] node0       crmd:     info: peer_update_callback:        node0 is now (null)
> Aug 16 13:11:12 [8982] node0       crmd:     info: crm_get_peer:        Node 1 has uuid 1
> Aug 16 13:11:12 [8982] node0       crmd:     info: crm_update_peer_proc:        cluster_connect_cpg: Node node0[1] - corosync-cpg is now online
> Aug 16 13:11:12 [8982] node0       crmd:     info: peer_update_callback:        Client node0/peer now has status [online] (DC=<null>)
> Aug 16 13:11:12 [8982] node0       crmd:     info: init_cs_connection_once:     Connection to 'corosync': established
> Aug 16 13:11:12 [8982] node0       crmd:   notice: cluster_connect_quorum:      Quorum acquired
> Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         Completed cib_modify operation for section nodes: OK (rc=0, origin=local/crmd/3, version=0.1.0)
> Aug 16 13:11:12 [8982] node0       crmd:     info: do_ha_control:       Connected to the cluster
> Aug 16 13:11:12 [8982] node0       crmd:     info: lrmd_ipc_connect:    Connecting to lrmd
> Aug 16 13:11:12 [8977] node0       lrmd:     info: crm_client_new:      Connecting 0x16025a0 for uid=189 gid=189 pid=8982 id=c6776327-b798-40f2-8ca4-a4b032cff2eb
> Aug 16 13:11:12 [8982] node0       crmd:     info: do_lrm_control:      LRM connection established
> Aug 16 13:11:12 [8982] node0       crmd:     info: do_started:  Delaying start, no membership data (0000000000100000)
> Aug 16 13:11:12 [8982] node0       crmd:     info: do_started:  Delaying start, no membership data (0000000000100000)
> Aug 16 13:11:12 [8975] node0        cib:     info: cib_process_request:         Completed cib_query operation for section crm_config: OK (rc=0, origin=local/crmd/4, version=0.1.0)
> Aug 16 13:11:12 [8982] node0       crmd:     info: pcmk_quorum_notification:    Membership 8571476093872377196: quorum retained (1)
> Aug 16 13:11:12 [8982] node0       crmd:   notice: crm_update_peer_state:       pcmk_quorum_notification: Node node0[1] - state is now member (was (null))
> Aug 16 13:11:12 [8982] node0       crmd:     info: peer_update_callback:        node0 is now member (was (null))
> Aug 16 13:11:12 [8982] node0       crmd:   notice: crm_update_peer_state:       pcmk_quorum_notification: Node node0[1] - state is now lost (was member)
> Aug 16 13:11:12 [8982] node0       crmd:     info: peer_update_callback:        node0 is now lost (was member)
> Aug 16 13:11:12 [8982] node0       crmd:    error: reap_dead_nodes:     We're not part of the cluster anymore
> Aug 16 13:11:12 [8982] node0       crmd:    error: do_log:      FSA: Input I_ERROR from reap_dead_nodes() received in state S_STARTING
> Aug 16 13:11:12 [8982] node0       crmd:   notice: do_state_transition:         State transition S_STARTING -> S_RECOVERY [ input=I_ERROR cause=C_FSA_INTERNAL origin=reap_dead_nodes ]
> Aug 16 13:11:12 [8982] node0       crmd:  warning: do_recover:  Fast-tracking shutdown in response to errors
> Aug 16 13:11:12 [8982] node0       crmd:    error: do_started:  Start cancelled... S_RECOVERY
> Aug 16 13:11:12 [8982] node0       crmd:    error: do_log:      FSA: Input I_TERMINATE from do_recover() received in state S_RECOVERY
> Aug 16 13:11:12 [8982] node0       crmd:     info: do_state_transition:         State transition S_RECOVERY -> S_TERMINATE [ input=I_TERMINATE cause=C_FSA_INTERNAL origin=do_recover ]
> Aug 16 13:11:12 [8982] node0       crmd:   notice: lrm_state_verify_stopped:    Stopped 0 recurring operations at shutdown (0 ops remaining)
> Aug 16 13:11:12 [8982] node0       crmd:     info: do_lrm_control:      Disconnecting from the LRM
> Aug 16 13:11:12 [8982] node0       crmd:     info: lrmd_api_disconnect:         Disconnecting from lrmd service
> Aug 16 13:11:12 [8982] node0       crmd:     info: lrmd_ipc_connection_destroy:         IPC connection destroyed
> Aug 16 13:11:12 [8977] node0       lrmd:     info: crm_client_destroy:  Destroying 0 events
> Aug 16 13:11:12 [8982] node0       crmd:     info: lrm_connection_destroy:      LRM Connection disconnected
> Aug 16 13:11:12 [8982] node0       crmd:     info: lrmd_api_disconnect:         Disconnecting from lrmd service
> Aug 16 13:11:12 [8982] node0       crmd:   notice: do_lrm_control:      Disconnected from the LRM
> Aug 16 13:11:12 [8982] node0       crmd:     info: crm_cluster_disconnect:      Disconnecting from cluster infrastructure: corosync
>
>
>
> wslium at 126.com
>
>
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org


-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.clusterlabs.org/pipermail/users/attachments/20160816/82121d65/attachment-0002.html>


More information about the Users mailing list