[ClusterLabs] Stack: unknown and all nodes offline

Ken Gaillot kgaillot at redhat.com
Thu Dec 10 20:41:12 UTC 2015


On 12/10/2015 12:45 PM, Louis Munro wrote:
> Hello all,
> 
> I am trying to get a Corosync 2 cluster going on CentOS 6.7 but I am running in a bit of a problem with either Corosync or Pacemaker.
> crm reports that all my nodes are offline and the stack is unknown (I am not sure if that is relevant).
> 
> I believe both nodes are actually present and seen in corosync, but they may not be considered as such by pacemaker.
> I have messages in the logs saying that the processes cannot get the node name and default to uname -n: 
> 
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:   notice: get_node_name:	Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: crm_get_peer:	Node 739513528 is now known as hack1.example.com
> 
> The uname -n is correct as far that is concerned.
> 
> 
> Does this mean anything to anyone here? 
> 
> 
> [Lots of details to follow]...
> 
> I compiled my own versions of Corosync, Pacemaker, crm and the resource-agents seemingly without problems.
> 
> Here is what I currently have installed:
> 
> # corosync -v
> Corosync Cluster Engine, version '2.3.5'
> Copyright (c) 2006-2009 Red Hat, Inc.
> 
> # pacemakerd -F
> Pacemaker 1.1.13 (Build: 5b41ae1)
>  Supporting v3.0.10:  generated-manpages agent-manpages ascii-docs ncurses libqb-logging libqb-ipc lha-fencing upstart nagios  corosync-native atomic-attrd libesmtp acls
> 
> # crm --version
> crm 2.2.0-rc3
> 
> 
> 
> Here is the output of crm status:
> 
> # crm status
> Last updated: Thu Dec 10 12:47:50 2015		Last change: Thu Dec 10 12:02:33 2015 by root via cibadmin on hack1.example.com
> Stack: unknown
> Current DC: NONE
> 2 nodes and 0 resources configured
> 
> OFFLINE: [ hack1.example.com hack2.example.com ]
> 
> Full list of resources:
> 
> {nothing to see here}
> 
> 
> 
> # corosync-cmapctl | grep members
> runtime.totem.pg.mrp.srp.members.739513528.config_version (u64) = 0
> runtime.totem.pg.mrp.srp.members.739513528.ip (str) = r(0) ip(172.20.20.184)
> runtime.totem.pg.mrp.srp.members.739513528.join_count (u32) = 1
> runtime.totem.pg.mrp.srp.members.739513528.status (str) = joined
> runtime.totem.pg.mrp.srp.members.739513590.config_version (u64) = 0
> runtime.totem.pg.mrp.srp.members.739513590.ip (str) = r(0) ip(172.20.20.246)
> runtime.totem.pg.mrp.srp.members.739513590.join_count (u32) = 1
> runtime.totem.pg.mrp.srp.members.739513590.status (str) = joined
> 
> 
> # uname -n
> hack1.example.com
> 
> # corosync-cfgtool -s
> Printing ring status.
> Local node ID 739513528
> RING ID 0
> 	id	= 172.20.20.184
> 	status	= ring 0 active with no faults
> 
> 
> # uname -n
> hack2.example.com
> 
> 
> # corosync-cfgtool -s
> Printing ring status.
> Local node ID 739513590
> RING ID 0
> 	id	= 172.20.20.246
> 	status	= ring 0 active with no faults
> 
> 
> 
> 
> Shouldn’t I see both nodes in the same ring?

They are in the same ring, but the cfgtool will only print the local id.

> My corosync config is currently defined as:
> 
> # egrep -v '#' /etc/corosync/corosync.conf
> totem {
> 	version: 2
> 
> 	crypto_cipher: none
> 	crypto_hash: none
> 	clear_node_high_bit: yes
> 	cluster_name: hack_cluster
> 	interface {
> 		ringnumber: 0
> 		bindnetaddr: 172.20.0.0
> 		mcastaddr: 239.255.1.1
> 		mcastport: 5405
> 		ttl: 1
> 	}
> 
> }
> 
> logging {
> 	fileline: on
> 	to_stderr: no
> 	to_logfile: yes
> 	logfile: /var/log/cluster/corosync.log
> 	to_syslog: yes
> 	debug: off
> 	timestamp: on
> 	logger_subsys {
> 		subsys: QUORUM
> 		debug: off
> 	}
> }
> 
> # cat /etc/corosync/service.d/pacemaker
> service {
>     name: pacemaker
>     ver: 1
> }

You don't want this section if you're using corosync 2. That's the old
"plugin" used with corosync 1.

> 
> 
> And here is my pacemaker configuration:
> 
> # crm config show xml
> <?xml version="1.0" ?>
> <cib num_updates="0" update-origin="hack1.example.com" crm_feature_set="3.0.10" validate-with="pacemaker-2.4" update-client="cibadmin" epoch="13" admin_epoch="0" update-user="root" cib-last-written="Thu Dec 10 13:35:06 2015">
>   <configuration>
>     <crm_config>
>       <cluster_property_set id="cib-bootstrap-options">
>         <nvpair name="stonith-enabled" value="false" id="cib-bootstrap-options-stonith-enabled"/>
>         <nvpair name="no-quorum-policy" value="ignore" id="cib-bootstrap-options-no-quorum-policy"/>
>       </cluster_property_set>
>     </crm_config>
>     <nodes>
>       <node uname="hack1.example.com" id="hack1.example.com">
>         <instance_attributes id="hack1.example.com-instance_attributes">
>           <nvpair name="standby" value="off" id="hack1.example.com-instance_attributes-standby"/>
>         </instance_attributes>
>       </node>
>       <node uname="hack2.example.com" id="hack2.example.com">
>         <instance_attributes id="hack2.example.com-instance_attributes">
>           <nvpair name="standby" value="off" id="hack2.example.com-instance_attributes-standby"/>
>         </instance_attributes>
>       </node>
>     </nodes>
>     <resources/>
>     <constraints/>
>   </configuration>
> </cib>
> 
> 
> 
> 
> 
> 
> And finally some logs that might be relevant: 
> 
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [MAIN  ] main.c:1227 Corosync Cluster Engine ('2.3.5'): started and ready to provide service.
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [MAIN  ] main.c:1228 Corosync built-in features: pie relro bindnow
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] totemnet.c:248 Initializing transport (UDP/IP Multicast).
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] totemcrypto.c:579 Initializing transmit/receive security (NSS) crypto: none hash: none
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] totemudp.c:671 The network interface [172.20.20.184] is now up.
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] service.c:174 Service engine loaded: corosync configuration map access [0]
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [QB    ] ipc_setup.c:377 server name: cmap
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] service.c:174 Service engine loaded: corosync configuration service [1]
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [QB    ] ipc_setup.c:377 server name: cfg
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] service.c:174 Service engine loaded: corosync cluster closed process group service v1.01 [2]
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [QB    ] ipc_setup.c:377 server name: cpg
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] service.c:174 Service engine loaded: corosync profile loading service [4]
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [SERV  ] service.c:174 Service engine loaded: corosync cluster quorum service v0.1 [3]
> Dec 10 13:38:50 [2227] hack1.example.com corosync info    [QB    ] ipc_setup.c:377 server name: quorum
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] totemsrp.c:2095 A new membership (172.20.20.184:300) was formed. Members joined: 739513528
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [MAIN  ] main.c:305 Completed service synchronization, ready to provide service.
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [TOTEM ] totemsrp.c:2095 A new membership (172.20.20.184:304) was formed. Members joined: 739513590
> Dec 10 13:38:50 [2227] hack1.example.com corosync notice  [MAIN  ] main.c:305 Completed service synchronization, ready to provide service.
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: mcp_read_config:	Configured corosync to accept connections from group 500: OK (1)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: main:	Starting Pacemaker 1.1.13 (Build: 5b41ae1):  generated-manpages agent-manpages ascii-docs ncurses libqb-logging libqb-ipc lha-fencing upstart nagios  corosync-native atomic-attrd libesmtp acls
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: main:	Maximum core file size is: 18446744073709551615
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: qb_ipcs_us_publish:	server name: pacemakerd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:	Created entry 212e3751-e79f-4a72-927b-6e0176a9b35c/0x1b68e50 for node (null)/739513528 (1 total)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:	Node 739513528 has uuid 739513528
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_update_peer_proc:	cluster_connect_cpg: Node (null)[739513528] - corosync-cpg is now online
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:    error: cluster_connect_quorum:	Corosync quorum is not configured
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: get_node_name:	Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:	Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Using uid=500 and group=500 for process cib
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Forked child 2231 for process cib
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Forked child 2232 for process stonith-ng
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Forked child 2233 for process lrmd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Using uid=500 and group=500 for process attrd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Forked child 2234 for process attrd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Using uid=500 and group=500 for process pengine
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Forked child 2235 for process pengine
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Using uid=500 and group=500 for process crmd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: start_child:	Forked child 2236 for process crmd
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: main:	Starting mainloop
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: pcmk_cpg_membership:	Node 739513528 joined group pacemakerd (counter=0.0)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: pcmk_cpg_membership:	Node 739513528 still member of group pacemakerd (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: crm_log_init:	Changed active directory to /var/lib/heartbeat/cores/hacluster
> Dec 10 13:38:52 [2231] hack1.example.com        cib:  warning: crm_is_writable:	/var/lib/pacemaker/cib should be owned and r/w by group haclient
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: get_cluster_type:	Verifying cluster type: 'corosync'
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: get_cluster_type:	Assuming an active 'corosync' cluster
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: retrieveCib:	Reading cluster configuration file /var/lib/pacemaker/cib/cib.xml (digest: /var/lib/pacemaker/cib/cib.xml.sig)
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: validate_with_relaxng:	Creating RNG parser context
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:     info: crm_log_init:	Changed active directory to /var/lib/heartbeat/cores/hacluster
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:     info: main:	Starting up
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:     info: get_cluster_type:	Verifying cluster type: 'corosync'
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:     info: get_cluster_type:	Assuming an active 'corosync' cluster
> Dec 10 13:38:52 [2234] hack1.example.com      attrd:   notice: crm_cluster_connect:	Connecting to cluster infrastructure: corosync
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:     info: crm_log_init:	Changed active directory to /var/lib/heartbeat/cores/hacluster
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:   notice: main:	CRM Git Version: 1.1.13 (5b41ae1)
> Dec 10 13:38:52 [2232] hack1.example.com stonith-ng:     info: crm_log_init:	Changed active directory to /var/lib/heartbeat/cores/root
> Dec 10 13:38:52 [2232] hack1.example.com stonith-ng:     info: get_cluster_type:	Verifying cluster type: 'corosync'
> Dec 10 13:38:52 [2232] hack1.example.com stonith-ng:     info: get_cluster_type:	Assuming an active 'corosync' cluster
> Dec 10 13:38:52 [2232] hack1.example.com stonith-ng:   notice: crm_cluster_connect:	Connecting to cluster infrastructure: corosync
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:  warning: crm_is_writable:	/var/lib/pacemaker/pengine should be owned and r/w by group haclient
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:  warning: crm_is_writable:	/var/lib/pacemaker/cib should be owned and r/w by group haclient

You definitely want to take care of these permissions issues.

> Dec 10 13:38:52 [2236] hack1.example.com       crmd:     info: do_log:	FSA: Input I_STARTUP from crmd_init() received in state S_STARTING
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:     info: get_cluster_type:	Verifying cluster type: 'corosync'
> Dec 10 13:38:52 [2236] hack1.example.com       crmd:     info: get_cluster_type:	Assuming an active 'corosync' cluster
> Dec 10 13:38:52 [2231] hack1.example.com        cib:     info: startCib:	CIB Initialization completed successfully
> Dec 10 13:38:52 [2231] hack1.example.com        cib:   notice: crm_cluster_connect:	Connecting to cluster infrastructure: corosync
> Dec 10 13:38:52 [2235] hack1.example.com    pengine:     info: crm_log_init:	Changed active directory to /var/lib/heartbeat/cores/hacluster
> Dec 10 13:38:52 [2235] hack1.example.com    pengine:  warning: crm_is_writable:	/var/lib/pacemaker/pengine should be owned and r/w by group haclient
> Dec 10 13:38:52 [2235] hack1.example.com    pengine:     info: qb_ipcs_us_publish:	server name: pengine
> Dec 10 13:38:52 [2235] hack1.example.com    pengine:     info: main:	Starting pengine
> Dec 10 13:38:52 [2233] hack1.example.com       lrmd:     info: crm_log_init:	Changed active directory to /var/lib/heartbeat/cores/root
> Dec 10 13:38:52 [2233] hack1.example.com       lrmd:     info: qb_ipcs_us_publish:	server name: lrmd
> Dec 10 13:38:52 [2233] hack1.example.com       lrmd:     info: main:	Starting
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: corosync_node_name:	Unable to get node name for nodeid 739513590
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:	Created entry f34a4388-23fc-4f86-8932-43c814eb5ad7/0x1b6a920 for node (null)/739513590 (2 total)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:	Node 739513590 has uuid 739513590
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: pcmk_cpg_membership:	Node 739513590 still member of group pacemakerd (peer=(null), counter=0.1)
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_update_peer_proc:	pcmk_cpg_membership: Node (null)[739513590] - corosync-cpg is now online
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: crm_get_peer:	Node 739513590 is now known as hack2.example.com
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: mcp_cpg_deliver:	Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: mcp_cpg_deliver:	Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: mcp_cpg_deliver:	Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: mcp_cpg_deliver:	Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: mcp_cpg_deliver:	Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: mcp_cpg_deliver:	Ignoring process list sent by peer for local node
> Dec 10 13:38:52 [2230] hack1.example.com pacemakerd:     info: mcp_cpg_deliver:	Ignoring process list sent by peer for local node
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: crm_get_peer:	Created entry 5fbd08d1-fdbf-4f8c-85d0-bbbed1240969/0x180eb70 for node (null)/739513528 (1 total)
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: crm_get_peer:	Node 739513528 has uuid 739513528
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: crm_update_peer_proc:	cluster_connect_cpg: Node (null)[739513528] - corosync-cpg is now online
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:   notice: crm_update_peer_state_iter:	crm_update_peer_proc: Node (null)[739513528] - state is now member (was (null))
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: init_cs_connection_once:	Connection to 'corosync': established
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:	Created entry c2d2ccef-c87c-4b3c-8004-659f09113048/0x11fa8e0 for node (null)/739513528 (1 total)
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:	Node 739513528 has uuid 739513528
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: crm_update_peer_proc:	cluster_connect_cpg: Node (null)[739513528] - corosync-cpg is now online
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:   notice: crm_update_peer_state_iter:	crm_update_peer_proc: Node (null)[739513528] - state is now member (was (null))
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: init_cs_connection_once:	Connection to 'corosync': established
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:   notice: get_node_name:	Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: crm_get_peer:	Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:   notice: get_node_name:	Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:	Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:	Created entry 97e43f18-db5c-4496-b8fc-bd7b4a6ea711/0x1deba80 for node (null)/739513528 (1 total)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:	Node 739513528 has uuid 739513528
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_update_peer_proc:	cluster_connect_cpg: Node (null)[739513528] - corosync-cpg is now online
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: crm_update_peer_state_iter:	crm_update_peer_proc: Node (null)[739513528] - state is now member (was (null))
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: init_cs_connection_once:	Connection to 'corosync': established
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: main:	Cluster connection active
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: qb_ipcs_us_publish:	server name: attrd
> Dec 10 13:38:53 [2234] hack1.example.com      attrd:     info: main:	Accepting attribute updates
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: get_node_name:	Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:	Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: qb_ipcs_us_publish:	server name: cib_ro
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: qb_ipcs_us_publish:	server name: cib_rw
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: qb_ipcs_us_publish:	server name: cib_shm
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: cib_init:	Starting cib mainloop
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: cib_file_backup:	Archived previous version as /var/lib/pacemaker/cib/cib-26.raw
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: cib_file_write_with_digest:	Wrote version 0.13.0 of the CIB to disk (digest: d37e6c1873e854e52dd9b2caaa5e4beb)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: cib_file_write_with_digest:	Reading cluster configuration file /var/lib/pacemaker/cib/cib.dqtKlA (digest: /var/lib/pacemaker/cib/cib.BkV3eX)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: pcmk_cpg_membership:	Node 739513528 joined group cib (counter=0.0)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: pcmk_cpg_membership:	Node 739513528 still member of group cib (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: corosync_node_name:	Unable to get node name for nodeid 739513590
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:	Created entry 87459e91-7cff-4781-9c9a-5f78385e4b7c/0x1dec1d0 for node (null)/739513590 (2 total)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_get_peer:	Node 739513590 has uuid 739513590
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: pcmk_cpg_membership:	Node 739513590 still member of group cib (peer=(null), counter=0.1)
> Dec 10 13:38:53 [2231] hack1.example.com        cib:     info: crm_update_peer_proc:	pcmk_cpg_membership: Node (null)[739513590] - corosync-cpg is now online
> Dec 10 13:38:53 [2231] hack1.example.com        cib:   notice: crm_update_peer_state_iter:	crm_update_peer_proc: Node (null)[739513590] - state is now member (was (null))
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: do_cib_control:	CIB connection established
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:   notice: crm_cluster_connect:	Connecting to cluster infrastructure: corosync
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: crm_get_peer:	Created entry 09d2eabe-fe50-4b40-b614-4f550dcbc677/0x156d0c0 for node (null)/739513528 (1 total)
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: crm_get_peer:	Node 739513528 has uuid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: crm_update_peer_proc:	cluster_connect_cpg: Node (null)[739513528] - corosync-cpg is now online
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: init_cs_connection_once:	Connection to 'corosync': established
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:   notice: get_node_name:	Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: crm_get_peer:	Node 739513528 is now known as hack1.example.com
> Dec 10 13:38:53 [2236] hack1.example.com       crmd:     info: peer_update_callback:	hack1.example.com is now in unknown state
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:    error: cluster_connect_quorum:	Corosync quorum is not configured
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: attrd_cib_connect:	Connected to the CIB after 2 attempts
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: main:	CIB connection active
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: pcmk_cpg_membership:	Node 739513528 joined group attrd (counter=0.0)
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: pcmk_cpg_membership:	Node 739513528 still member of group attrd (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: setup_cib:	Watching for stonith topology changes
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: qb_ipcs_us_publish:	server name: stonith-ng
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: main:	Starting stonith-ng mainloop
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: pcmk_cpg_membership:	Node 739513528 joined group stonith-ng (counter=0.0)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: pcmk_cpg_membership:	Node 739513528 still member of group stonith-ng (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: do_ha_control:	Connected to the cluster
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: lrmd_ipc_connect:	Connecting to lrmd
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: corosync_node_name:	Unable to get node name for nodeid 739513590
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: crm_get_peer:	Created entry 1edc2c5f-1438-4f5c-b5aa-e95861c4205e/0x1814190 for node (null)/739513590 (2 total)
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: crm_get_peer:	Node 739513590 has uuid 739513590
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: pcmk_cpg_membership:	Node 739513590 still member of group attrd (peer=(null), counter=0.1)
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:     info: crm_update_peer_proc:	pcmk_cpg_membership: Node (null)[739513590] - corosync-cpg is now online
> Dec 10 13:38:54 [2234] hack1.example.com      attrd:   notice: crm_update_peer_state_iter:	crm_update_peer_proc: Node (null)[739513590] - state is now member (was (null))
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: corosync_node_name:	Unable to get node name for nodeid 739513590
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:	Created entry 81f08c2d-6517-4b38-8fdc-51a4ada76af2/0x11fc5c0 for node (null)/739513590 (2 total)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:	Node 739513590 has uuid 739513590
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: pcmk_cpg_membership:	Node 739513590 still member of group stonith-ng (peer=(null), counter=0.1)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: crm_update_peer_proc:	pcmk_cpg_membership: Node (null)[739513590] - corosync-cpg is now online
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: crm_update_peer_state_iter:	crm_update_peer_proc: Node (null)[739513590] - state is now member (was (null))
> Dec 10 13:38:54 [2231] hack1.example.com        cib:     info: cib_process_request:	Forwarding cib_modify operation for section nodes to master (origin=local/crmd/3)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: do_lrm_control:	LRM connection established
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: do_started:	Delaying start, no membership data (0000000000100000)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: pcmk_cpg_membership:	Node 739513528 joined group crmd (counter=0.0)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: pcmk_cpg_membership:	Node 739513528 still member of group crmd (peer=hack1.example.com, counter=0.0)
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: get_node_name:	Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: init_cib_cache_cb:	Updating device list from the cib: init
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: cib_devices_update:	Updating devices to version 0.13.0
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:   notice: unpack_config:	On loss of CCM Quorum: Ignore
> Dec 10 13:38:54 [2232] hack1.example.com stonith-ng:     info: crm_get_peer:	Node 739513590 is now known as hack2.example.com
> Dec 10 13:38:54 [2231] hack1.example.com        cib:     info: corosync_node_name:	Unable to get node name for nodeid 739513528
> Dec 10 13:38:54 [2231] hack1.example.com        cib:   notice: get_node_name:	Defaulting to uname -n for the local corosync node name
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: corosync_node_name:	Unable to get node name for nodeid 739513590
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:   notice: get_node_name:	Could not obtain a node name for corosync nodeid 739513590
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: crm_get_peer:	Created entry b34ca282-9c15-47de-8c98-cb59d1d8bed4/0x16b5130 for node (null)/739513590 (2 total)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: crm_get_peer:	Node 739513590 has uuid 739513590
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: pcmk_cpg_membership:	Node 739513590 still member of group crmd (peer=(null), counter=0.1)
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: crm_update_peer_proc:	pcmk_cpg_membership: Node (null)[739513590] - corosync-cpg is now online
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:   notice: crmd_enable_notifications:	Notifications disabled
> Dec 10 13:38:54 [2236] hack1.example.com       crmd:     info: do_started:	Delaying start, no membership data (0000000000100000)
> Dec 10 13:38:54 [2231] hack1.example.com        cib:     info: cib_process_request:	Completed cib_modify operation for section nodes: OK (rc=0, origin=hack1.example.com/crmd/3, version=0.13.0)
> 
> 
> Any help would be much appreciated.
> 
> Best  regards,
> --
> Louis Munro
> 
> 
> 
> 
> 
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
> 





More information about the Users mailing list