[ClusterLabs] crmd error: Cannot route message to unknown node
Ferenc Wágner
wferi at niif.hu
Thu Apr 7 20:22:06 UTC 2016
Hi,
On a freshly rebooted cluster node (after crm_mon reports it as
'online'), I get the following:
wferi at vhbl08:~$ sudo crm_resource -r vm-cedar --cleanup
Cleaning up vm-cedar on vhbl03, removing fail-count-vm-cedar
Cleaning up vm-cedar on vhbl04, removing fail-count-vm-cedar
Cleaning up vm-cedar on vhbl05, removing fail-count-vm-cedar
Cleaning up vm-cedar on vhbl06, removing fail-count-vm-cedar
Cleaning up vm-cedar on vhbl07, removing fail-count-vm-cedar
Cleaning up vm-cedar on vhbl08, removing fail-count-vm-cedar
Waiting for 6 replies from the CRMd..No messages received in 60 seconds.. aborting
Meanwhile, this is written into syslog (I can also provide info level
logs if necessary):
22:03:02 vhbl08 crmd[8990]: error: Cannot route message to unknown node vhbl03
22:03:02 vhbl08 crmd[8990]: error: Cannot route message to unknown node vhbl04
22:03:02 vhbl08 crmd[8990]: error: Cannot route message to unknown node vhbl06
22:03:02 vhbl08 crmd[8990]: error: Cannot route message to unknown node vhbl07
22:03:04 vhbl08 crmd[8990]: notice: Operation vm-cedar_monitor_0: not running (node=vhbl08, call=626, rc=7, cib-update=169, confirmed=true)
For background:
wferi at vhbl08:~$ sudo cibadmin --scope=nodes -Q
<nodes>
<node id="167773707" uname="vhbl05">
<utilization id="nodes-167773707-utilization">
<nvpair id="nodes-167773707-utilization-memoryMiB" name="memoryMiB" value="124928"/>
</utilization>
<instance_attributes id="nodes-167773707"/>
</node>
<node id="167773708" uname="vhbl06">
<utilization id="nodes-167773708-utilization">
<nvpair id="nodes-167773708-utilization-memoryMiB" name="memoryMiB" value="124928"/>
</utilization>
<instance_attributes id="nodes-167773708"/>
</node>
<node id="167773706" uname="vhbl04">
<utilization id="nodes-167773706-utilization">
<nvpair id="nodes-167773706-utilization-memoryMiB" name="memoryMiB" value="124928"/>
</utilization>
<instance_attributes id="nodes-167773706"/>
</node>
<node id="167773705" uname="vhbl03">
<utilization id="nodes-167773705-utilization">
<nvpair id="nodes-167773705-utilization-memoryMiB" name="memoryMiB" value="124928"/>
</utilization>
<instance_attributes id="nodes-167773705"/>
</node>
<node id="167773709" uname="vhbl07">
<utilization id="nodes-167773709-utilization">
<nvpair id="nodes-167773709-utilization-memoryMiB" name="memoryMiB" value="124928"/>
</utilization>
<instance_attributes id="nodes-167773709"/>
</node>
<node id="167773710" uname="vhbl08">
<utilization id="nodes-167773710-utilization">
<nvpair id="nodes-167773710-utilization-memoryMiB" name="memoryMiB" value="124928"/>
</utilization>
</node>
</nodes>
Why does this happen? I've got no node names in corosync.conf, but
Pacemaker defaults to uname -n all right.
--
Thanks,
Feri
More information about the Users
mailing list