[ClusterLabs] getting "Totem is unable to form a cluster" error
Muhammad Sharfuddin
M.Sharfuddin at nds.com.pk
Thu Apr 7 20:24:13 CEST 2016
pacemaker 1.1.12-11.12
openais 1.1.4-5.24.5
corosync 1.4.7-0.23.5
Its a two node active/passive cluster and we just upgraded the SLES 11
SP 3 to SLES 11 SP 4(nothing else) but when we try to start the cluster
service we get the following error:
"Totem is unable to form a cluster because of an operating system or
network fault."
Firewall is stopped and disabled on both the nodes. Both nodes can
ping/ssh/vnc each other.
corosync.conf:
aisexec {
group: root
user: root
}
service {
use_mgmtd: yes
use_logd: yes
ver: 0
name: pacemaker
}
totem {
rrp_mode: none
join: 60
max_messages: 20
vsftype: none
token: 5000
consensus: 6000
interface {
bindnetaddr: 192.168.150.0
member {
memberaddr: 192.168.150.12
}
member {
memberaddr: 192.168.150.13
}
mcastport: 5405
ringnumber: 0
}
secauth: off
version: 2
transport: udpu
token_retransmits_before_loss_const: 10
clear_node_high_bit: new
}
logging {
to_logfile: no
to_syslog: yes
debug: off
timestamp: off
to_stderr: no
fileline: off
syslog_facility: daemon
}
amf {
mode: disable
}
/var/log/messages:
Apr 6 17:51:49 prd1 corosync[8672]: [MAIN ] Corosync Cluster Engine
('1.4.7'): started and ready to provide service.
Apr 6 17:51:49 prd1 corosync[8672]: [MAIN ] Corosync built-in
features: nss
Apr 6 17:51:49 prd1 corosync[8672]: [MAIN ] Successfully configured
openais services to load
Apr 6 17:51:49 prd1 corosync[8672]: [MAIN ] Successfully read main
configuration file '/etc/corosync/corosync.conf'.
Apr 6 17:51:49 prd1 corosync[8672]: [TOTEM ] Initializing transport
(UDP/IP Unicast).
Apr 6 17:51:49 prd1 corosync[8672]: [TOTEM ] Initializing
transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0).
Apr 6 17:51:49 prd1 corosync[8672]: [TOTEM ] The network interface is
down.
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
openais cluster membership service B.01.01
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
openais event service B.01.01
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
openais checkpoint service B.01.01
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
openais availability management framework B.01.01
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
openais message service B.03.01
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
openais distributed locking service B.03.01
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
openais timer service A.01.01
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: process_ais_conf:
Reading configure
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_init:
Local handle: 7685269064754659330 for logging
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
Processing additional logging options...
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: get_config_opt:
Found 'off' for option: debug
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: get_config_opt:
Found 'no' for option: to_logfile
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: get_config_opt:
Found 'yes' for option: to_syslog
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: get_config_opt:
Found 'daemon' for option: syslog_facility
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_init:
Local handle: 8535092201842016259 for quorum
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
No additional configuration supplied for: quorum
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: get_config_opt: No
default for option: provider
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_init:
Local handle: 8054506479773810692 for service
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
Processing additional service options...
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
Processing additional service options...
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
Processing additional service options...
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
Processing additional service options...
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
Processing additional service options...
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
Processing additional service options...
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
Processing additional service options...
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: config_find_next:
Processing additional service options...
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: get_config_opt:
Found '0' for option: ver
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: get_config_opt:
Defaulting to 'pcmk' for option: clustername
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: get_config_opt:
Found 'yes' for option: use_logd
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: get_config_opt:
Found 'yes' for option: use_mgmtd
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: pcmk_startup: CRM:
Initialized
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] Logging: Initialized
pcmk_startup
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: pcmk_startup:
Maximum core file size is: 18446744073709551615
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: pcmk_startup:
Service: 9
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: pcmk_startup: Local
hostname: prd1
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: pcmk_update_nodeid:
Local node id: 2130706433
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: update_member:
Creating entry for node 2130706433 born on 0
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: update_member:
0x64c9c0 Node 2130706433 now known as prd1 (was: (null))
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: update_member: Node
prd1 now has 1 quorum votes (was 0)
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: update_member: Node
2130706433/prd1 is now: member
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Using
uid=90 and group=90 for process cib
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Forked
child 8677 for process cib
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Forked
child 8678 for process stonith-ng
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Forked
child 8679 for process lrmd
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Using
uid=90 and group=90 for process attrd
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Forked
child 8680 for process attrd
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Using
uid=90 and group=90 for process pengine
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Forked
child 8681 for process pengine
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Using
uid=90 and group=90 for process crmd
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Forked
child 8682 for process crmd
Apr 6 17:51:49 prd1 corosync[8672]: [pcmk ] info: spawn_child: Forked
child 8683 for process mgmtd
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
Pacemaker Cluster Manager 1.1.12
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
corosync extended virtual synchrony service
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
corosync configuration service
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
corosync cluster closed process group service v1.01
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
corosync cluster config database access v1.01
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
corosync profile loading service
Apr 6 17:51:49 prd1 corosync[8672]: [SERV ] Service engine loaded:
corosync cluster quorum service v0.1
Apr 6 17:51:49 prd1 corosync[8672]: [MAIN ] Compatibility mode set to
whitetank. Using V1 and V2 of the synchronization engine.
Apr 6 17:51:49 prd1 corosync[8672]: [TOTEM ] adding new UDPU member
{192.168.150.12}
Apr 6 17:51:49 prd1 corosync[8672]: [TOTEM ] adding new UDPU member
{192.168.150.13}
Apr 6 17:51:50 prd1 lrmd[8679]: notice: crm_add_logfile: Additional
logging available in /var/log/pacemaker.log
Apr 6 17:51:50 prd1 mgmtd: [8683]: info: Pacemaker-mgmt Git Version:
969d213
Apr 6 17:51:50 prd1 mgmtd: [8683]: WARN: Core dumps could be lost if
multiple dumps occur.
Apr 6 17:51:50 prd1 mgmtd: [8683]: WARN: Consider setting non-default
value in /proc/sys/kernel/core_pattern (or equivalent) for maximum
supportability
Apr 6 17:51:50 prd1 mgmtd: [8683]: WARN: Consider setting
/proc/sys/kernel/core_uses_pid (or equivalent) to 1 for maximum
supportability
Apr 6 17:51:50 prd1 attrd[8680]: notice: crm_add_logfile: Additional
logging available in /var/log/pacemaker.log
Apr 6 17:51:50 prd1 pengine[8681]: notice: crm_add_logfile:
Additional logging available in /var/log/pacemaker.log
Apr 6 17:51:50 prd1 attrd[8680]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: classic openais (with plugin)
Apr 6 17:51:50 prd1 cib[8677]: notice: crm_add_logfile: Additional
logging available in /var/log/pacemaker.log
Apr 6 17:51:50 prd1 crmd[8682]: notice: crm_add_logfile: Additional
logging available in /var/log/pacemaker.log
Apr 6 17:51:50 prd1 attrd[8680]: notice: get_node_name: Defaulting to
uname -n for the local classic openais (with plugin) node name
Apr 6 17:51:50 prd1 corosync[8672]: [pcmk ] info: pcmk_ipc: Recorded
connection 0x7f944c04acf0 for attrd/8680
Apr 6 17:51:50 prd1 crmd[8682]: notice: main: CRM Git Version: f47ea56
Apr 6 17:51:50 prd1 attrd[8680]: notice: get_node_name: Defaulting to
uname -n for the local classic openais (with plugin) node name
Apr 6 17:51:50 prd1 attrd[8680]: notice: main: Starting mainloop...
Apr 6 17:51:50 prd1 stonith-ng[8678]: notice: crm_add_logfile:
Additional logging available in /var/log/pacemaker.log
Apr 6 17:51:50 prd1 stonith-ng[8678]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: classic openais (with plugin)
Apr 6 17:51:50 prd1 stonith-ng[8678]: notice: get_node_name:
Defaulting to uname -n for the local classic openais (with plugin) node name
Apr 6 17:51:50 prd1 corosync[8672]: [pcmk ] info: pcmk_ipc: Recorded
connection 0x658190 for stonith-ng/8678
Apr 6 17:51:50 prd1 corosync[8672]: [pcmk ] info: update_member: Node
prd1 now has process list: 00000000000000000000000000151312 (1381138)
Apr 6 17:51:50 prd1 corosync[8672]: [pcmk ] info: pcmk_ipc: Sending
membership update 0 to stonith-ng
Apr 6 17:51:50 prd1 stonith-ng[8678]: notice: get_node_name:
Defaulting to uname -n for the local classic openais (with plugin) node name
Apr 6 17:51:50 prd1 cib[8677]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: classic openais (with plugin)
Apr 6 17:51:50 prd1 cib[8677]: notice: get_node_name: Defaulting to
uname -n for the local classic openais (with plugin) node name
Apr 6 17:51:50 prd1 corosync[8672]: [pcmk ] info: pcmk_ipc: Recorded
connection 0x65d450 for cib/8677
Apr 6 17:51:50 prd1 corosync[8672]: [pcmk ] info: pcmk_ipc: Sending
membership update 0 to cib
Apr 6 17:51:50 prd1 cib[8677]: notice: get_node_name: Defaulting to
uname -n for the local classic openais (with plugin) node name
Apr 6 17:51:50 prd1 cib[8677]: notice: crm_update_peer_state:
cib_peer_update_callback: Node prd1[2130706433] - state is now lost (was
(null))
Apr 6 17:51:50 prd1 cib[8677]: notice: crm_update_peer_state:
plugin_handle_membership: Node prd1[2130706433] - state is now member
(was lost)
Apr 6 17:51:50 prd1 mgmtd: [8683]: info: Started.
Apr 6 17:51:51 prd1 crmd[8682]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: classic openais (with plugin)
Apr 6 17:51:51 prd1 crmd[8682]: notice: get_node_name: Defaulting to
uname -n for the local classic openais (with plugin) node name
Apr 6 17:51:51 prd1 corosync[8672]: [pcmk ] info: pcmk_ipc: Recorded
connection 0x661b00 for crmd/8682
Apr 6 17:51:51 prd1 corosync[8672]: [pcmk ] info: pcmk_ipc: Sending
membership update 0 to crmd
Apr 6 17:51:51 prd1 crmd[8682]: notice: get_node_name: Defaulting to
uname -n for the local classic openais (with plugin) node name
Apr 6 17:51:51 prd1 stonith-ng[8678]: notice: setup_cib: Watching for
stonith topology changes
Apr 6 17:51:51 prd1 stonith-ng[8678]: notice: crm_update_peer_state:
st_peer_update_callback: Node prd1[2130706433] - state is now lost (was
(null))
Apr 6 17:51:51 prd1 stonith-ng[8678]: notice: crm_update_peer_state:
plugin_handle_membership: Node prd1[2130706433] - state is now member
(was lost)
Apr 6 17:51:51 prd1 crmd[8682]: notice: crm_update_peer_state:
plugin_handle_membership: Node prd1[2130706433] - state is now member
(was (null))
Apr 6 17:51:51 prd1 crmd[8682]: notice: do_started: The local CRM is
operational
Apr 6 17:51:51 prd1 crmd[8682]: notice: do_state_transition: State
transition S_STARTING -> S_PENDING [ input=I_PENDING
cause=C_FSA_INTERNAL origin=do_started ]
Apr 6 17:51:51 prd1 stonith-ng[8678]: notice: unpack_config: On loss
of CCM Quorum: Ignore
Apr 6 17:52:12 prd1 crmd[8682]: warning: do_log: FSA: Input
I_DC_TIMEOUT from crm_timer_popped() received in state S_PENDING
Apr 6 17:52:35 prd1 corosync[8672]: [MAIN ] Totem is unable to form a
cluster because of an operating system or network fault. The most common
cause of this message is that the local firewall is configured improperly.
Apr 6 17:52:36 prd1 corosync[8672]: [MAIN ] Totem is unable to form a
cluster because of an operating system or network fault. The most common
cause of this message is that the local firewall is configured improperly.
--
Regards,
Muhammad Sharfuddin
<http://www.nds.com.pk>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://clusterlabs.org/pipermail/users/attachments/20160407/4fab02f8/attachment-0001.html>
More information about the Users
mailing list