[Pacemaker] Nodes unable to connect / find each other

Regendoerp, Achim Achim.Regendoerp at galacoral.com
Wed Mar 14 14:24:22 EDT 2012


Greetings everyone,

I am currently experiencing some troubles getting two cluster nodes in a corosync + pacemaker set up to talk with / find each other and set up properly.

I have read various howtos out in the net as well as based on the set up I've done last year too, but to no avail.

Basically I have two VMs in a VLAN, which has all necessary ports open inside the VLAN, and both are supposed to be clustered for an NFS system (which will involve DRBD too).

So far the problem appears to be on the corosync / pacemaker level as in the nodes can see each other (based on what the logs state), yet running 'crm configure' and setting various properties, the error always is that the remote node cannot be found. I am not able to get past this problem, and am running out of ideas.

Short and FQDN hostname are all set up too.
'crm configure' is run as root which is also part of the 'haclient' group.

/etc/hosts.allow and hosts.deny do not have any entries preventing connections

I'd be grateful for any ideas / advise / help etc. :)
Please let me know if any further logs or missing configuration print outs are needed. Thanks.

The package versions are:

rpm -qa | egrep "corosync|pacemaker|cluster|resource"

corosynclib-1.4.1-4.el6.x86_64
clusterlib-3.0.12.1-23.el6.x86_64
pacemaker-cluster-libs-1.1.6-3.el6.x86_64
resource-agents-3.9.2-7.el6.x86_64
cluster-glue-libs-1.0.5-2.el6.x86_64
pacemaker-libs-1.1.6-3.el6.x86_64
corosync-1.4.1-4.el6.x86_64
pacemaker-cli-1.1.6-3.el6.x86_64
cluster-glue-1.0.5-2.el6.x86_64
pacemaker-1.1.6-3.el6.x86_64

Below are the various configuration files and log messages:

### /etc/corosync/corosync.conf ###

compatibility: whitetank

totem {
        version: 2
        secauth: off
        threads: 0
        join:   1000
        consensus: 7500
        max_messages: 20
        interface {
                ringnumber: 0
bindnetaddr: 10.26.29.0 # colleague set these to .238 and .239 respectively on the other node, which are their eth0 NICs to test if it makes any difference
mcastaddr: 226.94.1.1
mcastport: 5405
                #ttl: 1
        }
}

logging {
        fileline: off
        to_stderr: off
        to_logfile: yes
        to_syslog: yes
        logfile: /var/log/cluster/corosync.log
        debug: on
        timestamp: on
        logger_subsys {
                subsys: AMF
                debug: off
        }
}

amf {
        mode: disabled
}

### /etc/corosync/service.d/pcmk  ###

service
{
        name: pacemaker
        ver: 1
        use_mgmtd: no
        use_logd: yes
}

### /etc/sysconfig/pacemaker  (all default) ###

# Variables for running child daemons under valgrind and/or checking for memory problems
#export G_SLICE=always-malloc
#export MALLOC_PERTURB_=221 # or 0
#export MALLOC_CHECK_=3     # or 0,1,2
#export HA_valgrind_enabled=yes
#export HA_valgrind_enabled=cib,crmd
#export HA_callgrind_enabled=yes
#export HA_callgrind_enabled=cib,crmd
#export VALGRIND_OPTS="--leak-check=full --trace-children=no --num-callers=25 --log-file=/tmp/pacemaker-%p.valgrind"

# Variables that control logging
#export PCMK_trace_functions=
#export PCMK_trace_formats=
#export PCMK_trace_files=


### crm error ###

crm(live)configure# property stonith-enabled="false"
crm(live)configure# commit
Call cib_replace failed (-41): Remote node did not respond
<null>
ERROR: could not replace cib
INFO: offending xml: <configuration>
        <crm_config>
                <cluster_property_set id="cib-bootstrap-options">
                        <nvpair id="cib-bootstrap-options-stonith-enabled" name="stonith-enabled" value="false"/>
                </cluster_property_set>
        </crm_config>
        <nodes/>
        <resources/>
        <constraints/>
</configuration>

### netstat node 1 ###
netstat -tulpen
udp        0      0 10.26.29.238:5404           0.0.0.0:*                               0          141073     9501/corosync
udp        0      0 10.26.29.238:5405           0.0.0.0:*                               0          141074     9501/corosync
udp        0      0 226.94.1.1:5405             0.0.0.0:*                               0          141072     9501/corosync
(and same for node 2 with its respective ip 10.26.29.239)

netstat -nlpa | grep corosync

udp        0      0 10.26.29.238:5404           0.0.0.0:*                               9501/corosync
udp        0      0 10.26.29.238:5405           0.0.0.0:*                               9501/corosync
udp        0      0 226.94.1.1:5405             0.0.0.0:*                               9501/corosync
unix  2      [ ACC ]     STREAM     LISTENING     141067 9501/corosync       @corosync.ipc
unix  3      [ ]         STREAM     CONNECTED     141236 9501/corosync       @corosync.ipc
unix  3      [ ]         STREAM     CONNECTED     141225 9501/corosync       @corosync.ipc
unix  3      [ ]         STREAM     CONNECTED     141200 9501/corosync       @corosync.ipc
unix  3      [ ]         STREAM     CONNECTED     141161 9501/corosync       @corosync.ipc
unix  3      [ ]         STREAM     CONNECTED     141152 9501/corosync       @corosync.ipc
unix  3      [ ]         STREAM     CONNECTED     141132 9501/corosync       @corosync.ipc
unix  3      [ ]         STREAM     CONNECTED     141124 9501/corosync       @corosync.ipc
unix  2      [ ]         DGRAM                    141064 9501/corosync


### /var/log/messages ###

Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [MAIN  ] Corosync Cluster Engine ('1.4.1'): started and ready to provide service.
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [MAIN  ] Corosync built-in features: nss dbus rdma snmp
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [MAIN  ] Successfully read main configuration file '/etc/corosync/corosync.conf'.
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [TOTEM ] Initializing transport (UDP/IP Multicast).
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [TOTEM ] Initializing transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0).
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [TOTEM ] The network interface [10.26.29.238] is now up.
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [SERV  ] Service engine loaded: corosync extended virtual synchrony service
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [SERV  ] Service engine loaded: corosync configuration service
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [SERV  ] Service engine loaded: corosync cluster closed process group service v1.01
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [SERV  ] Service engine loaded: corosync cluster config database access v1.01
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [SERV  ] Service engine loaded: corosync profile loading service
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [SERV  ] Service engine loaded: corosync cluster quorum service v0.1
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [MAIN  ] Compatibility mode set to whitetank.  Using V1 and V2 of the synchronization engine.
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [TOTEM ] A processor joined or left the membership and a new membership was formed.
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [CPG   ] chosen downlist: sender r(0) ip(10.26.29.238) ; members(old:0 left:0)
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [MAIN  ] Completed service synchronization, ready to provide service.
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [TOTEM ] A processor joined or left the membership and a new membership was formed.
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [CPG   ] chosen downlist: sender r(0) ip(10.26.29.238) ; members(old:1 left:0)
Mar 14 18:01:23 wkse13p1xynfs01 corosync[9501]:   [MAIN  ] Completed service synchronization, ready to provide service.
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: Invoked: pacemakerd
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/root
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: config_find_next: No additional configuration supplied for: service
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: config_find_next: No additional configuration supplied for: quorum
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: get_config_opt: No default for option: provider
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: get_cluster_type: Detected an active 'corosync' cluster
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: read_config: Reading configure for stack: corosync
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: config_find_next: Processing additional logging options...
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: get_config_opt: Found 'on' for option: debug
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: get_config_opt: Found 'yes' for option: to_logfile
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: get_config_opt: Found '/var/log/cluster/corosync.log' for option: logfile
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: get_config_opt: Found 'yes' for option: to_syslog
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9514]: info: get_config_opt: Defaulting to 'daemon' for option: syslog_facility
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/root
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: main: Starting Pacemaker 1.1.6-3.el6 (Build: a02c0f19a00c1eb2527ad38f146ebc0834814558):  generated-manpages agent-manpages ascii-docs publican-docs ncurses trace-logging cman corosync-quorum corosync
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: main: Maximum core file size is: 18446744073709551615
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: 0x1f41220 Node 3994884618 now known as wkse13p1xynfs01 (was: (null))
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs01 now has process list: 00000000000000000000000000000002 (was 00000000000000000000000000000000)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: G_main_add_SignalHandler: Added signal handler for signal 17
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: start_child: Forked child 9521 for process stonith-ng
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs01 now has process list: 00000000000000000000000000100002 (was 00000000000000000000000000000002)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: start_child: Forked child 9522 for process cib
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs01 now has process list: 00000000000000000000000000100102 (was 00000000000000000000000000100002)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: start_child: Forked child 9523 for process lrmd
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs01 now has process list: 00000000000000000000000000100112 (was 00000000000000000000000000100102)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: start_child: Forked child 9524 for process attrd
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs01 now has process list: 00000000000000000000000000101112 (was 00000000000000000000000000100112)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: start_child: Forked child 9525 for process pengine
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs01 now has process list: 00000000000000000000000000111112 (was 00000000000000000000000000101112)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: start_child: Forked child 9526 for process crmd
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs01 now has process list: 00000000000000000000000000111312 (was 00000000000000000000000000111112)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: main: Starting mainloop
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: 0x1f43510 Node 4011661834 now known as wkse13p1xynfs02 (was: (null))
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs02 now has process list: 00000000000000000000000000000002 (was 00000000000000000000000000000000)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs02 now has process list: 00000000000000000000000000100002 (was 00000000000000000000000000000002)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs02 now has process list: 00000000000000000000000000100102 (was 00000000000000000000000000100002)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs02 now has process list: 00000000000000000000000000100112 (was 00000000000000000000000000100102)
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs02 now has process list: 00000000000000000000000000101112 (was 00000000000000000000000000100112)
Mar 14 18:01:29 wkse13p1xynfs01 lrmd: [9523]: info: G_main_add_SignalHandler: Added signal handler for signal 15
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs02 now has process list: 00000000000000000000000000111112 (was 00000000000000000000000000101112)
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: Invoked: /usr/lib64/heartbeat/stonithd
Mar 14 18:01:29 wkse13p1xynfs01 pacemakerd: [9517]: info: update_node_processes: Node wkse13p1xynfs02 now has process list: 00000000000000000000000000111312 (was 00000000000000000000000000111112)
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/root
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: G_main_add_SignalHandler: Added signal handler for signal 17
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/hacluster
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: G_main_add_TriggerHandler: Added signal manual handler
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: G_main_add_SignalHandler: Added signal handler for signal 17
Mar 14 18:01:29 wkse13p1xynfs01 lrmd: [9523]: info: G_main_add_SignalHandler: Added signal handler for signal 17
Mar 14 18:01:29 wkse13p1xynfs01 lrmd: [9523]: info: enabling coredumps
Mar 14 18:01:29 wkse13p1xynfs01 lrmd: [9523]: info: G_main_add_SignalHandler: Added signal handler for signal 10
Mar 14 18:01:29 wkse13p1xynfs01 lrmd: [9523]: info: G_main_add_SignalHandler: Added signal handler for signal 12
Mar 14 18:01:29 wkse13p1xynfs01 lrmd: [9523]: info: Started.
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: get_cluster_type: Cluster type is: 'corosync'
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: Invoked: /usr/lib64/heartbeat/attrd
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: retrieveCib: Reading cluster configuration from: /var/lib/heartbeat/crm/cib.xml (digest: /var/lib/heartbeat/crm/cib.xml.sig)
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: validate_with_relaxng: Creating RNG parser context
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/hacluster
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: main: Starting up
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: get_cluster_type: Cluster type is: 'corosync'
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Mar 14 18:01:29 wkse13p1xynfs01 crmd: [9526]: info: Invoked: /usr/lib64/heartbeat/crmd
Mar 14 18:01:29 wkse13p1xynfs01 crmd: [9526]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/hacluster
Mar 14 18:01:29 wkse13p1xynfs01 pengine: [9525]: info: Invoked: /usr/lib64/heartbeat/pengine
Mar 14 18:01:29 wkse13p1xynfs01 pengine: [9525]: info: crm_log_init_worker: Changed active directory to /var/lib/heartbeat/cores/hacluster
Mar 14 18:01:29 wkse13p1xynfs01 crmd: [9526]: info: main: CRM Hg Version: a02c0f19a00c1eb2527ad38f146ebc0834814558
Mar 14 18:01:29 wkse13p1xynfs01 crmd: [9526]: info: crmd_init: Starting crmd
Mar 14 18:01:29 wkse13p1xynfs01 crmd: [9526]: info: G_main_add_SignalHandler: Added signal handler for signal 17
Mar 14 18:01:29 wkse13p1xynfs01 pengine: [9525]: info: main: Starting pengine
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: init_ais_connection_once: Connection to 'corosync': established
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: crm_new_peer: Node wkse13p1xynfs01 now has id: 3994884618
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: crm_new_peer: Node 3994884618 is now known as wkse13p1xynfs01
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: main: Starting stonith-ng mainloop
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: crm_update_peer: Node wkse13p1xynfs01: id=3994884618 state=unknown addr=(null) votes=0 born=0 seen=0 proc=00000000000000000000000000111312 (new)
Mar 14 18:01:29 wkse13p1xynfs01 stonith-ng: [9521]: info: crm_new_peer: Node 0 is now known as wkse13p1xynfs02
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: init_ais_connection_once: Connection to 'corosync': established
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: crm_new_peer: Node wkse13p1xynfs01 now has id: 3994884618
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: crm_new_peer: Node 3994884618 is now known as wkse13p1xynfs01
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: main: Cluster connection active
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: main: Accepting attribute updates
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: notice: main: Starting mainloop...
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: crm_update_peer: Node wkse13p1xynfs01: id=3994884618 state=unknown addr=(null) votes=0 born=0 seen=0 proc=00000000000000000000000000111312 (new)
Mar 14 18:01:29 wkse13p1xynfs01 attrd: [9524]: info: crm_new_peer: Node 0 is now known as wkse13p1xynfs02
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: startCib: CIB Initialization completed successfully
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: get_cluster_type: Cluster type is: 'corosync'
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: init_ais_connection_once: Connection to 'corosync': established
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: crm_new_peer: Node wkse13p1xynfs01 now has id: 3994884618
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: crm_new_peer: Node 3994884618 is now known as wkse13p1xynfs01
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: cib_init: Starting cib mainloop
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: crm_update_peer: Node wkse13p1xynfs01: id=3994884618 state=unknown addr=(null) votes=0 born=0 seen=0 proc=00000000000000000000000000111312 (new)
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: crm_new_peer: Node 0 is now known as wkse13p1xynfs02
Mar 14 18:01:29 wkse13p1xynfs01 cib: [9522]: info: Managed write_cib_contents process 9530 exited with return code 0.
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: do_cib_control: CIB connection established
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: get_cluster_type: Cluster type is: 'corosync'
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: notice: crm_cluster_connect: Connecting to cluster infrastructure: corosync
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: init_ais_connection_once: Connection to 'corosync': established
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: crm_new_peer: Node wkse13p1xynfs01 now has id: 3994884618
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: crm_new_peer: Node 3994884618 is now known as wkse13p1xynfs01
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: ais_status_callback: status: wkse13p1xynfs01 is now unknown
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: init_quorum_connection: Configuring Pacemaker to obtain quorum from Corosync
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: notice: init_quorum_connection: Quorum acquired
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: do_ha_control: Connected to the cluster
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: do_started: Delaying start, no membership data (0000000000100000)
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: crmd_init: Starting crmd's mainloop
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: config_query_callback: Shutdown escalation occurs after: 1200000ms
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: config_query_callback: Checking for expired actions every 900000ms
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: notice: crmd_peer_update: Status update: Client wkse13p1xynfs01/crmd now has status [online] (DC=<null>)
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: crm_update_peer: Node wkse13p1xynfs01: id=3994884618 state=unknown addr=(null) votes=0 born=0 seen=0 proc=00000000000000000000000000111312 (new)
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: crm_new_peer: Node 0 is now known as wkse13p1xynfs02
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: ais_status_callback: status: wkse13p1xynfs02 is now unknown
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: pcmk_quorum_notification: Membership 0: quorum retained (0)
Mar 14 18:01:30 wkse13p1xynfs01 crmd: [9526]: info: do_started: Delaying start, no membership data (0000000000100000)
Mar 14 18:01:34 wkse13p1xynfs01 attrd: [9524]: info: cib_connect: Connected to the CIB after 1 signon attempts
Mar 14 18:01:34 wkse13p1xynfs01 attrd: [9524]: info: cib_connect: Sending full refresh



Cheers,

Achim
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.clusterlabs.org/pipermail/pacemaker/attachments/20120314/068d957f/attachment-0002.html>


More information about the Pacemaker mailing list