[Pacemaker] Problems with corosync while forking processes during node startup.

Dan Frincu df.cluster at gmail.com
Tue Feb 19 09:23:12 EST 2013


Hi,

On Tue, Feb 19, 2013 at 3:29 PM, marc rosenbaum <da-nuke at web.de> wrote:
> Hi everyone,
> I am using SLES 11 sp2 64bit with corosync 1.4.5, Pacemaker 1.1.6.
> The cluster is using a simple two node configuration with NFS, samba, one clusterip and sfex for service fencing.
> This cluster did a good job, but a few days ago one clusternode hung during the shutdownprocedure. Because of this i had to reboot the node. Since hat the node is doing some trouble while trying to go online.
>
> For me it looks like corosync has problems while forking the processes like cib, crmd, attrd, pengine and mgmtd.
> If i start these processes manually as root they are doing their job but while starting via corosync they are not.
>
> I tried to switch the communication from multicast to broadcast to be sure that there is no networkproblem. It seems like corosync can communicate between the nodes with both configurations.
>
> Attached  you will find the configuration and the related logfile.
>
> It would be cool if someone could give me the hint to solve my problem.
> Thanks
>
> Marc
> Corosync.conf:
>
> # Please read the corosync.conf.5 manual page
> compatibility: whitetank
>
> aisexec {
>          # Run as root - this is necessary to be able to manage
>          # resources with Pacemaker
>          user:           root
>          group:          root
> }
>
> service {
>          # Load the Pacemaker Cluster Resource Manager
>          ver:            0

Change this to ver: 1, have pacemaker start from the init script.
For more details on why you should use ver:1 read
http://theclusterguy.clusterlabs.org/post/907043024/introducing-the-pacemaker-master-control-process-for

HTH,
Dan

>          name:           pacemaker
>          use_mgmtd:      yes
>          use_logd:       yes
> }
>
> totem {
>          # The only valid version is 2
>          version:        2
>
>          # How long before declaring a token lost (ms)
>          token:          5000
>
>          # How many token retransmits before forming a new configuration
>          token_retransmits_before_loss_const: 10
>
>          # How long to wait for join messages in the membership protocol
> (ms)
>          join:           60
>
>          # How long to wait for consensus to be achieved before starting
>          # a new round of membership configuration (ms)
>          consensus:      6000
>
>          # Turn off the virtual synchrony filter
>          vsftype:        none
>
>          # Number of messages that may be sent by one processor on
>          # receipt of the token
>          max_messages:   20
>
>          # Limit generated nodeids to 31-bits (positive signed integers)
>          clear_node_high_bit: yes
>
>          # Disable encryption
>          secauth:        on
>
>          # How many threads to use for encryption/decryption
>          threads:        0
>
>          # Optionally assign a fixed node id (integer)
>          # nodeid:       1234
>
>          interface {
>                  ringnumber: 0
>                  bindnetaddr: 10.10.36.0
>                  broadcast: yes
> #               mcastaddr: 226.94.1.1
> #                mcastaddr: 239.255.0.11
>                  mcastport: 5405
>                  ttl: 1
>          }
> }
>
> logging {
>          fileline: off
>          to_stderr: no
>          to_logfile: no
>          to_syslog: yes
> #       syslog_facility: daemon
>          syslog_facility: local3
>          debug: on
>          timestamp: off
>          logger_subsys {
>                  subsys: AMF
>                  debug: off
>          }
> }
>
> amf {
>          mode: disabled
> }
>
>
>
>
> Logfile:
>
>
>
>
> Feb 15 09:23:02 server3 corosync[22666]:  [MAIN  ] Corosync Cluster
> Engine ('1.4.5'): started and ready to provide service.
> Feb 15 09:23:02 server3 corosync[22666]:  [MAIN  ] Corosync built-in
> features: nss rdma
> Feb 15 09:23:02 server3 corosync[22666]:  [MAIN  ] Successfully
> configured openais services to load
> Feb 15 09:23:02 server3 corosync[22666]:  [MAIN  ] Successfully read
> main configuration file '/etc/corosync/corosync.conf'.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] waiting_trans_ack
> changed to 1
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Token Timeout (5000
> ms) retransmit timeout (490 ms)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] token hold (382 ms)
> retransmits before loss (10 retrans)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] join (60 ms)
> send_join (0 ms) consensus (6000 ms) merge (200 ms)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] downcheck (1000 ms)
> fail to recv const (2500 msgs)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] seqno unchanged const
> (30 rotations) Maximum network MTU 1402
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] window size per
> rotation (50 messages) maximum messages per rotation (20 messages)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] missed count const (5
> messages)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] send threads (0 threads)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] RRP token expired
> timeout (490 ms)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] RRP token problem
> counter (2000 ms)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] RRP threshold (10
> problem count)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] RRP multicast
> threshold (100 problem count)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] RRP automatic
> recovery check timeout (1000 ms)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] RRP mode set to none.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ]
> heartbeat_failures_allowed (0)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] max_network_delay (50 ms)
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] HeartBeat is
> Disabled. To enable set heartbeat_failures_allowed > 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Initializing
> transport (UDP/IP Multicast).
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Initializing
> transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0).
> Feb 15 09:23:02 server3 corosync[22666]:  [IPC   ] you are using ipc api v2
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Receive multicast
> socket recv buffer size (262142 bytes).
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Transmit multicast
> socket send buffer size (262142 bytes).
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Local receive
> multicast loop socket recv buffer size (262142 bytes).
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Local transmit
> multicast loop socket send buffer size (262142 bytes).
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] The network interface
> [10.10.36.1] is now up.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Created or loaded
> sequence id 350.10.10.36.1 for this ring.
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: openais cluster membership service B.01.01
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Evt exec init request
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: openais event service B.01.01
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: openais checkpoint service B.01.01
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: openais availability management framework B.01.01
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_exec_init_fn
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info: Invoked:
> /usr/lib64/heartbeat/stonithd
> Feb 15 09:23:02 server3 lrmd: [22674]: info: Signal sent to pid=4825,
> waiting for process to exit
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: openais message service B.03.01
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]: lck_exec_init_fn
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: openais distributed locking service B.03.01
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: openais timer service A.01.01
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_user_lookup: Cluster user root has uid=0 gid=0
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> process_ais_conf: Reading configure
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_init: Local handle: 8535092201842016258 for logging
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: Processing additional logging options...
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: get_config_opt:
> Found 'on' for option: debug
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: get_config_opt:
> Found 'no' for option: to_logfile
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: get_config_opt:
> Found 'yes' for option: to_syslog
> Feb 15 09:23:02 server3 lrmd: [4825]: info: lrmd is shutting down
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info: crm_log_init_worker:
> Changed active directory to /var/lib/heartbeat/cores/root
> Feb 15 09:23:02 server3 lrmd: [4825]: debug: [lrmd] stopped
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info: get_cluster_type:
> Cluster type is: 'openais'
> Feb 15 09:23:02 server3 stonith-ng: [22672]: notice:
> crm_cluster_connect: Connecting to cluster infrastructure: classic
> openais (with plugin)
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info:
> init_ais_connection_classic: Creating connection to our Corosync plugin
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: get_config_opt:
> Found 'local3' for option: syslog_facility
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_init: Local handle: 8054506479773810691 for quorum
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: No additional configuration supplied for: quorum
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: get_config_opt:
> No default for option: provider
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_init: Local handle: 7664968412203843588 for service
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: Processing additional service options...
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: Processing additional service options...
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: Processing additional service options...
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: Processing additional service options...
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: Processing additional service options...
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: Processing additional service options...
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: Processing additional service options...
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> config_find_next: Processing additional service options...
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: get_config_opt:
> Found '0' for option: ver
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: get_config_opt:
> Defaulting to 'pcmk' for option: clustername
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: get_config_opt:
> Found 'yes' for option: use_logd
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: get_config_opt:
> Found 'yes' for option: use_mgmtd
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: pcmk_startup:
> CRM: Initialized
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Logging: Initialized
> pcmk_startup
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: pcmk_startup:
> Maximum core file size is: 18446744073709551615
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_user_lookup: Cluster user hacluster has uid=90 gid=90
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: pcmk_startup:
> Service: 9
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: pcmk_startup:
> Local hostname: server3
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> pcmk_update_nodeid: Local node id: 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Creating entry for node 824445450 born on 0
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> 0x697310 Node 824445450 now known as server3 (was: (null))
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has 1 quorum votes (was 0)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node 824445450/server3 is now: member
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: spawn_child:
> Forked child 22672 for process stonith-ng
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_user_lookup: Cluster user hacluster has uid=90 gid=90
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: spawn_child:
> Forked child 22673 for process cib
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: spawn_child:
> Forked child 22674 for process lrmd
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_user_lookup: Cluster user hacluster has uid=90 gid=90
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: spawn_child:
> Forked child 22675 for process attrd
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_user_lookup: Cluster user hacluster has uid=90 gid=90
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: spawn_child:
> Forked child 22676 for process pengine
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_user_lookup: Cluster user hacluster has uid=90 gid=90
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: spawn_child:
> Forked child 22677 for process crmd
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: spawn_child:
> Forked child 22678 for process mgmtd
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: Pacemaker Cluster Manager 1.1.6
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: corosync extended virtual synchrony service
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: corosync configuration service
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: corosync cluster closed process group service v1.01
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: corosync cluster config database access v1.01
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: corosync profile loading service
> Feb 15 09:23:02 server3 corosync[22666]:  [SERV  ] Service engine
> loaded: corosync cluster quorum service v0.1
> Feb 15 09:23:02 server3 corosync[22666]:  [MAIN  ] Compatibility mode
> set to whitetank.  Using V1 and V2 of the synchronization engine.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] entering GATHER state
> from 15.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Creating commit token
> because I am the rep.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Saving state aru 0
> high seq received 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Storing new sequence
> id for ring 354
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] entering COMMIT state.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] got commit token
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] entering RECOVERY state.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] position [0] member
> 10.10.36.1:
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] previous ring seq 350
> rep 10.10.36.1
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] aru 0 high delivered
> 0 received flag 1
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Did not need to
> originate any messages in recovery.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] got commit token
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Sending initial ORF token
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] token retrans flag is
> 0 my set retrans flag0 retrans queue empty 1 count 0, aru 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] install seq 0 aru 0
> high seq received 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] token retrans flag is
> 0 my set retrans flag0 retrans queue empty 1 count 1, aru 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] install seq 0 aru 0
> high seq received 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] token retrans flag is
> 0 my set retrans flag0 retrans queue empty 1 count 2, aru 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] install seq 0 aru 0
> high seq received 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] token retrans flag is
> 0 my set retrans flag0 retrans queue empty 1 count 3, aru 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] install seq 0 aru 0
> high seq received 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] retrans flag count 4
> token aru 0 install seq 0 aru 0 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Resetting old ring state
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] recovery to regular 1-0
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] CLM CONFIGURATION CHANGE
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] New Configuration:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] Members Left:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] Members Joined:
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Evt conf change 1
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] m 0, j 0 l 0
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]: lck_confchg_fn
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_confchg_fn
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] notice:
> pcmk_peer_update: Transitional membership event on ring 852: memb=0,
> new=0, lost=0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] waiting_trans_ack
> changed to 1
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] CLM CONFIGURATION CHANGE
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] New Configuration:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ]     r(0) ip(10.10.36.1)
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] Members Left:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] Members Joined:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ]     r(0) ip(10.10.36.1)
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Evt conf change 0
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] m 1, j 1 l 0
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]: lck_confchg_fn
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_confchg_fn
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] notice:
> pcmk_peer_update: Stable membership event on ring 852: memb=1, new=1, lost=0
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> pcmk_peer_update: NEW:  server3 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_peer_update: Node 824445450 has address r(0) ip(10.10.36.1)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> pcmk_peer_update: MEMB: server3 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Leaving born-on unset: 852
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=0, seq=852
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000151312
> (1381138)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] This node is within
> the primary component and will provide service.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] entering OPERATIONAL
> state.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] A processor joined or
> left the membership and a new membership was formed.
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 1
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (openais cluster membership service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] got nodejoin message
> 10.10.36.1
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 1
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (openais cluster membership service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (dummy AMF service)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 1
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (dummy AMF service)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (openais checkpoint service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 1
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (openais checkpoint service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (openais event service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Evt synchronize
> initialization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] My node ID r(0)
> ip(10.10.36.1)
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Send max event ID updates
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Send open count updates
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] DONE Sending open counts
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Remote channel
> operation request
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] my node ID: 0x31240a0a
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Receive
> EVT_CONF_CHANGE_DONE from nodeid r(0) ip(10.10.36.1)  members 1 checked in 1
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] DONE Sending retained
> events
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Remote channel
> operation request
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] my node ID: 0x31240a0a
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Receive EVT_CONF_DONE
> from nodeid r(0) ip(10.10.36.1) , members 1 checked in 1
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Recovery complete
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 1
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Evt synchronize
> activation
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (openais event service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (corosync cluster closed process group service v1.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [CPG   ] comparing: sender
> r(0) ip(10.10.36.1) ; members(old:0 left:0)
> Feb 15 09:23:02 server3 corosync[22666]:  [CPG   ] chosen downlist:
> sender r(0) ip(10.10.36.1) ; members(old:0 left:0)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 1
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (corosync cluster closed process group service v1.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]: lck_sync_init
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
> lck_sync_resource_lock_timer_stop
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]: lck_sync_process
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNCV2] Committing
> synchronization for openais distributed locking service B.03.01
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
> lck_sync_activate
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
> lck_sync_resource_free
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
> lck_sync_resource_lock_timer_start
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
>   global_lock_count = 0
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_sync_init
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_queue_enter
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_sync_process
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_queue_iterate
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_group_enter
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_sync_process
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_group_iterate
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_reply_enter
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_sync_process
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_reply_iterate
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNCV2] Committing
> synchronization for openais message service B.03.01
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_activate
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_queue_free
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_group_free
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_reply_free
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_queue_timer_restart
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNCV2] Committing
> synchronization for openais availability management framework B.01.01
> Feb 15 09:23:02 server3 corosync[22666]:  [MAIN  ] Completed service
> synchronization, ready to provide service.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] waiting_trans_ack
> changed to 0
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: pcmk_ipc:
> Recorded connection 0x6a9370 for stonith-ng/22672
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> process_ais_message: Msg[0] (dest=local:ais,
> from=server3:stonith-ng.22672, remote=true, size=6): 22672
> Feb 15 09:23:02 server3 stonith-ng: [22672]: debug:
> init_ais_connection_classic: Adding fd=4 to mainloop
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info:
> init_ais_connection_classic: AIS connection established
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info: get_ais_nodeid:
> Server details: id=824445450 uname=server3 cname=pcmk
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info:
> init_ais_connection_once: Connection to 'classic openais (with plugin)':
> established
> Feb 15 09:23:02 server3 stonith-ng: [22672]: debug: crm_new_peer:
> Creating entry for node server3/824445450
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info: crm_new_peer: Node
> server3 now has id: 824445450
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info: crm_new_peer: Node
> 824445450 is now known as server3
> Feb 15 09:23:02 server3 stonith-ng: [22672]: info: main: Starting
> stonith-ng mainloop
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] entering GATHER state
> from 11.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Creating commit token
> because I am the rep.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Saving state aru 10
> high seq received 10
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Storing new sequence
> id for ring 358
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] entering COMMIT state.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] got commit token
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] entering RECOVERY state.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] TRANS [0] member
> 10.10.36.1:
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] position [0] member
> 10.10.36.1:
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] previous ring seq 354
> rep 10.10.36.1
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] aru 10 high delivered
> 10 received flag 1
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] position [1] member
> 10.10.36.2:
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] previous ring seq 354
> rep 10.10.36.2
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] aru 11 high delivered
> 11 received flag 1
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Did not need to
> originate any messages in recovery.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] got commit token
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Sending initial ORF token
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] token retrans flag is
> 0 my set retrans flag0 retrans queue empty 1 count 0, aru 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] install seq 0 aru 0
> high seq received 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] token retrans flag is
> 0 my set retrans flag0 retrans queue empty 1 count 1, aru 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] install seq 0 aru 0
> high seq received 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] token retrans flag is
> 0 my set retrans flag0 retrans queue empty 1 count 2, aru 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] install seq 0 aru 0
> high seq received 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] token retrans flag is
> 0 my set retrans flag0 retrans queue empty 1 count 3, aru 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] install seq 0 aru 0
> high seq received 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] retrans flag count 4
> token aru 0 install seq 0 aru 0 0
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] Resetting old ring state
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] recovery to regular 1-0
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] CLM CONFIGURATION CHANGE
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] New Configuration:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ]     r(0) ip(10.10.36.1)
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] Members Left:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] Members Joined:
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Evt conf change 1
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] m 1, j 0 l 0
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]: lck_confchg_fn
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_confchg_fn
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] notice:
> pcmk_peer_update: Transitional membership event on ring 856: memb=1,
> new=0, lost=0
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> pcmk_peer_update: memb: server3 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] waiting_trans_ack
> changed to 1
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] CLM CONFIGURATION CHANGE
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] New Configuration:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ]     r(0) ip(10.10.36.1)
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ]     r(0) ip(10.10.36.2)
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] Members Left:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] Members Joined:
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ]     r(0) ip(10.10.36.2)
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Evt conf change 0
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] m 2, j 1 l 0
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]: lck_confchg_fn
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_confchg_fn
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] notice:
> pcmk_peer_update: Stable membership event on ring 856: memb=2, new=1, lost=0
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Creating entry for node 841222666 born on 856
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node 841222666/unknown is now: member
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> pcmk_peer_update: NEW:  .pending. 841222666
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_peer_update: Node 841222666 has address r(0) ip(10.10.36.2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> pcmk_peer_update: MEMB: server3 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> pcmk_peer_update: MEMB: .pending. 841222666
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_peer_update: 1 nodes changed
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Born-on set to: 856 (peer)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> 0x697310 Node 824445450 ((null)) born on: 856
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] This node is within
> the primary component and will provide service.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] entering OPERATIONAL
> state.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] A processor joined or
> left the membership and a new membership was formed.
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server4 (1.1.6)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> 0x6920b0 Node 841222666 (server4) born on: 620
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> 0x6920b0 Node 841222666 now known as server4 (was: (null))
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server4 now has process list: 00000000000000000000000000111312
> (1118994)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server4 now has 1 quorum votes (was 0)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 841222666
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 0.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (openais cluster membership service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] got nodejoin message
> 10.10.36.1
> Feb 15 09:23:02 server3 corosync[22666]:  [CLM   ] got nodejoin message
> 10.10.36.2
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.crmd failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[438]
> (dest=local:crmd, from=server4:crmd.30563, remote=true, size=176):
> <create_request_adv origin="post_cache_update" t="crmd" version="3.0.5"
> subt="request" ref
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 0.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 841222666
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (openais cluster membership service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (dummy AMF service)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1234]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=834):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1235]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=851):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1236]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=955):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 841222666
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 0.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (dummy AMF service)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (openais checkpoint service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1237]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1320):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1238]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=888):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 0.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 841222666
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (openais checkpoint service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (openais event service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Evt synchronize
> initialization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Send max event ID updates
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Send set evt ID 0 to
> r(0) ip(10.10.36.1)
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Remote channel
> operation request
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] my node ID: 0x31240a0a
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Received Set event ID
> OP from nodeid 32240a0a to 0 for 31240a0a my addr r(0) ip(10.10.36.1)
> base 1
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Remote channel
> operation request
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] my node ID: 0x31240a0a
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Received Set event ID
> OP from nodeid 32240a0a to 0 for 32240a0a my addr r(0) ip(10.10.36.1)
> base 1
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.crmd failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[439]
> (dest=local:crmd, from=server4:crmd.30563, remote=true, size=176):
> <create_request_adv origin="post_cache_update" t="crmd" version="3.0.5"
> subt="request" ref
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1239]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=934):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1240]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1150):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Remote channel
> operation request
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] my node ID: 0x31240a0a
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Received Set event ID
> OP from nodeid 31240a0a to 0 for 31240a0a my addr r(0) ip(10.10.36.1)
> base 1
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Send open count updates
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] DONE Sending open counts
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Remote channel
> operation request
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] my node ID: 0x31240a0a
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Receive
> EVT_CONF_CHANGE_DONE from nodeid r(0) ip(10.10.36.2)  members 2 checked in 1
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Remote channel
> operation request
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] my node ID: 0x31240a0a
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Receive
> EVT_CONF_CHANGE_DONE from nodeid r(0) ip(10.10.36.1)  members 2 checked in 2
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] I am oldest in my
> transitional config
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Send retained event
> updates
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] DONE Sending retained
> events
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1241]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=955):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Remote channel
> operation request
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] my node ID: 0x31240a0a
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Receive EVT_CONF_DONE
> from nodeid r(0) ip(10.10.36.1) , members 2 checked in 1
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Wait for retained events
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Remote channel
> operation request
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] my node ID: 0x31240a0a
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Receive EVT_CONF_DONE
> from nodeid r(0) ip(10.10.36.2) , members 2 checked in 2
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Process Evt
> synchronization
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Recovery complete
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 841222666
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 0.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [EVT   ] Evt synchronize
> activation
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (openais event service B.01.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> actions starting for (corosync cluster closed process group service v1.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [CPG   ] comparing: sender
> r(0) ip(10.10.36.2) ; members(old:1 left:0)
> Feb 15 09:23:02 server3 corosync[22666]:  [CPG   ] comparing: sender
> r(0) ip(10.10.36.1) ; members(old:1 left:0)
> Feb 15 09:23:02 server3 corosync[22666]:  [CPG   ] chosen downlist:
> sender r(0) ip(10.10.36.1) ; members(old:1 left:0)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1242]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1096):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 824445450
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 0.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] confchg entries 2
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier Start
> Received From 841222666
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 824445450 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Barrier completion
> status for nodeid 841222666 = 1.
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Synchronization
> barrier completed
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNC  ] Committing
> synchronization for (corosync cluster closed process group service v1.01)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.crmd failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[440]
> (dest=local:crmd, from=server4:crmd.30563, remote=true, size=223):
> <create_request_adv origin="join_make_offer" t="crmd" version="3.0.5"
> subt="request" refer
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:02 server3 corosync[22666]:  [pcmk  ] Msg[1243]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=934):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]: lck_sync_init
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
> lck_sync_resource_lock_timer_stop
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]: lck_sync_process
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNCV2] Committing
> synchronization for openais distributed locking service B.03.01
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
> lck_sync_activate
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
> lck_sync_resource_free
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
> lck_sync_resource_lock_timer_start
> Feb 15 09:23:02 server3 corosync[22666]:  [LCK   ] [DEBUG]:
>   global_lock_count = 0
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_sync_init
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_queue_enter
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_sync_process
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_queue_iterate
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_group_enter
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_sync_process
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_group_iterate
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_reply_enter
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]: msg_sync_process
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_reply_iterate
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNCV2] Committing
> synchronization for openais message service B.03.01
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_activate
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_queue_free
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_group_free
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_sync_reply_free
> Feb 15 09:23:02 server3 corosync[22666]:  [MSG   ] [DEBUG]:
> msg_queue_timer_restart
> Feb 15 09:23:02 server3 corosync[22666]:  [SYNCV2] Committing
> synchronization for openais availability management framework B.01.01
> Feb 15 09:23:02 server3 corosync[22666]:  [MAIN  ] Completed service
> synchronization, ready to provide service.
> Feb 15 09:23:02 server3 corosync[22666]:  [TOTEM ] waiting_trans_ack
> changed to 0
> Feb 15 09:23:03 server3 lrmd: [22674]: info: enabling coredumps
> Feb 15 09:23:03 server3 lrmd: [22674]: WARN: Core dumps could be lost if
> multiple dumps occur.
> Feb 15 09:23:03 server3 lrmd: [22674]: WARN: Consider setting
> non-default value in /proc/sys/kernel/core_pattern (or equivalent) for
> maximum supportability
> Feb 15 09:23:03 server3 lrmd: [22674]: WARN: Consider setting
> /proc/sys/kernel/core_uses_pid (or equivalent) to 1 for maximum
> supportability
> Feb 15 09:23:03 server3 lrmd: [22674]: debug: main: run the loop...
> Feb 15 09:23:03 server3 lrmd: [22674]: info: Started.
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] ERROR:
> pcmk_wait_dispatch: Child process cib exited (pid=22673, rc=100)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] notice:
> pcmk_wait_dispatch: Child process cib no longer wishes to be respawned
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000151212
> (1380882)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] ERROR:
> pcmk_wait_dispatch: Child process crmd exited (pid=22677, rc=100)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] notice:
> pcmk_wait_dispatch: Child process crmd no longer wishes to be respawned
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000151012
> (1380370)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] ERROR:
> pcmk_wait_dispatch: Child process attrd exited (pid=22675, rc=100)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] notice:
> pcmk_wait_dispatch: Child process attrd no longer wishes to be respawned
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000150012
> (1376274)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] ERROR:
> pcmk_wait_dispatch: Child process pengine exited (pid=22676, rc=100)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] notice:
> pcmk_wait_dispatch: Child process pengine no longer wishes to be respawned
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000140012
> (1310738)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] ERROR:
> pcmk_wait_dispatch: Child process mgmtd exited (pid=22678, rc=100)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] notice:
> pcmk_wait_dispatch: Child process mgmtd no longer wishes to be respawned
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000100012
> (1048594)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> send_cluster_id: Local update: id=824445450, born=856, seq=856
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000151212
> (1380882)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000100012
> (1048594)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000151212
> (1380882)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000100012
> (1048594)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000151012
> (1380370)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000100012
> (1048594)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000151012
> (1380370)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000100012
> (1048594)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000150012
> (1376274)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000100012
> (1048594)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000150012
> (1376274)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000100012
> (1048594)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000140012
> (1310738)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000100012
> (1048594)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000140012
> (1310738)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info: update_member:
> Node server3 now has process list: 00000000000000000000000000100012
> (1048594)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] info:
> send_member_notification: Sending membership update 856 to 0 children
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] debug:
> pcmk_cluster_id_callback: Node update: server3 (1.1.6)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.crmd failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[443]
> (dest=local:crmd, from=server4:crmd.30563, remote=true, size=176):
> <create_request_adv origin="post_cache_update" t="crmd" version="3.0.5"
> subt="request" ref
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1244]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=955):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1245]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1096):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.crmd failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[444]
> (dest=local:crmd, from=server4:crmd.30563, remote=true, size=176):
> <create_request_adv origin="post_cache_update" t="crmd" version="3.0.5"
> subt="request" ref
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1246]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=934):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1247]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1151):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1248]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=955):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1249]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1097):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.crmd failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[445]
> (dest=local:crmd, from=server4:crmd.30563, remote=true, size=176):
> <create_request_adv origin="post_cache_update" t="crmd" version="3.0.5"
> subt="request" ref
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1250]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=934):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1251]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=955):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1252]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1097):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.crmd failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[446]
> (dest=local:crmd, from=server4:crmd.30563, remote=true, size=176):
> <create_request_adv origin="post_cache_update" t="crmd" version="3.0.5"
> subt="request" ref
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1253]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=934):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1254]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=955):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1255]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1097):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.crmd failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[447]
> (dest=local:crmd, from=server4:crmd.30563, remote=true, size=176):
> <create_request_adv origin="post_cache_update" t="crmd" version="3.0.5"
> subt="request" ref
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1256]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=934):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1257]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=955):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1258]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1097):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1259]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=934):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1260]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=9090): <copy
> __name__="cib_command" t="cib"
> cib_clientid="b7b07259-f8b4-4a2d-a66d-db2947a1cb36" c
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1261]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=886):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1262]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=4495):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1263]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=8505):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1264]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=955):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1265]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=1401):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1266]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=872):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="5297282d-f542-4178-89ab-2750df43
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.cib failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[1267]
> (dest=local:cib, from=server4:cib.30559, remote=true, size=913):
> <cib_command __name__="cib_command" t="cib"
> cib_async_id="90e34654-1b5c-48be-8acd-3a748f8a
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] WARN:
> route_ais_message: Sending message to local.attrd failed: ipc delivery
> failed (rc=-2)
> Feb 15 09:23:03 server3 corosync[22666]:  [pcmk  ] Msg[34]
> (dest=local:attrd, from=server4:attrd.30561, remote=true, size=179):
> <attrd_trigger_update t="attrd" src="server4" task="flush"
> attr_name="probe_complete" at
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org



-- 
Dan Frincu
CCNA, RHCE




More information about the Pacemaker mailing list