[ClusterLabs] recommendations for corosync totem timeout for CentOS 7 + VMware?

Jan Friesse jfriesse at redhat.com
Mon Mar 25 03:50:22 EDT 2019


Brian,

> On Fri, Mar 22, 2019 at 08:57:20AM +0100, Jan Friesse wrote:
>>> - If I manually set 'totem.token' to a higher value, am I responsible
>>>    for tracking the number of nodes in the cluster, to keep in
>>>    alignment with what Red Hat's page says?
>>
>> Nope. I've tried to explain what is really happening in the manpage
>> corosync.conf(5). totem.token and totem.token_coefficient are used in
>> the following formula:
> 
> I do see this under token_coefficient, thanks.
> 
>> Corosync used runtime.config.token.
> 
> Cool; thanks.  Bumping up totem.token to 2000 got me over this hump.
> 
>>> - Under these conditions, when corosync exits, why does it do so
>>>    with a zero status? It seems to me that if it exited at all,
>>
>> That's a good question. How reproducible is the issue? Corosync
>> shouldn't "exit" with zero status.
> 
> If I leave totem.token set to default, %100 in my case.
> 
> I stand corrected; yesterday, it was %100.  Today, I cannot reproduce
> this at all, even with reverting to the defaults.
> 

That's sad

> Here is a snippet of output from yesterday's experiments; this is
> based on a typescript capture file, so I apologize for the ANSI
> screen codes.
> 

Yep, np. Looks just fine.

> - by default, systemd doesn't report full log lines.
> 
> - by default, CentOS's config of systemd doesn't persist journaled
>    logs, so I can't directly review yesterday's efforts.
> 
> - and, it looks like I misinterpreted the 'exited' message; corosync
>    was enabled and running, but the 'Process' line doesn't report
>    on the 'corosync' process, but some systemd utility.
> 
> (Let me count the ways I'm coming to dislike systemd...)
> 
> I was able to recover logs from /var/log/messages, but other than
> the 'Consider token timeout increase' message, it looks hunky-dory.
> 
> With what I've since learned;
> 
> - I cannot explain why I can't reproduce the symptoms, even with
>    reverting to the defaults.
> 
> - And without being able to reproduce, I can't pursue why 'pcs
>    status cluster' was actually failing for me. :/
> 
> So, I appreciate your attention to this message, and I guess I'm
> off to further explore all of this.
> 
>    C]0;root at node1:~^G[root at node1 ~]# systemctl status corosync.service
>    ESC[1;32m●ESC[0m corosync.service - Corosync Cluster Engine
>     Loaded: loaded (/usr/lib/systemd/system/corosync.service; enabled; vendor
> preset: disabled)
>       Active: ESC[1;32mactive (running)ESC[0m since Thu 2019-03-21 14:26:56
> UTC; 1min 35s ago
>         Docs: man:corosync
>               man:corosync.conf
>               man:corosync_overview
>      Process: 5474 ExecStart=/usr/share/corosync/corosync start (code=exited,
> status=0/SUCCESS)
>     Main PID: 5490 (corosync)
>       CGroup: /system.slice/corosync.service
>             └─5490 corosync
> 
> 

As you can see, corosync service unit in COS 7 is executing init script 
which execs corosync and waits till connection to local IPC can be 
established. IPC connection can be established when corosync is ready. 
Initscript timeout for IPC is 1 minute and return code is 1 if 
connection cannot be established. On success initscript returns 0. So 
ExecStart (initscript) exited with 0/SUCESS = corosync was successfully 
started and it is running as a PID 5490.

Regards,
   Honza

>>    Honza
> 



More information about the Users mailing list