[ClusterLabs] "node is unclean" leads to gratuitous reboot

Tomas Jelinek tojeline at redhat.com
Mon Jul 15 08:16:09 EDT 2019


Dne 09. 07. 19 v 14:54 Michael Powell napsal(a):

...

> Here’s the configuration, from the first node –
> 
> [root at mgraid-16201289RN00023-0 bin]# pcs config
> 
> Cluster Name:
> 
> Corosync Nodes:
> 
> mgraid-16201289RN00023-0 mgraid-16201289RN00023-1
> 
> Pacemaker Nodes:
> 
> mgraid-16201289RN00023-0 mgraid-16201289RN00023-1
> 
> Resources:
> 
> Master: ms-SS16201289RN00023
> 
>    Meta Attrs: clone-max=2 notify=true globally-unique=false 
> target-role=Started
> 
>    Resource: SS16201289RN00023 (class=ocf provider=omneon type=ss)
> 
>     Attributes: ss_resource=SS16201289RN00023 
> ssconf=/var/omneon/config/config.16201289RN00023
> 
>     Operations: monitor interval=3s role=Master timeout=7s 
> (SS16201289RN00023-monitor-3s)
> 
>                 monitor interval=10s role=Slave timeout=7 
> (SS16201289RN00023-monitor-10s)
> 
>                 stop interval=0 timeout=20 (SS16201289RN00023-stop-0)
> 
>                 start interval=0 timeout=300 (SS16201289RN00023-start-0)
> 
> Clone: mgraid-stonith-clone
> 
>    Resource: mgraid-stonith (class=stonith type=mgpstonith)
> 
>     Operations: monitor interval=0 timeout=20s 
> (mgraid-stonith-monitor-interval-0)
> 
> Stonith Devices:
> 
> Fencing Levels:
> 
> Location Constraints:
> 
>    Resource: ms-SS16201289RN00023
> 
>      Constraint: ms-SS16201289RN00023-master-w1
> 
>        Rule: role=master score=100  (id:ms-SS16201289RN00023-master-w1-rule)
> 
>          Expression: #uname eq mgraid-16201289rn00023-0  
> (id:ms-SS16201289RN00023-master-w1-rule-expression)
> 
> Ordering Constraints:
> 
> Colocation Constraints:
> 
> Ticket Constraints:
> 
> Alerts:
> 
> No alerts defined
> 
> Resources Defaults:
> 
> failure-timeout: 1min
> 
> Operations Defaults:
> 
> No defaults set
> 
> Cluster Properties:
> 
> cluster-infrastructure: corosync
> 
> cluster-recheck-interval: 1min
> 
> dc-deadtime: 5s
> 
> dc-version: 1.1.19-8.el7-c3c624ea3d
> 
> have-watchdog: false
> 
> last-lrm-refresh: 1562513532
> 
> stonith-enabled: true
> 
> Quorum:
> 
>    Options:
> 
>      wait_for_all: 0
> 
> ------------------------------------------------------------------------
> 
> Interestingly, as you’ll note below, the “/two_node/” option is also set 
> to 1, but is not reported as such above.

Pcs prints only options configurable by the 'pcs quorum update' command 
there: auto_tie_breaker, last_man_standing, last_man_standing_window and 
wait_for_all. Quorum device settings are printed there as well, but you 
do not have any set.

We plan to make pcs print more corosync settings / options in the 
pcs-0.10 branch.


Regards,
Tomas


> Finally, here’s *//etc/corosync/corosync.conf/* –
> 
> totem {
> 
>          version: 2
> 
>          crypto_cipher: none
> 
>          crypto_hash: none
> 
>          interface {
> 
>                  ringnumber: 0
> 
>          bindnetaddr: 169.254.1.1
> 
>                  mcastaddr: 239.255.1.1
> 
>                  mcastport: 5405
> 
>                  ttl: 1
> 
>          }
> 
> }
> 
> logging {
> 
>          fileline: off
> 
>          to_stderr: no
> 
>          to_logfile: yes
> 
>          logfile: /var/log/cluster/corosync.log
> 
>          to_syslog: yes
> 
>          debug: on
> 
>          timestamp: on
> 
>          logger_subsys {
> 
>                  subsys: QUORUM
> 
>                  debug: on
> 
>          }
> 
> }
> 
> nodelist {
> 
>          node {
> 
>                  ring0_addr: mgraid-16201289RN00023-0
> 
>                  nodeid: 1
> 
>          }
> 
>          node {
> 
>                  ring0_addr: mgraid-16201289RN00023-1
> 
>                  nodeid: 2
> 
>          }
> 
> }
> 
> quorum {
> 
>          provider: corosync_votequorum
> 
>          two_node: 1
> 
>          wait_for_all: 0
> 
> }
> 
> I’d appreciate any insight you can offer into this behavior, and any 
> suggestions you may have.
> 
> Regards,
> 
>    Michael
> 
> cid:image004.gif at 01D15F05.08FA4670
> 
>      Michael Powell
> 
>      Sr. Staff Engineer
> 
>      15220 NW Greenbrier Pkwy
> 
>          Suite 290
> 
>      Beaverton, OR   97006
> 
>      T 503-372-7327    M 503-789-3019   H 503-625-5332
> 
> 
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
> 
> ClusterLabs home: https://www.clusterlabs.org/
> 


More information about the Users mailing list