[ClusterLabs] "node is unclean" leads to gratuitous reboot
Tomas Jelinek
tojeline at redhat.com
Mon Jul 15 08:16:09 EDT 2019
Dne 09. 07. 19 v 14:54 Michael Powell napsal(a):
...
> Here’s the configuration, from the first node –
>
> [root at mgraid-16201289RN00023-0 bin]# pcs config
>
> Cluster Name:
>
> Corosync Nodes:
>
> mgraid-16201289RN00023-0 mgraid-16201289RN00023-1
>
> Pacemaker Nodes:
>
> mgraid-16201289RN00023-0 mgraid-16201289RN00023-1
>
> Resources:
>
> Master: ms-SS16201289RN00023
>
> Meta Attrs: clone-max=2 notify=true globally-unique=false
> target-role=Started
>
> Resource: SS16201289RN00023 (class=ocf provider=omneon type=ss)
>
> Attributes: ss_resource=SS16201289RN00023
> ssconf=/var/omneon/config/config.16201289RN00023
>
> Operations: monitor interval=3s role=Master timeout=7s
> (SS16201289RN00023-monitor-3s)
>
> monitor interval=10s role=Slave timeout=7
> (SS16201289RN00023-monitor-10s)
>
> stop interval=0 timeout=20 (SS16201289RN00023-stop-0)
>
> start interval=0 timeout=300 (SS16201289RN00023-start-0)
>
> Clone: mgraid-stonith-clone
>
> Resource: mgraid-stonith (class=stonith type=mgpstonith)
>
> Operations: monitor interval=0 timeout=20s
> (mgraid-stonith-monitor-interval-0)
>
> Stonith Devices:
>
> Fencing Levels:
>
> Location Constraints:
>
> Resource: ms-SS16201289RN00023
>
> Constraint: ms-SS16201289RN00023-master-w1
>
> Rule: role=master score=100 (id:ms-SS16201289RN00023-master-w1-rule)
>
> Expression: #uname eq mgraid-16201289rn00023-0
> (id:ms-SS16201289RN00023-master-w1-rule-expression)
>
> Ordering Constraints:
>
> Colocation Constraints:
>
> Ticket Constraints:
>
> Alerts:
>
> No alerts defined
>
> Resources Defaults:
>
> failure-timeout: 1min
>
> Operations Defaults:
>
> No defaults set
>
> Cluster Properties:
>
> cluster-infrastructure: corosync
>
> cluster-recheck-interval: 1min
>
> dc-deadtime: 5s
>
> dc-version: 1.1.19-8.el7-c3c624ea3d
>
> have-watchdog: false
>
> last-lrm-refresh: 1562513532
>
> stonith-enabled: true
>
> Quorum:
>
> Options:
>
> wait_for_all: 0
>
> ------------------------------------------------------------------------
>
> Interestingly, as you’ll note below, the “/two_node/” option is also set
> to 1, but is not reported as such above.
Pcs prints only options configurable by the 'pcs quorum update' command
there: auto_tie_breaker, last_man_standing, last_man_standing_window and
wait_for_all. Quorum device settings are printed there as well, but you
do not have any set.
We plan to make pcs print more corosync settings / options in the
pcs-0.10 branch.
Regards,
Tomas
> Finally, here’s *//etc/corosync/corosync.conf/* –
>
> totem {
>
> version: 2
>
> crypto_cipher: none
>
> crypto_hash: none
>
> interface {
>
> ringnumber: 0
>
> bindnetaddr: 169.254.1.1
>
> mcastaddr: 239.255.1.1
>
> mcastport: 5405
>
> ttl: 1
>
> }
>
> }
>
> logging {
>
> fileline: off
>
> to_stderr: no
>
> to_logfile: yes
>
> logfile: /var/log/cluster/corosync.log
>
> to_syslog: yes
>
> debug: on
>
> timestamp: on
>
> logger_subsys {
>
> subsys: QUORUM
>
> debug: on
>
> }
>
> }
>
> nodelist {
>
> node {
>
> ring0_addr: mgraid-16201289RN00023-0
>
> nodeid: 1
>
> }
>
> node {
>
> ring0_addr: mgraid-16201289RN00023-1
>
> nodeid: 2
>
> }
>
> }
>
> quorum {
>
> provider: corosync_votequorum
>
> two_node: 1
>
> wait_for_all: 0
>
> }
>
> I’d appreciate any insight you can offer into this behavior, and any
> suggestions you may have.
>
> Regards,
>
> Michael
>
> cid:image004.gif at 01D15F05.08FA4670
>
> Michael Powell
>
> Sr. Staff Engineer
>
> 15220 NW Greenbrier Pkwy
>
> Suite 290
>
> Beaverton, OR 97006
>
> T 503-372-7327 M 503-789-3019 H 503-625-5332
>
>
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
>
More information about the Users
mailing list