[ClusterLabs] unable to start fence_scsi on a new add node
Stefan Sabolowitsch
Stefan.Sabolowitsch at felten-group.com
Sun Apr 19 06:38:26 EDT 2020
Privjet / Hello Andrei (happy easter to russia)
thanks for the tip, which made me even more brought something, but the volume is not mounted now on the new node.
[root at elastic ~]# pcs status
Cluster name: cluster_elastic
Stack: corosync
Current DC: elastic-02 (version 1.1.20-5.el7_7.2-3c4c782f70) - partition with quorum
Last updated: Sun Apr 19 12:21:01 2020
Last change: Sun Apr 19 12:15:29 2020 by root via cibadmin on elastic-03
3 nodes configured
10 resources configured
Online: [ elastic-01 elastic-02 elastic-03 ]
Full list of resources:
scsi (stonith:fence_scsi): Started elastic-01
Clone Set: dlm-clone [dlm]
Started: [ elastic-01 elastic-02 elastic-03 ]
Clone Set: clvmd-clone [clvmd]
Started: [ elastic-01 elastic-02 elastic-03 ]
Clone Set: fs_gfs2-clone [fs_gfs2]
Started: [ elastic-01 elastic-02 ]
Stopped: [ elastic-03 ]
Failed Resource Actions:
* fs_gfs2_start_0 on elastic-03 'unknown error' (1): call=53, status=complete, exitreason='Couldn't mount device [/dev/vg_cluster/lv_cluster] as /data-san',
last-rc-change='Sun Apr 19 12:02:44 2020', queued=0ms, exec=1015ms
Failed Fencing Actions:
* unfencing of elastic-03 failed: delegate=, client=crmd.5149, origin=elastic-02,
last-failed='Sun Apr 19 11:32:59 2020'
Daemon Status:
corosync: active/enabled
pacemaker: active/enabled
pcsd: active/enabled
my config:
Cluster Name: cluster_elastic
Corosync Nodes:
elastic-01 elastic-02 elastic-03
Pacemaker Nodes:
elastic-01 elastic-02 elastic-03
Resources:
Clone: dlm-clone
Meta Attrs: interleave=true ordered=true
Resource: dlm (class=ocf provider=pacemaker type=controld)
Operations: monitor interval=30s on-fail=fence (dlm-monitor-interval-30s)
start interval=0s timeout=90 (dlm-start-interval-0s)
stop interval=0s timeout=100 (dlm-stop-interval-0s)
Clone: clvmd-clone
Meta Attrs: interleave=true ordered=true
Resource: clvmd (class=ocf provider=heartbeat type=clvm)
Operations: monitor interval=30s on-fail=fence (clvmd-monitor-interval-30s)
start interval=0s timeout=90s (clvmd-start-interval-0s)
stop interval=0s timeout=90s (clvmd-stop-interval-0s)
Clone: fs_gfs2-clone
Meta Attrs: interleave=true
Resource: fs_gfs2 (class=ocf provider=heartbeat type=Filesystem)
Attributes: device=/dev/vg_cluster/lv_cluster directory=/data-san fstype=gfs2 options=noatime,nodiratime
Operations: monitor interval=10s on-fail=fence (fs_gfs2-monitor-interval-10s)
notify interval=0s timeout=60s (fs_gfs2-notify-interval-0s)
start interval=0s timeout=60s (fs_gfs2-start-interval-0s)
stop interval=0s timeout=60s (fs_gfs2-stop-interval-0s)
Stonith Devices:
Resource: scsi (class=stonith type=fence_scsi)
Attributes: pcmk_host_list="elastic-01 elastic-02 elastic-03" pcmk_monitor_action=metadata pcmk_reboot_action=offdevices=/dev/mapper/mpatha verbose
=true
Meta Attrs: provides=unfencing
Operations: monitor interval=60s (scsi-monitor-interval-60s)
Fencing Levels:
Location Constraints:
Ordering Constraints:
start dlm-clone then start clvmd-clone (kind:Mandatory) (id:order-dlm-clone-clvmd-clone-mandatory)
start clvmd-clone then start fs_gfs2-clone (kind:Mandatory) (id:order-clvmd-clone-fs_gfs2-clone-mandatory)
Colocation Constraints:
clvmd-clone with dlm-clone (score:INFINITY) (id:colocation-clvmd-clone-dlm-clone-INFINITY)
fs_gfs2-clone with clvmd-clone (score:INFINITY) (id:colocation-fs_gfs2-clone-clvmd-clone-INFINITY)
Ticket Constraints:
Alerts:
No alerts defined
Resources Defaults:
No defaults set
Operations Defaults:
No defaults set
Cluster Properties:
cluster-infrastructure: corosync
cluster-name: cluster_elastic
dc-version: 1.1.20-5.el7_7.2-3c4c782f70
have-watchdog: false
maintenance-mode: false
no-quorum-policy: ignore
Quorum:
Options:
More information about the Users
mailing list