[ClusterLabs] unable to start fence_scsi on a new add node

Stefan Sabolowitsch Stefan.Sabolowitsch at felten-group.com
Sun Apr 19 06:38:26 EDT 2020

Privjet / Hello Andrei (happy easter to russia)
thanks for the tip, which made me even more brought something, but the volume is not mounted now on the new node.

[root at elastic ~]# pcs status
Cluster name: cluster_elastic
Stack: corosync
Current DC: elastic-02 (version 1.1.20-5.el7_7.2-3c4c782f70) - partition with quorum
Last updated: Sun Apr 19 12:21:01 2020
Last change: Sun Apr 19 12:15:29 2020 by root via cibadmin on elastic-03

3 nodes configured
10 resources configured

Online: [ elastic-01 elastic-02 elastic-03 ]

Full list of resources:

 scsi	(stonith:fence_scsi):	Started elastic-01
 Clone Set: dlm-clone [dlm]
     Started: [ elastic-01 elastic-02 elastic-03 ]
 Clone Set: clvmd-clone [clvmd]
     Started: [ elastic-01 elastic-02 elastic-03 ]
 Clone Set: fs_gfs2-clone [fs_gfs2]
     Started: [ elastic-01 elastic-02 ]
     Stopped: [ elastic-03 ]

Failed Resource Actions:
* fs_gfs2_start_0 on elastic-03 'unknown error' (1): call=53, status=complete, exitreason='Couldn't mount device [/dev/vg_cluster/lv_cluster] as /data-san',
    last-rc-change='Sun Apr 19 12:02:44 2020', queued=0ms, exec=1015ms

Failed Fencing Actions:
* unfencing of elastic-03 failed: delegate=, client=crmd.5149, origin=elastic-02,
    last-failed='Sun Apr 19 11:32:59 2020'

Daemon Status:
  corosync: active/enabled
  pacemaker: active/enabled
  pcsd: active/enabled

my config:

Cluster Name: cluster_elastic
Corosync Nodes:
 elastic-01 elastic-02 elastic-03
Pacemaker Nodes:
 elastic-01 elastic-02 elastic-03

 Clone: dlm-clone
  Meta Attrs: interleave=true ordered=true
  Resource: dlm (class=ocf provider=pacemaker type=controld)
   Operations: monitor interval=30s on-fail=fence (dlm-monitor-interval-30s)
               start interval=0s timeout=90 (dlm-start-interval-0s)
               stop interval=0s timeout=100 (dlm-stop-interval-0s)
 Clone: clvmd-clone
  Meta Attrs: interleave=true ordered=true
  Resource: clvmd (class=ocf provider=heartbeat type=clvm)
   Operations: monitor interval=30s on-fail=fence (clvmd-monitor-interval-30s)
               start interval=0s timeout=90s (clvmd-start-interval-0s)
               stop interval=0s timeout=90s (clvmd-stop-interval-0s)
 Clone: fs_gfs2-clone
  Meta Attrs: interleave=true
  Resource: fs_gfs2 (class=ocf provider=heartbeat type=Filesystem)
   Attributes: device=/dev/vg_cluster/lv_cluster directory=/data-san fstype=gfs2 options=noatime,nodiratime
   Operations: monitor interval=10s on-fail=fence (fs_gfs2-monitor-interval-10s)
               notify interval=0s timeout=60s (fs_gfs2-notify-interval-0s)
               start interval=0s timeout=60s (fs_gfs2-start-interval-0s)
               stop interval=0s timeout=60s (fs_gfs2-stop-interval-0s)

Stonith Devices:
 Resource: scsi (class=stonith type=fence_scsi)
  Attributes: pcmk_host_list="elastic-01 elastic-02 elastic-03" pcmk_monitor_action=metadata pcmk_reboot_action=offdevices=/dev/mapper/mpatha verbose
  Meta Attrs: provides=unfencing
  Operations: monitor interval=60s (scsi-monitor-interval-60s)
Fencing Levels:

Location Constraints:
Ordering Constraints:
  start dlm-clone then start clvmd-clone (kind:Mandatory) (id:order-dlm-clone-clvmd-clone-mandatory)
  start clvmd-clone then start fs_gfs2-clone (kind:Mandatory) (id:order-clvmd-clone-fs_gfs2-clone-mandatory)
Colocation Constraints:
  clvmd-clone with dlm-clone (score:INFINITY) (id:colocation-clvmd-clone-dlm-clone-INFINITY)
  fs_gfs2-clone with clvmd-clone (score:INFINITY) (id:colocation-fs_gfs2-clone-clvmd-clone-INFINITY)
Ticket Constraints:

 No alerts defined

Resources Defaults:
 No defaults set
Operations Defaults:
 No defaults set

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: cluster_elastic
 dc-version: 1.1.20-5.el7_7.2-3c4c782f70
 have-watchdog: false
 maintenance-mode: false
 no-quorum-policy: ignore


More information about the Users mailing list