[ClusterLabs] Antw: [EXT] Re: SBD fencing not working on my two-node cluster
Ulrich Windl
Ulrich.Windl at rz.uni-regensburg.de
Tue Sep 29 07:50:37 EDT 2020
>>> Strahil Nikolov <hunter86_bg at yahoo.com> schrieb am 22.09.2020 um 07:23 in
Nachricht <1814286403.4657404.1600752191237 at mail.yahoo.com>:
> Replace /dev/sde1 with
> /dev/disk/by-id/scsi-36000c292840d37bd13eb6be46d3af4ab-part1 :
> - in /etc/sysconfig/sbd
> - in the cib (via crom configure edit)
>
> Also, I don't see 'stonith-enabled=true' which could be your actual
problem.
>
> I think you can set it via :
> crm configure property stonith-enabled=true
>
> P.S.: Consider setting the 'resource-stickiness' to '1'.Using partitions is
> not the best option but is better than nothing.
I think partitions are fine, especially when you have a modern SAN storage
where the smallest allocatable amount is 1GB.
The other thing I'd recommend is pre-allocating the message slots for each
cluster node.
Most importantly /dev/disk/by-id/scsi-36000c292840d37bd13eb6be46d3af4ab-part1
is probably the device to specify.
Regards,
Ulrich
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
>
>
> В вторник, 22 септември 2020 г., 02:06:10 Гринуич+3, Philippe M Stedman
> <pmstedma at us.ibm.com> написа:
>
>
>
>
>
> Hi Strahil,
>
> Here is the output of those commands.... I appreciate the help!
>
> # crm config show
> node 1: ceha03 \
> attributes ethmonitor-ens192=1
> node 2: ceha04 \
> attributes ethmonitor-ens192=1
> (...)
> primitive stonith_sbd stonith:fence_sbd \
> params devices="/dev/sde1" \
> meta is-managed=true
> (...)
> property cib-bootstrap-options: \
> have-watchdog=true \
> dc-version=2.0.2-1.el8-744a30d655 \
> cluster-infrastructure=corosync \
> cluster-name=ps_dom \
> stonith-enabled=true \
> no-quorum-policy=ignore \
> stop-all-resources=false \
> cluster-recheck-interval=60 \
> symmetric-cluster=true \
> stonith-watchdog-timeout=0
> rsc_defaults rsc-options: \
> is-managed=false \
> resource-stickiness=0 \
> failure-timeout=1min
>
> # cat /etc/sysconfig/sbd
> SBD_DEVICE="/dev/sde1"
> SBD_PACEMAKER=yes
> SBD_STARTMODE=always
> SBD_DELAY_START=no
> SBD_WATCHDOG_DEV=/dev/watchdog
> SBD_WATCHDOG_TIMEOUT=5
> SBD_TIMEOUT_ACTION=flush,reboot
> SBD_MOVE_TO_ROOT_CGROUP=auto
> SBD_OPTS=
>
> # systemctl status sbd
> sbd.service - Shared-storage based fencing daemon
> Loaded: loaded (/usr/lib/systemd/system/sbd.service; enabled; vendor preset:
> disabled)
> Active: active (running) since Mon 2020-09-21 18:36:28 EDT; 15min ago
> Docs: man:sbd(8)
> Process: 12810 ExecStart=/usr/sbin/sbd $SBD_OPTS -p /var/run/sbd.pid watch
> (code=exited, status=0/SUCCESS)
> Main PID: 12812 (sbd)
> Tasks: 4 (limit: 26213)
> Memory: 14.5M
> CGroup: /system.slice/sbd.service
> \u251c\u250012812 sbd: inquisitor
> \u251c\u250012814 sbd: watcher: /dev/sde1 - slot: 0 - uuid:
> 94d67f15-e301-4fa9-89ae-e3ce2e82c9e7
> \u251c\u250012815 sbd: watcher: Pacemaker
> \u2514\u250012816 sbd: watcher: Cluster
>
> Sep 21 18:36:27 ceha03.canlab.ibm.com systemd[1]: Starting Shared-storage
> based fencing daemon...
> Sep 21 18:36:27 ceha03.canlab.ibm.com sbd[12810]: notice: main: Doing flush
> + writing 'b' to sysrq on timeout
> Sep 21 18:36:27 ceha03.canlab.ibm.com sbd[12815]: pcmk: notice:
> servant_pcmk: Monitoring Pacemaker health
> Sep 21 18:36:27 ceha03.canlab.ibm.com sbd[12816]: cluster: notice:
> servant_cluster: Monitoring unknown cluster health
> Sep 21 18:36:27 ceha03.canlab.ibm.com sbd[12814]: /dev/sde1: notice:
> servant_md: Monitoring slot 0 on disk /dev/sde1
> Sep 21 18:36:28 ceha03.canlab.ibm.com sbd[12812]: notice: watchdog_init:
> Using watchdog device '/dev/watchdog'
> Sep 21 18:36:28 ceha03.canlab.ibm.com sbd[12816]: cluster: notice:
> sbd_get_two_node: Corosync is in 2Node-mode
> Sep 21 18:36:28 ceha03.canlab.ibm.com sbd[12812]: notice: inquisitor_child:
> Servant cluster is healthy (age: 0)
> Sep 21 18:36:28 ceha03.canlab.ibm.com systemd[1]: Started Shared-storage
> based fencing daemon.
>
> # sbd -d /dev/disk/by-id/scsi-<long_uuid> dump
> [root at ceha03 by-id]# sbd -d
> /dev/disk/by-id/scsi-36000c292840d37bd13eb6be46d3af4ab-part1 dump
> ==Dumping header on disk
> /dev/disk/by-id/scsi-36000c292840d37bd13eb6be46d3af4ab-part1
> Header version : 2.1
> UUID : 94d67f15-e301-4fa9-89ae-e3ce2e82c9e7
> Number of slots : 255
> Sector size : 512
> Timeout (watchdog) : 5
> Timeout (allocate) : 2
> Timeout (loop) : 1
> Timeout (msgwait) : 10
> ==Header on disk
> /dev/disk/by-id/scsi-36000c292840d37bd13eb6be46d3af4ab-part1 is dumped
>
>
> Thanks,
>
> Phil Stedman
> Db2 High Availability Development and Support
> Email: pmstedma at us.ibm.com
>
> Strahil Nikolov ---09/21/2020 01:41:10 PM---Can you provide (replace
> sensitive data) : crm configure show
>
> From: Strahil Nikolov <hunter86_bg at yahoo.com>
> To: "users at clusterlabs.org" <users at clusterlabs.org>
> Date: 09/21/2020 01:41 PM
> Subject: [EXTERNAL] Re: [ClusterLabs] SBD fencing not working on my two-node
> cluster
> Sent by: "Users" <users-bounces at clusterlabs.org>
> ________________________________
>
>
>
> Can you provide (replace sensitive data) :
>
> crm configure show
> cat /etc/sysconfig/sbd
> systemctl status sbd
> sbd -d /dev/disk/by-id/scsi-<long_uuid> dump
>
> P.S.: It is very bad practice to use "/dev/sdXYZ" as these are not
> permanent.Always use persistent names like those inside
> "/dev/disk/by-XYZ/ZZZZ". Also , SBD needs max 10MB block device and yours
> seems unnecessarily big.
>
>
> Most probably /dev/sde1 is your problem.
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
> В понеделник, 21 септември 2020 г., 23:19:47 Гринуич+3, Philippe M Stedman
> <pmstedma at us.ibm.com> написа:
>
>
>
>
>
> Hi,
>
> I have been following the instructions on the following page to try and
> configure SBD fencing on my two-node cluster:
> https://documentation.suse.com/sle-ha/15-SP1/html/SLE-HA-all/cha-ha-storage-
> protect.html
>
> I am able to get through all the steps successfully, I am using the
> following device (/dev/sde1) as my shared disk:
>
> Disk /dev/sde: 20 GiB, 21474836480 bytes, 41943040 sectors
> Units: sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disklabel type: gpt
> Disk identifier: 43987868-1C0B-41CE-8AF8-C522AB259655
>
> Device Start End Sectors Size Type
> /dev/sde1 48 41942991 41942944 20G IBM General Parallel Fs
>
> Since, I don't have a hardware watchdog at my disposal, I am using the
> software watchdog (softdog) instead. Having said this, I am able to get
> through all the steps successfully... I create the fence agent resource
> successfully, it shows as Started in crm status output:
>
> stonith_sbd (stonith:fence_sbd): Started ceha04
>
> The problem is when I run crm node fence ceha04 to test out fencing a host
> in my cluster. From the crm status output, I see that the reboot action has
> failed and furthermore, in the system logs, I see the following messages:
>
> Sep 21 14:12:33 ceha04 pacemaker-controld[24146]: notice: Requesting fencing
> (reboot) of node ceha04
> Sep 21 14:12:33 ceha04 pacemaker-fenced[24142]: notice: Client
> pacemaker-controld.24146.5ff1ac0c wants to fence (reboot) 'ceha04' with
> device '(any)'
> Sep 21 14:12:33 ceha04 pacemaker-fenced[24142]: notice: Requesting peer
> fencing (reboot) of ceha04
> Sep 21 14:12:33 ceha04 pacemaker-fenced[24142]: notice: Couldn't find anyone
> to fence (reboot) ceha04 with any device
> Sep 21 14:12:33 ceha04 pacemaker-fenced[24142]: error: Operation reboot of
> ceha04 by <no-one> for pacemaker-controld.24146 at ceha04.1bad3987: No such
device
> Sep 21 14:12:33 ceha04 pacemaker-controld[24146]: notice: Stonith operation
> 3/1:4317:0:ec560474-96ea-4984-b801-400d11b5b3ae: No such device (-19)
> Sep 21 14:12:33 ceha04 pacemaker-controld[24146]: notice: Stonith operation
> 3 for ceha04 failed (No such device): aborting transition.
> Sep 21 14:12:33 ceha04 pacemaker-controld[24146]: warning: No devices found
> in cluster to fence ceha04, giving up
> Sep 21 14:12:33 ceha04 pacemaker-controld[24146]: notice: Transition 4317
> aborted: Stonith failed
> Sep 21 14:12:33 ceha04 pacemaker-controld[24146]: notice: Peer ceha04 was
> not terminated (reboot) by <anyone> on behalf of pacemaker-controld.24146:
No
> such device
>
> I don't know why Pacemaker isn't able to discover my fencing resource, why
> isn't it able to find anyone to fence the host from the cluster?
>
> Any help is greatly appreciated. I can provide more details as required.
>
> Thanks,
>
> Phil Stedman
> Db2 High Availability Development and Support
> Email: pmstedma at us.ibm.com
>
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
>
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
>
>
>
>
>
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
More information about the Users
mailing list