[ClusterLabs] Antw: [EXT] Stonith failing

Strahil Nikolov hunter86_bg at yahoo.com
Thu Jul 30 01:42:28 EDT 2020

You got plenty of options:
-  IPMI based  fencing like  HP iLO,  DELL iDRAC
-  SCSI-3  persistent reservations (which can be extended to fence  the node when the reservation(s)  were  removed)

- Shared  disk (even iSCSI)  and using SBD (a.k.a. Poison pill) -> in case your hardware has  no watchdog,  you can use  softdog  kernel module  for linux.

Best  Regards,
Strahil Nikolov

На 29 юли 2020 г. 9:01:22 GMT+03:00, Gabriele Bulfon <gbulfon at sonicle.com> написа:
>That one was taken from a specific implementation on Solaris 11.
>The situation is a dual node server with shared storage controller:
>both nodes see the same disks concurrently.
>Here we must be sure that the two nodes are not going to import/mount
>the same zpool at the same time, or we will encounter data corruption:
>node 1 will be perferred for pool 1, node 2 for pool 2, only in case
>one of the node goes down or is taken offline the resources should be
>first free by the leaving node and taken by the other node.
>Would you suggest one of the available stonith in this case?
>Sonicle S.r.l. 
>Quantum Mechanics : 
>Da: Strahil Nikolov
>A: Cluster Labs - All topics related to open-source clustering welcomed
>Gabriele Bulfon
>Data: 29 luglio 2020 6.39.08 CEST
>Oggetto: Re: [ClusterLabs] Antw: [EXT] Stonith failing
>Do you have a reason not to use any stonith already available ?
>Best Regards,
>Strahil Nikolov
>На 28 юли 2020 г. 13:26:52 GMT+03:00, Gabriele Bulfon
>Thanks, I attach here the script.
>It basically runs ssh on the other node with no password (must be
>preconfigured via authorization keys) with commands.
>This was taken from a script by OpenIndiana (I think).
>As it stated in the comments, we don't want to halt or boot via ssh,
>only reboot.
>Maybe this is the problem, we should at least have it shutdown when
>asked for.
>Actually if I stop corosync in node 2, I don't want it to shutdown the
>system but just let node 1 keep control of all resources.
>Same if I just shutdown manually node 2, 
>node 1 should keep control of all resources and release them back on
>Instead, when I stopped corosync on node 2, log was showing the
>temptative to stonith node 2: why?
>Sonicle S.r.l. 
>Quantum Mechanics : 
>Reid Wahl
>Cluster Labs - All topics related to open-source clustering welcomed
>28 luglio 2020 12.03.46 CEST
>Re: [ClusterLabs] Antw: [EXT] Stonith failing
>"No route to host" is a somewhat generic error message when we can't
>find anyone to fence the node. It doesn't mean there's necessarily a
>network routing issue at fault; no need to focus on that error message.
>I agree with Ulrich about needing to know what the script does. But
>based on your initial message, it sounds like your custom fence agent
>returns 1 in response to "on" and "off" actions. Am I understanding
>correctly? If so, why does it behave that way? Pacemaker is trying to
>run a poweroff action based on the logs, so it needs your script to
>support an off action.
>On Tue, Jul 28, 2020 at 2:47 AM Ulrich Windl
>Ulrich.Windl at rz.uni-regensburg.de
>Gabriele Bulfon
>gbulfon at sonicle.com
>schrieb am 28.07.2020 um 10:56 in
>Hi, now I have my two nodes (xstha1 and xstha2) with IPs configured by
>To check how stonith would work, I turned off Corosync service on
>First node try to attempt to stonith 2nd node and take care of its
>resources, but this fails.
>Stonith action is configured to run a custom script to run ssh
>I think you should explain what that script does exactly.
>Manage your subscription:
>ClusterLabs home:
>Reid Wahl, RHCA
>Software Maintenance Engineer, Red Hat
>CEE - Platform Support Delivery - ClusterHA
>_______________________________________________Manage your
>home: https://www.clusterlabs.org/

More information about the Users mailing list