[ClusterLabs] Completely disabled resource failure triggered fencing
Ken Gaillot
kgaillot at redhat.com
Mon Jan 18 14:02:26 EST 2021
On Mon, 2021-01-18 at 13:01 +0000, Strahil Nikolov wrote:
> Have you tried on-fail=ignore option ?
>
> Best Regards,
> Strahil Nikolov
on-fail=ignore will act as if the operation succeeded, which probably
isn't desired here. It's usually used for flaky/buggy devices/agents
that sometimes (or always) report failure for successful starts or
monitors, or for noncritical resources where a monitor failure is
interesting (in status displays) but it's not worth doing anything
about.
on-fail=block does make more sense (essentially it means "wait for a
human to look into it"). Also I'm not sure whether on-fail=ignore is
allowed for stop.
>
>
>
>
>
> В неделя, 17 януари 2021 г., 20:45:27 Гринуич+2, Digimer <
> lists at alteeve.ca> написа:
>
>
>
>
>
> Hi all,
>
> I'm trying to figure out how to define a resource such that if it
> fails in any way, it will not cause pacemaker self self-fence. The
> reasoning being that there are relatively minor ways to fault a
> single
> resource (these are VMs, so for example, a bad edit to the XML
> definition renders it invalid, or the definition is accidentally
> removed).
>
> In a case like this, I fully expect that resource to enter a failed
> state. Of course, pacemaker won't be able to stop it, migrate it,
> etc.
> When this happens currently, it causes the host to self-fence, taking
> down all other hosted resources (servers). This is less than ideal.
>
> Is there a way to tell pacemaker that if it's unable to manage a
> resource, it flags it as failed and leaves it at that? I've been
> trying
> to do this and my config so far is;
>
> pcs resource create srv07-el6 ocf:alteeve:server name="srv07-el6" \
> meta allow-migrate="true" target-role="stopped" \
> op monitor interval="60" start timeout="INFINITY" \
> on-fail="block" stop timeout="INFINITY" on-fail="block" \
> migrate_to timeout="INFINITY"
>
> This is getting cumbersome and still, in testing, I'm finding cases
> where the node gets fenced when something breaks the resource in a
> creative way.
I'd expect the above to work. As discussed in the other thread, one
case where it can't work is when it's not there. :) If you've found
some other way where it doesn't work as expected, let me know. (Of
course, there's also the separate possibility of node failure, manual
or DLM-initiated fencing, etc. but I'm sure you're familiar with all
that.)
>
> Thanks for any insight/guidance!
>
--
Ken Gaillot <kgaillot at redhat.com>
More information about the Users
mailing list