[ClusterLabs] Trying to understand the default action of a fence agent
Bryan K. Walton
bwalton+1546953805 at leepfrog.com
Thu Jan 10 14:41:13 EST 2019
On Tue, Jan 08, 2019 at 01:29:51PM -0600, Bryan K. Walton wrote:
> On Tue, Jan 08, 2019 at 10:55:09AM -0600, Ken Gaillot wrote:
> >
> > FYI pcmk_off_action="off" is the default
> >
> > If you want the cluster to request an "off" command instead of a
> > "reboot" when fencing a node, set the stonith-action cluster property
> > to "off".
>
> Awesome! Thank you, Ken. I don't know how I've missed this, up to now.
> Setting this property is exactly what I needed.
I swear I had this working the other day. I'm still struggling with
this, apparently. I've set the default stonith-action to off:
[root at storage1 ~]# pcs config | grep -i stonith-action
stonith-action: off
[root at storage1 ~]# pcs config | grep -i stonith-enabled
stonith-enabled: true
But when I run "pcs stonith fence storage2" (from my storage1 node, the
fabric ports are getting sucessfully disabled, and then re-enabled:
Here are the logs that show stonith-ng issuing "off" commands
(successfully), and then following up with "on" commands:
Jan 10 13:31:55 storage1 stonith-ng[43051]: notice: Client
stonith_admin.44835.f958d69c wants to fence (reboot) 'storage2' with
device '(any)'
Jan 10 13:31:55 storage1 stonith-ng[43051]: notice: Requesting peer
fencing (off) of storage2
Jan 10 13:31:55 storage1 stonith-ng[43051]: notice:
fenceStorage2-millipede can fence (reboot) storage2: static-list
Jan 10 13:31:55 storage1 stonith-ng[43051]: notice:
fenceStorage2-centipede can fence (reboot) storage2: static-list
Jan 10 13:31:56 storage1 stonith-ng[43051]: notice: Operation 'off'
[44836] (call 2 from stonith_admin.44835) for host 'storage2' with
device 'fenceStorage2-centipede' returned: 0 (OK)
Jan 10 13:31:56 storage1 stonith-ng[43051]: notice: Call to
fenceStorage2-centipede for 'storage2 off' on behalf of
stonith_admin.44835 at storage1: OK (0)
Jan 10 13:31:57 storage1 stonith-ng[43051]: notice: Operation 'off'
[44930] (call 2 from stonith_admin.44835) for host 'storage2' with
device 'fenceStorage2-millipede' returned: 0 (OK)
Jan 10 13:31:57 storage1 stonith-ng[43051]: notice: Call to
fenceStorage2-millipede for 'storage2 off' on behalf of
stonith_admin.44835 at storage1: OK (0)
Jan 10 13:31:58 storage1 stonith-ng[43051]: notice: Operation 'on'
[44936] (call 2 from stonith_admin.44835) for host 'storage2' with
device 'fenceStorage2-centipede' returned: 0 (OK)
Jan 10 13:31:58 storage1 stonith-ng[43051]: notice: Call to
fenceStorage2-centipede for 'storage2 on' on behalf of
stonith_admin.44835 at storage1: OK (0)
Jan 10 13:32:00 storage1 stonith-ng[43051]: notice: Operation 'on'
[44942] (call 2 from stonith_admin.44835) for host 'storage2' with
device 'fenceStorage2-millipede' returned: 0 (OK)
Jan 10 13:32:00 storage1 stonith-ng[43051]: notice: Call to
fenceStorage2-millipede for 'storage2 on' on behalf of
stonith_admin.44835 at storage1: OK (0)
Jan 10 13:32:00 storage1 stonith-ng[43051]: notice: Operation reboot of
storage2 by storage1 for stonith_admin.44835 at storage1.0b0f51e0: OK
Jan 10 13:32:00 storage1 crmd[43055]: notice: Peer storage2 was
terminated (reboot) by storage1 on behalf of stonith_admin.44835: OK
Any ideas what I'm doing wrong? I'd be happy to provide more mogs, if
desired.
Thanks!
Bryan
More information about the Users
mailing list