[ClusterLabs] cluster doesn't do HA as expected, pingd doesn't help

Vladislav Bogdanov bubble at hoster-ok.com
Tue Dec 19 13:51:36 EST 2023


What if node (especially vm) freezes for several minutes and then continues 
to write to a shared disk where other nodes already put their data?
In my opinion, fencing, preferably two-level, is mandatory for lustre, 
trust me, I'd developed whole HA stack for both Exascaler and PangeaFS. 
We've seen so many points where data loss may occur...

On December 19, 2023 19:42:56 Artem <tyomikh at gmail.com> wrote:
> Andrei and Klaus thanks for prompt reply and clarification!
> As I understand, design and behavior of Pacemaker is tightly coupled with 
> the stonith concept. But isn't it too rigid?
>
> Is there a way to leverage self-monitoring or pingd rules to trigger 
> isolated node to umount its FS? Like vSphere High Availability host 
> isolation response.
> Can resource-stickiness=off (auto-failback) decrease risk of corruption by 
> unresponsive node coming back online?
> Is there a quorum feature not for cluster but for resource start/stop? Got 
> lock - is welcome to mount, unable to refresh lease - force unmount.
> Can on-fail=ignore break manual failover logic (stopped will be considered 
> as failed and thus ignored)?
>
> best regards,
> Artem
>
> On Tue, 19 Dec 2023 at 17:03, Klaus Wenninger <kwenning at redhat.com> wrote:
>
>
> On Tue, Dec 19, 2023 at 10:00 AM Andrei Borzenkov <arvidjaar at gmail.com> wrote:
> On Tue, Dec 19, 2023 at 10:41 AM Artem <tyomikh at gmail.com> wrote:
> ...
>> Dec 19 09:48:13 lustre-mds2.ntslab.ru pacemaker-schedulerd[785107] 
>> (update_resource_action_runnable)    warning: OST4_stop_0 on lustre4 is 
>> unrunnable (node is offline)
>> Dec 19 09:48:13 lustre-mds2.ntslab.ru pacemaker-schedulerd[785107] 
>> (recurring_op_for_active)    info: Start 20s-interval monitor for OST4 on 
>> lustre3
>> Dec 19 09:48:13 lustre-mds2.ntslab.ru pacemaker-schedulerd[785107] 
>> (log_list_item)      notice: Actions: Stop       OST4        (     lustre4 
>> )  blocked
>
> This is the default for the failed stop operation. The only way
> pacemaker can resolve failure to stop a resource is to fence the node
> where this resource was active. If it is not possible (and IIRC you
> refuse to use stonith), pacemaker has no other choice as to block it.
> If you insist, you can of course sert on-fail=ignore, but this means
> unreachable node will continue to run resources. Whether it can lead
> to some corruption in your case I cannot guess.
>
> Don't know if I'm reading that correctly but I understand what you had written
> above that you try to trigger the failover by stopping the VM (lustre4) without
> ordered shutdown.
> With fencing disabled what we are seeing is exactly what we would expect:
> The state of the resource is unknown - pacemaker tries to stop it - doesn't 
> work
> as the node is offline - no fencing configured - so everything it can do is 
> wait
> till there is info if the resource is up or not.
> I guess the strange output below is because of fencing disabled - quite an
> unusual - also not recommended - configuration and so this might not have
> shown up too often in that way.
>
> Klaus
>
>> Dec 19 09:48:13 lustre-mds2.ntslab.ru pacemaker-schedulerd[785107] 
>> (pcmk__create_graph)         crit: Cannot fence lustre4 because of OST4: 
>> blocked (OST4_stop_0)
>
> That is a rather strange phrase. The resource is blocked because the
> pacemaker could not fence the node, not the other way round.
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20231219/b4857639/attachment-0001.htm>


More information about the Users mailing list