[ClusterLabs] temporary loss of quorum when member starts to rejoin

Jan Friesse jfriesse at redhat.com
Wed Apr 8 02:56:05 EDT 2020


> On Tue, 7 Apr 2020 14:13:35 -0400
> Sherrard Burton <sb-clusterlabs at allafrica.com> wrote:
> 
>> On 4/7/20 1:16 PM, Andrei Borzenkov wrote:
>>> 07.04.2020 00:21, Sherrard Burton пишет:
>>>>>
>>>>> It looks like some timing issue or race condition. After reboot node
>>>>> manages to contact qnetd first, before connection to other node is
>>>>> established. Qnetd behaves as documented - it sees two equal size
>>>>> partitions and favors the partition that includes tie breaker (lowest
>>>>> node id). So existing node goes out of quorum. Second later both nodes
>>>>> see each other and so quorum is regained.
>>>>   
>>>
>>> Define the right problem to solve?
>>>
>>> Educated guess is that your problem is not corosync but pacemaker
>>> stopping resources. In this case just do what was done for years in two
>>> node cluster - set no-quorum-policy=ignore and rely on stonith to
>>> resolve split brain.
>>>
>>> I dropped idea to use qdevice in two node cluster. If you have reliable
>>> stonith device it is not needed and without stonith relying on watchdog
>>> suicide has too many problems.
>>>    
>>
>> Andrei,
>> in a two-node cluster with stonith only, but no qdevice, how do you
>> avoid the dreaded stonith death match, and the resultant flip-flopping
>> of services?
> 
> In my understanding, two_node and wait_for_all should avoid this.

Just a tiny comment. wait_for_all is enabled automatically when two_node 
mode is set (as long as wait_for_all is not explicitly disabled) so just 
setting two_node mode does the job.

> 
> After a node A has been fenced, the node B keeps the quorum thanks to two_node.
> When A comes back, as long as it is not able to join the corosync group, it will
> not be quorate thanks to wait_for_all. No quorum, no fencing allowed.
> 
> But the best protection is to disable pacemaker on boot so an admin can
> investigate the situation and join back the node safely.
> 
> Regards,
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
> 
> ClusterLabs home: https://www.clusterlabs.org/
> 



More information about the Users mailing list