[ClusterLabs] Antw: [EXT] Re: Our 2-Node Cluster with a Separate Qdevice Went Down Anyway?

Ulrich Windl Ulrich.Windl at rz.uni-regensburg.de
Mon Mar 1 02:10:11 EST 2021


>>> Digimer <lists at alteeve.ca> schrieb am 26.02.2021 um 17:34 in Nachricht
<699432c7-89a6-41bf-c805-f4a7a0a4a8cd at alteeve.ca>:
> On 2021‑02‑26 11:19 a.m., Eric Robinson wrote:
>> At 5:16 am Pacific time Monday, one of our cluster nodes failed and its
>> mysql services went down. The cluster did not automatically recover.
>> 
>> We’re trying to figure out:
>> 
>>  1. Why did it fail?
>>  2. Why did it not automatically recover?
>> 
>> The cluster did not recover until we manually executed…
>> 
>> # pcs resource cleanup p_mysql_622
>> 
>> OS: CentOS Linux release 7.5.1804 (Core)
>> 
>> Cluster version:
>> 
>> corosync.x86_64                  2.4.5‑4.el7                     @base
>> corosync‑qdevice.x86_64          2.4.5‑4.el7                     @base
>> pacemaker.x86_64                 1.1.21‑4.el7                    @base
>> 
>> Two nodes: 001db01a, 001db01b
>> 
>> The following log snippet is from node 001db01a:
>> 
>> [root at 001db01a cluster]# grep "Feb 22 05:1[67]" corosync.log‑20210223
> 
> <snip>
> 
>> Feb 22 05:16:30 [91682] 001db01a    pengine:  warning: cluster_status:  
> Fencing and resource management disabled due to lack of quorum
> 
> Seems like there was no quorum from this node's perspective, so it won't
> do anything. What does the other node's logs say?

@Digimer: The other node's log was included ;-)

Regards,
Ulrich





More information about the Users mailing list