[ClusterLabs] Antw: 2-Node Cluster, 2 Corosync Rings, Why Failover?
Ulrich Windl
Ulrich.Windl at rz.uni-regensburg.de
Mon Oct 10 06:19:16 UTC 2016
>>> Eric Robinson <eric.robinson at psmnv.com> schrieb am 09.10.2016 um 05:25 in
Nachricht
<DM5PR03MB2729771F3E4E94C4C49BF159FAD80 at DM5PR03MB2729.namprd03.prod.outlook.com>
> In a 2-node cluster where each node has two NICs connected to disjoint
> networks, and thus 2 corosync rings, why would loss of communication on one
> ring cause cluster failover?
>
> We have the following setup...
>
>
> LAN A
> /====SWITCH====SWITCH====\
> / \
>
> NODE_A NODE_B
>
> \ /
> \====SWITCH====SWITCH====/
> LAN B
>
>
> Everything on LAN A is physically separate from LAN B, different switches,
> cables, power, etc. For some reason, when either LAN A or LAN B suffers a
> failure, the cluster fails over. What would cause that?
Without logs that's hard to say. We are running a similar configuration (SLES11) without such problems.
>
> This happened yesterday at 2:05 pm Pacific time. I have the corosync and
> pacemaker logs from both nodes during that timeframe, but they are 20,000+
> lines. I can see the failover happening (because everything was going along
> normally, then the logs went nuts) but I don't understand why. Can someone
> tell me what clues I should be looking for?
>
> --
> Eric Robinson
>
>
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
More information about the Users
mailing list