[ClusterLabs] Interface confusion

Tomas Jelinek tojeline at redhat.com
Fri Mar 15 10:48:10 EDT 2019


Dne 15. 03. 19 v 15:09 Adam Budziński napsal(a):
> Hello Tomas,
> 
> Thank you! So far I  need to say how great this community is, would 
> never expect so much positive vibes! A big thank you your doing a great 
> job!
> 
> Now let's talk business :)
> 
> So if pcsd is using ring0 and it fails will ring1 not be used at all?

Pcs and pcsd never use ring1, but they are just tools for managing 
clusters. You can have a perfectly functioning cluster without pcs and 
pcsd running or even installed, it would be just more complicated to set 
it up and manage it.

Even if ring0 fails, you will be able to use pcs (in somehow limited 
manner) as most of its commands don't go through network anyway.

Corosync, which is the actual cluster messaging layer, will of course 
use ring1 in case of ring0 failure.

> 
> So in regards to VMware that would mean that the interface should be 
> configured with a network that can access the  vCenter to fence right? 
> But wouldn't it then use only ring0 so if that fails it wouldn't switch 
> to ring1?

If you are talking about pcmk_host_map, that does not really have 
anything to do with network interfaces of cluster nodes. It maps node 
names (parts before :) to "ports" of a fence device (parts after :). 
Pcs-0.9.x does not support defining custom node names, therefore node 
names are the same as ring0 addresses.

I am not an expert on fence agents / devices, but I'm sure someone else 
on this list will be able to help you with configuring fencing for your 
cluster.


Tomas

> 
> Thank you!
> 
> pt., 15.03.2019, 13:14 użytkownik Tomas Jelinek <tojeline at redhat.com 
> <mailto:tojeline at redhat.com>> napisał:
> 
>     Dne 15. 03. 19 v 12:32 Adam Budziński napsal(a):
>      > Hello Folks,____
>      >
>      > __ __
>      >
>      > Tow node active/passive VMware VM cluster.____
>      >
>      > __ __
>      >
>      > /etc/hosts____
>      >
>      > __ __
>      >
>      > 10.116.63.83    srv1____
>      >
>      > 10.116.63.84    srv2____
>      >
>      > 172.16.21.12    srv2cr1____
>      >
>      > 172.16.22.12    srv2cr2____
>      >
>      > 172.16.21.11    srv1cr1____
>      >
>      > 172.16.22.11    srv1cr2____
>      >
>      > __ __
>      >
>      > __ __
>      >
>      > I have 3 NIC’s on each VM:____
>      >
>      > __ __
>      >
>      > 10.116.63.83    srv1  and 10.116.63.84    srv2 are networks used to
>      > access the VM’s via SSH or any resource directly if not via a
>     VIP.____
>      >
>      > __ __
>      >
>      > Everything with cr in its name is used for corosync
>     communication, so
>      > basically I have two rings (this are two no routable networks
>     just for
>      > that).____
>      >
>      > __ __
>      >
>      > My questions are:____
>      >
>      > __ __
>      >
>      > __1.__With ‘pcs cluster auth’ which interface / interfaces should
>     I use
>      > ?____
> 
>     Hi Adam,
> 
>     I can see you are using pcs-0.9.x. In that case you should do:
>     pcs cluster auth srv1cr1 srv2cr1
> 
>     In other words, use the first address of each node.
>     Authenticating all the other addresses should not cause any issues. It
>     is pointless, though, as pcs only communicates via ring0 addresses.
> 
>      >
>      > __2.__With ‘pcs cluster setup –name’ I would use the corosync
>     interfaces
>      > e.g. ‘pcs cluster setup –name MyCluster srv1cr1,srv1cr2
>     srv2cr1,srv2cr2’
>      > right ?____
> 
>     Yes, that is correct.
> 
>      >
>      > __3.__With fence_vmware_soap inpcmk_host_map="X:VM_C;X:VM:OTRS_D"
>     which
>      > interface should replace X ?____
> 
>     X should be replaced by node names as seen by pacemaker. Once you
>     set up
>     and start your cluster, run 'pcs status' to get (amongs other info) the
>     node names. In your configuration, they should be srv1cr1 and srv2cr1.
> 
> 
>     Regards,
>     Tomas
> 
>      > __ __
>      >
>      > Thank you!
>      >
>      >
>      > _______________________________________________
>      > Users mailing list: Users at clusterlabs.org
>     <mailto:Users at clusterlabs.org>
>      > https://lists.clusterlabs.org/mailman/listinfo/users
>      >
>      > Project Home: http://www.clusterlabs.org
>      > Getting started:
>     http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>      > Bugs: http://bugs.clusterlabs.org
>      >
>     _______________________________________________
>     Users mailing list: Users at clusterlabs.org <mailto:Users at clusterlabs.org>
>     https://lists.clusterlabs.org/mailman/listinfo/users
> 
>     Project Home: http://www.clusterlabs.org
>     Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>     Bugs: http://bugs.clusterlabs.org
> 
> 
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> https://lists.clusterlabs.org/mailman/listinfo/users
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
> 


More information about the Users mailing list