[ClusterLabs] Antw: 2node cluster question

Ulrich Windl Ulrich.Windl at rz.uni-regensburg.de
Thu Aug 16 06:27:11 UTC 2018


>>> "Stefan K" <Shadow_7 at gmx.net> schrieb am 15.08.2018 um 10:58 in Nachricht
<trinity-80166f37-3ed1-474e-9df3-fcdff91a302c-1534323500296 at 3c-app-gmx-bs24>:
> Hello,
> 
> what is the 'best' 2‑node cluster config?
> What I want, if it run on nodeA and nodeA goes in standby or shut down, 
> everything must start at nodeB, if nodeA comes back, everything must still 
> run on nodeB.

Hi!

This is almost standard; you just have to add some default stickiness to
prevent "Load balancing" after the second node gets online again.

Regards,
Ulrich

> 
> pacemaker looks like:
>         have‑watchdog=false \
>         dc‑version=1.1.16‑94ff4df \
>         cluster‑infrastructure=corosync \
>         cluster‑name=zfs‑vmstorage \
>         no‑quorum‑policy=stop \
>         stonith‑enabled=true \
>         last‑lrm‑refresh=1528814481
> rsc_defaults rsc_defaults‑options: \
>         resource‑stickiness=100
> 
> and the corosync.config:
> totem {
>     version: 2
>     secauth: off
>     cluster_name: zfs‑vmstorage
>     transport: udpu
>     rrp_mode: passive
> }
> 
> nodelist {
>     node {
>         ring0_addr: zfs‑serv3
>         ring1_addr: 192.168.251.1
>         nodeid: 1
>     }
> 
>     node {
>         ring0_addr: zfs‑serv4
>         ring1_addr: 192.168.251.2
>         nodeid: 2
>     }
> }
> 
> quorum {
>     provider: corosync_votequorum
>     two_node: 1
> }
> 
> logging {
>     to_logfile: yes
>     logfile: /var/log/corosync/corosync.log
>     to_syslog: yes
> }
> 
> thanks in advance
> and best regards
> Stefan
> _______________________________________________
> Users mailing list: Users at clusterlabs.org 
> https://lists.clusterlabs.org/mailman/listinfo/users 
> 
> Project Home: http://www.clusterlabs.org 
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf 
> Bugs: http://bugs.clusterlabs.org 





More information about the Users mailing list