[Pacemaker] CoroSync's UDPu transport for public IP addresses?

Dmitry Koterov dmitry.koterov at gmail.com
Sun Dec 28 22:11:49 EST 2014


Hello.

I have a geographically distributed cluster, all machines have public IP
addresses. No virtual IP subnet exists, so no multicast is available.

I thought that UDPu transport can work in such environment, doesn't it?

To test everything in advance, I've set up a corosync+pacemaker on Ubuntu
14.04 with the following corosync.conf:

totem {
  transport: udpu
  interface {
        ringnumber: 0
        bindnetaddr: ip-address-of-the-current-machine
        mcastport: 5405
  }
}
nodelist {
  node {
    ring0_addr: node1
  }
  node {
    ring0_addr: node2
  }
}
...

(here node1 and node2 are hostnames from /etc/hosts on both machines).
After running "service corosync start; service pacemaker start" logs show
no problems, but actually both nodes are always offline:

root at node1:/etc/corosync# crm status | grep node
OFFLINE: [ node1 node2 ]

and "crm node online" (as all other attempts to make crm to do something)
are timed out with "communication error".

No iptables, selinux, apparmor and other bullshit are active: just pure
virtual machines with single public IP addresses on each. Also tcpdump
shows that UDB packets on port 5405 are going in and out, and if I e.g.
stop corosync at node1, the tcpdump output at node2 changes significantly.
So they see each other definitely.

And if I attach a gvpe adapter to these 2 machines with a private subnet
and switch transport to the default one, corosync + pacemaker begin to work.

So my question is: what am I doing wrong? Maybe UDPu is not suitable for
communications among machines with public IP addresses only?
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.clusterlabs.org/pipermail/pacemaker/attachments/20141229/b9800607/attachment-0002.html>


More information about the Pacemaker mailing list