[Pacemaker] cluster doesn't failover

Dejan Muhamedagic dejanmm at fastmail.fm
Fri Oct 30 14:12:37 UTC 2009


Hi,

On Fri, 30 Oct 2009 14:36 +0100, "Thomas Schneider"
<thomas.schneider at euskill.com> wrote:
> Hello,
> 
> I'm trying to setup a cluster for shared storage with pacemaker, drbd,
> and
> nfs.
> I use two server and there is two network interface one each server, eth0
> connected to the network, and eth1 direct link betwene the two node for
> drbd
> replication,
> 
> 
>                  Eth1: replication link
>  --------------      ----------------
>  | Server 1   |------| Server 2     |
>  --------------      ----------------
>       |                    |
>      Eth0                 eth0
>       |                    |  
>          External Network
> IPs
> - server1 (storage01): eth0:10.1.1.70 eth1:192.168.0.2
> - server2 (storage02): eth0:10.1.1.71 eth1:192.168.0.3
> 
> I can migrate the resource between the to node, but the problem is when i
> make a hard power-off of server01 ( the server where is running the
> resource) the second server (server02) doesn't make failover. (the
> resource
> doesn't start)
> May you can take a look to my config file in the following

Really hard to say without logs. Otherwise, this is probably an issue
somewhere in the drbd setup. I guess that it's better to talk to the
drbd list.

Thanks,

Dejan
> 
> Thanks for your help
> 
> Thomas Schneider
> 
> Drbd.conf
> ------------------------------------------------------------------------
> global {
>         usage-count no;
> }
> 
> common {
>         syncer { rate 35M; }
> }
> 
> resource nfs {
>         protocol C;
>         handlers {
>                 fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
>                 after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh";
>         }
>         startup {
>                 degr-wfc-timeout 120;    # 2 minutes.
>         }
>         disk {
>                 on-io-error   detach;
>                 fencing resource-only;
>         }
>         net {
>                 max-buffers     2048;
>                 max-epoch-size  2048;
>                 }
>         syncer {
>                 rate 35M;
>                 al-extents 257;
>         }
>         on storage01.myriapulse.local {
>                 device     /dev/drbd0;
>                 disk       /dev/sda4;
>                 address    192.168.0.2:7788;
>                 meta-disk  /dev/sda3[0];
>         }
>         on storage02.myriapulse.local {
>                 device    /dev/drbd0;
>                 disk      /dev/sda4;
>                 address   192.168.0.3:7788;
>                 meta-disk /dev/sda3[0];
>         }
> }
> ----------------------------------------------------------------------------
> ----------
> 
> storage01:~# crm configure show
> node storage01.myriapulse.local \
>         attributes standby="off"
> node storage02.myriapulse.local \
>         attributes standby="off"
> primitive drbd_nfs ocf:linbit:drbd \
>         params drbd_resource="nfs" \
>         op monitor interval="15s" \
>         meta target-role="Started"
> primitive fs_nfs ocf:heartbeat:Filesystem \
>         params device="/dev/drbd/by-res/nfs" directory="/share"
> fstype="ext3" \
>         meta is-managed="true"
> primitive ftp-server lsb:proftpd \
>         op monitor interval="1min"
> primitive ip_nfs ocf:heartbeat:IPaddr2 \
>         params ip="10.1.1.69" nic="eth0"
> primitive nfs-kernel-server lsb:nfs-kernel-server \
>         op monitor interval="1min"
> group nfs fs_nfs ip_nfs nfs-kernel-server \
>         meta target-role="Started"
> ms ms_drbd_nfs drbd_nfs \
>         meta master-max="1" master-node-max="1" clone-max="2"
> clone-node-max="1" notify="true"
> location cli-standby-nfs nfs \
>         rule $id="cli-standby-rule-nfs" -inf: #uname eq
> storage02.myriapulse.local
> colocation ftp_on_nfs inf: ftp-server nfs
> colocation nfs_on_drbd inf: nfs ms_drbd_nfs:Master
> order ftp_after_nfs inf: nfs ftp-server
> order nfs_after_drbd inf: ms_drbd_nfs:promote nfs:start
> property $id="cib-bootstrap-options" \
>         dc-version="1.0.5-unknown" \
>         cluster-infrastructure="openais" \
>         expected-quorum-votes="2" \
>         no-quorum-policy="ignore" \
>         stonith-enabled="false"
> storage01:~#
> ----------------------------------------------------------------------
> Openais.conf
> 
> storage02:~# cat /etc/openais/openais.conf
> # Please read the openais.conf.5 manual page
> 
> totem {
>         version: 2
> 
>         # How long before declaring a token lost (ms)
>         token: 3000
> 
>         # How many token retransmits before forming a new configuration
>         token_retransmits_before_loss_const: 10
> 
>         # How long to wait for join messages in the membership protocol
>         (ms)
>         join: 60
> 
>         # How long to wait for consensus to be achieved before starting a
> new round of membership configuration (ms)
>         consensus: 1500
> 
>         # Turn off the virtual synchrony filter
>         vsftype: none
> 
>         # Number of messages that may be sent by one processor on receipt
>         of
> the token
>         max_messages: 20
> 
>         # Limit generated nodeids to 31-bits (positive signed integers)
>         clear_node_high_bit: yes
> 
>         # Disable encryption
>         secauth: on
> 
>         # How many threads to use for encryption/decryption
>         threads: 0
> 
>         # Optionally assign a fixed node id (integer)
>         # nodeid: 1234
> 
>         # This specifies the mode of redundant ring, which may be none,
> active, or passive.
>         rrp_mode: passive
> 
>         interface {
>                 # The following values need to be set based on your
> environment
>                 ringnumber: 0
>                 bindnetaddr: 10.1.1.0
>                 mcastaddr: 226.94.1.1
>                 mcastport: 5405
>         }
>         interface {
>                 # The following values need to be set based on your
> environment
>                 ringnumber: 1
>                 bindnetaddr: 192.168.0.0
>                 mcastaddr: 226.94.2.1
>                 mcastport: 5405
>         }
> }
> 
> logging {
>         to_stderr: yes
>         debug: off
>         timestamp: on
>         to_file: /var/log/openais/openais.log
>         to_syslog: yes
>         syslog_facility: daemon
> }
> 
> amf {
>         mode: disabled
> }
> 
> service {
>         # Load the Pacemaker Cluster Resource Manager
>         ver:       0
>         name:      pacemaker
>         use_mgmtd: yes
> }
> 
> aisexec {
>         user:   root
>         group:  root
> }
> 
> 
> _______________________________________________
> Pacemaker mailing list
> Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker

-- 
http://www.fastmail.fm - Access your email from home and the web





More information about the Pacemaker mailing list