[Pacemaker] Question about cluster start-up in a 2 node cluster with a node offline.

James FLatten jflatten at iso-ne.com
Mon Feb 13 10:54:07 EST 2012


On 02/12/2012 04:55 PM, Andreas Kurz wrote:
> 	op monitor role="Master" interval="30s"
> 	op monitor role="Slave"  interval="31s"
> ipmi fencing device capable of
> fencing more than one node?
Andreas-

I applied both changes you mentioned and the behavior still exists.  
Here is my current configuration:

    node nodea \
         attributes standby="off"
    node nodeb \
         attributes standby="off"
    primitive ClusterIP ocf:heartbeat:IPaddr2 \
         params ip="192.168.1.3" cidr_netmask="32" \
         op monitor interval="30s"
    primitive datafs ocf:heartbeat:Filesystem \
         params device="/dev/drbd0" directory="/data" fstype="ext3" \
         meta target-role="Started"
    primitive drbd0 ocf:linbit:drbd \
         params drbd_resource="drbd0" \
         op monitor interval="31s" role="Slave" \
         op monitor interval="30s" role="Master"
    primitive drbd1 ocf:linbit:drbd \
         params drbd_resource="drbd1" \
         op monitor interval="31s" role="Slave" \
         op monitor interval="30s" role="Master"
    primitive fence-nodea stonith:fence_ipmilan \
         params pcmk_host_list="nodeb" ipaddr="xxx.xxx.xxx.xxx"
    login="xxxxxxx" passwd="xxxxxxxx" lanplus="1" timeout="4" auth="md5" \
         op monitor interval="60s"
    primitive fence-nodeb stonith:fence_ipmilan \
         params pcmk_host_list="nodea" ipaddr="xxx.xxx.xxx.xxx"
    login="xxxxxxx" passwd="xxxxxxxx" lanplus="1" timeout="4" auth="md5" \
         op monitor interval="60s"
    primitive httpd ocf:heartbeat:apache \
         params configfile="/etc/httpd/conf/httpd.conf" \
         op monitor interval="1min"
    primitive patchfs ocf:heartbeat:Filesystem \
         params device="/dev/drbd1" directory="/patch" fstype="ext3" \
         meta target-role="Started"
    group web datafs patchfs ClusterIP httpd
    ms drbd0clone drbd0 \
         meta master-max="1" master-node-max="1" clone-max="2"
    clone-node-max="1" notify="true" target-role="Master"
    ms drbd1clone drbd1 \
         meta master-max="1" master-node-max="1" clone-max="2"
    clone-node-max="1" notify="true" target-role="Master"
    location fence-on-nodea fence-nodea \
         rule $id="fence-on-nodea-rule" -inf: #uname ne nodea
    location fence-on-nodeb fence-nodeb \
         rule $id="fence-on-nodeb-rule" -inf: #uname ne nodeb
    colocation datafs-with-drbd0 inf: web drbd0clone:Master
    colocation patchfs-with-drbd1 inf: web drbd1clone:Master
    order datafs-after-drbd0 inf: drbd0clone:promote web:start
    order patchfs-after-drbd1 inf: drbd1clone:promote web:start
    property $id="cib-bootstrap-options" \
         dc-version="1.1.6-3.el6-a02c0f19a00c1eb2527ad38f146ebc0834814558" \
         cluster-infrastructure="openais" \
         expected-quorum-votes="2" \
         stonith-enabled="false" \
         no-quorum-policy="ignore" \
         last-lrm-refresh="1328556424"
    rsc_defaults $id="rsc-options" \
         resource-stickiness="100"

If the cluster is fully down, I start corosync and pacemaker on one 
node, the cluster fences the other node but the services do not come up 
until the cluster-recheck-interval occurs.  I have attached the 
corosync.log from this latest test.

-Davin
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20120213/1796ed8e/attachment-0003.html>
-------------- next part --------------
An embedded and charset-unspecified text was scrubbed...
Name: corosync.log.problem.1
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20120213/1796ed8e/attachment-0003.ksh>


More information about the Pacemaker mailing list