[Pacemaker] ocf:linbit:drbd monitor failed with stacked resource

Schmidt, Torsten torsten.schmidt at tecdoc.net
Mon Oct 19 06:32:25 EDT 2009


Hi list,

i configured pacemaker to take care of a three node drbd stetup but the monitor on the stacked drbd resouce failed and the resource never gets started.
Doing the necessary steps by hand everything works fine (on-node-with-ip: drbdadm --stacked adjust mysqlstack ; drbdadm --stacked primary mysqlstack)

is it possible that the drbd-monitor can't deal properly the a stacked-resource or do i miss something in my configuration?

ptest
-------------------------------------------------------------------------------
ptest[2395]: 2009/10/19_12:27:47 info: determine_online_status: Node mysqlHA2.cologne.tecdoc.local is online
ptest[2395]: 2009/10/19_12:27:47 info: unpack_rsc_op: res.drbd.mysqlstack:0_monitor_0 on mysqlHA2.cologne.tecdoc.local returned 6 (not configured) instead of the expected value: 7 (not running)
ptest[2395]: 2009/10/19_12:27:47 ERROR: unpack_rsc_op: Hard error - res.drbd.mysqlstack:0_monitor_0 failed with rc=6: Preventing ms.drbd.mysqlstack from re-starting anywhere in the cluster
ptest[2395]: 2009/10/19_12:27:47 info: unpack_rsc_op: res.ip.mysql_monitor_0 on mysqlHA2.cologne.tecdoc.local returned 0 (ok) instead of the expected value: 7 (not running)
ptest[2395]: 2009/10/19_12:27:47 notice: unpack_rsc_op: Operation res.ip.mysql_monitor_0 found resource res.ip.mysql active on mysqlHA2.cologne.tecdoc.local
ptest[2395]: 2009/10/19_12:27:47 info: determine_online_status: Node mysqlHA1.cologne.tecdoc.local is online
ptest[2395]: 2009/10/19_12:27:47 info: unpack_rsc_op: res.drbd.mysqlstack:0_monitor_0 on mysqlHA1.cologne.tecdoc.local returned 6 (not configured) instead of the expected value: 7 (not running)
ptest[2395]: 2009/10/19_12:27:47 ERROR: unpack_rsc_op: Hard error - res.drbd.mysqlstack:0_monitor_0 failed with rc=6: Preventing ms.drbd.mysqlstack from re-starting anywhere in the cluster
ptest[2395]: 2009/10/19_12:27:48 debug: native_assign_node: All nodes for resource res.drbd.mysqlstack:0 are unavailable, unclean or shutting down (mysqlHA2.cologne.tecdoc.local: 1, -1000000)
ptest[2395]: 2009/10/19_12:27:48 WARN: native_color: Resource res.drbd.mysqlstack:0 cannot run anywhere
-------------------------------------------------------------------------------


crm_mon
-------------------------------------------------------------------------------
============
Last updated: Mon Oct 19 12:10:48 2009
Stack: openais
Current DC: mysqlHA1.cologne.tecdoc.local - partition with quorum
Version: 1.0.5-462f1569a43740667daf7b0f6b521742e9eb8fa7
2 Nodes configured, 2 expected votes
3 Resources configured.
============

Online: [ mysqlHA2.cologne.tecdoc.local mysqlHA1.cologne.tecdoc.local ]

Master/Slave Set: ms.drbd.mysqldb
        Masters: [ mysqlHA2.cologne.tecdoc.local ]
        Slaves: [ mysqlHA1.cologne.tecdoc.local ]
res.ip.mysql    (ocf::heartbeat:IPaddr2):       Started mysqlHA2.cologne.tecdoc.local

Failed actions:
    res.drbd.mysqlstack:0_monitor_0 (node=(null), call=3, rc=6, status=complete): not configured
-------------------------------------------------------------------------------

cat /proc/drbd on node mysqlHA2:
-------------------------------------------------------------------------------
version: 8.3.2 (api:88/proto:86-90)
GIT-hash: dd7985327f146f33b86d4bff5ca8c94234ce840e build by mockbuild at v20z-x86-64.home.local, 2009-08-29 14:07:55
 0: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r----
    ns:4 nr:0 dw:0 dr:4 al:0 bm:1 lo:0 pe:0 ua:0 ap:0 ep:1 wo:b oos:0
-------------------------------------------------------------------------------


her my pacemaker config:
-------------------------------------------------------------------------------
node mysqlHA1.cologne.tecdoc.local
node mysqlHA2.cologne.tecdoc.local
primitive res.drbd.mysqldb ocf:linbit:drbd \
        params drbd_resource="mysqldb"
primitive res.drbd.mysqlstack ocf:linbit:drbd \
        params drbd_resource="mysqlstack" \
        meta target-role="Started"
primitive res.ip.mysql ocf:heartbeat:IPaddr2 \
        params ip="172.30.2.10" nic="eth0" netmask="255.255.255.0"
ms ms.drbd.mysqldb res.drbd.mysqldb \
        meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1" notify="true" globally_unique="false"
ms ms.drbd.mysqlstack res.drbd.mysqlstack \
        meta master-max="1" master-node-max="1" clone-max="1" clone-node-max="1" notify="true" globally_unique="false" t
colocation co.ms.drbd.mysqlstack_on_ms.drbd.mysqldb inf: ms.drbd.mysqlstack ms.drbd.mysqldb:Master
colocation co.res.ip.mysql_on_ms.drbd.mysqldb_master inf: res.ip.mysql ms.drbd.mysqldb:Master
order o.ip.mysql_before_ms.drbd.mysqlstack inf: res.ip.mysql ms.drbd.mysqlstack:start
order o.ms.drbd.mysqldb_before_ms.drbd.mysqlstack inf: ms.drbd.mysqldb:promote ms.drbd.mysqlstack:start
property $id="cib-bootstrap-options" \
        dc-version="1.0.5-462f1569a43740667daf7b0f6b521742e9eb8fa7" \
        cluster-infrastructure="openais" \
        expected-quorum-votes="2" \
        stonith-enabled="false" \
        last-lrm-refresh="1255946421"
-------------------------------------------------------------------------------


drbd.conf
-------------------------------------------------------------------------------
global {
  usage-count no;
}
common {
  syncer {
                rate 10M;
                verify-alg sha1;
        }
  protocol C;
}
resource mysqldb {
  handlers {
  fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
  after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh";
  split-brain   "/usr/lib/drbd/notify-split-brain.sh root";
  }
  startup {
    degr-wfc-timeout 60;    # 1 minute
    outdated-wfc-timeout 2;  # 2 seconds.
  }
  disk {
    on-io-error   detach;
  }
  net {
    cram-hmac-alg "sha1";
    shared-secret "supersecret";
    after-sb-0pri disconnect;
    after-sb-1pri disconnect;
    after-sb-2pri disconnect;
    rr-conflict disconnect;
  }
  syncer {
    rate 20M;
    al-extents 257;
  }
    device    /dev/drbd0;
    disk      /dev/sdb;
    meta-disk internal;
  on mysqlHA1.cologne.tecdoc.local {
    address    10.6.0.127:7788;
  }
  on mysqlHA2.cologne.tecdoc.local {
    address   10.6.0.128:7788;
  }
}
resource mysqlstack {
        protocol A;
        device          /dev/drbd10;
        stacked-on-top-of mysqldb {
                address         172.30.2.10:7789;  # Cluster IP
        }
        on mysqlHAoffsite.cologne.tecdoc.local {
                disk                    /dev/sdb;
                address         172.30.2.78:7789;  # public IP of backup-node
                meta-disk       internal;
        }
}
-------------------------------------------------------------------------------

Mit freundlichen Grüßen / with kind regards

Torsten Schmidt
System Manager

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.clusterlabs.org/pipermail/pacemaker/attachments/20091019/ab77658f/attachment.html>


More information about the Pacemaker mailing list