[Pacemaker] cibadmin set node offline?

James Gibbard thisbodydrop at gmail.com
Wed Aug 6 13:22:40 UTC 2014


Hi,

I have setup a 2 node cluster, using the following packages:

pacemaker                           1.1.10+git20130802-1ubuntu2
corosync                            2.3.3-1ubuntu1

My cluster config is as so:

node $id="12303" ldb03
node $id="12304" ldb04
primitive p_fence_ldb03 stonith:external/vcenter \
        params VI_SERVER="10.17.248.10"
VI_CREDSTORE="/root/.vmware/credstore/vicredentials.xml"
HOSTLIST="ldb03=ldb03" RESETPOWERON="0" pcmk_host_check="static-list"
pcmk_host_list="ldb03" \
        op start interval="0" timeout="500s"
primitive p_fence_ldb04 stonith:external/vcenter \
        params VI_SERVER="10.17.248.10"
VI_CREDSTORE="/root/.vmware/credstore/vicredentials.xml"
HOSTLIST="ldb04=ldb04" RESETPOWERON="0" pcmk_host_check="static-list"
pcmk_host_list="ldb04" \
        op start interval="0" timeout="500s"
primitive p_fs_mysql ocf:heartbeat:Filesystem \
        params device="nfsserver:/LDB_Cluster1" directory="/var/lib/mysql"
fstype="nfs"
options="relatime,rw,hard,nointr,rsize=32768,wsize=32768,bg,vers=3,proto=tcp"
\
        op start interval="0" timeout="60s" \
        op stop interval="0" timeout="120s" \
        op monitor interval="60s" timeout="60s" \
        meta is-managed="true"
primitive p_ip_1 ocf:heartbeat:IPaddr2 \
        params ip="10.10.10.11" cidr_netmask="25" \
        op monitor interval="30s" \
        meta target-role="Started" is-managed="true"
primitive p_ip_2 ocf:heartbeat:IPaddr2 \
        params ip="10.10.10.12" cidr_netmask="25" \
        op monitor interval="30s" \
        meta target-role="Started" is-managed="true"
primitive p_ip_3 ocf:heartbeat:IPaddr2 \
        params ip="10.10.10.13" cidr_netmask="25" \
        op monitor interval="30s" \
        meta target-role="Started" is-managed="true"
primitive p_mysql ocf:heartbeat:mysql \
        params datadir="/var/lib/mysql" binary="/usr/bin/mysqld_safe"
socket="/var/run/mysqld/mysqld.sock" \
        op start interval="0" timeout="120" \
        op stop interval="0" timeout="120" \
        op monitor interval="20" timeout="30" \
        meta target-role="Started" is-managed="true"
group g_mysql p_fs_mysql p_mysql p_ip_1 p_ip_2 p_ip_3 \
location l_fence_ldb03 p_fence_ldb03 -inf: ldb03
location l_fence_ldb04 p_fence_ldb04 -inf: ldb04
property $id="cib-bootstrap-options" \
        dc-version="1.1.10-42f2063" \
        cluster-infrastructure="corosync" \
        no-quorum-policy="ignore" \
        stonith-enabled="true" \
        stop-all-resources="false" \
        expected-quorum-votes="2" \
        last-lrm-refresh="1407325251"


This exact configuration has worked during the setup, but I have
encountered a problem with my inactive node ldb03. Corosync shows this node
as up:

root at ldb03:~# corosync-cmapctl | grep members
runtime.totem.pg.mrp.srp.members.12303.config_version (u64) = 0
runtime.totem.pg.mrp.srp.members.12303.ip (str) = r(0) ip(10.10.10.8)
runtime.totem.pg.mrp.srp.members.12303.join_count (u32) = 1
runtime.totem.pg.mrp.srp.members.12303.status (str) = joined
runtime.totem.pg.mrp.srp.members.12304.config_version (u64) = 0
runtime.totem.pg.mrp.srp.members.12304.ip (str) = r(0) ip(10.10.10.9)
runtime.totem.pg.mrp.srp.members.12304.join_count (u32) = 1
runtime.totem.pg.mrp.srp.members.12304.status (str) = joined

and crm status and crm node status show it as online:

Last updated: Wed Aug  6 14:16:24 2014
Last change: Wed Aug  6 14:02:00 2014 via crm_resource on ldb04
Stack: corosync
Current DC: ldb04 (12304) - partition with quorum
Version: 1.1.10-42f2063
2 Nodes configured
7 Resources configured
Online: [ ldb03 ldb04 ]

root at ldb03:~# crm node status
<nodes>
  <node id="12304" uname="ldb04"/>
  <node id="12303" uname="ldb03"/>
</nodes>


but....after seeing this entry in my logs:
Aug  6 13:26:23 ldb03 cibadmin[2140]:   notice: crm_log_args: Invoked:
cibadmin -M -c -o status --xml-text <node_state id="ldb03" uname="ldb03"
ha="active" in_ccm="false" crmd="offline" join="member" expected="down"
crm-debug-origin="manual_clear" shutdown="0"/>

I noticed that cibadmin shows it as normal(offline)
root at ldb03:~# crm node show
ldb04(12304): normal
ldb03(12303): normal(offline)

The offline state is not present in anything but cibadmin. Not the cib.xml,
not corosync-quorumtool and a tcpdump shows multicast traffic from both
hosts.

I tried (hesitantly) to delete the line using cibadmin, but I couldn't
quite get the syntax right. Any tips on how to get this node to show as
online and subsequently be able to run resources? Currently, when I run crm
resource move, this has no effect, no errors and nothing noticeable in the
logfiles either.

Sorry for long thread....I can attach more logs/config if necessary.

Thanks,

Jamie.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20140806/e64891af/attachment-0003.html>


More information about the Pacemaker mailing list