[ClusterLabs] group resources not grouped ?!?

zulucloud zulucloud at mailbox.org
Wed Oct 7 14:12:30 UTC 2015


Hi,
i got a problem i don't understand, maybe someone can give me a hint.

My 2-node cluster (named ali and baba) is configured to run mysql, an IP 
for mysql and the filesystem resource (on drbd master) together as a 
GROUP. After doing some crash-tests i ended up having filesystem and 
mysql running happily on one host (ali), and the related IP on the other 
(baba) .... although, the IP's not really up and running, crm_mon just 
SHOWS it as started there. In fact it's nowhere up, neither on ali nor 
on baba.

crm_mon shows that pacemaker tried to start it on baba, but gave up 
after fail-count=1000000.

Q1: why doesn't pacemaker put the IP on ali, where all the rest of it's 
group lives?
Q2: why doesn't pacemaker try to start the IP on ali, after max 
failcount had been reached on baba?
Q3: why is crm_mon showing the IP as "started", when it's down after 
100000 tries?

Thanks :)


config (some parts removed):
-------------------------------
node ali
node baba

primitive res_drbd ocf:linbit:drbd \
	params drbd_resource="r0" \
	op stop interval="0" timeout="100" \
	op start interval="0" timeout="240" \
	op promote interval="0" timeout="90" \
	op demote interval="0" timeout="90" \
	op notify interval="0" timeout="90" \
	op monitor interval="40" role="Slave" timeout="20" \
	op monitor interval="20" role="Master" timeout="20"
primitive res_fs ocf:heartbeat:Filesystem \
	params device="/dev/drbd0" directory="/drbd_mnt" fstype="ext4" \
	op monitor interval="30s"
primitive res_hamysql_ip ocf:heartbeat:IPaddr2 \
	params ip="XXX.XXX.XXX.224" nic="eth0" cidr_netmask="23" \
	op monitor interval="10s" timeout="20s" depth="0"
primitive res_mysql lsb:mysql \
	op start interval="0" timeout="15" \
	op stop interval="0" timeout="15" \
	op monitor start-delay="30" interval="15" time-out="15"

group gr_mysqlgroup res_fs res_mysql res_hamysql_ip \
	meta target-role="Started"
ms ms_drbd res_drbd \
	meta master-max="1" master-node-max="1" clone-max="2" 
clone-node-max="1" notify="true"

colocation col_fs_on_drbd_master inf: res_fs:Started ms_drbd:Master

order ord_drbd_master_then_fs inf: ms_drbd:promote res_fs:start

property $id="cib-bootstrap-options" \
	dc-version="1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b" \
	cluster-infrastructure="openais" \
	stonith-enabled="false" \
	no-quorum-policy="ignore" \
	expected-quorum-votes="2" \
	last-lrm-refresh="1438857246"


crm_mon -rnf (some parts removed):
---------------------------------
Node ali: online
         res_fs  (ocf::heartbeat:Filesystem) Started
         res_mysql       (lsb:mysql) Started
         res_drbd:0      (ocf::linbit:drbd) Master
Node baba: online
         res_hamysql_ip  (ocf::heartbeat:IPaddr2) Started
         res_drbd:1      (ocf::linbit:drbd) Slave

Inactive resources:

Migration summary:

* Node baba:
    res_hamysql_ip: migration-threshold=1000000 fail-count=1000000

Failed actions:
     res_hamysql_ip_stop_0 (node=a891vl107s, call=35, rc=1, 
status=complete): unknown error

corosync.log:
--------------
pengine: [1223]: WARN: should_dump_input: Ignoring requirement that 
res_hamysql_ip_stop_0 comeplete before gr_mysqlgroup_stopped_0: 
unmanaged failed resources cannot prevent shutdown

pengine: [1223]: WARN: should_dump_input: Ignoring requirement that 
res_hamysql_ip_stop_0 comeplete before gr_mysqlgroup_stopped_0: 
unmanaged failed resources cannot prevent shutdown

Software:
----------
corosync 1.2.1-4
pacemaker 1.0.9.1+hg15626-1
drbd8-utils 2:8.3.7-2.1
(for some reason it's not possible to update at this time)






More information about the Users mailing list