[Pacemaker] Resources not migrating on node failure?

Anton Altaparmakov aia21 at cam.ac.uk
Tue Nov 30 18:11:48 UTC 2010


Hi,

I have set up a three node cluster (running Ubuntu 10.04 LTS server with Corosync 1.2.0, Pacemaker 1.0.8, drbd 8.3.7), where one node is only present to provide quorum to the other two nodes in case one node fails but it itself cannot run any resources.  The other two nodes are running drbd in master/slave to provide replicated storage, then XFS file system on top of the drbd storage on the master, together with an NFS server on top of the XFS mount, and a service IP address on which the NFS export is shared.  This is all working brilliantly and I can cause the resources to move to the slave node by running "crm_standby -U cerberus -v on" where cerberus is the master node and everything then migrates to the slave node "minotaur".

My problem is if I pull the power plug on the master node "cerberus".  Then nothing happens!  minotaur continues to run as slave and it never takes over.

And I don't get why.  )-:

Also, a second question, possibly related to the first problem, is do I need to define monitor actions for each resource or is that done automatically?  If I need to do it specifically, how do I do that now that I have it all up and running without defining monitor actions?

My current configuration is shown below...

If anyone can give me any hints, that would be very much appreciated!  I have read the FAQ and all documentation I could find to no avail...

The configuration is:

# crm configure show
node cerberus \
	attributes standby="off"
node minotaur \
	attributes standby="off"
node yoda \
	attributes standby="off"
primitive res_drbd ocf:linbit:drbd \
	params drbd_resource="vmstore"
primitive res_filesystem ocf:heartbeat:Filesystem \
	params fstype="xfs" device="/dev/drbd0" directory="/vmstore" options="noatime,barrier,largeio,logbufs=8,logbsize=256k,swalloc" \
	meta target-role="Started"
primitive res_ip ocf:heartbeat:IPaddr2 \
	params ip="172.28.208.50" cidr_netmask="24" broadcast="172.28.208.255"
primitive res_nfs_server lsb:nfs-kernel-server
group group_vmstore res_filesystem res_nfs_server res_ip
ms ms_drbd res_drbd \
	meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1" notify="true" globally_unique="false"
location loc_drbd_cerberus ms_drbd 0: cerberus
location loc_drbd_minotaur ms_drbd 0: minotaur
location loc_vmstore_cerberus group_vmstore 0: cerberus
location loc_vmstore_minotaur group_vmstore 0: minotaur
colocation col_vmstore inf: group_vmstore ms_drbd:Master
order order_vmstore inf: ms_drbd:promote group_vmstore:start
property $id="cib-bootstrap-options" \
	expected-quorum-votes="3" \
	stonith-enabled="false" \
	symmetric-cluster="false" \
	pe-warn-series-max="100" \
	default-action-timeout="10" \
	pe-input-series-max="100" \
	dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \
	pe-error-series-max="100" \
	cluster-infrastructure="openais" \
	last-lrm-refresh="1290511853" \
	default-resource-stickiness="1000"
rsc_defaults $id="rsc-options" \
	target-role="started"

Best regards,

	Anton
-- 
Anton Altaparmakov <aia21 at cam.ac.uk> (replace at with @)
Unix Support, Computing Service, University of Cambridge, CB2 3QH, UK
Linux NTFS maintainer, http://www.linux-ntfs.org/





More information about the Pacemaker mailing list