[Pacemaker] [Problem]By the number of the resources, node do not succeed in resource movement.

renayama19661014 at ybb.ne.jp renayama19661014 at ybb.ne.jp
Wed Jun 22 02:51:23 EDT 2011


Hi all,

I tested the movement of the resource in the next procedure in environment like
PostgresSQL+drbd.

Step1) Start a cluster.
============
Last updated: Thu Jun 23 00:13:20 2011
Stack: Heartbeat
Current DC: srv02 (4247e5e4-b76c-4bcd-b81a-d971781fc802) - partition with
quorum
Version: 1.0.11-6e010d6b0d49a6b929d17c0114e9d2d934dc8e04
2 Nodes configured, unknown expected votes
6 Resources configured.
============

Online: [ srv01 srv02 ]

 Master/Slave Set: msGroup01
     Masters: [ srv01 ]
     Slaves: [ srv02 ]
 Resource Group: testGroup
     prmDummy1  (ocf::pacemaker:Dummy): Started srv01
 Clone Set: clnPingd
     Started: [ srv01 srv02 ]
 Clone Set: clnDiskd1
     Started: [ srv01 srv02 ]
 Resource Group: grpStonith1
     prmStonithN1       (stonith:external/ssh): Started srv02
 Resource Group: grpStonith2
     prmStonithN2       (stonith:external/ssh): Started srv01

Migration summary:
* Node srv02: 
* Node srv01: 

Step2) Move a group resource to slave node.

[root at srv01 ~]# crm_resource -M -r testGroup -H srv02 -f -Q

Step3) The movement succeeds. And the master node is replaced with slave node
definitely.

============
Last updated: Thu Jun 23 00:13:36 2011
Stack: Heartbeat
Current DC: srv02 (4247e5e4-b76c-4bcd-b81a-d971781fc802) - partition with
quorum
Version: 1.0.11-6e010d6b0d49a6b929d17c0114e9d2d934dc8e04
2 Nodes configured, unknown expected votes
6 Resources configured.
============

Online: [ srv01 srv02 ]

 Master/Slave Set: msGroup01
     Masters: [ srv02 ]
     Slaves: [ srv01 ]
 Resource Group: testGroup
     prmDummy1  (ocf::pacemaker:Dummy): Started srv02
 Clone Set: clnPingd
     Started: [ srv01 srv02 ]
 Clone Set: clnDiskd1
     Started: [ srv01 srv02 ]
 Resource Group: grpStonith1
     prmStonithN1       (stonith:external/ssh): Started srv02
 Resource Group: grpStonith2
     prmStonithN2       (stonith:external/ssh): Started srv01

Migration summary:
* Node srv02: 
* Node srv01: 


I increase the contents of the group resource to three next.
And the resource cannot move when I confirm it in the same procedure.

Step1) Start a cluster.

============
Last updated: Thu Jun 23 00:04:49 2011
Stack: Heartbeat
Current DC: srv02 (a105343e-c637-4d32-9a25-37d70439f669) - partition with
quorum
Version: 1.0.11-6e010d6b0d49a6b929d17c0114e9d2d934dc8e04
2 Nodes configured, unknown expected votes
6 Resources configured.
============

Online: [ srv01 srv02 ]

 Master/Slave Set: msGroup01
     Masters: [ srv01 ]
     Slaves: [ srv02 ]
 Resource Group: testGroup
     prmDummy1  (ocf::pacemaker:Dummy): Started srv01
     prmDummy2  (ocf::pacemaker:Dummy): Started srv01
     prmDummy3  (ocf::pacemaker:Dummy): Started srv01
 Clone Set: clnPingd
     Started: [ srv01 srv02 ]
 Clone Set: clnDiskd1
     Started: [ srv01 srv02 ]
 Resource Group: grpStonith1
     prmStonithN1       (stonith:external/ssh): Started srv02
 Resource Group: grpStonith2
     prmStonithN2       (stonith:external/ssh): Started srv01

Migration summary:
* Node srv02: 
* Node srv01: 


Step2) Move a group resource to slave node.

[root at srv01 ~]# crm_resource -M -r testGroup -H srv02 -f -Q
[root at srv01 ~]# crm_mon -1 -f 


Step3) The resource cannot move.


Last updated: Thu Jun 23 00:05:31 2011
Stack: Heartbeat
Current DC: srv02 (a105343e-c637-4d32-9a25-37d70439f669) - partition with
quorum
Version: 1.0.11-6e010d6b0d49a6b929d17c0114e9d2d934dc8e04
2 Nodes configured, unknown expected votes
6 Resources configured.
============

Online: [ srv01 srv02 ]

 Master/Slave Set: msGroup01
     Masters: [ srv01 ]
     Slaves: [ srv02 ]
 Clone Set: clnPingd
     Started: [ srv01 srv02 ]
 Clone Set: clnDiskd1
     Started: [ srv01 srv02 ]
 Resource Group: grpStonith1
     prmStonithN1       (stonith:external/ssh): Started srv02
 Resource Group: grpStonith2
     prmStonithN2       (stonith:external/ssh): Started srv01

Migration summary:
* Node srv02: 
* Node srv01: 

It is a problem not to move under the influence of the number of resources.


This problem does not happen with 1.0.10 of Pacemaker.
And the hit of 1.0.11 next correction of Pacemaker seems to influence it.
 * http://hg.clusterlabs.org/pacemaker/stable-1.0/rev/e6dd1746027f

I registered these contents and hb_report with Bugzilla.
 * http://developerbugs.linux-foundation.org/show_bug.cgi?id=2608

Best Regards,
Hideo Yamauchi.




More information about the Pacemaker mailing list