[Pacemaker] About replacement of clone and handling of the fail number of times.

Andrew Beekhof andrew at beekhof.net
Tue Mar 23 12:49:42 EDT 2010


2010/3/12  <renayama19661014 at ybb.ne.jp>:
> Hi,
>
> We tested the trouble of the clone.
>
> I confirmed it in the next procedure.
>
> Step1)I start all nodes and update cib.xml.
>
> ============
> Last updated: Fri Mar 12 14:53:38 2010
> Stack: openais
> Current DC: srv01 - partition with quorum
> Version: 1.0.7-049006f172774f407e165ec82f7ee09cb73fd0e7
> 4 Nodes configured, 2 expected votes
> 13 Resources configured.
> ============
>
> Online: [ srv01 srv02 srv03 srv04 ]
>
>  Resource Group: UMgroup01
>     UmVIPcheck (ocf::heartbeat:Dummy): Started srv01
>     UmIPaddr   (ocf::heartbeat:Dummy): Started srv01
>     UmDummy01  (ocf::heartbeat:Dummy): Started srv01
>     UmDummy02  (ocf::heartbeat:Dummy): Started srv01
>  Resource Group: OVDBgroup02-1
>     prmExPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
>     prmFsPostgreSQLDB1-1       (ocf::heartbeat:Dummy): Started srv01
>     prmFsPostgreSQLDB1-2       (ocf::heartbeat:Dummy): Started srv01
>     prmFsPostgreSQLDB1-3       (ocf::heartbeat:Dummy): Started srv01
>     prmIpPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
>     prmApPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
>  Resource Group: OVDBgroup02-2
>     prmExPostgreSQLDB2 (ocf::heartbeat:Dummy): Started srv02
>     prmFsPostgreSQLDB2-1       (ocf::heartbeat:Dummy): Started srv02
>     prmFsPostgreSQLDB2-2       (ocf::heartbeat:Dummy): Started srv02
>     prmFsPostgreSQLDB2-3       (ocf::heartbeat:Dummy): Started srv02
>     prmIpPostgreSQLDB2 (ocf::heartbeat:Dummy): Started srv02
>     prmApPostgreSQLDB2 (ocf::heartbeat:Dummy): Started srv02
>  Resource Group: OVDBgroup02-3
>     prmExPostgreSQLDB3 (ocf::heartbeat:Dummy): Started srv03
>     prmFsPostgreSQLDB3-1       (ocf::heartbeat:Dummy): Started srv03
>     prmFsPostgreSQLDB3-2       (ocf::heartbeat:Dummy): Started srv03
>     prmFsPostgreSQLDB3-3       (ocf::heartbeat:Dummy): Started srv03
>     prmIpPostgreSQLDB3 (ocf::heartbeat:Dummy): Started srv03
>     prmApPostgreSQLDB3 (ocf::heartbeat:Dummy): Started srv03
>  Resource Group: grpStonith1
>     prmStonithN1       (stonith:external/ssh): Started srv04
>  Resource Group: grpStonith2
>     prmStonithN2       (stonith:external/ssh): Started srv01
>  Resource Group: grpStonith3
>     prmStonithN3       (stonith:external/ssh): Started srv02
>  Resource Group: grpStonith4
>     prmStonithN4       (stonith:external/ssh): Started srv03
>  Clone Set: clnUMgroup01
>     Started: [ srv01 srv04 ]
>  Clone Set: clnPingd
>     Started: [ srv01 srv02 srv03 srv04 ]
>  Clone Set: clnDiskd1
>     Started: [ srv01 srv02 srv03 srv04 ]
>  Clone Set: clnG3dummy1
>     Started: [ srv01 srv02 srv03 srv04 ]
>  Clone Set: clnG3dummy2
>     Started: [ srv01 srv02 srv03 srv04 ]
>
> Step2)I generate the trouble of the clnUMgroup01 clone in a N1(srv01) node.
>
>  [root at srv01 ~]# rm -rf /var/run/heartbeat/rsctmp/Dummy-clnUMdummy02\:0.state
>
>  * The clone resources are replaced.
>
>  [root at srv01 ~]# ls /var/run/heartbeat/rsctmp/Dummy-clnUMdummy0*
>  /var/run/heartbeat/rsctmp/Dummy-clnUMdummy01:1.state
> /var/run/heartbeat/rsctmp/Dummy-clnUMdummy02:1.state
>
> Step3)Again...I generate the trouble of the clnUMgroup01 clone in a N1(srv01) node.
>
>  [root at srv01 ~]# rm -rf /var/run/heartbeat/rsctmp/Dummy-clnUMdummy02\:1.state
>
>  [root at srv01 ~]# ls /var/run/heartbeat/rsctmp/Dummy-clnUMdummy0*
>  /var/run/heartbeat/rsctmp/Dummy-clnUMdummy01:0.state
> /var/run/heartbeat/rsctmp/Dummy-clnUMdummy02:0.state
>
>  * The clone resources are replaced.
>
> ============
> Last updated: Fri Mar 12 14:56:19 2010
> Stack: openais
> Current DC: srv01 - partition with quorum
> Version: 1.0.7-049006f172774f407e165ec82f7ee09cb73fd0e7
> 4 Nodes configured, 2 expected votes
> 13 Resources configured.
> ============
> Online: [ srv01 srv02 srv03 srv04 ]
> (snip)
> Migration summary:
> * Node srv02:
> * Node srv03:
> * Node srv04:
> * Node srv01:
>   clnUMdummy02:0: migration-threshold=5 fail-count=1
>   clnUMdummy02:1: migration-threshold=5 fail-count=1
>
> Step4)I generate the trouble of the clnUMgroup01 clone in a N4(srv04) node.
>
> [root at srv04 ~]# rm -rf /var/run/heartbeat/rsctmp/Dummy-clnUMdummy02\:1.state
> [root at srv04 ~]# ls /var/run/heartbeat/rsctmp/Dummy-clnUMdummy02*
> /var/run/heartbeat/rsctmp/Dummy-clnUMdummy02:1.state
>
>  * The clone resources are not replaced.
>
> Step5)Again...I generate the trouble of the clnUMgroup01 clone in a N4(srv04) node.
>
>  * The clone resources are not replaced.
>
> (snip)
> Migration summary:
> * Node srv02:
> * Node srv03:
> * Node srv04:
>   clnUMdummy02:1: migration-threshold=5 fail-count=2
> * Node srv01:
>   clnUMdummy02:0: migration-threshold=5 fail-count=1
>   clnUMdummy02:1: migration-threshold=5 fail-count=1
>
>
> Step6)Again...I generate the trouble of the clnUMgroup01 clone in a N4(srv04) node and N1(srv01) node.
>
>  * In the N4 node, trouble of clnUMdummy02 is handled at five times, but, in the N1 node, it is
> processed at much number of times for replacement.
>
> (snip)
>  Clone Set: clnUMgroup01
>     Started: [ srv01 ]
>     Stopped: [ clnUmResource:1 ]
>  Clone Set: clnPingd
>     Started: [ srv01 srv02 srv03 srv04 ]
>  Clone Set: clnDiskd1
>     Started: [ srv01 srv02 srv03 srv04 ]
>  Clone Set: clnG3dummy1
>     Started: [ srv01 srv02 srv03 srv04 ]
>  Clone Set: clnG3dummy2
>     Started: [ srv01 srv02 srv03 srv04 ]
>
> Migration summary:
> * Node srv02:
> * Node srv03:
> * Node srv04:
>   clnUMdummy02:1: migration-threshold=5 fail-count=5
> * Node srv01:
>   clnUMdummy02:0: migration-threshold=5 fail-count=3
>   clnUMdummy02:1: migration-threshold=5 fail-count=3
>
> Of a clone rising in a N1(srv01) node at the time of "globally-unique=false" is replacing it right?
> In addition, is it right movement that replacement does not happen even if a clone breaks down in a
> N4(srv04) node?
>
> We think that, furthermore, there is a problem because the replacement is different.
>
> When it was assumed that the replacement of this clone is right, arrival to the trouble number of
> times is different from a N4(srv04) node in a N1(srv01) node.
>
> By this movement, we cannot set the limit of the trouble number of times of the clone well.

So if I can summarize, you're saying that clnUMdummy02 should not be
allowed to run on srv01 because the combined number of failures is 6
(and clnUMdummy02 is a non-unique clone).

And that the current behavior is that clnUMdummy02 continues to run.

Is that an accurate summary?
If so, then I agree its a bug.  Could you create a bugzilla entry for it please?

>
> This is specifications or bug? (Or is it already solved in the development version?)
> Is setting to operate definitely necessary for cib.xml?
>
> If there is a setting method of right cib.xml, please teach it.
>
> Because the size of the collection of hb_report result is big, I do not attach it.
> If there is information of hb_report which is necessary for the solution of the problem, give me
> comments.
>
> Best Regards,
> Hideo Yamauchi.
>
>
> _______________________________________________
> Pacemaker mailing list
> Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
>




More information about the Pacemaker mailing list