[Pacemaker] Problem : By colocations limitation, the resource appointment of the combination does not become effective.

renayama19661014 at ybb.ne.jp renayama19661014 at ybb.ne.jp
Fri Mar 5 00:42:53 UTC 2010


Hi All,

We test complicated colocation appointment.

We did resource appointment to start by limitation of colocation together.

But, the resource that set limitation starts when the resource that we appointed does not start in a
certain procedure.

We did the following appointment.

      <rsc_colocation id="rsc_colocation01-1" rsc="UMgroup01" with-rsc="clnPingd" score="1000"/>

When clnPingd did not start, we met with the phenomenon that UMgroup01 started.


The procedure to generate a phenomenon is as follows.

STEP1) Start corosync.
STEP2) Send cib.xml to Pacemaker.
STEP3) A cluster is stable.

[root at srv01 ~]# crm_mon -1
============
Last updated: Wed Mar  3 13:21:21 2010
Stack: openais
Current DC: srv01 - partition with quorum
Version: 1.0.7-6e1815972fc236825bf3658d7f8451d33227d420
4 Nodes configured, 4 expected votes
13 Resources configured.
============
Online: [ srv01 srv02 srv03 srv04 ]

 Resource Group: UMgroup01
     UmVIPcheck (ocf::heartbeat:Dummy): Started srv01
     UmIPaddr   (ocf::heartbeat:Dummy): Started srv01
     UmDummy01  (ocf::heartbeat:Dummy): Started srv01
     UmDummy02  (ocf::heartbeat:Dummy): Started srv01
 Resource Group: OVDBgroup02-1
     prmExPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
     prmFsPostgreSQLDB1-1       (ocf::heartbeat:Dummy): Started srv01
     prmFsPostgreSQLDB1-2       (ocf::heartbeat:Dummy): Started srv01
     prmFsPostgreSQLDB1-3       (ocf::heartbeat:Dummy): Started srv01
     prmIpPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
     prmApPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
 Resource Group: OVDBgroup02-2
     prmExPostgreSQLDB2 (ocf::heartbeat:Dummy): Started srv02
     prmFsPostgreSQLDB2-1       (ocf::heartbeat:Dummy): Started srv02
     prmFsPostgreSQLDB2-2       (ocf::heartbeat:Dummy): Started srv02
     prmFsPostgreSQLDB2-3       (ocf::heartbeat:Dummy): Started srv02
     prmIpPostgreSQLDB2 (ocf::heartbeat:Dummy): Started srv02
     prmApPostgreSQLDB2 (ocf::heartbeat:Dummy): Started srv02
 Resource Group: OVDBgroup02-3
     prmExPostgreSQLDB3 (ocf::heartbeat:Dummy): Started srv03
     prmFsPostgreSQLDB3-1       (ocf::heartbeat:Dummy): Started srv03
     prmFsPostgreSQLDB3-2       (ocf::heartbeat:Dummy): Started srv03
     prmFsPostgreSQLDB3-3       (ocf::heartbeat:Dummy): Started srv03
     prmIpPostgreSQLDB3 (ocf::heartbeat:Dummy): Started srv03
     prmApPostgreSQLDB3 (ocf::heartbeat:Dummy): Started srv03
 Resource Group: grpStonith1
     prmStonithN1       (stonith:external/ssh): Started srv04
 Resource Group: grpStonith2
     prmStonithN2       (stonith:external/ssh): Started srv01
 Resource Group: grpStonith3
     prmStonithN3       (stonith:external/ssh): Started srv02
 Resource Group: grpStonith4
     prmStonithN4       (stonith:external/ssh): Started srv03
 Clone Set: clnUMgroup01
     Started: [ srv01 srv04 ]
 Clone Set: clnPingd
     Started: [ srv01 srv02 srv03 srv04 ]
 Clone Set: clnDiskd1
     Started: [ srv01 srv02 srv03 srv04 ]
 Clone Set: clnG3dummy1
     Started: [ srv01 srv02 srv03 srv04 ]
 Clone Set: clnG3dummy2
     Started: [ srv01 srv02 srv03 srv04 ]

STEP4) Camouflage a stop error of pingd of the srv01 node.

pingd_stop() {
   exit $OCF_ERR_GENERIC
    if [ -f $OCF_RESKEY_pidfile ]; then
        pid=`cat $OCF_RESKEY_pidfile`
    fi

STEP5) Stop a clnPingd clone.

[root at srv01 ~]# crm
crm(live)# resource
crm(live)resource# stop clnPingd

[root at srv01 ~]# crm_mon -1 -f
============
Last updated: Wed Mar  3 13:24:16 2010
Stack: openais
Current DC: srv01 - partition with quorum
Version: 1.0.7-6e1815972fc236825bf3658d7f8451d33227d420
4 Nodes configured, 4 expected votes
13 Resources configured.
============

Online: [ srv01 srv02 srv03 srv04 ]

 Resource Group: UMgroup01
     UmVIPcheck (ocf::heartbeat:Dummy): Started srv01
     UmIPaddr   (ocf::heartbeat:Dummy): Started srv01
     UmDummy01  (ocf::heartbeat:Dummy): Started srv01
     UmDummy02  (ocf::heartbeat:Dummy): Started srv01
 Resource Group: OVDBgroup02-1
     prmExPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
     prmFsPostgreSQLDB1-1       (ocf::heartbeat:Dummy): Started srv01
     prmFsPostgreSQLDB1-2       (ocf::heartbeat:Dummy): Started srv01
     prmFsPostgreSQLDB1-3       (ocf::heartbeat:Dummy): Started srv01
     prmIpPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
     prmApPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
 Resource Group: grpStonith1
     prmStonithN1       (stonith:external/ssh): Started srv04
 Resource Group: grpStonith2
     prmStonithN2       (stonith:external/ssh): Started srv01
 Resource Group: grpStonith3
     prmStonithN3       (stonith:external/ssh): Started srv02
 Resource Group: grpStonith4
     prmStonithN4       (stonith:external/ssh): Started srv03
 Clone Set: clnUMgroup01
     Started: [ srv01 srv04 ]
 Clone Set: clnDiskd1
     Started: [ srv01 srv02 srv03 srv04 ]
 Clone Set: clnG3dummy1
     Started: [ srv01 srv02 srv03 srv04 ]
 Clone Set: clnG3dummy2
     Started: [ srv01 srv02 srv03 srv04 ]

Migration summary:
* Node srv02: 
* Node srv04: 
* Node srv03: 
* Node srv01: 
   clnPrmPingd:0: migration-threshold=10 fail-count=1000000

STEP6) Return a revision of pingd.

pingd_stop() {
#   exit $OCF_ERR_GENERIC
    if [ -f $OCF_RESKEY_pidfile ]; then
        pid=`cat $OCF_RESKEY_pidfile`
    fi

STEP7) Reboot a srv01 node.

STEP8) Wait for completion of STONITH.(STONITH is completed by a retry)

[root at srv02 ~]# crm_mon -1
============
Last updated: Wed Mar  3 13:34:12 2010
Stack: openais
Current DC: srv02 - partition with quorum
Version: 1.0.7-6e1815972fc236825bf3658d7f8451d33227d420
4 Nodes configured, 4 expected votes
13 Resources configured.
============

Online: [ srv02 srv03 srv04 ]
OFFLINE: [ srv01 ]

 Resource Group: grpStonith1
     prmStonithN1       (stonith:external/ssh): Started srv04
 Resource Group: grpStonith2
     prmStonithN2       (stonith:external/ssh): Started srv03
 Resource Group: grpStonith3
     prmStonithN3       (stonith:external/ssh): Started srv02
 Resource Group: grpStonith4
     prmStonithN4       (stonith:external/ssh): Started srv03
 Clone Set: clnUMgroup01
     Started: [ srv04 ]
     Stopped: [ clnUmResource:0 ]
 Clone Set: clnDiskd1
     Started: [ srv02 srv03 srv04 ]
     Stopped: [ clnPrmDiskd1:0 ]
 Clone Set: clnG3dummy1
     Started: [ srv02 srv03 srv04 ]
     Stopped: [ clnG3dummy01:0 ]
 Clone Set: clnG3dummy2
     Started: [ srv02 srv03 srv04 ]
     Stopped: [ clnG3dummy02:0 ]

STEP9) Start corosync in srv01 which rebooted. 
[root at srv02 ~]# crm_mon -1
============
Last updated: Wed Mar  3 13:37:57 2010
Stack: openais
Current DC: srv02 - partition with quorum
Version: 1.0.7-6e1815972fc236825bf3658d7f8451d33227d420
4 Nodes configured, 4 expected votes
13 Resources configured.
============

Online: [ srv01 srv02 srv03 srv04 ]

 Resource Group: UMgroup01
     UmVIPcheck (ocf::heartbeat:Dummy): Started srv01
     UmIPaddr   (ocf::heartbeat:Dummy): Started srv01
     UmDummy01  (ocf::heartbeat:Dummy): Started srv01
     UmDummy02  (ocf::heartbeat:Dummy): Started srv01
 Resource Group: OVDBgroup02-1
     prmExPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
     prmFsPostgreSQLDB1-1       (ocf::heartbeat:Dummy): Started srv01
     prmFsPostgreSQLDB1-2       (ocf::heartbeat:Dummy): Started srv01
     prmFsPostgreSQLDB1-3       (ocf::heartbeat:Dummy): Started srv01
     prmIpPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
     prmApPostgreSQLDB1 (ocf::heartbeat:Dummy): Started srv01
 Resource Group: grpStonith1
     prmStonithN1       (stonith:external/ssh): Started srv04
 Resource Group: grpStonith2
     prmStonithN2       (stonith:external/ssh): Started srv03
 Resource Group: grpStonith3
     prmStonithN3       (stonith:external/ssh): Started srv02
 Resource Group: grpStonith4
     prmStonithN4       (stonith:external/ssh): Started srv03
 Clone Set: clnUMgroup01
     Started: [ srv01 srv04 ]
 Clone Set: clnDiskd1
     Started: [ srv01 srv02 srv03 srv04 ]
 Clone Set: clnG3dummy1
     Started: [ srv01 srv02 srv03 srv04 ]
 Clone Set: clnG3dummy2
     Started: [ srv01 srv02 srv03 srv04 ]

[root at srv02 ~]# ptest -L -s | grep UmVIPcheck
group_color: UmVIPcheck allocation score on srv01: 300
group_color: UmVIPcheck allocation score on srv02: -1000000
group_color: UmVIPcheck allocation score on srv03: -1000000
group_color: UmVIPcheck allocation score on srv04: -1000000
native_color: UmVIPcheck allocation score on srv01: 1600
native_color: UmVIPcheck allocation score on srv02: -1000000
native_color: UmVIPcheck allocation score on srv03: -1000000
native_color: UmVIPcheck allocation score on srv04: -1000000


But clnPingd does not start in srv01, but UMgroup01 starts after this.
* Because there was colocation limitation, we did not expect start of UMgroup01.

Is there an error for my setting? 
Or is it a bug?
Or is this right movement?

I attached the thing which added a pengine directory of srv02 to a result of hb_report.
But, I delete it and attach it because a file is big as for the information of srv01,srv03,srv04.

Best Regards,
Hideo Yamauchi.
-------------- next part --------------
A non-text attachment was scrubbed...
Name: hb_report996.tar.gz
Type: application/x-gzip-compressed
Size: 256355 bytes
Desc: 2393974864-hb_report996.tar.gz
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20100305/1bfaaa37/attachment-0003.bin>


More information about the Pacemaker mailing list