[Pacemaker] Migrate/run resource only if m/s resource is master/promoted on target node (follow-up)

Martin Gombač martin at isg.si
Wed Dec 30 09:12:23 UTC 2009


I have it set already to ignore, because i do resource level fencing 
with custom app, which is run/triggered by drbd when it looses connection.
 >
property $id="cib-bootstrap-options" \
    dc-version="1.0.6-f709c638237cdff7556cb6ab615f32826c0f8c06" \
    cluster-infrastructure="Heartbeat" \
    stonith-enabled="false" \
    no-quorum-policy="ignore" \
    default-resource-stickiness="1000" \
    last-lrm-refresh="1262007932"
 >

I have just noticed an error with ms_drbd_r0.
 >
ms ms_drbd_r0 drbd_r0 \
    meta notify="true" master-max="2" inteleave="true"
 >
interleave is not inteleave, correct? :-)
If that won't help i'll also try node-master-max and/or 
clone(-master)-max meta options.

Will let you know if it works out.

Regards,
M.

hj lee wrote:
> Hi,
>
> Maybe this is related to no-quorum-policy. What is no-quorum-policy? 
> You can check no-quorum-policy in "crm configure show" command. If you 
> can not see in "crm configure show" command, then it is "stop" by 
> default. If that is your case, then please set it to ignore by "crm 
> configure property no-quorum-policy=ignore". and try it again.
>
> Thanks
> hj
>
> On Tue, Dec 29, 2009 at 6:43 AM, Martin Gombač <martin at isg.si 
> <mailto:martin at isg.si>> wrote:
>
>     Hi guys once again :-)
>
>     my resource Hosting on top of ms_drbd_r0 keeps restarting even if
>     the changes aren't local to the node.
>     By that, i mean Hosting get's restarted on node1 even if i restart
>     or outdate and demote node 2.
>
>     My constrains:
>     colocation Hosting_on_ms_drbd_r0 inf: Hosting ms_drbd_r0:Master
>     order ms_drbd_r0_b4_Hosting inf: ms_drbd_r0:promote Hosting:start
>
>     My resources:
>     primitive Hosting ocf:heartbeat:Xen \
>       params xmfile="/etc/xen/Hosting.cfg" \
>       meta target-role="Started" allow-migrate="true" is-managed="true" \
>       op monitor interval="120s" timeout="300s"
>     primitive drbd_r0 ocf:linbit:drbd \
>       params drbd_resource="r0" \
>       op monitor interval="15s" role="Master" timeout="30s" \
>       op monitor interval="30s" role="Slave" timeout="30"
>     ms ms_drbd_r0 drbd_r0 \
>       meta notify="true" master-max="2" inteleave="true"
>
>     I use location to:
>     location cli-prefer-Hosting Hosting \
>       rule $id="cli-prefer-rule-Hosting" inf: #uname eq ibm1
>
>     Example:
>     Shuting down heartbeat on node2/ibm2, while resource is running
>     perfectly fine on node1, makes resource Hosting restart on node1.
>
>     Dec 29 14:21:28 ibm1 pengine: [3716]: notice: native_print:
>     Hosting    (ocf::heartbeat:Xen):    Started ibm1
>     Dec 29 14:21:28 ibm1 pengine: [3716]: notice: clone_print:
>      Master/Slave Set: ms_drbd_r0
>     Dec 29 14:21:28 ibm1 pengine: [3716]: notice: short_print:    
>      Masters: [ ibm1 ibm2]
>     Dec 29 14:21:28 ibm1 pengine: [3716]: WARN: native_color: Resource
>     drbd_r0:1 cannot run anywhere
>     Dec 29 14:21:28 ibm1 pengine: [3716]: info: master_color:
>     Promoting drbd_r0:0 (Master ibm1)
>     Dec 29 14:21:28 ibm1 pengine: [3716]: info: master_color:
>     ms_drbd_r0: Promoted 1 instances of a possible 2 to master
>     Dec 29 14:21:28 ibm1 pengine: [3716]: info: master_color:
>     Promoting drbd_r0:0 (Master ibm1
>     Dec 29 14:21:28 ibm1 pengine: [3716]: info: master_color:
>     ms_drbd_r0: Promoted 1 instances of a possible 2 to master
>     Dec 29 14:21:28 ibm1 pengine: [3716]: info: stage6: Scheduling
>     Node ibm2 for shutdown
>     Dec 29 14:21:28 ibm1 pengine: [3716]: notice: LogActions: Restart
>     resource Hosting    (Started ibm1)
>     Dec 29 14:21:28 ibm1 pengine: [3716]: notice: LogActions: Leave
>     resource drbd_r0:0    (Master ibm1)
>     Dec 29 14:21:28 ibm1 pengine: [3716]: notice: LogActions: Demote
>     drbd_r0:1    (Master -> Stopped ibm2)
>     Dec 29 14:21:28 ibm1 pengine: [3716]: notice: LogActions: Stop
>     resource drbd_r0:1    (ibm2)
>     ...
>     Dec 29 14:21:28 ibm1 crmd: [3713]: info: te_rsc_command:
>     Initiating action 8: stop Hosting_stop_0 on ibm1(local)
>     Dec 29 14:21:28 ibm1 crmd: [3713]: info: do_lrm_rsc_op: Performing
>     key=8:4:0:13282265-c62f-4341-9fa5-363cd30ddd3e op=Hosting_stop_0 )
>     Dec 29 14:21:28 ibm1 lrmd: [3710]: info: rsc:Hosting:19: stop
>     ...
>     Dec 29 14:21:34 ibm1 crmd: [3713]: info: match_graph_event: Action
>     Hosting_stop_0 (8) confirmed on ibm1..si (rc=0)
>     Dec 29 14:21:34 ibm1 crmd: [3713]: info: te_rsc_command:
>     Initiating action 9: start Hosting_start_0 on ibm1..si (local)
>     Dec 29 14:21:34 ibm1 lrmd: [3710]: info: rsc:Hosting:21: start
>     ....
>     ...
>     .
>
>
>     Please advise me on how to configure ordering and collocation
>     constrains, to not get Hosting resource restarted each time when
>     something happens to backup drbd resource on the second node.
>
>
>     Thank you.
>     Martin
>
>     _______________________________________________
>     Pacemaker mailing list
>     Pacemaker at oss.clusterlabs.org <mailto:Pacemaker at oss.clusterlabs.org>
>     http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
>
>
>
> -- 
> Dream with longterm vision!
> kerdosa
> ------------------------------------------------------------------------
>
> _______________________________________________
> Pacemaker mailing list
> Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>   




More information about the Pacemaker mailing list