[Pacemaker] Stopping corosync on the slave in a 2 node cluster causes all the resources on the master to stop. Please Help.

Tom Pride tom.pride at gmail.com
Wed Oct 12 09:58:55 EDT 2011


Hi Florian,

Thank you for responding.

The config I provided in my initial posting is actually a cut down version
of what we are actually running.  There are really 11 activeMQ brokers
(resources) running on the cluster, so the xml output from cibadmin  is not
going to match the initial config I provided.

It might help if I give you some more detail on our setup and the history of
the testing I have done so far on this issue.  We actually have a total of 7
two node clusters (1 integration cluster, 1 test cluster, 2 staging
clusters, 2 live clusters and 1 stabe cluster) , all 7 use the same base crm
config of 4 resources before any activeMQ resources are added.  The problem
of the master shutting down all it's resources when corosync is stopped on
the slave exists on all 7 clusters.  It also occurs with just the 4 base
resources configured, so adding the activeMQ resources has no affect either
way when it comes to this problem. The 4 base resources are:

primitive activemq_drbd ocf:linbit:drbd
primitive cluster_IP ocf:heartbeat:IPaddr2
primitive drbd_fs ocf:heartbeat:Filesystem
primitive ping_gateway ocf:pacemaker:ping

This is what the output of crm_mon looks like on all the clusters when no
activeMQ resources have been added:

============
Last updated: Wed Oct 12 14:47:08 2011
Stack: openais
Current DC: mq009.back.stable.cwwtf.local - partition with quorum
Version: 1.0.11-1554a83db0d3c3e546cfd3aaff6af1184f79ee87
2 Nodes configured, 2 expected votes
4 Resources configured.
============

Online: [ mq009.back.stable.cwwtf.local mq010.back.stable.cwwtf.local ]

 Clone Set: ping_gateway_clone
     Started: [ mq009.back.stable.cwwtf.local mq010.back.stable.cwwtf.local
]
 Master/Slave Set: ActiveMQ_Data
     Masters: [ mq009.back.stable.cwwtf.local ]
     Slaves: [ mq010.back.stable.cwwtf.local ]
drbd_fs (ocf::heartbeat:Filesystem):    Started
mq009.back.stable.cwwtf.local
cluster_IP (ocf::heartbeat:IPaddr2): Started mq009.back.stable.cwwtf.local

To make it easier for you to help me diagnose this issue I have removed all
of the activeMQ broker resources from our "stable" cluster and have pasted
both the crm config and the cibadmin xml output from the cluster below.  Any
help you can give me on this would be very much appreciated.

root at mq010:~# crm configure show
node mq009.back.stable.cwwtf.local \
attributes standby="off"
node mq010.back.stable.cwwtf.local \
attributes standby="off"
primitive activemq_drbd ocf:linbit:drbd \
params drbd_resource="r0" \
op monitor interval="15s" timeout="20s" \
op start interval="0" timeout="240" \
op stop interval="0" timeout="100"
primitive cluster_IP ocf:heartbeat:IPaddr2 \
params ip="172.23.42.55" nic="eth0" \
op monitor interval="30s" timeout="90" \
op start interval="0" timeout="90" \
op stop interval="0" timeout="100"
primitive drbd_fs ocf:heartbeat:Filesystem \
params device="/dev/drbd1" directory="/drbd" fstype="ext3" \
op monitor interval="15s" timeout="40s" \
op start interval="0" timeout="60" \
op stop interval="0" timeout="60"
primitive ping_gateway ocf:pacemaker:ping \
params name="ping_gateway" host_list="172.23.40.2" multiplier="100" \
op monitor interval="15s" timeout="20s" \
op start interval="0" timeout="90" \
op stop interval="0" timeout="100"
ms ActiveMQ_Data activemq_drbd \
meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1"
notify="true" target-role="Master"
clone ping_gateway_clone ping_gateway
location ActiveMQ_Data_on_connected_node_only ActiveMQ_Data \
rule $id="ActiveMQ_Data_on_connected_node_only-rule" -inf: not_defined
ping_gateway or ping_gateway lte 0
location ActiveMQ_Data_prefer_mq009 ActiveMQ_Data \
rule $id="ActiveMQ_Data_prefer_mq009-rule" $role="Master" 500: #uname eq
mq009.back.stable.cwwtf.local
colocation cluster_IP_with_ActiveMQ_Data inf: cluster_IP
ActiveMQ_Data:Master
colocation drbd_fs_with_ActiveMQ_Data inf: drbd_fs ActiveMQ_Data:Master
order ActiveMQ_Data_after_ping_gateway_clone inf: ping_gateway_clone:start
ActiveMQ_Data:promote
order cluster_IP_after_drbd_fs inf: drbd_fs:start cluster_IP:start
order drbd_fs_after_ActiveMQ_Data inf: ActiveMQ_Data:promote drbd_fs:start
property $id="cib-bootstrap-options" \
dc-version="1.0.11-1554a83db0d3c3e546cfd3aaff6af1184f79ee87" \
cluster-infrastructure="openais" \
expected-quorum-votes="2" \
no-quorum-policy="ignore" \
stonith-enabled="false" \
last-lrm-refresh="1309359608"
rsc_defaults $id="rsc-options" \
resource-stickiness="100"



root at mq010:~# cibadmin -Q
<cib epoch="90" admin_epoch="0" validate-with="pacemaker-1.0"
crm_feature_set="3.0.1" have-quorum="1" num_updates="35"
cib-last-written="Tue Jul  5 12:23:25 2011"
dc-uuid="mq009.back.stable.cwwtf.local">
  <configuration>
    <crm_config>
      <cluster_property_set id="cib-bootstrap-options">
        <nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
value="1.0.11-1554a83db0d3c3e546cfd3aaff6af1184f79ee87"/>
        <nvpair id="cib-bootstrap-options-cluster-infrastructure"
name="cluster-infrastructure" value="openais"/>
        <nvpair id="cib-bootstrap-options-expected-quorum-votes"
name="expected-quorum-votes" value="2"/>
        <nvpair id="cib-bootstrap-options-no-quorum-policy"
name="no-quorum-policy" value="ignore"/>
        <nvpair id="cib-bootstrap-options-stonith-enabled"
name="stonith-enabled" value="false"/>
        <nvpair id="cib-bootstrap-options-last-lrm-refresh"
name="last-lrm-refresh" value="1309359608"/>
      </cluster_property_set>
    </crm_config>
    <nodes>
      <node id="mq009.back.stable.cwwtf.local" type="normal"
uname="mq009.back.stable.cwwtf.local">
        <instance_attributes id="nodes-mq009.back.stable.cwwtf.local">
          <nvpair id="nodes-mq009.back.stable.cwwtf.local-standby"
name="standby" value="off"/>
        </instance_attributes>
      </node>
      <node id="mq010.back.stable.cwwtf.local" type="normal"
uname="mq010.back.stable.cwwtf.local">
        <instance_attributes id="nodes-mq010.back.stable.cwwtf.local">
          <nvpair id="nodes-mq010.back.stable.cwwtf.local-standby"
name="standby" value="off"/>
        </instance_attributes>
      </node>
    </nodes>
    <resources>
      <clone id="ping_gateway_clone">
        <primitive class="ocf" id="ping_gateway" provider="pacemaker"
type="ping">
          <instance_attributes id="ping_gateway-instance_attributes">
            <nvpair id="ping_gateway-instance_attributes-name" name="name"
value="ping_gateway"/>
            <nvpair id="ping_gateway-instance_attributes-host_list"
name="host_list" value="172.23.40.2"/>
            <nvpair id="ping_gateway-instance_attributes-multiplier"
name="multiplier" value="100"/>
          </instance_attributes>
          <operations>
            <op id="ping_gateway-monitor-15s" interval="15s" name="monitor"
timeout="20s"/>
            <op id="ping_gateway-start-0" interval="0" name="start"
timeout="90"/>
            <op id="ping_gateway-stop-0" interval="0" name="stop"
timeout="100"/>
          </operations>
        </primitive>
      </clone>
      <master id="ActiveMQ_Data">
        <meta_attributes id="ActiveMQ_Data-meta_attributes">
          <nvpair id="ActiveMQ_Data-meta_attributes-master-max"
name="master-max" value="1"/>
          <nvpair id="ActiveMQ_Data-meta_attributes-master-node-max"
name="master-node-max" value="1"/>
          <nvpair id="ActiveMQ_Data-meta_attributes-clone-max"
name="clone-max" value="2"/>
          <nvpair id="ActiveMQ_Data-meta_attributes-clone-node-max"
name="clone-node-max" value="1"/>
          <nvpair id="ActiveMQ_Data-meta_attributes-notify" name="notify"
value="true"/>
          <nvpair id="ActiveMQ_Data-meta_attributes-target-role"
name="target-role" value="Master"/>
        </meta_attributes>
        <primitive class="ocf" id="activemq_drbd" provider="linbit"
type="drbd">
          <instance_attributes id="activemq_drbd-instance_attributes">
            <nvpair id="activemq_drbd-instance_attributes-drbd_resource"
name="drbd_resource" value="r0"/>
          </instance_attributes>
          <operations>
            <op id="activemq_drbd-monitor-15s" interval="15s" name="monitor"
timeout="20s"/>
            <op id="activemq_drbd-start-0" interval="0" name="start"
timeout="240"/>
            <op id="activemq_drbd-stop-0" interval="0" name="stop"
timeout="100"/>
          </operations>
        </primitive>
      </master>
      <primitive class="ocf" id="drbd_fs" provider="heartbeat"
type="Filesystem">
        <instance_attributes id="drbd_fs-instance_attributes">
          <nvpair id="drbd_fs-instance_attributes-device" name="device"
value="/dev/drbd1"/>
          <nvpair id="drbd_fs-instance_attributes-directory"
name="directory" value="/drbd"/>
          <nvpair id="drbd_fs-instance_attributes-fstype" name="fstype"
value="ext3"/>
        </instance_attributes>
        <operations>
          <op id="drbd_fs-monitor-15s" interval="15s" name="monitor"
timeout="40s"/>
          <op id="drbd_fs-start-0" interval="0" name="start" timeout="60"/>
          <op id="drbd_fs-stop-0" interval="0" name="stop" timeout="60"/>
        </operations>
      </primitive>
      <primitive class="ocf" id="cluster_IP" provider="heartbeat"
type="IPaddr2">
        <instance_attributes id="cluster_IP-instance_attributes">
          <nvpair id="cluster_IP-instance_attributes-ip" name="ip"
value="172.23.42.55"/>
          <nvpair id="cluster_IP-instance_attributes-nic" name="nic"
value="eth0"/>
        </instance_attributes>
        <operations>
          <op id="cluster_IP-monitor-30s" interval="30s" name="monitor"
timeout="90"/>
          <op id="cluster_IP-start-0" interval="0" name="start"
timeout="90"/>
          <op id="cluster_IP-stop-0" interval="0" name="stop"
timeout="100"/>
        </operations>
      </primitive>
    </resources>
    <constraints>
      <rsc_location id="ActiveMQ_Data_on_connected_node_only"
rsc="ActiveMQ_Data">
        <rule boolean-op="or" id="ActiveMQ_Data_on_connected_node_only-rule"
score="-INFINITY">
          <expression attribute="ping_gateway"
id="ActiveMQ_Data_on_connected_node_only-expression"
operation="not_defined"/>
          <expression attribute="ping_gateway"
id="ActiveMQ_Data_on_connected_node_only-expression-0" operation="lte"
value="0"/>
        </rule>
      </rsc_location>
      <rsc_location id="ActiveMQ_Data_prefer_mq009" rsc="ActiveMQ_Data">
        <rule id="ActiveMQ_Data_prefer_mq009-rule" role="Master"
score="500">
          <expression attribute="#uname"
id="ActiveMQ_Data_prefer_mq009-expression" operation="eq"
value="mq009.back.stable.cwwtf.local"/>
        </rule>
      </rsc_location>
      <rsc_colocation id="cluster_IP_with_ActiveMQ_Data" rsc="cluster_IP"
score="INFINITY" with-rsc="ActiveMQ_Data" with-rsc-role="Master"/>
      <rsc_colocation id="drbd_fs_with_ActiveMQ_Data" rsc="drbd_fs"
score="INFINITY" with-rsc="ActiveMQ_Data" with-rsc-role="Master"/>
      <rsc_order first="ping_gateway_clone" first-action="start"
id="ActiveMQ_Data_after_ping_gateway_clone" score="INFINITY"
then="ActiveMQ_Data" then-action="promote"/>
      <rsc_order first="drbd_fs" first-action="start"
id="cluster_IP_after_drbd_fs" score="INFINITY" then="cluster_IP"
then-action="start"/>
      <rsc_order first="ActiveMQ_Data" first-action="promote"
id="drbd_fs_after_ActiveMQ_Data" score="INFINITY" then="drbd_fs"
then-action="start"/>
    </constraints>
    <op_defaults/>
    <rsc_defaults>
      <meta_attributes id="rsc-options">
        <nvpair id="rsc-options-resource-stickiness"
name="resource-stickiness" value="100"/>
      </meta_attributes>
    </rsc_defaults>
  </configuration>
  <status>
    <node_state id="mq009.back.stable.cwwtf.local"
uname="mq009.back.stable.cwwtf.local" ha="active" in_ccm="true"
crmd="online" join="member" expected="member"
crm-debug-origin="do_update_resource" shutdown="0">
      <lrm id="mq009.back.stable.cwwtf.local">
        <lrm_resources>
          <lrm_resource id="ping_gateway:0" type="ping" class="ocf"
provider="pacemaker">
            <lrm_rsc_op id="ping_gateway:0_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="4:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:7;4:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="2" rc-code="7" op-status="0" interval="0" last-run="1309874967"
last-rc-change="1309874967" exec-time="20" queue-time="30"
op-digest="4faa58043b927d02be6352d6d4b5d1b3"/>
            <lrm_rsc_op id="ping_gateway:0_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="13:1:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;13:1:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="6" rc-code="0" op-status="0" interval="0" last-run="1309874968"
last-rc-change="1309874968" exec-time="4030" queue-time="20"
op-digest="4faa58043b927d02be6352d6d4b5d1b3"/>
            <lrm_rsc_op id="ping_gateway:0_monitor_15000"
operation="monitor" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.1"
transition-key="14:1:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;14:1:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="7" rc-code="0" op-status="0" interval="15000"
last-rc-change="1309874972" exec-time="4020" queue-time="0"
op-digest="43e591f50112aeb519555a036ffbf937"/>
          </lrm_resource>
          <lrm_resource id="activemq_drbd:0" type="drbd" class="ocf"
provider="linbit">
            <lrm_rsc_op id="activemq_drbd:0_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="5:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:7;5:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="3" rc-code="7" op-status="0" interval="0" last-run="1309874967"
last-rc-change="1309874967" exec-time="70" queue-time="20"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
            <lrm_rsc_op id="activemq_drbd:0_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="15:2:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;15:2:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="8" rc-code="0" op-status="0" interval="0" last-run="1309874978"
last-rc-change="1309874978" exec-time="270" queue-time="20"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
            <lrm_rsc_op id="activemq_drbd:0_post_notify_start_0"
operation="notify" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.1"
transition-key="56:2:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;56:2:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="9" rc-code="0" op-status="0" interval="0" last-run="1309874979"
last-rc-change="1309874979" exec-time="90" queue-time="50"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
            <lrm_rsc_op id="activemq_drbd:0_pre_notify_promote_0"
operation="notify" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.1"
transition-key="64:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;64:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="10" rc-code="0" op-status="0" interval="0" last-run="1309874980"
last-rc-change="1309874980" exec-time="30" queue-time="30"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
            <lrm_rsc_op id="activemq_drbd:0_promote_0" operation="promote"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="18:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;18:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="11" rc-code="0" op-status="0" interval="0" last-run="1309874981"
last-rc-change="1309874981" exec-time="70" queue-time="20"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
            <lrm_rsc_op id="activemq_drbd:0_post_notify_promote_0"
operation="notify" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.1"
transition-key="65:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;65:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="12" rc-code="0" op-status="0" interval="0" last-run="1309874981"
last-rc-change="1309874981" exec-time="70" queue-time="30"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
          </lrm_resource>
          <lrm_resource id="drbd_fs" type="Filesystem" class="ocf"
provider="heartbeat">
            <lrm_rsc_op id="drbd_fs_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="6:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:7;6:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="4" rc-code="7" op-status="0" interval="0" last-run="1309874967"
last-rc-change="1309874967" exec-time="40" queue-time="40"
op-digest="fed478d3b644235e16ce536a4a455e20"/>
            <lrm_rsc_op id="drbd_fs_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="47:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;47:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="13" rc-code="0" op-status="0" interval="0" last-run="1309874982"
last-rc-change="1309874982" exec-time="160" queue-time="20"
op-digest="fed478d3b644235e16ce536a4a455e20"/>
            <lrm_rsc_op id="drbd_fs_monitor_15000" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="48:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;48:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="14" rc-code="0" op-status="0" interval="15000"
last-rc-change="1309874983" exec-time="40" queue-time="20"
op-digest="b61254e5e6d321ff8ea886e90a81929d"/>
          </lrm_resource>
          <lrm_resource id="cluster_IP" type="IPaddr2" class="ocf"
provider="heartbeat">
            <lrm_rsc_op id="cluster_IP_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="7:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:7;7:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="5" rc-code="7" op-status="0" interval="0" last-run="1309874967"
last-rc-change="1309874967" exec-time="30" queue-time="30"
op-digest="e683599137b6cfe2a2737572c307b777"/>
            <lrm_rsc_op id="cluster_IP_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="49:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;49:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="15" rc-code="0" op-status="0" interval="0" last-run="1309874983"
last-rc-change="1309874983" exec-time="110" queue-time="20"
op-digest="e683599137b6cfe2a2737572c307b777"/>
            <lrm_rsc_op id="cluster_IP_monitor_30000" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="50:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;50:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="16" rc-code="0" op-status="0" interval="30000"
last-rc-change="1309874983" exec-time="30" queue-time="20"
op-digest="4e48b7523fb8821dbd8b003784ca14e5"/>
          </lrm_resource>
        </lrm_resources>
      </lrm>
      <transient_attributes id="mq009.back.stable.cwwtf.local">
        <instance_attributes id="status-mq009.back.stable.cwwtf.local">
          <nvpair id="status-mq009.back.stable.cwwtf.local-probe_complete"
name="probe_complete" value="true"/>
          <nvpair id="status-mq009.back.stable.cwwtf.local-ping_gateway"
name="ping_gateway" value="100"/>
          <nvpair
id="status-mq009.back.stable.cwwtf.local-master-activemq_drbd:0"
name="master-activemq_drbd:0" value="10000"/>
        </instance_attributes>
      </transient_attributes>
    </node_state>
    <node_state id="mq010.back.stable.cwwtf.local"
uname="mq010.back.stable.cwwtf.local" ha="active" in_ccm="true"
crmd="online" join="member" expected="member"
crm-debug-origin="do_update_resource" shutdown="0">
      <lrm id="mq010.back.stable.cwwtf.local">
        <lrm_resources>
          <lrm_resource id="ping_gateway:1" type="ping" class="ocf"
provider="pacemaker">
            <lrm_rsc_op id="ping_gateway:1_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="9:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:7;9:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="2" rc-code="7" op-status="0" interval="0" last-run="1309874967"
last-rc-change="1309874967" exec-time="20" queue-time="0"
op-digest="4faa58043b927d02be6352d6d4b5d1b3"/>
            <lrm_rsc_op id="ping_gateway:1_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="15:1:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;15:1:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="6" rc-code="0" op-status="0" interval="0" last-run="1309874968"
last-rc-change="1309874968" exec-time="4020" queue-time="0"
op-digest="4faa58043b927d02be6352d6d4b5d1b3"/>
            <lrm_rsc_op id="ping_gateway:1_monitor_15000"
operation="monitor" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.1"
transition-key="16:1:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;16:1:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="7" rc-code="0" op-status="0" interval="15000"
last-rc-change="1309874972" exec-time="4020" queue-time="0"
op-digest="43e591f50112aeb519555a036ffbf937"/>
          </lrm_resource>
          <lrm_resource id="activemq_drbd:0" type="drbd" class="ocf"
provider="linbit">
            <lrm_rsc_op id="activemq_drbd:0_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="10:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:7;10:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="3" rc-code="7" op-status="0" interval="0" last-run="1309874967"
last-rc-change="1309874967" exec-time="40" queue-time="0"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
          </lrm_resource>
          <lrm_resource id="drbd_fs" type="Filesystem" class="ocf"
provider="heartbeat">
            <lrm_rsc_op id="drbd_fs_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="11:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:7;11:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="4" rc-code="7" op-status="0" interval="0" last-run="1309874967"
last-rc-change="1309874967" exec-time="40" queue-time="0"
op-digest="fed478d3b644235e16ce536a4a455e20"/>
          </lrm_resource>
          <lrm_resource id="cluster_IP" type="IPaddr2" class="ocf"
provider="heartbeat">
            <lrm_rsc_op id="cluster_IP_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="12:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:7;12:1:7:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="5" rc-code="7" op-status="0" interval="0" last-run="1309874967"
last-rc-change="1309874967" exec-time="40" queue-time="0"
op-digest="e683599137b6cfe2a2737572c307b777"/>
          </lrm_resource>
          <lrm_resource id="activemq_drbd:1" type="drbd" class="ocf"
provider="linbit">
            <lrm_rsc_op id="activemq_drbd:1_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="17:2:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;17:2:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="8" rc-code="0" op-status="0" interval="0" last-run="1309874978"
last-rc-change="1309874978" exec-time="250" queue-time="0"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
            <lrm_rsc_op id="activemq_drbd:1_post_notify_start_0"
operation="notify" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.1"
transition-key="57:2:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;57:2:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="9" rc-code="0" op-status="0" interval="0" last-run="1309874979"
last-rc-change="1309874979" exec-time="70" queue-time="0"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
            <lrm_rsc_op id="activemq_drbd:1_pre_notify_promote_0"
operation="notify" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.1"
transition-key="66:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;66:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="10" rc-code="0" op-status="0" interval="0" last-run="1309874980"
last-rc-change="1309874980" exec-time="40" queue-time="0"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
            <lrm_rsc_op id="activemq_drbd:1_post_notify_promote_0"
operation="notify" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.1"
transition-key="67:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;67:3:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="11" rc-code="0" op-status="0" interval="0" last-run="1309874981"
last-rc-change="1309874981" exec-time="40" queue-time="0"
op-digest="c0e018b73fdf522b6cdd355e125af15e"/>
            <lrm_rsc_op id="activemq_drbd:1_monitor_15000"
operation="monitor" crm-debug-origin="do_update_resource"
crm_feature_set="3.0.1"
transition-key="22:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
transition-magic="0:0;22:4:0:c6cb2823-95a3-4c61-802a-8af4d9fea224"
call-id="12" rc-code="0" op-status="0" interval="15000"
last-rc-change="1309874982" exec-time="50" queue-time="0"
op-digest="db2fb3216fa7207fdc0b2bf31bbd09c1"/>
          </lrm_resource>
        </lrm_resources>
      </lrm>
      <transient_attributes id="mq010.back.stable.cwwtf.local">
        <instance_attributes id="status-mq010.back.stable.cwwtf.local">
          <nvpair id="status-mq010.back.stable.cwwtf.local-probe_complete"
name="probe_complete" value="true"/>
          <nvpair id="status-mq010.back.stable.cwwtf.local-ping_gateway"
name="ping_gateway" value="100"/>
          <nvpair
id="status-mq010.back.stable.cwwtf.local-master-activemq_drbd:1"
name="master-activemq_drbd:1" value="10000"/>
        </instance_attributes>
      </transient_attributes>
    </node_state>
  </status>
</cib>



On Wed, Oct 12, 2011 at 1:11 PM, Florian Haas <florian at hastexo.com> wrote:

> Tom,
>
> On 2011-10-12 14:02, Tom Pride wrote:
> > Hi Guys,
> >
> > I was hoping that someone might be able to do a quick review of my
> > cluster config, shown below, to work out why when I shutdown corosync on
> > the master all the resources failover to the slave without a problem,
> > but if I shutdown corosync on the slave all of the resources on the
> > master stop as well, leaving me with both nodes broken.  Obviously what
> > I want to happen is to be able to shutdown corosync on the slave and all
> > of the resources running on the master to remain untouched. I must have
> > something not quite right in the logic of my cluster config.
>
> As your issue may not only be caused by a problem in your cluster
> configuration (which you have provided), but also by your cluster status
> (which you haven't), I'd ask that you create a CIB dump with "cibadmin
> -Q > /tmp/cib.xml", upload that to pastebin or whatever your favorite
> service is, and share the URL.
>
> Cheers,
> Florian
>
> --
> Need help with Pacemaker?
> http://www.hastexo.com/now
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs:
> http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20111012/a8921813/attachment-0003.html>


More information about the Pacemaker mailing list