[ClusterLabs] Issue with Stonith Resource parameters

vija ar vjav78 at gmail.com
Wed Mar 9 03:19:45 EST 2016


here is the config ....


<cib epoch="20" num_updates="0" admin_epoch="0"
validate-with="pacemaker-1.2" cib-last-written="Wed Mar  9 00:56:57 2016"
update-origin="server02" update-client="cibadmin" update-user="hacluster"
crm_feature_set="3.0.8" have-quorum="1" dc-uuid="server01">
  <configuration>
    <crm_config>
      <cluster_property_set id="cib-bootstrap-options">
        <nvpair name="stonith-enabled" value="true"
id="cib-bootstrap-options-stonith-enabled"/>
        <nvpair name="no-quorum-policy" value="ignore"
id="cib-bootstrap-options-no-quorum-policy"/>
        <nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
value="1.1.11-3ca8c3b"/>
        <nvpair id="cib-bootstrap-options-cluster-infrastructure"
name="cluster-infrastructure" value="classic openais (with plugin)"/>
        <nvpair id="cib-bootstrap-options-expected-quorum-votes"
name="expected-quorum-votes" value="2"/>
        <nvpair name="stonith-action" value="reboot"
id="cib-bootstrap-options-stonith-action"/>
        <nvpair name="stonith-timeout" value="150s"
id="cib-bootstrap-options-stonith-timeout"/>
      </cluster_property_set>
    </crm_config>
    <nodes>
      <node id="server02" uname="server02"/>
      <node id="server01" uname="server01"/>
    </nodes>
    <resources>
      <primitive id="STONITH-server01" class="stonith" type="external/ipmi">
        <operations>
          <op name="monitor" interval="0" timeout="60s"
id="STONITH-server01-monitor-0"/>
          <op name="monitor" interval="300s" timeout="60s"
on-fail="restart" id="STONITH-server01-monitor-300s"/>
          <op name="start" interval="0" timeout="60s" on-fail="restart"
id="STONITH-server01-start-0"/>
        </operations>
        <instance_attributes id="STONITH-server01-instance_attributes">
          <nvpair name="hostname" value="server01"
id="STONITH-server01-instance_attributes-hostname"/>
          <nvpair name="ipaddr" value="server01-ipmi"
id="STONITH-server01-instance_attributes-ipaddr"/>
          <nvpair name="userid" value="administrator"
id="STONITH-server01-instance_attributes-userid"/>
          <nvpair name="passwd" value="To12"
id="STONITH-server01-instance_attributes-passwd"/>
          <nvpair name="interface" value="lanplus"
id="STONITH-server01-instance_attributes-interface"/>
        </instance_attributes>
      </primitive>
      <primitive id="STONITH-server02" class="stonith" type="external/ipmi">
        <operations>
          <op name="monitor" interval="0" timeout="60s"
id="STONITH-server02-monitor-0"/>
          <op name="monitor" interval="300s" timeout="60s"
on-fail="restart" id="STONITH-server02-monitor-300s"/>
          <op name="start" interval="0" timeout="60s" on-fail="restart"
id="STONITH-server02-start-0"/>
        </operations>
        <instance_attributes id="STONITH-server02-instance_attributes">
          <nvpair name="hostname" value="server02"
id="STONITH-server02-instance_attributes-hostname"/>
          <nvpair name="ipaddr" value="server02-ipmi"
id="STONITH-server02-instance_attributes-ipaddr"/>
          <nvpair name="userid" value="administrator"
id="STONITH-server02-instance_attributes-userid"/>
          <nvpair name="passwd" value="To12"
id="STONITH-server02-instance_attributes-passwd"/>
          <nvpair name="interface" value="lanplus"
id="STONITH-server02-instance_attributes-interface"/>
        </instance_attributes>
      </primitive>
      <primitive id="VIRTUAL-IP" class="ocf" provider="heartbeat"
type="IPaddr2">
        <instance_attributes id="VIRTUAL-IP-instance_attributes">
          <nvpair name="ip" value="10.0.0.44"
id="VIRTUAL-IP-instance_attributes-ip"/>
        </instance_attributes>
        <operations>
          <op name="monitor" timeout="20s" interval="10s"
id="VIRTUAL-IP-monitor-10s"/>
        </operations>
        <meta_attributes id="VIRTUAL-IP-meta_attributes">
          <nvpair name="is-managed" value="true"
id="VIRTUAL-IP-meta_attributes-is-managed"/>
          <nvpair name="target-role" value="Started"
id="VIRTUAL-IP-meta_attributes-target-role"/>
        </meta_attributes>
      </primitive>
    </resources>
    <constraints>
      <rsc_location id="LOC_STONITH_server01" rsc="STONITH-server01"
score="INFINITY" node="server02"/>
      <rsc_location id="LOC_STONITH_server02" rsc="STONITH-server02"
score="INFINITY" node="server01"/>
    </constraints>
    <rsc_defaults>
      <meta_attributes id="rsc-options">
        <nvpair name="migration-threshold" value="5000"
id="rsc-options-migration-threshold"/>
        <nvpair name="resource-stickiness" value="1000"
id="rsc-options-resource-stickiness"/>
      </meta_attributes>
    </rsc_defaults>
    <op_defaults>
      <meta_attributes id="op-options">
        <nvpair name="timeout" value="600" id="op-options-timeout"/>
        <nvpair name="record-pending" value="false"
id="op-options-record-pending"/>
      </meta_attributes>
    </op_defaults>
  </configuration>
</cib>


On Wed, Mar 9, 2016 at 1:25 PM, emmanuel segura <emi2fast at gmail.com> wrote:

> I think you should give the parameters to the stonith agent, anyway
> show your config.
>
> 2016-03-09 5:29 GMT+01:00 vija ar <vjav78 at gmail.com>:
> > I have configured SLEHA cluster on cisco ucs boxes with ipmi configured,
> i
> > have tested IPMI using impitool, however ipmitool to function neatly i
> have
> > to pass parameter -y i.e. <hex key> along with username and password,
> >
> > however  to configure stonith there is no parameter in pacemaker to pass
> > <hex key>? and due to which stonith is failing
> >
> > can you please let me know if there is any way to add it or is this a
> bug?
> >
> > *******************
> >
> >
> >
> > Mar  9 00:26:28 server02 stonith: external_status: 'ipmi status' failed
> with
> > rc 1
> > Mar  9 00:26:28 server02 stonith: external/ipmi device not accessible.
> > Mar  9 00:26:28 server02 stonith-ng[99114]:   notice: log_operation:
> > Operation 'monitor' [99200] for device 'STONITH-server02' returned: -201
> > (Generic Pacemaker error)
> > Mar  9 00:26:28 server02 stonith-ng[99114]:  warning: log_operation:
> > STONITH-server02:99200 [ Performing: stonith -t external/ipmi -S ]
> > Mar  9 00:26:28 server02 stonith-ng[99114]:  warning: log_operation:
> > STONITH-server02:99200 [ logd is not runningfailed:  1 ]
> > Mar  9 00:26:28 server02 crmd[99118]:    error: process_lrm_event: LRM
> > operation STONITH-server02_start_0 (call=13, status=4, cib-update=13,
> > confirmed=true) Error
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update:
> Sent
> > update 35: fail-count-STONITH-server02=INFINITY
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server02
> > (1457463388)
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update:
> Sent
> > update 37: last-failure-STONITH-server02=1457463388
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update:
> Sent
> > update 39: fail-count-STONITH-server02=INFINITY
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server02
> > (1457463388)
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update:
> Sent
> > update 41: last-failure-STONITH-server02=1457463388
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update:
> Sent
> > update 43: fail-count-STONITH-server02=INFINITY
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server02
> > (1457463388)
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update:
> Sent
> > update 45: last-failure-STONITH-server02=1457463388
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update:
> Sent
> > update 47: fail-count-STONITH-server02=INFINITY
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch:
> Update
> > relayed from server01
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server02
> > (1457463388)
> > Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update:
> Sent
> > update 49: last-failure-STONITH-server02=1457463388
> > Mar  9 00:26:28 server02 crmd[99118]:   notice: process_lrm_event: LRM
> > operation STONITH-server02_stop_0 (call=14, rc=0, cib-update=14,
> > confirmed=true) ok
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: status_from_rc: Action 9
> > (STONITH-server02_start_0) on server02 failed (target: 0 vs. rc: 1):
> Error
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount:
> Updating
> > failcount for STONITH-server02 on server02 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount:
> Updating
> > failcount for STONITH-server02 on server02 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: status_from_rc: Action 9
> > (STONITH-server02_start_0) on server02 failed (target: 0 vs. rc: 1):
> Error
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount:
> Updating
> > failcount for STONITH-server02 on server02 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount:
> Updating
> > failcount for STONITH-server02 on server02 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar  9 00:26:28 server01 stonith: external_status: 'ipmi status' failed
> with
> > rc 1
> > Mar  9 00:26:28 server01 stonith: external/ipmi device not accessible.
> > Mar  9 00:26:28 server01 stonith-ng[16805]:   notice: log_operation:
> > Operation 'monitor' [16891] for device 'STONITH-server01' returned: -201
> > (Generic Pacemaker error)
> > Mar  9 00:26:28 server01 stonith-ng[16805]:  warning: log_operation:
> > STONITH-server01:16891 [ Performing: stonith -t external/ipmi -S ]
> > Mar  9 00:26:28 server01 stonith-ng[16805]:  warning: log_operation:
> > STONITH-server01:16891 [ logd is not runningfailed:  1 ]
> > Mar  9 00:26:28 server01 crmd[16809]:    error: process_lrm_event: LRM
> > operation STONITH-server01_start_0 (call=13, status=4, cib-update=49,
> > confirmed=true) Error
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: status_from_rc: Action 7
> > (STONITH-server01_start_0) on server01 failed (target: 0 vs. rc: 1):
> Error
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount:
> Updating
> > failcount for STONITH-server01 on server01 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY)
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount:
> Updating
> > failcount for STONITH-server01 on server01 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: status_from_rc: Action 7
> > (STONITH-server01_start_0) on server01 failed (target: 0 vs. rc: 1):
> Error
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount:
> Updating
> > failcount for STONITH-server01 on server01 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount:
> Updating
> > failcount for STONITH-server01 on server01 after failed start: rc=1
> > (update=INFINITY, time=1457463388)
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update:
> Sent
> > update 47: fail-count-STONITH-server01=INFINITY
> > Mar  9 00:26:28 server01 crmd[16809]:   notice: run_graph: Transition 3
> > (Complete=5, Pending=0, Fired=0, Skipped=2, Incomplete=0,
> > Source=/var/lib/pacemaker/pengine/pe-input-70.bz2): Stopped
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server01
> > (1457463388)
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update:
> Sent
> > update 49: last-failure-STONITH-server01=1457463388
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: unpack_config: On
> loss of
> > CCM Quorum: Ignore
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server01: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server02: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server01 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server02 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server02 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Recover
> > STONITH-server01    (Started server01)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> > STONITH-server02    (server02)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: process_pe_message:
> > Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-71.bz2
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update:
> Sent
> > update 51: fail-count-STONITH-server01=INFINITY
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server01
> > (1457463388)
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update:
> Sent
> > update 53: last-failure-STONITH-server01=1457463388
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: unpack_config: On
> loss of
> > CCM Quorum: Ignore
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server01: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server02: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server01 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server01 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server02 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server02 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> > STONITH-server01    (server01)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> > STONITH-server02    (server02)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: process_pe_message:
> > Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-72.bz2
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update:
> Sent
> > update 55: fail-count-STONITH-server01=INFINITY
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> > Sending flush op to all hosts for: last-failure-STONITH-server01
> > (1457463388)
> > Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update:
> Sent
> > update 57: last-failure-STONITH-server01=1457463388
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: unpack_config: On
> loss of
> > CCM Quorum: Ignore
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server01: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server01: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server02 on server02: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> > Processing failed op start for STONITH-server01 on server02: unknown
> error
> > (1)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server01 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server01 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server01 away from server02 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:  warning:
> common_apply_stickiness:
> > Forcing STONITH-server02 away from server02 after 1000000 failures
> (max=3)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> > STONITH-server01    (server01)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> > STONITH-server02    (server02)
> > Mar  9 00:26:28 server01 pengine[16808]:   notice: process_pe_message:
> > Calculated Transition 6: /var/lib/pacemaker/pengine/pe-input-73.bz2
> > Mar  9 00:26:28 server01 crmd[16809]:   notice: do_te_invoke: Processing
> > graph 6 (ref=pe_calc-dc-1457463388-32) derived from
> > /var/lib/pacemaker/pengine/pe-input-73.bz2
> > Mar  9 00:26:28 server01 crmd[16809]:   notice: te_rsc_command:
> Initiating
> > action 1: stop STONITH-server01_stop_0 on server01 (local)
> > Mar  9 00:26:28 server01 crmd[16809]:   notice: te_rsc_command:
> Initiating
> > action 2: stop STONITH-server02_stop_0 on server02
> > Mar  9 00:26:28 server01 crmd[16809]:   notice: process_lrm_event: LRM
> > operation STONITH-server01_stop_0 (call=14, rc=0, cib-update=55,
> > confirmed=true) ok
> > Mar  9 00:26:28 server01 crmd[16809]:   notice: run_graph: Transition 6
> > (Complete=3, Pending=0, Fired=0, Skipped=0, Incomplete=0,
> > Source=/var/lib/pacemaker/pengine/pe-input-73.bz2): Complete
> > Mar  9 00:26:28 server01 crmd[16809]:   notice: do_state_transition:
> State
> > transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
> > cause=C_FSA_INTERNAL origin=notify_crmd ]
> >
> >
> >
> > _______________________________________________
> > Users mailing list: Users at clusterlabs.org
> > http://clusterlabs.org/mailman/listinfo/users
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> >
>
>
>
> --
>   .~.
>   /V\
>  //  \\
> /(   )\
> ^`~'^
>
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20160309/14c15c92/attachment-0003.html>


More information about the Users mailing list