[ClusterLabs] Issue with Stonith Resource parameters

emmanuel segura emi2fast at gmail.com
Wed Mar 9 02:55:15 EST 2016


I think you should give the parameters to the stonith agent, anyway
show your config.

2016-03-09 5:29 GMT+01:00 vija ar <vjav78 at gmail.com>:
> I have configured SLEHA cluster on cisco ucs boxes with ipmi configured, i
> have tested IPMI using impitool, however ipmitool to function neatly i have
> to pass parameter -y i.e. <hex key> along with username and password,
>
> however  to configure stonith there is no parameter in pacemaker to pass
> <hex key>? and due to which stonith is failing
>
> can you please let me know if there is any way to add it or is this a bug?
>
> *******************
>
>
>
> Mar  9 00:26:28 server02 stonith: external_status: 'ipmi status' failed with
> rc 1
> Mar  9 00:26:28 server02 stonith: external/ipmi device not accessible.
> Mar  9 00:26:28 server02 stonith-ng[99114]:   notice: log_operation:
> Operation 'monitor' [99200] for device 'STONITH-server02' returned: -201
> (Generic Pacemaker error)
> Mar  9 00:26:28 server02 stonith-ng[99114]:  warning: log_operation:
> STONITH-server02:99200 [ Performing: stonith -t external/ipmi -S ]
> Mar  9 00:26:28 server02 stonith-ng[99114]:  warning: log_operation:
> STONITH-server02:99200 [ logd is not runningfailed:  1 ]
> Mar  9 00:26:28 server02 crmd[99118]:    error: process_lrm_event: LRM
> operation STONITH-server02_start_0 (call=13, status=4, cib-update=13,
> confirmed=true) Error
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch: Update
> relayed from server01
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update: Sent
> update 35: fail-count-STONITH-server02=INFINITY
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch: Update
> relayed from server01
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: last-failure-STONITH-server02
> (1457463388)
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update: Sent
> update 37: last-failure-STONITH-server02=1457463388
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch: Update
> relayed from server01
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update: Sent
> update 39: fail-count-STONITH-server02=INFINITY
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch: Update
> relayed from server01
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: last-failure-STONITH-server02
> (1457463388)
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update: Sent
> update 41: last-failure-STONITH-server02=1457463388
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch: Update
> relayed from server01
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update: Sent
> update 43: fail-count-STONITH-server02=INFINITY
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch: Update
> relayed from server01
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: last-failure-STONITH-server02
> (1457463388)
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update: Sent
> update 45: last-failure-STONITH-server02=1457463388
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch: Update
> relayed from server01
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: fail-count-STONITH-server02 (INFINITY)
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update: Sent
> update 47: fail-count-STONITH-server02=INFINITY
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_cs_dispatch: Update
> relayed from server01
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: last-failure-STONITH-server02
> (1457463388)
> Mar  9 00:26:28 server02 attrd[99116]:   notice: attrd_perform_update: Sent
> update 49: last-failure-STONITH-server02=1457463388
> Mar  9 00:26:28 server02 crmd[99118]:   notice: process_lrm_event: LRM
> operation STONITH-server02_stop_0 (call=14, rc=0, cib-update=14,
> confirmed=true) ok
> Mar  9 00:26:28 server01 crmd[16809]:  warning: status_from_rc: Action 9
> (STONITH-server02_start_0) on server02 failed (target: 0 vs. rc: 1): Error
> Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount: Updating
> failcount for STONITH-server02 on server02 after failed start: rc=1
> (update=INFINITY, time=1457463388)
> Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount: Updating
> failcount for STONITH-server02 on server02 after failed start: rc=1
> (update=INFINITY, time=1457463388)
> Mar  9 00:26:28 server01 crmd[16809]:  warning: status_from_rc: Action 9
> (STONITH-server02_start_0) on server02 failed (target: 0 vs. rc: 1): Error
> Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount: Updating
> failcount for STONITH-server02 on server02 after failed start: rc=1
> (update=INFINITY, time=1457463388)
> Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount: Updating
> failcount for STONITH-server02 on server02 after failed start: rc=1
> (update=INFINITY, time=1457463388)
> Mar  9 00:26:28 server01 stonith: external_status: 'ipmi status' failed with
> rc 1
> Mar  9 00:26:28 server01 stonith: external/ipmi device not accessible.
> Mar  9 00:26:28 server01 stonith-ng[16805]:   notice: log_operation:
> Operation 'monitor' [16891] for device 'STONITH-server01' returned: -201
> (Generic Pacemaker error)
> Mar  9 00:26:28 server01 stonith-ng[16805]:  warning: log_operation:
> STONITH-server01:16891 [ Performing: stonith -t external/ipmi -S ]
> Mar  9 00:26:28 server01 stonith-ng[16805]:  warning: log_operation:
> STONITH-server01:16891 [ logd is not runningfailed:  1 ]
> Mar  9 00:26:28 server01 crmd[16809]:    error: process_lrm_event: LRM
> operation STONITH-server01_start_0 (call=13, status=4, cib-update=49,
> confirmed=true) Error
> Mar  9 00:26:28 server01 crmd[16809]:  warning: status_from_rc: Action 7
> (STONITH-server01_start_0) on server01 failed (target: 0 vs. rc: 1): Error
> Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount: Updating
> failcount for STONITH-server01 on server01 after failed start: rc=1
> (update=INFINITY, time=1457463388)
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY)
> Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount: Updating
> failcount for STONITH-server01 on server01 after failed start: rc=1
> (update=INFINITY, time=1457463388)
> Mar  9 00:26:28 server01 crmd[16809]:  warning: status_from_rc: Action 7
> (STONITH-server01_start_0) on server01 failed (target: 0 vs. rc: 1): Error
> Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount: Updating
> failcount for STONITH-server01 on server01 after failed start: rc=1
> (update=INFINITY, time=1457463388)
> Mar  9 00:26:28 server01 crmd[16809]:  warning: update_failcount: Updating
> failcount for STONITH-server01 on server01 after failed start: rc=1
> (update=INFINITY, time=1457463388)
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update: Sent
> update 47: fail-count-STONITH-server01=INFINITY
> Mar  9 00:26:28 server01 crmd[16809]:   notice: run_graph: Transition 3
> (Complete=5, Pending=0, Fired=0, Skipped=2, Incomplete=0,
> Source=/var/lib/pacemaker/pengine/pe-input-70.bz2): Stopped
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: last-failure-STONITH-server01
> (1457463388)
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update: Sent
> update 49: last-failure-STONITH-server01=1457463388
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: unpack_config: On loss of
> CCM Quorum: Ignore
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server02 on server01: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server01 on server01: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server01 on server01: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server02 on server02: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server02 on server02: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server01 on server02: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server02 away from server01 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server01 away from server02 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server02 away from server02 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Recover
> STONITH-server01    (Started server01)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> STONITH-server02    (server02)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: process_pe_message:
> Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-71.bz2
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update: Sent
> update 51: fail-count-STONITH-server01=INFINITY
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: last-failure-STONITH-server01
> (1457463388)
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update: Sent
> update 53: last-failure-STONITH-server01=1457463388
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: fail-count-STONITH-server01 (INFINITY)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: unpack_config: On loss of
> CCM Quorum: Ignore
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server02 on server01: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server01 on server01: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server01 on server01: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server02 on server02: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server02 on server02: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server01 on server02: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server01 away from server01 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server02 away from server01 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server01 away from server02 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server02 away from server02 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> STONITH-server01    (server01)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> STONITH-server02    (server02)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: process_pe_message:
> Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-72.bz2
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update: Sent
> update 55: fail-count-STONITH-server01=INFINITY
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: last-failure-STONITH-server01
> (1457463388)
> Mar  9 00:26:28 server01 attrd[16807]:   notice: attrd_perform_update: Sent
> update 57: last-failure-STONITH-server01=1457463388
> Mar  9 00:26:28 server01 pengine[16808]:   notice: unpack_config: On loss of
> CCM Quorum: Ignore
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server02 on server01: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server01 on server01: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server01 on server01: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server02 on server02: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server02 on server02: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: unpack_rsc_op_failure:
> Processing failed op start for STONITH-server01 on server02: unknown error
> (1)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server01 away from server01 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server02 away from server01 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server01 away from server02 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:  warning: common_apply_stickiness:
> Forcing STONITH-server02 away from server02 after 1000000 failures (max=3)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> STONITH-server01    (server01)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: LogActions: Stop
> STONITH-server02    (server02)
> Mar  9 00:26:28 server01 pengine[16808]:   notice: process_pe_message:
> Calculated Transition 6: /var/lib/pacemaker/pengine/pe-input-73.bz2
> Mar  9 00:26:28 server01 crmd[16809]:   notice: do_te_invoke: Processing
> graph 6 (ref=pe_calc-dc-1457463388-32) derived from
> /var/lib/pacemaker/pengine/pe-input-73.bz2
> Mar  9 00:26:28 server01 crmd[16809]:   notice: te_rsc_command: Initiating
> action 1: stop STONITH-server01_stop_0 on server01 (local)
> Mar  9 00:26:28 server01 crmd[16809]:   notice: te_rsc_command: Initiating
> action 2: stop STONITH-server02_stop_0 on server02
> Mar  9 00:26:28 server01 crmd[16809]:   notice: process_lrm_event: LRM
> operation STONITH-server01_stop_0 (call=14, rc=0, cib-update=55,
> confirmed=true) ok
> Mar  9 00:26:28 server01 crmd[16809]:   notice: run_graph: Transition 6
> (Complete=3, Pending=0, Fired=0, Skipped=0, Incomplete=0,
> Source=/var/lib/pacemaker/pengine/pe-input-73.bz2): Complete
> Mar  9 00:26:28 server01 crmd[16809]:   notice: do_state_transition: State
> transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
> cause=C_FSA_INTERNAL origin=notify_crmd ]
>
>
>
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>



-- 
  .~.
  /V\
 //  \\
/(   )\
^`~'^




More information about the Users mailing list