[Pacemaker] R: R: Stonith external/sbd problem

Nicola Sabatelli n.sabatelli at ct.rupar.puglia.it
Thu Apr 29 14:23:40 UTC 2010


Yes, I create the disk and allocate the node, and I create a resource on
cluster in this way:

<clone id="cl_external_sbd_1">

        <meta_attributes id="cl_external_sbd_1-meta_attributes">

          <nvpair id="cl_external_sbd_1-meta_attributes-clone-max"
name="clone-max" value="2"/>

        </meta_attributes>

        <primitive class="stonith" type="external/sbd"
id="stonith_external_sbd_LOCK_LUN">

          <instance_attributes
id="stonith_external_sbd_LOCK_LUN-instance_attributes">

            <nvpair id="nvpair-stonith_external_sbd_LOCK_LUN-sbd_device"
name="sbd_device" value="/dev/mapper/mpath1p1"/>

          </instance_attributes>

          <operations id="stonith_external_sbd_LOCK_LUN-operations">

            <op id="op-stonith_external_sbd_LOCK_LUN-stop" interval="0"
name="stop" timeout="60"/>

            <op id="op-stonith_external_sbd_LOCK_LUN-monitor" interval="60"
name="monitor" start-delay="0" timeout="60"/>

            <op id="op-stonith_external_sbd_LOCK_LUN-start" interval="0"
name="start" timeout="60"/>

          </operations>

          <meta_attributes
id="stonith_external_sbd_LOCK_LUN-meta_attributes">

            <nvpair name="target-role"
id="stonith_external_sbd_LOCK_LUN-meta_attributes-target-role"
value="stopped"/>

          </meta_attributes>

        </primitive>

      </clone>

 

 

Ciao, Nicola.

  _____  

Da: Vit Pelcak [mailto:vpelcak at suse.cz] 
Inviato: giovedì 29 aprile 2010 16.08
A: pacemaker at oss.clusterlabs.org
Oggetto: Re: [Pacemaker] R: Stonith external/sbd problem

 

Also, it is needed to add stonith to cib:

crm configure primitive sbd_stonith stonith:external/sbd meta
target-role="Started" op monitor interval="15" timeout="15" start-delay="15"
params sbd_device="/dev/sda1"


Dne 29.4.2010 15:46, Nicola Sabatelli napsal(a): 

I have done exactly the configuration in the SBD_Fencing documentation.

That is:

/etc/sysconfig/sbd

SBD_DEVICE="/dev/mapper/mpath1p1"

SBD_OPTS="-W"

And I start the demon in this manner:

/usr/sbin/sbd -d /dev/mapper/mpath1p1 -D -W watch

Is correct?

 

Ciao, Nicola.

  _____  

Da: Vit Pelcak [mailto:vpelcak at suse.cz] 
Inviato: giovedì 29 aprile 2010 15.02
A: pacemaker at oss.clusterlabs.org
Oggetto: Re: [Pacemaker] Stonith external/sbd problem

 

cat /etc/sysconfig/sbd

SBD_DEVICE="/dev/sda1"
SBD_OPTS="-W"


sbd -d /dev/shared_disk create
sbd -d /dev/shared_disk allocate your_machine


Dne 29.4.2010 14:55, Michael Brown napsal(a): 

Oh, I forgot a piece: I had simular trouble until I actually properly
started sbd and then it worked.

M.

  _____  

From: Michael Brown 
To: pacemaker at oss.clusterlabs.org 
Sent: Thu Apr 29 08:53:32 2010
Subject: Re: [Pacemaker] Stonith external/sbd problem 




I just set this up myself and it worked fine for me.

Did you follow the guide? You need to configure the sbd daemon to run on
bootup with appropriate options before external/sbd can use it.

M.

  _____  

From: Nicola Sabatelli 
To: pacemaker at oss.clusterlabs.org 
Sent: Thu Apr 29 08:47:04 2010
Subject: [Pacemaker] Stonith external/sbd problem 


I have a problem with STONITH plugin external/sbd.

I have configured the system in according to directive that I find at url
http://www.linux-ha.org/wiki/SBD_Fencing, and the device that I use is
configured with multipath software because this disk is residend on a
storage system.

I have create a resurse on my cluster using clove directive.

But when I try to start the resurse I have these errors:

 

from ha-log file:

 

Apr 29 14:37:51 clover-h stonithd: [16811]: info: external_run_cmd: Calling
'/usr/lib64/stonith/plugins/external/sbd status' returned 256

Apr 29 14:37:51 clover-h stonithd: [16811]: CRIT: external_status: 'sbd
status' failed with rc 256

Apr 29 14:37:51 clover-h stonithd: [10615]: WARN: start
stonith_external_sbd_LOCK_LUN:0 failed, because its hostlist is empty

 

from crm_verify:

 

crm_verify[18607]: 2010/04/29_14:39:27 info: main: =#=#=#=#= Getting XML
=#=#=#=#=

crm_verify[18607]: 2010/04/29_14:39:27 info: main: Reading XML from: live
cluster

crm_verify[18607]: 2010/04/29_14:39:27 notice: unpack_config: On loss of CCM
Quorum: Ignore

crm_verify[18607]: 2010/04/29_14:39:27 info: unpack_config: Node scores:
'red' = -INFINITY, 'yellow' = 0, 'green' = 0

crm_verify[18607]: 2010/04/29_14:39:27 info: determine_online_status: Node
clover-a.rsr.rupar.puglia.it is online

crm_verify[18607]: 2010/04/29_14:39:27 WARN: unpack_rsc_op: Processing
failed op stonith_external_sbd_LOCK_LUN:1_start_0 on
clover-a.rsr.rupar.puglia.it: unknown error (1)

crm_verify[18607]: 2010/04/29_14:39:27 info: find_clone: Internally renamed
stonith_external_sbd_LOCK_LUN:0 on clover-a.rsr.rupar.puglia.it to
stonith_external_sbd_LOCK_LUN:2 (ORPHAN)

crm_verify[18607]: 2010/04/29_14:39:27 info: determine_online_status: Node
clover-h.rsr.rupar.puglia.it is online

crm_verify[18607]: 2010/04/29_14:39:27 WARN: unpack_rsc_op: Processing
failed op stonith_external_sbd_LOCK_LUN:0_start_0 on
clover-h.rsr.rupar.puglia.it: unknown error (1)

crm_verify[18607]: 2010/04/29_14:39:27 notice: clone_print:  Master/Slave
Set: ms_drbd_1

crm_verify[18607]: 2010/04/29_14:39:27 notice: short_print:      Stopped: [
res_drbd_1:0 res_drbd_1:1 ]

crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print:
res_Filesystem_TEST        (ocf::heartbeat:Filesystem):    Stopped

crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print:
res_IPaddr2_ip_clover      (ocf::heartbeat:IPaddr2):       Stopped

crm_verify[18607]: 2010/04/29_14:39:27 notice: clone_print:  Clone Set:
cl_external_sbd_1

crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print:
stonith_external_sbd_LOCK_LUN:0       (stonith:external/sbd): Started
clover-h.rsr.rupar.puglia.it FAILED

crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print:
stonith_external_sbd_LOCK_LUN:1       (stonith:external/sbd): Started
clover-a.rsr.rupar.puglia.it FAILED

crm_verify[18607]: 2010/04/29_14:39:27 info: get_failcount:
cl_external_sbd_1 has failed 1000000 times on clover-h.rsr.rupar.puglia.it

crm_verify[18607]: 2010/04/29_14:39:27 WARN: common_apply_stickiness:
Forcing cl_external_sbd_1 away from clover-h.rsr.rupar.puglia.it after
1000000 failures (max=1000000)

crm_verify[18607]: 2010/04/29_14:39:27 info: get_failcount:
cl_external_sbd_1 has failed 1000000 times on clover-a.rsr.rupar.puglia.it

crm_verify[18607]: 2010/04/29_14:39:27 WARN: common_apply_stickiness:
Forcing cl_external_sbd_1 away from clover-a.rsr.rupar.puglia.it after
1000000 failures (max=1000000)

crm_verify[18607]: 2010/04/29_14:39:27 info: native_merge_weights:
ms_drbd_1: Rolling back scores from res_Filesystem_TEST

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource
res_drbd_1:0 cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource
res_drbd_1:1 cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 info: native_merge_weights:
ms_drbd_1: Rolling back scores from res_Filesystem_TEST

crm_verify[18607]: 2010/04/29_14:39:27 info: master_color: ms_drbd_1:
Promoted 0 instances of a possible 1 to master

crm_verify[18607]: 2010/04/29_14:39:27 info: master_color: ms_drbd_1:
Promoted 0 instances of a possible 1 to master

crm_verify[18607]: 2010/04/29_14:39:27 info: native_merge_weights:
res_Filesystem_TEST: Rolling back scores from res_IPaddr2_ip_clover

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource
res_Filesystem_TEST cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource
res_IPaddr2_ip_clover cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource
stonith_external_sbd_LOCK_LUN:0 cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource
stonith_external_sbd_LOCK_LUN:1 cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource
res_drbd_1:0  (Stopped)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource
res_drbd_1:1  (Stopped)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource
res_Filesystem_TEST   (Stopped)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource
res_IPaddr2_ip_clover (Stopped)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Stop resource
stonith_external_sbd_LOCK_LUN:0        (clover-h.rsr.rupar.puglia.it)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Stop resource
stonith_external_sbd_LOCK_LUN:1        (clover-a.rsr.rupar.puglia.it)

Warnings found during check: config may not be valid

 

and from crm_mon:

 

============

Last updated: Thu Apr 29 14:39:57 2010

Stack: Heartbeat

Current DC: clover-h.rsr.rupar.puglia.it
(e39bb201-2a6f-457a-a308-be6bfe71309c) - partition with quorum

Version: 1.0.8-9881a7350d6182bae9e8e557cf20a3cc5dac3ee7

2 Nodes configured, unknown expected votes

4 Resources configured.

============

 

Online: [ clover-h.rsr.rupar.puglia.it clover-a.rsr.rupar.puglia.it ]

 

 Clone Set: cl_external_sbd_1

     stonith_external_sbd_LOCK_LUN:0    (stonith:external/sbd): Started
clover-h.rsr.rupar.puglia.it FAILED

     stonith_external_sbd_LOCK_LUN:1    (stonith:external/sbd): Started
clover-a.rsr.rupar.puglia.it FAILED

 

Operations:

* Node clover-a.rsr.rupar.puglia.it:

   stonith_external_sbd_LOCK_LUN:1: migration-threshold=1000000
fail-count=1000000

    + (24) start: rc=1 (unknown error)

* Node clover-h.rsr.rupar.puglia.it:

   stonith_external_sbd_LOCK_LUN:0: migration-threshold=1000000
fail-count=1000000

    + (25) start: rc=1 (unknown error)

 

Failed actions:

    stonith_external_sbd_LOCK_LUN:1_start_0
(node=clover-a.rsr.rupar.puglia.it, call=24, rc=1, status=complete): unknown
error

    stonith_external_sbd_LOCK_LUN:0_start_0
(node=clover-h.rsr.rupar.puglia.it, call=25, rc=1, status=complete): unknown
error

 

 

 

 

Ciao, Nicola.

 
 
_______________________________________________
Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
 
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf

 

 
 
_______________________________________________
Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
 
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf

 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20100429/b1346862/attachment-0002.htm>


More information about the Pacemaker mailing list