[Pacemaker] R: R: R: R: Stonith external/sbd problem

Nicola Sabatelli n.sabatelli at ct.rupar.puglia.it
Mon May 10 07:21:03 UTC 2010


Hi,

I have solved my problem.

I find a little problem in the script ‘/usr/lib64/stonith/plugins/external/sbd’ when it retrieve the hosts list.

I substitute this lines:

 

nodes=$(

if is_heartbeat; then

    crm_node -H -p

else

    crm_node -p

fi)

 

Whit these:

 

if is_heartbeat; then

    nodes=$(crm_node -H -p)

else

    nodes=$(crm_node -p)

fi

 

and now the resource ‘external/sbd’ function very well.

 

 

 

Best regards, Nicola.

 

  _____  

Da: Michael Brown [mailto:michael at netdirect.ca] 
Inviato: giovedì 29 aprile 2010 16.53
A: n.sabatelli at ct.rupar.puglia.it
Oggetto: Re: R: [Pacemaker] R: R: Stonith external/sbd problem

 

Hrm, my limited knowledge is exhausted. Good luck!

M.

  _____  

From: Nicola Sabatelli 
To: 'Michael Brown' 
Sent: Thu Apr 29 10:36:15 2010
Subject: R: [Pacemaker] R: R: Stonith external/sbd problem

The response to a query

/usr/sbin/sbd -d /dev/mapper/mpath1p1 list

is

0       clover-a.rsr.rupar.puglia.it    clear

1       clover-h.rsr.rupar.puglia.it    clear

 

 

Ciao, Nicola.

  _____  

Da: Michael Brown [mailto:michael at netdirect.ca] 
Inviato: giovedì 29 aprile 2010 16.33
A: The Pacemaker cluster resource manager
Cc: Nicola Sabatelli
Oggetto: Re: [Pacemaker] R: R: Stonith external/sbd problem

 

FWIW, here's my setup for sbd on shared storage:

in /etc/init.d/boot.local:
sbd -d /dev/disk/by-id/dm-uuid-part2-mpath-3600a0b8000266f7e000035414bd00428 -D -W watch

xenhost1:~ # sbd -d /dev/disk/by-id/dm-uuid-part2-mpath-3600a0b8000266f7e000035414bd00428 list
0       xenhost1        clear
1       xenhost2        clear

excerpt from 'crm configure show':
primitive sbd stonith:external/sbd \
        operations $id="sbd-operations" \
        op monitor interval="15" timeout="15" start-delay="15" \
        params sbd_device="/dev/disk/by-id/dm-uuid-part2-mpath-3600a0b8000266f7e000035414bd00428"
clone sbd-clone sbd \
        meta interleave="true"

What do you see if you run '/usr/sbin/sbd -d /dev/mapper/mpath1p1 list'?

M.

On 04/29/2010 10:23 AM, Nicola Sabatelli wrote: 

Yes, I create the disk and allocate the node, and I create a resource on cluster in this way:

<clone id="cl_external_sbd_1">

        <meta_attributes id="cl_external_sbd_1-meta_attributes">

          <nvpair id="cl_external_sbd_1-meta_attributes-clone-max" name="clone-max" value="2"/>

        </meta_attributes>

        <primitive class="stonith" type="external/sbd" id="stonith_external_sbd_LOCK_LUN">

          <instance_attributes id="stonith_external_sbd_LOCK_LUN-instance_attributes">

            <nvpair id="nvpair-stonith_external_sbd_LOCK_LUN-sbd_device" name="sbd_device" value="/dev/mapper/mpath1p1"/>

          </instance_attributes>

          <operations id="stonith_external_sbd_LOCK_LUN-operations">

            <op id="op-stonith_external_sbd_LOCK_LUN-stop" interval="0" name="stop" timeout="60"/>

            <op id="op-stonith_external_sbd_LOCK_LUN-monitor" interval="60" name="monitor" start-delay="0" timeout="60"/>

            <op id="op-stonith_external_sbd_LOCK_LUN-start" interval="0" name="start" timeout="60"/>

          </operations>

          <meta_attributes id="stonith_external_sbd_LOCK_LUN-meta_attributes">

            <nvpair name="target-role" id="stonith_external_sbd_LOCK_LUN-meta_attributes-target-role" value="stopped"/>

          </meta_attributes>

        </primitive>

      </clone>

 

 

Ciao, Nicola.

  _____  

Da: Vit Pelcak [mailto:vpelcak at suse.cz] 
Inviato: giovedì 29 aprile 2010 16.08
A: pacemaker at oss.clusterlabs.org
Oggetto: Re: [Pacemaker] R: Stonith external/sbd problem

 

Also, it is needed to add stonith to cib:

crm configure primitive sbd_stonith stonith:external/sbd meta target-role="Started" op monitor interval="15" timeout="15" start-delay="15" params sbd_device="/dev/sda1"


Dne 29.4.2010 15:46, Nicola Sabatelli napsal(a): 

I have done exactly the configuration in the SBD_Fencing documentation.

That is:

/etc/sysconfig/sbd

SBD_DEVICE="/dev/mapper/mpath1p1"

SBD_OPTS="-W"

And I start the demon in this manner:

/usr/sbin/sbd -d /dev/mapper/mpath1p1 -D -W watch

Is correct?

 

Ciao, Nicola.

  _____  

Da: Vit Pelcak [mailto:vpelcak at suse.cz] 
Inviato: giovedì 29 aprile 2010 15.02
A: pacemaker at oss.clusterlabs.org
Oggetto: Re: [Pacemaker] Stonith external/sbd problem

 

cat /etc/sysconfig/sbd

SBD_DEVICE="/dev/sda1"
SBD_OPTS="-W"


sbd -d /dev/shared_disk create
sbd -d /dev/shared_disk allocate your_machine


Dne 29.4.2010 14:55, Michael Brown napsal(a): 

Oh, I forgot a piece: I had simular trouble until I actually properly started sbd and then it worked.

M.

  _____  

From: Michael Brown 
To: pacemaker at oss.clusterlabs.org 
Sent: Thu Apr 29 08:53:32 2010
Subject: Re: [Pacemaker] Stonith external/sbd problem 




I just set this up myself and it worked fine for me.

Did you follow the guide? You need to configure the sbd daemon to run on bootup with appropriate options before external/sbd can use it.

M.

  _____  

From: Nicola Sabatelli 
To: pacemaker at oss.clusterlabs.org 
Sent: Thu Apr 29 08:47:04 2010
Subject: [Pacemaker] Stonith external/sbd problem 


I have a problem with STONITH plugin external/sbd.

I have configured the system in according to directive that I find at url http://www.linux-ha.org/wiki/SBD_Fencing, and the device that I use is configured with multipath software because this disk is residend on a storage system.

I have create a resurse on my cluster using clove directive.

But when I try to start the resurse I have these errors:

 

from ha-log file:

 

Apr 29 14:37:51 clover-h stonithd: [16811]: info: external_run_cmd: Calling '/usr/lib64/stonith/plugins/external/sbd status' returned 256

Apr 29 14:37:51 clover-h stonithd: [16811]: CRIT: external_status: 'sbd status' failed with rc 256

Apr 29 14:37:51 clover-h stonithd: [10615]: WARN: start stonith_external_sbd_LOCK_LUN:0 failed, because its hostlist is empty

 

from crm_verify:

 

crm_verify[18607]: 2010/04/29_14:39:27 info: main: =#=#=#=#= Getting XML =#=#=#=#=

crm_verify[18607]: 2010/04/29_14:39:27 info: main: Reading XML from: live cluster

crm_verify[18607]: 2010/04/29_14:39:27 notice: unpack_config: On loss of CCM Quorum: Ignore

crm_verify[18607]: 2010/04/29_14:39:27 info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0

crm_verify[18607]: 2010/04/29_14:39:27 info: determine_online_status: Node clover-a.rsr.rupar.puglia.it is online

crm_verify[18607]: 2010/04/29_14:39:27 WARN: unpack_rsc_op: Processing failed op stonith_external_sbd_LOCK_LUN:1_start_0 on clover-a.rsr.rupar.puglia.it: unknown error (1)

crm_verify[18607]: 2010/04/29_14:39:27 info: find_clone: Internally renamed stonith_external_sbd_LOCK_LUN:0 on clover-a.rsr.rupar.puglia.it to stonith_external_sbd_LOCK_LUN:2 (ORPHAN)

crm_verify[18607]: 2010/04/29_14:39:27 info: determine_online_status: Node clover-h.rsr.rupar.puglia.it is online

crm_verify[18607]: 2010/04/29_14:39:27 WARN: unpack_rsc_op: Processing failed op stonith_external_sbd_LOCK_LUN:0_start_0 on clover-h.rsr.rupar.puglia.it: unknown error (1)

crm_verify[18607]: 2010/04/29_14:39:27 notice: clone_print:  Master/Slave Set: ms_drbd_1

crm_verify[18607]: 2010/04/29_14:39:27 notice: short_print:      Stopped: [ res_drbd_1:0 res_drbd_1:1 ]

crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print: res_Filesystem_TEST        (ocf::heartbeat:Filesystem):    Stopped

crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print: res_IPaddr2_ip_clover      (ocf::heartbeat:IPaddr2):       Stopped

crm_verify[18607]: 2010/04/29_14:39:27 notice: clone_print:  Clone Set: cl_external_sbd_1

crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print:      stonith_external_sbd_LOCK_LUN:0       (stonith:external/sbd): Started clover-h.rsr.rupar.puglia.it FAILED

crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print:      stonith_external_sbd_LOCK_LUN:1       (stonith:external/sbd): Started clover-a.rsr.rupar.puglia.it FAILED

crm_verify[18607]: 2010/04/29_14:39:27 info: get_failcount: cl_external_sbd_1 has failed 1000000 times on clover-h.rsr.rupar.puglia.it

crm_verify[18607]: 2010/04/29_14:39:27 WARN: common_apply_stickiness: Forcing cl_external_sbd_1 away from clover-h.rsr.rupar.puglia.it after 1000000 failures (max=1000000)

crm_verify[18607]: 2010/04/29_14:39:27 info: get_failcount: cl_external_sbd_1 has failed 1000000 times on clover-a.rsr.rupar.puglia.it

crm_verify[18607]: 2010/04/29_14:39:27 WARN: common_apply_stickiness: Forcing cl_external_sbd_1 away from clover-a.rsr.rupar.puglia.it after 1000000 failures (max=1000000)

crm_verify[18607]: 2010/04/29_14:39:27 info: native_merge_weights: ms_drbd_1: Rolling back scores from res_Filesystem_TEST

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource res_drbd_1:0 cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource res_drbd_1:1 cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 info: native_merge_weights: ms_drbd_1: Rolling back scores from res_Filesystem_TEST

crm_verify[18607]: 2010/04/29_14:39:27 info: master_color: ms_drbd_1: Promoted 0 instances of a possible 1 to master

crm_verify[18607]: 2010/04/29_14:39:27 info: master_color: ms_drbd_1: Promoted 0 instances of a possible 1 to master

crm_verify[18607]: 2010/04/29_14:39:27 info: native_merge_weights: res_Filesystem_TEST: Rolling back scores from res_IPaddr2_ip_clover

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource res_Filesystem_TEST cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource res_IPaddr2_ip_clover cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource stonith_external_sbd_LOCK_LUN:0 cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource stonith_external_sbd_LOCK_LUN:1 cannot run anywhere

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource res_drbd_1:0  (Stopped)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource res_drbd_1:1  (Stopped)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource res_Filesystem_TEST   (Stopped)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource res_IPaddr2_ip_clover (Stopped)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Stop resource stonith_external_sbd_LOCK_LUN:0        (clover-h.rsr.rupar.puglia.it)

crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Stop resource stonith_external_sbd_LOCK_LUN:1        (clover-a.rsr.rupar.puglia.it)

Warnings found during check: config may not be valid

 

and from crm_mon:

 

============

Last updated: Thu Apr 29 14:39:57 2010

Stack: Heartbeat

Current DC: clover-h.rsr.rupar.puglia.it (e39bb201-2a6f-457a-a308-be6bfe71309c) - partition with quorum

Version: 1.0.8-9881a7350d6182bae9e8e557cf20a3cc5dac3ee7

2 Nodes configured, unknown expected votes

4 Resources configured.

============

 

Online: [ clover-h.rsr.rupar.puglia.it clover-a.rsr.rupar.puglia.it ]

 

 Clone Set: cl_external_sbd_1

     stonith_external_sbd_LOCK_LUN:0    (stonith:external/sbd): Started clover-h.rsr.rupar.puglia.it FAILED

     stonith_external_sbd_LOCK_LUN:1    (stonith:external/sbd): Started clover-a.rsr.rupar.puglia.it FAILED

 

Operations:

* Node clover-a.rsr.rupar.puglia.it:

   stonith_external_sbd_LOCK_LUN:1: migration-threshold=1000000 fail-count=1000000

    + (24) start: rc=1 (unknown error)

* Node clover-h.rsr.rupar.puglia.it:

   stonith_external_sbd_LOCK_LUN:0: migration-threshold=1000000 fail-count=1000000

    + (25) start: rc=1 (unknown error)

 

Failed actions:

    stonith_external_sbd_LOCK_LUN:1_start_0 (node=clover-a.rsr.rupar.puglia.it, call=24, rc=1, status=complete): unknown error

    stonith_external_sbd_LOCK_LUN:0_start_0 (node=clover-h.rsr.rupar.puglia.it, call=25, rc=1, status=complete): unknown error

 

 

 

 

Ciao, Nicola.

 
 
_______________________________________________
Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
 
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf

 

 
 
_______________________________________________
Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
 
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf

 

 
 
_______________________________________________
Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
 
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf





-- 
Michael Brown               | `One of the main causes of the fall of
Systems Consultant          | the Roman Empire was that, lacking zero,
Net Direct Inc.             | they had no way to indicate successful
☎: +1 519 883 1172 x5106    | termination of their C programs.' - Firth
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20100510/8abc297c/attachment-0001.html>


More information about the Pacemaker mailing list