[Pacemaker] R: R: R: R: Stonith external/sbd problem
Dejan Muhamedagic
dejanmm at fastmail.fm
Mon May 10 10:39:03 UTC 2010
Hi,
On Mon, May 10, 2010 at 09:21:03AM +0200, Nicola Sabatelli wrote:
> Hi,
>
> I have solved my problem.
>
> I find a little problem in the script ‘/usr/lib64/stonith/plugins/external/sbd’ when it retrieve the hosts list.
>
> I substitute this lines:
>
>
>
> nodes=$(
>
> if is_heartbeat; then
>
> crm_node -H -p
>
> else
>
> crm_node -p
>
> fi)
>
>
>
> Whit these:
>
>
>
> if is_heartbeat; then
>
> nodes=$(crm_node -H -p)
>
> else
>
> nodes=$(crm_node -p)
>
> fi
Fixed now.
Cheers,
Dejan
>
>
> and now the resource ‘external/sbd’ function very well.
>
>
>
>
>
>
>
> Best regards, Nicola.
>
>
>
> _____
>
> Da: Michael Brown [mailto:michael at netdirect.ca]
> Inviato: giovedì 29 aprile 2010 16.53
> A: n.sabatelli at ct.rupar.puglia.it
> Oggetto: Re: R: [Pacemaker] R: R: Stonith external/sbd problem
>
>
>
> Hrm, my limited knowledge is exhausted. Good luck!
>
> M.
>
> _____
>
> From: Nicola Sabatelli
> To: 'Michael Brown'
> Sent: Thu Apr 29 10:36:15 2010
> Subject: R: [Pacemaker] R: R: Stonith external/sbd problem
>
> The response to a query
>
> /usr/sbin/sbd -d /dev/mapper/mpath1p1 list
>
> is
>
> 0 clover-a.rsr.rupar.puglia.it clear
>
> 1 clover-h.rsr.rupar.puglia.it clear
>
>
>
>
>
> Ciao, Nicola.
>
> _____
>
> Da: Michael Brown [mailto:michael at netdirect.ca]
> Inviato: giovedì 29 aprile 2010 16.33
> A: The Pacemaker cluster resource manager
> Cc: Nicola Sabatelli
> Oggetto: Re: [Pacemaker] R: R: Stonith external/sbd problem
>
>
>
> FWIW, here's my setup for sbd on shared storage:
>
> in /etc/init.d/boot.local:
> sbd -d /dev/disk/by-id/dm-uuid-part2-mpath-3600a0b8000266f7e000035414bd00428 -D -W watch
>
> xenhost1:~ # sbd -d /dev/disk/by-id/dm-uuid-part2-mpath-3600a0b8000266f7e000035414bd00428 list
> 0 xenhost1 clear
> 1 xenhost2 clear
>
> excerpt from 'crm configure show':
> primitive sbd stonith:external/sbd \
> operations $id="sbd-operations" \
> op monitor interval="15" timeout="15" start-delay="15" \
> params sbd_device="/dev/disk/by-id/dm-uuid-part2-mpath-3600a0b8000266f7e000035414bd00428"
> clone sbd-clone sbd \
> meta interleave="true"
>
> What do you see if you run '/usr/sbin/sbd -d /dev/mapper/mpath1p1 list'?
>
> M.
>
> On 04/29/2010 10:23 AM, Nicola Sabatelli wrote:
>
> Yes, I create the disk and allocate the node, and I create a resource on cluster in this way:
>
> <clone id="cl_external_sbd_1">
>
> <meta_attributes id="cl_external_sbd_1-meta_attributes">
>
> <nvpair id="cl_external_sbd_1-meta_attributes-clone-max" name="clone-max" value="2"/>
>
> </meta_attributes>
>
> <primitive class="stonith" type="external/sbd" id="stonith_external_sbd_LOCK_LUN">
>
> <instance_attributes id="stonith_external_sbd_LOCK_LUN-instance_attributes">
>
> <nvpair id="nvpair-stonith_external_sbd_LOCK_LUN-sbd_device" name="sbd_device" value="/dev/mapper/mpath1p1"/>
>
> </instance_attributes>
>
> <operations id="stonith_external_sbd_LOCK_LUN-operations">
>
> <op id="op-stonith_external_sbd_LOCK_LUN-stop" interval="0" name="stop" timeout="60"/>
>
> <op id="op-stonith_external_sbd_LOCK_LUN-monitor" interval="60" name="monitor" start-delay="0" timeout="60"/>
>
> <op id="op-stonith_external_sbd_LOCK_LUN-start" interval="0" name="start" timeout="60"/>
>
> </operations>
>
> <meta_attributes id="stonith_external_sbd_LOCK_LUN-meta_attributes">
>
> <nvpair name="target-role" id="stonith_external_sbd_LOCK_LUN-meta_attributes-target-role" value="stopped"/>
>
> </meta_attributes>
>
> </primitive>
>
> </clone>
>
>
>
>
>
> Ciao, Nicola.
>
> _____
>
> Da: Vit Pelcak [mailto:vpelcak at suse.cz]
> Inviato: giovedì 29 aprile 2010 16.08
> A: pacemaker at oss.clusterlabs.org
> Oggetto: Re: [Pacemaker] R: Stonith external/sbd problem
>
>
>
> Also, it is needed to add stonith to cib:
>
> crm configure primitive sbd_stonith stonith:external/sbd meta target-role="Started" op monitor interval="15" timeout="15" start-delay="15" params sbd_device="/dev/sda1"
>
>
> Dne 29.4.2010 15:46, Nicola Sabatelli napsal(a):
>
> I have done exactly the configuration in the SBD_Fencing documentation.
>
> That is:
>
> /etc/sysconfig/sbd
>
> SBD_DEVICE="/dev/mapper/mpath1p1"
>
> SBD_OPTS="-W"
>
> And I start the demon in this manner:
>
> /usr/sbin/sbd -d /dev/mapper/mpath1p1 -D -W watch
>
> Is correct?
>
>
>
> Ciao, Nicola.
>
> _____
>
> Da: Vit Pelcak [mailto:vpelcak at suse.cz]
> Inviato: giovedì 29 aprile 2010 15.02
> A: pacemaker at oss.clusterlabs.org
> Oggetto: Re: [Pacemaker] Stonith external/sbd problem
>
>
>
> cat /etc/sysconfig/sbd
>
> SBD_DEVICE="/dev/sda1"
> SBD_OPTS="-W"
>
>
> sbd -d /dev/shared_disk create
> sbd -d /dev/shared_disk allocate your_machine
>
>
> Dne 29.4.2010 14:55, Michael Brown napsal(a):
>
> Oh, I forgot a piece: I had simular trouble until I actually properly started sbd and then it worked.
>
> M.
>
> _____
>
> From: Michael Brown
> To: pacemaker at oss.clusterlabs.org
> Sent: Thu Apr 29 08:53:32 2010
> Subject: Re: [Pacemaker] Stonith external/sbd problem
>
>
>
>
> I just set this up myself and it worked fine for me.
>
> Did you follow the guide? You need to configure the sbd daemon to run on bootup with appropriate options before external/sbd can use it.
>
> M.
>
> _____
>
> From: Nicola Sabatelli
> To: pacemaker at oss.clusterlabs.org
> Sent: Thu Apr 29 08:47:04 2010
> Subject: [Pacemaker] Stonith external/sbd problem
>
>
> I have a problem with STONITH plugin external/sbd.
>
> I have configured the system in according to directive that I find at url http://www.linux-ha.org/wiki/SBD_Fencing, and the device that I use is configured with multipath software because this disk is residend on a storage system.
>
> I have create a resurse on my cluster using clove directive.
>
> But when I try to start the resurse I have these errors:
>
>
>
> from ha-log file:
>
>
>
> Apr 29 14:37:51 clover-h stonithd: [16811]: info: external_run_cmd: Calling '/usr/lib64/stonith/plugins/external/sbd status' returned 256
>
> Apr 29 14:37:51 clover-h stonithd: [16811]: CRIT: external_status: 'sbd status' failed with rc 256
>
> Apr 29 14:37:51 clover-h stonithd: [10615]: WARN: start stonith_external_sbd_LOCK_LUN:0 failed, because its hostlist is empty
>
>
>
> from crm_verify:
>
>
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: main: =#=#=#=#= Getting XML =#=#=#=#=
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: main: Reading XML from: live cluster
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: unpack_config: On loss of CCM Quorum: Ignore
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: determine_online_status: Node clover-a.rsr.rupar.puglia.it is online
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: unpack_rsc_op: Processing failed op stonith_external_sbd_LOCK_LUN:1_start_0 on clover-a.rsr.rupar.puglia.it: unknown error (1)
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: find_clone: Internally renamed stonith_external_sbd_LOCK_LUN:0 on clover-a.rsr.rupar.puglia.it to stonith_external_sbd_LOCK_LUN:2 (ORPHAN)
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: determine_online_status: Node clover-h.rsr.rupar.puglia.it is online
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: unpack_rsc_op: Processing failed op stonith_external_sbd_LOCK_LUN:0_start_0 on clover-h.rsr.rupar.puglia.it: unknown error (1)
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: clone_print: Master/Slave Set: ms_drbd_1
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: short_print: Stopped: [ res_drbd_1:0 res_drbd_1:1 ]
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print: res_Filesystem_TEST (ocf::heartbeat:Filesystem): Stopped
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print: res_IPaddr2_ip_clover (ocf::heartbeat:IPaddr2): Stopped
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: clone_print: Clone Set: cl_external_sbd_1
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print: stonith_external_sbd_LOCK_LUN:0 (stonith:external/sbd): Started clover-h.rsr.rupar.puglia.it FAILED
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: native_print: stonith_external_sbd_LOCK_LUN:1 (stonith:external/sbd): Started clover-a.rsr.rupar.puglia.it FAILED
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: get_failcount: cl_external_sbd_1 has failed 1000000 times on clover-h.rsr.rupar.puglia.it
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: common_apply_stickiness: Forcing cl_external_sbd_1 away from clover-h.rsr.rupar.puglia.it after 1000000 failures (max=1000000)
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: get_failcount: cl_external_sbd_1 has failed 1000000 times on clover-a.rsr.rupar.puglia.it
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: common_apply_stickiness: Forcing cl_external_sbd_1 away from clover-a.rsr.rupar.puglia.it after 1000000 failures (max=1000000)
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: native_merge_weights: ms_drbd_1: Rolling back scores from res_Filesystem_TEST
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource res_drbd_1:0 cannot run anywhere
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource res_drbd_1:1 cannot run anywhere
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: native_merge_weights: ms_drbd_1: Rolling back scores from res_Filesystem_TEST
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: master_color: ms_drbd_1: Promoted 0 instances of a possible 1 to master
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: master_color: ms_drbd_1: Promoted 0 instances of a possible 1 to master
>
> crm_verify[18607]: 2010/04/29_14:39:27 info: native_merge_weights: res_Filesystem_TEST: Rolling back scores from res_IPaddr2_ip_clover
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource res_Filesystem_TEST cannot run anywhere
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource res_IPaddr2_ip_clover cannot run anywhere
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource stonith_external_sbd_LOCK_LUN:0 cannot run anywhere
>
> crm_verify[18607]: 2010/04/29_14:39:27 WARN: native_color: Resource stonith_external_sbd_LOCK_LUN:1 cannot run anywhere
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource res_drbd_1:0 (Stopped)
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource res_drbd_1:1 (Stopped)
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource res_Filesystem_TEST (Stopped)
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Leave resource res_IPaddr2_ip_clover (Stopped)
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Stop resource stonith_external_sbd_LOCK_LUN:0 (clover-h.rsr.rupar.puglia.it)
>
> crm_verify[18607]: 2010/04/29_14:39:27 notice: LogActions: Stop resource stonith_external_sbd_LOCK_LUN:1 (clover-a.rsr.rupar.puglia.it)
>
> Warnings found during check: config may not be valid
>
>
>
> and from crm_mon:
>
>
>
> ============
>
> Last updated: Thu Apr 29 14:39:57 2010
>
> Stack: Heartbeat
>
> Current DC: clover-h.rsr.rupar.puglia.it (e39bb201-2a6f-457a-a308-be6bfe71309c) - partition with quorum
>
> Version: 1.0.8-9881a7350d6182bae9e8e557cf20a3cc5dac3ee7
>
> 2 Nodes configured, unknown expected votes
>
> 4 Resources configured.
>
> ============
>
>
>
> Online: [ clover-h.rsr.rupar.puglia.it clover-a.rsr.rupar.puglia.it ]
>
>
>
> Clone Set: cl_external_sbd_1
>
> stonith_external_sbd_LOCK_LUN:0 (stonith:external/sbd): Started clover-h.rsr.rupar.puglia.it FAILED
>
> stonith_external_sbd_LOCK_LUN:1 (stonith:external/sbd): Started clover-a.rsr.rupar.puglia.it FAILED
>
>
>
> Operations:
>
> * Node clover-a.rsr.rupar.puglia.it:
>
> stonith_external_sbd_LOCK_LUN:1: migration-threshold=1000000 fail-count=1000000
>
> + (24) start: rc=1 (unknown error)
>
> * Node clover-h.rsr.rupar.puglia.it:
>
> stonith_external_sbd_LOCK_LUN:0: migration-threshold=1000000 fail-count=1000000
>
> + (25) start: rc=1 (unknown error)
>
>
>
> Failed actions:
>
> stonith_external_sbd_LOCK_LUN:1_start_0 (node=clover-a.rsr.rupar.puglia.it, call=24, rc=1, status=complete): unknown error
>
> stonith_external_sbd_LOCK_LUN:0_start_0 (node=clover-h.rsr.rupar.puglia.it, call=25, rc=1, status=complete): unknown error
>
>
>
>
>
>
>
>
>
> Ciao, Nicola.
>
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>
>
>
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>
>
>
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>
>
>
>
>
> --
> Michael Brown | `One of the main causes of the fall of
> Systems Consultant | the Roman Empire was that, lacking zero,
> Net Direct Inc. | they had no way to indicate successful
> ☎: +1 519 883 1172 x5106 | termination of their C programs.' - Firth
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
More information about the Pacemaker
mailing list