[Pacemaker] Migrate/run resource only if m/s resource is master/promoted on target node (follow-up)
Martin Gombač
martin at isg.si
Wed Dec 30 09:12:23 UTC 2009
I have it set already to ignore, because i do resource level fencing
with custom app, which is run/triggered by drbd when it looses connection.
>
property $id="cib-bootstrap-options" \
dc-version="1.0.6-f709c638237cdff7556cb6ab615f32826c0f8c06" \
cluster-infrastructure="Heartbeat" \
stonith-enabled="false" \
no-quorum-policy="ignore" \
default-resource-stickiness="1000" \
last-lrm-refresh="1262007932"
>
I have just noticed an error with ms_drbd_r0.
>
ms ms_drbd_r0 drbd_r0 \
meta notify="true" master-max="2" inteleave="true"
>
interleave is not inteleave, correct? :-)
If that won't help i'll also try node-master-max and/or
clone(-master)-max meta options.
Will let you know if it works out.
Regards,
M.
hj lee wrote:
> Hi,
>
> Maybe this is related to no-quorum-policy. What is no-quorum-policy?
> You can check no-quorum-policy in "crm configure show" command. If you
> can not see in "crm configure show" command, then it is "stop" by
> default. If that is your case, then please set it to ignore by "crm
> configure property no-quorum-policy=ignore". and try it again.
>
> Thanks
> hj
>
> On Tue, Dec 29, 2009 at 6:43 AM, Martin Gombač <martin at isg.si
> <mailto:martin at isg.si>> wrote:
>
> Hi guys once again :-)
>
> my resource Hosting on top of ms_drbd_r0 keeps restarting even if
> the changes aren't local to the node.
> By that, i mean Hosting get's restarted on node1 even if i restart
> or outdate and demote node 2.
>
> My constrains:
> colocation Hosting_on_ms_drbd_r0 inf: Hosting ms_drbd_r0:Master
> order ms_drbd_r0_b4_Hosting inf: ms_drbd_r0:promote Hosting:start
>
> My resources:
> primitive Hosting ocf:heartbeat:Xen \
> params xmfile="/etc/xen/Hosting.cfg" \
> meta target-role="Started" allow-migrate="true" is-managed="true" \
> op monitor interval="120s" timeout="300s"
> primitive drbd_r0 ocf:linbit:drbd \
> params drbd_resource="r0" \
> op monitor interval="15s" role="Master" timeout="30s" \
> op monitor interval="30s" role="Slave" timeout="30"
> ms ms_drbd_r0 drbd_r0 \
> meta notify="true" master-max="2" inteleave="true"
>
> I use location to:
> location cli-prefer-Hosting Hosting \
> rule $id="cli-prefer-rule-Hosting" inf: #uname eq ibm1
>
> Example:
> Shuting down heartbeat on node2/ibm2, while resource is running
> perfectly fine on node1, makes resource Hosting restart on node1.
>
> Dec 29 14:21:28 ibm1 pengine: [3716]: notice: native_print:
> Hosting (ocf::heartbeat:Xen): Started ibm1
> Dec 29 14:21:28 ibm1 pengine: [3716]: notice: clone_print:
> Master/Slave Set: ms_drbd_r0
> Dec 29 14:21:28 ibm1 pengine: [3716]: notice: short_print:
> Masters: [ ibm1 ibm2]
> Dec 29 14:21:28 ibm1 pengine: [3716]: WARN: native_color: Resource
> drbd_r0:1 cannot run anywhere
> Dec 29 14:21:28 ibm1 pengine: [3716]: info: master_color:
> Promoting drbd_r0:0 (Master ibm1)
> Dec 29 14:21:28 ibm1 pengine: [3716]: info: master_color:
> ms_drbd_r0: Promoted 1 instances of a possible 2 to master
> Dec 29 14:21:28 ibm1 pengine: [3716]: info: master_color:
> Promoting drbd_r0:0 (Master ibm1
> Dec 29 14:21:28 ibm1 pengine: [3716]: info: master_color:
> ms_drbd_r0: Promoted 1 instances of a possible 2 to master
> Dec 29 14:21:28 ibm1 pengine: [3716]: info: stage6: Scheduling
> Node ibm2 for shutdown
> Dec 29 14:21:28 ibm1 pengine: [3716]: notice: LogActions: Restart
> resource Hosting (Started ibm1)
> Dec 29 14:21:28 ibm1 pengine: [3716]: notice: LogActions: Leave
> resource drbd_r0:0 (Master ibm1)
> Dec 29 14:21:28 ibm1 pengine: [3716]: notice: LogActions: Demote
> drbd_r0:1 (Master -> Stopped ibm2)
> Dec 29 14:21:28 ibm1 pengine: [3716]: notice: LogActions: Stop
> resource drbd_r0:1 (ibm2)
> ...
> Dec 29 14:21:28 ibm1 crmd: [3713]: info: te_rsc_command:
> Initiating action 8: stop Hosting_stop_0 on ibm1(local)
> Dec 29 14:21:28 ibm1 crmd: [3713]: info: do_lrm_rsc_op: Performing
> key=8:4:0:13282265-c62f-4341-9fa5-363cd30ddd3e op=Hosting_stop_0 )
> Dec 29 14:21:28 ibm1 lrmd: [3710]: info: rsc:Hosting:19: stop
> ...
> Dec 29 14:21:34 ibm1 crmd: [3713]: info: match_graph_event: Action
> Hosting_stop_0 (8) confirmed on ibm1..si (rc=0)
> Dec 29 14:21:34 ibm1 crmd: [3713]: info: te_rsc_command:
> Initiating action 9: start Hosting_start_0 on ibm1..si (local)
> Dec 29 14:21:34 ibm1 lrmd: [3710]: info: rsc:Hosting:21: start
> ....
> ...
> .
>
>
> Please advise me on how to configure ordering and collocation
> constrains, to not get Hosting resource restarted each time when
> something happens to backup drbd resource on the second node.
>
>
> Thank you.
> Martin
>
> _______________________________________________
> Pacemaker mailing list
> Pacemaker at oss.clusterlabs.org <mailto:Pacemaker at oss.clusterlabs.org>
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
>
>
>
> --
> Dream with longterm vision!
> kerdosa
> ------------------------------------------------------------------------
>
> _______________________________________________
> Pacemaker mailing list
> Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
More information about the Pacemaker
mailing list