[Pacemaker] Resources move on Pacemaker + Corosync cluster with set stickiness
Andrew Beekhof
andrew at beekhof.net
Fri May 23 02:42:34 CEST 2014
On 22 May 2014, at 9:00 pm, Danilo Malcangio <d.malcangio at eletech.it> wrote:
> Hi Andrew, first of all thanks for answering.
>
>> Almost certainly the node is configured to start those resources at bootup.
>> Don't do that :)
>>
>
> Are you advicing me to delete the location constraint? (location prefer-et-ipbx-1 cluster-group 100: BX-1)
> Or is it something else that starts the resources on BX-1 Node??
By the looks of it, yes.
Try crm_mon -o and look for monitor operations that return 0 instead of 7
> I've removed all the scripts from the startup sequence with "update-rc.d -f SCRIPT_NAME remove"
>
> Thank you
>
> Il 22/05/2014 12.25, Andrew Beekhof ha scritto:
>> On 22 May 2014, at 5:31 pm, Danilo Malcangio <d.malcangio at eletech.it>
>> wrote:
>>
>>
>>> Hi everyone,
>>> I've created an active/passive 2 node cluster following the documentation on clusterlabs.
>>> My cluster has the following characteristics
>>> Debian Wheezy 7.2.0
>>> Pacemaker 1.1.7
>>> Corosync 1.4.2
>>>
>>> I've made it with the following configuration
>>>
>>> node BX-1
>>> node BX-2
>>> primitive cluster-apache2 ocf:heartbeat:apache \
>>> params configfile="/etc/apache2/apache2.conf" httpd="/usr/sbin/apache2" port="80" \
>>> op monitor interval="10s" timeout="60s" \
>>> op start interval="0" timeout="40s" \
>>> op stop interval="0" timeout="60s"
>>> primitive cluster-asterisk lsb:asterisk \
>>> op monitor interval="30" \
>>> op start interval="0" timeout="120s" \
>>> op stop interval="0" timeout="120s"
>>> primitive cluster-ip ocf:heartbeat:IPaddr2 \
>>> params ip="10.2.30.10" cidr_netmask="20" \
>>> op monitor interval="10s" timeout="20s" \
>>> op start interval="0" timeout="20s" \
>>> op stop interval="0" timeout="20s"
>>> primitive cluster-ntp lsb:ntp \
>>> op monitor interval="30" \
>>> op start interval="0" timeout="120s" \
>>> op stop interval="0" timeout="120s"
>>> primitive cluster-tftp lsb:tftpd-hpa \
>>> op monitor interval="30" \
>>> op start interval="0" timeout="120s" \
>>> op stop interval="0" timeout="120s"
>>> group cluster-group cluster-ip cluster-asterisk cluster-apache2 cluster-tftp cluster-ntp \
>>> meta resource-stickiness="101"
>>> location prefer-et-ipbx-1 cluster-group 100: BX-1
>>> colocation cluster-dependency inf: cluster-ip cluster-asterisk cluster-apache2 cluster-tftp cluster-ntp
>>> property $id="cib-bootstrap-options" \
>>> dc-version="1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff" \
>>> cluster-infrastructure="openais" \
>>> expected-quorum-votes="2" \
>>> stonith-enabled="false" \
>>> no-quorum-policy="ignore" \
>>> default-resource-stickiness="1"
>>> rsc_defaults $id="rsc-options" \
>>> failure-timeout="60s"
>>>
>>> I've set location constraint to have BX-1 as the preferred node at cluster startup, and group stickiness at 101 to avoid moving resources when the master node comes back up (following this guide
>>> http://foaa.de/old-blog/2010/10/intro-to-pacemaker-part-2-advanced-topics/trackback/index.html
>>> ).
>>>
>>> I've got the following problem: resources move when I reboot nodes.
>>>
>>> If i stop corosync on BX-1 resources move to BX-2 and when i restart corosync on BX-1 they stay on BX-2 (as I expected).
>>> But all this doesn't happen with reboot (when I reboot BX-1).
>>>
>> Almost certainly the node is configured to start those resources at bootup.
>> Don't do that :)
>>
>>
>>> After the reboot of BX-1 resources move to BX-2 and when BX-1 comes back up the resources move back to BX-1.
>>>
>>> What am I missing in the configuration??
>>>
>>> Thank you very much for the support
>>>
>>>
>>>
>>> P.s. I attach also the allocation scores obtained with ptest -sL
>>>
>>> BX-1 has resources
>>>
>>> Allocation scores:
>>> group_color: et-cluster allocation score on ET-IPBX-1: 100
>>> group_color: et-cluster allocation score on ET-IPBX-2: 0
>>> group_color: cluster-ip allocation score on ET-IPBX-1: 201
>>> group_color: cluster-ip allocation score on ET-IPBX-2: 0
>>> group_color: cluster-asterisk allocation score on ET-IPBX-1: 101
>>> group_color: cluster-asterisk allocation score on ET-IPBX-2: 0
>>> group_color: cluster-apache2 allocation score on ET-IPBX-1: 101
>>> group_color: cluster-apache2 allocation score on ET-IPBX-2: 0
>>> group_color: cluster-tftp allocation score on ET-IPBX-1: 101
>>> group_color: cluster-tftp allocation score on ET-IPBX-2: 0
>>> group_color: cluster-ntp allocation score on ET-IPBX-1: 101
>>> group_color: cluster-ntp allocation score on ET-IPBX-2: 0
>>> native_color: cluster-ip allocation score on ET-IPBX-1: 3231
>>> native_color: cluster-ip allocation score on ET-IPBX-2: 0
>>> native_color: cluster-asterisk allocation score on ET-IPBX-1: 1515
>>> native_color: cluster-asterisk allocation score on ET-IPBX-2: -INFINITY
>>> native_color: cluster-apache2 allocation score on ET-IPBX-1: 707
>>> native_color: cluster-apache2 allocation score on ET-IPBX-2: -INFINITY
>>> native_color: cluster-tftp allocation score on ET-IPBX-1: 303
>>> native_color: cluster-tftp allocation score on ET-IPBX-2: -INFINITY
>>> native_color: cluster-ntp allocation score on ET-IPBX-1: 101
>>> native_color: cluster-ntp allocation score on ET-IPBX-2: -INFINITY
>>>
>>> BX-2 has resources (while BX-1 reboot)
>>>
>>> Allocation scores:
>>> group_color: et-cluster allocation score on ET-IPBX-1: -INFINITY
>>> group_color: et-cluster allocation score on ET-IPBX-2: 0
>>> group_color: cluster-ip allocation score on ET-IPBX-1: -INFINITY
>>> group_color: cluster-ip allocation score on ET-IPBX-2: 101
>>> group_color: cluster-asterisk allocation score on ET-IPBX-1: 0
>>> group_color: cluster-asterisk allocation score on ET-IPBX-2: 101
>>> group_color: cluster-apache2 allocation score on ET-IPBX-1: 0
>>> group_color: cluster-apache2 allocation score on ET-IPBX-2: 101
>>> group_color: cluster-tftp allocation score on ET-IPBX-1: 0
>>> group_color: cluster-tftp allocation score on ET-IPBX-2: 101
>>> group_color: cluster-ntp allocation score on ET-IPBX-1: 0
>>> group_color: cluster-ntp allocation score on ET-IPBX-2: 101
>>> native_color: cluster-ip allocation score on ET-IPBX-1: -INFINITY
>>> native_color: cluster-ip allocation score on ET-IPBX-2: 3131
>>> native_color: cluster-asterisk allocation score on ET-IPBX-1: -INFINITY
>>> native_color: cluster-asterisk allocation score on ET-IPBX-2: 1515
>>> native_color: cluster-apache2 allocation score on ET-IPBX-1: -INFINITY
>>> native_color: cluster-apache2 allocation score on ET-IPBX-2: 707
>>> native_color: cluster-tftp allocation score on ET-IPBX-1: -INFINITY
>>> native_color: cluster-tftp allocation score on ET-IPBX-2: 303
>>> native_color: cluster-ntp allocation score on ET-IPBX-1: -INFINITY
>>> native_color: cluster-ntp allocation score on ET-IPBX-2: 101
>>> _______________________________________________
>>> Pacemaker mailing list:
>>> Pacemaker at oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>
>>>
>>> Project Home:
>>> http://www.clusterlabs.org
>>>
>>> Getting started:
>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>
>>> Bugs:
>>> http://bugs.clusterlabs.org
>>
>>
>> _______________________________________________
>> Pacemaker mailing list:
>> Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>>
>> Project Home:
>> http://www.clusterlabs.org
>>
>> Getting started:
>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>
>> Bugs:
>> http://bugs.clusterlabs.org
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 841 bytes
Desc: Message signed with OpenPGP using GPGMail
URL: <http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20140523/d15c3d86/attachment.sig>
More information about the Pacemaker
mailing list