[ClusterLabs] resource-stickiness
Rakovec Jost
Jost.Rakovec at snt.si
Wed Aug 26 12:09:27 UTC 2015
Sorry one typo: problem is the same....
location cli-prefer-aapche aapche role=Started 10: sles2
to:
location cli-prefer-aapche aapche role=Started inf: sles2
It keep change to infinity.
my configuration is:
node sles1
node sles2
primitive filesystem Filesystem \
params fstype=ext3 directory="/srv/www/vhosts" device="/dev/xvdd1" \
op start interval=0 timeout=60 \
op stop interval=0 timeout=60 \
op monitor interval=20 timeout=40
primitive myip IPaddr2 \
params ip=x.x.x.x \
op start interval=0 timeout=20s \
op stop interval=0 timeout=20s \
op monitor interval=10s timeout=20s
primitive stonith_sbd stonith:external/sbd \
params pcmk_delay_max=30
primitive web apache \
params configfile="/etc/apache2/httpd.conf" \
op start interval=0 timeout=40s \
op stop interval=0 timeout=60s \
op monitor interval=10 timeout=20s
group aapche filesystem myip web \
meta target-role=Started is-managed=true resource-stickiness=1000
location cli-prefer-aapche aapche role=Started 10: sles2
property cib-bootstrap-options: \
stonith-enabled=true \
no-quorum-policy=ignore \
placement-strategy=balanced \
expected-quorum-votes=2 \
dc-version=1.1.12-f47ea56 \
cluster-infrastructure="classic openais (with plugin)" \
last-lrm-refresh=1440502955 \
stonith-timeout=40s
rsc_defaults rsc-options: \
resource-stickiness=1000 \
migration-threshold=3
op_defaults op-options: \
timeout=600 \
record-pending=true
and after migration:
node sles1
node sles2
primitive filesystem Filesystem \
params fstype=ext3 directory="/srv/www/vhosts" device="/dev/xvdd1" \
op start interval=0 timeout=60 \
op stop interval=0 timeout=60 \
op monitor interval=20 timeout=40
primitive myip IPaddr2 \
params ip=10.9.131.86 \
op start interval=0 timeout=20s \
op stop interval=0 timeout=20s \
op monitor interval=10s timeout=20s
primitive stonith_sbd stonith:external/sbd \
params pcmk_delay_max=30
primitive web apache \
params configfile="/etc/apache2/httpd.conf" \
op start interval=0 timeout=40s \
op stop interval=0 timeout=60s \
op monitor interval=10 timeout=20s
group aapche filesystem myip web \
meta target-role=Started is-managed=true resource-stickiness=1000
location cli-prefer-aapche aapche role=Started inf: sles2
property cib-bootstrap-options: \
stonith-enabled=true \
no-quorum-policy=ignore \
placement-strategy=balanced \
expected-quorum-votes=2 \
dc-version=1.1.12-f47ea56 \
cluster-infrastructure="classic openais (with plugin)" \
last-lrm-refresh=1440502955 \
stonith-timeout=40s
rsc_defaults rsc-options: \
resource-stickiness=1000 \
migration-threshold=3
op_defaults op-options: \
timeout=600 \
record-pending=true
________________________________
From: Rakovec Jost
Sent: Wednesday, August 26, 2015 1:33 PM
To: users at clusterlabs.org
Subject: resource-stickiness
Hi list,
I have configure simple cluster on sles 11 sp4 and have a problem with “auto_failover off". The problem is that when ever I migrate resource group via HAWK my configuration change from:
location cli-prefer-aapche aapche role=Started 10: sles2
to:
location cli-ban-aapche-on-sles1 aapche role=Started -inf: sles1
It keep change to inf.
and then after fance node, resource is moving back to original node which I don't want. How can I avoid this situation?
my configuration is:
node sles1
node sles2
primitive filesystem Filesystem \
params fstype=ext3 directory="/srv/www/vhosts" device="/dev/xvdd1" \
op start interval=0 timeout=60 \
op stop interval=0 timeout=60 \
op monitor interval=20 timeout=40
primitive myip IPaddr2 \
params ip=x.x.x.x \
op start interval=0 timeout=20s \
op stop interval=0 timeout=20s \
op monitor interval=10s timeout=20s
primitive stonith_sbd stonith:external/sbd \
params pcmk_delay_max=30
primitive web apache \
params configfile="/etc/apache2/httpd.conf" \
op start interval=0 timeout=40s \
op stop interval=0 timeout=60s \
op monitor interval=10 timeout=20s
group aapche filesystem myip web \
meta target-role=Started is-managed=true resource-stickiness=1000
location cli-prefer-aapche aapche role=Started 10: sles2
property cib-bootstrap-options: \
stonith-enabled=true \
no-quorum-policy=ignore \
placement-strategy=balanced \
expected-quorum-votes=2 \
dc-version=1.1.12-f47ea56 \
cluster-infrastructure="classic openais (with plugin)" \
last-lrm-refresh=1440502955 \
stonith-timeout=40s
rsc_defaults rsc-options: \
resource-stickiness=1000 \
migration-threshold=3
op_defaults op-options: \
timeout=600 \
record-pending=true
and after migration:
node sles1
node sles2
primitive filesystem Filesystem \
params fstype=ext3 directory="/srv/www/vhosts" device="/dev/xvdd1" \
op start interval=0 timeout=60 \
op stop interval=0 timeout=60 \
op monitor interval=20 timeout=40
primitive myip IPaddr2 \
params ip=10.9.131.86 \
op start interval=0 timeout=20s \
op stop interval=0 timeout=20s \
op monitor interval=10s timeout=20s
primitive stonith_sbd stonith:external/sbd \
params pcmk_delay_max=30
primitive web apache \
params configfile="/etc/apache2/httpd.conf" \
op start interval=0 timeout=40s \
op stop interval=0 timeout=60s \
op monitor interval=10 timeout=20s
group aapche filesystem myip web \
meta target-role=Started is-managed=true resource-stickiness=1000
location cli-ban-aapche-on-sles1 aapche role=Started -inf: sles1
location cli-prefer-aapche aapche role=Started 10: sles2
property cib-bootstrap-options: \
stonith-enabled=true \
no-quorum-policy=ignore \
placement-strategy=balanced \
expected-quorum-votes=2 \
dc-version=1.1.12-f47ea56 \
cluster-infrastructure="classic openais (with plugin)" \
last-lrm-refresh=1440502955 \
stonith-timeout=40s
rsc_defaults rsc-options: \
resource-stickiness=1000 \
migration-threshold=3
op_defaults op-options: \
timeout=600 \
record-pending=true
thanks
Best Regards
Jost
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20150826/2bb56b30/attachment.htm>
More information about the Users
mailing list