[Pacemaker] unexpected demote request on master
K Mehta
kiranmehta1981 at gmail.com
Wed May 28 04:13:31 UTC 2014
Chris,
Here is the required information
[root at vsanqa11 ~]# rpm -qa | grep pcs ; rpm -qa | grep pacemaker ; uname -a
; cat /etc/redhat-release
pcs-0.9.90-2.el6.centos.2.noarch
pacemaker-cli-1.1.10-14.el6_5.3.x86_64
pacemaker-libs-1.1.10-14.el6_5.3.x86_64
pacemaker-1.1.10-14.el6_5.3.x86_64
pacemaker-cluster-libs-1.1.10-14.el6_5.3.x86_64
Linux vsanqa11 2.6.32-279.el6.x86_64 #1 SMP Fri Jun 22 12:19:21 UTC 2012
x86_64 x86_64 x86_64 GNU/Linux
CentOS release 6.3 (Final)
Regards,
Kiran
On Wed, May 28, 2014 at 2:47 AM, Chris Feist <cfeist at redhat.com> wrote:
> On 05/27/14 05:38, K Mehta wrote:
>
>> One more question.
>> With crmsh, it was easy to add constraint to avoid a resource from
>> running only
>> a subset(say vsanqa11 and vsanqa12) of nodes using the following command
>>
>> crm configure location ms-${uuid}-nodes ms-$uuid rule -inf: \#uname ne
>> vsanqa11
>> and \#uname ne vsanqa12
>> [root at vsanqa11 ~]# pcs constraint show --full
>> Location Constraints:
>> Resource: ms-c6933988-9e5c-419e-8fdf-744100d76ad6
>> Constraint: ms-c6933988-9e5c-419e-8fdf-744100d76ad6-nodes
>> Rule: score=-INFINITY
>> (id:ms-c6933988-9e5c-419e-8fdf-744100d76ad6-nodes-rule)
>> Expression: #uname ne vsanqa11
>> (id:ms-c6933988-9e5c-419e-8fdf-744100d76ad6-nodes-expression)
>> Expression: #uname ne vsanqa12
>> (id:ms-c6933988-9e5c-419e-8fdf-744100d76ad6-nodes-expression-0)
>> Ordering Constraints:
>> Colocation Constraints:
>>
>> So, both expression are part of the same rule as expected.
>>
>>
>>
>> With pcs, I am not sure how to use avoid constraints if I need a resource
>> to run
>> on vsanqa11 and vsanqa12 and not on any other node.
>> So I tried adding location constraint as follows:
>> pcs -f $CLUSTER_CREATE_LOG constraint location vha-$uuid rule
>> score=-INFINITY
>> \#uname ne vsanqa11 and \#uname ne vsanqa12
>> Even though no error is thrown, the condition after "and" is silently
>> dropped as
>> shown below
>>
>> [root at vsanqa11 ~]# pcs constraint show --full
>> Location Constraints:
>> Resource: ms-c6933988-9e5c-419e-8fdf-744100d76ad6
>> Constraint: location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6
>> Rule: score=-INFINITY
>> (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-rule)
>> Expression: #uname ne vsanqa11
>> (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-rule-expr)
>> Ordering Constraints:
>> Colocation Constraints:
>>
>>
>> Then I tried the following
>> pcs -f $CLUSTER_CREATE_LOG constraint location vha-$uuid rule
>> score=-INFINITY
>> \#uname ne vsanqa11
>> pcs -f $CLUSTER_CREATE_LOG constraint location vha-$uuid rule
>> score=-INFINITY
>> \#uname ne vsanqa12
>>
>> but running these two commands did not help either. Expressions were
>> added to
>> separate rules.
>>
>> [root at vsanqa11 ~]# pcs constraint show --full
>> Location Constraints:
>> Resource: ms-c6933988-9e5c-419e-8fdf-744100d76ad6
>> Constraint: location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-1
>> Rule: score=-INFINITY
>> (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-1-rule)
>> Expression: #uname ne vsanqa12
>> (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-1-rule-expr)
>> Constraint: location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6
>> Rule: score=-INFINITY
>> (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-rule)
>> Expression: #uname ne vsanqa11
>> (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-rule-expr)
>> Ordering Constraints:
>> Colocation Constraints:
>>
>>
>> Also, tried using multistate resource name
>> [root at vsanqa11 ~]# pcs constraint location
>> ms-c6933988-9e5c-419e-8fdf-744100d76ad6 rule score=-INFINITY \#uname ne
>> vsanqa11
>> Error: 'ms-c6933988-9e5c-419e-8fdf-744100d76ad6' is not a resource
>>
>>
>> Can anyone let me correct command for this ?
>>
>
> Which version of pcs are you using (and what distribution)? This has been
> fixed upstream. (Below is a test from my system using the upstream pcs).
>
> [root at rh7-1 pcs]# pcs constraint location D1 rule score=-INFINITY \#uname
> ne vsanqa11 and \#uname ne vsanqa12
> [root at rh7-1 pcs]# pcs constraint
> Location Constraints:
> Resource: D1
> Constraint: location-D1
> Rule: score=-INFINITY boolean-op=and
> Expression: #uname ne vsanqa11
> Expression: #uname ne vsanqa12
>
> Thanks,
> Chris
>
>
>
>>
>>
>>
>> On Tue, May 27, 2014 at 11:01 AM, Andrew Beekhof <andrew at beekhof.net
>> <mailto:andrew at beekhof.net>> wrote:
>>
>>
>> On 27 May 2014, at 2:37 pm, K Mehta <kiranmehta1981 at gmail.com
>> <mailto:kiranmehta1981 at gmail.com>> wrote:
>>
>> > So is globally-unique=false correct in my case ?
>>
>> yes
>>
>> >
>> >
>> > On Tue, May 27, 2014 at 5:30 AM, Andrew Beekhof <
>> andrew at beekhof.net
>> <mailto:andrew at beekhof.net>> wrote:
>> >
>> > On 26 May 2014, at 9:56 pm, K Mehta <kiranmehta1981 at gmail.com
>> <mailto:kiranmehta1981 at gmail.com>> wrote:
>> >
>> > > What I understand from "globally-unique=false" is as follows
>> > > Agent handling the resource does exactly same processing on all
>> nodes.
>> For processing this resource, agent on all nodes will use exactly same
>> resources (files, processes, same parameters to agent entry points,
>> etc).
>> > >
>> > > In case of my resource, agent on all nodes execute same
>> "command" to
>> find score.
>> > > Driver present on all nodes will make sure that the node that is
>> to be
>> promoted is the one that reports highest score as output of the
>> "command".
>> Score is reported to CM using ( /usr/sbin/crm_master -Q -l reboot -v
>> $score)
>> in monitor entry point. Until this score
>> > > is reported, agent on other node will just delete the score using
>> /usr/sbin/crm_master -Q -l reboot -D in monitor entry point
>> > >
>> > >
>> > >
>> > >
>> > > I want to make sure that the resource does not run on nodes
>> other than
>> $node1 and $node2. To achieve this i use the following commands.
>> > >
>> > > pcs -f $CLUSTER_CREATE_LOG constraint location
>> vha-${uuid}
>> prefers $node1
>> > > pcs -f $CLUSTER_CREATE_LOG constraint location
>> vha-${uuid}
>> prefers $node2
>> > > pcs -f $CLUSTER_CREATE_LOG constraint location
>> ms-${uuid}
>> prefers $node1
>> > > pcs -f $CLUSTER_CREATE_LOG constraint location
>> ms-${uuid}
>> prefers $node2
>> > >
>> > > Any issue here ?
>> >
>> > Perhaps this is not intuitive but you'd need to specify 'avoids'
>> constraints for the nodes it must not run on.
>> > 'prefers' only says that of all the available nodes, this one is
>> the best.
>> >
>> > >
>> > > Regards,
>> > > Kiran
>> > >
>> > >
>> > >
>> > > On Mon, May 26, 2014 at 8:54 AM, Andrew Beekhof <
>> andrew at beekhof.net
>> <mailto:andrew at beekhof.net>> wrote:
>> > >
>> > > On 22 May 2014, at 11:20 pm, K Mehta <kiranmehta1981 at gmail.com
>> <mailto:kiranmehta1981 at gmail.com>> wrote:
>> > >
>> > > > > May 13 01:38:36 vsanqa28 pengine[4310]: notice: LogActions:
>> Promote vha-924bf029-93a2-41a0-adcf-f1c1a42956e5:0#011(Slave ->
>> Master vsanqa28)
>> > > > > May 13 01:38:36 vsanqa28 pengine[4310]: notice: LogActions:
>> Demote vha-924bf029-93a2-41a0-adcf-f1c1a42956e5:1#011(Master ->
>> Slave
>> vsanqa27) <<<<< Why did this happen ?
>> > > >
>> > > > attach the file mentioned on the next line and we might be
>> able to
>> find out
>> > > >
>> > >
>> > > Quick question, do you understand what globally-unique=false
>> means and
>> are you sure you want it?
>> > > If the answer is 'yes and yes', are you sure that your agent is
>> using
>> crm_master correctly?
>> > >
>> > > If I run, 'tools/crm_simulate -Sx ~/Downloads/pe-input-818.bz2
>> -s |
>> grep vha-924bf029-93a2-41a0-adcf-f1c1a42956e5', I see:
>> > >
>> > > vha-924bf029-93a2-41a0-adcf-f1c1a42956e5:0 promotion score on
>> vsanqa28:
>> INFINITY
>> > > vha-924bf029-93a2-41a0-adcf-f1c1a42956e5:1 promotion score on
>> vsanqa27: 2
>> > >
>> > >
>> > > Although much of the 'INFINITY' is probably from:
>> > >
>> > > <rsc_location
>> id="location-ms-924bf029-93a2-41a0-adcf-f1c1a42956e5-
>> vsanqa28-INFINITY"
>> node="vsanqa28" rsc="ms-924bf029-93a2-41a0-adcf-f1c1a42956e5"
>> score="INFINITY"/>
>> > >
>> > > This is somewhat odd to include for a clone/master resource.
>> > >
>> > > _______________________________________________
>> > > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> <mailto:Pacemaker at oss.clusterlabs.org>
>>
>> > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> > >
>> > > Project Home: http://www.clusterlabs.org
>> > > Getting started: http://www.clusterlabs.org/
>> doc/Cluster_from_Scratch.pdf
>> > > Bugs: http://bugs.clusterlabs.org
>> > >
>> > >
>> > > _______________________________________________
>> > > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> <mailto:Pacemaker at oss.clusterlabs.org>
>>
>> > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> > >
>> > > Project Home: http://www.clusterlabs.org
>> > > Getting started: http://www.clusterlabs.org/
>> doc/Cluster_from_Scratch.pdf
>> > > Bugs: http://bugs.clusterlabs.org
>> >
>> >
>> > _______________________________________________
>> > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> <mailto:Pacemaker at oss.clusterlabs.org>
>>
>> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> >
>> > Project Home: http://www.clusterlabs.org
>> > Getting started: http://www.clusterlabs.org/
>> doc/Cluster_from_Scratch.pdf
>> > Bugs: http://bugs.clusterlabs.org
>> >
>> >
>> > _______________________________________________
>> > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> <mailto:Pacemaker at oss.clusterlabs.org>
>>
>> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> >
>> > Project Home: http://www.clusterlabs.org
>> > Getting started: http://www.clusterlabs.org/
>> doc/Cluster_from_Scratch.pdf
>> > Bugs: http://bugs.clusterlabs.org
>>
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> <mailto:Pacemaker at oss.clusterlabs.org>
>>
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/
>> doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>>
>>
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20140528/ab9fce29/attachment.htm>
More information about the Pacemaker
mailing list