[Pacemaker] unexpected demote request on master

K Mehta kiranmehta1981 at gmail.com
Tue May 27 10:38:40 UTC 2014


One more question.
With crmsh, it was easy to add constraint to avoid a resource from running
only a subset(say vsanqa11 and vsanqa12) of nodes using the following
command

crm configure location ms-${uuid}-nodes ms-$uuid rule -inf: \#uname ne
vsanqa11 and \#uname ne  vsanqa12
[root at vsanqa11 ~]# pcs constraint show --full
Location Constraints:
  Resource: ms-c6933988-9e5c-419e-8fdf-744100d76ad6
    Constraint: ms-c6933988-9e5c-419e-8fdf-744100d76ad6-nodes
      Rule: score=-INFINITY
 (id:ms-c6933988-9e5c-419e-8fdf-744100d76ad6-nodes-rule)
        Expression: #uname ne vsanqa11
 (id:ms-c6933988-9e5c-419e-8fdf-744100d76ad6-nodes-expression)
        Expression: #uname ne vsanqa12
 (id:ms-c6933988-9e5c-419e-8fdf-744100d76ad6-nodes-expression-0)
Ordering Constraints:
Colocation Constraints:

So, both expression are part of the same rule as expected.



With pcs, I am not sure how to use avoid constraints if I need a resource
to run on vsanqa11 and vsanqa12 and not on any other node.
So I tried adding location constraint as follows:
pcs -f $CLUSTER_CREATE_LOG constraint location vha-$uuid rule
score=-INFINITY \#uname ne vsanqa11 and \#uname ne vsanqa12
Even though no error is thrown, the condition after "and" is silently
dropped as shown below

[root at vsanqa11 ~]# pcs constraint show --full
Location Constraints:
  Resource: ms-c6933988-9e5c-419e-8fdf-744100d76ad6
    Constraint: location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6
      Rule: score=-INFINITY
 (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-rule)
        Expression: #uname ne vsanqa11
 (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-rule-expr)
Ordering Constraints:
Colocation Constraints:


Then I tried the following
pcs -f $CLUSTER_CREATE_LOG constraint location vha-$uuid rule
score=-INFINITY \#uname ne vsanqa11
pcs -f $CLUSTER_CREATE_LOG constraint location vha-$uuid rule
score=-INFINITY \#uname ne vsanqa12

but running these two commands did not help either. Expressions were added
to separate rules.

[root at vsanqa11 ~]# pcs constraint show --full
Location Constraints:
  Resource: ms-c6933988-9e5c-419e-8fdf-744100d76ad6
    Constraint: location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-1
      Rule: score=-INFINITY
 (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-1-rule)
        Expression: #uname ne vsanqa12
 (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-1-rule-expr)
    Constraint: location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6
      Rule: score=-INFINITY
 (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-rule)
        Expression: #uname ne vsanqa11
 (id:location-vha-c6933988-9e5c-419e-8fdf-744100d76ad6-rule-expr)
Ordering Constraints:
Colocation Constraints:


Also, tried using multistate resource name
[root at vsanqa11 ~]# pcs constraint location
ms-c6933988-9e5c-419e-8fdf-744100d76ad6 rule score=-INFINITY \#uname ne
vsanqa11
Error: 'ms-c6933988-9e5c-419e-8fdf-744100d76ad6' is not a resource


Can anyone let me correct command for this ?




On Tue, May 27, 2014 at 11:01 AM, Andrew Beekhof <andrew at beekhof.net> wrote:

>
> On 27 May 2014, at 2:37 pm, K Mehta <kiranmehta1981 at gmail.com> wrote:
>
> > So is globally-unique=false correct in my case ?
>
> yes
>
> >
> >
> > On Tue, May 27, 2014 at 5:30 AM, Andrew Beekhof <andrew at beekhof.net>
> wrote:
> >
> > On 26 May 2014, at 9:56 pm, K Mehta <kiranmehta1981 at gmail.com> wrote:
> >
> > > What I understand from "globally-unique=false" is as follows
> > > Agent handling the resource does exactly same processing on all nodes.
> For processing this resource, agent on all nodes will use exactly same
> resources (files, processes, same parameters to agent entry points, etc).
> > >
> > > In case of my resource, agent on all nodes execute same "command" to
> find score.
> > > Driver present on all nodes will make sure that the node that is to be
> promoted is the one that reports highest score as output of the "command".
> Score is reported to CM using ( /usr/sbin/crm_master -Q -l reboot -v
> $score) in monitor entry point. Until this score
> > > is reported, agent on other node will just delete the score using
> /usr/sbin/crm_master -Q -l reboot -D in monitor entry point
> > >
> > >
> > >
> > >
> > > I want to make sure that the resource does not run on nodes other than
> $node1 and $node2. To achieve this i use the following commands.
> > >
> > >         pcs -f $CLUSTER_CREATE_LOG constraint  location vha-${uuid}
>  prefers $node1
> > >         pcs -f $CLUSTER_CREATE_LOG constraint  location vha-${uuid}
>  prefers $node2
> > >         pcs -f $CLUSTER_CREATE_LOG constraint  location ms-${uuid}
>  prefers $node1
> > >         pcs -f $CLUSTER_CREATE_LOG constraint  location ms-${uuid}
>  prefers $node2
> > >
> > > Any issue here ?
> >
> > Perhaps this is not intuitive but you'd need to specify 'avoids'
> constraints for the nodes it must not run on.
> > 'prefers' only says that of all the available nodes, this one is the
> best.
> >
> > >
> > > Regards,
> > >  Kiran
> > >
> > >
> > >
> > > On Mon, May 26, 2014 at 8:54 AM, Andrew Beekhof <andrew at beekhof.net>
> wrote:
> > >
> > > On 22 May 2014, at 11:20 pm, K Mehta <kiranmehta1981 at gmail.com> wrote:
> > >
> > > > > May 13 01:38:36 vsanqa28 pengine[4310]:   notice: LogActions:
> Promote vha-924bf029-93a2-41a0-adcf-f1c1a42956e5:0#011(Slave -> Master
> vsanqa28)
> > > > > May 13 01:38:36 vsanqa28 pengine[4310]:   notice: LogActions:
> Demote  vha-924bf029-93a2-41a0-adcf-f1c1a42956e5:1#011(Master -> Slave
> vsanqa27)  <<<<< Why did this happen ?
> > > >
> > > > attach the file mentioned on the next line and we might be able to
> find out
> > > >
> > >
> > > Quick question, do you understand what globally-unique=false means and
> are you sure you want it?
> > > If the answer is 'yes and yes', are you sure that your agent is using
> crm_master correctly?
> > >
> > > If I run, 'tools/crm_simulate -Sx ~/Downloads/pe-input-818.bz2 -s |
> grep vha-924bf029-93a2-41a0-adcf-f1c1a42956e5', I see:
> > >
> > > vha-924bf029-93a2-41a0-adcf-f1c1a42956e5:0 promotion score on
> vsanqa28: INFINITY
> > > vha-924bf029-93a2-41a0-adcf-f1c1a42956e5:1 promotion score on
> vsanqa27: 2
> > >
> > >
> > > Although much of the 'INFINITY' is probably from:
> > >
> > >       <rsc_location
> id="location-ms-924bf029-93a2-41a0-adcf-f1c1a42956e5-vsanqa28-INFINITY"
> node="vsanqa28" rsc="ms-924bf029-93a2-41a0-adcf-f1c1a42956e5"
> score="INFINITY"/>
> > >
> > > This is somewhat odd to include for a clone/master resource.
> > >
> > > _______________________________________________
> > > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> > >
> > > Project Home: http://www.clusterlabs.org
> > > Getting started:
> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > > Bugs: http://bugs.clusterlabs.org
> > >
> > >
> > > _______________________________________________
> > > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> > >
> > > Project Home: http://www.clusterlabs.org
> > > Getting started:
> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > > Bugs: http://bugs.clusterlabs.org
> >
> >
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> >
> >
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20140527/4d0a0fc9/attachment.htm>


More information about the Pacemaker mailing list