[ClusterLabs] Antwort: Re: Antwort: Re: clone resource - pacemaker remote
philipp.achmueller at arz.at
philipp.achmueller at arz.at
Wed Dec 7 07:26:11 EST 2016
> Von: Ken Gaillot <kgaillot at redhat.com>
> An: philipp.achmueller at arz.at, Cluster Labs - All topics related to
> open-source clustering welcomed <users at clusterlabs.org>
> Datum: 05.12.2016 17:38
> Betreff: Re: Antwort: Re: [ClusterLabs] clone resource - pacemaker
remote
>
> On 12/05/2016 09:20 AM, philipp.achmueller at arz.at wrote:
> > Ken Gaillot <kgaillot at redhat.com> schrieb am 02.12.2016 19:27:09:
> >
> >> Von: Ken Gaillot <kgaillot at redhat.com>
> >> An: users at clusterlabs.org
> >> Datum: 02.12.2016 19:32
> >> Betreff: Re: [ClusterLabs] clone resource - pacemaker remote
> >>
> >> On 12/02/2016 07:08 AM, philipp.achmueller at arz.at wrote:
> >> > hi,
> >> >
> >> > what is best way to prevent clone resource trying to run on
remote/guest
> >> > nodes?
> >>
> >> location constraints with a negative score:
> >>
> >> http://clusterlabs.org/doc/en-US/Pacemaker/1.1-pcs/html-single/
> >>
Pacemaker_Explained/index.html#_deciding_which_nodes_a_resource_can_run_on
> >>
> >>
> >> you can even use a single constraint with a rule based on #kind ne
> >> cluster, so you don't need a separate constraint for each node:
> >>
> >> http://clusterlabs.org/doc/en-US/Pacemaker/1.1-pcs/html-single/
> >> Pacemaker_Explained/index.html#_node_attribute_expressions
> >>
> >>
> >> alternatively, you can set symmetric-cluster=false and use positive
> >> constraints for cluster nodes only
> >>
> >
> > set constraint to single primitive, group, or on clone resource?
> > are there any advantages/disadvantages using one of these methods?
>
> When a resource is cloned, you want to refer to the clone name in any
> constraints, rather than the primitive name.
>
> For a group, it doesn't really matter, but it's simplest to use the
> group name in constraints -- mainly that keeps you from accidentally
> setting conflicting constraints on different members of the group. And
> of course group members are automatically ordered/colocated with each
> other, so you don't need individual constraints for that.
>
set location constraint to group didn't work:
ERROR: error: unpack_location_tags: Constraint 'location-base-group':
Invalid reference to 'base-group'
for clone it works like expected.
but crm_mon is showing "disallowed" set as "stopped". is this "works as
designed" or how to prevent this?
crm configure show
...
location location-base-clone base-clone resource-discovery=never \
rule -inf: #kind ne cluster
...
crm_mon -r
Clone Set: base-clone [base-group]
Started: [ lnx0223a lnx0223b ]
Stopped: [ vm-lnx0106a vm-lnx0107a ]
> >
> >> >
> >> > ...
> >> > node 167873318: lnx0223a \
> >> > attributes maintenance=off
> >> > node 167873319: lnx0223b \
> >> > attributes maintenance=off
> >> > ...
> >> > /primitive vm-lnx0107a VirtualDomain \/
> >> > / params hypervisor="qemu:///system"
> >> > config="/etc/kvm/lnx0107a.xml" \/
> >> > / meta remote-node=lnx0107a238 \/
> >> > / utilization cpu=1 hv_memory=4096/
> >> > /primitive remote-lnx0106a ocf:pacemaker:remote \/
> >> > / params server=xx.xx.xx.xx \/
> >> > / meta target-role=Started/
> >> > /group base-group dlm clvm vg1/
> >> > /clone base-clone base-group \/
> >> > / meta interleave=true target-role=Started/
> >> > /.../
> >> >
> >> > /Dec 1 14:32:57 lnx0223a crmd[9826]: notice: Initiating start
> >> > operation dlm_start_0 on lnx0107a238/
> >> > /Dec 1 14:32:58 lnx0107a pacemaker_remoted[1492]: notice:
executing -
> >> > rsc:dlm action:start call_id:7/
> >> > /Dec 1 14:32:58 lnx0107a pacemaker_remoted[1492]: notice:
finished -
> >> > rsc:dlm action:start call_id:7 exit-code:5 exec-time:16ms
> > queue-time:0ms/
> >> > /Dec 1 14:32:58 lnx0223b crmd[9328]: error: Result of start
> >> > operation for dlm on lnx0107a238: Not installed/
> >> > /Dec 1 14:32:58 lnx0223a crmd[9826]: warning: Action 31
(dlm_start_0)
> >> > on lnx0107a238 failed (target: 0 vs. rc: 5): Error/
> >> > /Dec 1 14:32:58 lnx0223a crmd[9826]: warning: Action 31
(dlm_start_0)
> >> > on lnx0107a238 failed (target: 0 vs. rc: 5): Error/
> >> > /Dec 1 14:34:07 lnx0223a pengine[9824]: warning: Processing
failed op
> >> > start for dlm:2 on lnx0107a238: not installed (5)/
> >> > /Dec 1 14:34:07 lnx0223a pengine[9824]: warning: Processing
failed op
> >> > start for dlm:2 on lnx0107a238: not installed (5)/
> >> > /.../
> >> > /Dec 1 14:32:49 lnx0223a pengine[9824]: notice: Start
> >> > dlm:3#011(remote-lnx0106a)/
> >> > /Dec 1 14:32:49 lnx0223a crmd[9826]: notice: Initiating monitor
> >> > operation dlm_monitor_0 locally on remote-lnx0106a/
> >> > /Dec 1 14:32:50 lnx0223a crmd[9826]: error: Result of probe
> >> > operation for dlm on remote-lnx0106a: Not installed/
> >> > /Dec 1 14:32:50 lnx0223a crmd[9826]: warning: Action 5
(dlm_monitor_0)
> >> > on remote-lnx0106a failed (target: 7 vs. rc: 5): Error/
> >> > /Dec 1 14:32:50 lnx0223a crmd[9826]: warning: Action 5
(dlm_monitor_0)
> >> > on remote-lnx0106a failed (target: 7 vs. rc: 5): Error/
> >> > /.../
> >> >
> >> > ---
> >> > env: pacemaker-1.1.15-19.15.x86_64
> >> >
> >> > thank you!
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20161207/6b95f00d/attachment-0003.html>
More information about the Users
mailing list