[ClusterLabs] Antwort: Re: clone resource - pacemaker remote
Ken Gaillot
kgaillot at redhat.com
Mon Dec 5 11:29:44 EST 2016
On 12/05/2016 09:20 AM, philipp.achmueller at arz.at wrote:
> Ken Gaillot <kgaillot at redhat.com> schrieb am 02.12.2016 19:27:09:
>
>> Von: Ken Gaillot <kgaillot at redhat.com>
>> An: users at clusterlabs.org
>> Datum: 02.12.2016 19:32
>> Betreff: Re: [ClusterLabs] clone resource - pacemaker remote
>>
>> On 12/02/2016 07:08 AM, philipp.achmueller at arz.at wrote:
>> > hi,
>> >
>> > what is best way to prevent clone resource trying to run on remote/guest
>> > nodes?
>>
>> location constraints with a negative score:
>>
>> http://clusterlabs.org/doc/en-US/Pacemaker/1.1-pcs/html-single/
>> Pacemaker_Explained/index.html#_deciding_which_nodes_a_resource_can_run_on
>>
>>
>> you can even use a single constraint with a rule based on #kind ne
>> cluster, so you don't need a separate constraint for each node:
>>
>> http://clusterlabs.org/doc/en-US/Pacemaker/1.1-pcs/html-single/
>> Pacemaker_Explained/index.html#_node_attribute_expressions
>>
>>
>> alternatively, you can set symmetric-cluster=false and use positive
>> constraints for cluster nodes only
>>
>
> set constraint to single primitive, group, or on clone resource?
> are there any advantages/disadvantages using one of these methods?
When a resource is cloned, you want to refer to the clone name in any
constraints, rather than the primitive name.
For a group, it doesn't really matter, but it's simplest to use the
group name in constraints -- mainly that keeps you from accidentally
setting conflicting constraints on different members of the group. And
of course group members are automatically ordered/colocated with each
other, so you don't need individual constraints for that.
>
>> >
>> > ...
>> > node 167873318: lnx0223a \
>> > attributes maintenance=off
>> > node 167873319: lnx0223b \
>> > attributes maintenance=off
>> > ...
>> > /primitive vm-lnx0107a VirtualDomain \/
>> > / params hypervisor="qemu:///system"
>> > config="/etc/kvm/lnx0107a.xml" \/
>> > / meta remote-node=lnx0107a238 \/
>> > / utilization cpu=1 hv_memory=4096/
>> > /primitive remote-lnx0106a ocf:pacemaker:remote \/
>> > / params server=xx.xx.xx.xx \/
>> > / meta target-role=Started/
>> > /group base-group dlm clvm vg1/
>> > /clone base-clone base-group \/
>> > / meta interleave=true target-role=Started/
>> > /.../
>> >
>> > /Dec 1 14:32:57 lnx0223a crmd[9826]: notice: Initiating start
>> > operation dlm_start_0 on lnx0107a238/
>> > /Dec 1 14:32:58 lnx0107a pacemaker_remoted[1492]: notice: executing -
>> > rsc:dlm action:start call_id:7/
>> > /Dec 1 14:32:58 lnx0107a pacemaker_remoted[1492]: notice: finished -
>> > rsc:dlm action:start call_id:7 exit-code:5 exec-time:16ms
> queue-time:0ms/
>> > /Dec 1 14:32:58 lnx0223b crmd[9328]: error: Result of start
>> > operation for dlm on lnx0107a238: Not installed/
>> > /Dec 1 14:32:58 lnx0223a crmd[9826]: warning: Action 31 (dlm_start_0)
>> > on lnx0107a238 failed (target: 0 vs. rc: 5): Error/
>> > /Dec 1 14:32:58 lnx0223a crmd[9826]: warning: Action 31 (dlm_start_0)
>> > on lnx0107a238 failed (target: 0 vs. rc: 5): Error/
>> > /Dec 1 14:34:07 lnx0223a pengine[9824]: warning: Processing failed op
>> > start for dlm:2 on lnx0107a238: not installed (5)/
>> > /Dec 1 14:34:07 lnx0223a pengine[9824]: warning: Processing failed op
>> > start for dlm:2 on lnx0107a238: not installed (5)/
>> > /.../
>> > /Dec 1 14:32:49 lnx0223a pengine[9824]: notice: Start
>> > dlm:3#011(remote-lnx0106a)/
>> > /Dec 1 14:32:49 lnx0223a crmd[9826]: notice: Initiating monitor
>> > operation dlm_monitor_0 locally on remote-lnx0106a/
>> > /Dec 1 14:32:50 lnx0223a crmd[9826]: error: Result of probe
>> > operation for dlm on remote-lnx0106a: Not installed/
>> > /Dec 1 14:32:50 lnx0223a crmd[9826]: warning: Action 5 (dlm_monitor_0)
>> > on remote-lnx0106a failed (target: 7 vs. rc: 5): Error/
>> > /Dec 1 14:32:50 lnx0223a crmd[9826]: warning: Action 5 (dlm_monitor_0)
>> > on remote-lnx0106a failed (target: 7 vs. rc: 5): Error/
>> > /.../
>> >
>> > ---
>> > env: pacemaker-1.1.15-19.15.x86_64
>> >
>> > thank you!
More information about the Users
mailing list