[ClusterLabs] Pacemaker

Ken Gaillot kgaillot at redhat.com
Wed Nov 2 13:21:29 EDT 2016


On 11/01/2016 02:31 AM, Siwakoti, Ganesh wrote:
> Hi,
> 
> 
> i'm using CentOS release 6.8 (Final) as a KVM and i configured 3
> nodes(PM1.local,PM2.local and PM3.local), and using
> CMAN clustering. Resources running at two nodes as Active node then
> another one node is for Fail-over resource as a Standby node. Configured
> resources are groupA and groupB, 5 resources are configured in each
> resource group. groupA is Basically run at PM1.local and groupB at
> PM2.local. If fail any resource group it'll fail-over at PM3.local. On
> PM3.local can run only one Resource group at a time. groupA should not
> run at PM2.local and groupB should not run at PM1.local. groupA or
> groupB resources not move back on own nodes from PM3.local
> automatically, if need to move back on own nodes, can move by
> manually.someone pease help me.

Your configuration looks good to me. I don't think you need the -inf
rules for fail-counts; migration-threshold=1 handles that. Are you
seeing any problems?

> 
> 
> version information:
> 
> CentOS release 6.8 (Final) (KVM)
> 
> Pacemaker 1.1.14-8.el6_8.1
> 
> Cman Version: 6.2.0
> 
> 
> I want to be sure that my configuration will work properly or not.
> 
> 
> my Cluster configuration and crm_configuration are on below:
> 
> 
> cluster.conf:
> 
> <?xml version="1.0"?>
> 
> <cluster config_version="23" name="clusterpm666">
> 
>         <fence_daemon/>
> 
>         <clusternodes>
> 
>                 <clusternode name="pm1.local" nodeid="1">
> 
>                         <fence>
> 
>                                 <method name="pcmk-redirect">
> 
>                                         <device name="pcmk"
> port="pm1.local"/>
> 
>                                 </method>
> 
>                         </fence>
> 
>                         <altname name="192.168.14.210"/>
> 
>                 </clusternode>
> 
>                 <clusternode name="pm2.local" nodeid="2">
> 
>                         <fence>
> 
>                                 <method name="pcmk-redirect">
> 
>                                         <device name="pcmk"
> port="pm2.local"/>
> 
>                                 </method>
> 
>                         </fence>
> 
>                         <altname name="192.168.14.211"/>
> 
>                 </clusternode>
> 
>                 <clusternode name="pm3.local" nodeid="3">
> 
>                         <fence>
> 
>                                 <method name="pcmk-redirect">
> 
>                                         <device name="pcmk"
> port="pm3.local"/>
> 
>                                 </method>
> 
>                         </fence>
> 
>                         <altname name="192.168.14.212"/>
> 
>                 </clusternode>
> 
>         </clusternodes>
> 
>         <fencedevices>
> 
>                 <fencedevice agent="fence_pcmk" name="pcmk"/>
> 
>         </fencedevices>
> 
>         <rm>
> 
>                 <failoverdomains/>
> 
>                 <resources/>
> 
>         </rm>
> 
> </cluster>
> 
> 
> crm_configuration:
> 
> 
> node pm1.local \
> 
> attributes \
> 
> utilization capacity=1
> 
> node pm2.local \
> 
> attributes \
> 
> utilization capacity=1
> 
> node pm3.local \
> 
> attributes \
> 
> utilization capacity=1
> 
> primitive asteriskA asterisk \
> 
> params binary="/usr/sbin/asterisk" canary_binary=astcanary
> additional_parameters="-vvvg -I" config="/etc/asteri
> 
> sk_pm1/asterisk.conf" user=root group=root additional_parameters="-vvvg
> -I" realtime=true maxfiles=32768 \
> 
> meta migration-threshold=2 \
> 
> op monitor interval=20s start-delay=5s timeout=30s \
> 
> op stop interval=0s on-fail=ignore
> 
> primitive asteriskB asterisk \
> 
> params binary="/usr/sbin/asterisk" canary_binary=astcanary
> additional_parameters="-vvvg -I" config="/etc/asteri
> 
> sk_pm2/asterisk.conf" user=root group=root additional_parameters="-vvvg
> -I" realtime=true maxfiles=32768 \
> 
> meta migration-threshold=2 \
> 
> op monitor interval=20s start-delay=5s timeout=30s \
> 
> op stop interval=0s on-fail=ignore
> 
> primitive changeSrcIpA ocf:pacemaker:changeSrcIp \
> 
> params vip=192.168.12.215 mask=23 device=eth0 \
> 
> op start interval=0s timeout=0 \
> 
> op monitor interval=10s \
> 
> op stop interval=0s on-fail=ignore
> 
> primitive changeSrcIpB ocf:pacemaker:changeSrcIp \
> 
> params vip=192.168.12.216 mask=23 device=eth0 \
> 
> op start interval=0s timeout=0 \
> 
> op monitor interval=10s \
> 
> op stop interval=0s on-fail=ignore
> 
> primitive cronA lsb:crond \
> 
> meta migration-threshold=2 \
> 
> op monitor interval=20s start-delay=5s timeout=15s \
> 
> op stop interval=0s on-fail=ignore
> 
> primitive cronB lsb:crond \
> 
> meta migration-threshold=2 \
> 
> op monitor interval=20s start-delay=5s timeout=15s \
> 
> op stop interval=0s on-fail=ignore
> 
> primitive vip-local-checkA VIPcheck \
> 
> params target_ip=192.168.12.215 count=1 wait=5 \
> 
> op start interval=0s on-fail=restart timeout=60s \
> 
> op monitor interval=10s timeout=60s \
> 
> op stop interval=0s on-fail=ignore timeout=60s \
> 
> utilization capacity=1
> 
> primitive vip-local-checkB VIPcheck \
> 
> params target_ip=192.168.12.216 count=1 wait=5 \
> 
> op start interval=0s on-fail=restart timeout=60s \
> 
> op monitor interval=10s timeout=60s \
> 
> op stop interval=0s on-fail=ignore timeout=60s \
> 
> utilization capacity=1
> 
> primitive vip-localA IPaddr2 \
> 
> params ip=192.168.12.215 cidr_netmask=23 nic=eth0 iflabel=0
> broadcast=192.168.13.255 \
> 
> op start interval=0s timeout=20s \
> 
> op monitor interval=5s timeout=20s \
> 
> op stop interval=0s on-fail=ignore
> 
> primitive vip-localB IPaddr2 \
> 
> params ip=192.168.12.216 cidr_netmask=23 nic=eth0 iflabel=0
> broadcast=192.168.13.255 \
> 
> op start interval=0s timeout=20s \
> 
> op monitor interval=5s timeout=20s \
> 
> op stop interval=0s on-fail=ignore
> 
> group groupA vip-local-checkA vip-localA changeSrcIpA cronA asteriskA \
> 
> meta target-role=Started
> 
> group groupB vip-local-checkB vip-localB changeSrcIpB cronB asteriskB \
> 
> meta
> 
> location location-groupA-avoid-failed-node groupA \
> 
> rule -inf: defined fail-count-vip-local-checkA \
> 
> rule -inf: defined fail-count-vip-localA \
> 
> rule -inf: defined fail-count-changeSrcIpA \
> 
> rule -inf: defined fail-count-cronA \
> 
> rule -inf: defined fail-count-asteriskA \
> 
> rule -inf: defined fail-count-vip-local-checkB \
> 
> rule -inf: defined fail-count-vip-localB \
> 
> rule -inf: defined fail-count-changeSrcIpB \
> 
> rule -inf: defined fail-count-cronB \
> 
> rule -inf: defined fail-count-asteriskB
> 
> location location-groupA-pm1.local-10000 groupA 10000: pm1.local
> 
> location location-groupA-pm2.local--inf groupA resource-discovery=never
> -inf: pm2.local
> 
> location location-groupA-pm3.local-1000 groupA 1000: pm3.local
> 
> location location-groupB-avoid-failed-node groupB \
> 
> rule -inf: defined fail-count-vip-local-checkA \
> 
> rule -inf: defined fail-count-vip-localA \
> 
> rule -inf: defined fail-count-changeSrcIpA \
> 
> rule -inf: defined fail-count-cronA \
> 
> rule -inf: defined fail-count-asteriskA \
> 
> rule -inf: defined fail-count-vip-local-checkB \
> 
> rule -inf: defined fail-count-vip-localB \
> 
> rule -inf: defined fail-count-changeSrcIpB \
> 
> rule -inf: defined fail-count-cronB \
> 
> rule -inf: defined fail-count-asteriskB
> 
> location location-groupB-pm1.local--inf groupB resource-discovery=never
> -inf: pm1.local
> 
> location location-groupB-pm2.local-10000 groupB 10000: pm2.local
> 
> location location-groupB-pm3.local-1000 groupB 1000: pm3.local
> 
> property cib-bootstrap-options: \
> 
> have-watchdog=false \
> 
> dc-version=1.1.14-8.el6_8.1-70404b0 \
> 
> cluster-infrastructure=cman \
> 
> stonith-enabled=false \
> 
> no-quorum-policy=ignore \
> 
> enable-startup-probes=true \
> 
> maintenance-mode=false \
> 
> startup-fencing=false \
> 
> dc-deadtime=20s \
> 
> last-lrm-refresh=1475445380 \
> 
> placement-strategy=balanced
> 
> rsc_defaults rsc_defaults-options: \
> 
> migration-threshold=1 \
> 
> resource-stickiness=INFINITY




More information about the Users mailing list