[ClusterLabs] Pacemaker
Ken Gaillot
kgaillot at redhat.com
Wed Nov 2 17:21:29 UTC 2016
On 11/01/2016 02:31 AM, Siwakoti, Ganesh wrote:
> Hi,
>
>
> i'm using CentOS release 6.8 (Final) as a KVM and i configured 3
> nodes(PM1.local,PM2.local and PM3.local), and using
> CMAN clustering. Resources running at two nodes as Active node then
> another one node is for Fail-over resource as a Standby node. Configured
> resources are groupA and groupB, 5 resources are configured in each
> resource group. groupA is Basically run at PM1.local and groupB at
> PM2.local. If fail any resource group it'll fail-over at PM3.local. On
> PM3.local can run only one Resource group at a time. groupA should not
> run at PM2.local and groupB should not run at PM1.local. groupA or
> groupB resources not move back on own nodes from PM3.local
> automatically, if need to move back on own nodes, can move by
> manually.someone pease help me.
Your configuration looks good to me. I don't think you need the -inf
rules for fail-counts; migration-threshold=1 handles that. Are you
seeing any problems?
>
>
> version information:
>
> CentOS release 6.8 (Final) (KVM)
>
> Pacemaker 1.1.14-8.el6_8.1
>
> Cman Version: 6.2.0
>
>
> I want to be sure that my configuration will work properly or not.
>
>
> my Cluster configuration and crm_configuration are on below:
>
>
> cluster.conf:
>
> <?xml version="1.0"?>
>
> <cluster config_version="23" name="clusterpm666">
>
> <fence_daemon/>
>
> <clusternodes>
>
> <clusternode name="pm1.local" nodeid="1">
>
> <fence>
>
> <method name="pcmk-redirect">
>
> <device name="pcmk"
> port="pm1.local"/>
>
> </method>
>
> </fence>
>
> <altname name="192.168.14.210"/>
>
> </clusternode>
>
> <clusternode name="pm2.local" nodeid="2">
>
> <fence>
>
> <method name="pcmk-redirect">
>
> <device name="pcmk"
> port="pm2.local"/>
>
> </method>
>
> </fence>
>
> <altname name="192.168.14.211"/>
>
> </clusternode>
>
> <clusternode name="pm3.local" nodeid="3">
>
> <fence>
>
> <method name="pcmk-redirect">
>
> <device name="pcmk"
> port="pm3.local"/>
>
> </method>
>
> </fence>
>
> <altname name="192.168.14.212"/>
>
> </clusternode>
>
> </clusternodes>
>
> <fencedevices>
>
> <fencedevice agent="fence_pcmk" name="pcmk"/>
>
> </fencedevices>
>
> <rm>
>
> <failoverdomains/>
>
> <resources/>
>
> </rm>
>
> </cluster>
>
>
> crm_configuration:
>
>
> node pm1.local \
>
> attributes \
>
> utilization capacity=1
>
> node pm2.local \
>
> attributes \
>
> utilization capacity=1
>
> node pm3.local \
>
> attributes \
>
> utilization capacity=1
>
> primitive asteriskA asterisk \
>
> params binary="/usr/sbin/asterisk" canary_binary=astcanary
> additional_parameters="-vvvg -I" config="/etc/asteri
>
> sk_pm1/asterisk.conf" user=root group=root additional_parameters="-vvvg
> -I" realtime=true maxfiles=32768 \
>
> meta migration-threshold=2 \
>
> op monitor interval=20s start-delay=5s timeout=30s \
>
> op stop interval=0s on-fail=ignore
>
> primitive asteriskB asterisk \
>
> params binary="/usr/sbin/asterisk" canary_binary=astcanary
> additional_parameters="-vvvg -I" config="/etc/asteri
>
> sk_pm2/asterisk.conf" user=root group=root additional_parameters="-vvvg
> -I" realtime=true maxfiles=32768 \
>
> meta migration-threshold=2 \
>
> op monitor interval=20s start-delay=5s timeout=30s \
>
> op stop interval=0s on-fail=ignore
>
> primitive changeSrcIpA ocf:pacemaker:changeSrcIp \
>
> params vip=192.168.12.215 mask=23 device=eth0 \
>
> op start interval=0s timeout=0 \
>
> op monitor interval=10s \
>
> op stop interval=0s on-fail=ignore
>
> primitive changeSrcIpB ocf:pacemaker:changeSrcIp \
>
> params vip=192.168.12.216 mask=23 device=eth0 \
>
> op start interval=0s timeout=0 \
>
> op monitor interval=10s \
>
> op stop interval=0s on-fail=ignore
>
> primitive cronA lsb:crond \
>
> meta migration-threshold=2 \
>
> op monitor interval=20s start-delay=5s timeout=15s \
>
> op stop interval=0s on-fail=ignore
>
> primitive cronB lsb:crond \
>
> meta migration-threshold=2 \
>
> op monitor interval=20s start-delay=5s timeout=15s \
>
> op stop interval=0s on-fail=ignore
>
> primitive vip-local-checkA VIPcheck \
>
> params target_ip=192.168.12.215 count=1 wait=5 \
>
> op start interval=0s on-fail=restart timeout=60s \
>
> op monitor interval=10s timeout=60s \
>
> op stop interval=0s on-fail=ignore timeout=60s \
>
> utilization capacity=1
>
> primitive vip-local-checkB VIPcheck \
>
> params target_ip=192.168.12.216 count=1 wait=5 \
>
> op start interval=0s on-fail=restart timeout=60s \
>
> op monitor interval=10s timeout=60s \
>
> op stop interval=0s on-fail=ignore timeout=60s \
>
> utilization capacity=1
>
> primitive vip-localA IPaddr2 \
>
> params ip=192.168.12.215 cidr_netmask=23 nic=eth0 iflabel=0
> broadcast=192.168.13.255 \
>
> op start interval=0s timeout=20s \
>
> op monitor interval=5s timeout=20s \
>
> op stop interval=0s on-fail=ignore
>
> primitive vip-localB IPaddr2 \
>
> params ip=192.168.12.216 cidr_netmask=23 nic=eth0 iflabel=0
> broadcast=192.168.13.255 \
>
> op start interval=0s timeout=20s \
>
> op monitor interval=5s timeout=20s \
>
> op stop interval=0s on-fail=ignore
>
> group groupA vip-local-checkA vip-localA changeSrcIpA cronA asteriskA \
>
> meta target-role=Started
>
> group groupB vip-local-checkB vip-localB changeSrcIpB cronB asteriskB \
>
> meta
>
> location location-groupA-avoid-failed-node groupA \
>
> rule -inf: defined fail-count-vip-local-checkA \
>
> rule -inf: defined fail-count-vip-localA \
>
> rule -inf: defined fail-count-changeSrcIpA \
>
> rule -inf: defined fail-count-cronA \
>
> rule -inf: defined fail-count-asteriskA \
>
> rule -inf: defined fail-count-vip-local-checkB \
>
> rule -inf: defined fail-count-vip-localB \
>
> rule -inf: defined fail-count-changeSrcIpB \
>
> rule -inf: defined fail-count-cronB \
>
> rule -inf: defined fail-count-asteriskB
>
> location location-groupA-pm1.local-10000 groupA 10000: pm1.local
>
> location location-groupA-pm2.local--inf groupA resource-discovery=never
> -inf: pm2.local
>
> location location-groupA-pm3.local-1000 groupA 1000: pm3.local
>
> location location-groupB-avoid-failed-node groupB \
>
> rule -inf: defined fail-count-vip-local-checkA \
>
> rule -inf: defined fail-count-vip-localA \
>
> rule -inf: defined fail-count-changeSrcIpA \
>
> rule -inf: defined fail-count-cronA \
>
> rule -inf: defined fail-count-asteriskA \
>
> rule -inf: defined fail-count-vip-local-checkB \
>
> rule -inf: defined fail-count-vip-localB \
>
> rule -inf: defined fail-count-changeSrcIpB \
>
> rule -inf: defined fail-count-cronB \
>
> rule -inf: defined fail-count-asteriskB
>
> location location-groupB-pm1.local--inf groupB resource-discovery=never
> -inf: pm1.local
>
> location location-groupB-pm2.local-10000 groupB 10000: pm2.local
>
> location location-groupB-pm3.local-1000 groupB 1000: pm3.local
>
> property cib-bootstrap-options: \
>
> have-watchdog=false \
>
> dc-version=1.1.14-8.el6_8.1-70404b0 \
>
> cluster-infrastructure=cman \
>
> stonith-enabled=false \
>
> no-quorum-policy=ignore \
>
> enable-startup-probes=true \
>
> maintenance-mode=false \
>
> startup-fencing=false \
>
> dc-deadtime=20s \
>
> last-lrm-refresh=1475445380 \
>
> placement-strategy=balanced
>
> rsc_defaults rsc_defaults-options: \
>
> migration-threshold=1 \
>
> resource-stickiness=INFINITY
More information about the Users
mailing list