[ClusterLabs] Cannot clone clvmd resource
Anne Nicolas
ennael1 at gmail.com
Wed Mar 1 18:09:40 EST 2017
Le 01/03/2017 à 23:20, Ken Gaillot a écrit :
> On 03/01/2017 03:49 PM, Anne Nicolas wrote:
>> Hi there
>>
>>
>> I'm testing quite an easy configuration to work on clvm. I'm just
>> getting crazy as it seems clmd cannot be cloned on other nodes.
>>
>> clvmd start well on node1 but fails on both node2 and node3.
>
> Your config looks fine, so I'm going to guess there's some local
> difference on the nodes.
>
>> In pacemaker journalctl I get the following message
>> Mar 01 16:34:36 node3 pidofproc[27391]: pidofproc: cannot stat /clvmd:
>> No such file or directory
>> Mar 01 16:34:36 node3 pidofproc[27392]: pidofproc: cannot stat
>> /cmirrord: No such file or directory
>
> I have no idea where the above is coming from. pidofproc is an LSB
> function, but (given journalctl) I'm assuming you're using systemd. I
> don't think anything in pacemaker or resource-agents uses pidofproc (at
> least not currently, not sure about the older version you're using).
Thanks for your feedback. I finally checked the RA script and found the
error
in clvm2 RA script on non working nodes I got
# Common variables
DAEMON="${sbindir}/clvmd"
CMIRRORD="${sbindir}/cmirrord"
LVMCONF="${sbindir}/lvmconf"
on working node
DAEMON="/usr/sbin/clvmd"
CMIRRORD="/usr/sbin/cmirrord"
Looks like it was path variables were not interpreted. I just have to
check why I did get those versions
THanks again for your answer
>
>> Mar 01 16:34:36 node3 lrmd[2174]: notice: finished - rsc:p-clvmd
>> action:stop call_id:233 pid:27384 exit-code:0 exec-time:45ms queue-time:0ms
>> Mar 01 16:34:36 node3 crmd[2177]: notice: Operation p-clvmd_stop_0: ok
>> (node=node3, call=233, rc=0, cib-update=541, confirmed=true)
>> Mar 01 16:34:36 node3 crmd[2177]: notice: Initiating action 72: stop
>> p-dlm_stop_0 on node3 (local)
>> Mar 01 16:34:36 node3 lrmd[2174]: notice: executing - rsc:p-dlm
>> action:stop call_id:235
>> Mar 01 16:34:36 node3 crmd[2177]: notice: Initiating action 67: stop
>> p-dlm_stop_0 on node2
>>
>> Here is my configuration
>>
>> node 739312139: node1
>> node 739312140: node2
>> node 739312141: node3
>> primitive admin_addr IPaddr2 \
>> params ip=172.17.2.10 \
>> op monitor interval=10 timeout=20 \
>> meta target-role=Started
>> primitive p-clvmd ocf:lvm2:clvmd \
>> op start timeout=90 interval=0 \
>> op stop timeout=100 interval=0 \
>> op monitor interval=30 timeout=90
>> primitive p-dlm ocf:pacemaker:controld \
>> op start timeout=90 interval=0 \
>> op stop timeout=100 interval=0 \
>> op monitor interval=60 timeout=90
>> primitive stonith-sbd stonith:external/sbd
>> group g-clvm p-dlm p-clvmd
>> clone c-clvm g-clvm meta interleave=true
>> property cib-bootstrap-options: \
>> have-watchdog=true \
>> dc-version=1.1.13-14.7-6f22ad7 \
>> cluster-infrastructure=corosync \
>> cluster-name=hacluster \
>> stonith-enabled=true \
>> placement-strategy=balanced \
>> no-quorum-policy=freeze \
>> last-lrm-refresh=1488404073
>> rsc_defaults rsc-options: \
>> resource-stickiness=1 \
>> migration-threshold=10
>> op_defaults op-options: \
>> timeout=600 \
>> record-pending=true
>>
>> Thanks in advance for your input
>>
>> Cheers
>>
>
>
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://lists.clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
--
Anne Nicolas
http://mageia.org
More information about the Users
mailing list