[ClusterLabs] Cannot clone clvmd resource

Ken Gaillot kgaillot at redhat.com
Wed Mar 1 22:20:08 UTC 2017


On 03/01/2017 03:49 PM, Anne Nicolas wrote:
> Hi there
> 
> 
> I'm testing quite an easy configuration to work on clvm. I'm just
> getting crazy as it seems clmd cannot be cloned on other nodes.
> 
> clvmd start well on node1 but fails on both node2 and node3.

Your config looks fine, so I'm going to guess there's some local
difference on the nodes.

> In pacemaker journalctl I get the following message
> Mar 01 16:34:36 node3 pidofproc[27391]: pidofproc: cannot stat /clvmd:
> No such file or directory
> Mar 01 16:34:36 node3 pidofproc[27392]: pidofproc: cannot stat
> /cmirrord: No such file or directory

I have no idea where the above is coming from. pidofproc is an LSB
function, but (given journalctl) I'm assuming you're using systemd. I
don't think anything in pacemaker or resource-agents uses pidofproc (at
least not currently, not sure about the older version you're using).

> Mar 01 16:34:36 node3 lrmd[2174]: notice: finished - rsc:p-clvmd
> action:stop call_id:233 pid:27384 exit-code:0 exec-time:45ms queue-time:0ms
> Mar 01 16:34:36 node3 crmd[2177]: notice: Operation p-clvmd_stop_0: ok
> (node=node3, call=233, rc=0, cib-update=541, confirmed=true)
> Mar 01 16:34:36 node3 crmd[2177]: notice: Initiating action 72: stop
> p-dlm_stop_0 on node3 (local)
> Mar 01 16:34:36 node3 lrmd[2174]: notice: executing - rsc:p-dlm
> action:stop call_id:235
> Mar 01 16:34:36 node3 crmd[2177]: notice: Initiating action 67: stop
> p-dlm_stop_0 on node2
> 
> Here is my configuration
> 
> node 739312139: node1
> node 739312140: node2
> node 739312141: node3
> primitive admin_addr IPaddr2 \
>         params ip=172.17.2.10 \
>         op monitor interval=10 timeout=20 \
>         meta target-role=Started
> primitive p-clvmd ocf:lvm2:clvmd \
>         op start timeout=90 interval=0 \
>         op stop timeout=100 interval=0 \
>         op monitor interval=30 timeout=90
> primitive p-dlm ocf:pacemaker:controld \
>         op start timeout=90 interval=0 \
>         op stop timeout=100 interval=0 \
>         op monitor interval=60 timeout=90
> primitive stonith-sbd stonith:external/sbd
> group g-clvm p-dlm p-clvmd
> clone c-clvm g-clvm meta interleave=true
> property cib-bootstrap-options: \
>         have-watchdog=true \
>         dc-version=1.1.13-14.7-6f22ad7 \
>         cluster-infrastructure=corosync \
>         cluster-name=hacluster \
>         stonith-enabled=true \
>         placement-strategy=balanced \
>         no-quorum-policy=freeze \
>         last-lrm-refresh=1488404073
> rsc_defaults rsc-options: \
>         resource-stickiness=1 \
>         migration-threshold=10
> op_defaults op-options: \
>         timeout=600 \
>         record-pending=true
> 
> Thanks in advance for your input
> 
> Cheers
> 





More information about the Users mailing list