[ClusterLabs] Disabled resource is hard logging

Klaus Wenninger kwenning at redhat.com
Thu Feb 16 14:04:19 CET 2017


On 02/16/2017 01:55 PM, Oscar Segarra wrote:
> Hi Klaus, 
>
> Thanks a lot, I will try to delete the stop monitor.
>
> Nevertheless, I have 6 domains configured exactly the same... Is there
> any reason why just this domain has this behaviour ?

None I can see here at least.
Does the problem persist as is if you shut down the whole cluster and
bring it up again? Just to sort out history issues ...

>
> Thanks a lot.
>
> 2017-02-16 11:12 GMT+01:00 Klaus Wenninger <kwenning at redhat.com
> <mailto:kwenning at redhat.com>>:
>
>     On 02/16/2017 11:02 AM, Oscar Segarra wrote:
>     > Hi Kaluss
>     >
>     > Which is your proposal to fix this behavior?
>
>     First you can try to remove the monitor op for role=stopped.
>     Then the startup-probing will probably still fail but for that
>     the behaviour is different.
>     The startup-probing can be disabled globally via cluster-property
>     enable-startup-probes that is defaulting to true.
>     But be aware that the cluster then wouldn't be able to react
>     properly if services are already up when pacemaker is starting.
>     It should be possible to disable the probing on a per resource
>     or node basis as well iirc. But I can't tell you out of my mind
>     how that worked - there was a discussion a few weeks ago
>     on the list iirc.
>
>     Regards,
>     Klaus
>
>     >
>     > Thanks a lot!
>     >
>     >
>     > El 16 feb. 2017 10:57 a. m., "Klaus Wenninger"
>     <kwenning at redhat.com <mailto:kwenning at redhat.com>
>     > <mailto:kwenning at redhat.com <mailto:kwenning at redhat.com>>> escribió:
>     >
>     >     On 02/16/2017 09:05 AM, Oscar Segarra wrote:
>     >     > Hi,
>     >     >
>     >     > In my environment I have deployed 5 VirtualDomains as one can
>     >     see below:
>     >     > [root at vdicnode01 ~]# pcs status
>     >     > Cluster name: vdic-cluster
>     >     > Stack: corosync
>     >     > Current DC: vdicnode01-priv (version
>     1.1.15-11.el7_3.2-e174ec8) -
>     >     > partition with quorum
>     >     > Last updated: Thu Feb 16 09:02:53 2017          Last
>     change: Thu Feb
>     >     > 16 08:20:53 2017 by root via crm_attribute on vdicnode02-priv
>     >     >
>     >     > 2 nodes and 14 resources configured: 5 resources DISABLED
>     and 0
>     >     > BLOCKED from being started due to failures
>     >     >
>     >     > Online: [ vdicnode01-priv vdicnode02-priv ]
>     >     >
>     >     > Full list of resources:
>     >     >
>     >     >  nfs-vdic-mgmt-vm-vip   (ocf::heartbeat:IPaddr):       
>     Started
>     >     > vdicnode01-priv
>     >     >  Clone Set: nfs_setup-clone [nfs_setup]
>     >     >      Started: [ vdicnode01-priv vdicnode02-priv ]
>     >     >  Clone Set: nfs-mon-clone [nfs-mon]
>     >     >      Started: [ vdicnode01-priv vdicnode02-priv ]
>     >     >  Clone Set: nfs-grace-clone [nfs-grace]
>     >     >      Started: [ vdicnode01-priv vdicnode02-priv ]
>     >     >  vm-vdicone01   (ocf::heartbeat:VirtualDomain): FAILED
>     (disabled)[
>     >     > vdicnode02-priv vdicnode01-priv ]
>     >     >  vm-vdicsunstone01      (ocf::heartbeat:VirtualDomain): FAILED
>     >     > vdicnode01-priv (disabled)
>     >     >  vm-vdicdb01    (ocf::heartbeat:VirtualDomain): FAILED
>     (disabled)[
>     >     > vdicnode02-priv vdicnode01-priv ]
>     >     >  vm-vdicudsserver       (ocf::heartbeat:VirtualDomain): FAILED
>     >     > (disabled)[ vdicnode02-priv vdicnode01-priv ]
>     >     >  vm-vdicudstuneler      (ocf::heartbeat:VirtualDomain): FAILED
>     >     > vdicnode01-priv (disabled)
>     >     >  Clone Set: nfs-vdic-images-vip-clone [nfs-vdic-images-vip]
>     >     >      Stopped: [ vdicnode01-priv vdicnode02-priv ]
>     >     >
>     >     > Failed Actions:
>     >     > * vm-vdicone01_monitor_20000 on vdicnode02-priv 'not
>     installed' (5):
>     >     > call=2322, status=complete, exitreason='Configuration file
>     >     > /mnt/nfs-vdic-mgmt-vm/vdicone01.xml does not exist or is not
>     >     readable.',
>     >     >     last-rc-change='Thu Feb 16 09:02:07 2017', queued=0ms,
>     exec=21ms
>     >     > * vm-vdicsunstone01_monitor_20000 on vdicnode02-priv 'not
>     installed'
>     >     > (5): call=2310, status=complete, exitreason='Configuration
>     file
>     >     > /mnt/nfs-vdic-mgmt-vm/vdicsunstone01.xml does not exist or
>     is not
>     >     > readable.',
>     >     >     last-rc-change='Thu Feb 16 09:02:07 2017', queued=0ms,
>     exec=37ms
>     >     > * vm-vdicdb01_monitor_20000 on vdicnode02-priv 'not
>     installed' (5):
>     >     > call=2320, status=complete, exitreason='Configuration file
>     >     > /mnt/nfs-vdic-mgmt-vm/vdicdb01.xml does not exist or is not
>     >     readable.',
>     >     >     last-rc-change='Thu Feb 16 09:02:07 2017', queued=0ms,
>     exec=35ms
>     >     > * vm-vdicudsserver_monitor_20000 on vdicnode02-priv 'not
>     installed'
>     >     > (5): call=2321, status=complete, exitreason='Configuration
>     file
>     >     > /mnt/nfs-vdic-mgmt-vm/vdicudsserver.xml does not exist or
>     is not
>     >     > readable.',
>     >     >     last-rc-change='Thu Feb 16 09:02:07 2017', queued=0ms,
>     exec=42ms
>     >     > * vm-vdicudstuneler_monitor_20000 on vdicnode01-priv 'not
>     installed'
>     >     > (5): call=1987183, status=complete,
>     exitreason='Configuration file
>     >     > /mnt/nfs-vdic-mgmt-vm/vdicudstuneler.xml does not exist or
>     is not
>     >     > readable.',
>     >     >     last-rc-change='Thu Feb 16 04:00:25 2017', queued=0ms,
>     exec=30ms
>     >     > * vm-vdicdb01_monitor_20000 on vdicnode01-priv 'not
>     installed' (5):
>     >     > call=2550049, status=complete, exitreason='Configuration file
>     >     > /mnt/nfs-vdic-mgmt-vm/vdicdb01.xml does not exist or is not
>     >     readable.',
>     >     >     last-rc-change='Thu Feb 16 08:13:37 2017', queued=0ms,
>     exec=44ms
>     >     > * nfs-mon_monitor_10000 on vdicnode01-priv 'unknown error'
>     (1):
>     >     > call=1984009, status=Timed Out, exitreason='none',
>     >     >     last-rc-change='Thu Feb 16 04:24:30 2017', queued=0ms,
>     exec=0ms
>     >     > * vm-vdicsunstone01_monitor_20000 on vdicnode01-priv 'not
>     installed'
>     >     > (5): call=2552050, status=complete,
>     exitreason='Configuration file
>     >     > /mnt/nfs-vdic-mgmt-vm/vdicsunstone01.xml does not exist or
>     is not
>     >     > readable.',
>     >     >     last-rc-change='Thu Feb 16 08:14:07 2017', queued=0ms,
>     exec=22ms
>     >     > * vm-vdicone01_monitor_20000 on vdicnode01-priv 'not
>     installed' (5):
>     >     > call=2620052, status=complete, exitreason='Configuration file
>     >     > /mnt/nfs-vdic-mgmt-vm/vdicone01.xml does not exist or is not
>     >     readable.',
>     >     >     last-rc-change='Thu Feb 16 09:02:53 2017', queued=0ms,
>     exec=45ms
>     >     > * vm-vdicudsserver_monitor_20000 on vdicnode01-priv 'not
>     installed'
>     >     > (5): call=2550052, status=complete,
>     exitreason='Configuration file
>     >     > /mnt/nfs-vdic-mgmt-vm/vdicudsserver.xml does not exist or
>     is not
>     >     > readable.',
>     >     >     last-rc-change='Thu Feb 16 08:13:37 2017', queued=0ms,
>     exec=48ms
>     >     >
>     >     >
>     >     > Al VirtualDomain resources are configured the same:
>     >     >
>     >     > [root at vdicnode01 cluster]# pcs resource show vm-vdicone01
>     >     >  Resource: vm-vdicone01 (class=ocf provider=heartbeat
>     >     type=VirtualDomain)
>     >     >   Attributes: hypervisor=qemu:///system
>     >     > config=/mnt/nfs-vdic-mgmt-vm/vdicone01.xml
>     >     > migration_network_suffix=tcp:// migration_transport=ssh
>     >     >   Meta Attrs: allow-migrate=true target-role=Stopped
>     >     >   Utilization: cpu=1 hv_memory=512
>     >     >   Operations: start interval=0s timeout=90
>     >     > (vm-vdicone01-start-interval-0s)
>     >     >               stop interval=0s timeout=90
>     >     (vm-vdicone01-stop-interval-0s)
>     >     >               monitor interval=20s role=Stopped
>     >     > (vm-vdicone01-monitor-interval-20s)
>     >     >               monitor interval=30s
>     >     (vm-vdicone01-monitor-interval-30s)
>     >     > [root at vdicnode01 cluster]# pcs resource show vm-vdicdb01
>     >     >  Resource: vm-vdicdb01 (class=ocf provider=heartbeat
>     >     type=VirtualDomain)
>     >     >   Attributes: hypervisor=qemu:///system
>     >     > config=/mnt/nfs-vdic-mgmt-vm/vdicdb01.xml
>     >     > migration_network_suffix=tcp:// migration_transport=ssh
>     >     >   Meta Attrs: allow-migrate=true target-role=Stopped
>     >     >   Utilization: cpu=1 hv_memory=512
>     >     >   Operations: start interval=0s timeout=90
>     >     (vm-vdicdb01-start-interval-0s)
>     >     >               stop interval=0s timeout=90
>     >     (vm-vdicdb01-stop-interval-0s)
>     >     >               monitor interval=20s role=Stopped
>     >     > (vm-vdicdb01-monitor-interval-20s)
>     >     >               monitor interval=30s
>     >     (vm-vdicdb01-monitor-interval-30s)
>     >     >
>     >     >
>     >     >
>     >     > Nevertheless, one of the virtual domains is logging hardly and
>     >     > fulfilling my hard disk:
>     >     >
>     >     > VirtualDomain(vm-vdicone01)[116359]:   
>     2017/02/16_08:52:27 INFO:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not
>     readable,
>     >     > resource considered stopped.
>     >     > VirtualDomain(vm-vdicone01)[116401]:   
>     2017/02/16_08:52:27 ERROR:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml
>     does not
>     >     exist
>     >     > or is not readable.
>     >     > VirtualDomain(vm-vdicone01)[116423]:   
>     2017/02/16_08:52:27 INFO:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not
>     readable,
>     >     > resource considered stopped.
>     >     > VirtualDomain(vm-vdicone01)[116444]:   
>     2017/02/16_08:52:27 ERROR:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml
>     does not
>     >     exist
>     >     > or is not readable.
>     >     > VirtualDomain(vm-vdicone01)[116466]:   
>     2017/02/16_08:52:27 INFO:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not
>     readable,
>     >     > resource considered stopped.
>     >     > VirtualDomain(vm-vdicone01)[116487]:   
>     2017/02/16_08:52:27 ERROR:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml
>     does not
>     >     exist
>     >     > or is not readable.
>     >     > VirtualDomain(vm-vdicone01)[116509]:   
>     2017/02/16_08:52:27 INFO:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not
>     readable,
>     >     > resource considered stopped.
>     >     > VirtualDomain(vm-vdicone01)[116530]:   
>     2017/02/16_08:52:27 ERROR:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml
>     does not
>     >     exist
>     >     > or is not readable.
>     >     > VirtualDomain(vm-vdicone01)[116552]:   
>     2017/02/16_08:52:27 INFO:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not
>     readable,
>     >     > resource considered stopped.
>     >     > VirtualDomain(vm-vdicone01)[116573]:   
>     2017/02/16_08:52:27 ERROR:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml
>     does not
>     >     exist
>     >     > or is not readable.
>     >     > VirtualDomain(vm-vdicone01)[116595]:   
>     2017/02/16_08:52:27 INFO:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not
>     readable,
>     >     > resource considered stopped.
>     >     > VirtualDomain(vm-vdicone01)[116616]:   
>     2017/02/16_08:52:27 ERROR:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml
>     does not
>     >     exist
>     >     > or is not readable.
>     >     > VirtualDomain(vm-vdicone01)[116638]:   
>     2017/02/16_08:52:27 INFO:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not
>     readable,
>     >     > resource considered stopped.
>     >     > VirtualDomain(vm-vdicone01)[116659]:   
>     2017/02/16_08:52:27 ERROR:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml
>     does not
>     >     exist
>     >     > or is not readable.
>     >     > VirtualDomain(vm-vdicone01)[116681]:   
>     2017/02/16_08:52:27 INFO:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml not
>     readable,
>     >     > resource considered stopped.
>     >     > VirtualDomain(vm-vdicone01)[116702]:   
>     2017/02/16_08:52:27 ERROR:
>     >     > Configuration file /mnt/nfs-vdic-mgmt-vm/vdicone01.xml
>     does not
>     >     exist
>     >     > or is not readable.
>     >     > [root at vdicnode01 cluster]# pcs status
>     >     >
>     >     >
>     >     > Note! Is it normal the error as I have not mounted the nfs
>     >     > resource /mnt/nfs-vdic-mgmt-vm/vdicone01.xml yet.
>     >
>     >     Well that is probably the explanation already:
>     >     The resource should be stopped. The config-file is not
>     available.
>     >     But the resource needs the config-file to verify that it is
>     really
>     >     stopped.
>     >     So the probe is failing and as you have a monitoring op for
>     >     role="stopped" it
>     >     is doing that over and over again.
>     >
>     >     >
>     >     > ¿Is there any explanation for this heavy logging?
>     >     >
>     >     > Thanks a lot!
>     >     >
>     >     >
>     >     > _______________________________________________
>     >     > Users mailing list: Users at clusterlabs.org
>     <mailto:Users at clusterlabs.org>
>     >     <mailto:Users at clusterlabs.org <mailto:Users at clusterlabs.org>>
>     >     > http://lists.clusterlabs.org/mailman/listinfo/users
>     <http://lists.clusterlabs.org/mailman/listinfo/users>
>     >     <http://lists.clusterlabs.org/mailman/listinfo/users
>     <http://lists.clusterlabs.org/mailman/listinfo/users>>
>     >     >
>     >     > Project Home: http://www.clusterlabs.org
>     >     > Getting started:
>     >     http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>     <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf>
>     >     <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>     <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf>>
>     >     > Bugs: http://bugs.clusterlabs.org
>     >
>     >
>     >
>     >     _______________________________________________
>     >     Users mailing list: Users at clusterlabs.org
>     <mailto:Users at clusterlabs.org>
>     >     <mailto:Users at clusterlabs.org <mailto:Users at clusterlabs.org>>
>     >     http://lists.clusterlabs.org/mailman/listinfo/users
>     <http://lists.clusterlabs.org/mailman/listinfo/users>
>     >     <http://lists.clusterlabs.org/mailman/listinfo/users
>     <http://lists.clusterlabs.org/mailman/listinfo/users>>
>     >
>     >     Project Home: http://www.clusterlabs.org
>     >     Getting started:
>     >     http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>     <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf>
>     >     <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>     <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf>>
>     >     Bugs: http://bugs.clusterlabs.org
>     >
>     >
>
>
>     _______________________________________________
>     Users mailing list: Users at clusterlabs.org
>     <mailto:Users at clusterlabs.org>
>     http://lists.clusterlabs.org/mailman/listinfo/users
>     <http://lists.clusterlabs.org/mailman/listinfo/users>
>
>     Project Home: http://www.clusterlabs.org
>     Getting started:
>     http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>     <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf>
>     Bugs: http://bugs.clusterlabs.org
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.clusterlabs.org/pipermail/users/attachments/20170216/5c6785f5/attachment-0001.html>


More information about the Users mailing list