[ClusterLabs] Antw: [EXT] VirtualDomain - started but... not really

lejeczek peljasz at yahoo.co.uk
Tue Dec 14 09:48:36 EST 2021


> Hi!
>
> My guess is that you checked the corresponding logs already; why not show them here?
> I can imagine that the VMs die rather early after start.
>
> Regards,
> Ulrich
>
>>>> lejeczek via Users <users at clusterlabs.org> schrieb am 10.12.2021 um 17:33 in
> Nachricht <df8eac8f-a58e-28e5-53b5-73eb1fe432b2 at yahoo.co.uk>:
>> Hi guys.
>>
>> I quite often.. well, to frequently in my mind, see a VM
>> which cluster says:
>> -> $ pcs resource status | grep -v disabled
>> ...
>>     * c8kubermaster2    (ocf::heartbeat:VirtualDomain):
>>    Started dzien
>> ..
>>
>> but that is false, also cluster itself confirms it:
>> -> $ pcs resource debug-monitor c8kubermaster2
>> crm_resource: Error performing operation: Not running
>> Operation force-check for c8kubermaster2
>> (ocf:heartbeat:VirtualDomain) returned: 'not running' (7)
>>
>> What is the issue here, might be & how best to troubleshoot it?
>>
>> -> $ pcs resource config c8kubermaster2
>>    Resource: c8kubermaster2 (class=ocf provider=heartbeat
>> type=VirtualDomain)
>>     Attributes:
>> config=/var/lib/pacemaker/conf.d/c8kubermaster2.xml
>> hypervisor=qemu:///system migration_transport=ssh
>>     Meta Attrs: allow-migrate=true failure-timeout=30s
>>     Operations: migrate_from interval=0s timeout=180s
>> (c8kubermaster2-migrate_from-interval-0s)
>>                 migrate_to interval=0s timeout=180s
>> (c8kubermaster2-migrate_to-interval-0s)
>>                 monitor interval=30s
>> (c8kubermaster2-monitor-interval-30s)
>>                 start interval=0s timeout=90s
>> (c8kubermaster2-start-interval-0s)
>>                 stop interval=0s timeout=90s
>> (c8kubermaster2-stop-interval-0s)
>>
>> many thanks, L.
>> _______________________________________________
>> Manage your subscription:
>> https://lists.clusterlabs.org/mailman/listinfo/users
>>
>> ClusterLabs home: https://www.clusterlabs.org/
Not much there in the logs I could see (which is probably 
why cluster decides the resource is okey)
What is the resource's monitor for it not for that exactly - 
to check the state of resource - whether it dies early or 
late should not matter.
What suffices in order to "fix" such resource 
false-positive, I do quick dis/enable the resource or as in 
this very instance rpm updates which restarted node.
Again, how cluster might think resource is okey while 
debug-monitor shows it's not.
I only do not know how to reproduce this in a controlled, 
orderly manner.

thanks, L


More information about the Users mailing list