[ClusterLabs] HA problem: No live migration when setting node on standby
Vladislav Bogdanov
bubble at hoster-ok.com
Wed Apr 12 06:21:47 EDT 2023
Hi,
Just add a Master role for drbd resource in the colocation. Default is
Started (or Slave).
Philip Schiller <p.schiller at plusoptix.de> 12 апреля 2023 г. 11:28:57 написал:
> Hi All, I am using a simple two-nodes cluster with Zvol -> DRBD -> Virsh in
> primary/primary mode (necessary for live migration). My configuration:
> primitive pri-vm-alarmanlage VirtualDomain \ params
> config="/etc/libvirt/qemu/alarmanlage.xml" hypervisor="qemu:///system"
> migration_transport=ssh \ meta allow-migrate=true target-role=Started
> is-managed=true \ op monitor interval=0 timeout=120 \ op start interval=0
> timeout=120 \ op stop interval=0 timeout=1800 \ op migrate_to interval=0
> timeout=1800 \ op migrate_from interval=0 timeout=1800 \ utilization cpu=2
> hv_memory=4096
> ms mas-drbd-alarmanlage pri-drbd-alarmanlage \ meta clone-max=2
> promoted-max=2 notify=true promoted-node-max=1 clone-node-max=1
> interleave=true target-role=Started is-managed=true
> colocation colo_mas_drbd_alarmanlage_with_clo_pri_zfs_drbd-storage inf:
> mas-drbd-alarmanlage clo-pri-zfs-drbd_storage
> location location-pri-vm-alarmanlage-s0-200 pri-vm-alarmanlage 200: s1
> order ord_pri-alarmanlage-after-mas-drbd-alarmanlage Mandatory:
> mas-drbd-alarmanlage:promote pri-vm-alarmanlage:start So to summerize:
> - A resource for Virsh
> - A Master/Slave DRBD ressources for the VM filesystem . - a "order"
> directive to start the VM after drbd has been promoted. Node startup is ok,
> the VM is started after DRBD is promoted.
> Migration with virsh or over crm <crm resource move pri-vm-alarmanlage s0>
> works fine. Node standby is problematic. Assuming the Virsh VM runs on node
> s1 : When puting node s1 in standby when node s0 is active, a live
> migration is started, BUT in the same second, pacemaker tries to demote
> DRBD volumes on s1 (while live migration is in progress). All this results
> in "stopping the vm" on s1 and starting the "vm on s0". I do not understand
> why pacemaker does demote/stop DRBD volumes before VM is migrated.
> Do i need additional constraints? Setup is done with - Corosync Cluster
> Engine, version '3.1.6'
> - Pacemaker 2.1.2
> - Ubuntu 22.04.2 LTS Thanks for your help, with kind regards Philip
>
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20230412/f14d2834/attachment.htm>
More information about the Users
mailing list