[pve-devel] Bug 1458 - PVE 5 live migration downtime degraded to several seconds (compared to PVE 4)
Alexandre DERUMIER
aderumier at odiso.com
Thu Jul 27 19:41:43 CEST 2017
I have done ping test, on proxmox 4, with migrate_insecure (so "cont" don't apply at vm_start),
without disk migration
and I'm around 60ms of loss.
Will try proxmox 5 tomorrow.
----- Mail original -----
De: "aderumier" <aderumier at odiso.com>
À: "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Jeudi 27 Juillet 2017 16:30:37
Objet: Re: [pve-devel] Bug 1458 - PVE 5 live migration downtime degraded to several seconds (compared to PVE 4)
looking at user migration log:
Jul 24 18:12:37 start migrate command to unix:/run/qemu-server/100.migrate
Jul 24 18:12:39 migration speed: 256.00 MB/s - downtime 39 ms
Seem that the vm have very low memory, as migration take 2second between the begin and the end.
so maybe the usleep lowering is not working here.
----- Mail original -----
De: "aderumier" <aderumier at odiso.com>
À: "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Jeudi 27 Juillet 2017 16:08:35
Objet: Re: [pve-devel] Bug 1458 - PVE 5 live migration downtime degraded to several seconds (compared to PVE 4)
Thanks for the explain Fabian. (I'm always using migration insecure, so I didn't notice this bug)
>>when live-migrating over a unix socket, PVE 5 takes up to a few seconds
>>between completing the RAM transfer and pausing the source VM, and
>>resuming the target VM. in PVE 4, the same migration has a downtime of
>>almost 0.
few seconds seem so huge ... (user talk about 4s)....
>>AFAICT, the reason for this is a bug fix in PVE 5's qemu-server which
>>was required to support storage live migration in Qemu 2.9.
any commit reference ?
>>originally in PVE 4, the target VM in a live migration was started in
>>incoming migration mode and NOT continued on startup (whereas VMs rolled
>>back to a RAM snapshot where started in the same mode, but immediately
>>continued).
>>in June 2016[3], migration over ssh-forwarded unix sockets was
>>implemented. the check for skipping the continue command on startup of
>>the target VM was overlooked, so now VMs migrated over unix sockets were
>>started in incoming migration mode, but continued on startup.
But this seem to be a bug, fixed later here ?
https://git.proxmox.com/?p=qemu-server.git;a=commit;h=b37ecfe6ae7f7b557db7712ee6988cb0397306e9
>>I wonder whether going the "immediately cont" route for live migrations
>>without local storage can cause any issues besides the obvious "moving
>>the conf file failed and VM is now active on the wrong node" one?
I don't known if it could be great to have some kind of temporary conf file where a kvm process is running.
(here we could see vm on source host with state running, and vm on source target with state migrating for example).
Like this if something bad happen at the end of migration, user could still stop the target kvm process with gui.
But maybe it's too complex to implement, don't known...
>>if not, I propose doing just that. otherwise, we could think about lowering
>>the polling interval when waiting for RAM migration to complete (in
>>phase2) - that should shave off a bit of the downtime as well.
I wonder where exactly it take so much time..
$downtime seem to be low, but as it's coming from status, maybe are we missing some query migrate .
Also I think we already try to lowering usleep at the end
#reduce sleep if remainig memory if lower than the everage transfert
$usleep = 300000 if $avglstat && $rem < $avglstat;
maybe this don't work correctly ?
I think a proper way could be catch qemu events, instead pooling status. (but require maybe lot of work)
----- Mail original -----
De: "Fabian Grünbichler" <f.gruenbichler at proxmox.com>
À: "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Jeudi 27 Juillet 2017 14:45:43
Objet: [pve-devel] Bug 1458 - PVE 5 live migration downtime degraded to several seconds (compared to PVE 4)
the following issue was reported on the forum[1] and as bug #1458[2],
moving this here for further discussion of potential fixes.
when live-migrating over a unix socket, PVE 5 takes up to a few seconds
between completing the RAM transfer and pausing the source VM, and
resuming the target VM. in PVE 4, the same migration has a downtime of
almost 0.
AFAICT, the reason for this is a bug fix in PVE 5's qemu-server which
was required to support storage live migration in Qemu 2.9.
originally in PVE 4, the target VM in a live migration was started in
incoming migration mode and NOT continued on startup (whereas VMs rolled
back to a RAM snapshot where started in the same mode, but immediately
continued).
in June 2016[3], migration over ssh-forwarded unix sockets was
implemented. the check for skipping the continue command on startup of
the target VM was overlooked, so now VMs migrated over unix sockets were
started in incoming migration mode, but continued on startup. this does
not change the behaviour on startup, as a VM in incoming migration mode
is not actually running until a migration has happened. this does mean
that the downtime is vastly reduced for such migrations, as Qemu will
continue the target VM automatically as soon as the migration job is
completed.
the only things that happen after this automatic resume is
- finish tunnel
- moving the conf file logically between nodes
- resuming on the target side (which is a no-op in this case)
so the risk for inconsistencies seems pretty small.
later on, we introduced live-storage migration. in those cases, we now
have the following scenario:
- start storage migration jobs
- start RAM migration
- wait for RAM to be completed
- finish tunnel
- finish block jobs
- update conf file
- move the conf file logically between the nodes
- resume on target node
so depending on whether the migration goes over tcp (OK) or over unix
(not so much) we have very different behaviour and risk for
inconsistencies.
with the introduction to PVE 5, this different behaviour was fixed /
made consistent, by adapting the "manual resume" stance. this was needed
because Qemu 2.9 does not allow the storage migration over NBD and the
target VM itself to have write access to the same disks at the same
time. this fix was not backported to PVE 4, which means that storage
live-migration is potentially buggy there, but live-migration over unix
sockets is faster.
I wonder whether going the "immediately cont" route for live migrations
without local storage can cause any issues besides the obvious "moving
the conf file failed and VM is now active on the wrong node" one? if
not, I propose doing just that. otherwise, we could think about lowering
the polling interval when waiting for RAM migration to complete (in
phase2) - that should shave off a bit of the downtime as well.
in any case, I think we need to backport the manual resume in case of
local storage live migration fix to PVE 4.
1: https://forum.proxmox.com/threads/pve-5-live-migration-downtime-degradation-2-4-sec.35890
2: https://bugzilla.proxmox.com/show_bug.cgi?id=1458
3: 1c9d54bfd05e0d017a6e2ac5524d75466b1a4455
_______________________________________________
pve-devel mailing list
pve-devel at pve.proxmox.com
https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel
_______________________________________________
pve-devel mailing list
pve-devel at pve.proxmox.com
https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel
_______________________________________________
pve-devel mailing list
pve-devel at pve.proxmox.com
https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel
More information about the pve-devel
mailing list