[pve-devel] [PATCH qemu-server] cfg2cmd: turn smm off when SeaBIOS and serial display are used
Fabian Ebner
f.ebner at proxmox.com
Mon Nov 8 12:36:02 CET 2021
Am 05.11.21 um 15:47 schrieb Fabian Ebner:
> Am 05.11.21 um 14:12 schrieb Thomas Lamprecht:
>> On 05.11.21 14:06, Fabian Ebner wrote:
>>> Since commit 277d33454f77ec1d1e0bc04e37621e4dd2424b67 in pve-qemu,
>>> smm=off is no longer the default, but with SeaBIOS and serial display,
>>> this can lead to a boot loop.
>>>
>>> Reported in the community forum [0] and reproduced with a Debian 10
>>> VM.
>>>
>>> [0]:
>>> https://forum.proxmox.com/threads/pve-7-0-all-vms-with-cloud-init-seabios-fail-during-boot-process-bootloop-disk-not-found.97310/post-427129
>>>
>>>
>>> Signed-off-by: Fabian Ebner <f.ebner at proxmox.com>
>>> ---
>>> PVE/QemuServer.pm | 12 ++++++++++
>>> test/cfg2cmd/seabios_serial.conf | 16 ++++++++++++++
>>> test/cfg2cmd/seabios_serial.conf.cmd | 33 ++++++++++++++++++++++++++++
>>> 3 files changed, 61 insertions(+)
>>> create mode 100644 test/cfg2cmd/seabios_serial.conf
>>> create mode 100644 test/cfg2cmd/seabios_serial.conf.cmd
>>>
>>> diff --git a/PVE/QemuServer.pm b/PVE/QemuServer.pm
>>> index 763c412..9b76512 100644
>>> --- a/PVE/QemuServer.pm
>>> +++ b/PVE/QemuServer.pm
>>> @@ -3403,6 +3403,16 @@ my sub get_cpuunits {
>>> my ($conf) = @_;
>>> return $conf->{cpuunits} // (PVE::CGroup::cgroup_mode() == 2 ?
>>> 100 : 1024);
>>> }
>>> +
>>> +# Since commit 277d33454f77ec1d1e0bc04e37621e4dd2424b67 in pve-qemu,
>>> smm is not off by default
>>> +# anymore. But smm=off seems to be required when using SeaBIOS and
>>> serial display.
>>> +my sub should_disable_smm {
>>> + my ($conf, $vga) = @_;
>>> +
>>> + return (!defined($conf->{bios}) || $conf->{bios} eq 'seabios') &&
>>> + $vga->{type} && $vga->{type} =~ m/^serial\d+$/;
>>> +}
>>> +
>>> sub config_to_command {
>>> my ($storecfg, $vmid, $conf, $defaults, $forcemachine, $forcecpu,
>>> $pbs_backing) = @_;
>>> @@ -4002,6 +4012,8 @@ sub config_to_command {
>>> push @$machineFlags, 'accel=tcg';
>>> }
>>> + push @$machineFlags, 'smm=off' if should_disable_smm($conf, $vga);
>>
>> doesn't that breaks live migration or do we know that it could never
>> work with smm=on
>> under seabios + display=serial, so that there can no running VM that
>> could be live
>> migrated?
No, we can't say that. I found that the same configuration does work
with CentOS 7.
>>
>
> Stefan tested migration in both direction when the pve-qemu patch was
> dropped [0], but he did sound more confident about the smm=off -> smm=on
> direction. And now the scenario Wolfgang mentioned would actually be
> relevant, because we can get into a situation where smm=off -> smm=on ->
> smm=off by migrating from older -> old -> new. I'll test around some
> more and see if I can break it.
>
> [0]: https://lists.proxmox.com/pipermail/pve-devel/2021-August/049778.html
>
With the following nodes:
N1: pve-qemu 6.0.0-3 -> smm=off
N2: pve-qemu 6.1.0-1 -> smm=on
N3: patched qemu-server to always turn smm off -> smm=off
I tested:
1. Migration N1 -> N2
2. soft reboot
3. migration N2 -> N3
4. soft reboot
For Windows 10 and CentOS 7 everything seemed to work. I didn't test
Windows with serial output as I don't know how to set that up.
Still, the same might not be true for more complex (virtual) hardware
setups.
For the problematic Debian 10 VM, step 2 exhibits the problem, i.e.
leading to a boot loop. When I skipped step 2, it seemed to work too.
If there is a problem with live migration smm=on -> smm=off, it's at
least not obvious and would only affect VMs running pve-qemu >= 6.0.0-4
(and SeaBIOS with serial display). On the other hand, the patch does fix
a user-reported regression from pve-qemu 6.0.0-4 (although it's also not
clear how widespread it is).
>
>
> _______________________________________________
> pve-devel mailing list
> pve-devel at lists.proxmox.com
> https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel
>
>
More information about the pve-devel
mailing list