[pve-devel] avoiding VMID reuse

Alexandre DERUMIER aderumier at odiso.com
Fri Apr 13 04:04:22 CEST 2018


Hi,

I'm jumping in the conversion.

I personally think that it could be great that proxmox use UUID for vmid,
to be sure that we have unique id across different cluster, or when we delete/recreate a vm with same id on same cluster.

for example, currently, if we delete and recreate a vm with same id, new vm have access to old vm backups.
(think of multi-users security, that's pretty bad).


I'm already using custom ceph backup script (with diff backup), using UUID of smbios for qemu machine. but for lxc, we don't have unique UUID.


I don't known if it's a big task to switch to UUID, or make both working at the same time for vmid management.

Regards,

Alexandre



----- Mail original -----
De: "Lauri Tirkkonen" <lauri at tuxera.com>
À: "pve-devel" <pve-devel at pve.proxmox.com>, "Thomas Lamprecht" <t.lamprecht at proxmox.com>, phil at tuxera.com
Envoyé: Jeudi 12 Avril 2018 14:41:44
Objet: Re: [pve-devel] avoiding VMID reuse

On Thu, Apr 12 2018 14:26:53 +0200, Fabian Grünbichler wrote: 
> > Sure, it's not a guarantee (because it isn't an error to use an unused 
> > ID less than nextid -- it would be easy to convert the warning to an 
> > error though). But we don't especially need it to be a guarantee, we 
> > just want casual web interface use to not end us up in a situation where 
> > backups break or data is lost, so it's enough to just fix the suggestion 
> > made by the web interface (which is what /cluster/nextid does 
> > precisely). 
> 
> but it does change the semantics and introduces a new class of problem 
> (the guest ID cannot get arbitrarily high, and you only go up and never 
> back down). reusing "holes" avoids this altogether. 

I consider this an unlikely problem, and a much smaller one than 
problems that can arise from not having unique identifiers. So yes, with 
these patches we are trading the problem class "it is not possible to 
uniquely identify virtual machines over time" to "IDs may get high and 
the ID namespace may have holes". 

I don't consider it an especially bad thing that holes are possible. 
There's precedent in other systems, for example PostgreSQL uses the 
"serial" type for auto-incrementing numeric id fields, and holes happen 
there similarly if data is deleted. Yes, it limits the "lifetime" of 
id's to 2^31, but that *is* a lot. 

> > > > another approach would be to adapt your snapshot/sync scripts to remove 
> > > > sync targets if the source gets removed, or do a forceful full sync if 
> > > > an ID gets re-used. the latter is how PVE's builtin ZFS replication 
> > > > works if it fails to find a snapshot combination that allows incremental 
> > > > sending. 
> > 
> > That sounds super dangerous. If I delete a VM and then someone creates a 
> > new one that now gets the same ID, I also lose all backups of my deleted 
> > VM! 
> 
> replication != backup. replication in PVE is for fast disaster recovery. 
> when you delete the source, the replicated copies also get deleted. 

Sure, but I'm specifically talking about backups - just pointing out 
that your advice does not apply there. 

> > > > I am a bit hesitant to introduce such special case heuristics, 
> > > > especially since we don't know if anybody relies on the current 
> > > > semantics of /cluster/nextid 
> > > 
> > > that point still stands though ;) 
> > 
> > I didn't make this configurable, because I don't really see how someone 
> > could be relying on id's getting reused (unless there's an upper limit 
> > to id numbers that could be argued to be reachable). 
> 
> guest IDs are unsigned ints (32 bit) internally. the API limits that 
> further to the range of 100-999999999. while that might seem like a lot, 
> with your proposed change a user just needs to "allocate" the maximum ID 
> to break your scheme (intentionally or otherwise). 

Sure, you could allocate the maximum id right away and then the 
suggestion in the web UI would break. But that is fixable by just 
editing the nextid file, and in at least our deployment we don't really 
worry that users creating VMs want to make each other's life more 
difficult. I could change it so that you are not allowed to use an id 
that is some amount larger than the current highest one if you think 
it's a problem though. 

Or would you propose something different instead? 
_______________________________________________ 
pve-devel mailing list 
pve-devel at pve.proxmox.com 
https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel 



More information about the pve-devel mailing list