[pve-devel] [PATCH 2/2] hugepages: use hostnodes value as numanode for topology
Alexandre DERUMIER
aderumier at odiso.com
Fri Jul 29 11:31:51 CEST 2016
>>Looks like we still need to do some more mapping though, as the foreach
>>function still uses '$i % $sockets' for the numa node index.
Oh, I didn't review the hotplug memory part yet (I have tested succefully my patch without memory hotplug for now)
>>If the above makes sense to you I can prepare a patch for you to review?
I think it's ok, I'll check that next week.
----- Mail original -----
De: "Wolfgang Bumiller" <w.bumiller at proxmox.com>
À: "aderumier" <aderumier at odiso.com>
Cc: "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Vendredi 29 Juillet 2016 11:14:22
Objet: Re: [pve-devel] [PATCH 2/2] hugepages: use hostnodes value as numanode for topology
On Fri, Jul 29, 2016 at 10:40:37AM +0200, Alexandre DERUMIER wrote:
> >>I don't really see the relation there, why would this restriction apply
> >>specifically with hugepages enabled? Or is it just a performance
> >>concern?
>
> yes,it's only when hugepages is enabled.
>
> As we reserved hugepages, I really don't known how qemu will try to split memory with
>
> hostnodes=0;1,memory=4096
>
> or
>
> hostnodes=0;2-3,memory=4096
>
>
> I think it's random, so in this case we should reserved pages on all nodes to be sure.
> Seem to be a little be complex to manage.
Makes sense.
Looks like we still need to do some more mapping though, as the foreach
function still uses '$i % $sockets' for the numa node index.
With custom numa entries this matches neither those nor the host nodes,
especially with hostnodes defined.
I think we can do it this way:
* in foreach_dimm: Count existing $conf->{numaX} and if any are
defined use that count instead of $sockets.
* go through foreach* uses and make sure numa nodes are mapped
accordingly:
- qemu_memory_hotplug(): foreach_dimm's passed $numanode parameter is
used in two ways:
| my $hugepages_topology->{$hugepages_size}->{$numanode} = hugepages_nr($dimm_size, $hugepages_size);
This needs to be mapped to $conf->{numa$numanode}'s hostnode.
| eval { PVE::QemuServer::vm_mon_cmd($vmid, "device_add", driver => "pc-dimm", id => "$name", memdev => "mem-$name", node => $numanode) };
This should be fine once foreach_dimm() takes numaX entries into
account as it refers to the guest's node index AFAIK.
- config(): Should be fine, refers to guest node index.
- hugepages_topology(): Needs to be mapped to hostnode.
If the above makes sense to you I can prepare a patch for you to review?
>
>
> ----- Mail original -----
> De: "Wolfgang Bumiller" <w.bumiller at proxmox.com>
> À: "aderumier" <aderumier at odiso.com>
> Cc: "pve-devel" <pve-devel at pve.proxmox.com>
> Envoyé: Vendredi 29 Juillet 2016 09:48:04
> Objet: Re: [pve-devel] [PATCH 2/2] hugepages: use hostnodes value as numanode for topology
>
> On Fri, Jul 29, 2016 at 05:17:34AM +0200, Alexandre Derumier wrote:
> > also disallow more than 1 hostnode value
>
> I don't really see the relation there, why would this restriction apply
> specifically with hugepages enabled? Or is it just a performance
> concern?
>
> Patch seems fine otherwise.
>
> >
> > Signed-off-by: Alexandre Derumier <aderumier at odiso.com>
> > ---
> > PVE/QemuServer/Memory.pm | 35 +++++++++++++++++++++++------------
> > 1 file changed, 23 insertions(+), 12 deletions(-)
> >
> > diff --git a/PVE/QemuServer/Memory.pm b/PVE/QemuServer/Memory.pm
> > index 58c7217..37f285a 100644
> > --- a/PVE/QemuServer/Memory.pm
> > +++ b/PVE/QemuServer/Memory.pm
> > @@ -230,17 +230,8 @@ sub config {
> > # hostnodes
> > my $hostnodelists = $numa->{hostnodes};
> > if (defined($hostnodelists)) {
> > - my $hostnodes;
> > - foreach my $hostnoderange (@$hostnodelists) {
> > - my ($start, $end) = @$hostnoderange;
> > - $hostnodes .= ',' if $hostnodes;
> > - $hostnodes .= $start;
> > - $hostnodes .= "-$end" if defined($end);
> > - $end //= $start;
> > - for (my $i = $start; $i <= $end; ++$i ) {
> > - die "host NUMA node$i doesn't exist\n" if ! -d "/sys/devices/system/node/node$i/";
> > - }
> > - }
> > +
> > + my $hostnodes = print_numa_hostnodes($hostnodelists);
> >
> > # policy
> > my $policy = $numa->{policy};
> > @@ -311,6 +302,23 @@ sub print_mem_object {
> >
> > }
> >
> > +sub print_numa_hostnodes {
> > + my ($hostnodelists) = @_;
> > +
> > + my $hostnodes;
> > + foreach my $hostnoderange (@$hostnodelists) {
> > + my ($start, $end) = @$hostnoderange;
> > + $hostnodes .= ',' if $hostnodes;
> > + $hostnodes .= $start;
> > + $hostnodes .= "-$end" if defined($end);
> > + $end //= $start;
> > + for (my $i = $start; $i <= $end; ++$i ) {
> > + die "host NUMA node$i doesn't exist\n" if ! -d "/sys/devices/system/node/node$i/";
> > + }
> > + }
> > + return $hostnodes;
> > +}
> > +
> > sub hugepages_mount {
> >
> > my $mountdata = PVE::ProcFSTools::parse_proc_mounts();
> > @@ -406,9 +414,12 @@ sub hugepages_topology {
> >
> > $numa_custom_topology = 1;
> > my $numa_memory = $numa->{memory};
> > + my $hostnodelists = $numa->{hostnodes};
> > + my $hostnodes = print_numa_hostnodes($hostnodelists);
> >
> > + die "more than 1 hostnode value in numa node is not supported when hugepages are enabled" if $hostnodes !~ m/^(\d)$/;
> > my $hugepages_size = hugepages_size($conf, $numa_memory);
> > - $hugepages_topology->{$hugepages_size}->{$i} += hugepages_nr($numa_memory, $hugepages_size);
> > + $hugepages_topology->{$hugepages_size}->{$hostnodes} += hugepages_nr($numa_memory, $hugepages_size);
> >
> > }
> >
> > --
> > 2.1.4
> >
> > _______________________________________________
> > pve-devel mailing list
> > pve-devel at pve.proxmox.com
> > http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel
> >
>
>
More information about the pve-devel
mailing list