[pve-devel] The network performance future for VMs
Alexandre DERUMIER
aderumier at odiso.com
Wed Aug 19 11:08:57 CEST 2015
>>Page 18:
>>Config I - DPDK with VFIO device assignment (future)
>>According to the graph, it is functional to QEMU, but I think that not for
>>VMs Windows... is it correct???
I really don't known ;) I thinked that qemu support was enough, maybe not ...
>>Moreover, all this questions is due to that i want to improve the speed of
>>the network in a VM Win2k8 r2
>>... is there anything that we can do for get better performance?
>>(I would like to get a maximum of 20 Gb/s due to that i have 2 ports of 10
>>Gb/s each one with LACP configured and with the Linux stack)
>>Note: I know that i will need two connections simultaneous of net for get 10
>>Gb/s in each link.
>From my tests, I get a lot better performance with win2012r2 and win2K8r2.
and also, you can try nic multiqueue feature, it should improve performance with multiple streams.
Don't remember, but I think I was around 7-8gigabit for 1 windows vm.
But still a lot lower than linux vm.
now, it's clear that dpdk should improve performance for high pps.
----- Mail original -----
De: "Cesar Peschiera" <brain at click.com.py>
À: "aderumier" <aderumier at odiso.com>
Cc: "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Mercredi 19 Août 2015 10:50:37
Objet: Re: [pve-devel] The network performance future for VMs
Thanks Alexandre for your prompt response.
> Seem to be easy with vhost-user virtual network card, and this one can't
> work with linux bridge, because it's userland
I forgot say that OVS was configured in two ports for the LAN link, and
other two ports with the Linux stack for DRBD in blance-rr NIC-to-NIC (OVS
not have the option balance-rr).
In this case is that i had problems with DRBD, then, I preferred disable
totally in my servers the OVS setup.
> I'm not sure, but maybe dpkg on linux stack can only work with host
> physical interfaces and not qemu virtual interfaces.
dpkg?, i assume that you want to say DPDK.
Please see in this link:
https://videos.cdn.redhat.com/summit2015/presentations/12752_red-hat-enterprise-virtualization-hypervisor-kvm-now-in-the-future.pdf
Page 18:
Config I - DPDK with VFIO device assignment (future)
According to the graph, it is functional to QEMU, but I think that not for
VMs Windows... is it correct???
Moreover, all this questions is due to that i want to improve the speed of
the network in a VM Win2k8 r2
... is there anything that we can do for get better performance?
(I would like to get a maximum of 20 Gb/s due to that i have 2 ports of 10
Gb/s each one with LACP configured and with the Linux stack)
Note: I know that i will need two connections simultaneous of net for get 10
Gb/s in each link.
----- Original Message -----
From: "Alexandre DERUMIER" <aderumier at odiso.com>
To: "Cesar Peschiera" <brain at click.com.py>
Cc: "pve-devel" <pve-devel at pve.proxmox.com>
Sent: Wednesday, August 19, 2015 1:39 AM
Subject: Re: [pve-devel] The network performance future for VMs
>>So now my question is if DPDK can be activated also with the Linux stack?.
I need to dig a little more about this.
Intel seem to push the ovs-dpdk in all conferenfece I have see.
(Seem to be easy with vhost-user virtual network card, and this one can't
work with linux bridge, because it's userland)
I'm not sure, but maybe dpkg on linux stack can only work with host physical
interfaces and not qemu virtual interfaces.
----- Mail original -----
De: "Cesar Peschiera" <brain at click.com.py>
À: "aderumier" <aderumier at odiso.com>
Cc: "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Mardi 18 Août 2015 21:25:46
Objet: Re: [pve-devel] The network performance future for VMs
Oh, ok.
In the past, i had problems with DRBD 8.4.5 when OVS is enabled, so i had
that change my setup from OVS to the Linux stack, after of it, i had no more
problems with DRBD.
About of the problem with OVS and DRBD, i did not test in depth the problem
(in the season of preproduction phase), but if i not bad remember, maybe the
problem appears when "OVS Intport" is enabled, or maybe only when OVS is
enabled in the setup.
I was using PVE 3.3
So now my question is if DPDK can be activated also with the Linux stack?.
----- Original Message -----
From: "Alexandre DERUMIER" <aderumier at odiso.com>
To: "Cesar Peschiera" <brain at click.com.py>
Cc: "pve-devel" <pve-devel at pve.proxmox.com>
Sent: Tuesday, August 18, 2015 8:57 AM
Subject: Re: [pve-devel] The network performance future for VMs
>>So, i would like to ask about of the future of PVE in network performance
>>terms.
dpdk will be implemented in openvswitch through vhost-user,
I'm waiting for ovs 2.4 to look at this.
----- Mail original -----
De: "Cesar Peschiera" <brain at click.com.py>
À: "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Mardi 18 Août 2015 13:00:59
Objet: [pve-devel] The network performance future for VMs
Hi developers of PVE
I would like to talk about of the network speed for VMs:
I see in this link (Web official of Red Hat):
https://videos.cdn.redhat.com/summit2015/presentations/12752_red-hat-enterprise-virtualization-hypervisor-kvm-now-in-the-future.pdf
In the page 19 of this pdf, i see a interesting info:
Network Function Virtualization (NFV)
Throughput and Packets/sec "RHEL7.x + DPDK (Data Plane Development Kit)":
Millons packets per second:
KVM = 208
Docker = 215
Bare-metal = 218
HW maximum = 225
Between KVM and Bare-metal, the difference is little: 10
Also i see a list of HW NICs compatibility on this link:
http://dpdk.org/doc/nics
So, i would like to ask about of the future of PVE in network performance
terms.
Best regards
Cesar
_______________________________________________
pve-devel mailing list
pve-devel at pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel
More information about the pve-devel
mailing list