[PVE-User] High CPU and Memory utilzation
Gilberto Nunes
gilberto.nunes32 at gmail.com
Wed Jul 13 14:36:13 CEST 2016
Hello PVE guys...
I have one host, which is a Dell PowerEdge R430, with 12 cores and 48 GB of
memory.
The disk are all SAS 10K RPM.
I have just one KVM VM with Ubuntu 14.04, which is our Zimbra Mail Server.
The qcow2 image is reside in another server, also a Dell PowerEdge R430,
with 15 GB of memory. This server, act like a Storage, a GlusterFS Server,
set up like NFS.
There's 3 SAS disk 10K RPM. Between this to servers, I have three 1 GB NIC
set as bond0, with Bonding Mode: load balancing.
This the iperf3 output:
iperf3 -c storage100
Connecting to host storage100, port 5201
[ 4] local 10.1.1.140 port 36098 connected to 10.1.1.180 port 5201
[ ID] Interval Transfer Bandwidth Retr Cwnd
[ 4] 0.00-1.00 sec 309 MBytes 2.60 Gbits/sec 101 386 KBytes
[ 4] 1.00-2.00 sec 250 MBytes 2.10 Gbits/sec 20 387 KBytes
[ 4] 2.00-3.00 sec 247 MBytes 2.07 Gbits/sec 20 390 KBytes
[ 4] 3.00-4.00 sec 306 MBytes 2.56 Gbits/sec 0 390 KBytes
[ 4] 4.00-5.00 sec 314 MBytes 2.63 Gbits/sec 13 390 KBytes
[ 4] 5.00-6.00 sec 314 MBytes 2.63 Gbits/sec 0 390 KBytes
[ 4] 6.00-7.00 sec 295 MBytes 2.48 Gbits/sec 0 390 KBytes
[ 4] 7.00-8.00 sec 280 MBytes 2.35 Gbits/sec 90 468 KBytes
[ 4] 8.00-9.00 sec 297 MBytes 2.49 Gbits/sec 0 468 KBytes
[ 4] 9.00-10.00 sec 304 MBytes 2.55 Gbits/sec 142 485 KBytes
- - - - - - - - - - - - - - - - - - - - - - - - -
[ ID] Interval Transfer Bandwidth Retr
[ 4] 0.00-10.00 sec 2.85 GBytes 2.45 Gbits/sec 386 sender
[ 4] 0.00-10.00 sec 2.85 GBytes 2.44 Gbits/sec
receiver
iperf Done.
Sometimes, and this is more often than I wish, I run into high cpu and
memory utilization. The CPU reach %CPU 250-300 and %MEM reach 77-80.
And I don't exactly this is the cause of slow access to disk, but in fact,
I get more than 50 MB/s in diskwrite. That is showing to me, from Disk IO
rrdtool graphics in PVE web console.
In this period of time, the Zimbra Mail Server has slow access and my users
complain a lot of time...
I had make some research in Google and try improve performance in glusterFS.
Here is something I working on:
gluster vol info
Volume Name: storage
Type: Distribute
Volume ID: 183b40bc-9e1d-4f2c-a772-ec8e15367485
Status: Started
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: storage100:/data/volume
Options Reconfigured:
performance.write-behind-window-size: 1024MB
performance.cache-size: 1GB
performance.io-thread-count: 16
performance.flush-behind: on
performance.cache-refresh-timeout: 10
performance.quick-read: off
performance.read-ahead: enable
network.ping-timeout: 2
performance.cache-max-file-size: 4MB
performance.md-cache-timeout: 1
nfs.addr-namelookup: off
performance.client-io-threads: on
performance.nfs.io-cache: off
performance.cache-priority: *.qcow2:1,*.raw:1
Is there some other thing I can do improve performance?
I'll appreciated any advice!...
Thanks a lot.
More information about the pve-user
mailing list