[PVE-User] Proxmox Ceph high memory usage

Alwin Antreich a.antreich at proxmox.com
Wed Jan 16 14:07:26 CET 2019


Hello Gilberto,

On Wed, Jan 16, 2019 at 10:11:06AM -0200, Gilberto Nunes wrote:
> Hi there
> 
> Anybody else experiment hight memory usage in Proxmox CEPH Storage Server?
> I have a 6 node PVE CEPH and after upgrade, I have noticed this high memory
> usage...
> All server has 16GB of ram. I know this is not recomended, but that what I
> have at the moment.
> In fact, just 3 servers ran with about 90% of memory usage.
> All server is IBM x3200 m2 with SATA disks...
> Here's ceph osd tree
> ceph osd tree
> ID  CLASS WEIGHT   TYPE NAME           STATUS REWEIGHT PRI-AFF
>  -1       38.50000 root default
>  -3        4.00000     host pve-ceph01
>   0   hdd  1.00000         osd.0           up  1.00000       0
>   1   hdd  1.00000         osd.1           up  1.00000       0
>   2   hdd  1.00000         osd.2           up  1.00000       0
>   3   hdd  1.00000         osd.3           up  1.00000       0
>  -5        8.00000     host pve-ceph02
>   4   hdd  2.00000         osd.4           up  1.00000 1.00000
>   5   hdd  2.00000         osd.5           up  1.00000 1.00000
>   6   hdd  2.00000         osd.6           up  1.00000 1.00000
>   7   hdd  2.00000         osd.7           up  1.00000 1.00000
>  -7        9.00000     host pve-ceph03
>   8   hdd  3.00000         osd.8           up  1.00000 1.00000
>   9   hdd  3.00000         osd.9           up  1.00000 1.00000
>  10   hdd  3.00000         osd.10          up  1.00000 1.00000
>  -9       12.00000     host pve-ceph04
>  11   hdd  3.00000         osd.11          up  1.00000 1.00000
>  12   hdd  3.00000         osd.12          up  1.00000 1.00000
>  13   hdd  3.00000         osd.13          up  1.00000 1.00000
>  14   hdd  3.00000         osd.14          up  1.00000 1.00000
> -11        1.00000     host pve-ceph05
>  15   hdd  0.50000         osd.15          up  1.00000       0
>  16   hdd  0.50000         osd.16          up  1.00000       0
> -13        4.50000     host pve-ceph06
>  17   hdd  1.00000         osd.17          up  1.00000       0
>  18   hdd  1.00000         osd.18          up  1.00000       0
>  20   hdd  1.00000         osd.20          up  1.00000       0
>  21   hdd  1.50000         osd.21          up  1.00000 1.00000
> 
Did you see the changle on package upgrade? It explains why it is using
more memory then before.
http://download.proxmox.com/debian/ceph-luminous/dists/stretch/main/binary-amd64/ceph_12.2.10-pve1.changelog

--
Cheers,
Alwin




More information about the pve-user mailing list