[PVE-User] [ceph-users] Ceph on Proxmox VE
Eneko Lacunza
elacunza at binovo.es
Thu Jan 26 15:25:01 CET 2017
Hi Martin,
El 26/01/17 a las 14:08, Martin Maurer escribió:
> We just created a new tutorial for installing Ceph Jewel on Proxmox VE.
>
> The Ceph Server integration in Proxmox VE is already available since
> three years and is a widely used component for smaller deployments to
> get a real open source hyper-converged virtualization and storage setup,
> highly scalable and without limits.
>
> Video Tutorial
> https://youtu.be/jFFLINtNnXs
>
> Documentation
> https://pve.proxmox.com/wiki/Ceph_Server
>
Replying in pve-user not to bother ceph-users with Proxmox especific stuff.
Is Jewel officially supported in Proxmox? I didn't see any notice about
this here nor in release notes.
Also, seems new Ceph_server suggested specs are a bit high on some
fronts, and lacking in others. Maybe the problem is you're describing
the lab setup in that section?:
- 10G network will be overkill if using magnetic drives for OSD storage,
if you don't have tens of OSD per server. We have various setups with
only 1G network that work like a charm, even in one cluster with one SSD
OSD disks per server, we're not able to saturate 1G link.
- I don't think it is wise to recommend Intel SSD DC S3520 drives for
production use. They have very low warranted endurance. S3610 drives
aren't much pricier and offer x10 times the endurance.
Also, fast CPU is a must if OSDs are SSD; if using many magnetic OSDs
maybe having more cores is better than more Ghz.
Again, I think the problem is you're describing the test lab setup in
the recommended hardware section?
Thanks
Eneko
--
Zuzendari Teknikoa / Director Técnico
Binovo IT Human Project, S.L.
Telf. 943493611
943324914
Astigarraga bidea 2, planta 6 dcha., ofi. 3-2; 20180 Oiartzun (Gipuzkoa)
www.binovo.es
More information about the pve-user
mailing list