[pve-devel] applied: [PATCH docs v2] Expand the Precondition section

Thomas Lamprecht t.lamprecht at proxmox.com
Thu Apr 4 14:33:31 CEST 2019


On 4/4/19 11:23 AM, Alwin Antreich wrote:
> This patch adds more information about hardware preconditions and
> practices.
> 
> Signed-off-by: Alwin Antreich <a.antreich at proxmox.com>
> ---
> V1 -> V2: tried to simplify english, as discussed off list.

applied, thanks! it would be great if you "namespace" such additions with
a short keyword, e.g., for your series I always add "ceph: " in front of the
commit subject, makes scrolling through commit log a bit more enjoyable.
Just as side remark :)

> 
>  pveceph.adoc | 60 ++++++++++++++++++++++++++++++++++++++++++++++++++++++------
>  1 file changed, 54 insertions(+), 6 deletions(-)
> 
> diff --git a/pveceph.adoc b/pveceph.adoc
> index d330dea..bfe6a62 100644
> --- a/pveceph.adoc
> +++ b/pveceph.adoc
> @@ -72,16 +72,59 @@ footnote:[Ceph glossary http://docs.ceph.com/docs/luminous/glossary].
>  Precondition
>  ------------
>  
> -To build a Proxmox Ceph Cluster there should be at least three (preferably)
> -identical servers for the setup.
> -
> -A 10Gb network, exclusively used for Ceph, is recommended. A meshed network
> -setup is also an option if there are no 10Gb switches available, see our wiki
> -article footnote:[Full Mesh Network for Ceph {webwiki-url}Full_Mesh_Network_for_Ceph_Server] .
> +To build a hyper-converged Proxmox + Ceph Cluster there should be at least
> +three (preferably) identical servers for the setup.
>  
>  Check also the recommendations from
>  http://docs.ceph.com/docs/luminous/start/hardware-recommendations/[Ceph's website].
>  
> +.CPU
> +As higher the core frequency the better, this will reduce latency.  Among other
> +things, this benefits the services of Ceph, as they can process data faster.
> +To simplify planning, you should assign a CPU core (or thread) to each Ceph
> +service to provide enough resources for stable and durable Ceph performance.
> +
> +.Memory
> +Especially in a hyper-converged setup, the memory consumption needs to be
> +carefully monitored. In addition to the intended workload (VM / Container),
> +Ceph needs enough memory to provide good and stable performance. As a rule of
> +thumb, for roughly 1TiB of data, 1 GiB of memory will be used by an OSD. With
> +additionally needed memory for OSD caching.
> +
> +.Network
> +We recommend a network bandwidth of at least 10 GbE or more, which is used
> +exclusively for Ceph. A meshed network setup
> +footnote:[Full Mesh Network for Ceph {webwiki-url}Full_Mesh_Network_for_Ceph_Server]
> +is also an option if there are no 10 GbE switches available.
> +
> +To be explicit about the network, since Ceph is a distributed network storage,
> +its traffic must be put on its own physical network. The volume of traffic
> +especially during recovery will interfere with other services on the same
> +network.
> +
> +Further, estimate your bandwidth needs. While one HDD might not saturate a 1 Gb
> +link, a SSD or a NVMe SSD certainly can. Modern NVMe SSDs will even saturate 10
> +Gb of bandwidth. You also should consider higher bandwidths, as these tend to
> +come with lower latency.
> +
> +.Disks
> +When planning the size of your Ceph cluster, it is important to take the
> +recovery time into consideration. Especially with small clusters, the recovery
> +might take long. It is recommended that you use SSDs instead of HDDs in small
> +setups to reduce recovery time, minimizing the likelihood of a subsequent
> +failure event during recovery.
> +
> +In general SSDs will provide more IOPs then spinning disks. This fact and the
> +higher cost may make a xref:pve_ceph_device_classes[class based] separation of
> +pools appealing.  Another possibility to speedup OSDs is to use a faster disk
> +as journal or DB/WAL device, see xref:pve_ceph_osds[creating Ceph OSDs]. If a
> +faster disk is used for multiple OSDs, a proper balance between OSD and WAL /
> +DB (or journal) disk must be selected, otherwise the faster disk becomes the
> +bottleneck for all linked OSDs.
> +
> +Aside from the disk type, Ceph best performs with an even sized and distributed
> +amount of disks per node. For example, 4x disks à 500 GB in each node.
> +
>  .Avoid RAID
>  As Ceph handles data object redundancy and multiple parallel writes to disks
>  (OSDs) on its own, using a RAID controller normally doesn’t improve
> @@ -93,6 +136,10 @@ the ones from Ceph.
>  
>  WARNING: Avoid RAID controller, use host bus adapter (HBA) instead.
>  
> +NOTE: Above recommendations should be seen as a rough guidance for choosing
> +hardware. Therefore, it is still essential to test your setup and monitor
> +health & performance.
> +
>  
>  [[pve_ceph_install]]
>  Installation of Ceph Packages
> @@ -316,6 +363,7 @@ operation footnote:[Ceph pool operation
>  http://docs.ceph.com/docs/luminous/rados/operations/pools/]
>  manual.
>  
> +[[pve_ceph_device_classes]]
>  Ceph CRUSH & device classes
>  ---------------------------
>  The foundation of Ceph is its algorithm, **C**ontrolled **R**eplication
> 






More information about the pve-devel mailing list