[pve-devel] applied: [PATCH docs] change links from master/mimic to luminous

Wolfgang Bumiller w.bumiller at proxmox.com
Mon Feb 18 10:58:39 CET 2019


applied

On Wed, Feb 13, 2019 at 10:38:14AM +0100, David Limbeck wrote:
> Signed-off-by: David Limbeck <d.limbeck at proxmox.com>
> ---
>  pveceph.adoc | 6 +++---
>  1 file changed, 3 insertions(+), 3 deletions(-)
> 
> diff --git a/pveceph.adoc b/pveceph.adoc
> index c90a92e..3e35bb0 100644
> --- a/pveceph.adoc
> +++ b/pveceph.adoc
> @@ -58,7 +58,7 @@ and VMs on the same node is possible.
>  To simplify management, we provide 'pveceph' - a tool to install and
>  manage {ceph} services on {pve} nodes.
>  
> -.Ceph consists of a couple of Daemons footnote:[Ceph intro http://docs.ceph.com/docs/master/start/intro/], for use as a RBD storage:
> +.Ceph consists of a couple of Daemons footnote:[Ceph intro http://docs.ceph.com/docs/luminous/start/intro/], for use as a RBD storage:
>  - Ceph Monitor (ceph-mon)
>  - Ceph Manager (ceph-mgr)
>  - Ceph OSD (ceph-osd; Object Storage Daemon)
> @@ -470,7 +470,7 @@ Since Luminous (12.2.x) you can also have multiple active metadata servers
>  running, but this is normally only useful for a high count on parallel clients,
>  as else the `MDS` seldom is the bottleneck. If you want to set this up please
>  refer to the ceph documentation. footnote:[Configuring multiple active MDS
> -daemons http://docs.ceph.com/docs/mimic/cephfs/multimds/]
> +daemons http://docs.ceph.com/docs/luminous/cephfs/multimds/]
>  
>  [[pveceph_fs_create]]
>  Create a CephFS
> @@ -502,7 +502,7 @@ This creates a CephFS named `'cephfs'' using a pool for its data named
>  Check the xref:pve_ceph_pools[{pve} managed Ceph pool chapter] or visit the
>  Ceph documentation for more information regarding a fitting placement group
>  number (`pg_num`) for your setup footnote:[Ceph Placement Groups
> -http://docs.ceph.com/docs/mimic/rados/operations/placement-groups/].
> +http://docs.ceph.com/docs/luminous/rados/operations/placement-groups/].
>  Additionally, the `'--add-storage'' parameter will add the CephFS to the {pve}
>  storage configuration after it was created successfully.
>  
> -- 
> 2.11.0



More information about the pve-devel mailing list