[pve-devel] [PATCH docs 11/11] pveceph: add attribute ceph_codename

Alwin Antreich a.antreich at proxmox.com
Mon Nov 4 14:52:16 CET 2019


To change the codename for Ceph in one place, the patch adds the
asciidoc attribute 'ceph_codename'. Replaces the outdated references to
luminous and the http -> https on the links in pveceph.adoc.

Signed-off-by: Alwin Antreich <a.antreich at proxmox.com>
---
 pveceph.adoc               | 26 +++++++++++++-------------
 asciidoc/asciidoc-pve.conf |  2 +-
 2 files changed, 14 insertions(+), 14 deletions(-)

diff --git a/pveceph.adoc b/pveceph.adoc
index 127e3bb..3316389 100644
--- a/pveceph.adoc
+++ b/pveceph.adoc
@@ -58,15 +58,15 @@ and VMs on the same node is possible.
 To simplify management, we provide 'pveceph' - a tool to install and
 manage {ceph} services on {pve} nodes.
 
-.Ceph consists of a couple of Daemons footnote:[Ceph intro http://docs.ceph.com/docs/luminous/start/intro/], for use as a RBD storage:
+.Ceph consists of a couple of Daemons footnote:[Ceph intro https://docs.ceph.com/docs/{ceph_codename}/start/intro/], for use as a RBD storage:
 - Ceph Monitor (ceph-mon)
 - Ceph Manager (ceph-mgr)
 - Ceph OSD (ceph-osd; Object Storage Daemon)
 
 TIP: We highly recommend to get familiar with Ceph's architecture
-footnote:[Ceph architecture http://docs.ceph.com/docs/luminous/architecture/]
+footnote:[Ceph architecture https://docs.ceph.com/docs/{ceph_codename}/architecture/]
 and vocabulary
-footnote:[Ceph glossary http://docs.ceph.com/docs/luminous/glossary].
+footnote:[Ceph glossary https://docs.ceph.com/docs/{ceph_codename}/glossary].
 
 
 Precondition
@@ -76,7 +76,7 @@ To build a hyper-converged Proxmox + Ceph Cluster there should be at least
 three (preferably) identical servers for the setup.
 
 Check also the recommendations from
-http://docs.ceph.com/docs/luminous/start/hardware-recommendations/[Ceph's website].
+https://docs.ceph.com/docs/{ceph_codename}/start/hardware-recommendations/[Ceph's website].
 
 .CPU
 Higher CPU core frequency reduce latency and should be preferred. As a simple
@@ -237,7 +237,7 @@ configuration file.
 Ceph Monitor
 -----------
 The Ceph Monitor (MON)
-footnote:[Ceph Monitor http://docs.ceph.com/docs/luminous/start/intro/]
+footnote:[Ceph Monitor https://docs.ceph.com/docs/{ceph_codename}/start/intro/]
 maintains a master copy of the cluster map. For high availability you need to
 have at least 3 monitors. One monitor will already be installed if you
 used the installation wizard. You won't need more than 3 monitors as long
@@ -284,7 +284,7 @@ Ceph Manager
 ------------
 The Manager daemon runs alongside the monitors, providing an interface for
 monitoring the cluster. Since the Ceph luminous release at least one ceph-mgr
-footnote:[Ceph Manager http://docs.ceph.com/docs/luminous/mgr/] daemon is
+footnote:[Ceph Manager https://docs.ceph.com/docs/{ceph_codename}/mgr/] daemon is
 required.
 
 Creating Manager
@@ -479,7 +479,7 @@ mark the checkbox "Add storages" in the GUI or use the command line option
 
 Further information on Ceph pool handling can be found in the Ceph pool
 operation footnote:[Ceph pool operation
-http://docs.ceph.com/docs/luminous/rados/operations/pools/]
+https://docs.ceph.com/docs/{ceph_codename}/rados/operations/pools/]
 manual.
 
 
@@ -515,7 +515,7 @@ advantage that no central index service is needed. CRUSH works with a map of
 OSDs, buckets (device locations) and rulesets (data replication) for pools.
 
 NOTE: Further information can be found in the Ceph documentation, under the
-section CRUSH map footnote:[CRUSH map http://docs.ceph.com/docs/luminous/rados/operations/crush-map/].
+section CRUSH map footnote:[CRUSH map https://docs.ceph.com/docs/{ceph_codename}/rados/operations/crush-map/].
 
 This map can be altered to reflect different replication hierarchies. The object
 replicas can be separated (eg. failure domains), while maintaining the desired
@@ -661,7 +661,7 @@ Since Luminous (12.2.x) you can also have multiple active metadata servers
 running, but this is normally only useful for a high count on parallel clients,
 as else the `MDS` seldom is the bottleneck. If you want to set this up please
 refer to the ceph documentation. footnote:[Configuring multiple active MDS
-daemons http://docs.ceph.com/docs/luminous/cephfs/multimds/]
+daemons https://docs.ceph.com/docs/{ceph_codename}/cephfs/multimds/]
 
 [[pveceph_fs_create]]
 Creating CephFS
@@ -693,7 +693,7 @@ This creates a CephFS named `'cephfs'' using a pool for its data named
 Check the xref:pve_ceph_pools[{pve} managed Ceph pool chapter] or visit the
 Ceph documentation for more information regarding a fitting placement group
 number (`pg_num`) for your setup footnote:[Ceph Placement Groups
-http://docs.ceph.com/docs/luminous/rados/operations/placement-groups/].
+https://docs.ceph.com/docs/{ceph_codename}/rados/operations/placement-groups/].
 Additionally, the `'--add-storage'' parameter will add the CephFS to the {pve}
 storage configuration after it was created successfully.
 
@@ -775,7 +775,7 @@ object in a PG for its health. There are two forms of Scrubbing, daily
 (metadata compare) and weekly. The latter reads the object and uses checksums
 to ensure data integrity. If a running scrub interferes with business needs,
 you can adjust the time of execution of Scrub footnote:[Ceph scrubbing
-https://docs.ceph.com/docs/nautilus/rados/configuration/osd-config-ref/#scrubbing].
+https://docs.ceph.com/docs/{ceph_codename}/rados/configuration/osd-config-ref/#scrubbing].
 
 
 Ceph monitoring and troubleshooting
@@ -798,10 +798,10 @@ pve# ceph -w
 
 To get a more detailed view, every ceph service has a log file under
 `/var/log/ceph/` and if there is not enough detail, the log level can be
-adjusted footnote:[Ceph log and debugging http://docs.ceph.com/docs/luminous/rados/troubleshooting/log-and-debug/].
+adjusted footnote:[Ceph log and debugging https://docs.ceph.com/docs/{ceph_codename}/rados/troubleshooting/log-and-debug/].
 
 You can find more information about troubleshooting
-footnote:[Ceph troubleshooting http://docs.ceph.com/docs/luminous/rados/troubleshooting/]
+footnote:[Ceph troubleshooting https://docs.ceph.com/docs/{ceph_codename}/rados/troubleshooting/]
 a Ceph cluster on the official website.
 
 
diff --git a/asciidoc/asciidoc-pve.conf b/asciidoc/asciidoc-pve.conf
index 0716442..2229ea9 100644
--- a/asciidoc/asciidoc-pve.conf
+++ b/asciidoc/asciidoc-pve.conf
@@ -15,4 +15,4 @@ author=Proxmox Server Solutions Gmbh
 email=support at proxmox.com
 endif::docinfo1[]
 ceph=http://ceph.com[Ceph]
-
+ceph_codename=nautilus
-- 
2.20.1





More information about the pve-devel mailing list