From 2798d126f3ca639cd57e98d0c97ffe0a9ce0e099 Mon Sep 17 00:00:00 2001 From: Alwin Antreich Date: Wed, 6 Nov 2019 15:09:11 +0100 Subject: [PATCH] pveceph: add attribute ceph_codename To change the codename for Ceph in one place, the patch adds the asciidoc attribute 'ceph_codename'. Replaces the outdated references to luminous and the http -> https on the links in pveceph.adoc. Signed-off-by: Alwin Antreich --- asciidoc/asciidoc-pve.conf | 1 + pveceph.adoc | 30 +++++++++++++++--------------- 2 files changed, 16 insertions(+), 15 deletions(-) diff --git a/asciidoc/asciidoc-pve.conf b/asciidoc/asciidoc-pve.conf index 0716442..2e355cc 100644 --- a/asciidoc/asciidoc-pve.conf +++ b/asciidoc/asciidoc-pve.conf @@ -15,4 +15,5 @@ author=Proxmox Server Solutions Gmbh email=support@proxmox.com endif::docinfo1[] ceph=http://ceph.com[Ceph] +ceph_codename=nautilus diff --git a/pveceph.adoc b/pveceph.adoc index f6fe3fa..10a6ccd 100644 --- a/pveceph.adoc +++ b/pveceph.adoc @@ -58,15 +58,15 @@ and VMs on the same node is possible. To simplify management, we provide 'pveceph' - a tool to install and manage {ceph} services on {pve} nodes. -.Ceph consists of a couple of Daemons footnote:[Ceph intro http://docs.ceph.com/docs/luminous/start/intro/], for use as a RBD storage: +.Ceph consists of a couple of Daemons footnote:[Ceph intro https://docs.ceph.com/docs/{ceph_codename}/start/intro/], for use as a RBD storage: - Ceph Monitor (ceph-mon) - Ceph Manager (ceph-mgr) - Ceph OSD (ceph-osd; Object Storage Daemon) TIP: We highly recommend to get familiar with Ceph's architecture -footnote:[Ceph architecture http://docs.ceph.com/docs/luminous/architecture/] +footnote:[Ceph architecture https://docs.ceph.com/docs/{ceph_codename}/architecture/] and vocabulary -footnote:[Ceph glossary http://docs.ceph.com/docs/luminous/glossary]. +footnote:[Ceph glossary https://docs.ceph.com/docs/{ceph_codename}/glossary]. Precondition @@ -76,7 +76,7 @@ To build a hyper-converged Proxmox + Ceph Cluster there should be at least three (preferably) identical servers for the setup. Check also the recommendations from -http://docs.ceph.com/docs/luminous/start/hardware-recommendations/[Ceph's website]. +https://docs.ceph.com/docs/{ceph_codename}/start/hardware-recommendations/[Ceph's website]. .CPU Higher CPU core frequency reduce latency and should be preferred. As a simple @@ -237,7 +237,7 @@ configuration file. Ceph Monitor ----------- The Ceph Monitor (MON) -footnote:[Ceph Monitor http://docs.ceph.com/docs/luminous/start/intro/] +footnote:[Ceph Monitor https://docs.ceph.com/docs/{ceph_codename}/start/intro/] maintains a master copy of the cluster map. For high availability you need to have at least 3 monitors. One monitor will already be installed if you used the installation wizard. You won't need more than 3 monitors as long @@ -282,7 +282,7 @@ Ceph Manager ------------ The Manager daemon runs alongside the monitors. It provides an interface to monitor the cluster. Since the Ceph luminous release at least one ceph-mgr -footnote:[Ceph Manager http://docs.ceph.com/docs/luminous/mgr/] daemon is +footnote:[Ceph Manager https://docs.ceph.com/docs/{ceph_codename}/mgr/] daemon is required. Create Manager @@ -355,7 +355,7 @@ WARNING: The above command will destroy data on the disk! Starting with the Ceph Kraken release, a new Ceph OSD storage type was introduced, the so called Bluestore -footnote:[Ceph Bluestore http://ceph.com/community/new-luminous-bluestore/]. +footnote:[Ceph Bluestore https://ceph.com/community/new-luminous-bluestore/]. This is the default when creating OSDs since Ceph Luminous. [source,bash] @@ -452,7 +452,7 @@ NOTE: The default number of PGs works for 2-5 disks. Ceph throws a It is advised to calculate the PG number depending on your setup, you can find the formula and the PG calculator footnote:[PG calculator -http://ceph.com/pgcalc/] online. While PGs can be increased later on, they can +https://ceph.com/pgcalc/] online. While PGs can be increased later on, they can never be decreased. @@ -470,7 +470,7 @@ mark the checkbox "Add storages" in the GUI or use the command line option Further information on Ceph pool handling can be found in the Ceph pool operation footnote:[Ceph pool operation -http://docs.ceph.com/docs/luminous/rados/operations/pools/] +https://docs.ceph.com/docs/{ceph_codename}/rados/operations/pools/] manual. @@ -503,7 +503,7 @@ advantage that no central index service is needed. CRUSH works with a map of OSDs, buckets (device locations) and rulesets (data replication) for pools. NOTE: Further information can be found in the Ceph documentation, under the -section CRUSH map footnote:[CRUSH map http://docs.ceph.com/docs/luminous/rados/operations/crush-map/]. +section CRUSH map footnote:[CRUSH map https://docs.ceph.com/docs/{ceph_codename}/rados/operations/crush-map/]. This map can be altered to reflect different replication hierarchies. The object replicas can be separated (eg. failure domains), while maintaining the desired @@ -649,7 +649,7 @@ Since Luminous (12.2.x) you can also have multiple active metadata servers running, but this is normally only useful for a high count on parallel clients, as else the `MDS` seldom is the bottleneck. If you want to set this up please refer to the ceph documentation. footnote:[Configuring multiple active MDS -daemons http://docs.ceph.com/docs/luminous/cephfs/multimds/] +daemons https://docs.ceph.com/docs/{ceph_codename}/cephfs/multimds/] [[pveceph_fs_create]] Create CephFS @@ -681,7 +681,7 @@ This creates a CephFS named `'cephfs'' using a pool for its data named Check the xref:pve_ceph_pools[{pve} managed Ceph pool chapter] or visit the Ceph documentation for more information regarding a fitting placement group number (`pg_num`) for your setup footnote:[Ceph Placement Groups -http://docs.ceph.com/docs/luminous/rados/operations/placement-groups/]. +https://docs.ceph.com/docs/{ceph_codename}/rados/operations/placement-groups/]. Additionally, the `'--add-storage'' parameter will add the CephFS to the {pve} storage configuration after it was created successfully. @@ -763,7 +763,7 @@ object in a PG for its health. There are two forms of Scrubbing, daily (metadata compare) and weekly. The weekly reads the objects and uses checksums to ensure data integrity. If a running scrub interferes with business needs, you can adjust the time when scrubs footnote:[Ceph scrubbing -https://docs.ceph.com/docs/nautilus/rados/configuration/osd-config-ref/#scrubbing] +https://docs.ceph.com/docs/{ceph_codename}/rados/configuration/osd-config-ref/#scrubbing] are executed. @@ -787,10 +787,10 @@ pve# ceph -w To get a more detailed view, every ceph service has a log file under `/var/log/ceph/` and if there is not enough detail, the log level can be -adjusted footnote:[Ceph log and debugging http://docs.ceph.com/docs/luminous/rados/troubleshooting/log-and-debug/]. +adjusted footnote:[Ceph log and debugging https://docs.ceph.com/docs/{ceph_codename}/rados/troubleshooting/log-and-debug/]. You can find more information about troubleshooting -footnote:[Ceph troubleshooting http://docs.ceph.com/docs/luminous/rados/troubleshooting/] +footnote:[Ceph troubleshooting https://docs.ceph.com/docs/{ceph_codename}/rados/troubleshooting/] a Ceph cluster on the official website. -- 2.39.2