X-Git-Url: https://git.proxmox.com/?a=blobdiff_plain;f=pveceph.adoc;h=b3bbadfda7f5bd817195f029d8fbfc5aad87e0bd;hb=c6e098a291471715218db3edb6b90f09b3dd8f33;hp=5b9e1990d79bb3e40491cc6fc6e560e59c0223fa;hpb=8a38333f6400f2ddc71f97b81b4827d8102aeabc;p=pve-docs.git diff --git a/pveceph.adoc b/pveceph.adoc index 5b9e199..b3bbadf 100644 --- a/pveceph.adoc +++ b/pveceph.adoc @@ -18,8 +18,8 @@ DESCRIPTION ----------- endif::manvolnum[] ifndef::manvolnum[] -Manage Ceph Services on Proxmox VE Nodes -======================================== +Deploy Hyper-Converged Ceph Cluster +=================================== :pve-toplevel: endif::manvolnum[] @@ -58,15 +58,15 @@ and VMs on the same node is possible. To simplify management, we provide 'pveceph' - a tool to install and manage {ceph} services on {pve} nodes. -.Ceph consists of a couple of Daemons footnote:[Ceph intro http://docs.ceph.com/docs/luminous/start/intro/], for use as a RBD storage: +.Ceph consists of a couple of Daemons footnote:[Ceph intro https://docs.ceph.com/docs/{ceph_codename}/start/intro/], for use as a RBD storage: - Ceph Monitor (ceph-mon) - Ceph Manager (ceph-mgr) - Ceph OSD (ceph-osd; Object Storage Daemon) TIP: We highly recommend to get familiar with Ceph's architecture -footnote:[Ceph architecture http://docs.ceph.com/docs/luminous/architecture/] +footnote:[Ceph architecture https://docs.ceph.com/docs/{ceph_codename}/architecture/] and vocabulary -footnote:[Ceph glossary http://docs.ceph.com/docs/luminous/glossary]. +footnote:[Ceph glossary https://docs.ceph.com/docs/{ceph_codename}/glossary]. Precondition @@ -76,7 +76,7 @@ To build a hyper-converged Proxmox + Ceph Cluster there should be at least three (preferably) identical servers for the setup. Check also the recommendations from -http://docs.ceph.com/docs/luminous/start/hardware-recommendations/[Ceph's website]. +https://docs.ceph.com/docs/{ceph_codename}/start/hardware-recommendations/[Ceph's website]. .CPU Higher CPU core frequency reduce latency and should be preferred. As a simple @@ -101,7 +101,7 @@ services on the same network and may even break the {pve} cluster stack. Further, estimate your bandwidth needs. While one HDD might not saturate a 1 Gb link, multiple HDD OSDs per node can, and modern NVMe SSDs will even saturate -10 Gbps of bandwidth quickly. Deploying a network capable of even more bandwith +10 Gbps of bandwidth quickly. Deploying a network capable of even more bandwidth will ensure that it isn't your bottleneck and won't be anytime soon, 25, 40 or even 100 GBps are possible. @@ -237,7 +237,7 @@ configuration file. Ceph Monitor ----------- The Ceph Monitor (MON) -footnote:[Ceph Monitor http://docs.ceph.com/docs/luminous/start/intro/] +footnote:[Ceph Monitor https://docs.ceph.com/docs/{ceph_codename}/start/intro/] maintains a master copy of the cluster map. For high availability you need to have at least 3 monitors. One monitor will already be installed if you used the installation wizard. You won't need more than 3 monitors as long @@ -245,6 +245,7 @@ as your cluster is small to midsize, only really large clusters will need more than that. +[[pveceph_create_mon]] Create Monitors ~~~~~~~~~~~~~~~ @@ -259,7 +260,7 @@ create it by using the 'Ceph -> Monitor' tab in the GUI or run. pveceph mon create ---- - +[[pveceph_destroy_mon]] Destroy Monitors ~~~~~~~~~~~~~~~~ @@ -282,9 +283,10 @@ Ceph Manager ------------ The Manager daemon runs alongside the monitors. It provides an interface to monitor the cluster. Since the Ceph luminous release at least one ceph-mgr -footnote:[Ceph Manager http://docs.ceph.com/docs/luminous/mgr/] daemon is +footnote:[Ceph Manager https://docs.ceph.com/docs/{ceph_codename}/mgr/] daemon is required. +[[pveceph_create_mgr]] Create Manager ~~~~~~~~~~~~~~ @@ -299,6 +301,7 @@ NOTE: It is recommended to install the Ceph Manager on the monitor nodes. For high availability install more then one manager. +[[pveceph_destroy_mgr]] Destroy Manager ~~~~~~~~~~~~~~~ @@ -325,6 +328,7 @@ network. It is recommended to use one OSD per physical disk. NOTE: By default an object is 4 MiB in size. +[[pve_ceph_osd_create]] Create OSDs ~~~~~~~~~~~ @@ -354,7 +358,7 @@ WARNING: The above command will destroy data on the disk! Starting with the Ceph Kraken release, a new Ceph OSD storage type was introduced, the so called Bluestore -footnote:[Ceph Bluestore http://ceph.com/community/new-luminous-bluestore/]. +footnote:[Ceph Bluestore https://ceph.com/community/new-luminous-bluestore/]. This is the default when creating OSDs since Ceph Luminous. [source,bash] @@ -374,7 +378,7 @@ pveceph osd create /dev/sd[X] -db_dev /dev/sd[Y] -wal_dev /dev/sd[Z] ---- You can directly choose the size for those with the '-db_size' and '-wal_size' -paremeters respectively. If they are not given the following values (in order) +parameters respectively. If they are not given the following values (in order) will be used: * bluestore_block_{db,wal}_size from ceph configuration... @@ -401,6 +405,7 @@ Starting with Ceph Nautilus, {pve} does not support creating such OSDs with ceph-volume lvm create --filestore --data /dev/sd[X] --journal /dev/sd[Y] ---- +[[pve_ceph_osd_destroy]] Destroy OSDs ~~~~~~~~~~~~ @@ -450,7 +455,7 @@ NOTE: The default number of PGs works for 2-5 disks. Ceph throws a It is advised to calculate the PG number depending on your setup, you can find the formula and the PG calculator footnote:[PG calculator -http://ceph.com/pgcalc/] online. While PGs can be increased later on, they can +https://ceph.com/pgcalc/] online. While PGs can be increased later on, they can never be decreased. @@ -468,7 +473,7 @@ mark the checkbox "Add storages" in the GUI or use the command line option Further information on Ceph pool handling can be found in the Ceph pool operation footnote:[Ceph pool operation -http://docs.ceph.com/docs/luminous/rados/operations/pools/] +https://docs.ceph.com/docs/{ceph_codename}/rados/operations/pools/] manual. @@ -501,7 +506,7 @@ advantage that no central index service is needed. CRUSH works with a map of OSDs, buckets (device locations) and rulesets (data replication) for pools. NOTE: Further information can be found in the Ceph documentation, under the -section CRUSH map footnote:[CRUSH map http://docs.ceph.com/docs/luminous/rados/operations/crush-map/]. +section CRUSH map footnote:[CRUSH map https://docs.ceph.com/docs/{ceph_codename}/rados/operations/crush-map/]. This map can be altered to reflect different replication hierarchies. The object replicas can be separated (eg. failure domains), while maintaining the desired @@ -647,7 +652,7 @@ Since Luminous (12.2.x) you can also have multiple active metadata servers running, but this is normally only useful for a high count on parallel clients, as else the `MDS` seldom is the bottleneck. If you want to set this up please refer to the ceph documentation. footnote:[Configuring multiple active MDS -daemons http://docs.ceph.com/docs/luminous/cephfs/multimds/] +daemons https://docs.ceph.com/docs/{ceph_codename}/cephfs/multimds/] [[pveceph_fs_create]] Create CephFS @@ -679,7 +684,7 @@ This creates a CephFS named `'cephfs'' using a pool for its data named Check the xref:pve_ceph_pools[{pve} managed Ceph pool chapter] or visit the Ceph documentation for more information regarding a fitting placement group number (`pg_num`) for your setup footnote:[Ceph Placement Groups -http://docs.ceph.com/docs/luminous/rados/operations/placement-groups/]. +https://docs.ceph.com/docs/{ceph_codename}/rados/operations/placement-groups/]. Additionally, the `'--add-storage'' parameter will add the CephFS to the {pve} storage configuration after it was created successfully. @@ -712,6 +717,67 @@ pveceph pool destroy NAME ---- +Ceph maintenance +---------------- + +Replace OSDs +~~~~~~~~~~~~ + +One of the common maintenance tasks in Ceph is to replace a disk of an OSD. If +a disk is already in a failed state, then you can go ahead and run through the +steps in xref:pve_ceph_osd_destroy[Destroy OSDs]. Ceph will recreate those +copies on the remaining OSDs if possible. This rebalancing will start as soon +as an OSD failure is detected or an OSD was actively stopped. + +NOTE: With the default size/min_size (3/2) of a pool, recovery only starts when +`size + 1` nodes are available. The reason for this is that the Ceph object +balancer xref:pve_ceph_device_classes[CRUSH] defaults to a full node as +`failure domain'. + +To replace a still functioning disk, on the GUI go through the steps in +xref:pve_ceph_osd_destroy[Destroy OSDs]. The only addition is to wait until +the cluster shows 'HEALTH_OK' before stopping the OSD to destroy it. + +On the command line use the following commands. +---- +ceph osd out osd. +---- + +You can check with the command below if the OSD can be safely removed. +---- +ceph osd safe-to-destroy osd. +---- + +Once the above check tells you that it is save to remove the OSD, you can +continue with following commands. +---- +systemctl stop ceph-osd@.service +pveceph osd destroy +---- + +Replace the old disk with the new one and use the same procedure as described +in xref:pve_ceph_osd_create[Create OSDs]. + +Trim/Discard +~~~~~~~~~~~~ +It is a good measure to run 'fstrim' (discard) regularly on VMs or containers. +This releases data blocks that the filesystem isn’t using anymore. It reduces +data usage and resource load. Most modern operating systems issue such discard +commands to their disks regularly. You only need to ensure that the Virtual +Machines enable the xref:qm_hard_disk_discard[disk discard option]. + +[[pveceph_scrub]] +Scrub & Deep Scrub +~~~~~~~~~~~~~~~~~~ +Ceph ensures data integrity by 'scrubbing' placement groups. Ceph checks every +object in a PG for its health. There are two forms of Scrubbing, daily +cheap metadata checks and weekly deep data checks. The weekly deep scrub reads +the objects and uses checksums to ensure data integrity. If a running scrub +interferes with business (performance) needs, you can adjust the time when +scrubs footnote:[Ceph scrubbing https://docs.ceph.com/docs/{ceph_codename}/rados/configuration/osd-config-ref/#scrubbing] +are executed. + + Ceph monitoring and troubleshooting ----------------------------------- A good start is to continuosly monitor the ceph health from the start of @@ -732,10 +798,10 @@ pve# ceph -w To get a more detailed view, every ceph service has a log file under `/var/log/ceph/` and if there is not enough detail, the log level can be -adjusted footnote:[Ceph log and debugging http://docs.ceph.com/docs/luminous/rados/troubleshooting/log-and-debug/]. +adjusted footnote:[Ceph log and debugging https://docs.ceph.com/docs/{ceph_codename}/rados/troubleshooting/log-and-debug/]. You can find more information about troubleshooting -footnote:[Ceph troubleshooting http://docs.ceph.com/docs/luminous/rados/troubleshooting/] +footnote:[Ceph troubleshooting https://docs.ceph.com/docs/{ceph_codename}/rados/troubleshooting/] a Ceph cluster on the official website.