X-Git-Url: https://git.proxmox.com/?p=pve-docs.git;a=blobdiff_plain;f=pve-storage-rbd.adoc;h=7361935b16e88da905173a23b43f53a865fde6e5;hp=f8edf85e75807bda2e4c437731ca5fbd3d8d188d;hb=edc73653fb32890096a3b2fa15a7a95ff22e0882;hpb=8c1189b640ae7d10119ff1c046580f48749d38bd diff --git a/pve-storage-rbd.adoc b/pve-storage-rbd.adoc index f8edf85..7361935 100644 --- a/pve-storage-rbd.adoc +++ b/pve-storage-rbd.adoc @@ -2,6 +2,10 @@ Ceph RADOS Block Devices (RBD) ------------------------------ include::attributes.txt[] +ifdef::wiki[] +:pve-toplevel: +endif::wiki[] + Storage pool type: `rbd` http://ceph.com[Ceph] is a distributed object store and file system @@ -16,7 +20,7 @@ storage, and you get the following advantages: * self healing * no single point of failure * scalable to the exabyte level -* kernel and unser space implementation available +* kernel and user space implementation available NOTE: For smaller deployments, it is also possible to run Ceph services directly on your {pve} nodes. Recent hardware has plenty