X-Git-Url: https://git.proxmox.com/?a=blobdiff_plain;f=ceph%2Fdoc%2Fcephfs%2Frecover-fs-after-mon-store-loss.rst;fp=ceph%2Fdoc%2Fcephfs%2Frecover-fs-after-mon-store-loss.rst;h=d44269c831880da15ec91ff946fa01e7146df62b;hb=20effc670b57271cb089376d6d0800990e5218d5;hp=0000000000000000000000000000000000000000;hpb=a71831dadd1e1f3e0fa70405511f65cc33db0498;p=ceph.git diff --git a/ceph/doc/cephfs/recover-fs-after-mon-store-loss.rst b/ceph/doc/cephfs/recover-fs-after-mon-store-loss.rst new file mode 100644 index 000000000..d44269c83 --- /dev/null +++ b/ceph/doc/cephfs/recover-fs-after-mon-store-loss.rst @@ -0,0 +1,51 @@ +Recovering the file system after catastrophic Monitor store loss +================================================================ + +During rare occasions, all the monitor stores of a cluster may get corrupted +or lost. To recover the cluster in such a scenario, you need to rebuild the +monitor stores using the OSDs (see :ref:`mon-store-recovery-using-osds`), +and get back the pools intact (active+clean state). However, the rebuilt monitor +stores don't restore the file system maps ("FSMap"). Additional steps are required +to bring back the file system. The steps to recover a multiple active MDS file +system or multiple file systems are yet to be identified. Currently, only the steps +to recover a **single active MDS** file system with no additional file systems +in the cluster have been identified and tested. Briefly the steps are: +recreate the FSMap with basic defaults; and allow MDSs to recover from +the journal/metadata stored in the filesystem's pools. The steps are described +in more detail below. + +First up, recreate the file system using the recovered file system pools. The +new FSMap will have the filesystem's default settings. However, the user defined +file system settings such as ``standby_count_wanted``, ``required_client_features``, +extra data pools, etc., are lost and need to be reapplied later. + +:: + + ceph fs new --force --recover + +The ``recover`` flag sets the state of file system's rank 0 to existing but +failed. So when a MDS daemon eventually picks up rank 0, the daemon reads the +existing in-RADOS metadata and doesn't overwrite it. The flag also prevents the +standby MDS daemons to activate the file system. + +The file system cluster ID, fscid, of the file system will not be preserved. +This behaviour may not be desirable for certain applications (e.g., Ceph CSI) +that expect the file system to be unchanged across recovery. To fix this, you +can optionally set the ``fscid`` option in the above command (see +:ref:`advanced-cephfs-admin-settings`). + +Allow standby MDS daemons to join the file system. + +:: + + ceph fs set joinable true + + +Check that the file system is no longer in degraded state and has an active +MDS. + +:: + + ceph fs status + +Reapply any other custom file system settings.