-#!/bin/bash
+#!/usr/bin/env bash
+# -*- mode:sh; tab-width:4; sh-basic-offset:4; indent-tabs-mode:nil -*-
+# vim: softtabstop=4 shiftwidth=4 expandtab
# abort on failure
set -e
+quoted_print() {
+ for s in "$@"; do
+ if [[ "$s" =~ \ ]]; then
+ printf -- "'%s' " "$s"
+ else
+ printf -- "$s "
+ fi
+ done
+ printf '\n'
+}
+
+debug() {
+ "$@" >&2
+}
+
+prunb() {
+ debug quoted_print "$@" '&'
+ "$@" &
+}
+
+prun() {
+ debug quoted_print "$@"
+ "$@"
+}
+
+
if [ -n "$VSTART_DEST" ]; then
- SRC_PATH=`dirname $0`
- SRC_PATH=`(cd $SRC_PATH; pwd)`
+ SRC_PATH=`dirname $0`
+ SRC_PATH=`(cd $SRC_PATH; pwd)`
- CEPH_DIR=$SRC_PATH
- CEPH_BIN=${PWD}/bin
- CEPH_LIB=${PWD}/lib
+ CEPH_DIR=$SRC_PATH
+ CEPH_BIN=${PWD}/bin
+ CEPH_LIB=${PWD}/lib
- CEPH_CONF_PATH=$VSTART_DEST
- CEPH_DEV_DIR=$VSTART_DEST/dev
- CEPH_OUT_DIR=$VSTART_DEST/out
+ CEPH_CONF_PATH=$VSTART_DEST
+ CEPH_DEV_DIR=$VSTART_DEST/dev
+ CEPH_OUT_DIR=$VSTART_DEST/out
+ CEPH_ASOK_DIR=$VSTART_DEST/out
fi
+get_cmake_variable() {
+ local variable=$1
+ grep "${variable}:" CMakeCache.txt | cut -d "=" -f 2
+}
+
# for running out of the CMake build directory
if [ -e CMakeCache.txt ]; then
- # Out of tree build, learn source location from CMakeCache.txt
- CEPH_ROOT=`grep ceph_SOURCE_DIR CMakeCache.txt | cut -d "=" -f 2`
- CEPH_BUILD_DIR=`pwd`
- [ -z "$MGR_PYTHON_PATH" ] && MGR_PYTHON_PATH=$CEPH_ROOT/src/pybind/mgr
+ # Out of tree build, learn source location from CMakeCache.txt
+ CEPH_ROOT=$(get_cmake_variable ceph_SOURCE_DIR)
+ CEPH_BUILD_DIR=`pwd`
+ [ -z "$MGR_PYTHON_PATH" ] && MGR_PYTHON_PATH=$CEPH_ROOT/src/pybind/mgr
fi
-# use CEPH_BUILD_ROOT to vstart from a 'make install'
+# use CEPH_BUILD_ROOT to vstart from a 'make install'
if [ -n "$CEPH_BUILD_ROOT" ]; then
- [ -z "$CEPH_BIN" ] && CEPH_BIN=$CEPH_BUILD_ROOT/bin
- [ -z "$CEPH_LIB" ] && CEPH_LIB=$CEPH_BUILD_ROOT/lib
- [ -z "$EC_PATH" ] && EC_PATH=$CEPH_LIB/erasure-code
- [ -z "$OBJCLASS_PATH" ] && OBJCLASS_PATH=$CEPH_LIB/rados-classes
+ [ -z "$CEPH_BIN" ] && CEPH_BIN=$CEPH_BUILD_ROOT/bin
+ [ -z "$CEPH_LIB" ] && CEPH_LIB=$CEPH_BUILD_ROOT/lib
+ [ -z "$EC_PATH" ] && EC_PATH=$CEPH_LIB/erasure-code
+ [ -z "$OBJCLASS_PATH" ] && OBJCLASS_PATH=$CEPH_LIB/rados-classes
+ # make install should install python extensions into PYTHONPATH
elif [ -n "$CEPH_ROOT" ]; then
- [ -z "$PYBIND" ] && PYBIND=$CEPH_ROOT/src/pybind
- [ -z "$CEPH_BIN" ] && CEPH_BIN=$CEPH_BUILD_DIR/bin
- [ -z "$CEPH_ADM" ] && CEPH_ADM=$CEPH_BIN/ceph
- [ -z "$INIT_CEPH" ] && INIT_CEPH=$CEPH_BUILD_DIR/bin/init-ceph
- [ -z "$CEPH_LIB" ] && CEPH_LIB=$CEPH_BUILD_DIR/lib
- [ -z "$OBJCLASS_PATH" ] && OBJCLASS_PATH=$CEPH_LIB
- [ -z "$EC_PATH" ] && EC_PATH=$CEPH_LIB
+ [ -z "$CEPHFS_SHELL" ] && CEPHFS_SHELL=$CEPH_ROOT/src/tools/cephfs/cephfs-shell
+ [ -z "$PYBIND" ] && PYBIND=$CEPH_ROOT/src/pybind
+ [ -z "$CEPH_BIN" ] && CEPH_BIN=$CEPH_BUILD_DIR/bin
+ [ -z "$CEPH_ADM" ] && CEPH_ADM=$CEPH_BIN/ceph
+ [ -z "$INIT_CEPH" ] && INIT_CEPH=$CEPH_BIN/init-ceph
+ [ -z "$CEPH_LIB" ] && CEPH_LIB=$CEPH_BUILD_DIR/lib
+ [ -z "$OBJCLASS_PATH" ] && OBJCLASS_PATH=$CEPH_LIB
+ [ -z "$EC_PATH" ] && EC_PATH=$CEPH_LIB
+ [ -z "$CEPH_PYTHON_COMMON" ] && CEPH_PYTHON_COMMON=$CEPH_ROOT/src/python-common
fi
if [ -z "${CEPH_VSTART_WRAPPER}" ]; then
[ -z "$PYBIND" ] && PYBIND=./pybind
-export PYTHONPATH=$PYBIND:$CEPH_LIB/cython_modules/lib.2:$PYTHONPATH
+[ -n "$CEPH_PYTHON_COMMON" ] && CEPH_PYTHON_COMMON="$CEPH_PYTHON_COMMON:"
+CYTHON_PYTHONPATH="$CEPH_LIB/cython_modules/lib.3"
+export PYTHONPATH=$PYBIND:$CYTHON_PYTHONPATH:$CEPH_PYTHON_COMMON$PYTHONPATH
+
export LD_LIBRARY_PATH=$CEPH_LIB:$LD_LIBRARY_PATH
export DYLD_LIBRARY_PATH=$CEPH_LIB:$DYLD_LIBRARY_PATH
# Suppress logging for regular use that indicated that we are using a
-# development version. vstart.sh is only used during testing and
+# development version. vstart.sh is only used during testing and
# development
export CEPH_DEV=1
[ -z "$CEPH_NUM_MGR" ] && CEPH_NUM_MGR="$MGR"
[ -z "$CEPH_NUM_FS" ] && CEPH_NUM_FS="$FS"
[ -z "$CEPH_NUM_RGW" ] && CEPH_NUM_RGW="$RGW"
+[ -z "$GANESHA_DAEMON_NUM" ] && GANESHA_DAEMON_NUM="$NFS"
# if none of the CEPH_NUM_* number is specified, kill the existing
# cluster.
if [ -z "$CEPH_NUM_MON" -a \
-z "$CEPH_NUM_OSD" -a \
-z "$CEPH_NUM_MDS" -a \
- -z "$CEPH_NUM_MGR" ]; then
+ -z "$CEPH_NUM_MGR" -a \
+ -z "$GANESHA_DAEMON_NUM" ]; then
kill_all=1
else
kill_all=0
[ -z "$CEPH_NUM_FS" ] && CEPH_NUM_FS=1
[ -z "$CEPH_MAX_MDS" ] && CEPH_MAX_MDS=1
[ -z "$CEPH_NUM_RGW" ] && CEPH_NUM_RGW=0
+[ -z "$GANESHA_DAEMON_NUM" ] && GANESHA_DAEMON_NUM=0
[ -z "$CEPH_DIR" ] && CEPH_DIR="$PWD"
[ -z "$CEPH_DEV_DIR" ] && CEPH_DEV_DIR="$CEPH_DIR/dev"
debug=0
ip=""
nodaemon=0
+redirect=0
smallmds=0
short=0
ec=0
+cephadm=0
+parallel=true
hitset=""
-overwrite_conf=1
+overwrite_conf=0
cephx=1 #turn cephx on by default
+gssapi_authx=0
cache=""
-memstore=0
-bluestore=0
-rgw_frontend="civetweb"
+if [ `uname` = FreeBSD ]; then
+ objectstore="filestore"
+else
+ objectstore="bluestore"
+fi
+ceph_osd=ceph-osd
+rgw_frontend="beast"
rgw_compression=""
lockdep=${LOCKDEP:-1}
+spdk_enabled=0 #disable SPDK by default
+pci_id=""
+
+with_mgr_dashboard=true
+if [[ "$(get_cmake_variable WITH_MGR_DASHBOARD_FRONTEND)" != "ON" ]] ||
+ [[ "$(get_cmake_variable WITH_RBD)" != "ON" ]]; then
+ debug echo "ceph-mgr dashboard not built - disabling."
+ with_mgr_dashboard=false
+fi
filestore_path=
+kstore_path=
+bluestore_dev=
VSTART_SEC="client.vstart.sh"
keyring_fn="$CEPH_CONF_PATH/keyring"
osdmap_fn="/tmp/ceph_osdmap.$$"
monmap_fn="/tmp/ceph_monmap.$$"
+inc_osd_num=0
-usage="usage: $0 [option]... \nex: $0 -n -d --mon_num 3 --osd_num 3 --mds_num 1 --rgw_num 1\n"
+msgr="21"
+
+usage="usage: $0 [option]... \nex: MON=3 OSD=1 MDS=1 MGR=1 RGW=1 NFS=1 $0 -n -d\n"
usage=$usage"options:\n"
usage=$usage"\t-d, --debug\n"
usage=$usage"\t-s, --standby_mds: Generate standby-replay MDS for each active\n"
usage=$usage"\t-l, --localhost: use localhost instead of hostname\n"
usage=$usage"\t-i <ip>: bind to specific ip\n"
usage=$usage"\t-n, --new\n"
-usage=$usage"\t-N, --not-new: reuse existing cluster config (default)\n"
usage=$usage"\t--valgrind[_{osd,mds,mon,rgw}] 'toolname args...'\n"
usage=$usage"\t--nodaemon: use ceph-run as wrapper for mon/osd/mds\n"
-usage=$usage"\t--smallmds: limit mds cache size\n"
+usage=$usage"\t--redirect-output: only useful with nodaemon, directs output to log file\n"
+usage=$usage"\t--smallmds: limit mds cache memory limit\n"
usage=$usage"\t-m ip:port\t\tspecify monitor address\n"
-usage=$usage"\t-k keep old configuration files\n"
+usage=$usage"\t-k keep old configuration files (default)\n"
usage=$usage"\t-x enable cephx (on by default)\n"
usage=$usage"\t-X disable cephx\n"
+usage=$usage"\t-g --gssapi enable Kerberos/GSSApi authentication\n"
+usage=$usage"\t-G disable Kerberos/GSSApi authentication\n"
usage=$usage"\t--hitset <pool> <hit_set_type>: enable hitset tracking\n"
usage=$usage"\t-e : create an erasure pool\n";
usage=$usage"\t-o config\t\t add extra config parameters to all sections\n"
-usage=$usage"\t--mon_num specify ceph monitor count\n"
-usage=$usage"\t--osd_num specify ceph osd count\n"
-usage=$usage"\t--mds_num specify ceph mds count\n"
-usage=$usage"\t--rgw_num specify ceph rgw count\n"
-usage=$usage"\t--mgr_num specify ceph mgr count\n"
usage=$usage"\t--rgw_port specify ceph rgw http listen port\n"
usage=$usage"\t--rgw_frontend specify the rgw frontend configuration\n"
usage=$usage"\t--rgw_compression specify the rgw compression plugin\n"
-usage=$usage"\t-b, --bluestore use bluestore as the osd objectstore backend\n"
+usage=$usage"\t-b, --bluestore use bluestore as the osd objectstore backend (default)\n"
+usage=$usage"\t-f, --filestore use filestore as the osd objectstore backend\n"
+usage=$usage"\t-K, --kstore use kstore as the osd objectstore backend\n"
usage=$usage"\t--memstore use memstore as the osd objectstore backend\n"
usage=$usage"\t--cache <pool>: enable cache tiering on pool\n"
usage=$usage"\t--short: short object names only; necessary for ext4 dev\n"
usage=$usage"\t--nolockdep disable lockdep\n"
usage=$usage"\t--multimds <count> allow multimds with maximum active count\n"
+usage=$usage"\t--without-dashboard: do not run using mgr dashboard\n"
+usage=$usage"\t--bluestore-spdk <vendor>:<device>: enable SPDK and specify the PCI-ID of the NVME device\n"
+usage=$usage"\t--msgr1: use msgr1 only\n"
+usage=$usage"\t--msgr2: use msgr2 only\n"
+usage=$usage"\t--msgr21: use msgr2 and msgr1\n"
+usage=$usage"\t--crimson: use crimson-osd instead of ceph-osd\n"
+usage=$usage"\t--osd-args: specify any extra osd specific options\n"
+usage=$usage"\t--bluestore-devs: comma-separated list of blockdevs to use for bluestore\n"
+usage=$usage"\t--inc-osd: append some more osds into existing vcluster\n"
+usage=$usage"\t--cephadm: enable cephadm orchestrator with ~/.ssh/id_rsa[.pub]\n"
+usage=$usage"\t--no-parallel: dont start all OSDs in parallel\n"
usage_exit() {
- printf "$usage"
- exit
+ printf "$usage"
+ exit
}
while [ $# -ge 1 ]; do
case $1 in
-d | --debug )
- debug=1
- ;;
+ debug=1
+ ;;
-s | --standby_mds)
- standby=1
- ;;
+ standby=1
+ ;;
-l | --localhost )
- ip="127.0.0.1"
- ;;
+ ip="127.0.0.1"
+ ;;
-i )
- [ -z "$2" ] && usage_exit
- ip="$2"
- shift
- ;;
+ [ -z "$2" ] && usage_exit
+ ip="$2"
+ shift
+ ;;
-e )
- ec=1
- ;;
+ ec=1
+ ;;
--new | -n )
- new=1
- ;;
- --not-new | -N )
- new=0
- ;;
+ new=1
+ ;;
+ --inc-osd )
+ new=0
+ kill_all=0
+ inc_osd_num=$2
+ if [ "$inc_osd_num" == "" ]; then
+ inc_osd_num=1
+ else
+ shift
+ fi
+ ;;
--short )
- short=1
- ;;
+ short=1
+ ;;
+ --crimson )
+ ceph_osd=crimson-osd
+ ;;
+ --osd-args )
+ extra_osd_args="$2"
+ shift
+ ;;
+ --msgr1 )
+ msgr="1"
+ ;;
+ --msgr2 )
+ msgr="2"
+ ;;
+ --msgr21 )
+ msgr="21"
+ ;;
+ --cephadm )
+ cephadm=1
+ ;;
+ --no-parallel )
+ parallel=false
+ ;;
--valgrind )
- [ -z "$2" ] && usage_exit
- valgrind=$2
- shift
- ;;
+ [ -z "$2" ] && usage_exit
+ valgrind=$2
+ shift
+ ;;
--valgrind_args )
- valgrind_args="$2"
- shift
- ;;
+ valgrind_args="$2"
+ shift
+ ;;
--valgrind_mds )
- [ -z "$2" ] && usage_exit
- valgrind_mds=$2
- shift
- ;;
+ [ -z "$2" ] && usage_exit
+ valgrind_mds=$2
+ shift
+ ;;
--valgrind_osd )
- [ -z "$2" ] && usage_exit
- valgrind_osd=$2
- shift
- ;;
+ [ -z "$2" ] && usage_exit
+ valgrind_osd=$2
+ shift
+ ;;
--valgrind_mon )
- [ -z "$2" ] && usage_exit
- valgrind_mon=$2
- shift
- ;;
+ [ -z "$2" ] && usage_exit
+ valgrind_mon=$2
+ shift
+ ;;
--valgrind_mgr )
- [ -z "$2" ] && usage_exit
- valgrind_mgr=$2
- shift
- ;;
+ [ -z "$2" ] && usage_exit
+ valgrind_mgr=$2
+ shift
+ ;;
--valgrind_rgw )
- [ -z "$2" ] && usage_exit
- valgrind_rgw=$2
- shift
- ;;
+ [ -z "$2" ] && usage_exit
+ valgrind_rgw=$2
+ shift
+ ;;
--nodaemon )
- nodaemon=1
- ;;
+ nodaemon=1
+ ;;
+ --redirect-output)
+ redirect=1
+ ;;
--smallmds )
- smallmds=1
- ;;
- --mon_num )
- echo "mon_num:$2"
- CEPH_NUM_MON="$2"
- shift
- ;;
- --osd_num )
- CEPH_NUM_OSD=$2
- shift
- ;;
- --mds_num )
- CEPH_NUM_MDS=$2
- shift
- ;;
- --rgw_num )
- CEPH_NUM_RGW=$2
- shift
- ;;
- --mgr_num )
- CEPH_NUM_MGR=$2
- shift
- ;;
+ smallmds=1
+ ;;
--rgw_port )
- CEPH_RGW_PORT=$2
- shift
- ;;
+ CEPH_RGW_PORT=$2
+ shift
+ ;;
--rgw_frontend )
- rgw_frontend=$2
- shift
- ;;
+ rgw_frontend=$2
+ shift
+ ;;
--rgw_compression )
- rgw_compression=$2
- shift
- ;;
+ rgw_compression=$2
+ shift
+ ;;
+ --kstore_path )
+ kstore_path=$2
+ shift
+ ;;
--filestore_path )
- filestore_path=$2
- shift
- ;;
+ filestore_path=$2
+ shift
+ ;;
-m )
- [ -z "$2" ] && usage_exit
- MON_ADDR=$2
- shift
- ;;
+ [ -z "$2" ] && usage_exit
+ MON_ADDR=$2
+ shift
+ ;;
-x )
- cephx=1 # this is on be default, flag exists for historical consistency
- ;;
+ cephx=1 # this is on be default, flag exists for historical consistency
+ ;;
-X )
- cephx=0
- ;;
+ cephx=0
+ ;;
+
+ -g | --gssapi)
+ gssapi_authx=1
+ ;;
+ -G)
+ gssapi_authx=0
+ ;;
+
-k )
- if [ ! -r $conf_fn ]; then
- echo "cannot use old configuration: $conf_fn not readable." >&2
- exit
- fi
- overwrite_conf=0
- ;;
+ if [ ! -r $conf_fn ]; then
+ echo "cannot use old configuration: $conf_fn not readable." >&2
+ exit
+ fi
+ new=0
+ ;;
--memstore )
- memstore=1
- ;;
+ objectstore="memstore"
+ ;;
-b | --bluestore )
- bluestore=1
- ;;
+ objectstore="bluestore"
+ ;;
+ -f | --filestore )
+ objectstore="filestore"
+ ;;
+ -K | --kstore )
+ objectstore="kstore"
+ ;;
--hitset )
- hitset="$hitset $2 $3"
- shift
- shift
- ;;
+ hitset="$hitset $2 $3"
+ shift
+ shift
+ ;;
-o )
- extra_conf="$extra_conf $2
+ extra_conf="$extra_conf $2
"
- shift
- ;;
+ shift
+ ;;
--cache )
- if [ -z "$cache" ]; then
- cache="$2"
- else
- cache="$cache $2"
- fi
- shift
- ;;
+ if [ -z "$cache" ]; then
+ cache="$2"
+ else
+ cache="$cache $2"
+ fi
+ shift
+ ;;
--nolockdep )
- lockdep=0
- ;;
+ lockdep=0
+ ;;
--multimds)
CEPH_MAX_MDS="$2"
shift
;;
+ --without-dashboard)
+ with_mgr_dashboard=false
+ ;;
+ --bluestore-spdk )
+ [ -z "$2" ] && usage_exit
+ pci_id="$2"
+ spdk_enabled=1
+ shift
+ ;;
+ --bluestore-devs )
+ IFS=',' read -r -a bluestore_dev <<< "$2"
+ for dev in "${bluestore_dev[@]}"; do
+ if [ ! -b $dev -o ! -w $dev ]; then
+ echo "All --bluestore-devs must refer to writable block devices"
+ exit 1
+ fi
+ done
+ shift
+ ;;
* )
- usage_exit
+ usage_exit
esac
shift
done
$SUDO $INIT_CEPH stop
fi
-if [ "$overwrite_conf" -eq 0 ]; then
- MON=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC num_mon 2>/dev/null` && \
+if [ "$new" -eq 0 ]; then
+ if [ -z "$CEPH_ASOK_DIR" ]; then
+ CEPH_ASOK_DIR=`dirname $($CEPH_BIN/ceph-conf -c $conf_fn --show-config-value admin_socket)`
+ fi
+ mkdir -p $CEPH_ASOK_DIR
+ MON=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC --lookup num_mon 2>/dev/null` && \
CEPH_NUM_MON="$MON"
- OSD=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC num_osd 2>/dev/null` && \
+ OSD=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC --lookup num_osd 2>/dev/null` && \
CEPH_NUM_OSD="$OSD"
- MDS=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC num_mds 2>/dev/null` && \
+ MDS=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC --lookup num_mds 2>/dev/null` && \
CEPH_NUM_MDS="$MDS"
- MGR=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC num_mgr 2>/dev/null` && \
+ MGR=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC --lookup num_mgr 2>/dev/null` && \
CEPH_NUM_MGR="$MGR"
- RGW=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC num_rgw 2>/dev/null` && \
+ RGW=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC --lookup num_rgw 2>/dev/null` && \
CEPH_NUM_RGW="$RGW"
+ NFS=`$CEPH_BIN/ceph-conf -c $conf_fn --name $VSTART_SEC --lookup num_ganesha 2>/dev/null` && \
+ GANESHA_DAEMON_NUM="$NFS"
else
- if [ "$new" -ne 0 ]; then
- # only delete if -n
- [ -e "$conf_fn" ] && rm -- "$conf_fn"
- else
- # -k is implied... (doesn't make sense otherwise)
- overwrite_conf=0
+ # only delete if -n
+ if [ -e "$conf_fn" ]; then
+ asok_dir=`dirname $($CEPH_BIN/ceph-conf -c $conf_fn --show-config-value admin_socket)`
+ rm -- "$conf_fn"
+ if [ $asok_dir != /var/run/ceph ]; then
+ [ -d $asok_dir ] && rm -f $asok_dir/* && rmdir $asok_dir
+ fi
+ fi
+ if [ -z "$CEPH_ASOK_DIR" ]; then
+ CEPH_ASOK_DIR=`mktemp -u -d "${TMPDIR:-/tmp}/ceph-asok.XXXXXX"`
fi
fi
ARGS="-c $conf_fn"
-prunb() {
- echo "$* &"
- "$@" &
-}
-
-prun() {
- echo "$*"
- "$@"
-}
-
run() {
type=$1
shift
+ num=$1
+ shift
eval "valg=\$valgrind_$type"
[ -z "$valg" ] && valg="$valgrind"
else
if [ "$nodaemon" -eq 0 ]; then
prun "$@"
+ elif [ "$redirect" -eq 0 ]; then
+ prunb ${CEPH_ROOT}/src/ceph-run "$@" -f
else
- prunb ./ceph-run "$@" -f
+ ( prunb ${CEPH_ROOT}/src/ceph-run "$@" -f ) >$CEPH_OUT_DIR/$type.$num.stdout 2>&1
fi
fi
}
wconf() {
- if [ "$overwrite_conf" -eq 1 ]; then
- cat >> "$conf_fn"
- fi
+ if [ "$new" -eq 1 -o "$overwrite_conf" -eq 1 ]; then
+ cat >> "$conf_fn"
+ fi
+}
+
+get_pci_selector() {
+ which_pci=$1
+ lspci -mm -n -D -d $pci_id | cut -d ' ' -f 1 | sed -n $which_pci'p'
+}
+
+get_pci_selector_num() {
+ lspci -mm -n -D -d $pci_id | cut -d' ' -f 1 | wc -l
+}
+
+do_rgw_conf() {
+
+ if [ $CEPH_NUM_RGW -eq 0 ]; then
+ return 0
+ fi
+
+ # setup each rgw on a sequential port, starting at $CEPH_RGW_PORT.
+ # individual rgw's ids will be their ports.
+ current_port=$CEPH_RGW_PORT
+ for n in $(seq 1 $CEPH_NUM_RGW); do
+ wconf << EOF
+[client.rgw.${current_port}]
+ rgw frontends = $rgw_frontend port=${current_port}
+ admin socket = ${CEPH_OUT_DIR}/radosgw.${current_port}.asok
+EOF
+ current_port=$((current_port + 1))
+done
+
}
prepare_conf() {
local DAEMONOPTS="
log file = $CEPH_OUT_DIR/\$name.log
- admin socket = $CEPH_OUT_DIR/\$name.asok
+ admin socket = $CEPH_ASOK_DIR/\$name.asok
chdir = \"\"
pid file = $CEPH_OUT_DIR/\$name.pid
heartbeat file = $CEPH_OUT_DIR/\$name.heartbeat
"
+ local mgr_modules="restful iostat"
+ if $with_mgr_dashboard; then
+ mgr_modules="dashboard $mgr_modules"
+ fi
+
+ local msgr_conf=''
+ if [ $msgr -eq 21 ]; then
+ msgr_conf="
+ ms bind msgr2 = true
+ ms bind msgr1 = true
+";
+ fi
+ if [ $msgr -eq 2 ]; then
+ msgr_conf="
+ ms bind msgr2 = true
+ ms bind msgr1 = false
+";
+ fi
+ if [ $msgr -eq 1 ]; then
+ msgr_conf="
+ ms bind msgr2 = false
+ ms bind msgr1 = true
+";
+ fi
+
wconf <<EOF
; generated by vstart.sh on `date`
[$VSTART_SEC]
num mds = $CEPH_NUM_MDS
num mgr = $CEPH_NUM_MGR
num rgw = $CEPH_NUM_RGW
+ num ganesha = $GANESHA_DAEMON_NUM
[global]
fsid = $(uuidgen)
- osd pg bits = 3
- osd pgp bits = 5 ; (invalid, but ceph should cope!)
- osd pool default size = $OSD_POOL_DEFAULT_SIZE
- osd crush chooseleaf type = 0
- osd pool default min size = 1
osd failsafe full ratio = .99
- mon osd reporter subtree level = osd
mon osd full ratio = .99
- mon data avail warn = 10
- mon data avail crit = 1
+ mon osd nearfull ratio = .99
+ mon osd backfillfull ratio = .99
+ mon_max_pg_per_osd = ${MON_MAX_PG_PER_OSD:-1000}
erasure code dir = $EC_PATH
plugin dir = $CEPH_LIB
- osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd
- rgw frontends = $rgw_frontend port=$CEPH_RGW_PORT
- ; needed for s3tests
- rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= testkey-2=aWIKTWFrZWZpbGUKbWFuCm91dApzcmMKVGVzdGluZwo=
- rgw crypt require ssl = false
- rgw lc debug interval = 10
filestore fd cache size = 32
run dir = $CEPH_OUT_DIR
+ crash dir = $CEPH_OUT_DIR
enable experimental unrecoverable data corrupting features = *
+ osd_crush_chooseleaf_type = 0
+ debug asok assert abort = true
+$msgr_conf
+$extra_conf
EOF
- if [ "$lockdep" -eq 1 ] ; then
- wconf <<EOF
+ if [ "$lockdep" -eq 1 ] ; then
+ wconf <<EOF
lockdep = true
EOF
- fi
- if [ "$cephx" -eq 1 ] ; then
- wconf <<EOF
+ fi
+ if [ "$cephx" -eq 1 ] ; then
+ wconf <<EOF
auth cluster required = cephx
auth service required = cephx
auth client required = cephx
EOF
- else
- wconf <<EOF
- auth cluster required = none
- auth service required = none
- auth client required = none
+ elif [ "$gssapi_authx" -eq 1 ] ; then
+ wconf <<EOF
+ auth cluster required = gss
+ auth service required = gss
+ auth client required = gss
+ gss ktab client file = $CEPH_DEV_DIR/gss_\$name.keytab
EOF
- fi
- if [ "$short" -eq 1 ]; then
- COSDSHORT=" osd max object name len = 460
+ else
+ wconf <<EOF
+ auth cluster required = none
+ auth service required = none
+ auth client required = none
+EOF
+ fi
+ if [ "$short" -eq 1 ]; then
+ COSDSHORT=" osd max object name len = 460
osd max object namespace len = 64"
- fi
- wconf <<EOF
+ fi
+ if [ "$objectstore" == "bluestore" ]; then
+ if [ "$spdk_enabled" -eq 1 ]; then
+ if [ "$(get_pci_selector_num)" -eq 0 ]; then
+ echo "Not find the specified NVME device, please check." >&2
+ exit
+ fi
+ if [ $(get_pci_selector_num) -lt $CEPH_NUM_OSD ]; then
+ echo "OSD number ($CEPH_NUM_OSD) is greater than NVME SSD number ($(get_pci_selector_num)), please check." >&2
+ exit
+ fi
+ BLUESTORE_OPTS=" bluestore_block_db_path = \"\"
+ bluestore_block_db_size = 0
+ bluestore_block_db_create = false
+ bluestore_block_wal_path = \"\"
+ bluestore_block_wal_size = 0
+ bluestore_block_wal_create = false
+ bluestore_spdk_mem = 2048"
+ else
+ BLUESTORE_OPTS=" bluestore block db path = $CEPH_DEV_DIR/osd\$id/block.db.file
+ bluestore block db size = 1073741824
+ bluestore block db create = true
+ bluestore block wal path = $CEPH_DEV_DIR/osd\$id/block.wal.file
+ bluestore block wal size = 1048576000
+ bluestore block wal create = true"
+ fi
+ fi
+ wconf <<EOF
[client]
keyring = $keyring_fn
log file = $CEPH_OUT_DIR/\$name.\$pid.log
- admin socket = $CEPH_OUT_DIR/\$name.\$pid.asok
+ admin socket = $CEPH_ASOK_DIR/\$name.\$pid.asok
+
+ ; needed for s3tests
+ rgw crypt s3 kms backend = testing
+ rgw crypt s3 kms encryption keys = testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= testkey-2=aWIKTWFrZWZpbGUKbWFuCm91dApzcmMKVGVzdGluZwo=
+ rgw crypt require ssl = false
+ ; uncomment the following to set LC days as the value in seconds;
+ ; needed for passing lc time based s3-tests (can be verbose)
+ ; rgw lc debug interval = 10
-[client.rgw]
+$extra_conf
+EOF
+wconf <<EOF
+[cephfs-shell]
+ debug shell = true
+
+$extra_conf
+EOF
+ do_rgw_conf
+ wconf << EOF
[mds]
$DAEMONOPTS
-$CMDSDEBUG
- mds debug frag = true
- mds debug auth pins = true
- mds debug subtrees = true
mds data = $CEPH_DEV_DIR/mds.\$id
mds root ino uid = `id -u`
mds root ino gid = `id -g`
[mgr]
mgr data = $CEPH_DEV_DIR/mgr.\$id
mgr module path = $MGR_PYTHON_PATH
- mon reweight min pgs per osd = 4
- mon pg warn min per osd = 3
+ cephadm path = $CEPH_ROOT/src/cephadm/cephadm
$DAEMONOPTS
-$CMGRDEBUG
$extra_conf
[osd]
$DAEMONOPTS
osd class dir = $OBJCLASS_PATH
osd class load list = *
osd class default list = *
- osd scrub load threshold = 2000.0
- osd debug op order = true
- osd debug misdirected ops = true
+ osd fast shutdown = false
+
filestore wbthrottle xfs ios start flusher = 10
filestore wbthrottle xfs ios hard limit = 20
filestore wbthrottle xfs inodes hard limit = 30
filestore wbthrottle btrfs ios start flusher = 10
filestore wbthrottle btrfs ios hard limit = 20
filestore wbthrottle btrfs inodes hard limit = 30
- osd copyfrom max chunk = 524288
bluestore fsck on mount = true
bluestore block create = true
- bluestore block db size = 67108864
- bluestore block db create = true
- bluestore block wal size = 1048576000
- bluestore block wal create = true
-$COSDDEBUG
-$COSDMEMSTORE
+$BLUESTORE_OPTS
+
+ ; kstore
+ kstore fsck on mount = true
+ osd objectstore = $objectstore
$COSDSHORT
$extra_conf
[mon]
- mgr initial modules = restful status dashboard
- mon pg warn min per osd = 3
- mon osd allow primary affinity = true
- mon reweight min pgs per osd = 4
- mon osd prime pg temp = true
- crushtool = $CEPH_BIN/crushtool
- mon allow pool delete = true
+ mgr initial modules = $mgr_modules
$DAEMONOPTS
$CMONDEBUG
$extra_conf
mon cluster log file = $CEPH_OUT_DIR/cluster.mon.\$id.log
-[global]
-$extra_conf
+ osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd
EOF
}
do
[ $count -eq $CEPH_NUM_MON ] && break;
count=$(($count + 1))
- if [ -z "$MONS" ];
- then
- MONS="$f"
+ if [ -z "$MONS" ]; then
+ MONS="$f"
else
- MONS="$MONS $f"
+ MONS="$MONS $f"
fi
done
if [ "$new" -eq 1 ]; then
- if [ `echo $IP | grep '^127\\.'` ]
- then
- echo
- echo "NOTE: hostname resolves to loopback; remote hosts will not be able to"
- echo " connect. either adjust /etc/hosts, or edit this script to use your"
- echo " machine's real IP."
- echo
- fi
-
- prun $SUDO "$CEPH_BIN/ceph-authtool" --create-keyring --gen-key --name=mon. "$keyring_fn" --cap mon 'allow *'
- prun $SUDO "$CEPH_BIN/ceph-authtool" --gen-key --name=client.admin --set-uid=0 \
- --cap mon 'allow *' \
- --cap osd 'allow *' \
- --cap mds 'allow *' \
- --cap mgr 'allow *' \
- "$keyring_fn"
-
- prun $SUDO "$CEPH_BIN/ceph-authtool" --gen-key --name=client.rgw \
- --cap mon 'allow rw' \
- --cap osd 'allow rwx' \
- --cap mgr 'allow rw' \
- "$keyring_fn"
-
- # build a fresh fs monmap, mon fs
- local str=""
- local count=0
- for f in $MONS
- do
- str="$str --add $f $IP:$(($CEPH_PORT+$count))"
- wconf <<EOF
+ if [ `echo $IP | grep '^127\\.'` ]; then
+ echo
+ echo "NOTE: hostname resolves to loopback; remote hosts will not be able to"
+ echo " connect. either adjust /etc/hosts, or edit this script to use your"
+ echo " machine's real IP."
+ echo
+ fi
+
+ prun $SUDO "$CEPH_BIN/ceph-authtool" --create-keyring --gen-key --name=mon. "$keyring_fn" --cap mon 'allow *'
+ prun $SUDO "$CEPH_BIN/ceph-authtool" --gen-key --name=client.admin \
+ --cap mon 'allow *' \
+ --cap osd 'allow *' \
+ --cap mds 'allow *' \
+ --cap mgr 'allow *' \
+ "$keyring_fn"
+
+ # build a fresh fs monmap, mon fs
+ local params=()
+ local count=0
+ local mon_host=""
+ for f in $MONS
+ do
+ if [ $msgr -eq 1 ]; then
+ A="v1:$IP:$(($CEPH_PORT+$count+1))"
+ fi
+ if [ $msgr -eq 2 ]; then
+ A="v2:$IP:$(($CEPH_PORT+$count+1))"
+ fi
+ if [ $msgr -eq 21 ]; then
+ A="[v2:$IP:$(($CEPH_PORT+$count)),v1:$IP:$(($CEPH_PORT+$count+1))]"
+ fi
+ params+=("--addv" "$f" "$A")
+ mon_host="$mon_host $A"
+ wconf <<EOF
[mon.$f]
host = $HOSTNAME
mon data = $CEPH_DEV_DIR/mon.$f
- mon addr = $IP:$(($CEPH_PORT+$count))
EOF
- count=$(($count + 1))
- done
- prun "$CEPH_BIN/monmaptool" --create --clobber $str --print "$monmap_fn"
-
- for f in $MONS
- do
- prun rm -rf -- "$CEPH_DEV_DIR/mon.$f"
- prun mkdir -p "$CEPH_DEV_DIR/mon.$f"
- prun "$CEPH_BIN/ceph-mon" --mkfs -c "$conf_fn" -i "$f" --monmap="$monmap_fn" --keyring="$keyring_fn"
- done
-
- prun rm -- "$monmap_fn"
- fi
+ count=$(($count + 2))
+ done
+ wconf <<EOF
+[global]
+ mon host = $mon_host
+EOF
+ prun "$CEPH_BIN/monmaptool" --create --clobber "${params[@]}" --print "$monmap_fn"
+
+ for f in $MONS
+ do
+ prun rm -rf -- "$CEPH_DEV_DIR/mon.$f"
+ prun mkdir -p "$CEPH_DEV_DIR/mon.$f"
+ prun "$CEPH_BIN/ceph-mon" --mkfs -c "$conf_fn" -i "$f" --monmap="$monmap_fn" --keyring="$keyring_fn"
+ done
- # start monitors
- for f in $MONS
- do
- run 'mon' $CEPH_BIN/ceph-mon -i $f $ARGS $CMON_ARGS
- done
+ prun rm -- "$monmap_fn"
+ fi
+
+ # start monitors
+ for f in $MONS
+ do
+ run 'mon' $f $CEPH_BIN/ceph-mon -i $f $ARGS $CMON_ARGS
+ done
}
start_osd() {
- for osd in `seq 0 $((CEPH_NUM_OSD-1))`
+ if [ $inc_osd_num -gt 0 ]; then
+ old_maxosd=$($CEPH_BIN/ceph osd getmaxosd | sed -e 's/max_osd = //' -e 's/ in epoch.*//')
+ start=$old_maxosd
+ end=$(($start-1+$inc_osd_num))
+ overwrite_conf=1 # fake wconf
+ else
+ start=0
+ end=$(($CEPH_NUM_OSD-1))
+ fi
+ local osds_wait
+ for osd in `seq $start $end`
do
- if [ "$new" -eq 1 ]; then
- wconf <<EOF
+ local extra_seastar_args
+ if [ "$ceph_osd" == "crimson-osd" ]; then
+ # designate a single CPU node $osd for osd.$osd
+ extra_seastar_args="--smp 1 --cpuset $osd"
+ if [ "$debug" -ne 0 ]; then
+ extra_seastar_args+=" --debug"
+ fi
+ fi
+ if [ "$new" -eq 1 -o $inc_osd_num -gt 0 ]; then
+ wconf <<EOF
[osd.$osd]
host = $HOSTNAME
EOF
+ if [ "$spdk_enabled" -eq 1 ]; then
+ wconf <<EOF
+ bluestore_block_path = spdk:$(get_pci_selector $((osd+1)))
+EOF
+ fi
rm -rf $CEPH_DEV_DIR/osd$osd || true
if command -v btrfs > /dev/null; then
for f in $CEPH_DEV_DIR/osd$osd/*; do btrfs sub delete $f &> /dev/null || true; done
fi
- if [ -n "$filestore_path" ]; then
- ln -s $filestore_path $CEPH_DEV_DIR/osd$osd
- else
- mkdir -p $CEPH_DEV_DIR/osd$osd
- fi
+ if [ -n "$filestore_path" ]; then
+ ln -s $filestore_path $CEPH_DEV_DIR/osd$osd
+ elif [ -n "$kstore_path" ]; then
+ ln -s $kstore_path $CEPH_DEV_DIR/osd$osd
+ else
+ mkdir -p $CEPH_DEV_DIR/osd$osd
+ if [ -n "${bluestore_dev[$osd]}" ]; then
+ dd if=/dev/zero of=${bluestore_dev[$osd]} bs=1M count=1
+ ln -s ${bluestore_dev[$osd]} $CEPH_DEV_DIR/osd$osd/block
+ wconf <<EOF
+ bluestore fsck on mount = false
+EOF
+ fi
+ fi
local uuid=`uuidgen`
echo "add osd$osd $uuid"
- ceph_adm osd create $uuid
- ceph_adm osd crush add osd.$osd 1.0 host=$HOSTNAME root=default
- $SUDO $CEPH_BIN/ceph-osd -i $osd $ARGS --mkfs --mkkey --osd-uuid $uuid
+ OSD_SECRET=$($CEPH_BIN/ceph-authtool --gen-print-key)
+ echo "{\"cephx_secret\": \"$OSD_SECRET\"}" > $CEPH_DEV_DIR/osd$osd/new.json
+ ceph_adm osd new $uuid -i $CEPH_DEV_DIR/osd$osd/new.json
+ rm $CEPH_DEV_DIR/osd$osd/new.json
+ $SUDO $CEPH_BIN/$ceph_osd $extra_osd_args -i $osd $ARGS --mkfs --key $OSD_SECRET --osd-uuid $uuid $extra_seastar_args
local key_fn=$CEPH_DEV_DIR/osd$osd/keyring
- echo adding osd$osd key to auth repository
- ceph_adm -i "$key_fn" auth add osd.$osd osd "allow *" mon "allow profile osd" mgr "allow profile osd"
+ cat > $key_fn<<EOF
+[osd.$osd]
+ key = $OSD_SECRET
+EOF
fi
echo start osd.$osd
- run 'osd' $SUDO $CEPH_BIN/ceph-osd -i $osd $ARGS $COSD_ARGS
+ local osd_pid
+ run 'osd' $osd $SUDO $CEPH_BIN/$ceph_osd \
+ $extra_seastar_args $extra_osd_args \
+ -i $osd $ARGS $COSD_ARGS &
+ osd_pid=$!
+ if $parallel; then
+ osds_wait=$osd_pid
+ else
+ wait $osd_pid
+ fi
done
+ if $parallel; then
+ for p in $osds_wait; do
+ wait $p
+ done
+ debug echo OSDs started
+ fi
+ if [ $inc_osd_num -gt 0 ]; then
+ # update num osd
+ new_maxosd=$($CEPH_BIN/ceph osd getmaxosd | sed -e 's/max_osd = //' -e 's/ in epoch.*//')
+ sed -i "s/num osd = .*/num osd = $new_maxosd/g" $conf_fn
+ fi
}
start_mgr() {
local mgr=0
+ local ssl=${DASHBOARD_SSL:-1}
# avoid monitors on nearby ports (which test/*.sh use extensively)
MGR_PORT=$(($CEPH_PORT + 1000))
+ PROMETHEUS_PORT=9283
for name in x y z a b c d e f g h i j k l m n o p
do
[ $mgr -eq $CEPH_NUM_MGR ] && break
key_fn=$CEPH_DEV_DIR/mgr.$name/keyring
$SUDO $CEPH_BIN/ceph-authtool --create-keyring --gen-key --name=mgr.$name $key_fn
ceph_adm -i $key_fn auth add mgr.$name mon 'allow profile mgr' mds 'allow *' osd 'allow *'
- fi
- wconf <<EOF
+ wconf <<EOF
[mgr.$name]
host = $HOSTNAME
EOF
- ceph_adm config-key put mgr/dashboard/$name/server_port $MGR_PORT
- DASH_URLS+="http://$IP:$MGR_PORT/"
- MGR_PORT=$(($MGR_PORT + 1000))
+ if $with_mgr_dashboard ; then
+ local port_option="ssl_server_port"
+ local http_proto="https"
+ if [ "$ssl" == "0" ]; then
+ port_option="server_port"
+ http_proto="http"
+ ceph_adm config set mgr mgr/dashboard/ssl false --force
+ fi
+ ceph_adm config set mgr mgr/dashboard/$name/$port_option $MGR_PORT --force
+ if [ $mgr -eq 1 ]; then
+ DASH_URLS="$http_proto://$IP:$MGR_PORT"
+ else
+ DASH_URLS+=", $http_proto://$IP:$MGR_PORT"
+ fi
+ fi
+ MGR_PORT=$(($MGR_PORT + 1000))
+ ceph_adm config set mgr mgr/prometheus/$name/server_port $PROMETHEUS_PORT --force
+ PROMETHEUS_PORT=$(($PROMETHEUS_PORT + 1000))
+
+ ceph_adm config set mgr mgr/restful/$name/server_port $MGR_PORT --force
+ if [ $mgr -eq 1 ]; then
+ RESTFUL_URLS="https://$IP:$MGR_PORT"
+ else
+ RESTFUL_URLS+=", https://$IP:$MGR_PORT"
+ fi
+ MGR_PORT=$(($MGR_PORT + 1000))
+ fi
- ceph_adm config-key put mgr/restful/$name/server_port $MGR_PORT
+ debug echo "Starting mgr.${name}"
+ run 'mgr' $name $CEPH_BIN/ceph-mgr -i $name $ARGS
+ done
- RESTFUL_URLS+="https://$IP:$MGR_PORT"
- MGR_PORT=$(($MGR_PORT + 1000))
+ if [ "$new" -eq 1 ]; then
+ # setting login credentials for dashboard
+ if $with_mgr_dashboard; then
+ while ! ceph_adm -h | grep -c -q ^dashboard ; do
+ debug echo 'waiting for mgr dashboard module to start'
+ sleep 1
+ done
+ ceph_adm dashboard ac-user-create --force-password admin admin administrator
+ if [ "$ssl" != "0" ]; then
+ if ! ceph_adm dashboard create-self-signed-cert; then
+ debug echo dashboard module not working correctly!
+ fi
+ fi
+ fi
- echo "Starting mgr.${name}"
- run 'mgr' $CEPH_BIN/ceph-mgr -i $name $ARGS
- done
+ while ! ceph_adm -h | grep -c -q ^restful ; do
+ debug echo 'waiting for mgr restful module to start'
+ sleep 1
+ done
+ if ceph_adm restful create-self-signed-cert; then
+ SF=`mktemp`
+ ceph_adm restful create-key admin -o $SF
+ RESTFUL_SECRET=`cat $SF`
+ rm $SF
+ else
+ debug echo MGR Restful is not working, perhaps the package is not installed?
+ fi
+ fi
- if ceph_adm tell mgr restful create-self-signed-cert; then
- SF=`mktemp`
- ceph_adm tell mgr restful create-key admin -o $SF
- RESTFUL_SECRET=`cat $SF`
- rm $SF
- else
- echo MGR Restful is not working, perhaps the package is not installed?
+ if [ "$cephadm" -eq 1 ]; then
+ debug echo Enabling cephadm orchestrator
+ if [ "$new" -eq 1 ]; then
+ digest=$(curl -s \
+ https://registry.hub.docker.com/v2/repositories/ceph/daemon-base/tags/latest-master-devel \
+ | jq -r '.images[].digest')
+ ceph_adm config set global container_image "docker.io/ceph/daemon-base@$digest"
+ fi
+ ceph_adm config-key set mgr/cephadm/ssh_identity_key -i ~/.ssh/id_rsa
+ ceph_adm config-key set mgr/cephadm/ssh_identity_pub -i ~/.ssh/id_rsa.pub
+ ceph_adm mgr module enable cephadm
+ ceph_adm orch set backend cephadm
+ ceph_adm orch host add "$(hostname)"
+ ceph_adm orch apply crash '*'
+ ceph_adm config set mgr mgr/cephadm/allow_ptrace true
fi
}
start_mds() {
+ local mds=0
+ for name in a b c d e f g h i j k l m n o p
+ do
+ [ $mds -eq $CEPH_NUM_MDS ] && break
+ mds=$(($mds + 1))
+
+ if [ "$new" -eq 1 ]; then
+ prun mkdir -p "$CEPH_DEV_DIR/mds.$name"
+ key_fn=$CEPH_DEV_DIR/mds.$name/keyring
+ wconf <<EOF
+[mds.$name]
+ host = $HOSTNAME
+EOF
+ if [ "$standby" -eq 1 ]; then
+ mkdir -p $CEPH_DEV_DIR/mds.${name}s
+ wconf <<EOF
+ mds standby for rank = $mds
+[mds.${name}s]
+ mds standby replay = true
+ mds standby for name = ${name}
+EOF
+ fi
+ prun $SUDO "$CEPH_BIN/ceph-authtool" --create-keyring --gen-key --name="mds.$name" "$key_fn"
+ ceph_adm -i "$key_fn" auth add "mds.$name" mon 'allow profile mds' osd 'allow rw tag cephfs *=*' mds 'allow' mgr 'allow profile mds'
+ if [ "$standby" -eq 1 ]; then
+ prun $SUDO "$CEPH_BIN/ceph-authtool" --create-keyring --gen-key --name="mds.${name}s" \
+ "$CEPH_DEV_DIR/mds.${name}s/keyring"
+ ceph_adm -i "$CEPH_DEV_DIR/mds.${name}s/keyring" auth add "mds.${name}s" \
+ mon 'allow profile mds' osd 'allow *' mds 'allow' mgr 'allow profile mds'
+ fi
+ fi
+
+ run 'mds' $name $CEPH_BIN/ceph-mds -i $name $ARGS $CMDS_ARGS
+ if [ "$standby" -eq 1 ]; then
+ run 'mds' $name $CEPH_BIN/ceph-mds -i ${name}s $ARGS $CMDS_ARGS
+ fi
+
+ #valgrind --tool=massif $CEPH_BIN/ceph-mds $ARGS --mds_log_max_segments 2 --mds_thrash_fragments 0 --mds_thrash_exports 0 > m #--debug_ms 20
+ #$CEPH_BIN/ceph-mds -d $ARGS --mds_thrash_fragments 0 --mds_thrash_exports 0 #--debug_ms 20
+ #ceph_adm mds set max_mds 2
+ done
+
if [ $new -eq 1 ]; then
if [ "$CEPH_NUM_FS" -gt "0" ] ; then
+ sleep 5 # time for MDS to come up as standby to avoid health warnings on fs creation
if [ "$CEPH_NUM_FS" -gt "1" ] ; then
ceph_adm fs flag set enable_multiple true --yes-i-really-mean-it
fi
+ # wait for volume module to load
+ while ! ceph_adm fs volume ls ; do sleep 1 ; done
local fs=0
for name in a b c d e f g h i j k l m n o p
do
- ceph_adm osd pool create "cephfs_data_${name}" 8
- ceph_adm osd pool create "cephfs_metadata_${name}" 8
- ceph_adm fs new "cephfs_${name}" "cephfs_metadata_${name}" "cephfs_data_${name}"
+ ceph_adm fs volume create ${name}
+ ceph_adm fs authorize ${name} "client.fs_${name}" / rwp >> "$keyring_fn"
fs=$(($fs + 1))
[ $fs -eq $CEPH_NUM_FS ] && break
done
fi
fi
- local mds=0
+}
+
+# Ganesha Daemons requires nfs-ganesha nfs-ganesha-ceph nfs-ganesha-rados-grace
+# nfs-ganesha-rados-urls (version 3.3 and above) packages installed. On
+# Fedora>=31 these packages can be installed directly with 'dnf'. For CentOS>=8
+# the packages are available at
+# https://wiki.centos.org/SpecialInterestGroup/Storage
+# Similarly for Ubuntu>=16.04 follow the instructions on
+# https://launchpad.net/~nfs-ganesha
+
+start_ganesha() {
+ cluster_id="vstart"
+ GANESHA_PORT=$(($CEPH_PORT + 4000))
+ local ganesha=0
+ test_user="ganesha-$cluster_id"
+ pool_name="nfs-ganesha"
+ namespace=$cluster_id
+ url="rados://$pool_name/$namespace/conf-nfs.$test_user"
+
+ prun ceph_adm auth get-or-create client.$test_user \
+ mon "allow r" \
+ osd "allow rw pool=$pool_name namespace=$namespace, allow rw tag cephfs data=a" \
+ mds "allow rw path=/" \
+ >> "$keyring_fn"
+
+ ceph_adm mgr module enable test_orchestrator
+ ceph_adm orch set backend test_orchestrator
+ ceph_adm test_orchestrator load_data -i $CEPH_ROOT/src/pybind/mgr/test_orchestrator/dummy_data.json
+ prun ceph_adm nfs cluster create cephfs $cluster_id
+ prun ceph_adm nfs export create cephfs "a" $cluster_id "/cephfs"
+
for name in a b c d e f g h i j k l m n o p
do
- [ $mds -eq $CEPH_NUM_MDS ] && break
- mds=$(($mds + 1))
-
- if [ "$new" -eq 1 ]; then
- prun mkdir -p "$CEPH_DEV_DIR/mds.$name"
- key_fn=$CEPH_DEV_DIR/mds.$name/keyring
- wconf <<EOF
-[mds.$name]
+ [ $ganesha -eq $GANESHA_DAEMON_NUM ] && break
+
+ port=$(($GANESHA_PORT + ganesha))
+ ganesha=$(($ganesha + 1))
+ ganesha_dir="$CEPH_DEV_DIR/ganesha.$name"
+ prun rm -rf $ganesha_dir
+ prun mkdir -p $ganesha_dir
+
+ echo "NFS_CORE_PARAM {
+ Enable_NLM = false;
+ Enable_RQUOTA = false;
+ Protocols = 4;
+ NFS_Port = $port;
+ }
+
+ MDCACHE {
+ Dir_Chunk = 0;
+ }
+
+ NFSv4 {
+ RecoveryBackend = rados_cluster;
+ Minor_Versions = 1, 2;
+ }
+
+ %url $url
+
+ RADOS_KV {
+ pool = $pool_name;
+ namespace = $namespace;
+ UserId = $test_user;
+ nodeid = $name;
+ }
+
+ RADOS_URLS {
+ Userid = $test_user;
+ watch_url = \"$url\";
+ }" > "$ganesha_dir/ganesha-$name.conf"
+ wconf <<EOF
+[ganesha.$name]
host = $HOSTNAME
+ ip = $IP
+ port = $port
+ ganesha data = $ganesha_dir
+ pid file = $ganesha_dir/ganesha-$name.pid
EOF
- if [ "$standby" -eq 1 ]; then
- mkdir -p $CEPH_DEV_DIR/mds.${name}s
- wconf <<EOF
- mds standby for rank = $mds
-[mds.${name}s]
- mds standby replay = true
- mds standby for name = ${name}
-EOF
- fi
- prun $SUDO "$CEPH_BIN/ceph-authtool" --create-keyring --gen-key --name="mds.$name" "$key_fn"
- ceph_adm -i "$key_fn" auth add "mds.$name" mon 'allow profile mds' osd 'allow *' mds 'allow' mgr 'allow profile mds'
- if [ "$standby" -eq 1 ]; then
- prun $SUDO "$CEPH_BIN/ceph-authtool" --create-keyring --gen-key --name="mds.${name}s" \
- "$CEPH_DEV_DIR/mds.${name}s/keyring"
- ceph_adm -i "$CEPH_DEV_DIR/mds.${name}s/keyring" auth add "mds.${name}s" \
- mon 'allow profile mds' osd 'allow *' mds 'allow' mgr 'allow profile mds'
- fi
- fi
- run 'mds' $CEPH_BIN/ceph-mds -i $name $ARGS $CMDS_ARGS
- if [ "$standby" -eq 1 ]; then
- run 'mds' $CEPH_BIN/ceph-mds -i ${name}s $ARGS $CMDS_ARGS
- fi
+ prun env CEPH_CONF="${conf_fn}" ganesha-rados-grace --userid $test_user -p $pool_name -n $namespace add $name
+ prun env CEPH_CONF="${conf_fn}" ganesha-rados-grace --userid $test_user -p $pool_name -n $namespace
- #valgrind --tool=massif $CEPH_BIN/ceph-mds $ARGS --mds_log_max_segments 2 --mds_thrash_fragments 0 --mds_thrash_exports 0 > m #--debug_ms 20
- #$CEPH_BIN/ceph-mds -d $ARGS --mds_thrash_fragments 0 --mds_thrash_exports 0 #--debug_ms 20
- #ceph_adm mds set max_mds 2
+ prun env CEPH_CONF="${conf_fn}" ganesha.nfsd -L "$CEPH_OUT_DIR/ganesha-$name.log" -f "$ganesha_dir/ganesha-$name.conf" -p "$CEPH_OUT_DIR/ganesha-$name.pid" -N NIV_DEBUG
+
+ # Wait few seconds for grace period to be removed
+ sleep 2
+
+ prun env CEPH_CONF="${conf_fn}" ganesha-rados-grace --userid $test_user -p $pool_name -n $namespace
+
+ if $with_mgr_dashboard; then
+ $CEPH_BIN/rados -p $pool_name put "conf-$name" "$ganesha_dir/ganesha-$name.conf"
+ fi
+
+ echo "$test_user ganesha daemon $name started on port: $port"
done
+
+ if $with_mgr_dashboard; then
+ ceph_adm dashboard set-ganesha-clusters-rados-pool-namespace $pool_name
+ fi
}
if [ "$debug" -eq 0 ]; then
CMONDEBUG='
debug mon = 10
debug ms = 1'
- COSDDEBUG='
- debug ms = 1'
- CMDSDEBUG='
- debug ms = 1'
- CMGRDEBUG='
- debug ms = 1'
else
- echo "** going verbose **"
+ debug echo "** going verbose **"
CMONDEBUG='
debug mon = 20
debug paxos = 20
debug auth = 20
- debug mgrc = 20
- debug ms = 1'
- COSDDEBUG='
- debug ms = 1
- debug osd = 25
- debug objecter = 20
- debug monc = 20
debug mgrc = 20
- debug journal = 20
- debug filestore = 20
- debug bluestore = 30
- debug bluefs = 20
- debug rocksdb = 10
- debug bdev = 20
- debug rgw = 20
- debug objclass = 20'
- CMDSDEBUG='
- debug ms = 1
- debug mds = 20
- debug auth = 20
- debug monc = 20
- debug mgrc = 20
- mds debug scatterstat = true
- mds verify scatter = true
- mds log max segments = 2'
- CMGRDEBUG='
- debug ms = 1
- debug monc = 20
- debug mon = 20
- debug mgr = 20'
+ debug ms = 1'
fi
if [ -n "$MON_ADDR" ]; then
- CMON_ARGS=" -m "$MON_ADDR
- COSD_ARGS=" -m "$MON_ADDR
- CMDS_ARGS=" -m "$MON_ADDR
-fi
-
-if [ "$memstore" -eq 1 ]; then
- COSDMEMSTORE='
- osd objectstore = memstore'
-fi
-if [ "$bluestore" -eq 1 ]; then
- COSDMEMSTORE='
- osd objectstore = bluestore'
+ CMON_ARGS=" -m "$MON_ADDR
+ COSD_ARGS=" -m "$MON_ADDR
+ CMDS_ARGS=" -m "$MON_ADDR
fi
if [ -z "$CEPH_PORT" ]; then
- CEPH_PORT=6789
- [ -e ".ceph_port" ] && CEPH_PORT=`cat .ceph_port`
+ while [ true ]
+ do
+ CEPH_PORT="$(echo $(( RANDOM % 1000 + 40000 )))"
+ ss -a -n | egrep "\<LISTEN\>.+:${CEPH_PORT}\s+" 1>/dev/null 2>&1 || break
+ done
fi
[ -z "$INIT_CEPH" ] && INIT_CEPH=$CEPH_BIN/init-ceph
# sudo if btrfs
-test -d $CEPH_DEV_DIR/osd0/. && test -e $CEPH_DEV_DIR/sudo && SUDO="sudo"
+[ -d $CEPH_DEV_DIR/osd0/. ] && [ -e $CEPH_DEV_DIR/sudo ] && SUDO="sudo"
-prun $SUDO rm -f core*
+if [ $inc_osd_num -eq 0 ]; then
+ prun $SUDO rm -f core*
+fi
-test -d $CEPH_OUT_DIR || mkdir $CEPH_OUT_DIR
-test -d $CEPH_DEV_DIR || mkdir $CEPH_DEV_DIR
-$SUDO rm -rf $CEPH_OUT_DIR/*
-test -d gmon && $SUDO rm -rf gmon/*
+[ -d $CEPH_ASOK_DIR ] || mkdir -p $CEPH_ASOK_DIR
+[ -d $CEPH_OUT_DIR ] || mkdir -p $CEPH_OUT_DIR
+[ -d $CEPH_DEV_DIR ] || mkdir -p $CEPH_DEV_DIR
+if [ $inc_osd_num -eq 0 ]; then
+ $SUDO rm -rf $CEPH_OUT_DIR/*
+fi
+[ -d gmon ] && $SUDO rm -rf gmon/*
-[ "$cephx" -eq 1 ] && [ "$new" -eq 1 ] && test -e $keyring_fn && rm $keyring_fn
+[ "$cephx" -eq 1 ] && [ "$new" -eq 1 ] && [ -e $keyring_fn ] && rm $keyring_fn
# figure machine's ip
else
echo hostname $HOSTNAME
if [ -x "$(which ip 2>/dev/null)" ]; then
- IP_CMD="ip addr"
+ IP_CMD="ip addr"
else
- IP_CMD="ifconfig"
+ IP_CMD="ifconfig"
fi
- # filter out IPv6 and localhost addresses
+ # filter out IPv4 and localhost addresses
IP="$($IP_CMD | sed -En 's/127.0.0.1//;s/.*inet (addr:)?(([0-9]*\.){3}[0-9]*).*/\2/p' | head -n1)"
# if nothing left, try using localhost address, it might work
if [ -z "$IP" ]; then IP="127.0.0.1"; fi
fi
}
+if [ $inc_osd_num -gt 0 ]; then
+ start_osd
+ exit
+fi
+
if [ "$new" -eq 1 ]; then
prepare_conf
fi
if [ $CEPH_NUM_MON -gt 0 ]; then
start_mon
+
+ debug echo Populating config ...
+ cat <<EOF | $CEPH_BIN/ceph -c $conf_fn config assimilate-conf -i -
+[global]
+osd_pool_default_size = $OSD_POOL_DEFAULT_SIZE
+osd_pool_default_min_size = 1
+
+[mon]
+mon_osd_reporter_subtree_level = osd
+mon_data_avail_warn = 2
+mon_data_avail_crit = 1
+mon_allow_pool_delete = true
+
+[osd]
+osd_scrub_load_threshold = 2000
+osd_debug_op_order = true
+osd_debug_misdirected_ops = true
+osd_copyfrom_max_chunk = 524288
+
+[mds]
+mds_debug_frag = true
+mds_debug_auth_pins = true
+mds_debug_subtrees = true
+
+[mgr]
+mgr/telemetry/nag = false
+mgr/telemetry/enable = false
+
+EOF
+
+ if [ "$debug" -ne 0 ]; then
+ debug echo Setting debug configs ...
+ cat <<EOF | $CEPH_BIN/ceph -c $conf_fn config assimilate-conf -i -
+[mgr]
+debug_ms = 1
+debug_mgr = 20
+debug_monc = 20
+debug_mon = 20
+
+[osd]
+debug_ms = 1
+debug_osd = 25
+debug_objecter = 20
+debug_monc = 20
+debug_mgrc = 20
+debug_journal = 20
+debug_filestore = 20
+debug_bluestore = 20
+debug_bluefs = 20
+debug_rocksdb = 20
+debug_bdev = 20
+debug_reserver = 10
+debug_objclass = 20
+
+[mds]
+debug_ms = 1
+debug_mds = 20
+debug_monc = 20
+debug_mgrc = 20
+mds_debug_scatterstat = true
+mds_verify_scatter = true
+EOF
+ fi
+ if [ "$cephadm" -gt 0 ]; then
+ debug echo Setting mon public_network ...
+ public_network=$(ip route list | grep -w "$IP" | awk '{print $1}')
+ ceph_adm config set mon public_network $public_network
+ fi
+fi
+
+if [ $CEPH_NUM_MGR -gt 0 ]; then
+ start_mgr
fi
# osd
if [ "$smallmds" -eq 1 ]; then
wconf <<EOF
[mds]
- mds log max segments = 2
- mds cache size = 10000
+ mds log max segments = 2
+ # Default 'mds cache memory limit' is 1GiB, and here we set it to 100MiB.
+ mds cache memory limit = 100M
EOF
fi
if [ $CEPH_NUM_MDS -gt 0 ]; then
start_mds
+ # key with access to all FS
+ ceph_adm fs authorize \* "client.fs" / rwp >> "$keyring_fn"
fi
# Don't set max_mds until all the daemons are started, otherwise
[ $fs -eq $CEPH_NUM_FS ] && break
fs=$(($fs + 1))
if [ "$CEPH_MAX_MDS" -gt 1 ]; then
- ceph_adm fs set "cephfs_${name}" allow_multimds true --yes-i-really-mean-it
ceph_adm fs set "cephfs_${name}" max_mds "$CEPH_MAX_MDS"
fi
done
# mgr
-if [ $CEPH_NUM_MGR -gt 0 ]; then
- start_mgr
-fi
-
if [ "$ec" -eq 1 ]; then
ceph_adm <<EOF
osd erasure-code-profile set ec-profile m=2 k=2
-osd pool create ec 8 8 erasure ec-profile
+osd pool create ec erasure ec-profile
EOF
fi
+# Ganesha Daemons
+if [ $GANESHA_DAEMON_NUM -gt 0 ]; then
+ pseudo_path="/cephfs"
+ if [ "$cephadm" -gt 0 ]; then
+ cluster_id="vstart"
+ prun ceph_adm nfs cluster create cephfs $cluster_id
+ prun ceph_adm nfs export create cephfs "a" $cluster_id $pseudo_path
+ port="2049"
+ else
+ start_ganesha
+ port="<ganesha-port-num>"
+ fi
+ echo "Mount using: mount -t nfs -o port=$port $IP:$pseudo_path mountpoint"
+fi
+
do_cache() {
while [ -n "$*" ]; do
- p="$1"
- shift
- echo "creating cache for pool $p ..."
- ceph_adm <<EOF
-osd pool create ${p}-cache 8
+ p="$1"
+ shift
+ debug echo "creating cache for pool $p ..."
+ ceph_adm <<EOF
+osd pool create ${p}-cache
osd tier add $p ${p}-cache
osd tier cache-mode ${p}-cache writeback
osd tier set-overlay $p ${p}-cache
do_hitsets() {
while [ -n "$*" ]; do
- pool="$1"
- type="$2"
- shift
- shift
- echo "setting hit_set on pool $pool type $type ..."
- ceph_adm <<EOF
+ pool="$1"
+ type="$2"
+ shift
+ shift
+ debug echo "setting hit_set on pool $pool type $type ..."
+ ceph_adm <<EOF
osd pool set $pool hit_set_type $type
osd pool set $pool hit_set_count 8
osd pool set $pool hit_set_period 30
# Create S3 user
local akey='0555b35654ad1656d804'
local skey='h7GhxuBLTrlhVUyxSPUKUV8r/2EI4ngqJxD7iBdBYLhwluN30JaT3Q=='
- echo "setting up user testid"
+ debug echo "setting up user testid"
$CEPH_BIN/radosgw-admin user create --uid testid --access-key $akey --secret $skey --display-name 'M. Tester' --email tester@ceph.com -c $conf_fn > /dev/null
# Create S3-test users
# See: https://github.com/ceph/s3-tests
- echo "setting up s3-test users"
+ debug echo "setting up s3-test users"
$CEPH_BIN/radosgw-admin user create \
--uid 0123456789abcdef0123456789abcdef0123456789abcdef0123456789abcdef \
--access-key ABCDEFGHIJKLMNOPQRST \
--email tenanteduser@example.com -c $conf_fn > /dev/null
# Create Swift user
- echo "setting up user tester"
+ debug echo "setting up user tester"
$CEPH_BIN/radosgw-admin user create -c $conf_fn --subuser=test:tester --display-name=Tester-Subuser --key-type=swift --secret=testing --access=full > /dev/null
echo ""
do_rgw()
{
if [ "$new" -eq 1 ]; then
- do_rgw_create_users
+ do_rgw_create_users
if [ -n "$rgw_compression" ]; then
- echo "setting compression type=$rgw_compression"
+ debug echo "setting compression type=$rgw_compression"
$CEPH_BIN/radosgw-admin zone placement modify -c $conf_fn --rgw-zone=default --placement-id=default-placement --compression=$rgw_compression > /dev/null
fi
fi
# Start server
RGWDEBUG=""
if [ "$debug" -ne 0 ]; then
- RGWDEBUG="--debug-rgw=20"
+ RGWDEBUG="--debug-rgw=20 --debug-ms=1"
fi
+ local CEPH_RGW_PORT_NUM="${CEPH_RGW_PORT}"
+ local CEPH_RGW_HTTPS="${CEPH_RGW_PORT: -1}"
+ if [[ "${CEPH_RGW_HTTPS}" = "s" ]]; then
+ CEPH_RGW_PORT_NUM="${CEPH_RGW_PORT::-1}"
+ else
+ CEPH_RGW_HTTPS=""
+ fi
RGWSUDO=
- [ $CEPH_RGW_PORT -lt 1024 ] && RGWSUDO=sudo
- n=$(($CEPH_NUM_RGW - 1))
- i=0
- for rgw in j k l m n o p q r s t u v; do
- echo start rgw on http://localhost:$((CEPH_RGW_PORT + i))
- run 'rgw' $RGWSUDO $CEPH_BIN/radosgw -c $conf_fn --log-file=${CEPH_OUT_DIR}/rgw.$rgw.log ${RGWDEBUG} --debug-ms=1 -n client.rgw "--rgw_frontends=${rgw_frontend} port=$((CEPH_RGW_PORT + i))"
- i=$(($i + 1))
+ [ $CEPH_RGW_PORT_NUM -lt 1024 ] && RGWSUDO=sudo
+
+ current_port=$CEPH_RGW_PORT
+ for n in $(seq 1 $CEPH_NUM_RGW); do
+ rgw_name="client.rgw.${current_port}"
+
+ ceph_adm auth get-or-create $rgw_name \
+ mon 'allow rw' \
+ osd 'allow rwx' \
+ mgr 'allow rw' \
+ >> "$keyring_fn"
+
+ debug echo start rgw on http${CEPH_RGW_HTTPS}://localhost:${current_port}
+ run 'rgw' $current_port $RGWSUDO $CEPH_BIN/radosgw -c $conf_fn \
+ --log-file=${CEPH_OUT_DIR}/radosgw.${current_port}.log \
+ --admin-socket=${CEPH_OUT_DIR}/radosgw.${current_port}.asok \
+ --pid-file=${CEPH_OUT_DIR}/radosgw.${current_port}.pid \
+ ${RGWDEBUG} \
+ -n ${rgw_name} \
+ "--rgw_frontends=${rgw_frontend} port=${current_port}${CEPH_RGW_HTTPS}"
+
+ i=$(($i + 1))
[ $i -eq $CEPH_NUM_RGW ] && break
+
+ current_port=$((current_port+1))
done
}
if [ "$CEPH_NUM_RGW" -gt 0 ]; then
do_rgw
fi
-echo "started. stop.sh to stop. see out/* (e.g. 'tail -f out/????') for debug output."
+debug echo "vstart cluster complete. Use stop.sh to stop. See out/* (e.g. 'tail -f out/????') for debug output."
echo ""
-echo "dashboard urls: $DASH_URLS"
-echo " restful urls: $RESTFUL_URLS"
-echo " w/ user/pass: admin / $RESTFUL_SECRET"
+if [ "$new" -eq 1 ]; then
+ if $with_mgr_dashboard; then
+ echo "dashboard urls: $DASH_URLS"
+ echo " w/ user/pass: admin / admin"
+ fi
+ echo "restful urls: $RESTFUL_URLS"
+ echo " w/ user/pass: admin / $RESTFUL_SECRET"
+ echo ""
+fi
echo ""
-echo "export PYTHONPATH=./pybind:$PYTHONPATH"
-echo "export LD_LIBRARY_PATH=$CEPH_LIB"
+# add header to the environment file
+{
+ echo "#"
+ echo "# source this file into your shell to set up the environment."
+ echo "# For example:"
+ echo "# $ . $CEPH_DIR/vstart_environment.sh"
+ echo "#"
+} > $CEPH_DIR/vstart_environment.sh
+{
+ echo "export PYTHONPATH=$PYBIND:$CYTHON_PYTHONPATH:$CEPH_PYTHON_COMMON\$PYTHONPATH"
+ echo "export LD_LIBRARY_PATH=$CEPH_LIB:\$LD_LIBRARY_PATH"
-if [ "$CEPH_DIR" != "$PWD" ]; then
- echo "export CEPH_CONF=$conf_fn"
- echo "export CEPH_KEYRING=$keyring_fn"
-fi
+ if [ "$CEPH_DIR" != "$PWD" ]; then
+ echo "export CEPH_CONF=$conf_fn"
+ echo "export CEPH_KEYRING=$keyring_fn"
+ fi
+
+ if [ -n "$CEPHFS_SHELL" ]; then
+ echo "alias cephfs-shell=$CEPHFS_SHELL"
+ fi
+} | tee -a $CEPH_DIR/vstart_environment.sh
echo "CEPH_DEV=1"
+# always keep this section at the very bottom of this file
+STRAY_CONF_PATH="/etc/ceph/ceph.conf"
+if [ -f "$STRAY_CONF_PATH" -a -n "$conf_fn" -a ! "$conf_fn" -ef "$STRAY_CONF_PATH" ]; then
+ echo ""
+ echo ""
+ echo "WARNING:"
+ echo " Please remove stray $STRAY_CONF_PATH if not needed."
+ echo " Your conf files $conf_fn and $STRAY_CONF_PATH may not be in sync"
+ echo " and may lead to undesired results."
+ echo ""
+ echo "NOTE:"
+ echo " Remember to restart cluster after removing $STRAY_CONF_PATH"
+fi