X-Git-Url: https://git.proxmox.com/?a=blobdiff_plain;f=ceph%2Fsrc%2Fmon%2FMonCommands.h;h=21635974c606836df105c284e13a3ebd103afc43;hb=e306af509c4d4816a1f73b17a825ea5186fa0030;hp=bd255fe5b2e5470774f7876893f801ee825f4a69;hpb=224ce89bce8186937e77bdbda572a650953f8c23;p=ceph.git diff --git a/ceph/src/mon/MonCommands.h b/ceph/src/mon/MonCommands.h index bd255fe5b..21635974c 100644 --- a/ceph/src/mon/MonCommands.h +++ b/ceph/src/mon/MonCommands.h @@ -1,18 +1,18 @@ -// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*- +// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*- // vim: ts=8 sw=2 smarttab /* * Ceph - scalable distributed file system * - * Copyright (C) 2013 Inktank Storage, Inc. + * Copyright (C) 2013 Inktank Storage, Inc. * Copyright (C) 2013,2014 Cloudwatt * * Author: Loic Dachary * * This is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public - * License version 2.1, as published by the Free Software + * License version 2.1, as published by the Free Software * Foundation. See file COPYING. - * + * */ /* no guard; may be included multiple times */ @@ -32,7 +32,6 @@ * mds, osd, pg (osd), mon, auth, log, config-key, mgr * req perms: required permission in that modulename space to execute command * this also controls what type of REST command is accepted - * availability: cli, rest, or both * * The commands describe themselves completely enough for the separate * frontend(s) to be able to accept user input and validate it against @@ -43,11 +42,11 @@ * each descriptor is either a literal string, which can contain no spaces or * '=' signs (for instance, in "pg stat", both "pg" and "stat" are literal * strings representing one descriptor each), or a list of key=val[,key=val...] - * which also includes no spaces. + * which also includes no spaces. * * The key=val form describes a non-literal parameter. Each will have at * least a name= and type=, and each type can have its own type-specific - * parameters. The parser is the arbiter of these types and their + * parameters. The parser is the arbiter of these types and their * interpretation. A few more non-type-specific key=val pairs exist: * * req=false marks an optional parameter (default for req is 'true') @@ -76,10 +75,10 @@ * * Example: * - * COMMAND("auth add " \ - * "name=entity,type=CephString " \ - * "name=caps,type=CephString,n=N,req=false", \ - * "add auth info for from input file, or random key " \ + * COMMAND("auth add " + * "name=entity,type=CephString " + * "name=caps,type=CephString,n=N,req=false", + * "add auth info for from input file, or random key " * "if no input given, and/or any caps specified in the command") * * defines a command "auth add" that takes a required argument "entity" @@ -89,7 +88,7 @@ * enters auth add client.admin 'mon rwx' 'osd *'. The result will be a * JSON object like {"prefix":"auth add", "entity":"client.admin", * "caps":["mon rwx", "osd *"]}. - * Note that + * Note that * - string literals are accumulated into 'prefix' * - n=1 descriptors are given normal string or int object values * - n=N descriptors are given array values @@ -112,32 +111,23 @@ * OBSOLETE - command is considered obsolete * DEPRECATED - command is considered deprecated * MGR - command goes to ceph-mgr (for luminous+) + * POLL - command is intended to be called periodically by the + * client (see iostat) + * HIDDEN - command is hidden (no reported by help etc) + * TELL - tell/asok command. it's an alias of (NOFORWARD | HIDDEN) * * A command should always be first considered DEPRECATED before being * considered OBSOLETE, giving due consideration to users and conforming * to any guidelines regarding deprecating commands. */ -/* - * pg commands PGMonitor.cc - */ - -// note: this should be replaced shortly! -COMMAND("pg force_create_pg name=pgid,type=CephPgid", \ - "force creation of pg ", "pg", "rw", "cli,rest") -COMMAND_WITH_FLAG("pg set_full_ratio name=ratio,type=CephFloat,range=0.0|1.0", \ - "set ratio at which pgs are considered full", \ - "pg", "rw", "cli,rest", FLAG(DEPRECATED)) -COMMAND_WITH_FLAG("pg set_nearfull_ratio " \ - "name=ratio,type=CephFloat,range=0.0|1.0", \ - "set ratio at which pgs are considered nearly full", \ - "pg", "rw", "cli,rest", FLAG(DEPRECATED)) - COMMAND("pg map name=pgid,type=CephPgid", "show mapping of pg to osds", \ - "pg", "r", "cli,rest") + "pg", "r") +COMMAND("pg repeer name=pgid,type=CephPgid", "force a PG to repeer", + "osd", "rw") COMMAND("osd last-stat-seq name=id,type=CephOsdName", \ "get the last pg stats sequence number reported for this osd", \ - "osd", "r", "cli,rest") + "osd", "r") /* * auth commands AuthMonitor.cc @@ -145,837 +135,1162 @@ COMMAND("osd last-stat-seq name=id,type=CephOsdName", \ COMMAND("auth export name=entity,type=CephString,req=false", \ "write keyring for requested entity, or master keyring if none given", \ - "auth", "rx", "cli,rest") + "auth", "rx") COMMAND("auth get name=entity,type=CephString", \ - "write keyring file with requested key", "auth", "rx", "cli,rest") + "write keyring file with requested key", "auth", "rx") COMMAND("auth get-key name=entity,type=CephString", "display requested key", \ - "auth", "rx", "cli,rest") + "auth", "rx") COMMAND("auth print-key name=entity,type=CephString", "display requested key", \ - "auth", "rx", "cli,rest") + "auth", "rx") COMMAND("auth print_key name=entity,type=CephString", "display requested key", \ - "auth", "rx", "cli,rest") -COMMAND("auth list", "list authentication state", "auth", "rx", "cli,rest") -COMMAND("auth import", "auth import: read keyring file from -i ", \ - "auth", "rwx", "cli,rest") -COMMAND("auth add " \ - "name=entity,type=CephString " \ - "name=caps,type=CephString,n=N,req=false", \ - "add auth info for from input file, or random key if no " \ + "auth", "rx") +COMMAND_WITH_FLAG("auth list", "list authentication state", "auth", "rx", + FLAG(DEPRECATED)) +COMMAND("auth ls", "list authentication state", "auth", "rx") +COMMAND("auth import", "auth import: read keyring file from -i ", + "auth", "rwx") +COMMAND("auth add " + "name=entity,type=CephString " + "name=caps,type=CephString,n=N,req=false", + "add auth info for from input file, or random key if no " "input is given, and/or any caps specified in the command", - "auth", "rwx", "cli,rest") -COMMAND("auth get-or-create-key " \ - "name=entity,type=CephString " \ - "name=caps,type=CephString,n=N,req=false", \ - "get, or add, key for from system/caps pairs specified in the command. If key already exists, any given caps must match the existing caps for that key.", \ - "auth", "rwx", "cli,rest") -COMMAND("auth get-or-create " \ - "name=entity,type=CephString " \ - "name=caps,type=CephString,n=N,req=false", \ - "add auth info for from input file, or random key if no input given, and/or any caps specified in the command", \ - "auth", "rwx", "cli,rest") -COMMAND("auth caps " \ - "name=entity,type=CephString " \ - "name=caps,type=CephString,n=N", \ - "update caps for from caps specified in the command", \ - "auth", "rwx", "cli,rest") -COMMAND("auth del " \ - "name=entity,type=CephString", \ - "delete all caps for ", \ - "auth", "rwx", "cli,rest") -COMMAND("auth rm " \ - "name=entity,type=CephString", \ - "remove all caps for ", \ - "auth", "rwx", "cli,rest") + "auth", "rwx") +COMMAND("auth get-or-create-key " + "name=entity,type=CephString " + "name=caps,type=CephString,n=N,req=false", + "get, or add, key for from system/caps pairs specified in the command. If key already exists, any given caps must match the existing caps for that key.", + "auth", "rwx") +COMMAND("auth get-or-create " + "name=entity,type=CephString " + "name=caps,type=CephString,n=N,req=false", + "add auth info for from input file, or random key if no input given, and/or any caps specified in the command", + "auth", "rwx") +COMMAND("fs authorize " + "name=filesystem,type=CephString " + "name=entity,type=CephString " + "name=caps,type=CephString,n=N", + "add auth for to access file system based on following directory and permissions pairs", + "auth", "rwx") +COMMAND("auth caps " + "name=entity,type=CephString " + "name=caps,type=CephString,n=N", + "update caps for from caps specified in the command", + "auth", "rwx") +COMMAND_WITH_FLAG("auth del " + "name=entity,type=CephString", + "delete all caps for ", + "auth", "rwx", + FLAG(DEPRECATED)) +COMMAND("auth rm " + "name=entity,type=CephString", + "remove all caps for ", + "auth", "rwx") /* * Monitor commands (Monitor.cc) */ -COMMAND_WITH_FLAG("compact", "cause compaction of monitor's leveldb storage", \ - "mon", "rw", "cli,rest", \ - FLAG(NOFORWARD)|FLAG(DEPRECATED)) -COMMAND_WITH_FLAG("scrub", "scrub the monitor stores", \ - "mon", "rw", "cli,rest", \ - FLAG(DEPRECATED)) -COMMAND("fsid", "show cluster FSID/UUID", "mon", "r", "cli,rest") -COMMAND("log name=logtext,type=CephString,n=N", \ - "log supplied text to the monitor log", "mon", "rw", "cli,rest") +COMMAND_WITH_FLAG("compact", "cause compaction of monitor's leveldb/rocksdb storage", + "mon", "rw", + FLAG(TELL)) +COMMAND_WITH_FLAG("scrub", "scrub the monitor stores", + "mon", "rw", + FLAG(OBSOLETE)) +COMMAND("fsid", "show cluster FSID/UUID", "mon", "r") +COMMAND("log name=logtext,type=CephString,n=N", + "log supplied text to the monitor log", "mon", "rw") COMMAND("log last " "name=num,type=CephInt,range=1,req=false " "name=level,type=CephChoices,strings=debug|info|sec|warn|error,req=false " - "name=channel,type=CephChoices,strings=*|cluster|audit,req=false", \ - "print last few lines of the cluster log", \ - "mon", "rw", "cli,rest") -COMMAND_WITH_FLAG("injectargs " \ - "name=injected_args,type=CephString,n=N", \ - "inject config arguments into monitor", "mon", "rw", "cli,rest", - FLAG(NOFORWARD)) -COMMAND("status", "show cluster status", "mon", "r", "cli,rest") -COMMAND("health name=detail,type=CephChoices,strings=detail,req=false", \ - "show cluster health", "mon", "r", "cli,rest") -COMMAND("time-sync-status", "show time sync status", "mon", "r", "cli,rest") -COMMAND("df name=detail,type=CephChoices,strings=detail,req=false", \ - "show cluster free space stats", "mon", "r", "cli,rest") -COMMAND("report name=tags,type=CephString,n=N,req=false", \ - "report full status of cluster, optional title tag strings", \ - "mon", "r", "cli,rest") -COMMAND("features", "report of connected features", \ - "mon", "r", "cli,rest") -COMMAND("quorum_status", "report status of monitor quorum", \ - "mon", "r", "cli,rest") + "name=channel,type=CephChoices,strings=*|cluster|audit|cephadm,req=false", + "print last few lines of the cluster log", + "mon", "r") + +COMMAND("status", "show cluster status", "mon", "r") +COMMAND("health name=detail,type=CephChoices,strings=detail,req=false", + "show cluster health", "mon", "r") +COMMAND("health mute "\ + "name=code,type=CephString " + "name=ttl,type=CephString,req=false " + "name=sticky,type=CephBool,req=false", + "mute health alert", "mon", "w") +COMMAND("health unmute "\ + "name=code,type=CephString,req=false", + "unmute existing health alert mute(s)", "mon", "w") +COMMAND("time-sync-status", "show time sync status", "mon", "r") +COMMAND("df name=detail,type=CephChoices,strings=detail,req=false", + "show cluster free space stats", "mon", "r") +COMMAND("report name=tags,type=CephString,n=N,req=false", + "report full status of cluster, optional title tag strings", + "mon", "r") +COMMAND("features", "report of connected features", + "mon", "r") +COMMAND("quorum_status", "report status of monitor quorum", + "mon", "r") +COMMAND("mon ok-to-stop " + "name=ids,type=CephString,n=N", + "check whether mon(s) can be safely stopped without reducing immediate " + "availability", + "mon", "r") +COMMAND("mon ok-to-add-offline", + "check whether adding a mon and not starting it would break quorum", + "mon", "r") +COMMAND("mon ok-to-rm " + "name=id,type=CephString", + "check whether removing the specified mon would break quorum", + "mon", "r") -COMMAND_WITH_FLAG("mon_status", "report status of monitors", "mon", "r", "cli,rest", - FLAG(NOFORWARD)) -COMMAND_WITH_FLAG("sync force " \ - "name=validate1,type=CephChoices,strings=--yes-i-really-mean-it,req=false " \ - "name=validate2,type=CephChoices,strings=--i-know-what-i-am-doing,req=false", \ - "force sync of and clear monitor store", \ - "mon", "rw", "cli,rest", \ - FLAG(NOFORWARD)|FLAG(DEPRECATED)) -COMMAND_WITH_FLAG("heap " \ - "name=heapcmd,type=CephChoices,strings=dump|start_profiler|stop_profiler|release|stats", \ - "show heap usage info (available only if compiled with tcmalloc)", \ - "mon", "rw", "cli,rest", FLAG(NOFORWARD)) -COMMAND("quorum name=quorumcmd,type=CephChoices,strings=enter|exit,n=1", \ - "enter or exit quorum", "mon", "rw", "cli,rest") -COMMAND("tell " \ - "name=target,type=CephName " \ - "name=args,type=CephString,n=N", \ - "send a command to a specific daemon", "mon", "rw", "cli,rest") -COMMAND_WITH_FLAG("version", "show mon daemon version", "mon", "r", "cli,rest", - FLAG(NOFORWARD)) +COMMAND("tell " + "name=target,type=CephName " + "name=args,type=CephString,n=N", + "send a command to a specific daemon", "mon", "rw") +COMMAND_WITH_FLAG("version", "show mon daemon version", "mon", "r", + FLAG(TELL)) -COMMAND("node ls " \ - "name=type,type=CephChoices,strings=all|osd|mon|mds,req=false", - "list all nodes in cluster [type]", "mon", "r", "cli,rest") +COMMAND("node ls " + "name=type,type=CephChoices,strings=all|osd|mon|mds|mgr,req=false", + "list all nodes in cluster [type]", "mon", "r") /* * Monitor-specific commands under module 'mon' */ -COMMAND_WITH_FLAG("mon compact", \ - "cause compaction of monitor's leveldb storage", \ - "mon", "rw", "cli,rest", \ - FLAG(NOFORWARD)) COMMAND_WITH_FLAG("mon scrub", - "scrub the monitor stores", \ - "mon", "rw", "cli,rest", \ + "scrub the monitor stores", + "mon", "rw", FLAG(NONE)) -COMMAND_WITH_FLAG("mon sync force " \ - "name=validate1,type=CephChoices,strings=--yes-i-really-mean-it,req=false " \ - "name=validate2,type=CephChoices,strings=--i-know-what-i-am-doing,req=false", \ - "force sync of and clear monitor store", \ - "mon", "rw", "cli,rest", \ - FLAG(NOFORWARD)) COMMAND("mon metadata name=id,type=CephString,req=false", "fetch metadata for mon ", - "mon", "r", "cli,rest") + "mon", "r") COMMAND("mon count-metadata name=property,type=CephString", "count mons by metadata field property", - "mon", "r", "cli,rest") + "mon", "r") COMMAND("mon versions", "check running versions of monitors", - "mon", "r", "cli,rest") + "mon", "r") +COMMAND("versions", + "check running versions of ceph daemons", + "mon", "r") + /* * MDS commands (MDSMonitor.cc) */ -COMMAND("mds stat", "show MDS status", "mds", "r", "cli,rest") +COMMAND_WITH_FLAG("mds stat", "show MDS status", "mds", "r", FLAG(HIDDEN)) COMMAND_WITH_FLAG("mds dump " - "name=epoch,type=CephInt,req=false,range=0", \ + "name=epoch,type=CephInt,req=false,range=0", "dump legacy MDS cluster info, optionally from epoch", - "mds", "r", "cli,rest", FLAG(DEPRECATED)) + "mds", "r", FLAG(OBSOLETE)) COMMAND("fs dump " - "name=epoch,type=CephInt,req=false,range=0", \ - "dump all CephFS status, optionally from epoch", "mds", "r", "cli,rest") -COMMAND_WITH_FLAG("mds getmap " \ - "name=epoch,type=CephInt,req=false,range=0", \ - "get MDS map, optionally from epoch", "mds", "r", "cli,rest", FLAG(DEPRECATED)) + "name=epoch,type=CephInt,req=false,range=0", + "dump all CephFS status, optionally from epoch", "mds", "r") +COMMAND_WITH_FLAG("mds getmap " + "name=epoch,type=CephInt,req=false,range=0", + "get MDS map, optionally from epoch", "mds", "r", FLAG(OBSOLETE)) COMMAND("mds metadata name=who,type=CephString,req=false", - "fetch metadata for mds ", - "mds", "r", "cli,rest") + "fetch metadata for mds ", + "mds", "r") COMMAND("mds count-metadata name=property,type=CephString", "count MDSs by metadata field property", - "mds", "r", "cli,rest") + "mds", "r") COMMAND("mds versions", "check running versions of MDSs", - "mds", "r", "cli,rest") -COMMAND_WITH_FLAG("mds tell " \ - "name=who,type=CephString " \ - "name=args,type=CephString,n=N", \ - "send command to particular mds", "mds", "rw", "cli,rest", FLAG(OBSOLETE)) -COMMAND("mds compat show", "show mds compatibility settings", \ - "mds", "r", "cli,rest") -COMMAND_WITH_FLAG("mds stop name=who,type=CephString", "stop mds", \ - "mds", "rw", "cli,rest", FLAG(DEPRECATED)) -COMMAND("mds deactivate name=who,type=CephString", - "clean up specified MDS rank (use with `set max_mds` to shrink cluster)", \ - "mds", "rw", "cli,rest") -COMMAND_WITH_FLAG("mds set_max_mds " \ - "name=maxmds,type=CephInt,range=0", \ - "set max MDS index", "mds", "rw", "cli,rest", FLAG(DEPRECATED)) -COMMAND_WITH_FLAG("mds set " \ - "name=var,type=CephChoices,strings=max_mds|max_file_size" - "|allow_new_snaps|inline_data|allow_multimds|allow_dirfrags " \ - "name=val,type=CephString " \ - "name=confirm,type=CephString,req=false", \ - "set mds parameter to ", "mds", "rw", "cli,rest", FLAG(DEPRECATED)) + "mds", "r") +COMMAND_WITH_FLAG("mds tell " + "name=who,type=CephString " + "name=args,type=CephString,n=N", + "send command to particular mds", "mds", "rw", FLAG(OBSOLETE)) +COMMAND("mds compat show", "show mds compatibility settings", + "mds", "r") +COMMAND_WITH_FLAG("mds stop name=role,type=CephString", "stop mds", + "mds", "rw", FLAG(OBSOLETE)) +COMMAND_WITH_FLAG("mds deactivate name=role,type=CephString", + "clean up specified MDS rank (use with `set max_mds` to shrink cluster)", + "mds", "rw", FLAG(OBSOLETE)) +COMMAND("mds ok-to-stop name=ids,type=CephString,n=N", + "check whether stopping the specified MDS would reduce immediate availability", + "mds", "r") +COMMAND_WITH_FLAG("mds set_max_mds " + "name=maxmds,type=CephInt,range=0", + "set max MDS index", "mds", "rw", FLAG(OBSOLETE)) +COMMAND_WITH_FLAG("mds set " + "name=var,type=CephChoices,strings=max_mds|max_file_size|inline_data|" + "allow_new_snaps|allow_multimds|allow_multimds_snaps|allow_dirfrags " + "name=val,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "set mds parameter to ", "mds", "rw", FLAG(OBSOLETE)) +COMMAND_WITH_FLAG("mds freeze name=role_or_gid,type=CephString" + " name=val,type=CephString", + "freeze MDS yes/no", "mds", "rw", FLAG(HIDDEN)) // arbitrary limit 0-20 below; worth standing on head to make it // relate to actual state definitions? // #include "include/ceph_fs.h" -COMMAND("mds set_state " \ - "name=gid,type=CephInt,range=0 " \ - "name=state,type=CephInt,range=0|20", \ - "set mds state of to ", "mds", "rw", "cli,rest") -COMMAND("mds fail name=who,type=CephString", \ +COMMAND_WITH_FLAG("mds set_state " + "name=gid,type=CephInt,range=0 " + "name=state,type=CephInt,range=0|20", + "set mds state of to ", "mds", "rw", FLAG(HIDDEN)) +COMMAND("mds fail name=role_or_gid,type=CephString", "Mark MDS failed: trigger a failover if a standby is available", - "mds", "rw", "cli,rest") -COMMAND("mds repaired name=rank,type=CephString", \ - "mark a damaged MDS rank as no longer damaged", "mds", "rw", "cli,rest") -COMMAND("mds rm " \ - "name=gid,type=CephInt,range=0", \ - "remove nonactive mds", "mds", "rw", "cli,rest") -COMMAND("mds rmfailed name=who,type=CephString name=confirm,type=CephString,req=false", \ - "remove failed mds", "mds", "rw", "cli,rest") -COMMAND_WITH_FLAG("mds cluster_down", "take MDS cluster down", "mds", "rw", "cli,rest", FLAG(DEPRECATED)) -COMMAND_WITH_FLAG("mds cluster_up", "bring MDS cluster up", "mds", "rw", "cli,rest", FLAG(DEPRECATED)) -COMMAND("mds compat rm_compat " \ - "name=feature,type=CephInt,range=0", \ - "remove compatible feature", "mds", "rw", "cli,rest") -COMMAND("mds compat rm_incompat " \ - "name=feature,type=CephInt,range=0", \ - "remove incompatible feature", "mds", "rw", "cli,rest") -COMMAND_WITH_FLAG("mds add_data_pool " \ - "name=pool,type=CephString", \ - "add data pool ", "mds", "rw", "cli,rest", FLAG(DEPRECATED)) -COMMAND_WITH_FLAG("mds remove_data_pool " \ - "name=pool,type=CephString", \ - "remove data pool ", "mds", "rw", "cli,rest", FLAG(DEPRECATED)) -COMMAND_WITH_FLAG("mds rm_data_pool " \ - "name=pool,type=CephString", \ - "remove data pool ", "mds", "rw", "cli,rest", FLAG(DEPRECATED)) -COMMAND_WITH_FLAG("mds newfs " \ - "name=metadata,type=CephInt,range=0 " \ - "name=data,type=CephInt,range=0 " \ - "name=sure,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "make new filesystem using pools and ", \ - "mds", "rw", "cli,rest", FLAG(OBSOLETE)) -COMMAND("fs new " \ - "name=fs_name,type=CephString " \ - "name=metadata,type=CephString " \ - "name=data,type=CephString " \ - "name=force,type=CephChoices,strings=--force,req=false " \ - "name=sure,type=CephChoices,strings=--allow-dangerous-metadata-overlay,req=false", \ - "make new filesystem using named pools and ", \ - "fs", "rw", "cli,rest") -COMMAND("fs rm " \ - "name=fs_name,type=CephString " \ - "name=sure,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "disable the named filesystem", \ - "fs", "rw", "cli,rest") -COMMAND("fs reset " \ - "name=fs_name,type=CephString " \ - "name=sure,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "disaster recovery only: reset to a single-MDS map", \ - "fs", "rw", "cli,rest") -COMMAND("fs ls ", \ - "list filesystems", \ - "fs", "r", "cli,rest") -COMMAND("fs get name=fs_name,type=CephString", \ - "get info about one filesystem", \ - "fs", "r", "cli,rest") -COMMAND("fs set " \ - "name=fs_name,type=CephString " \ + "mds", "rw") +COMMAND("mds repaired name=role,type=CephString", + "mark a damaged MDS rank as no longer damaged", "mds", "rw") +COMMAND("mds rm " + "name=gid,type=CephInt,range=0", + "remove nonactive mds", "mds", "rw") +COMMAND_WITH_FLAG("mds rmfailed name=role,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "remove failed rank", "mds", "rw", FLAG(HIDDEN)) +COMMAND_WITH_FLAG("mds cluster_down", "take MDS cluster down", "mds", "rw", FLAG(OBSOLETE)) +COMMAND_WITH_FLAG("mds cluster_up", "bring MDS cluster up", "mds", "rw", FLAG(OBSOLETE)) +COMMAND("mds compat rm_compat " + "name=feature,type=CephInt,range=0", + "remove compatible feature", "mds", "rw") +COMMAND("mds compat rm_incompat " + "name=feature,type=CephInt,range=0", + "remove incompatible feature", "mds", "rw") +COMMAND_WITH_FLAG("mds add_data_pool " + "name=pool,type=CephString", + "add data pool ", "mds", "rw", FLAG(OBSOLETE)) +COMMAND_WITH_FLAG("mds rm_data_pool " + "name=pool,type=CephString", + "remove data pool ", "mds", "rw", FLAG(OBSOLETE)) +COMMAND_WITH_FLAG("mds remove_data_pool " + "name=pool,type=CephString", + "remove data pool ", "mds", "rw", FLAG(OBSOLETE)) +COMMAND_WITH_FLAG("mds newfs " + "name=metadata,type=CephInt,range=0 " + "name=data,type=CephInt,range=0 " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "make new filesystem using pools and ", + "mds", "rw", FLAG(OBSOLETE)) +COMMAND("fs new " + "name=fs_name,type=CephString " + "name=metadata,type=CephString " + "name=data,type=CephString " + "name=force,type=CephBool,req=false " + "name=allow_dangerous_metadata_overlay,type=CephBool,req=false", + "make new filesystem using named pools and ", + "fs", "rw") +COMMAND("fs fail " + "name=fs_name,type=CephString ", + "bring the file system down and all of its ranks", + "fs", "rw") +COMMAND("fs rm " + "name=fs_name,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "disable the named filesystem", + "fs", "rw") +COMMAND("fs reset " + "name=fs_name,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "disaster recovery only: reset to a single-MDS map", + "fs", "rw") +COMMAND("fs ls ", + "list filesystems", + "fs", "r") +COMMAND("fs get name=fs_name,type=CephString", + "get info about one filesystem", + "fs", "r") +COMMAND("fs set " + "name=fs_name,type=CephString " "name=var,type=CephChoices,strings=max_mds|max_file_size" - "|allow_new_snaps|inline_data|cluster_down|allow_multimds|allow_dirfrags|balancer" \ - "|standby_count_wanted " \ - "name=val,type=CephString " \ - "name=confirm,type=CephString,req=false", \ - "set mds parameter to ", "mds", "rw", "cli,rest") + "|allow_new_snaps|inline_data|cluster_down|allow_dirfrags|balancer" + "|standby_count_wanted|session_timeout|session_autoclose" + "|allow_standby_replay|down|joinable|min_compat_client " + "name=val,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false " + "name=yes_i_really_really_mean_it,type=CephBool,req=false", + "set fs parameter to ", "mds", "rw") COMMAND("fs flag set name=flag_name,type=CephChoices,strings=enable_multiple " - "name=val,type=CephString " \ - "name=confirm,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "Set a global CephFS flag", \ - "fs", "rw", "cli,rest") -COMMAND("fs add_data_pool name=fs_name,type=CephString " \ - "name=pool,type=CephString", \ - "add data pool ", "mds", "rw", "cli,rest") -COMMAND("fs rm_data_pool name=fs_name,type=CephString " \ - "name=pool,type=CephString", \ - "remove data pool ", "mds", "rw", "cli,rest") -COMMAND_WITH_FLAG("fs set_default name=fs_name,type=CephString", \ - "set the default to the named filesystem", \ - "fs", "rw", "cli,rest", \ + "name=val,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "Set a global CephFS flag", + "fs", "rw") +COMMAND("fs add_data_pool name=fs_name,type=CephString " + "name=pool,type=CephString", + "add data pool ", "mds", "rw") +COMMAND("fs rm_data_pool name=fs_name,type=CephString " + "name=pool,type=CephString", + "remove data pool ", "mds", "rw") +COMMAND_WITH_FLAG("fs set_default name=fs_name,type=CephString", + "set the default to the named filesystem", + "fs", "rw", FLAG(DEPRECATED)) -COMMAND("fs set-default name=fs_name,type=CephString", \ - "set the default to the named filesystem", \ - "fs", "rw", "cli,rest") +COMMAND("fs set-default name=fs_name,type=CephString", + "set the default to the named filesystem", + "fs", "rw") /* * Monmap commands */ -COMMAND("mon dump " \ - "name=epoch,type=CephInt,range=0,req=false", \ - "dump formatted monmap (optionally from epoch)", \ - "mon", "r", "cli,rest") -COMMAND("mon stat", "summarize monitor status", "mon", "r", "cli,rest") -COMMAND("mon getmap " \ - "name=epoch,type=CephInt,range=0,req=false", \ - "get monmap", "mon", "r", "cli,rest") -COMMAND("mon add " \ - "name=name,type=CephString " \ - "name=addr,type=CephIPAddr", \ - "add new monitor named at ", "mon", "rw", "cli,rest") -COMMAND("mon remove " \ - "name=name,type=CephString", \ - "remove monitor named ", "mon", "rw", "cli,rest") -COMMAND("mon rm " \ - "name=name,type=CephString", \ - "remove monitor named ", "mon", "rw", "cli,rest") -COMMAND("mon feature ls " \ - "name=with_value,type=CephChoices,strings=--with-value,req=false", \ - "list available mon map features to be set/unset", \ - "mon", "r", "cli,rest") -COMMAND("mon feature set " \ - "name=feature_name,type=CephString " \ - "name=sure,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "set provided feature on mon map", \ - "mon", "rw", "cli,rest") +COMMAND("mon dump " + "name=epoch,type=CephInt,range=0,req=false", + "dump formatted monmap (optionally from epoch)", + "mon", "r") +COMMAND("mon stat", "summarize monitor status", "mon", "r") +COMMAND("mon getmap " + "name=epoch,type=CephInt,range=0,req=false", + "get monmap", "mon", "r") +COMMAND("mon add " + "name=name,type=CephString " + "name=addr,type=CephIPAddr", + "add new monitor named at ", "mon", "rw") +COMMAND("mon rm " + "name=name,type=CephString", + "remove monitor named ", "mon", "rw") +COMMAND_WITH_FLAG("mon remove " + "name=name,type=CephString", + "remove monitor named ", "mon", "rw", + FLAG(DEPRECATED)) +COMMAND("mon feature ls " + "name=with_value,type=CephChoices,strings=--with-value,req=false", + "list available mon map features to be set/unset", + "mon", "r") +COMMAND("mon feature set " + "name=feature_name,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "set provided feature on mon map", + "mon", "rw") +COMMAND("mon set-rank " + "name=name,type=CephString " + "name=rank,type=CephInt", + "set the rank for the specified mon", + "mon", "rw") +COMMAND("mon set-addrs " + "name=name,type=CephString " + "name=addrs,type=CephString", + "set the addrs (IPs and ports) a specific monitor binds to", + "mon", "rw") +COMMAND("mon set-weight " + "name=name,type=CephString " + "name=weight,type=CephInt,range=0|65535", + "set the weight for the specified mon", + "mon", "rw") +COMMAND("mon enable-msgr2", + "enable the msgr2 protocol on port 3300", + "mon", "rw") /* * OSD commands */ -COMMAND("osd stat", "print summary of OSD map", "osd", "r", "cli,rest") -COMMAND("osd dump " \ +COMMAND("osd stat", "print summary of OSD map", "osd", "r") +COMMAND("osd dump " + "name=epoch,type=CephInt,range=0,req=false", + "print summary of OSD map", "osd", "r") +COMMAND("osd info " + "name=id,type=CephOsdName,req=false", + "print osd's {id} information (instead of all osds from map)", + "osd", "r") +COMMAND("osd tree " + "name=epoch,type=CephInt,range=0,req=false " + "name=states,type=CephChoices,strings=up|down|in|out|destroyed,n=N,req=false", + "print OSD tree", "osd", "r") +COMMAND("osd tree-from " + "name=epoch,type=CephInt,range=0,req=false " + "name=bucket,type=CephString " + "name=states,type=CephChoices,strings=up|down|in|out|destroyed,n=N,req=false", + "print OSD tree in bucket", "osd", "r") +COMMAND("osd ls " + "name=epoch,type=CephInt,range=0,req=false", + "show all OSD ids", "osd", "r") +COMMAND("osd getmap " "name=epoch,type=CephInt,range=0,req=false", - "print summary of OSD map", "osd", "r", "cli,rest") -COMMAND("osd tree " \ - "name=epoch,type=CephInt,range=0,req=false " \ - "name=states,type=CephChoices,strings=up|down|in|out,n=N,req=false", \ - "print OSD tree", "osd", "r", "cli,rest") -COMMAND("osd ls " \ - "name=epoch,type=CephInt,range=0,req=false", \ - "show all OSD ids", "osd", "r", "cli,rest") -COMMAND("osd getmap " \ - "name=epoch,type=CephInt,range=0,req=false", \ - "get OSD map", "osd", "r", "cli,rest") -COMMAND("osd getcrushmap " \ - "name=epoch,type=CephInt,range=0,req=false", \ - "get CRUSH map", "osd", "r", "cli,rest") -COMMAND("osd getmaxosd", "show largest OSD id", "osd", "r", "cli,rest") -COMMAND("osd ls-tree " \ + "get OSD map", "osd", "r") +COMMAND("osd getcrushmap " + "name=epoch,type=CephInt,range=0,req=false", + "get CRUSH map", "osd", "r") +COMMAND("osd getmaxosd", "show largest OSD id", "osd", "r") +COMMAND("osd ls-tree " "name=epoch,type=CephInt,range=0,req=false " - "name=name,type=CephString,req=true", \ - "show OSD ids under bucket in the CRUSH map", \ - "osd", "r", "cli,rest") -COMMAND("osd find " \ - "name=id,type=CephOsdName", \ - "find osd in the CRUSH map and show its location", \ - "osd", "r", "cli,rest") -COMMAND("osd metadata " \ - "name=id,type=CephOsdName,req=false", \ - "fetch metadata for osd {id} (default all)", \ - "osd", "r", "cli,rest") + "name=name,type=CephString,req=true", + "show OSD ids under bucket in the CRUSH map", + "osd", "r") +COMMAND("osd find " + "name=id,type=CephOsdName", + "find osd in the CRUSH map and show its location", + "osd", "r") +COMMAND("osd metadata " + "name=id,type=CephOsdName,req=false", + "fetch metadata for osd {id} (default all)", + "osd", "r") COMMAND("osd count-metadata name=property,type=CephString", "count OSDs by metadata field property", - "osd", "r", "cli,rest") -COMMAND("osd versions", \ + "osd", "r") +COMMAND("osd versions", "check running versions of OSDs", - "osd", "r", "cli,rest") -COMMAND("osd map " \ - "name=pool,type=CephPoolname " \ - "name=object,type=CephObjectname " \ - "name=nspace,type=CephString,req=false", \ - "find pg for in with [namespace]", "osd", "r", "cli,rest") -COMMAND("osd lspools " \ - "name=auid,type=CephInt,req=false", \ - "list pools", "osd", "r", "cli,rest") -COMMAND("osd blacklist ls", "show blacklisted clients", "osd", "r", "cli,rest") -COMMAND("osd blacklist clear", "clear all blacklisted clients", "osd", "rw", - "cli,rest") -COMMAND("osd crush rule list", "list crush rules", "osd", "r", "cli,rest") -COMMAND("osd crush rule ls", "list crush rules", "osd", "r", "cli,rest") -COMMAND("osd crush rule dump " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.],req=false", \ - "dump crush rule (default all)", \ - "osd", "r", "cli,rest") -COMMAND("osd crush dump", \ - "dump crush map", \ - "osd", "r", "cli,rest") -COMMAND("osd setcrushmap name=prior_version,type=CephInt,req=false", \ - "set crush map from input file", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush set name=prior_version,type=CephInt,req=false", \ - "set crush map from input file", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush add-bucket " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=type,type=CephString", \ - "add no-parent (probably root) crush bucket of type ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush rename-bucket " \ - "name=srcname,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=dstname,type=CephString,goodchars=[A-Za-z0-9-_.]", \ - "rename bucket to ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush set " \ - "name=id,type=CephOsdName " \ - "name=weight,type=CephFloat,range=0.0 " \ - "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", \ - "update crushmap position and weight for to with location ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush add " \ - "name=id,type=CephOsdName " \ - "name=weight,type=CephFloat,range=0.0 " \ - "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", \ - "add or update crushmap position and weight for with and location ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush set-device-class " \ - "name=class,type=CephString " \ - "name=ids,type=CephString,n=N", \ - "set the of the osd(s) [...]," \ - "or use to set all.", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush create-or-move " \ - "name=id,type=CephOsdName " \ - "name=weight,type=CephFloat,range=0.0 " \ - "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", \ - "create entry or move existing entry for at/to location ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush move " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", \ - "move existing entry for to location ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush swap-bucket " \ - "name=source,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=dest,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=force,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "swap existing bucket contents from (orphan) bucket and ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush link " \ - "name=name,type=CephString " \ - "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", \ - "link existing entry for under location ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush rm " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=ancestor,type=CephString,req=false,goodchars=[A-Za-z0-9-_.]", \ + "osd", "r") +COMMAND("osd numa-status", + "show NUMA status of OSDs", + "osd", "r") +COMMAND("osd map " + "name=pool,type=CephPoolname " + "name=object,type=CephObjectname " + "name=nspace,type=CephString,req=false", + "find pg for in with [namespace]", "osd", "r") +COMMAND_WITH_FLAG("osd lspools", + "list pools", "osd", "r", FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("osd crush rule list", "list crush rules", "osd", "r", + FLAG(DEPRECATED)) +COMMAND("osd crush rule ls", "list crush rules", "osd", "r") +COMMAND("osd crush rule ls-by-class " + "name=class,type=CephString,goodchars=[A-Za-z0-9-_.]", + "list all crush rules that reference the same ", + "osd", "r") +COMMAND("osd crush rule dump " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.],req=false", + "dump crush rule (default all)", + "osd", "r") +COMMAND("osd crush dump", + "dump crush map", + "osd", "r") +COMMAND("osd setcrushmap name=prior_version,type=CephInt,req=false", + "set crush map from input file", + "osd", "rw") +COMMAND("osd crush set name=prior_version,type=CephInt,req=false", + "set crush map from input file", + "osd", "rw") +COMMAND("osd crush add-bucket " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=type,type=CephString " + "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=],req=false", + "add no-parent (probably root) crush bucket of type " + "to location ", + "osd", "rw") +COMMAND("osd crush rename-bucket " + "name=srcname,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=dstname,type=CephString,goodchars=[A-Za-z0-9-_.]", + "rename bucket to ", + "osd", "rw") +COMMAND("osd crush set " + "name=id,type=CephOsdName " + "name=weight,type=CephFloat,range=0.0 " + "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", + "update crushmap position and weight for to with location ", + "osd", "rw") +COMMAND("osd crush add " + "name=id,type=CephOsdName " + "name=weight,type=CephFloat,range=0.0 " + "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", + "add or update crushmap position and weight for with and location ", + "osd", "rw") +COMMAND("osd crush set-all-straw-buckets-to-straw2", + "convert all CRUSH current straw buckets to use the straw2 algorithm", + "osd", "rw") +COMMAND("osd crush class create " + "name=class,type=CephString,goodchars=[A-Za-z0-9-_]", + "create crush device class ", + "osd", "rw") +COMMAND("osd crush class rm " + "name=class,type=CephString,goodchars=[A-Za-z0-9-_]", + "remove crush device class ", + "osd", "rw") +COMMAND("osd crush set-device-class " + "name=class,type=CephString " + "name=ids,type=CephString,n=N", + "set the of the osd(s) [...]," + "or use to set all.", + "osd", "rw") +COMMAND("osd crush rm-device-class " + "name=ids,type=CephString,n=N", + "remove class of the osd(s) [...]," + "or use to remove all.", + "osd", "rw") +COMMAND("osd crush class rename " + "name=srcname,type=CephString,goodchars=[A-Za-z0-9-_] " + "name=dstname,type=CephString,goodchars=[A-Za-z0-9-_]", + "rename crush device class to ", + "osd", "rw") +COMMAND("osd crush create-or-move " + "name=id,type=CephOsdName " + "name=weight,type=CephFloat,range=0.0 " + "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", + "create entry or move existing entry for at/to location ", + "osd", "rw") +COMMAND("osd crush move " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", + "move existing entry for to location ", + "osd", "rw") +COMMAND("osd crush swap-bucket " + "name=source,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=dest,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "swap existing bucket contents from (orphan) bucket and ", + "osd", "rw") +COMMAND("osd crush link " + "name=name,type=CephString " + "name=args,type=CephString,n=N,goodchars=[A-Za-z0-9-_.=]", + "link existing entry for under location ", + "osd", "rw") +COMMAND("osd crush rm " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=ancestor,type=CephString,req=false,goodchars=[A-Za-z0-9-_.]", "remove from crush map (everywhere, or just at )",\ - "osd", "rw", "cli,rest") -COMMAND("osd crush remove " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=ancestor,type=CephString,req=false,goodchars=[A-Za-z0-9-_.]", \ - "remove from crush map (everywhere, or just at )", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush unlink " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=ancestor,type=CephString,req=false,goodchars=[A-Za-z0-9-_.]", \ - "unlink from crush map (everywhere, or just at )", \ - "osd", "rw", "cli,rest") + "osd", "rw") +COMMAND_WITH_FLAG("osd crush remove " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=ancestor,type=CephString,req=false,goodchars=[A-Za-z0-9-_.]", + "remove from crush map (everywhere, or just at )", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND("osd crush unlink " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=ancestor,type=CephString,req=false,goodchars=[A-Za-z0-9-_.]", + "unlink from crush map (everywhere, or just at )", + "osd", "rw") COMMAND("osd crush reweight-all", "recalculate the weights for the tree to ensure they sum correctly", - "osd", "rw", "cli,rest") -COMMAND("osd crush reweight " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=weight,type=CephFloat,range=0.0", \ - "change 's weight to in crush map", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush reweight-subtree " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=weight,type=CephFloat,range=0.0", \ - "change all leaf items beneath to in crush map", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush tunables " \ - "name=profile,type=CephChoices,strings=legacy|argonaut|bobtail|firefly|hammer|jewel|optimal|default", \ - "set crush tunables values to ", "osd", "rw", "cli,rest") -COMMAND("osd crush set-tunable " \ - "name=tunable,type=CephChoices,strings=straw_calc_version " \ + "osd", "rw") +COMMAND("osd crush reweight " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=weight,type=CephFloat,range=0.0", + "change 's weight to in crush map", + "osd", "rw") +COMMAND("osd crush reweight-subtree " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=weight,type=CephFloat,range=0.0", + "change all leaf items beneath to in crush map", + "osd", "rw") +COMMAND("osd crush tunables " + "name=profile,type=CephChoices,strings=legacy|argonaut|bobtail|firefly|hammer|jewel|optimal|default", + "set crush tunables values to ", "osd", "rw") +COMMAND("osd crush set-tunable " + "name=tunable,type=CephChoices,strings=straw_calc_version " "name=value,type=CephInt", "set crush tunable to ", - "osd", "rw", "cli,rest") -COMMAND("osd crush get-tunable " \ + "osd", "rw") +COMMAND("osd crush get-tunable " "name=tunable,type=CephChoices,strings=straw_calc_version", "get crush tunable ", - "osd", "rw", "cli,rest") -COMMAND("osd crush show-tunables", \ - "show current crush tunables", "osd", "r", "cli,rest") -COMMAND("osd crush rule create-simple " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=root,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=type,type=CephString,goodchars=[A-Za-z0-9-_.] " \ + "osd", "r") +COMMAND("osd crush show-tunables", + "show current crush tunables", "osd", "r") +COMMAND("osd crush rule create-simple " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=root,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=type,type=CephString,goodchars=[A-Za-z0-9-_.] " "name=mode,type=CephChoices,strings=firstn|indep,req=false", - "create crush rule to start from , replicate across buckets of type , using a choose mode of (default firstn; indep best for erasure pools)", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush rule create-replicated " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=root,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=type,type=CephString,goodchars=[A-Za-z0-9-_.] " \ + "create crush rule to start from , replicate across buckets of type , using a choose mode of (default firstn; indep best for erasure pools)", + "osd", "rw") +COMMAND("osd crush rule create-replicated " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=root,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=type,type=CephString,goodchars=[A-Za-z0-9-_.] " "name=class,type=CephString,goodchars=[A-Za-z0-9-_.],req=false", - "create crush rule for replicated pool to start from , replicate across buckets of type , using a choose mode of (default firstn; indep best for erasure pools)", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush rule create-erasure " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=profile,type=CephString,req=false,goodchars=[A-Za-z0-9-_.=]", \ - "create crush rule for erasure coded pool created with (default default)", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush rule rm " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] ", \ - "remove crush rule ", "osd", "rw", "cli,rest") -COMMAND("osd crush tree", + "create crush rule for replicated pool to start from , replicate across buckets of type , use devices of type (ssd or hdd)", + "osd", "rw") +COMMAND("osd crush rule create-erasure " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=profile,type=CephString,req=false,goodchars=[A-Za-z0-9-_.=]", + "create crush rule for erasure coded pool created with (default default)", + "osd", "rw") +COMMAND("osd crush rule rm " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] ", + "remove crush rule ", "osd", "rw") +COMMAND("osd crush rule rename " + "name=srcname,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=dstname,type=CephString,goodchars=[A-Za-z0-9-_.]", + "rename crush rule to ", + "osd", "rw") +COMMAND("osd crush tree " + "name=shadow,type=CephChoices,strings=--show-shadow,req=false", "dump crush buckets and items in a tree view", - "osd", "r", "cli,rest") -COMMAND("osd crush class create " \ - "name=class,type=CephString,goodchars=[A-Za-z0-9-_]", \ - "create crush device class ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush class rm " \ - "name=class,type=CephString,goodchars=[A-Za-z0-9-_]", \ - "remove crush device class ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush class rename " \ - "name=srcname,type=CephString,goodchars=[A-Za-z0-9-_] " \ - "name=dstname,type=CephString,goodchars=[A-Za-z0-9-_]", \ - "rename crush device class to ", \ - "osd", "rw", "cli,rest") -COMMAND("osd crush class ls", \ - "list all crush device classes", \ - "osd", "r", "cli,rest") -COMMAND("osd crush class ls-osd " \ - "name=class,type=CephString,goodchars=[A-Za-z0-9-_]", \ - "list all osds belonging to the specific ", \ - "osd", "r", "cli,rest") -COMMAND("osd setmaxosd " \ - "name=newmax,type=CephInt,range=0", \ - "set new maximum osd value", "osd", "rw", "cli,rest") -COMMAND("osd set-full-ratio " \ - "name=ratio,type=CephFloat,range=0.0|1.0", \ + "osd", "r") +COMMAND("osd crush ls name=node,type=CephString,goodchars=[A-Za-z0-9-_.]", + "list items beneath a node in the CRUSH tree", + "osd", "r") +COMMAND("osd crush class ls", + "list all crush device classes", + "osd", "r") +COMMAND("osd crush class ls-osd " + "name=class,type=CephString,goodchars=[A-Za-z0-9-_]", + "list all osds belonging to the specific ", + "osd", "r") +COMMAND("osd crush get-device-class " + "name=ids,type=CephString,n=N", + "get classes of specified osd(s) [...]", + "osd", "r") +COMMAND("osd crush weight-set ls", + "list crush weight sets", + "osd", "r") +COMMAND("osd crush weight-set dump", + "dump crush weight sets", + "osd", "r") +COMMAND("osd crush weight-set create-compat", + "create a default backward-compatible weight-set", + "osd", "rw") +COMMAND("osd crush weight-set create " + "name=pool,type=CephPoolname "\ + "name=mode,type=CephChoices,strings=flat|positional", + "create a weight-set for a given pool", + "osd", "rw") +COMMAND("osd crush weight-set rm name=pool,type=CephPoolname", + "remove the weight-set for a given pool", + "osd", "rw") +COMMAND("osd crush weight-set rm-compat", + "remove the backward-compatible weight-set", + "osd", "rw") +COMMAND("osd crush weight-set reweight " + "name=pool,type=CephPoolname " + "name=item,type=CephString " + "name=weight,type=CephFloat,range=0.0,n=N", + "set weight for an item (bucket or osd) in a pool's weight-set", + "osd", "rw") +COMMAND("osd crush weight-set reweight-compat " + "name=item,type=CephString " + "name=weight,type=CephFloat,range=0.0,n=N", + "set weight for an item (bucket or osd) in the backward-compatible weight-set", + "osd", "rw") +COMMAND("osd setmaxosd " + "name=newmax,type=CephInt,range=0", + "set new maximum osd value", "osd", "rw") +COMMAND("osd set-full-ratio " + "name=ratio,type=CephFloat,range=0.0|1.0", "set usage ratio at which OSDs are marked full", - "osd", "rw", "cli,rest") -COMMAND("osd set-backfillfull-ratio " \ - "name=ratio,type=CephFloat,range=0.0|1.0", \ + "osd", "rw") +COMMAND("osd set-backfillfull-ratio " + "name=ratio,type=CephFloat,range=0.0|1.0", "set usage ratio at which OSDs are marked too full to backfill", - "osd", "rw", "cli,rest") -COMMAND("osd set-nearfull-ratio " \ - "name=ratio,type=CephFloat,range=0.0|1.0", \ + "osd", "rw") +COMMAND("osd set-nearfull-ratio " + "name=ratio,type=CephFloat,range=0.0|1.0", "set usage ratio at which OSDs are marked near-full", - "osd", "rw", "cli,rest") -COMMAND("osd set-require-min-compat-client " \ - "name=version,type=CephString " \ - "name=sure,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ + "osd", "rw") +COMMAND("osd get-require-min-compat-client", + "get the minimum client version we will maintain compatibility with", + "osd", "r") +COMMAND("osd set-require-min-compat-client " + "name=version,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false", "set the minimum client version we will maintain compatibility with", - "osd", "rw", "cli,rest") -COMMAND("osd pause", "pause osd", "osd", "rw", "cli,rest") -COMMAND("osd unpause", "unpause osd", "osd", "rw", "cli,rest") -COMMAND("osd erasure-code-profile set " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " \ - "name=profile,type=CephString,n=N,req=false", \ - "create erasure code profile with [ ...] pairs. Add a --force at the end to override an existing profile (VERY DANGEROUS)", \ - "osd", "rw", "cli,rest") -COMMAND("osd erasure-code-profile get " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.]", \ - "get erasure code profile ", \ - "osd", "r", "cli,rest") -COMMAND("osd erasure-code-profile rm " \ - "name=name,type=CephString,goodchars=[A-Za-z0-9-_.]", \ - "remove erasure code profile ", \ - "osd", "rw", "cli,rest") -COMMAND("osd erasure-code-profile ls", \ - "list all erasure code profiles", \ - "osd", "r", "cli,rest") -COMMAND("osd set " \ - "name=key,type=CephChoices,strings=full|pause|noup|nodown|noout|noin|nobackfill|norebalance|norecover|noscrub|nodeep-scrub|notieragent|sortbitwise|require_jewel_osds|require_kraken_osds", \ - "set ", "osd", "rw", "cli,rest") -COMMAND("osd unset " \ - "name=key,type=CephChoices,strings=full|pause|noup|nodown|noout|noin|nobackfill|norebalance|norecover|noscrub|nodeep-scrub|notieragent", \ - "unset ", "osd", "rw", "cli,rest") + "osd", "rw") +COMMAND("osd pause", "pause osd", "osd", "rw") +COMMAND("osd unpause", "unpause osd", "osd", "rw") +COMMAND("osd erasure-code-profile set " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=profile,type=CephString,n=N,req=false " + "name=force,type=CephBool,req=false", + "create erasure code profile with [ ...] pairs. Add a --force at the end to override an existing profile (VERY DANGEROUS)", + "osd", "rw") +COMMAND("osd erasure-code-profile get " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.]", + "get erasure code profile ", + "osd", "r") +COMMAND("osd erasure-code-profile rm " + "name=name,type=CephString,goodchars=[A-Za-z0-9-_.]", + "remove erasure code profile ", + "osd", "rw") +COMMAND("osd erasure-code-profile ls", + "list all erasure code profiles", + "osd", "r") +COMMAND("osd set " + "name=key,type=CephChoices,strings=full|pause|noup|nodown|" + "noout|noin|nobackfill|norebalance|norecover|noscrub|nodeep-scrub|" + "notieragent|nosnaptrim|pglog_hardlimit " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "set ", "osd", "rw") +COMMAND("osd unset " + "name=key,type=CephChoices,strings=full|pause|noup|nodown|"\ + "noout|noin|nobackfill|norebalance|norecover|noscrub|nodeep-scrub|" + "notieragent|nosnaptrim", + "unset ", "osd", "rw") COMMAND("osd require-osd-release "\ - "name=release,type=CephChoices,strings=luminous", + "name=release,type=CephChoices,strings=luminous|mimic|nautilus|octopus " + "name=yes_i_really_mean_it,type=CephBool,req=false", "set the minimum allowed OSD release to participate in the cluster", - "osd", "rw", "cli,rest") -COMMAND("osd cluster_snap", "take cluster snapshot (disabled)", \ - "osd", "r", "") -COMMAND("osd down " \ - "type=CephString,name=ids,n=N", \ - "set osd(s) [...] down, " \ - "or use to set all osds down", \ - "osd", "rw", "cli,rest") -COMMAND("osd out " \ - "name=ids,type=CephString,n=N", \ - "set osd(s) [...] out, " \ - "or use to set all osds out", \ - "osd", "rw", "cli,rest") -COMMAND("osd in " \ - "name=ids,type=CephString,n=N", \ + "osd", "rw") +COMMAND("osd down " + "name=ids,type=CephString,n=N " + "name=definitely_dead,type=CephBool,req=false", + "set osd(s) [...] down, " + "or use to set all osds down", + "osd", "rw") +COMMAND("osd stop " + "type=CephString,name=ids,n=N", + "stop the corresponding osd daemons and mark them as down", + "osd", "rw") +COMMAND("osd out " + "name=ids,type=CephString,n=N", + "set osd(s) [...] out, " + "or use to set all osds out", + "osd", "rw") +COMMAND("osd in " + "name=ids,type=CephString,n=N", "set osd(s) [...] in, " - "can use to automatically set all previously out osds in", \ - "osd", "rw", "cli,rest") -COMMAND("osd rm " \ - "name=ids,type=CephString,n=N", \ + "can use to automatically set all previously out osds in", + "osd", "rw") +COMMAND_WITH_FLAG("osd rm " + "name=ids,type=CephString,n=N", "remove osd(s) [...], " - "or use to remove all osds", \ - "osd", "rw", "cli,rest") -COMMAND("osd add-noup " \ - "name=ids,type=CephString,n=N", \ - "mark osd(s) [...] as noup, " \ - "or use to mark all osds as noup", \ - "osd", "rw", "cli,rest") -COMMAND("osd add-nodown " \ - "name=ids,type=CephString,n=N", \ - "mark osd(s) [...] as nodown, " \ - "or use to mark all osds as nodown", \ - "osd", "rw", "cli,rest") -COMMAND("osd add-noin " \ - "name=ids,type=CephString,n=N", \ - "mark osd(s) [...] as noin, " \ - "or use to mark all osds as noin", \ - "osd", "rw", "cli,rest") -COMMAND("osd add-noout " \ - "name=ids,type=CephString,n=N", \ - "mark osd(s) [...] as noout, " \ - "or use to mark all osds as noout", \ - "osd", "rw", "cli,rest") -COMMAND("osd rm-noup " \ - "name=ids,type=CephString,n=N", \ - "allow osd(s) [...] to be marked up " \ - "(if they are currently marked as noup), " \ - "can use to automatically filter out all noup osds", \ - "osd", "rw", "cli,rest") -COMMAND("osd rm-nodown " \ - "name=ids,type=CephString,n=N", \ - "allow osd(s) [...] to be marked down " \ - "(if they are currently marked as nodown), " \ - "can use to automatically filter out all nodown osds", \ - "osd", "rw", "cli,rest") -COMMAND("osd rm-noin " \ - "name=ids,type=CephString,n=N", \ - "allow osd(s) [...] to be marked in " \ - "(if they are currently marked as noin), " \ - "can use to automatically filter out all noin osds", \ - "osd", "rw", "cli,rest") -COMMAND("osd rm-noout " \ - "name=ids,type=CephString,n=N", \ - "allow osd(s) [...] to be marked out " \ - "(if they are currently marked as noout), " \ - "can use to automatically filter out all noout osds", \ - "osd", "rw", "cli,rest") -COMMAND("osd reweight " \ - "name=id,type=CephOsdName " \ - "type=CephFloat,name=weight,range=0.0|1.0", \ - "reweight osd to 0.0 < < 1.0", "osd", "rw", "cli,rest") -COMMAND("osd reweightn " \ + "or use to remove all osds", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("osd add-noup " + "name=ids,type=CephString,n=N", + "mark osd(s) [...] as noup, " + "or use to mark all osds as noup", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("osd add-nodown " + "name=ids,type=CephString,n=N", + "mark osd(s) [...] as nodown, " + "or use to mark all osds as nodown", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("osd add-noin " + "name=ids,type=CephString,n=N", + "mark osd(s) [...] as noin, " + "or use to mark all osds as noin", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("osd add-noout " + "name=ids,type=CephString,n=N", + "mark osd(s) [...] as noout, " + "or use to mark all osds as noout", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("osd rm-noup " + "name=ids,type=CephString,n=N", + "allow osd(s) [...] to be marked up " + "(if they are currently marked as noup), " + "can use to automatically filter out all noup osds", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("osd rm-nodown " + "name=ids,type=CephString,n=N", + "allow osd(s) [...] to be marked down " + "(if they are currently marked as nodown), " + "can use to automatically filter out all nodown osds", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("osd rm-noin " + "name=ids,type=CephString,n=N", + "allow osd(s) [...] to be marked in " + "(if they are currently marked as noin), " + "can use to automatically filter out all noin osds", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("osd rm-noout " + "name=ids,type=CephString,n=N", + "allow osd(s) [...] to be marked out " + "(if they are currently marked as noout), " + "can use to automatically filter out all noout osds", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND("osd set-group " + "name=flags,type=CephString " + "name=who,type=CephString,n=N", + "set for batch osds or crush nodes, " + " must be a comma-separated subset of {noup,nodown,noin,noout}", + "osd", "rw") +COMMAND("osd unset-group " + "name=flags,type=CephString " + "name=who,type=CephString,n=N", + "unset for batch osds or crush nodes, " + " must be a comma-separated subset of {noup,nodown,noin,noout}", + "osd", "rw") +COMMAND("osd reweight " + "name=id,type=CephOsdName " + "type=CephFloat,name=weight,range=0.0|1.0", + "reweight osd to 0.0 < < 1.0", "osd", "rw") +COMMAND("osd reweightn " "name=weights,type=CephString", "reweight osds with {: ,...})", - "osd", "rw", "cli,rest") -COMMAND("osd pg-temp " \ - "name=pgid,type=CephPgid " \ - "name=id,type=CephOsdName,n=N,req=false", \ - "set pg_temp mapping pgid:[ [...]] (developers only)", \ - "osd", "rw", "cli,rest") -COMMAND("osd pg-upmap " \ - "name=pgid,type=CephPgid " \ - "name=id,type=CephOsdName,n=N", \ - "set pg_upmap mapping :[ [...]] (developers only)", \ - "osd", "rw", "cli,rest") -COMMAND("osd rm-pg-upmap " \ - "name=pgid,type=CephPgid", \ - "clear pg_upmap mapping for (developers only)", \ - "osd", "rw", "cli,rest") -COMMAND("osd pg-upmap-items " \ - "name=pgid,type=CephPgid " \ - "name=id,type=CephOsdName,n=N", \ - "set pg_upmap_items mapping :{ to , [...]} (developers only)", \ - "osd", "rw", "cli,rest") -COMMAND("osd rm-pg-upmap-items " \ - "name=pgid,type=CephPgid", \ - "clear pg_upmap_items mapping for (developers only)", \ - "osd", "rw", "cli,rest") -COMMAND("osd primary-temp " \ - "name=pgid,type=CephPgid " \ - "name=id,type=CephOsdName", \ - "set primary_temp mapping pgid:|-1 (developers only)", \ - "osd", "rw", "cli,rest") -COMMAND("osd primary-affinity " \ - "name=id,type=CephOsdName " \ - "type=CephFloat,name=weight,range=0.0|1.0", \ - "adjust osd primary-affinity from 0.0 <= <= 1.0", \ - "osd", "rw", "cli,rest") -COMMAND("osd destroy " \ - "name=id,type=CephOsdName " \ - "name=sure,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "mark osd as being destroyed. Keeps the ID intact (allowing reuse), " \ + "osd", "rw") +COMMAND("osd force-create-pg " + "name=pgid,type=CephPgid "\ + "name=yes_i_really_mean_it,type=CephBool,req=false", + "force creation of pg ", + "osd", "rw") +COMMAND("osd pg-temp " + "name=pgid,type=CephPgid " + "name=id,type=CephOsdName,n=N,req=false", + "set pg_temp mapping pgid:[ [...]] (developers only)", + "osd", "rw") +COMMAND("osd pg-upmap " + "name=pgid,type=CephPgid " + "name=id,type=CephOsdName,n=N", + "set pg_upmap mapping :[ [...]] (developers only)", + "osd", "rw") +COMMAND("osd rm-pg-upmap " + "name=pgid,type=CephPgid", + "clear pg_upmap mapping for (developers only)", + "osd", "rw") +COMMAND("osd pg-upmap-items " + "name=pgid,type=CephPgid " + "name=id,type=CephOsdName,n=N", + "set pg_upmap_items mapping :{ to , [...]} (developers only)", + "osd", "rw") +COMMAND("osd rm-pg-upmap-items " + "name=pgid,type=CephPgid", + "clear pg_upmap_items mapping for (developers only)", + "osd", "rw") +COMMAND("osd primary-temp " + "name=pgid,type=CephPgid " + "name=id,type=CephOsdName", + "set primary_temp mapping pgid:|-1 (developers only)", + "osd", "rw") +COMMAND("osd primary-affinity " + "name=id,type=CephOsdName " + "type=CephFloat,name=weight,range=0.0|1.0", + "adjust osd primary-affinity from 0.0 <= <= 1.0", + "osd", "rw") +COMMAND_WITH_FLAG("osd destroy-actual " + "name=id,type=CephOsdName " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "mark osd as being destroyed. Keeps the ID intact (allowing reuse), " "but removes cephx keys, config-key data and lockbox keys, "\ - "rendering data permanently unreadable.", \ - "osd", "rw", "cli,rest") -COMMAND("osd purge " \ - "name=id,type=CephOsdName " \ - "name=sure,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "purge all osd data from the monitors. Combines `osd destroy`, " \ - "`osd rm`, and `osd crush rm`.", \ - "osd", "rw", "cli,rest") -COMMAND("osd lost " \ - "name=id,type=CephOsdName " \ - "name=sure,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "mark osd as permanently lost. THIS DESTROYS DATA IF NO MORE REPLICAS EXIST, BE CAREFUL", \ - "osd", "rw", "cli,rest") -COMMAND_WITH_FLAG("osd create " \ - "name=uuid,type=CephUUID,req=false " \ - "name=id,type=CephOsdName,req=false", \ - "create new osd (with optional UUID and ID)", "osd", "rw", "cli,rest", + "rendering data permanently unreadable.", + "osd", "rw", FLAG(HIDDEN)) +COMMAND("osd purge-new " + "name=id,type=CephOsdName " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "purge all traces of an OSD that was partially created but never " + "started", + "osd", "rw") +COMMAND_WITH_FLAG("osd purge-actual " + "name=id,type=CephOsdName " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "purge all osd data from the monitors. Combines `osd destroy`, " + "`osd rm`, and `osd crush rm`.", + "osd", "rw", FLAG(HIDDEN)) +COMMAND("osd lost " + "name=id,type=CephOsdName " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "mark osd as permanently lost. THIS DESTROYS DATA IF NO MORE REPLICAS EXIST, BE CAREFUL", + "osd", "rw") +COMMAND_WITH_FLAG("osd create " + "name=uuid,type=CephUUID,req=false " + "name=id,type=CephOsdName,req=false", + "create new osd (with optional UUID and ID)", "osd", "rw", FLAG(DEPRECATED)) -COMMAND("osd new " \ - "name=uuid,type=CephUUID,req=true " \ - "name=id,type=CephOsdName,req=false", \ - "Create a new OSD. If supplied, the `id` to be replaced needs to " \ - "exist and have been previously destroyed. " \ - "Reads secrets from JSON file via `-i ` (see man page).", \ - "osd", "rw", "cli,rest") -COMMAND("osd blacklist " \ - "name=blacklistop,type=CephChoices,strings=add|rm " \ - "name=addr,type=CephEntityAddr " \ - "name=expire,type=CephFloat,range=0.0,req=false", \ - "add (optionally until seconds from now) or remove from blacklist", \ - "osd", "rw", "cli,rest") -COMMAND("osd pool mksnap " \ - "name=pool,type=CephPoolname " \ - "name=snap,type=CephString", \ - "make snapshot in ", "osd", "rw", "cli,rest") -COMMAND("osd pool rmsnap " \ - "name=pool,type=CephPoolname " \ - "name=snap,type=CephString", \ - "remove snapshot from ", "osd", "rw", "cli,rest") -COMMAND("osd pool ls " \ - "name=detail,type=CephChoices,strings=detail,req=false", \ - "list pools", "osd", "r", "cli,rest") -COMMAND("osd pool create " \ - "name=pool,type=CephPoolname " \ - "name=pg_num,type=CephInt,range=0 " \ - "name=pgp_num,type=CephInt,range=0,req=false " \ - "name=pool_type,type=CephChoices,strings=replicated|erasure,req=false " \ - "name=erasure_code_profile,type=CephString,req=false,goodchars=[A-Za-z0-9-_.] " \ - "name=rule,type=CephString,req=false " \ - "name=expected_num_objects,type=CephInt,req=false", \ - "create pool", "osd", "rw", "cli,rest") -COMMAND("osd pool delete " \ - "name=pool,type=CephPoolname " \ - "name=pool2,type=CephPoolname,req=false " \ - "name=sure,type=CephString,req=false", \ - "delete pool", \ - "osd", "rw", "cli,rest") -COMMAND("osd pool rm " \ - "name=pool,type=CephPoolname " \ - "name=pool2,type=CephPoolname,req=false " \ - "name=sure,type=CephString,req=false", \ - "remove pool", \ - "osd", "rw", "cli,rest") -COMMAND("osd pool rename " \ - "name=srcpool,type=CephPoolname " \ - "name=destpool,type=CephPoolname", \ - "rename to ", "osd", "rw", "cli,rest") -COMMAND("osd pool get " \ - "name=pool,type=CephPoolname " \ - "name=var,type=CephChoices,strings=size|min_size|crash_replay_interval|pg_num|pgp_num|crush_rule|hashpspool|nodelete|nopgchange|nosizechange|write_fadvise_dontneed|noscrub|nodeep-scrub|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|use_gmt_hitset|auid|target_max_objects|target_max_bytes|cache_target_dirty_ratio|cache_target_dirty_high_ratio|cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|erasure_code_profile|min_read_recency_for_promote|all|min_write_recency_for_promote|fast_read|hit_set_grade_decay_rate|hit_set_search_last_n|scrub_min_interval|scrub_max_interval|deep_scrub_interval|recovery_priority|recovery_op_priority|scrub_priority|compression_mode|compression_algorithm|compression_required_ratio|compression_max_blob_size|compression_min_blob_size|csum_type|csum_min_block|csum_max_block", \ - "get pool parameter ", "osd", "r", "cli,rest") -COMMAND("osd pool set " \ - "name=pool,type=CephPoolname " \ - "name=var,type=CephChoices,strings=size|min_size|crash_replay_interval|pg_num|pgp_num|crush_rule|hashpspool|nodelete|nopgchange|nosizechange|write_fadvise_dontneed|noscrub|nodeep-scrub|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|use_gmt_hitset|target_max_bytes|target_max_objects|cache_target_dirty_ratio|cache_target_dirty_high_ratio|cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|auid|min_read_recency_for_promote|min_write_recency_for_promote|fast_read|hit_set_grade_decay_rate|hit_set_search_last_n|scrub_min_interval|scrub_max_interval|deep_scrub_interval|recovery_priority|recovery_op_priority|scrub_priority|compression_mode|compression_algorithm|compression_required_ratio|compression_max_blob_size|compression_min_blob_size|csum_type|csum_min_block|csum_max_block|allow_ec_overwrites " \ - "name=val,type=CephString " \ - "name=force,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "set pool parameter to ", "osd", "rw", "cli,rest") +COMMAND("osd new " + "name=uuid,type=CephUUID,req=true " + "name=id,type=CephOsdName,req=false", + "Create a new OSD. If supplied, the `id` to be replaced needs to " + "exist and have been previously destroyed. " + "Reads secrets from JSON file via `-i ` (see man page).", + "osd", "rw") +COMMAND("osd blacklist " + "name=blacklistop,type=CephChoices,strings=add|rm " + "name=addr,type=CephEntityAddr " + "name=expire,type=CephFloat,range=0.0,req=false", + "add (optionally until seconds from now) or remove from blacklist", + "osd", "rw") +COMMAND("osd blacklist ls", "show blacklisted clients", "osd", "r") +COMMAND("osd blacklist clear", "clear all blacklisted clients", "osd", "rw") +COMMAND("osd pool mksnap " + "name=pool,type=CephPoolname " + "name=snap,type=CephString", + "make snapshot in ", "osd", "rw") +COMMAND("osd pool rmsnap " + "name=pool,type=CephPoolname " + "name=snap,type=CephString", + "remove snapshot from ", "osd", "rw") +COMMAND("osd pool ls " + "name=detail,type=CephChoices,strings=detail,req=false", + "list pools", "osd", "r") +COMMAND("osd pool create " + "name=pool,type=CephPoolname " + "name=pg_num,type=CephInt,range=0,req=false " + "name=pgp_num,type=CephInt,range=0,req=false " + "name=pool_type,type=CephChoices,strings=replicated|erasure,req=false " + "name=erasure_code_profile,type=CephString,req=false,goodchars=[A-Za-z0-9-_.] " + "name=rule,type=CephString,req=false " + "name=expected_num_objects,type=CephInt,range=0,req=false " + "name=size,type=CephInt,range=0,req=false " + "name=pg_num_min,type=CephInt,range=0,req=false " + "name=autoscale_mode,type=CephChoices,strings=on|off|warn,req=false " + "name=target_size_bytes,type=CephInt,range=0,req=false " + "name=target_size_ratio,type=CephFloat,range=0|1,req=false",\ + "create pool", "osd", "rw") +COMMAND_WITH_FLAG("osd pool delete " + "name=pool,type=CephPoolname " + "name=pool2,type=CephPoolname,req=false " + "name=yes_i_really_really_mean_it,type=CephBool,req=false " + "name=yes_i_really_really_mean_it_not_faking,type=CephBool,req=false ", + "delete pool", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND("osd pool rm " + "name=pool,type=CephPoolname " + "name=pool2,type=CephPoolname,req=false " + "name=yes_i_really_really_mean_it,type=CephBool,req=false " + "name=yes_i_really_really_mean_it_not_faking,type=CephBool,req=false ", + "remove pool", + "osd", "rw") +COMMAND("osd pool rename " + "name=srcpool,type=CephPoolname " + "name=destpool,type=CephPoolname", + "rename to ", "osd", "rw") +COMMAND("osd pool get " + "name=pool,type=CephPoolname " + "name=var,type=CephChoices,strings=size|min_size|pg_num|pgp_num|crush_rule|hashpspool|nodelete|nopgchange|nosizechange|write_fadvise_dontneed|noscrub|nodeep-scrub|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|use_gmt_hitset|target_max_objects|target_max_bytes|cache_target_dirty_ratio|cache_target_dirty_high_ratio|cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|erasure_code_profile|min_read_recency_for_promote|all|min_write_recency_for_promote|fast_read|hit_set_grade_decay_rate|hit_set_search_last_n|scrub_min_interval|scrub_max_interval|deep_scrub_interval|recovery_priority|recovery_op_priority|scrub_priority|compression_mode|compression_algorithm|compression_required_ratio|compression_max_blob_size|compression_min_blob_size|csum_type|csum_min_block|csum_max_block|allow_ec_overwrites|fingerprint_algorithm|pg_autoscale_mode|pg_autoscale_bias|pg_num_min|target_size_bytes|target_size_ratio", + "get pool parameter ", "osd", "r") +COMMAND("osd pool set " + "name=pool,type=CephPoolname " + "name=var,type=CephChoices,strings=size|min_size|pg_num|pgp_num|pgp_num_actual|crush_rule|hashpspool|nodelete|nopgchange|nosizechange|write_fadvise_dontneed|noscrub|nodeep-scrub|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|use_gmt_hitset|target_max_bytes|target_max_objects|cache_target_dirty_ratio|cache_target_dirty_high_ratio|cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|min_read_recency_for_promote|min_write_recency_for_promote|fast_read|hit_set_grade_decay_rate|hit_set_search_last_n|scrub_min_interval|scrub_max_interval|deep_scrub_interval|recovery_priority|recovery_op_priority|scrub_priority|compression_mode|compression_algorithm|compression_required_ratio|compression_max_blob_size|compression_min_blob_size|csum_type|csum_min_block|csum_max_block|allow_ec_overwrites|fingerprint_algorithm|pg_autoscale_mode|pg_autoscale_bias|pg_num_min|target_size_bytes|target_size_ratio " + "name=val,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "set pool parameter to ", "osd", "rw") // 'val' is a CephString because it can include a unit. Perhaps // there should be a Python type for validation/conversion of strings // with units. -COMMAND("osd pool set-quota " \ - "name=pool,type=CephPoolname " \ - "name=field,type=CephChoices,strings=max_objects|max_bytes " \ +COMMAND("osd pool set-quota " + "name=pool,type=CephPoolname " + "name=field,type=CephChoices,strings=max_objects|max_bytes " "name=val,type=CephString", - "set object or byte limit on pool", "osd", "rw", "cli,rest") -COMMAND("osd pool get-quota " \ + "set object or byte limit on pool", "osd", "rw") +COMMAND("osd pool get-quota " "name=pool,type=CephPoolname ", "obtain object or byte limits for pool", - "osd", "r", "cli,rest") + "osd", "r") +COMMAND("osd pool application enable " + "name=pool,type=CephPoolname " + "name=app,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "enable use of an application [cephfs,rbd,rgw] on pool ", + "osd", "rw") +COMMAND("osd pool application disable " + "name=pool,type=CephPoolname " + "name=app,type=CephString " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "disables use of an application on pool ", + "osd", "rw") +COMMAND("osd pool application set " + "name=pool,type=CephPoolname " + "name=app,type=CephString " + "name=key,type=CephString,goodchars=[A-Za-z0-9-_.] " + "name=value,type=CephString,goodchars=[A-Za-z0-9-_.=]", + "sets application metadata key to on pool ", + "osd", "rw") +COMMAND("osd pool application rm " + "name=pool,type=CephPoolname " + "name=app,type=CephString " + "name=key,type=CephString", + "removes application metadata key on pool ", + "osd", "rw") +COMMAND("osd pool application get " + "name=pool,type=CephPoolname,req=fasle " + "name=app,type=CephString,req=false " + "name=key,type=CephString,req=false", + "get value of key of application on pool ", + "osd", "r") COMMAND("osd utilization", "get basic pg distribution stats", - "osd", "r", "cli,rest") + "osd", "r") // tiering -COMMAND("osd tier add " \ - "name=pool,type=CephPoolname " \ - "name=tierpool,type=CephPoolname " \ +COMMAND("osd tier add " + "name=pool,type=CephPoolname " + "name=tierpool,type=CephPoolname " "name=force_nonempty,type=CephChoices,strings=--force-nonempty,req=false", - "add the tier (the second one) to base pool (the first one)", \ - "osd", "rw", "cli,rest") -COMMAND("osd tier remove " \ - "name=pool,type=CephPoolname " \ + "add the tier (the second one) to base pool (the first one)", + "osd", "rw") +COMMAND("osd tier rm " + "name=pool,type=CephPoolname " "name=tierpool,type=CephPoolname", - "remove the tier (the second one) from base pool (the first one)", \ - "osd", "rw", "cli,rest") -COMMAND("osd tier rm " \ - "name=pool,type=CephPoolname " \ + "remove the tier (the second one) from base pool (the first one)", + "osd", "rw") +COMMAND_WITH_FLAG("osd tier remove " + "name=pool,type=CephPoolname " "name=tierpool,type=CephPoolname", - "remove the tier (the second one) from base pool (the first one)", \ - "osd", "rw", "cli,rest") -COMMAND("osd tier cache-mode " \ - "name=pool,type=CephPoolname " \ - "name=mode,type=CephChoices,strings=none|writeback|forward|readonly|readforward|proxy|readproxy " \ - "name=sure,type=CephChoices,strings=--yes-i-really-mean-it,req=false", \ - "specify the caching mode for cache tier ", "osd", "rw", "cli,rest") -COMMAND("osd tier set-overlay " \ - "name=pool,type=CephPoolname " \ - "name=overlaypool,type=CephPoolname", \ - "set the overlay pool for base pool to be ", "osd", "rw", "cli,rest") -COMMAND("osd tier remove-overlay " \ - "name=pool,type=CephPoolname ", \ - "remove the overlay pool for base pool ", "osd", "rw", "cli,rest") -COMMAND("osd tier rm-overlay " \ - "name=pool,type=CephPoolname ", \ - "remove the overlay pool for base pool ", "osd", "rw", "cli,rest") + "remove the tier (the second one) from base pool (the first one)", + "osd", "rw", + FLAG(DEPRECATED)) +COMMAND("osd tier cache-mode " + "name=pool,type=CephPoolname " + "name=mode,type=CephChoices,strings=writeback|readproxy|readonly|none " + "name=yes_i_really_mean_it,type=CephBool,req=false", + "specify the caching mode for cache tier ", "osd", "rw") +COMMAND("osd tier set-overlay " + "name=pool,type=CephPoolname " + "name=overlaypool,type=CephPoolname", + "set the overlay pool for base pool to be ", "osd", "rw") +COMMAND("osd tier rm-overlay " + "name=pool,type=CephPoolname ", + "remove the overlay pool for base pool ", "osd", "rw") +COMMAND_WITH_FLAG("osd tier remove-overlay " + "name=pool,type=CephPoolname ", + "remove the overlay pool for base pool ", "osd", "rw", + FLAG(DEPRECATED)) -COMMAND("osd tier add-cache " \ - "name=pool,type=CephPoolname " \ - "name=tierpool,type=CephPoolname " \ - "name=size,type=CephInt,range=0", \ - "add a cache (the second one) of size to existing pool (the first one)", \ - "osd", "rw", "cli,rest") +COMMAND("osd tier add-cache " + "name=pool,type=CephPoolname " + "name=tierpool,type=CephPoolname " + "name=size,type=CephInt,range=0", + "add a cache (the second one) of size to existing pool (the first one)", + "osd", "rw") /* * mon/ConfigKeyService.cc */ -COMMAND("config-key get " \ - "name=key,type=CephString", \ - "get ", "config-key", "r", "cli,rest") -COMMAND("config-key put " \ - "name=key,type=CephString " \ - "name=val,type=CephString,req=false", \ - "put , value ", "config-key", "rw", "cli,rest") -COMMAND("config-key del " \ - "name=key,type=CephString", \ - "delete ", "config-key", "rw", "cli,rest") -COMMAND("config-key rm " \ - "name=key,type=CephString", \ - "rm ", "config-key", "rw", "cli,rest") -COMMAND("config-key exists " \ - "name=key,type=CephString", \ - "check for 's existence", "config-key", "r", "cli,rest") -COMMAND("config-key list ", "list keys", "config-key", "r", "cli,rest") -COMMAND("config-key dump", "dump keys and values", "config-key", "r", "cli,rest") +COMMAND("config-key get " + "name=key,type=CephString", + "get ", "config-key", "r") +COMMAND("config-key set " + "name=key,type=CephString " + "name=val,type=CephString,req=false", + "set to value ", "config-key", "rw") +COMMAND_WITH_FLAG("config-key put " + "name=key,type=CephString " + "name=val,type=CephString,req=false", + "put , value ", "config-key", "rw", + FLAG(DEPRECATED)) +COMMAND_WITH_FLAG("config-key del " + "name=key,type=CephString", + "delete ", "config-key", "rw", + FLAG(DEPRECATED)) +COMMAND("config-key rm " + "name=key,type=CephString", + "rm ", "config-key", "rw") +COMMAND("config-key exists " + "name=key,type=CephString", + "check for 's existence", "config-key", "r") +COMMAND_WITH_FLAG("config-key list ", "list keys", "config-key", "r", + FLAG(DEPRECATED)) +COMMAND("config-key ls ", "list keys", "config-key", "r") +COMMAND("config-key dump " + "name=key,type=CephString,req=false", "dump keys and values (with optional prefix)", "config-key", "r") /* * mon/MgrMonitor.cc */ -COMMAND("mgr dump " \ - "name=epoch,type=CephInt,range=0,req=false", \ - "dump the latest MgrMap", \ - "mgr", "r", "cli,rest") -COMMAND("mgr fail name=who,type=CephString", \ - "treat the named manager daemon as failed", "mgr", "rw", "cli,rest") +COMMAND("mgr dump " + "name=epoch,type=CephInt,range=0,req=false", + "dump the latest MgrMap", + "mgr", "r") +COMMAND("mgr fail name=who,type=CephString,req=false", + "treat the named manager daemon as failed", "mgr", "rw") COMMAND("mgr module ls", - "list active mgr modules", "mgr", "r", "cli,rest") -COMMAND("mgr module enable " \ - "name=module,type=CephString " \ + "list active mgr modules", "mgr", "r") +COMMAND("mgr services", + "list service endpoints provided by mgr modules", + "mgr", "r") +COMMAND("mgr module enable " + "name=module,type=CephString " "name=force,type=CephChoices,strings=--force,req=false", - "enable mgr module", "mgr", "rw", "cli,rest") -COMMAND("mgr module disable " \ + "enable mgr module", "mgr", "rw") +COMMAND("mgr module disable " "name=module,type=CephString", - "disable mgr module", "mgr", "rw", "cli,rest") + "disable mgr module", "mgr", "rw") +COMMAND("mgr metadata name=who,type=CephString,req=false", + "dump metadata for all daemons or a specific daemon", + "mgr", "r") +COMMAND("mgr count-metadata name=property,type=CephString", + "count ceph-mgr daemons by metadata field property", + "mgr", "r") +COMMAND("mgr versions", + "check running versions of ceph-mgr daemons", + "mgr", "r") + +// ConfigMonitor +COMMAND("config set" + " name=who,type=CephString" + " name=name,type=CephString" + " name=value,type=CephString" + " name=force,type=CephBool,req=false", + "Set a configuration option for one or more entities", + "config", "rw") +COMMAND("config rm" + " name=who,type=CephString" + " name=name,type=CephString", + "Clear a configuration option for one or more entities", + "config", "rw") +COMMAND("config get " + "name=who,type=CephString " + "name=key,type=CephString,req=False", + "Show configuration option(s) for an entity", + "config", "r") +COMMAND("config dump", + "Show all configuration option(s)", + "mon", "r") +COMMAND("config help " + "name=key,type=CephString", + "Describe a configuration option", + "config", "r") +COMMAND("config ls", + "List available configuration options", + "config", "r") +COMMAND("config assimilate-conf", + "Assimilate options from a conf, and return a new, minimal conf file", + "config", "rw") +COMMAND("config log name=num,type=CephInt,req=False", + "Show recent history of config changes", + "config", "r") +COMMAND("config reset " + "name=num,type=CephInt,range=0", + "Revert configuration to a historical version specified by ", + "config", "rw") +COMMAND("config generate-minimal-conf", + "Generate a minimal ceph.conf file", + "config", "r") + + + + +// these are tell commands that were implemented as CLI commands in +// the broken pre-octopus way that we want to allow to work when a +// monitor has upgraded to octopus+ but the monmap min_mon_release is +// still < octopus. we exclude things that weren't well supported +// before and that aren't implemented by the octopus mon anymore. +// +// the command set below matches the kludge in Monitor::handle_command +// that shunts these off to the asok machinery. + +COMMAND_WITH_FLAG("injectargs " + "name=injected_args,type=CephString,n=N", + "inject config arguments into monitor", "mon", "rw", + FLAG(TELL)) +COMMAND_WITH_FLAG("smart name=devid,type=CephString,req=false", + "Query health metrics for underlying device", + "mon", "rw", + FLAG(TELL)) +COMMAND_WITH_FLAG("mon_status", + "report status of monitors", + "mon", "r", + FLAG(TELL)) +COMMAND_WITH_FLAG("heap " + "name=heapcmd,type=CephChoices,strings=dump|start_profiler|stop_profiler|release|stats " + "name=value,type=CephString,req=false", + "show heap usage info (available only if compiled with tcmalloc)", + "mon", "rw", + FLAG(TELL)) +COMMAND_WITH_FLAG("sync_force " + "name=validate,type=CephChoices,strings=--yes-i-really-mean-it,req=false", + "force sync of and clear monitor store", + "mon", "rw", + FLAG(TELL)) +COMMAND_WITH_FLAG("add_bootstrap_peer_hint " + "name=addr,type=CephIPAddr", + "add peer address as potential bootstrap " + "peer for cluster bringup", + "mon", "rw", + FLAG(TELL)) +COMMAND_WITH_FLAG("add_bootstrap_peer_hintv " + "name=addrv,type=CephString", + "add peer address vector as potential bootstrap " + "peer for cluster bringup", + "mon", "rw", + FLAG(TELL)) +COMMAND_WITH_FLAG("quorum enter ", + "force monitor back into quorum", + "mon", "rw", + FLAG(TELL)) +COMMAND_WITH_FLAG("quorum exit", + "force monitor out of the quorum", + "mon", "rw", + FLAG(TELL)) +COMMAND_WITH_FLAG("ops", + "show the ops currently in flight", + "mon", "r", + FLAG(TELL)) +COMMAND_WITH_FLAG("sessions", + "list existing sessions", + "mon", "r", + FLAG(TELL)) +COMMAND_WITH_FLAG("dump_historic_ops", + "dump_historic_ops", + "mon", "r", + FLAG(TELL))