:orphan: ================================== ceph -- ceph administration tool ================================== .. program:: ceph Synopsis ======== | **ceph** **auth** [ *add* \| *caps* \| *del* \| *export* \| *get* \| *get-key* \| *get-or-create* \| *get-or-create-key* \| *import* \| *list* \| *print-key* \| *print_key* ] ... | **ceph** **compact** | **ceph** **config-key** [ *del* | *exists* | *get* | *list* | *put* ] ... | **ceph** **daemon** ** \| ** ** ... | **ceph** **daemonperf** ** \| ** [ *interval* [ *count* ] ] | **ceph** **df** *{detail}* | **ceph** **fs** [ *ls* \| *new* \| *reset* \| *rm* ] ... | **ceph** **fsid** | **ceph** **health** *{detail}* | **ceph** **heap** [ *dump* \| *start_profiler* \| *stop_profiler* \| *release* \| *stats* ] ... | **ceph** **injectargs** ** [ **... ] | **ceph** **log** ** [ **... ] | **ceph** **mds** [ *compat* \| *deactivate* \| *fail* \| *rm* \| *rmfailed* \| *set_state* \| *stat* \| *tell* ] ... | **ceph** **mon** [ *add* \| *dump* \| *getmap* \| *remove* \| *stat* ] ... | **ceph** **mon_status** | **ceph** **osd** [ *blacklist* \| *blocked-by* \| *create* \| *deep-scrub* \| *df* \| *down* \| *dump* \| *erasure-code-profile* \| *find* \| *getcrushmap* \| *getmap* \| *getmaxosd* \| *in* \| *lspools* \| *map* \| *metadata* \| *out* \| *pause* \| *perf* \| *pg-temp* \| *primary-affinity* \| *primary-temp* \| *repair* \| *reweight* \| *reweight-by-pg* \| *rm* \| *scrub* \| *set* \| *setcrushmap* \| *setmaxosd* \| *stat* \| *thrash* \| *tree* \| *unpause* \| *unset* ] ... | **ceph** **osd** **crush** [ *add* \| *add-bucket* \| *create-or-move* \| *dump* \| *get-tunable* \| *link* \| *move* \| *remove* \| *rename-bucket* \| *reweight* \| *reweight-all* \| *reweight-subtree* \| *rm* \| *rule* \| *set* \| *set-tunable* \| *show-tunables* \| *tunables* \| *unlink* ] ... | **ceph** **osd** **pool** [ *create* \| *delete* \| *get* \| *get-quota* \| *ls* \| *mksnap* \| *rename* \| *rmsnap* \| *set* \| *set-quota* \| *stats* ] ... | **ceph** **osd** **tier** [ *add* \| *add-cache* \| *cache-mode* \| *remove* \| *remove-overlay* \| *set-overlay* ] ... | **ceph** **pg** [ *debug* \| *deep-scrub* \| *dump* \| *dump_json* \| *dump_pools_json* \| *dump_stuck* \| *force_create_pg* \| *getmap* \| *ls* \| *ls-by-osd* \| *ls-by-pool* \| *ls-by-primary* \| *map* \| *repair* \| *scrub* \| *set_full_ratio* \| *set_nearfull_ratio* \| *stat* ] ... | **ceph** **quorum** [ *enter* \| *exit* ] | **ceph** **quorum_status** | **ceph** **report** { ** [ *...* ] } | **ceph** **scrub** | **ceph** **status** | **ceph** **sync** **force** {--yes-i-really-mean-it} {--i-know-what-i-am-doing} | **ceph** **tell** * [...]* | **ceph** **version** Description =========== :program:`ceph` is a control utility which is used for manual deployment and maintenance of a Ceph cluster. It provides a diverse set of commands that allows deployment of monitors, OSDs, placement groups, MDS and overall maintenance, administration of the cluster. Commands ======== auth ---- Manage authentication keys. It is used for adding, removing, exporting or updating of authentication keys for a particular entity such as a monitor or OSD. It uses some additional subcommands. Subcommand ``add`` adds authentication info for a particular entity from input file, or random key if no input is given and/or any caps specified in the command. Usage:: ceph auth add { [...]} Subcommand ``caps`` updates caps for **name** from caps specified in the command. Usage:: ceph auth caps [...] Subcommand ``del`` deletes all caps for ``name``. Usage:: ceph auth del Subcommand ``export`` writes keyring for requested entity, or master keyring if none given. Usage:: ceph auth export {} Subcommand ``get`` writes keyring file with requested key. Usage:: ceph auth get Subcommand ``get-key`` displays requested key. Usage:: ceph auth get-key Subcommand ``get-or-create`` adds authentication info for a particular entity from input file, or random key if no input given and/or any caps specified in the command. Usage:: ceph auth get-or-create { [...]} Subcommand ``get-or-create-key`` gets or adds key for ``name`` from system/caps pairs specified in the command. If key already exists, any given caps must match the existing caps for that key. Usage:: ceph auth get-or-create-key { [...]} Subcommand ``import`` reads keyring from input file. Usage:: ceph auth import Subcommand ``list`` lists authentication state. Usage:: ceph auth list Subcommand ``print-key`` displays requested key. Usage:: ceph auth print-key Subcommand ``print_key`` displays requested key. Usage:: ceph auth print_key compact ------- Causes compaction of monitor's leveldb storage. Usage:: ceph compact config-key ---------- Manage configuration key. It uses some additional subcommands. Subcommand ``del`` deletes configuration key. Usage:: ceph config-key del Subcommand ``exists`` checks for configuration keys existence. Usage:: ceph config-key exists Subcommand ``get`` gets the configuration key. Usage:: ceph config-key get Subcommand ``list`` lists configuration keys. Usage:: ceph config-key list Subcommand ``put`` puts configuration key and values. Usage:: ceph config-key put {} daemon ------ Submit admin-socket commands. Usage:: ceph daemon {daemon_name|socket_path} {command} ... Example:: ceph daemon osd.0 help daemonperf ---------- Watch performance counters from a Ceph daemon. Usage:: ceph daemonperf {daemon_name|socket_path} [{interval} [{count}]] df -- Show cluster's free space status. Usage:: ceph df {detail} fs -- Manage cephfs filesystems. It uses some additional subcommands. Subcommand ``ls`` to list filesystems Usage:: ceph fs ls Subcommand ``new`` to make a new filesystem using named pools and Usage:: ceph fs new Subcommand ``reset`` is used for disaster recovery only: reset to a single-MDS map Usage:: ceph fs reset {--yes-i-really-mean-it} Subcommand ``rm`` to disable the named filesystem Usage:: ceph fs rm {--yes-i-really-mean-it} fsid ---- Show cluster's FSID/UUID. Usage:: ceph fsid health ------ Show cluster's health. Usage:: ceph health {detail} heap ---- Show heap usage info (available only if compiled with tcmalloc) Usage:: ceph heap dump|start_profiler|stop_profiler|release|stats injectargs ---------- Inject configuration arguments into monitor. Usage:: ceph injectargs [...] log --- Log supplied text to the monitor log. Usage:: ceph log [...] mds --- Manage metadata server configuration and administration. It uses some additional subcommands. Subcommand ``compat`` manages compatible features. It uses some additional subcommands. Subcommand ``rm_compat`` removes compatible feature. Usage:: ceph mds compat rm_compat Subcommand ``rm_incompat`` removes incompatible feature. Usage:: ceph mds compat rm_incompat Subcommand ``show`` shows mds compatibility settings. Usage:: ceph mds compat show Subcommand ``deactivate`` stops mds. Usage:: ceph mds deactivate Subcommand ``fail`` forces mds to status fail. Usage:: ceph mds fail Subcommand ``rm`` removes inactive mds. Usage:: ceph mds rm (type.id)> Subcommand ``rmfailed`` removes failed mds. Usage:: ceph mds rmfailed Subcommand ``set_state`` sets mds state of to . Usage:: ceph mds set_state Subcommand ``stat`` shows MDS status. Usage:: ceph mds stat Subcommand ``tell`` sends command to particular mds. Usage:: ceph mds tell [...] mon --- Manage monitor configuration and administration. It uses some additional subcommands. Subcommand ``add`` adds new monitor named at . Usage:: ceph mon add Subcommand ``dump`` dumps formatted monmap (optionally from epoch) Usage:: ceph mon dump {} Subcommand ``getmap`` gets monmap. Usage:: ceph mon getmap {} Subcommand ``remove`` removes monitor named . Usage:: ceph mon remove Subcommand ``stat`` summarizes monitor status. Usage:: ceph mon stat mon_status ---------- Reports status of monitors. Usage:: ceph mon_status osd --- Manage OSD configuration and administration. It uses some additional subcommands. Subcommand ``blacklist`` manage blacklisted clients. It uses some additional subcommands. Subcommand ``add`` add to blacklist (optionally until seconds from now) Usage:: ceph osd blacklist add {} Subcommand ``ls`` show blacklisted clients Usage:: ceph osd blacklist ls Subcommand ``rm`` remove from blacklist Usage:: ceph osd blacklist rm Subcommand ``blocked-by`` prints a histogram of which OSDs are blocking their peers Usage:: ceph osd blocked-by Subcommand ``create`` creates new osd (with optional UUID and ID). Usage:: ceph osd create {} {} Subcommand ``crush`` is used for CRUSH management. It uses some additional subcommands. Subcommand ``add`` adds or updates crushmap position and weight for with and location . Usage:: ceph osd crush add [...] Subcommand ``add-bucket`` adds no-parent (probably root) crush bucket of type . Usage:: ceph osd crush add-bucket Subcommand ``create-or-move`` creates entry or moves existing entry for at/to location . Usage:: ceph osd crush create-or-move [...] Subcommand ``dump`` dumps crush map. Usage:: ceph osd crush dump Subcommand ``get-tunable`` get crush tunable straw_calc_version Usage:: ceph osd crush get-tunable straw_calc_version Subcommand ``link`` links existing entry for under location . Usage:: ceph osd crush link [...] Subcommand ``move`` moves existing entry for to location . Usage:: ceph osd crush move [...] Subcommand ``remove`` removes from crush map (everywhere, or just at ). Usage:: ceph osd crush remove {} Subcommand ``rename-bucket`` renames buchket to Usage:: ceph osd crush rename-bucket Subcommand ``reweight`` change 's weight to in crush map. Usage:: ceph osd crush reweight Subcommand ``reweight-all`` recalculate the weights for the tree to ensure they sum correctly Usage:: ceph osd crush reweight-all Subcommand ``reweight-subtree`` changes all leaf items beneath to in crush map Usage:: ceph osd crush reweight-subtree Subcommand ``rm`` removes from crush map (everywhere, or just at ). Usage:: ceph osd crush rm {} Subcommand ``rule`` is used for creating crush rules. It uses some additional subcommands. Subcommand ``create-erasure`` creates crush rule for erasure coded pool created with (default default). Usage:: ceph osd crush rule create-erasure {} Subcommand ``create-simple`` creates crush rule to start from , replicate across buckets of type , using a choose mode of (default firstn; indep best for erasure pools). Usage:: ceph osd crush rule create-simple {firstn|indep} Subcommand ``dump`` dumps crush rule (default all). Usage:: ceph osd crush rule dump {} Subcommand ``list`` lists crush rules. Usage:: ceph osd crush rule list Subcommand ``ls`` lists crush rules. Usage:: ceph osd crush rule ls Subcommand ``rm`` removes crush rule . Usage:: ceph osd crush rule rm Subcommand ``set`` used alone, sets crush map from input file. Usage:: ceph osd crush set Subcommand ``set`` with osdname/osd.id update crushmap position and weight for to with location . Usage:: ceph osd crush set [...] Subcommand ``set-tunable`` set crush tunable to . The only tunable that can be set is straw_calc_version. Usage:: ceph osd crush set-tunable straw_calc_version Subcommand ``show-tunables`` shows current crush tunables. Usage:: ceph osd crush show-tunables Subcommand ``tree`` shows the crush buckets and items in a tree view. Usage:: ceph osd crush tree Subcommand ``tunables`` sets crush tunables values to . Usage:: ceph osd crush tunables legacy|argonaut|bobtail|firefly|hammer|optimal|default Subcommand ``unlink`` unlinks from crush map (everywhere, or just at ). Usage:: ceph osd crush unlink {} Subcommand ``df`` shows OSD utilization Usage:: ceph osd df {plain|tree} Subcommand ``deep-scrub`` initiates deep scrub on specified osd. Usage:: ceph osd deep-scrub Subcommand ``down`` sets osd(s) [...] down. Usage:: ceph osd down [...] Subcommand ``dump`` prints summary of OSD map. Usage:: ceph osd dump {} Subcommand ``erasure-code-profile`` is used for managing the erasure code profiles. It uses some additional subcommands. Subcommand ``get`` gets erasure code profile . Usage:: ceph osd erasure-code-profile get Subcommand ``ls`` lists all erasure code profiles. Usage:: ceph osd erasure-code-profile ls Subcommand ``rm`` removes erasure code profile . Usage:: ceph osd erasure-code-profile rm Subcommand ``set`` creates erasure code profile with [ ...] pairs. Add a --force at the end to override an existing profile (IT IS RISKY). Usage:: ceph osd erasure-code-profile set { [...]} Subcommand ``find`` find osd in the CRUSH map and shows its location. Usage:: ceph osd find Subcommand ``getcrushmap`` gets CRUSH map. Usage:: ceph osd getcrushmap {} Subcommand ``getmap`` gets OSD map. Usage:: ceph osd getmap {} Subcommand ``getmaxosd`` shows largest OSD id. Usage:: ceph osd getmaxosd Subcommand ``in`` sets osd(s) [...] in. Usage:: ceph osd in [...] Subcommand ``lost`` marks osd as permanently lost. THIS DESTROYS DATA IF NO MORE REPLICAS EXIST, BE CAREFUL. Usage:: ceph osd lost {--yes-i-really-mean-it} Subcommand ``ls`` shows all OSD ids. Usage:: ceph osd ls {} Subcommand ``lspools`` lists pools. Usage:: ceph osd lspools {} Subcommand ``map`` finds pg for in . Usage:: ceph osd map Subcommand ``metadata`` fetches metadata for osd . Usage:: ceph osd metadata {int[0-]} (default all) Subcommand ``out`` sets osd(s) [...] out. Usage:: ceph osd out [...] Subcommand ``pause`` pauses osd. Usage:: ceph osd pause Subcommand ``perf`` prints dump of OSD perf summary stats. Usage:: ceph osd perf Subcommand ``pg-temp`` set pg_temp mapping pgid:[ [...]] (developers only). Usage:: ceph osd pg-temp { [...]} Subcommand ``pool`` is used for managing data pools. It uses some additional subcommands. Subcommand ``create`` creates pool. Usage:: ceph osd pool create {} {replicated|erasure} {} {} {} Subcommand ``delete`` deletes pool. Usage:: ceph osd pool delete {} {--yes-i-really-really-mean-it} Subcommand ``get`` gets pool parameter . Usage:: ceph osd pool get size|min_size|crash_replay_interval|pg_num| pgp_num|crush_ruleset|auid|write_fadvise_dontneed Only for tiered pools:: ceph osd pool get hit_set_type|hit_set_period|hit_set_count|hit_set_fpp| target_max_objects|target_max_bytes|cache_target_dirty_ratio|cache_target_dirty_high_ratio| cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age| min_read_recency_for_promote|hit_set_grade_decay_rate|hit_set_search_last_n Only for erasure coded pools:: ceph osd pool get erasure_code_profile Use ``all`` to get all pool parameters that apply to the pool's type:: ceph osd pool get all Subcommand ``get-quota`` obtains object or byte limits for pool. Usage:: ceph osd pool get-quota Subcommand ``ls`` list pools Usage:: ceph osd pool ls {detail} Subcommand ``mksnap`` makes snapshot in . Usage:: ceph osd pool mksnap Subcommand ``rename`` renames to . Usage:: ceph osd pool rename Subcommand ``rmsnap`` removes snapshot from . Usage:: ceph osd pool rmsnap Subcommand ``set`` sets pool parameter to . Usage:: ceph osd pool set size|min_size|crash_replay_interval|pg_num| pgp_num|crush_ruleset|hashpspool|nodelete|nopgchange|nosizechange| hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|debug_fake_ec_pool| target_max_bytes|target_max_objects|cache_target_dirty_ratio| cache_target_dirty_high_ratio| cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|auid| min_read_recency_for_promote|write_fadvise_dontneed|hit_set_grade_decay_rate| hit_set_search_last_n {--yes-i-really-mean-it} Subcommand ``set-quota`` sets object or byte limit on pool. Usage:: ceph osd pool set-quota max_objects|max_bytes Subcommand ``stats`` obtain stats from all pools, or from specified pool. Usage:: ceph osd pool stats {} Subcommand ``primary-affinity`` adjust osd primary-affinity from 0.0 <= <= 1.0 Usage:: ceph osd primary-affinity Subcommand ``primary-temp`` sets primary_temp mapping pgid:|-1 (developers only). Usage:: ceph osd primary-temp Subcommand ``repair`` initiates repair on a specified osd. Usage:: ceph osd repair Subcommand ``reweight`` reweights osd to 0.0 < < 1.0. Usage:: osd reweight Subcommand ``reweight-by-pg`` reweight OSDs by PG distribution [overload-percentage-for-consideration, default 120]. Usage:: ceph osd reweight-by-pg {} { [} {--no-increasing} Subcommand ``rm`` removes osd(s) [...] in the cluster. Usage:: ceph osd rm [...] Subcommand ``scrub`` initiates scrub on specified osd. Usage:: ceph osd scrub Subcommand ``set`` sets . Usage:: ceph osd set full|pause|noup|nodown|noout|noin|nobackfill| norebalance|norecover|noscrub|nodeep-scrub|notieragent Subcommand ``setcrushmap`` sets crush map from input file. Usage:: ceph osd setcrushmap Subcommand ``setmaxosd`` sets new maximum osd value. Usage:: ceph osd setmaxosd Subcommand ``stat`` prints summary of OSD map. Usage:: ceph osd stat Subcommand ``tier`` is used for managing tiers. It uses some additional subcommands. Subcommand ``add`` adds the tier (the second one) to base pool (the first one). Usage:: ceph osd tier add {--force-nonempty} Subcommand ``add-cache`` adds a cache (the second one) of size to existing pool (the first one). Usage:: ceph osd tier add-cache Subcommand ``cache-mode`` specifies the caching mode for cache tier . Usage:: ceph osd tier cache-mode none|writeback|forward|readonly| readforward|readproxy Subcommand ``remove`` removes the tier (the second one) from base pool (the first one). Usage:: ceph osd tier remove Subcommand ``remove-overlay`` removes the overlay pool for base pool . Usage:: ceph osd tier remove-overlay Subcommand ``set-overlay`` set the overlay pool for base pool to be . Usage:: ceph osd tier set-overlay Subcommand ``tree`` prints OSD tree. Usage:: ceph osd tree {} Subcommand ``unpause`` unpauses osd. Usage:: ceph osd unpause Subcommand ``unset`` unsets . Usage:: ceph osd unset full|pause|noup|nodown|noout|noin|nobackfill| norebalance|norecover|noscrub|nodeep-scrub|notieragent pg -- It is used for managing the placement groups in OSDs. It uses some additional subcommands. Subcommand ``debug`` shows debug info about pgs. Usage:: ceph pg debug unfound_objects_exist|degraded_pgs_exist Subcommand ``deep-scrub`` starts deep-scrub on . Usage:: ceph pg deep-scrub Subcommand ``dump`` shows human-readable versions of pg map (only 'all' valid with plain). Usage:: ceph pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief} [{all|summary|sum|delta|pools|osds|pgs|pgs_brief...]} Subcommand ``dump_json`` shows human-readable version of pg map in json only. Usage:: ceph pg dump_json {all|summary|sum|delta|pools|osds|pgs|pgs_brief} [{all|summary|sum|delta|pools|osds|pgs|pgs_brief...]} Subcommand ``dump_pools_json`` shows pg pools info in json only. Usage:: ceph pg dump_pools_json Subcommand ``dump_stuck`` shows information about stuck pgs. Usage:: ceph pg dump_stuck {inactive|unclean|stale|undersized|degraded [inactive|unclean|stale|undersized|degraded...]} {} Subcommand ``force_create_pg`` forces creation of pg . Usage:: ceph pg force_create_pg Subcommand ``getmap`` gets binary pg map to -o/stdout. Usage:: ceph pg getmap Subcommand ``ls`` lists pg with specific pool, osd, state Usage:: ceph pg ls {} {active|clean|down|replay|splitting| scrubbing|scrubq|degraded|inconsistent|peering|repair| recovery|backfill_wait|incomplete|stale| remapped| deep_scrub|backfill|backfill_toofull|recovery_wait| undersized [active|clean|down|replay|splitting| scrubbing|scrubq|degraded|inconsistent|peering|repair| recovery|backfill_wait|incomplete|stale|remapped| deep_scrub|backfill|backfill_toofull|recovery_wait| undersized...]} Subcommand ``ls-by-osd`` lists pg on osd [osd] Usage:: ceph pg ls-by-osd {} {active|clean|down|replay|splitting| scrubbing|scrubq|degraded|inconsistent|peering|repair| recovery|backfill_wait|incomplete|stale| remapped| deep_scrub|backfill|backfill_toofull|recovery_wait| undersized [active|clean|down|replay|splitting| scrubbing|scrubq|degraded|inconsistent|peering|repair| recovery|backfill_wait|incomplete|stale|remapped| deep_scrub|backfill|backfill_toofull|recovery_wait| undersized...]} Subcommand ``ls-by-pool`` lists pg with pool = [poolname] Usage:: ceph pg ls-by-pool {} {active| clean|down|replay|splitting| scrubbing|scrubq|degraded|inconsistent|peering|repair| recovery|backfill_wait|incomplete|stale| remapped| deep_scrub|backfill|backfill_toofull|recovery_wait| undersized [active|clean|down|replay|splitting| scrubbing|scrubq|degraded|inconsistent|peering|repair| recovery|backfill_wait|incomplete|stale|remapped| deep_scrub|backfill|backfill_toofull|recovery_wait| undersized...]} Subcommand ``ls-by-primary`` lists pg with primary = [osd] Usage:: ceph pg ls-by-primary {} {active|clean|down|replay|splitting| scrubbing|scrubq|degraded|inconsistent|peering|repair| recovery|backfill_wait|incomplete|stale| remapped| deep_scrub|backfill|backfill_toofull|recovery_wait| undersized [active|clean|down|replay|splitting| scrubbing|scrubq|degraded|inconsistent|peering|repair| recovery|backfill_wait|incomplete|stale|remapped| deep_scrub|backfill|backfill_toofull|recovery_wait| undersized...]} Subcommand ``map`` shows mapping of pg to osds. Usage:: ceph pg map Subcommand ``repair`` starts repair on . Usage:: ceph pg repair Subcommand ``scrub`` starts scrub on . Usage:: ceph pg scrub Subcommand ``set_full_ratio`` sets ratio at which pgs are considered full. Usage:: ceph pg set_full_ratio Subcommand ``set_nearfull_ratio`` sets ratio at which pgs are considered nearly full. Usage:: ceph pg set_nearfull_ratio Subcommand ``stat`` shows placement group status. Usage:: ceph pg stat quorum ------ Cause MON to enter or exit quorum. Usage:: ceph quorum enter|exit Note: this only works on the MON to which the ``ceph`` command is connected. If you want a specific MON to enter or exit quorum, use this syntax:: ceph tell mon. quorum enter|exit quorum_status ------------- Reports status of monitor quorum. Usage:: ceph quorum_status report ------ Reports full status of cluster, optional title tag strings. Usage:: ceph report { [...]} scrub ----- Scrubs the monitor stores. Usage:: ceph scrub status ------ Shows cluster status. Usage:: ceph status sync force ---------- Forces sync of and clear monitor store. Usage:: ceph sync force {--yes-i-really-mean-it} {--i-know-what-i-am-doing} tell ---- Sends a command to a specific daemon. Usage:: ceph tell [...] version ------- Show mon daemon version Usage:: ceph version Options ======= .. option:: -i infile will specify an input file to be passed along as a payload with the command to the monitor cluster. This is only used for specific monitor commands. .. option:: -o outfile will write any payload returned by the monitor cluster with its reply to outfile. Only specific monitor commands (e.g. osd getmap) return a payload. .. option:: -c ceph.conf, --conf=ceph.conf Use ceph.conf configuration file instead of the default ``/etc/ceph/ceph.conf`` to determine monitor addresses during startup. .. option:: --id CLIENT_ID, --user CLIENT_ID Client id for authentication. .. option:: --name CLIENT_NAME, -n CLIENT_NAME Client name for authentication. .. option:: --cluster CLUSTER Name of the Ceph cluster. .. option:: --admin-daemon ADMIN_SOCKET, daemon DAEMON_NAME Submit admin-socket commands via admin sockets in /var/run/ceph. .. option:: --admin-socket ADMIN_SOCKET_NOPE You probably mean --admin-daemon .. option:: -s, --status Show cluster status. .. option:: -w, --watch Watch live cluster changes. .. option:: --watch-debug Watch debug events. .. option:: --watch-info Watch info events. .. option:: --watch-sec Watch security events. .. option:: --watch-warn Watch warning events. .. option:: --watch-error Watch error events. .. option:: --version, -v Display version. .. option:: --verbose Make verbose. .. option:: --concise Make less verbose. .. option:: -f {json,json-pretty,xml,xml-pretty,plain}, --format Format of output. .. option:: --connect-timeout CLUSTER_TIMEOUT Set a timeout for connecting to the cluster. .. option:: --no-increasing ``--no-increasing`` is off by default. So increasing the osd weight is allowed using the ``reweight-by-utilization`` or ``test-reweight-by-utilization`` commands. If this option is used with these commands, it will help not to increase osd weight even the osd is under utilized. Availability ============ :program:`ceph` is part of Ceph, a massively scalable, open-source, distributed storage system. Please refer to the Ceph documentation at http://ceph.com/docs for more information. See also ======== :doc:`ceph-mon `\(8), :doc:`ceph-osd `\(8), :doc:`ceph-mds `\(8)