mirror of
https://github.com/ceph/ceph
synced 2024-12-16 08:26:25 +00:00
76da87a64c
Signed-off-by: Nilamdyuti Goswami <ngoswami@redhat.com>
817 lines
20 KiB
ReStructuredText
817 lines
20 KiB
ReStructuredText
==================================
|
|
ceph -- ceph administration tool
|
|
==================================
|
|
|
|
.. program:: ceph
|
|
|
|
Synopsis
|
|
========
|
|
|
|
| **ceph** **auth** *add* *<entity>* {*<caps>* [*<caps>*...]}
|
|
|
|
| **ceph** **auth** *export* *<entity>*
|
|
|
|
| **ceph** **config-key** *get* *<key>*
|
|
|
|
| **ceph** **mds** *add_data_pool* *<pool>*
|
|
|
|
| **ceph** **mds** *getmap* {*<int[0-]>*}
|
|
|
|
| **ceph** **mon** *add* *<name>* <*IPaddr[:port]*>
|
|
|
|
| **ceph** **mon_status**
|
|
|
|
| **ceph** **osd** *create* {*<uuid>*}
|
|
|
|
| **ceph** **osd** **crush** *add* *<osdname (id|osd.id)>*
|
|
*<float[0.0-]>* *<args>* [*<args>*...]
|
|
|
|
| **ceph** **pg** *force_create_pg* *<pgid>*
|
|
|
|
| **ceph** **pg** *stat*
|
|
|
|
| **ceph** **quorum_status**
|
|
|
|
Description
|
|
===========
|
|
|
|
**ceph** is a control utility which is used for manual deployment and maintenance
|
|
of a Ceph cluster. It provides a diverse set of commands that allows deployment of
|
|
monitors, OSDs, placement groups, MDS and overall maintenance, administration
|
|
of the cluster.
|
|
|
|
Commands
|
|
========
|
|
|
|
**auth**: Manage authentication keys. It is used for adding, removing, exporting
|
|
or updating of authentication keys for a particular entity such as a monitor or
|
|
OSD. It uses some additional subcommands.
|
|
|
|
Subcommand **add** adds authentication info for a particular entity from input
|
|
file, or random key if no input given and/or any caps specified in the command.
|
|
|
|
Usage: ceph auth add <entity> {<caps> [<caps>...]}
|
|
|
|
Subcommand **caps** updates caps for **name** from caps specified in the command.
|
|
|
|
Usage: ceph auth caps <entity> <caps> [<caps>...]
|
|
|
|
Subcommand **del** deletes all caps for **name**.
|
|
|
|
Usage: ceph auth del <entity>
|
|
|
|
Subcommand **export** writes keyring for requested entity, or master keyring if
|
|
none given.
|
|
|
|
Usage: ceph auth export {<entity>}
|
|
|
|
Subcommand **get** writes keyring file with requested key.
|
|
|
|
Usage: ceph auth get <entity>
|
|
|
|
Subcommand **get-key** displays requested key.
|
|
|
|
Usage: ceph auth get-key <entity>
|
|
|
|
Subcommand **get-or-create** adds authentication info for a particular entity
|
|
from input file, or random key if no input given and/or any caps specified in the
|
|
command.
|
|
|
|
Usage: ceph auth get-or-create <entity> {<caps> [<caps>...]}
|
|
|
|
Subcommand **get-or-create-key** gets or adds key for **name** from system/caps
|
|
pairs specified in the command. If key already exists, any given caps must match
|
|
the existing caps for that key.
|
|
|
|
Subcommand **import** reads keyring from input file.
|
|
|
|
Usage: ceph auth import
|
|
|
|
Subcommand **list** lists authentication state.
|
|
|
|
Usage: ceph auth list
|
|
|
|
Subcommand **print-key** displays requested key.
|
|
|
|
Usage: ceph auth print-key <entity>
|
|
|
|
Subcommand **print_key** displays requested key.
|
|
|
|
Usage: ceph auth print_key <entity>
|
|
|
|
**compact**: Causes compaction of monitor's leveldb storage.
|
|
|
|
Usage: ceph compact
|
|
|
|
**config-key**: Manage configuration key. It uses some additional subcommands.
|
|
|
|
Subcommand **get** gets the configuration key.
|
|
|
|
Usage: ceph config-key get <key>
|
|
|
|
Subcommand **put** puts configuration key and values.
|
|
|
|
Usage: ceph config-key put <key> {<val>}
|
|
|
|
Subcommand **exists** checks for configuration keys existence.
|
|
|
|
Usage: ceph config-key exists <key>
|
|
|
|
Subcommand **list** lists configuration keys.
|
|
|
|
Usage: ceph config-key list
|
|
|
|
Subcommand **del** deletes configuration key.
|
|
|
|
Usage: ceph config-key del <key>
|
|
|
|
**df**: Show cluster's free space status.
|
|
|
|
Usage: ceph df
|
|
|
|
**fsid**: Show cluster's FSID/UUID.
|
|
|
|
Usage: ceph fsid
|
|
|
|
**health**: Show cluster's health.
|
|
|
|
Usage: ceph health
|
|
|
|
**heap**: Show heap usage info (available only if compiled with tcmalloc)
|
|
|
|
Usage: ceph heap dump|start_profiler|stop_profiler|release|stats
|
|
|
|
**injectargs**: Inject configuration arguments into monitor.
|
|
|
|
Usage: ceph injectargs <injected_args> [<injected_args>...]
|
|
|
|
**log**: Log supplied text to the monitor log.
|
|
|
|
Usage: ceph log <logtext> [<logtext>...]
|
|
|
|
**mds**: Manage metadata server configuration and administration. It uses some
|
|
additional subcommands.
|
|
|
|
Subcommand **add_data_pool** adds data pool.
|
|
|
|
Usage: ceph mds add_data_pool <pool>
|
|
|
|
Subcommand **cluster_down** takes mds cluster down.
|
|
|
|
Usage: ceph mds cluster_down
|
|
|
|
Subcommand **cluster_up** brings mds cluster up.
|
|
|
|
Usage: ceph mds cluster_up
|
|
|
|
Subcommand **compat** manages compatible features. It uses some additional
|
|
subcommands.
|
|
|
|
Subcommand **rm_compat** removes compatible feature.
|
|
|
|
Usage: ceph mds compat rm_compat <int[0-]>
|
|
|
|
Subcommand **rm_incompat** removes incompatible feature.
|
|
|
|
Usage: ceph mds compat rm_incompat <int[0-]>
|
|
|
|
Subcommand **show** shows mds compatibility settings.
|
|
|
|
Usage: ceph mds compat show
|
|
|
|
Subcommand **deactivate** stops mds.
|
|
|
|
Usage: ceph mds deactivate <who>
|
|
|
|
Subcommand **dump** dumps information, optionally from epoch.
|
|
|
|
Usage: ceph mds dump {<int[0-]>}
|
|
|
|
Subcommand **fail** forces mds to status fail.
|
|
|
|
Usage: ceph mds fail <who>
|
|
|
|
Subcommand **getmap** gets MDS map, optionally from epoch.
|
|
|
|
Usage: ceph mds getmap {<int[0-]>}
|
|
|
|
Subcommand **newfs** makes new filesystem using pools <metadata> and <data>.
|
|
|
|
Usage: ceph mds newfs <int[0-]> <int[0-]> {--yes-i-really-mean-it}
|
|
|
|
Subcommand **remove_data_pool** removes data pool.
|
|
|
|
Usage: ceph mds remove_data_pool <pool>
|
|
|
|
Subcommand **rm** removes inactive mds.
|
|
|
|
Usage: ceph mds rm <int[0-]> <name> (type.id)>
|
|
|
|
Subcommand **rmfailed** removes failed mds.
|
|
|
|
Usage: ceph mds rmfailed <int[0-]>
|
|
|
|
Subcommand **set_max_mds** sets max MDS index.
|
|
|
|
Usage: ceph mds set_max_mds <int[0-]>
|
|
|
|
Subcommand **set_state** sets mds state of <gid> to <numeric-state>.
|
|
|
|
Usage: ceph mds set_state <int[0-]> <int[0-20]>
|
|
|
|
Subcommand **setmap** sets mds map; must supply correct epoch number.
|
|
|
|
Usage: ceph mds setmap <int[0-]>
|
|
|
|
Subcommand **stat** shows MDS status.
|
|
|
|
Usage: ceph mds stat
|
|
|
|
Subcommand **stop** stops mds.
|
|
|
|
Usage: ceph mds stop <who>
|
|
|
|
Subcommand **tell** sends command to particular mds.
|
|
|
|
Usage: ceph mds tell <who> <args> [<args>...]
|
|
|
|
**mon**: Manage monitor configuration and administration. It uses some
|
|
additional subcommands.
|
|
|
|
Subcommand **add** adds new monitor named <name> at <addr>.
|
|
|
|
Usage: ceph mon add <name> <IPaddr[:port]>
|
|
|
|
Subcommand **dump** dumps formatted monmap (optionally from epoch)
|
|
|
|
Usage: ceph mon dump {<int[0-]>}
|
|
|
|
Subcommand **getmap** gets monmap.
|
|
|
|
Usage: ceph mon getmap {<int[0-]>}
|
|
|
|
Subcommand **remove** removes monitor named <name>.
|
|
|
|
Usage: ceph mon remove <name>
|
|
|
|
Subcommand **stat** summarizes monitor status.
|
|
|
|
Usage: ceph mon stat
|
|
|
|
Subcommand **mon_status** reports status of monitors.
|
|
|
|
Usage: ceph mon_status
|
|
|
|
**osd**: Manage OSD configuration and administration. It uses some additional
|
|
subcommands.
|
|
|
|
Subcommand **create** creates new osd (with optional UUID).
|
|
|
|
Usage: ceph osd create {<uuid>}
|
|
|
|
Subcommand **crush** is used for CRUSH management. It uses some additional
|
|
subcommands.
|
|
|
|
Subcommand **add** adds or updates crushmap position and weight for <name> with
|
|
<weight> and location <args>.
|
|
|
|
Usage: ceph osd crush add <osdname (id|osd.id)> <float[0.0-]> <args> [<args>...]
|
|
|
|
Subcommand **add-bucket** adds no-parent (probably root) crush bucket <name> of
|
|
type <type>.
|
|
|
|
Usage: ceph osd crush add-bucket <name> <type>
|
|
|
|
Subcommand **create-or-move** creates entry or moves existing entry for <name>
|
|
<weight> at/to location <args>.
|
|
|
|
Usage: ceph osd crush create-or-move <osdname (id|osd.id)> <float[0.0-]> <args>
|
|
[<args>...]
|
|
|
|
Subcommand **dump** dumps crush map.
|
|
|
|
Usage: ceph osd crush dump
|
|
|
|
Subcommand **link** links existing entry for <name> under location <args>.
|
|
|
|
Usage: ceph osd crush link <name> <args> [<args>...]
|
|
|
|
Subcommand **move** moves existing entry for <name> to location <args>.
|
|
|
|
Usage: ceph osd crush move <name> <args> [<args>...]
|
|
|
|
Subcommand **remove** removes <name> from crush map (everywhere, or just at
|
|
<ancestor>).
|
|
|
|
Usage: ceph osd crush remove <name> {<ancestor>}
|
|
|
|
Subcommand **reweight** change <name>'s weight to <weight> in crush map.
|
|
|
|
Usage: ceph osd crush reweight <name> <float[0.0-]>
|
|
|
|
Subcommand **rm** removes <name> from crush map (everywhere, or just at
|
|
<ancestor>).
|
|
|
|
Usage: ceph osd crush rm <name> {<ancestor>}
|
|
|
|
Subcommand **rule** is used for creating crush rules. It uses some additional
|
|
subcommands.
|
|
|
|
Subcommand **create-erasure** creates crush rule <name> for erasure coded pool
|
|
created with <profile> (default default).
|
|
|
|
Usage: ceph osd crush rule create-erasure <name> {<profile>}
|
|
|
|
Subcommand **create-simple** creates crush rule <name> to start from <root>,
|
|
replicate across buckets of type <type>, using a choose mode of <firstn|indep>
|
|
(default firstn; indep best for erasure pools).
|
|
|
|
Usage: ceph osd crush rule create-simple <name> <root> <type> {firstn|indep}
|
|
|
|
Subcommand **dump** dumps crush rule <name> (default all).
|
|
|
|
Usage: ceph osd crush rule dump {<name>}
|
|
|
|
Subcommand **list** lists crush rules.
|
|
|
|
Usage: ceph osd crush rule list
|
|
|
|
Subcommand **ls** lists crush rules.
|
|
|
|
Usage: ceph osd crush rule ls
|
|
|
|
Subcommand **rm** removes crush rule <name>.
|
|
|
|
Usage: ceph osd crush rule rm <name>
|
|
|
|
Subcommand **set** sets crush map from input file.
|
|
|
|
Usage: ceph osd crush set
|
|
|
|
Subcommand **set** with osdname/osd.id update crushmap position and weight
|
|
for <name> to <weight> with location <args>.
|
|
|
|
Usage: ceph osd crush set <osdname (id|osd.id)> <float[0.0-]> <args> [<args>...]
|
|
|
|
Subcommand **show-tunables** shows current crush tunables.
|
|
|
|
Usage: ceph osd crush show-tunables
|
|
|
|
Subcommand **tunables** sets crush tunables values to <profile>.
|
|
|
|
Usage: ceph osd crush tunables legacy|argonaut|bobtail|firefly|optimal|default
|
|
|
|
Subcommand **unlink** unlinks <name> from crush map (everywhere, or just at
|
|
<ancestor>).
|
|
|
|
Usage: ceph osd crush unlink <name> {<ancestor>}
|
|
|
|
Subcommand **deep-scrub** initiates deep scrub on specified osd.
|
|
|
|
Usage: ceph osd deep-scrub <who>
|
|
|
|
Subcommand **down** sets osd(s) <id> [<id>...] down.
|
|
|
|
Usage: ceph osd down <ids> [<ids>...]
|
|
|
|
Subcommand **dump** prints summary of OSD map.
|
|
|
|
Usage: ceph osd dump {<int[0-]>}
|
|
|
|
Subcommand **erasure-code-profile** is used for managing the erasure code
|
|
profiles. It uses some additional subcommands.
|
|
|
|
Subcommand **get** gets erasure code profile <name>.
|
|
|
|
Usage: ceph osd erasure-code-profile get <name>
|
|
|
|
Subcommand **ls** lists all erasure code profiles.
|
|
|
|
Usage: ceph osd erasure-code-profile ls
|
|
|
|
Subcommand **rm** removes erasure code profile <name>.
|
|
|
|
Usage: ceph osd erasure-code-profile rm <name>
|
|
|
|
Subcommand **set** creates erasure code profile <name> with [<key[=value]> ...]
|
|
pairs. Add a --force at the end to override an existing profile (IT IS RISKY).
|
|
|
|
Usage: ceph osd erasure-code-profile set <name> {<profile> [<profile>...]}
|
|
|
|
Subcommand **find** find osd <id> in the CRUSH map and shows its location.
|
|
|
|
Usage: ceph osd find <int[0-]>
|
|
|
|
Subcommand **getcrushmap** gets CRUSH map.
|
|
|
|
Usage: ceph osd getcrushmap {<int[0-]>}
|
|
|
|
Subcommand **getmap** gets OSD map.
|
|
|
|
Usage: ceph osd getmap {<int[0-]>}
|
|
|
|
Subcommand **getmaxosd** shows largest OSD id.
|
|
|
|
Usage: ceph osd getmaxosd
|
|
|
|
Subcommand **in** sets osd(s) <id> [<id>...] in.
|
|
|
|
Usage: ceph osd in <ids> [<ids>...]
|
|
|
|
Subcommand **lost** marks osd as permanently lost. THIS DESTROYS DATA IF NO
|
|
MORE REPLICAS EXIST, BE CAREFUL.
|
|
|
|
Usage: ceph osd lost <int[0-]> {--yes-i-really-mean-it}
|
|
|
|
Subcommand **ls** shows all OSD ids.
|
|
|
|
Usage: ceph osd ls {<int[0-]>}
|
|
|
|
Subcommand **lspools** lists pools.
|
|
|
|
Usage: ceph osd lspools {<int>}
|
|
|
|
Subcommand **map** finds pg for <object> in <pool>.
|
|
|
|
Usage: ceph osd map <poolname> <objectname>
|
|
|
|
Subcommand **metadata** fetches metadata for osd <id>.
|
|
|
|
Usage: ceph osd metadata <int[0-]>
|
|
|
|
Subcommand **out** sets osd(s) <id> [<id>...] out.
|
|
|
|
Usage: ceph osd out <ids> [<ids>...]
|
|
|
|
Subcommand **pause** pauses osd.
|
|
|
|
Usage: ceph osd pause
|
|
|
|
Subcommand **perf** prints dump of OSD perf summary stats.
|
|
|
|
Usage: ceph osd perf
|
|
|
|
Subcommand **pg-temp** set pg_temp mapping pgid:[<id> [<id>...]] (developers
|
|
only).
|
|
|
|
Usage: ceph osd pg-temp <pgid> {<id> [<id>...]}
|
|
|
|
Subcommand **pool** is used for managing data pools. It uses some additional
|
|
subcommands.
|
|
|
|
Subcommand **create** creates pool.
|
|
|
|
Usage: ceph osd pool create <poolname> <int[0-]> {<int[0-]>} {replicated|erasure}
|
|
{<erasure_code_profile>} {<ruleset>}
|
|
|
|
Subcommand **delete** deletes pool.
|
|
|
|
Usage: ceph osd pool delete <poolname> {<poolname>} {--yes-i-really-really-mean-it}
|
|
|
|
Subcommand **get** gets pool parameter <var>.
|
|
|
|
Usage: ceph osd pool get <poolname> size|min_size|crash_replay_interval|pg_num|
|
|
pgp_num|crush_ruleset|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|
|
|
|
|
ceph osd pool get <poolname> auid|target_max_objects|target_max_bytes
|
|
|
|
ceph osd pool get <poolname> cache_target_dirty_ratio|cache_target_full_ratio
|
|
|
|
ceph osd pool get <poolname> cache_min_flush_age|cache_min_evict_age|
|
|
erasure_code_profile
|
|
|
|
Subcommand **get-quota** obtains object or byte limits for pool.
|
|
|
|
Usage: ceph osd pool get-quota <poolname>
|
|
|
|
Subcommand **mksnap** makes snapshot <snap> in <pool>.
|
|
|
|
Usage: ceph osd pool mksnap <poolname> <snap>
|
|
|
|
Subcommand **rename** renames <srcpool> to <destpool>.
|
|
|
|
Usage: ceph osd pool rename <poolname> <poolname>
|
|
|
|
Subcommand **rmsnap** removes snapshot <snap> from <pool>.
|
|
|
|
Usage: ceph osd pool rmsnap <poolname> <snap>
|
|
|
|
Subcommand **set** sets pool parameter <var> to <val>.
|
|
|
|
Usage: ceph osd pool set <poolname> size|min_size|crash_replay_interval|pg_num|
|
|
pgp_num|crush_ruleset|hashpspool|hit_set_type|hit_set_period|
|
|
|
|
ceph osd pool set <poolname> hit_set_count|hit_set_fpp|debug_fake_ec_pool
|
|
|
|
ceph osd pool set <poolname> target_max_bytes|target_max_objects
|
|
|
|
ceph osd pool set <poolname> cache_target_dirty_ratio|cache_target_full_ratio
|
|
|
|
ceph osd pool set <poolname> cache_min_flush_age
|
|
|
|
ceph osd pool set <poolname> cache_min_evict_age|auid <val> {--yes-i-really-mean-it}
|
|
|
|
Subcommand **set-quota** sets object or byte limit on pool.
|
|
|
|
Usage: ceph osd pool set-quota <poolname> max_objects|max_bytes <val>
|
|
|
|
Subcommand **stats** obtain stats from all pools, or from specified pool.
|
|
|
|
Usage: ceph osd pool stats {<name>}
|
|
|
|
Subcommand **primary-affinity** adjust osd primary-affinity from 0.0 <=<weight>
|
|
<= 1.0
|
|
|
|
Usage: ceph osd primary-affinity <osdname (id|osd.id)> <float[0.0-1.0]>
|
|
|
|
Subcommand **primary-temp** sets primary_temp mapping pgid:<id>|-1 (developers
|
|
only).
|
|
|
|
Usage: ceph osd primary-temp <pgid> <id>
|
|
|
|
Subcommand **repair** initiates repair on a specified osd.
|
|
|
|
Usage: ceph osd repair <who>
|
|
|
|
Subcommand **reweight** reweights osd to 0.0 < <weight> < 1.0.
|
|
|
|
Usage: osd reweight <int[0-]> <float[0.0-1.0]>
|
|
|
|
Subcommand **reweight-by-utilization** reweight OSDs by utilization
|
|
[overload-percentage-for-consideration, default 120].
|
|
|
|
Usage: ceph osd reweight-by-utilization {<int[100-]>}
|
|
|
|
Subcommand **rm** removes osd(s) <id> [<id>...] in the cluster.
|
|
|
|
Usage: ceph osd rm <ids> [<ids>...]
|
|
|
|
Subcommand **scrub** initiates scrub on specified osd.
|
|
|
|
Usage: ceph osd scrub <who>
|
|
|
|
Subcommand **set** sets <key>.
|
|
|
|
Usage: ceph osd set pause|noup|nodown|noout|noin|nobackfill|norecover|noscrub|
|
|
nodeep-scrub|notieragent
|
|
|
|
Subcommand **setcrushmap** sets crush map from input file.
|
|
|
|
Usage: ceph osd setcrushmap
|
|
|
|
Subcommand **setmaxosd** sets new maximum osd value.
|
|
|
|
Usage: ceph osd setmaxosd <int[0-]>
|
|
|
|
Subcommand **stat** prints summary of OSD map.
|
|
|
|
Usage: ceph osd stat
|
|
|
|
Subcommand **thrash** thrashes OSDs for <num_epochs>.
|
|
|
|
Usage: ceph osd thrash <int[0-]>
|
|
|
|
Subcommand **tier** is used for managing tiers. It uses some additional
|
|
subcommands.
|
|
|
|
Subcommand **add** adds the tier <tierpool> (the second one) to base pool <pool>
|
|
(the first one).
|
|
|
|
Usage: ceph osd tier add <poolname> <poolname> {--force-nonempty}
|
|
|
|
Subcommand **add-cache** adds a cache <tierpool> (the second one) of size <size>
|
|
to existing pool <pool> (the first one).
|
|
|
|
Usage: ceph osd tier add-cache <poolname> <poolname> <int[0-]>
|
|
|
|
Subcommand **cache-mode** specifies the caching mode for cache tier <pool>.
|
|
|
|
Usage: ceph osd tier cache-mode <poolname> none|writeback|forward|readonly
|
|
|
|
Subcommand **remove** removes the tier <tierpool> (the second one) from base pool
|
|
<pool> (the first one).
|
|
|
|
Usage: ceph osd tier remove <poolname> <poolname>
|
|
|
|
Subcommand **remove-overlay** removes the overlay pool for base pool <pool>.
|
|
|
|
Usage: ceph osd tier remove-overlay <poolname>
|
|
|
|
Subcommand **set-overlay** set the overlay pool for base pool <pool> to be
|
|
<overlaypool>.
|
|
|
|
Usage: ceph osd tier set-overlay <poolname> <poolname>
|
|
|
|
Subcommand **tree** prints OSD tree.
|
|
|
|
Usage: ceph osd tree {<int[0-]>}
|
|
|
|
Subcommand **unpause** unpauses osd.
|
|
|
|
Usage: ceph osd unpause
|
|
|
|
Subcommand **unset** unsets <key>.
|
|
|
|
Usage: osd unset pause|noup|nodown|noout|noin|nobackfill|norecover|noscrub|
|
|
nodeep-scrub|notieragent
|
|
|
|
**pg**: It is used for managing the placement groups in OSDs. It uses some
|
|
additional subcommands.
|
|
|
|
Subcommand **debug** shows debug info about pgs.
|
|
|
|
Usage: ceph pg debug unfound_objects_exist|degraded_pgs_exist
|
|
|
|
Subcommand **deep-scrub** starts deep-scrub on <pgid>.
|
|
|
|
Usage: ceph pg deep-scrub <pgid>
|
|
|
|
Subcommand **dump** shows human-readable versions of pg map (only 'all' valid with
|
|
plain).
|
|
|
|
Usage: ceph pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief}
|
|
|
|
ceph pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief...}
|
|
|
|
Subcommand **dump_json** shows human-readable version of pg map in json only.
|
|
|
|
Usage: ceph pg dump_json {all|summary|sum|pools|osds|pgs[all|summary|sum|pools|
|
|
osds|pgs...]}
|
|
|
|
Subcommand **dump_pools_json** shows pg pools info in json only.
|
|
|
|
Usage: ceph pg dump_pools_json
|
|
|
|
Subcommand **dump_stuck** shows information about stuck pgs.
|
|
|
|
Usage: ceph pg dump_stuck {inactive|unclean|stale[inactive|unclean|stale...]}
|
|
{<int>}
|
|
|
|
Subcommand **force_create_pg** forces creation of pg <pgid>.
|
|
|
|
Usage: ceph pg force_create_pg <pgid>
|
|
|
|
Subcommand **getmap** gets binary pg map to -o/stdout.
|
|
|
|
Usage: ceph pg getmap
|
|
|
|
Subcommand **map** shows mapping of pg to osds.
|
|
|
|
Usage: ceph pg map <pgid>
|
|
|
|
Subcommand **repair** starts repair on <pgid>.
|
|
|
|
Usage: ceph pg repair <pgid>
|
|
|
|
Subcommand **scrub** starts scrub on <pgid>.
|
|
|
|
Usage: ceph pg scrub <pgid>
|
|
|
|
Subcommand **send_pg_creates** triggers pg creates to be issued.
|
|
|
|
Usage: ceph pg send_pg_creates
|
|
|
|
Subcommand **set_full_ratio** sets ratio at which pgs are considered full.
|
|
|
|
Usage: ceph pg set_full_ratio <float[0.0-1.0]>
|
|
|
|
Subcommand **set_nearfull_ratio** sets ratio at which pgs are considered nearly
|
|
full.
|
|
|
|
Usage: ceph pg set_nearfull_ratio <float[0.0-1.0]>
|
|
|
|
Subcommand **stat** shows placement group status.
|
|
|
|
Usage: ceph pg stat
|
|
|
|
**quorum**: Enter or exit quorum.
|
|
|
|
Usage: ceph quorum enter|exit
|
|
|
|
**quorum_status**: Reports status of monitor quorum.
|
|
|
|
Usage: ceph quorum_status
|
|
|
|
**report**: Reports full status of cluster, optional title tag strings.
|
|
|
|
Usage: ceph report {<tags> [<tags>...]}
|
|
|
|
**scrub**: Scrubs the monitor stores.
|
|
|
|
Usage: ceph scrub
|
|
|
|
**status**: Shows cluster status.
|
|
|
|
Usage: ceph status
|
|
|
|
**sync force**: Forces sync of and clear monitor store.
|
|
|
|
Usage: ceph sync force {--yes-i-really-mean-it} {--i-know-what-i-am-doing}
|
|
|
|
**tell**: Sends a command to a specific daemon.
|
|
|
|
Usage: ceph tell <name (type.id)> <args> [<args>...]
|
|
|
|
Options
|
|
=======
|
|
|
|
.. option:: -i infile
|
|
|
|
will specify an input file to be passed along as a payload with the
|
|
command to the monitor cluster. This is only used for specific
|
|
monitor commands.
|
|
|
|
.. option:: -o outfile
|
|
|
|
will write any payload returned by the monitor cluster with its
|
|
reply to outfile. Only specific monitor commands (e.g. osd getmap)
|
|
return a payload.
|
|
|
|
.. option:: -c ceph.conf, --conf=ceph.conf
|
|
|
|
Use ceph.conf configuration file instead of the default
|
|
/etc/ceph/ceph.conf to determine monitor addresses during startup.
|
|
|
|
.. option:: --id CLIENT_ID, --user CLIENT_ID
|
|
|
|
Client id for authentication.
|
|
|
|
.. option:: --name CLIENT_NAME, -n CLIENT_NAME
|
|
|
|
Client name for authentication.
|
|
|
|
.. option:: --cluster CLUSTER
|
|
|
|
Name of the Ceph cluster.
|
|
|
|
.. option:: --admin-daemon ADMIN_SOCKET
|
|
|
|
Submit admin-socket commands.
|
|
|
|
.. option:: --admin-socket ADMIN_SOCKET_NOPE
|
|
|
|
You probably mean --admin-daemon
|
|
|
|
.. option:: -s, --status
|
|
|
|
Show cluster status.
|
|
|
|
.. option:: -w, --watch
|
|
|
|
Watch live cluster changes.
|
|
|
|
.. option:: --watch-debug
|
|
|
|
Watch debug events.
|
|
|
|
.. option:: --watch-info
|
|
|
|
Watch info events.
|
|
|
|
.. option:: --watch-sec
|
|
|
|
Watch security events.
|
|
|
|
.. option:: --watch-warn
|
|
|
|
Watch warning events.
|
|
|
|
.. option:: --watch-error
|
|
|
|
Watch error events.
|
|
|
|
.. option:: --version, -v
|
|
|
|
Display version.
|
|
|
|
.. option:: --verbose
|
|
|
|
Make verbose.
|
|
|
|
.. option:: --concise
|
|
|
|
Make less verbose.
|
|
|
|
.. option:: -f {json,json-pretty,xml,xml-pretty,plain}, --format
|
|
|
|
Format of output.
|
|
|
|
.. option:: --connect-timeout CLUSTER_TIMEOUT
|
|
|
|
Set a timeout for connecting to the cluster.
|
|
|
|
|
|
Availability
|
|
============
|
|
|
|
**ceph** is part of the Ceph distributed storage system. Please refer to the Ceph documentation at
|
|
http://ceph.com/docs for more information.
|
|
|
|
|
|
See also
|
|
========
|
|
|
|
:doc:`ceph-mon <ceph-mon>`\(8),
|
|
:doc:`ceph-osd <ceph-osd>`\(8),
|
|
:doc:`ceph-mds <ceph-mds>`\(8)
|