================================== ceph -- ceph administration tool ================================== .. program:: ceph Synopsis ======== | **ceph** **auth** *add* ** {** [**...]} | **ceph** **auth** *export* ** | **ceph** **config-key** *get* ** | **ceph** **mds** *add_data_pool* ** | **ceph** **mds** *getmap* {**} | **ceph** **mon** *add* ** <*IPaddr[:port]*> | **ceph** **mon_status** | **ceph** **osd** *create* {**} | **ceph** **osd** **crush** *add* ** ** ** [**...] | **ceph** **pg** *force_create_pg* ** | **ceph** **pg** *stat* | **ceph** **quorum_status** Description =========== **ceph** is a control utility which is used for manual deployment and maintenance of a Ceph cluster. It provides a diverse set of commands that allows deployment of monitors, OSDs, placement groups, MDS and overall maintenance, administration of the cluster. Commands ======== **auth**: Manage authentication keys. It is used for adding, removing, exporting or updating of authentication keys for a particular entity such as a monitor or OSD. It uses some additional subcommands. Subcommand **add** adds authentication info for a particular entity from input file, or random key if no input given and/or any caps specified in the command. Usage: ceph auth add { [...]} Subcommand **caps** updates caps for **name** from caps specified in the command. Usage: ceph auth caps [...] Subcommand **del** deletes all caps for **name**. Usage: ceph auth del Subcommand **export** writes keyring for requested entity, or master keyring if none given. Usage: ceph auth export {} Subcommand **get** writes keyring file with requested key. Usage: ceph auth get Subcommand **get-key** displays requested key. Usage: ceph auth get-key Subcommand **get-or-create** adds authentication info for a particular entity from input file, or random key if no input given and/or any caps specified in the command. Usage: ceph auth get-or-create { [...]} Subcommand **get-or-create-key** gets or adds key for **name** from system/caps pairs specified in the command. If key already exists, any given caps must match the existing caps for that key. Subcommand **import** reads keyring from input file. Usage: ceph auth import Subcommand **list** lists authentication state. Usage: ceph auth list Subcommand **print-key** displays requested key. Usage: ceph auth print-key Subcommand **print_key** displays requested key. Usage: ceph auth print_key **compact**: Causes compaction of monitor's leveldb storage. Usage: ceph compact **config-key**: Manage configuration key. It uses some additional subcommands. Subcommand **get** gets the configuration key. Usage: ceph config-key get Subcommand **put** puts configuration key and values. Usage: ceph config-key put {} Subcommand **exists** checks for configuration keys existence. Usage: ceph config-key exists Subcommand **list** lists configuration keys. Usage: ceph config-key list Subcommand **del** deletes configuration key. Usage: ceph config-key del **df**: Show cluster's free space status. Usage: ceph df **fsid**: Show cluster's FSID/UUID. Usage: ceph fsid **health**: Show cluster's health. Usage: ceph health **heap**: Show heap usage info (available only if compiled with tcmalloc) Usage: ceph heap dump|start_profiler|stop_profiler|release|stats **injectargs**: Inject configuration arguments into monitor. Usage: ceph injectargs [...] **log**: Log supplied text to the monitor log. Usage: ceph log [...] **mds**: Manage metadata server configuration and administration. It uses some additional subcommands. Subcommand **add_data_pool** adds data pool. Usage: ceph mds add_data_pool Subcommand **cluster_down** takes mds cluster down. Usage: ceph mds cluster_down Subcommand **cluster_up** brings mds cluster up. Usage: ceph mds cluster_up Subcommand **compat** manages compatible features. It uses some additional subcommands. Subcommand **rm_compat** removes compatible feature. Usage: ceph mds compat rm_compat Subcommand **rm_incompat** removes incompatible feature. Usage: ceph mds compat rm_incompat Subcommand **show** shows mds compatibility settings. Usage: ceph mds compat show Subcommand **deactivate** stops mds. Usage: ceph mds deactivate Subcommand **dump** dumps information, optionally from epoch. Usage: ceph mds dump {} Subcommand **fail** forces mds to status fail. Usage: ceph mds fail Subcommand **getmap** gets MDS map, optionally from epoch. Usage: ceph mds getmap {} Subcommand **newfs** makes new filesystem using pools and . Usage: ceph mds newfs {--yes-i-really-mean-it} Subcommand **remove_data_pool** removes data pool. Usage: ceph mds remove_data_pool Subcommand **rm** removes inactive mds. Usage: ceph mds rm (type.id)> Subcommand **rmfailed** removes failed mds. Usage: ceph mds rmfailed Subcommand **set_max_mds** sets max MDS index. Usage: ceph mds set_max_mds Subcommand **set_state** sets mds state of to . Usage: ceph mds set_state Subcommand **setmap** sets mds map; must supply correct epoch number. Usage: ceph mds setmap Subcommand **stat** shows MDS status. Usage: ceph mds stat Subcommand **stop** stops mds. Usage: ceph mds stop Subcommand **tell** sends command to particular mds. Usage: ceph mds tell [...] **mon**: Manage monitor configuration and administration. It uses some additional subcommands. Subcommand **add** adds new monitor named at . Usage: ceph mon add Subcommand **dump** dumps formatted monmap (optionally from epoch) Usage: ceph mon dump {} Subcommand **getmap** gets monmap. Usage: ceph mon getmap {} Subcommand **remove** removes monitor named . Usage: ceph mon remove Subcommand **stat** summarizes monitor status. Usage: ceph mon stat Subcommand **mon_status** reports status of monitors. Usage: ceph mon_status **osd**: Manage OSD configuration and administration. It uses some additional subcommands. Subcommand **create** creates new osd (with optional UUID). Usage: ceph osd create {} Subcommand **crush** is used for CRUSH management. It uses some additional subcommands. Subcommand **add** adds or updates crushmap position and weight for with and location . Usage: ceph osd crush add [...] Subcommand **add-bucket** adds no-parent (probably root) crush bucket of type . Usage: ceph osd crush add-bucket Subcommand **create-or-move** creates entry or moves existing entry for at/to location . Usage: ceph osd crush create-or-move [...] Subcommand **dump** dumps crush map. Usage: ceph osd crush dump Subcommand **link** links existing entry for under location . Usage: ceph osd crush link [...] Subcommand **move** moves existing entry for to location . Usage: ceph osd crush move [...] Subcommand **remove** removes from crush map (everywhere, or just at ). Usage: ceph osd crush remove {} Subcommand **reweight** change 's weight to in crush map. Usage: ceph osd crush reweight Subcommand **rm** removes from crush map (everywhere, or just at ). Usage: ceph osd crush rm {} Subcommand **rule** is used for creating crush rules. It uses some additional subcommands. Subcommand **create-erasure** creates crush rule for erasure coded pool created with (default default). Usage: ceph osd crush rule create-erasure {} Subcommand **create-simple** creates crush rule to start from , replicate across buckets of type , using a choose mode of (default firstn; indep best for erasure pools). Usage: ceph osd crush rule create-simple {firstn|indep} Subcommand **dump** dumps crush rule (default all). Usage: ceph osd crush rule dump {} Subcommand **list** lists crush rules. Usage: ceph osd crush rule list Subcommand **ls** lists crush rules. Usage: ceph osd crush rule ls Subcommand **rm** removes crush rule . Usage: ceph osd crush rule rm Subcommand **set** sets crush map from input file. Usage: ceph osd crush set Subcommand **set** with osdname/osd.id update crushmap position and weight for to with location . Usage: ceph osd crush set [...] Subcommand **show-tunables** shows current crush tunables. Usage: ceph osd crush show-tunables Subcommand **tunables** sets crush tunables values to . Usage: ceph osd crush tunables legacy|argonaut|bobtail|firefly|optimal|default Subcommand **unlink** unlinks from crush map (everywhere, or just at ). Usage: ceph osd crush unlink {} Subcommand **deep-scrub** initiates deep scrub on specified osd. Usage: ceph osd deep-scrub Subcommand **down** sets osd(s) [...] down. Usage: ceph osd down [...] Subcommand **dump** prints summary of OSD map. Usage: ceph osd dump {} Subcommand **erasure-code-profile** is used for managing the erasure code profiles. It uses some additional subcommands. Subcommand **get** gets erasure code profile . Usage: ceph osd erasure-code-profile get Subcommand **ls** lists all erasure code profiles. Usage: ceph osd erasure-code-profile ls Subcommand **rm** removes erasure code profile . Usage: ceph osd erasure-code-profile rm Subcommand **set** creates erasure code profile with [ ...] pairs. Add a --force at the end to override an existing profile (IT IS RISKY). Usage: ceph osd erasure-code-profile set { [...]} Subcommand **find** find osd in the CRUSH map and shows its location. Usage: ceph osd find Subcommand **getcrushmap** gets CRUSH map. Usage: ceph osd getcrushmap {} Subcommand **getmap** gets OSD map. Usage: ceph osd getmap {} Subcommand **getmaxosd** shows largest OSD id. Usage: ceph osd getmaxosd Subcommand **in** sets osd(s) [...] in. Usage: ceph osd in [...] Subcommand **lost** marks osd as permanently lost. THIS DESTROYS DATA IF NO MORE REPLICAS EXIST, BE CAREFUL. Usage: ceph osd lost {--yes-i-really-mean-it} Subcommand **ls** shows all OSD ids. Usage: ceph osd ls {} Subcommand **lspools** lists pools. Usage: ceph osd lspools {} Subcommand **map** finds pg for in . Usage: ceph osd map Subcommand **metadata** fetches metadata for osd . Usage: ceph osd metadata Subcommand **out** sets osd(s) [...] out. Usage: ceph osd out [...] Subcommand **pause** pauses osd. Usage: ceph osd pause Subcommand **perf** prints dump of OSD perf summary stats. Usage: ceph osd perf Subcommand **pg-temp** set pg_temp mapping pgid:[ [...]] (developers only). Usage: ceph osd pg-temp { [...]} Subcommand **pool** is used for managing data pools. It uses some additional subcommands. Subcommand **create** creates pool. Usage: ceph osd pool create {} {replicated|erasure} {} {} Subcommand **delete** deletes pool. Usage: ceph osd pool delete {} {--yes-i-really-really-mean-it} Subcommand **get** gets pool parameter . Usage: ceph osd pool get size|min_size|crash_replay_interval|pg_num| pgp_num|crush_ruleset|hit_set_type|hit_set_period|hit_set_count|hit_set_fpp| ceph osd pool get auid|target_max_objects|target_max_bytes ceph osd pool get cache_target_dirty_ratio|cache_target_full_ratio ceph osd pool get cache_min_flush_age|cache_min_evict_age| erasure_code_profile Subcommand **get-quota** obtains object or byte limits for pool. Usage: ceph osd pool get-quota Subcommand **mksnap** makes snapshot in . Usage: ceph osd pool mksnap Subcommand **rename** renames to . Usage: ceph osd pool rename Subcommand **rmsnap** removes snapshot from . Usage: ceph osd pool rmsnap Subcommand **set** sets pool parameter to . Usage: ceph osd pool set size|min_size|crash_replay_interval|pg_num| pgp_num|crush_ruleset|hashpspool|hit_set_type|hit_set_period| ceph osd pool set hit_set_count|hit_set_fpp|debug_fake_ec_pool ceph osd pool set target_max_bytes|target_max_objects ceph osd pool set cache_target_dirty_ratio|cache_target_full_ratio ceph osd pool set cache_min_flush_age ceph osd pool set cache_min_evict_age|auid {--yes-i-really-mean-it} Subcommand **set-quota** sets object or byte limit on pool. Usage: ceph osd pool set-quota max_objects|max_bytes Subcommand **stats** obtain stats from all pools, or from specified pool. Usage: ceph osd pool stats {} Subcommand **primary-affinity** adjust osd primary-affinity from 0.0 <= <= 1.0 Usage: ceph osd primary-affinity Subcommand **primary-temp** sets primary_temp mapping pgid:|-1 (developers only). Usage: ceph osd primary-temp Subcommand **repair** initiates repair on a specified osd. Usage: ceph osd repair Subcommand **reweight** reweights osd to 0.0 < < 1.0. Usage: osd reweight Subcommand **reweight-by-utilization** reweight OSDs by utilization [overload-percentage-for-consideration, default 120]. Usage: ceph osd reweight-by-utilization {} Subcommand **rm** removes osd(s) [...] in the cluster. Usage: ceph osd rm [...] Subcommand **scrub** initiates scrub on specified osd. Usage: ceph osd scrub Subcommand **set** sets . Usage: ceph osd set pause|noup|nodown|noout|noin|nobackfill|norecover|noscrub| nodeep-scrub|notieragent Subcommand **setcrushmap** sets crush map from input file. Usage: ceph osd setcrushmap Subcommand **setmaxosd** sets new maximum osd value. Usage: ceph osd setmaxosd Subcommand **stat** prints summary of OSD map. Usage: ceph osd stat Subcommand **thrash** thrashes OSDs for . Usage: ceph osd thrash Subcommand **tier** is used for managing tiers. It uses some additional subcommands. Subcommand **add** adds the tier (the second one) to base pool (the first one). Usage: ceph osd tier add {--force-nonempty} Subcommand **add-cache** adds a cache (the second one) of size to existing pool (the first one). Usage: ceph osd tier add-cache Subcommand **cache-mode** specifies the caching mode for cache tier . Usage: ceph osd tier cache-mode none|writeback|forward|readonly Subcommand **remove** removes the tier (the second one) from base pool (the first one). Usage: ceph osd tier remove Subcommand **remove-overlay** removes the overlay pool for base pool . Usage: ceph osd tier remove-overlay Subcommand **set-overlay** set the overlay pool for base pool to be . Usage: ceph osd tier set-overlay Subcommand **tree** prints OSD tree. Usage: ceph osd tree {} Subcommand **unpause** unpauses osd. Usage: ceph osd unpause Subcommand **unset** unsets . Usage: osd unset pause|noup|nodown|noout|noin|nobackfill|norecover|noscrub| nodeep-scrub|notieragent **pg**: It is used for managing the placement groups in OSDs. It uses some additional subcommands. Subcommand **debug** shows debug info about pgs. Usage: ceph pg debug unfound_objects_exist|degraded_pgs_exist Subcommand **deep-scrub** starts deep-scrub on . Usage: ceph pg deep-scrub Subcommand **dump** shows human-readable versions of pg map (only 'all' valid with plain). Usage: ceph pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief} ceph pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief...} Subcommand **dump_json** shows human-readable version of pg map in json only. Usage: ceph pg dump_json {all|summary|sum|pools|osds|pgs[all|summary|sum|pools| osds|pgs...]} Subcommand **dump_pools_json** shows pg pools info in json only. Usage: ceph pg dump_pools_json Subcommand **dump_stuck** shows information about stuck pgs. Usage: ceph pg dump_stuck {inactive|unclean|stale[inactive|unclean|stale...]} {} Subcommand **force_create_pg** forces creation of pg . Usage: ceph pg force_create_pg Subcommand **getmap** gets binary pg map to -o/stdout. Usage: ceph pg getmap Subcommand **map** shows mapping of pg to osds. Usage: ceph pg map Subcommand **repair** starts repair on . Usage: ceph pg repair Subcommand **scrub** starts scrub on . Usage: ceph pg scrub Subcommand **send_pg_creates** triggers pg creates to be issued. Usage: ceph pg send_pg_creates Subcommand **set_full_ratio** sets ratio at which pgs are considered full. Usage: ceph pg set_full_ratio Subcommand **set_nearfull_ratio** sets ratio at which pgs are considered nearly full. Usage: ceph pg set_nearfull_ratio Subcommand **stat** shows placement group status. Usage: ceph pg stat **quorum**: Enter or exit quorum. Usage: ceph quorum enter|exit **quorum_status**: Reports status of monitor quorum. Usage: ceph quorum_status **report**: Reports full status of cluster, optional title tag strings. Usage: ceph report { [...]} **scrub**: Scrubs the monitor stores. Usage: ceph scrub **status**: Shows cluster status. Usage: ceph status **sync force**: Forces sync of and clear monitor store. Usage: ceph sync force {--yes-i-really-mean-it} {--i-know-what-i-am-doing} **tell**: Sends a command to a specific daemon. Usage: ceph tell [...] Options ======= .. option:: -i infile will specify an input file to be passed along as a payload with the command to the monitor cluster. This is only used for specific monitor commands. .. option:: -o outfile will write any payload returned by the monitor cluster with its reply to outfile. Only specific monitor commands (e.g. osd getmap) return a payload. .. option:: -c ceph.conf, --conf=ceph.conf Use ceph.conf configuration file instead of the default /etc/ceph/ceph.conf to determine monitor addresses during startup. .. option:: --id CLIENT_ID, --user CLIENT_ID Client id for authentication. .. option:: --name CLIENT_NAME, -n CLIENT_NAME Client name for authentication. .. option:: --cluster CLUSTER Name of the Ceph cluster. .. option:: --admin-daemon ADMIN_SOCKET Submit admin-socket commands. .. option:: --admin-socket ADMIN_SOCKET_NOPE You probably mean --admin-daemon .. option:: -s, --status Show cluster status. .. option:: -w, --watch Watch live cluster changes. .. option:: --watch-debug Watch debug events. .. option:: --watch-info Watch info events. .. option:: --watch-sec Watch security events. .. option:: --watch-warn Watch warning events. .. option:: --watch-error Watch error events. .. option:: --version, -v Display version. .. option:: --verbose Make verbose. .. option:: --concise Make less verbose. .. option:: -f {json,json-pretty,xml,xml-pretty,plain}, --format Format of output. .. option:: --connect-timeout CLUSTER_TIMEOUT Set a timeout for connecting to the cluster. Availability ============ **ceph** is part of the Ceph distributed storage system. Please refer to the Ceph documentation at http://ceph.com/docs for more information. See also ======== :doc:`ceph-mon `\(8), :doc:`ceph-osd `\(8), :doc:`ceph-mds `\(8)