Commit Graph

13 Commits

Author SHA1 Message Date
Patrick Donnelly
e7a7cf429e
doc: filesystem to file system
"Filesystem" is not a word (although fairly common in use).

Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
2019-09-10 08:43:28 -07:00
Patrick Donnelly
6962b4eb9b
doc: create mds state documentation
Fixes: http://tracker.ceph.com/issues/22989

Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
2018-07-24 13:17:21 -07:00
John Spray
82f9960162 doc/cephfs: make scary DR bits less prominent
I'm sure people will still find them, but let's at least
force people to click through one more time to get to the
commands that can damage your cluster.

Also, the ".. danger" directive at the top of the page
wasn't actually getting special formatting, so I changed
it to a ".. warning" which is red.

Signed-off-by: John Spray <john.spray@redhat.com>
2018-07-10 10:52:52 +01:00
Nathan Cutler
aea9fa01ae doc: globally change CRUSH ruleset to CRUSH rule
Since kraken, Ceph enforces a 1:1 correspondence between CRUSH ruleset and
CRUSH rule, so effectively ruleset and rule are the same thing, although
the term "ruleset" still survives - notably in the CRUSH rule itself, where it
effectively denotes the number of the rule.

This commit updates the documentation to more faithfully reflect the current
state of the code.

Fixes: http://tracker.ceph.com/issues/20559
Signed-off-by: Nathan Cutler <ncutler@suse.com>
2017-12-11 17:15:23 +01:00
Yan, Zheng
1ef6aec7e4 doc: add some docs about 'cephfs-data-scan scan_links'
Signed-off-by: "Yan, Zheng" <zyan@redhat.com>
2017-07-07 20:06:18 +08:00
Douglas Fuller
cb86740a5f cephfs: Permit recovering metadata into a new RADOS pool
Add a procedure that permits reconstructing metadata in a potentially
damaged cephfs metadata pool and writing the results into a
freshly-initialized pool that refers to the same data pool. Add option
flags to override checks that would ordinarily prevent this and add
options to the recovery tools to write output to a separate pool instead of
the one selected for recovery. See docs/cephfs/disaster-recovery.rst for
details.

Fixes: http://tracker.ceph.com/issues/15068
Fixes: http://tracker.ceph.com/issues/15069
Signed-off-by: Douglas Fuller <dfuller@redhat.com>
2017-04-04 12:29:01 -07:00
Wido den Hollander
d0c8d8cfe0
doc: Update CephFS disaster recovery documentation
Better documentation about spawning multiple workers to speed
up the recovery proces.

Signed-off-by: Wido den Hollander <wido@42on.com>
2017-01-30 10:07:35 +01:00
John Spray
3c7c65c669 Merge pull request #12337 from vishalkanaujia/wip_12132
tools: cleanup phase of cephfs-data-scan

Reviewed-by: John Spray <john.spray@redhat.com>
2016-12-14 13:52:35 +00:00
Wido den Hollander
82c2f840b6
doc: Fix for worker arguments to cephfs-data-scan tool
The docs were incorrect and these arguments should work.

Signed-off-by: Wido den Hollander <wido@42on.com>
2016-12-07 09:49:09 +01:00
Vishal Kanaujia
8e8f8dd260 tools: cleanup phase of cephfs-data-scan
cephfs-data-scan has a new optional phase to delete xattrs generated
during recovery.

Signed-off-by: Vishal Kanaujia <vishal.kanaujia@sandisk.com>
2016-12-06 18:43:52 +05:30
John Spray
c83cb7a11d tools/cephfs: add pg_files command
Sometimes users know that particular data pool PGs
have been damaged, and they would like to scan
their files to work out which ones might have
been affected.

Fixes: http://tracker.ceph.com/issues/17249
Signed-off-by: John Spray <john.spray@redhat.com>
2016-10-19 23:06:41 +01:00
John Spray
8eaa2f263d doc: add some docs about cephfs-data-scan
These are deliberately fairly sparse, because:
 * These tools are for experts
 * These tools may well be wrapped in a higher
   level recovery tool that orchestrates parallel
   workers at some stage.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-07-13 14:05:17 +01:00
John Spray
a007c529e8 doc: add cephfs disaster recovery guidance
This is a place to put some useful notes about
the new offline recovery tooling.

Signed-off-by: John Spray <john.spray@redhat.com>
2015-02-02 09:56:14 +01:00