Documentation for new pg force-recovery, pg force-backfill,
pg-cancel-force-recovery and pg-cancel-force-backfill.
Signed-off-by: Piotr Dałek <piotr.dalek@corp.ovh.com>
* 8/ceph.rst: `rm` subcommand removes osd from osdmap, not the cluster.
the latter is more ambiguous in different contexts.
* rados/operations/add-or-rm-osds.rst: add a subsection of "Replacing an
OSD". update the subsection of "Removing the OSD" with "ceph osd
purge" command
* release-notes.rst: link from it to the new subsection in add-or-rm-osds.rst
Signed-off-by: Kefu Chai <kchai@redhat.com>
* doc/release-notes.rst: escape asterisks not used for inline emphasis
fix bad hyper links
* doc/rados/troubleshooting/troubleshooting-osd.rst: escape asterisks
not used for inline emphasis
* doc/radosgw/index.rst: add orphan docs to toc
* doc/dev/perf_histograms.rst: indent block quote
* doc/install/manual-freebsd-deployment.rst: fix block quotes
* doc/mgr/administrator.rst: escape asterisks not used for inline emphasis
* doc/start/quick-rbd.rst: add missing hyper link target
Signed-off-by: Kefu Chai <kchai@redhat.com>
1) ruleset is an obsolete term, and
2) crush-{rule,failure-domain,...} is more descriptive.
Note that we are changing the names of the erasure code profile keys
from ruleset-* to crush-*. We will update this on upgrade when the
luminous flag is set, but that means that during mon upgrade you cannot
create EC pools that use these fields.
When the upgrade completes (users sets require_osd_release = luminous)
existing ec profiles are updated automatically.
Signed-off-by: Sage Weil <sage@redhat.com>
We don't use pool ID 0 any more because CephFS can't handle it.
Update the user quick-start to include it, and add some links to support that process.
Fix the tests to handle the new state of life.
Signed-off-by: Greg Farnum <gfarnum@redhat.com>
Add some brief documentation on updating their caps and link to it.
Fixes: http://tracker.ceph.com/issues/20296
Signed-off-by: Greg Farnum <gfarnum@redhat.com>
With bluestore, making the smallest write match min_alloc_size avoids
write amplification. With EC pools this is the stripe unit, or
stripe_width / num_data_chunks. Rather than requiring people to divide
by k to get the smallest ec write, allow it to be specified directly
via stripe_unit. Store it in the ec profile so changing a monitor
config option isn't necessary to set it.
This is particularly important for ec overwrites since they allow random i/o
which should match bluestore's checksum granularity (aka min_alloc_size).
Signed-off-by: Josh Durgin <jdurgin@redhat.com>
The class qualifier is documented for devices and step take
only. Although it shows in the buckets and could be set by the user, it
would be very error prone to do so.
Refs: http://tracker.ceph.com/issues/18943
Signed-off-by: Loic Dachary <ldachary@redhat.com>
If we only fill OSDs to 95%, we should factor that into
the MAX AVAIL calculation for the pool.
Fixes: http://tracker.ceph.com/issues/18522
Signed-off-by: Sage Weil <sage@redhat.com>
Tell users they need to set this to true before Monitors will allow
pools to be removed.
Also update the Pending Release Notes so that users can find this change
there.
This was changed with commit 5d7f4ea
Signed-off-by: Wido den Hollander <wido@42on.com>
This mostly is just removing the commands from the man page ceph(1). I
left the legacy section in doc/cephfs/administration.rst as-is.
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>