ceph/doc/mgr/localpool.rst
Sage Weil 1a0f42b70a pybind/mgr/localpool: module to automagically create localized pools
By default, this will create a pool per rack, 3x replication, with a host
failure domain.  Those parameters can be customized via mgr config-key
options.

Signed-off-by: Sage Weil <sage@redhat.com>
2017-09-11 11:57:35 -05:00

35 lines
1.2 KiB
ReStructuredText

Local pool plugin
=================
The *localpool* plugin can automatically create RADOS pools that are
localized to a subset of the overall cluster. For example, by default, it will
create a pool for each distinct rack in the cluster. This can be useful for some
deployments that want to distribute some data locally as well as globally across the cluster .
Enabling
--------
The *localpool* module is enabled with::
ceph mgr module enable localpool
Configuring
-----------
The *localpool* module understands the following options:
* **subtree** (default: `rack`): which CRUSH subtree type the module
should create a pool for.
* **failure_domain** (default: `host`): what failure domain we should
separate data replicas across.
* **pg_num** (default: `128`): number of PGs to create for each pool
* **num_rep** (default: `3`): number of replicas for each pool.
(Currently, pools are always replicated.)
* **prefix** (default: `by-$subtreetype-`): prefix for the pool name.
These options are set via the config-key interface. For example, to
change the replication level to 2x with only 64 PGs, ::
ceph config-key set mgr/localpool/num_rep 2
ceph config-key set mgr/localpool/pg_num 64