mirror of
https://github.com/ceph/ceph
synced 2024-12-25 04:43:17 +00:00
0cb56e0f13
RTD does not support installing system packages, the only ways to install dependencies are setuptools and pip. while ditaa is a tool written in Java. so we need to find a native python tool allowing us to render ditaa images. plantweb is able to the web service for rendering the ditaa diagram. so let's use it as a fallback if "ditaa" is not around. also start a new line after the directive, otherwise planweb server will return 500 at seeing the diagram. Signed-off-by: Kefu Chai <kchai@redhat.com>
73 lines
2.4 KiB
ReStructuredText
73 lines
2.4 KiB
ReStructuredText
==========================
|
|
Block Device Quick Start
|
|
==========================
|
|
|
|
To use this guide, you must have executed the procedures in the `Storage
|
|
Cluster Quick Start`_ guide first. Ensure your :term:`Ceph Storage Cluster` is
|
|
in an ``active + clean`` state before working with the :term:`Ceph Block
|
|
Device`.
|
|
|
|
.. note:: The Ceph Block Device is also known as :term:`RBD` or :term:`RADOS`
|
|
Block Device.
|
|
|
|
|
|
.. ditaa::
|
|
|
|
/------------------\ /----------------\
|
|
| Admin Node | | ceph-client |
|
|
| +-------->+ cCCC |
|
|
| ceph-deploy | | ceph |
|
|
\------------------/ \----------------/
|
|
|
|
|
|
You may use a virtual machine for your ``ceph-client`` node, but do not
|
|
execute the following procedures on the same physical node as your Ceph
|
|
Storage Cluster nodes (unless you use a VM). See `FAQ`_ for details.
|
|
|
|
Create a Block Device Pool
|
|
==========================
|
|
|
|
#. On the admin node, use the ``ceph`` tool to `create a pool`_
|
|
(we recommend the name 'rbd').
|
|
|
|
#. On the admin node, use the ``rbd`` tool to initialize the pool for use by RBD::
|
|
|
|
rbd pool init <pool-name>
|
|
|
|
Configure a Block Device
|
|
========================
|
|
|
|
#. On the ``ceph-client`` node, create a block device image. ::
|
|
|
|
rbd create foo --size 4096 --image-feature layering [-m {mon-IP}] [-k /path/to/ceph.client.admin.keyring] [-p {pool-name}]
|
|
|
|
#. On the ``ceph-client`` node, map the image to a block device. ::
|
|
|
|
sudo rbd map foo --name client.admin [-m {mon-IP}] [-k /path/to/ceph.client.admin.keyring] [-p {pool-name}]
|
|
|
|
#. Use the block device by creating a file system on the ``ceph-client``
|
|
node. ::
|
|
|
|
sudo mkfs.ext4 -m0 /dev/rbd/{pool-name}/foo
|
|
|
|
This may take a few moments.
|
|
|
|
#. Mount the file system on the ``ceph-client`` node. ::
|
|
|
|
sudo mkdir /mnt/ceph-block-device
|
|
sudo mount /dev/rbd/{pool-name}/foo /mnt/ceph-block-device
|
|
cd /mnt/ceph-block-device
|
|
|
|
#. Optionally configure the block device to be automatically mapped and mounted
|
|
at boot (and unmounted/unmapped at shutdown) - see the `rbdmap manpage`_.
|
|
|
|
|
|
See `block devices`_ for additional details.
|
|
|
|
.. _Storage Cluster Quick Start: ../quick-ceph-deploy
|
|
.. _create a pool: ../../rados/operations/pools/#create-a-pool
|
|
.. _block devices: ../../rbd
|
|
.. _FAQ: http://wiki.ceph.com/How_Can_I_Give_Ceph_a_Try
|
|
.. _OS Recommendations: ../os-recommendations
|
|
.. _rbdmap manpage: ../../man/8/rbdmap
|