mirror of
https://github.com/ceph/ceph
synced 2024-12-24 20:33:27 +00:00
137 lines
5.9 KiB
ReStructuredText
137 lines
5.9 KiB
ReStructuredText
=============================
|
|
Block Devices and CloudStack
|
|
=============================
|
|
|
|
You may use Ceph Block Device images with CloudStack 4.0 and higher through
|
|
``libvirt``, which configures the QEMU interface to ``librbd``. Ceph stripes
|
|
block device images as objects across the cluster, which means that large Ceph
|
|
Block Device images have better performance than a standalone server!
|
|
|
|
To use Ceph Block Devices with CloudStack 4.0 and higher, you must install QEMU,
|
|
``libvirt``, and CloudStack first. We recommend using a separate physical host
|
|
for your CloudStack installation. CloudStack recommends a minimum of 4GB of RAM
|
|
and a dual-core processor, but more CPU and RAM will perform better. The
|
|
following diagram depicts the CloudStack/Ceph technology stack.
|
|
|
|
|
|
.. ditaa:: +---------------------------------------------------+
|
|
| CloudStack |
|
|
+---------------------------------------------------+
|
|
| libvirt |
|
|
+------------------------+--------------------------+
|
|
|
|
|
| configures
|
|
v
|
|
+---------------------------------------------------+
|
|
| QEMU |
|
|
+---------------------------------------------------+
|
|
| librbd |
|
|
+---------------------------------------------------+
|
|
| librados |
|
|
+------------------------+-+------------------------+
|
|
| OSDs | | Monitors |
|
|
+------------------------+ +------------------------+
|
|
|
|
.. important:: To use Ceph Block Devices with CloudStack, you must have
|
|
access to a running Ceph Storage Cluster.
|
|
|
|
CloudStack integrates with Ceph's block devices to provide CloudStack with a
|
|
back end for CloudStack's Primary Storage. The instructions below detail the
|
|
setup for CloudStack Primary Storage.
|
|
|
|
.. note:: We recommend installing with Ubuntu 13.04 or later so that
|
|
you can use package installation instead of having to compile
|
|
libvirt from source.
|
|
|
|
.. note:: Make sure the /tmp partition on your hypervisors is at least 25GB.
|
|
When deploying from a template from the first time /tmp will be used for
|
|
converting the template from QCOW2 to RAW for storage on RBD.
|
|
|
|
Installing and configuring QEMU for use with CloudStack doesn't require any
|
|
special handling. Ensure that you have a running Ceph Storage Cluster. Install
|
|
QEMU and configure it for use with Ceph; then, install ``libvirt`` version
|
|
0.9.13 or higher (you may need to compile from source) and ensure it is running
|
|
with Ceph.
|
|
|
|
#. `Install and Configure QEMU`_.
|
|
#. `Install and Configure libvirt`_ version 0.9.13 or higher.
|
|
#. Also see `KVM Hypervisor Host Installation`_.
|
|
|
|
|
|
.. note:: Raring Ringtail (13.04) will have ``libvirt`` version 0.9.13 or higher
|
|
with RBD storage pool support enabled by default.
|
|
|
|
index:: pools; CloudStack
|
|
|
|
Create a Pool
|
|
=============
|
|
|
|
By default, Ceph block devices use the ``rbd`` pool. Create a pool for
|
|
CloudStack NFS Primary Storage. Ensure your Ceph cluster is running, then create
|
|
the pool. ::
|
|
|
|
ceph osd pool create cloudstack
|
|
|
|
See `Create a Pool`_ for details on specifying the number of placement groups
|
|
for your pools, and `Placement Groups`_ for details on the number of placement
|
|
groups you should set for your pools.
|
|
|
|
Create a CephX user
|
|
===================
|
|
|
|
To access the Ceph cluster we require a CephX user which has the correct credentials
|
|
to access the ``cloudstack`` pool we just created.
|
|
|
|
Although we could use ``client.admin`` for this, it's recommended to create a user
|
|
with only access to the ``cloudstack`` pool.
|
|
|
|
ceph auth get-or-create client.cloudstack mon 'allow r' osd 'allow rwx pool=cloudstack'
|
|
|
|
Use the information returned by the command in the next step when adding the Primary Storage.
|
|
|
|
Add Primary Storage
|
|
===================
|
|
|
|
To add primary storage, refer to `Add Primary Storage (4.2.0)`_ to add a Ceph block device, the steps
|
|
include:
|
|
|
|
#. Log in to the CloudStack UI.
|
|
#. Click **Infrastructure** on the left side navigation bar.
|
|
#. Select the Zone you want to use for Primary Storage.
|
|
#. Click the **Compute** tab.
|
|
#. Select **View All** on the `Primary Storage` node in the diagram.
|
|
#. Click **Add Primary Storage**.
|
|
#. Follow the CloudStack instructions.
|
|
|
|
- For **Protocol**, select ``RBD``.
|
|
- Add cluster information (cephx is supported). Note: Do not include the ``client.`` part of the user.
|
|
- Add ``rbd`` as a tag.
|
|
|
|
|
|
Create a Disk Offering
|
|
======================
|
|
|
|
To create a new disk offering, refer to `Create a New Disk Offering (4.2.0)`_.
|
|
Create a disk offering so that it matches the ``rbd`` tag.
|
|
The ``StoragePoolAllocator`` will choose the ``rbd``
|
|
pool when searching for a suitable storage pool. If the disk offering doesn't
|
|
match the ``rbd`` tag, the ``StoragePoolAllocator`` may select the pool you
|
|
created (e.g., ``cloudstack``).
|
|
|
|
|
|
Limitations
|
|
===========
|
|
|
|
- CloudStack will only bind to one monitor (You can however create a Round Robin DNS record over multiple monitors)
|
|
- You may need to compile ``libvirt`` to use version 0.9.13 with Ubuntu.
|
|
|
|
|
|
|
|
.. _Create a Pool: ../../rados/operations/pools#createpool
|
|
.. _Placement Groups: ../../rados/operations/placement-groups
|
|
.. _Install and Configure QEMU: ../qemu-rbd
|
|
.. _Install and Configure libvirt: ../libvirt
|
|
.. _KVM Hypervisor Host Installation: http://cloudstack.apache.org/docs/en-US/Apache_CloudStack/4.2.0/html/Installation_Guide/hypervisor-kvm-install-flow.html
|
|
.. _Add Primary Storage (4.2.0): http://cloudstack.apache.org/docs/en-US/Apache_CloudStack/4.2.0/html/Admin_Guide/primary-storage-add.html
|
|
.. _Create a New Disk Offering (4.2.0): http://cloudstack.apache.org/docs/en-US/Apache_CloudStack/4.2.0/html/Admin_Guide/compute-disk-service-offerings.html#creating-disk-offerings
|