2012-10-10 20:22:47 +00:00
=============================
Block Devices and OpenStack
=============================
2012-09-18 20:00:58 +00:00
2013-06-14 23:57:48 +00:00
.. index :: Ceph Block Device; OpenStack
You may use Ceph Block Device images with OpenStack through `` libvirt `` , which
2012-10-10 20:22:47 +00:00
configures the QEMU interface to `` librbd `` . Ceph stripes block device images as
2013-06-14 23:57:48 +00:00
objects across the cluster, which means that large Ceph Block Device images have
2012-10-10 20:22:47 +00:00
better performance than a standalone server!
2012-09-18 20:00:58 +00:00
2013-06-14 23:57:48 +00:00
To use Ceph Block Devices with OpenStack, you must install QEMU, `` libvirt `` ,
2013-10-31 00:21:14 +00:00
and OpenStack first. We recommend using a separate physical node for your
2012-10-10 20:22:47 +00:00
OpenStack installation. OpenStack recommends a minimum of 8GB of RAM and a
quad-core processor. The following diagram depicts the OpenStack/Ceph
technology stack.
2012-09-18 20:00:58 +00:00
.. ditaa :: +---------------------------------------------------+
| OpenStack |
2012-10-09 21:06:40 +00:00
+---------------------------------------------------+
2012-09-18 20:00:58 +00:00
| libvirt |
2012-10-09 21:06:40 +00:00
+------------------------+--------------------------+
|
| configures
v
+---------------------------------------------------+
| QEMU |
2012-09-18 20:00:58 +00:00
+---------------------------------------------------+
| librbd |
+---------------------------------------------------+
2012-10-01 21:53:57 +00:00
| librados |
2012-09-26 00:19:50 +00:00
+------------------------+-+------------------------+
| OSDs | | Monitors |
+------------------------+ +------------------------+
2012-06-18 21:29:04 +00:00
2013-12-06 14:43:37 +00:00
.. important :: To use Ceph Block Devices with OpenStack, you must have
2013-06-14 23:57:48 +00:00
access to a running Ceph Storage Cluster.
2012-06-18 21:29:04 +00:00
2013-12-06 14:43:37 +00:00
Three parts of OpenStack integrate with Ceph's block devices:
2012-09-18 20:00:58 +00:00
2012-10-10 20:22:47 +00:00
- **Images** : OpenStack Glance manages images for VMs. Images
are immutable. OpenStack treats images as binary blobs and
2013-12-06 14:43:37 +00:00
downloads them accordingly.
2012-10-01 18:39:54 +00:00
2012-10-10 20:22:47 +00:00
- **Volumes** : Volumes are block devices. OpenStack uses volumes
2013-12-06 14:43:37 +00:00
to boot VMs, or to attach volumes to running VMs. OpenStack manages
2013-10-31 00:21:14 +00:00
volumes using Cinder services.
2012-10-01 18:39:54 +00:00
2013-12-06 14:43:37 +00:00
- **Guest Disks** : Guest disks are guest operating system disks.
By default, when you boot a virtual machine,
its disk appears as a file on the filesystem of the hypervisor
(usually under `` /var/lib/nova/instances/<uuid>/ `` ). Prior OpenStack
Havana, the only way to boot a VM in Ceph was to use the boot from volume
functionality from Cinder. However, now it is possible to
directly boot every virtual machine inside Ceph without using Cinder.
This is really handy because it allows us to easily perform
maintenance operation with the live-migration process.
On the other hand, if your hypervisor dies it is also really convenient
to trigger `` nova evacuate `` and almost seamlessly run the virtual machine
somewhere else.
You can use OpenStack Glance to store images in a Ceph Block Device, and you
2013-10-31 00:21:14 +00:00
can use Cinder to boot a VM using a copy-on-write clone of an image.
2012-10-10 20:22:47 +00:00
2013-12-06 14:43:37 +00:00
The instructions below detail the setup for Glance, Cinder and Nova, although
2012-10-10 20:22:47 +00:00
they do not have to be used together. You may store images in Ceph block devices
while running VMs using a local disk, or vice versa.
2012-06-18 21:29:04 +00:00
2013-12-06 14:43:37 +00:00
.. tip :: This document describes using Ceph Block Devices with OpenStack Havana.
For earlier versions of OpenStack see
2013-10-31 00:44:24 +00:00
`Block Devices and OpenStack (Dumpling)`_ .
2013-06-14 23:57:48 +00:00
.. index :: pools; OpenStack
2012-06-18 21:29:04 +00:00
Create a Pool
2012-09-18 20:00:58 +00:00
=============
2012-10-10 20:22:47 +00:00
By default, Ceph block devices use the `` rbd `` pool. You may use any available
2013-10-31 00:21:14 +00:00
pool. We recommend creating a pool for Cinder and a pool for Glance. Ensure
2012-10-10 20:22:47 +00:00
your Ceph cluster is running, then create the pools. ::
2012-06-18 21:29:04 +00:00
2012-11-12 20:01:07 +00:00
ceph osd pool create volumes 128
ceph osd pool create images 128
2013-12-06 14:43:37 +00:00
ceph osd pool create backups 128
2012-09-18 20:00:58 +00:00
2012-10-10 20:22:47 +00:00
See `Create a Pool`_ for detail on specifying the number of placement groups for
your pools, and `Placement Groups`_ for details on the number of placement
2012-10-01 18:39:54 +00:00
groups you should set for your pools.
2012-12-03 20:22:37 +00:00
.. _Create a Pool: ../../rados/operations/pools#createpool
.. _Placement Groups: ../../rados/operations/placement-groups
2012-10-01 18:39:54 +00:00
2012-10-10 20:22:47 +00:00
Configure OpenStack Ceph Clients
2012-10-01 18:39:54 +00:00
================================
2013-12-06 14:43:37 +00:00
The nodes running `` glance-api `` , `` cinder-volume `` , `` nova-compute `` and `` cinder-backup `` act as Ceph clients. Each
2013-10-31 00:21:14 +00:00
requires the `` ceph.conf `` file::
2012-10-01 18:39:54 +00:00
2012-10-10 20:22:47 +00:00
ssh {your-openstack-server} sudo tee /etc/ceph/ceph.conf </etc/ceph/ceph.conf
2012-10-01 18:39:54 +00:00
Install Ceph client packages
----------------------------
2013-12-06 14:43:37 +00:00
On the `` glance-api `` node, you'll need the Python bindings for `` librbd `` ::
2012-10-01 18:39:54 +00:00
sudo apt-get install python-ceph
2013-10-31 00:21:14 +00:00
sudo yum install python-ceph
2012-10-01 18:39:54 +00:00
2013-12-06 14:43:37 +00:00
On the `` nova-compute `` , `` cinder-backup `` and on the `` cinder-volume `` node, use both
the Python bindings and the client command line tools::
2012-10-01 18:39:54 +00:00
sudo apt-get install ceph-common
2013-10-31 00:21:14 +00:00
sudo yum install ceph
2012-10-01 18:39:54 +00:00
2012-10-10 20:22:47 +00:00
Setup Ceph Client Authentication
2012-10-01 18:39:54 +00:00
--------------------------------
2012-11-13 16:14:09 +00:00
If you have `cephx authentication`_ enabled, create a new user for Nova/Cinder
2013-10-31 00:21:14 +00:00
and Glance. Execute the following::
2012-12-30 07:57:01 +00:00
2013-12-06 14:43:37 +00:00
ceph auth get-or-create client.cinder mon 'allow r' osd 'allow class-read object_prefix rbd_children, allow rwx pool=volumes, allow rx pool=images'
ceph auth get-or-create client.glance mon 'allow r' osd 'allow class-read object_prefix rbd_children, allow rwx pool=images'
ceph auth get-or-create client.cinder-backup mon 'allow r' osd 'allow class-read object_prefix rbd_children, allow rwx pool=backups'
2012-11-13 16:14:09 +00:00
2013-12-06 14:43:37 +00:00
Add the keyrings for `` client.cinder `` , `` client.glance `` , and `` client.cinder-backup `` to the appropriate
2013-10-31 00:21:14 +00:00
nodes and change their ownership::
2012-10-01 18:39:54 +00:00
2013-12-06 14:43:37 +00:00
ceph auth get-or-create client.glance | ssh {your-glance-api-server} sudo tee /etc/ceph/ceph.client.glance.keyring
ssh {your-glance-api-server} sudo chown glance:glance /etc/ceph/ceph.client.glance.keyring
ceph auth get-or-create client.cinder | ssh {your-volume-server} sudo tee /etc/ceph/ceph.client.cinder.keyring
ssh {your-cinder-volume-server} sudo chown cinder:cinder /etc/ceph/ceph.client.cinder.keyring
ceph auth get-or-create client.cinder-backup | ssh {your-cinder-backup-server} sudo tee /etc/ceph/ceph.client.cinder-backup.keyring
ssh {your-cinder-backup-server} sudo chown cinder:cinder /etc/ceph/ceph.client.cinder-backup.keyring
2012-06-18 21:29:04 +00:00
2013-12-06 14:43:37 +00:00
Nodes running `` nova-compute `` need the keyring file for the `` nova-compute `` process. They
also need to store the secret key of the `` client.cinder `` user in `` libvirt `` . The libvirt
process needs it to access the cluster while attaching a block device from Cinder.
Create a temporary copy of the secret
2013-10-31 00:21:14 +00:00
key on the nodes running `` nova-compute `` ::
2012-09-18 20:00:58 +00:00
2013-12-06 14:43:37 +00:00
ceph auth get-key client.cinder | ssh {your-compute-node} tee client.cinder.key
2012-06-18 21:29:04 +00:00
2013-10-31 00:21:14 +00:00
Then, on the compute nodes, add the secret key to `` libvirt `` and remove the
2013-03-12 21:25:44 +00:00
temporary copy of the key::
2012-06-18 21:29:04 +00:00
2013-12-06 14:43:37 +00:00
uuidgen
457eb676-33da-42ec-9a8c-9293d545c337
2012-10-01 18:39:54 +00:00
cat > secret.xml <<EOF
<secret ephemeral='no' private='no'>
2013-12-06 14:43:37 +00:00
<uuid>457eb676-33da-42ec-9a8c-9293d545c337</uuid>
2012-10-01 18:39:54 +00:00
<usage type='ceph'>
2013-12-06 14:43:37 +00:00
<name>client.cinder secret</name>
2012-10-01 18:39:54 +00:00
</usage>
</secret>
EOF
sudo virsh secret-define --file secret.xml
2013-12-06 14:43:37 +00:00
Secret 457eb676-33da-42ec-9a8c-9293d545c337 created
sudo virsh secret-set-value --secret 457eb676-33da-42ec-9a8c-9293d545c337 --base64 $(cat client.cinder.key) && rm client.cinder.key secret.xml
2012-09-18 20:00:58 +00:00
2012-10-10 20:22:47 +00:00
Save the uuid of the secret for configuring `` nova-compute `` later.
2012-09-18 20:00:58 +00:00
2013-12-06 14:43:37 +00:00
.. important :: You don't necessarily need the UUID on all the compute nodes. However from a platform consistency perspective it's better to keep the same UUID.
2012-12-03 20:22:37 +00:00
.. _cephx authentication: ../../rados/operations/authentication
2012-10-09 20:53:37 +00:00
2012-10-01 18:39:54 +00:00
2012-10-10 20:22:47 +00:00
Configure OpenStack to use Ceph
===============================
2012-10-01 18:39:54 +00:00
Configuring Glance
------------------
2012-10-10 20:22:47 +00:00
Glance can use multiple back ends to store images. To use Ceph block devices by
2012-10-01 18:39:54 +00:00
default, edit `` /etc/glance/glance-api.conf `` and add::
default_store=rbd
2013-12-06 14:43:37 +00:00
rbd_store_user=glance
2012-10-01 18:39:54 +00:00
rbd_store_pool=images
2013-10-31 00:53:37 +00:00
If want to enable copy-on-write cloning of images into
2013-10-31 00:44:24 +00:00
volumes, also add::
2013-12-06 14:43:37 +00:00
show_image_direct_url=True
2013-10-31 00:44:24 +00:00
Note that this exposes the back end location via Glance’ s API, so the endpoint
with this option enabled should not be publicly accessible.
2012-10-10 20:22:47 +00:00
2013-10-31 00:21:14 +00:00
Configuring Cinder
------------------
2012-10-10 20:22:47 +00:00
OpenStack requires a driver to interact with Ceph block devices. You must also
2013-10-31 00:21:14 +00:00
specify the pool name for the block device. On your OpenStack node,
edit `` /etc/cinder/cinder.conf `` by adding::
2013-05-30 21:17:35 +00:00
volume_driver=cinder.volume.drivers.rbd.RBDDriver
rbd_pool=volumes
2013-12-06 14:43:37 +00:00
rbd_ceph_conf=/etc/ceph/ceph.conf
rbd_flatten_volume_from_snapshot=false
rbd_max_clone_depth=5
2013-05-30 21:17:35 +00:00
glance_api_version=2
2012-06-18 21:29:04 +00:00
2012-10-01 18:39:54 +00:00
If you're using `cephx authentication`_ , also configure the user and
2013-12-06 14:43:37 +00:00
uuid of the secret you added to `` libvirt `` as documented earlier::
rbd_user=cinder
rbd_secret_uuid=457eb676-33da-42ec-9a8c-9293d545c337
Configuring Cinder Backup
-------------------------
OpenStack Cinder Backup requires a specific daemon so don't forget to install it.
On your Cinder Backup node, edit `` /etc/cinder/cinder.conf `` and add::
backup_driver=cinder.backup.drivers.ceph
backup_ceph_conf=/etc/ceph/ceph.conf
backup_ceph_user=cinder-backup
backup_ceph_chunk_size=134217728
backup_ceph_pool=backups
backup_ceph_stripe_unit=0
backup_ceph_stripe_count=0
restore_discard_excess_bytes=true
2012-10-01 18:39:54 +00:00
2012-09-18 20:00:58 +00:00
2013-12-06 14:43:37 +00:00
Configuring Nova
----------------
In order to boot all the virtual machines directly into Ceph Nova must be configured.
On every Compute nodes, edit `` /etc/nova/nova.conf `` and add::
libvirt_images_type=rbd
libvirt_images_rbd_pool=volumes
libvirt_images_rbd_ceph_conf=/etc/ceph/ceph.conf
rbd_user=cinder
rbd_secret_uuid=457eb676-33da-42ec-9a8c-9293d545c337
It is also a good practice to disable any file injection.
Usually, while booting an instance Nova attempts to open the rootfs of the virtual machine.
Then, it injects directly into the filesystem things like: password, ssh keys etc...
At this point, it is better to rely on the metadata service and cloud-init.
On every Compute nodes, edit `` /etc/nova/nova.conf `` and add::
libvirt_inject_password=false
libvirt_inject_key=false
libvirt_inject_partition=-2
2012-11-13 16:14:09 +00:00
2012-10-10 20:22:47 +00:00
2012-09-26 00:21:52 +00:00
Restart OpenStack
2012-09-18 20:00:58 +00:00
=================
2012-10-10 20:22:47 +00:00
To activate the Ceph block device driver and load the block device pool name
2013-12-06 14:43:37 +00:00
into the configuration, you must restart OpenStack. Thus, for Debian based systems
execute these commands on the appropriate nodes::
2012-06-18 21:29:04 +00:00
2013-12-06 14:43:37 +00:00
sudo glance-control api restart
2012-10-09 20:53:37 +00:00
sudo service nova-compute restart
sudo service cinder-volume restart
2013-12-06 14:43:37 +00:00
sudo service cinder-backup restart
For Red Hat based systems execute::
sudo service openstack-glance-api restart
sudo service openstack-nova-compute restart
sudo service openstack-cinder-volume restart
sudo service openstack-cinder-backup restart
2012-06-18 21:29:04 +00:00
2013-12-06 14:43:37 +00:00
Once OpenStack is up and running, you should be able to create a volume
and boot from it.
2012-10-10 20:22:47 +00:00
2012-10-01 18:39:54 +00:00
2012-10-10 20:22:47 +00:00
Booting from a Block Device
===========================
2012-10-01 18:39:54 +00:00
2013-10-31 00:21:14 +00:00
You can create a volume from an image using the Cinder command line tool::
2012-10-01 18:39:54 +00:00
cinder create --image-id {id of image} --display-name {name of volume} {size of volume}
2012-10-24 23:19:21 +00:00
Note that image must be raw format. You can use `qemu-img`_ to convert
from one format to another, i.e.::
qemu-img convert -f qcow2 -O raw precise-cloudimg.img precise-cloudimg.raw
2013-10-31 00:21:14 +00:00
When Glance and Cinder are both using Ceph block devices, the image is a
2012-10-01 18:39:54 +00:00
copy-on-write clone, so volume creation is very fast.
2012-10-10 20:22:47 +00:00
In the OpenStack dashboard you can then boot from that volume by launching a new
instance, choosing the image that you created the volume from, and selecting
'boot from volume' and the volume you created.
2012-10-24 23:19:21 +00:00
.. _qemu-img: ../qemu-rbd/#running-qemu-with-rbd
2013-12-06 14:43:37 +00:00
.. _Block Devices and OpenStack (Dumpling): http://ceph.com/docs/dumpling/rbd/rbd-openstack