============================= Block Devices and OpenStack ============================= .. index:: Ceph Block Device; OpenStack You can attach Ceph Block Device images to OpenStack instances through ``libvirt``, which configures the QEMU interface to ``librbd``. Ceph stripes block volumes across multiple OSDs within the cluster, which means that large volumes can realize better performance than local drives on a standalone server! To use Ceph Block Devices with OpenStack, you must install QEMU, ``libvirt``, and OpenStack first. We recommend using a separate physical node for your OpenStack installation. OpenStack recommends a minimum of 8GB of RAM and a quad-core processor. The following diagram depicts the OpenStack/Ceph technology stack. .. ditaa:: +---------------------------------------------------+ | OpenStack | +---------------------------------------------------+ | libvirt | +------------------------+--------------------------+ | | configures v +---------------------------------------------------+ | QEMU | +---------------------------------------------------+ | librbd | +---------------------------------------------------+ | librados | +------------------------+-+------------------------+ | OSDs | | Monitors | +------------------------+ +------------------------+ .. important:: To use Ceph Block Devices with OpenStack, you must have access to a running Ceph Storage Cluster. Three parts of OpenStack integrate with Ceph's block devices: - **Images**: OpenStack Glance manages images for VMs. Images are immutable. OpenStack treats images as binary blobs and downloads them accordingly. - **Volumes**: Volumes are block devices. OpenStack uses volumes to boot VMs, or to attach volumes to running VMs. OpenStack manages volumes using Cinder services. - **Guest Disks**: Guest disks are guest operating system disks. By default, when you boot a virtual machine, its disk appears as a file on the file system of the hypervisor (usually under ``/var/lib/nova/instances//``). Prior to OpenStack Havana, the only way to boot a VM in Ceph was to use the boot-from-volume functionality of Cinder. However, now it is possible to boot every virtual machine inside Ceph directly without using Cinder, which is advantageous because it allows you to perform maintenance operations easily with the live-migration process. Additionally, if your hypervisor dies it is also convenient to trigger ``nova evacuate`` and reinstate the virtual machine elsewhere almost seamlessly. In doing so, :ref:`exclusive locks ` prevent multiple compute nodes from concurrently accessing the guest disk. You can use OpenStack Glance to store images as Ceph Block Devices, and you can use Cinder to boot a VM using a copy-on-write clone of an image. The instructions below detail the setup for Glance, Cinder and Nova, although they do not have to be used together. You may store images in Ceph block devices while running VMs using a local disk, or vice versa. .. important:: Using QCOW2 for hosting a virtual machine disk is NOT recommended. If you want to boot virtual machines in Ceph (ephemeral backend or boot from volume), please use the ``raw`` image format within Glance. .. index:: pools; OpenStack Create a Pool ============= By default, Ceph block devices live within the ``rbd`` pool. You may use any suitable pool by specifying it explicitly. We recommend creating a pool for Cinder and a pool for Glance. Ensure your Ceph cluster is running, then create the pools. :: ceph osd pool create volumes ceph osd pool create images ceph osd pool create backups ceph osd pool create vms See `Create a Pool`_ for detail on specifying the number of placement groups for your pools, and `Placement Groups`_ for details on the number of placement groups you should set for your pools. Newly created pools must be initialized prior to use. Use the ``rbd`` tool to initialize the pools:: rbd pool init volumes rbd pool init images rbd pool init backups rbd pool init vms .. _Create a Pool: ../../rados/operations/pools#createpool .. _Placement Groups: ../../rados/operations/placement-groups Configure OpenStack Ceph Clients ================================ The nodes running ``glance-api``, ``cinder-volume``, ``nova-compute`` and ``cinder-backup`` act as Ceph clients. Each requires the ``ceph.conf`` file:: ssh {your-openstack-server} sudo tee /etc/ceph/ceph.conf secret.xml < 457eb676-33da-42ec-9a8c-9293d545c337 client.cinder secret EOF sudo virsh secret-define --file secret.xml Secret 457eb676-33da-42ec-9a8c-9293d545c337 created sudo virsh secret-set-value --secret 457eb676-33da-42ec-9a8c-9293d545c337 --base64 $(cat client.cinder.key) && rm client.cinder.key secret.xml Save the uuid of the secret for configuring ``nova-compute`` later. .. important:: You don't necessarily need the UUID on all the compute nodes. However from a platform consistency perspective, it's better to keep the same UUID. .. _cephx authentication: ../../rados/configuration/auth-config-ref/#enabling-disabling-cephx Configure OpenStack to use Ceph =============================== Configuring Glance ------------------ Glance can use multiple back ends to store images. To use Ceph block devices by default, configure Glance like the following. Kilo and after ~~~~~~~~~~~~~~ Edit ``/etc/glance/glance-api.conf`` and add under the ``[glance_store]`` section:: [glance_store] stores = rbd default_store = rbd rbd_store_pool = images rbd_store_user = glance rbd_store_ceph_conf = /etc/ceph/ceph.conf rbd_store_chunk_size = 8 For more information about the configuration options available in Glance please refer to the OpenStack Configuration Reference: http://docs.openstack.org/. Enable copy-on-write cloning of images ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Note that this exposes the back end location via Glance's API, so the endpoint with this option enabled should not be publicly accessible. Any OpenStack version except Mitaka ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ If you want to enable copy-on-write cloning of images, also add under the ``[DEFAULT]`` section:: show_image_direct_url = True Disable cache management (any OpenStack version) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Disable the Glance cache management to avoid images getting cached under ``/var/lib/glance/image-cache/``, assuming your configuration file has ``flavor = keystone+cachemanagement``:: [paste_deploy] flavor = keystone Image properties ~~~~~~~~~~~~~~~~ We recommend to use the following properties for your images: - ``hw_scsi_model=virtio-scsi``: add the virtio-scsi controller and get better performance and support for discard operation - ``hw_disk_bus=scsi``: connect every cinder block devices to that controller - ``hw_qemu_guest_agent=yes``: enable the QEMU guest agent - ``os_require_quiesce=yes``: send fs-freeze/thaw calls through the QEMU guest agent Configuring Cinder ------------------ OpenStack requires a driver to interact with Ceph block devices. You must also specify the pool name for the block device. On your OpenStack node, edit ``/etc/cinder/cinder.conf`` by adding:: [DEFAULT] ... enabled_backends = ceph glance_api_version = 2 ... [ceph] volume_driver = cinder.volume.drivers.rbd.RBDDriver volume_backend_name = ceph rbd_pool = volumes rbd_ceph_conf = /etc/ceph/ceph.conf rbd_flatten_volume_from_snapshot = false rbd_max_clone_depth = 5 rbd_store_chunk_size = 4 rados_connect_timeout = -1 If you are using `cephx authentication`_, also configure the user and uuid of the secret you added to ``libvirt`` as documented earlier:: [ceph] ... rbd_user = cinder rbd_secret_uuid = 457eb676-33da-42ec-9a8c-9293d545c337 Note that if you are configuring multiple cinder back ends, ``glance_api_version = 2`` must be in the ``[DEFAULT]`` section. Configuring Cinder Backup ------------------------- OpenStack Cinder Backup requires a specific daemon so don't forget to install it. On your Cinder Backup node, edit ``/etc/cinder/cinder.conf`` and add:: backup_driver = cinder.backup.drivers.ceph backup_ceph_conf = /etc/ceph/ceph.conf backup_ceph_user = cinder-backup backup_ceph_chunk_size = 134217728 backup_ceph_pool = backups backup_ceph_stripe_unit = 0 backup_ceph_stripe_count = 0 restore_discard_excess_bytes = true Configuring Nova to attach Ceph RBD block device ------------------------------------------------ In order to attach Cinder devices (either normal block or by issuing a boot from volume), you must tell Nova (and libvirt) which user and UUID to refer to when attaching the device. libvirt will refer to this user when connecting and authenticating with the Ceph cluster. :: [libvirt] ... rbd_user = cinder rbd_secret_uuid = 457eb676-33da-42ec-9a8c-9293d545c337 These two flags are also used by the Nova ephemeral back end. Configuring Nova ---------------- In order to boot virtual machines directly from Ceph volumes, you must configure the ephemeral backend for Nova. It is recommended to enable the RBD cache in your Ceph configuration file; this has been enabled by default since the Giant release. Moreover, enabling the client admin socket allows the collection of metrics and can be invaluable for troubleshooting. This socket can be accessed on the hypvervisor (Nova compute) node:: ceph daemon /var/run/ceph/ceph-client.cinder.19195.32310016.asok help To enable RBD cache and admin sockets, ensure that on each hypervisor's ``ceph.conf`` contains:: [client] rbd cache = true rbd cache writethrough until flush = true admin socket = /var/run/ceph/guests/$cluster-$type.$id.$pid.$cctid.asok log file = /var/log/qemu/qemu-guest-$pid.log rbd concurrent management ops = 20 Configure permissions for these directories:: mkdir -p /var/run/ceph/guests/ /var/log/qemu/ chown qemu:libvirtd /var/run/ceph/guests /var/log/qemu/ Note that user ``qemu`` and group ``libvirtd`` can vary depending on your system. The provided example works for RedHat based systems. .. tip:: If your virtual machine is already running you can simply restart it to enable the admin socket Restart OpenStack ================= To activate the Ceph block device driver and load the block device pool name into the configuration, you must restart the related OpenStack services. For Debian based systems execute these commands on the appropriate nodes:: sudo glance-control api restart sudo service nova-compute restart sudo service cinder-volume restart sudo service cinder-backup restart For Red Hat based systems execute:: sudo service openstack-glance-api restart sudo service openstack-nova-compute restart sudo service openstack-cinder-volume restart sudo service openstack-cinder-backup restart Once OpenStack is up and running, you should be able to create a volume and boot from it. Booting from a Block Device =========================== You can create a volume from an image using the Cinder command line tool:: cinder create --image-id {id of image} --display-name {name of volume} {size of volume} You can use `qemu-img`_ to convert from one format to another. For example:: qemu-img convert -f {source-format} -O {output-format} {source-filename} {output-filename} qemu-img convert -f qcow2 -O raw precise-cloudimg.img precise-cloudimg.raw When Glance and Cinder are both using Ceph block devices, the image is a copy-on-write clone, so new volumes are created quickly. In the OpenStack dashboard, you can boot from that volume by performing the following steps: #. Launch a new instance. #. Choose the image associated to the copy-on-write clone. #. Select 'boot from volume'. #. Select the volume you created. .. _qemu-img: ../qemu-rbd/#running-qemu-with-rbd