Written in conjunction with scuttlemonkey (Patrick McGarry), of the Ceph Project, and cross-posted on the Ceph blog.
scuttlemonkey: “The Ceph team has been extremely blessed with the number of new people who choose to become involved with our community in some way. Even more exciting are the sheer numbers of people committing code and integration work, and the folks from OpenNebula are a great example of this in action.
At the end of February, one of the OpenNebula developers reached out to let us know that their integration work with Ceph was nearly complete. Below you can find a brief overview of how Ceph behaves in an OpenNebula environment, as well as a link to how to get it set up. Read on for details!”
OpenNebula continues with its growing support of new storage technologies. OpenNebula 4.0 comes with a fresh integration with Ceph, an impressive distributed object store and file system.
OpenNebula provides an interface for Ceph RBDs (RADOS Block Device), which allows registering images directly in a Ceph pool, and running VMs using that backend.
There is an extensive Ceph for OpenNebula guide, but it can be summarized as follows:
- OpenNebula worker nodes should be part of a working Ceph cluster.
- The ”one” Ceph pool should be available (the name is configurable).
- Use Libvirt/KVM as the hypervisor. Xen is not yet supported.
Once we have that up and running using it is extremely simple!
- Make sure we have the ”one” Ceph pool$ ceph osd lspools 0 data,1 metadata,2 rbd,3 one,
- Create a Ceph datastore$ cat ceph.one NAME = ceph DS_MAD = ceph TM_MAD = ceph DISK_TYPE = RBD HOST = ceph0 POOL_NAME = one SAFE_DIRS="/" $ onedatastore create ceph.one ID: 101
- Register an image$ oneimage create --name centos-6.4 --path /tmp/centos-6.4.img -d ceph ID: 4
- Run your VM$ onevm create --name centos-6.4 --cpu 1 --memory 768 --nic 1 --disk centos-6.4 --nic net_172 ID: 10
What happens behind the scenes is that OpenNebula interacts with the Ceph cluster and clones the base image. The Libvirt/KVM deployment file uses that clone image as the OS:<disk type='network' device='disk'> <source protocol='rbd' name='one/one-4-10-0' /> <target dev='hdb' bus='ide'/> </disk>
All the image handling and manipulation (cloning, renaming, removing, etc…) is performed in a specific server defined in the datastore template, in our case ”HOST = ceph0”, using the ”rbd” capabilities of ”qemu-img” the
registration of new images.
scuttlemonkey: “Thanks to the OpenNebula team for hammering out this integration, we love to see new use cases for Ceph!
If you have your own use case story, we would love to hear about it (and share it with the world). Feel free to drop us your thoughts in the form of a suggested blog title, raw data that you wish us to write some prose about, or a fully formed blog post that we can push out to the community. Thanks, and happy Ceph-ing.”