I've been trying to take advantage of the code additions made by Josh Durgin to OpenStack Folsom for combining boot-from-volume and Ceph RBD. First off, nice work Josh! I'm hoping you folks can help me out with something strange I am seeing. The question may be more OpenStack related than Ceph, though, but hear me out first. I created a new volume (to use for boot-from-volume) from an existing image like so: #cinder create --display-name uec-test-vol --image-id 699137a2-a864-4a87-98fa-1684d7677044 5 This completes just fine. Later I try to boot from it, that fails. Cutting to the chase, here is why: kvm: -drive file=rbd:nova-volume/volume-9f4e4b70-7fbb-4d81-b912-b1c6fcf86c8b,if=none,id=drive-virtio-disk0,format=raw,cache=none: error reading header from volume-9f4e4b70-7fbb-4d81-b912-b1c6fcf86c8b kvm: -drive file=rbd:nova-volume/volume-9f4e4b70-7fbb-4d81-b912-b1c6fcf86c8b,if=none,id=drive-virtio-disk0,format=raw,cache=none: could not open disk image rbd:nova-volume/volume-9f4e4b70-7fbb-4d81-b912-b1c6fcf86c8b: No such file or directory It's weird that creating the volume was successful, but that KVM can't read it. Poking around a bit more, it was clear why: # rbd -n client.novavolume --pool nova-volume ls <returns nothing> # rbd -n client.novavolume ls volume-9f4e4b70-7fbb-4d81-b912-b1c6fcf86c8b Okay, the volume is the "rbd" pool! That's really weird, though. Here is the my nova.conf entries: volume_driver=nova.volume.driver.RBDDriver rbd_pool=nova-volume rbd_user=novavolume AND, here are the log entries from nova-volume.log (cleaned up a little): rbd create --pool nova-volume --size 5120 volume-9f4e4b70-7fbb-4d81-b912-b1c6fcf86c8b rbd rm --pool nova-volume volume-9f4e4b70-7fbb-4d81-b912-b1c6fcf86c8b rbd import --pool nova-volume /tmp/tmplQUwzt volume-9f4e4b70-7fbb-4d81-b912-b1c6fcf86c8b I'm not sure why it goes create/delete/import, but regardless all of that worked. More importantly, all these commands used --pool nova-volume. So how the heck did that RBD end up in the "rbd" pool instead of the "nova-volume" pool? Any ideas? Before I hit "send", I figured I should at least test this myself. Watch this: #rbd create -n client.novavolume --pool nova-volume --size 1024 test # rbd ls -n client.novavolume --pool nova-volume test # rbd export -n client.novavolume --pool nova-volume test /tmp/test Exporting image: 100% complete...done. # rbd rm -n client.novavolume --pool nova-volume test Removing image: 100% complete...done. # rbd import -n client.novavolume --pool nova-volume /tmp/test test Importing image: 100% complete...done. # rbd ls -n client.novavolume --pool nova-volume # rbd ls -n client.novavolume --pool rbd test So it seems that "rbd import" doesn't honor the --pool argument? I am using 0.53 on the backend, but my client is 0.48.2. I'll upgrade that and see if that makes a different. - Travis -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html