Check that cinder has access to read your ceph.conf file. I’ve had to 644 mine. From: ceph-users-bounces@xxxxxxxxxxxxxx [mailto:ceph-users-bounces@xxxxxxxxxxxxxx]
On Behalf Of bigbird Lim hi, I am trying to get ceph working with cinder. I have an existing openstack setup with one nova-controller and five compute-nodes. I have setup another three separate servers as a ceph cluster. Following the instruction at http://ceph.com/docs/master/rbd/rbd-openstack/,
I am getting this error when starting cinder-volume: 2013-12-18 09:06:49.756 12380 AUDIT cinder.service [-] Starting cinder-volume node (version 2013.2) 2013-12-18 09:06:50.286 12380 INFO cinder.openstack.common.rpc.common [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Connected to AMQP server on localhost:5672 2013-12-18 09:06:50.297 12380 INFO cinder.volume.manager [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Starting volume driver RBDDriver (1.1.0) 2013-12-18 09:06:50.316 12380 ERROR cinder.volume.drivers.rbd [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] error connecting to ceph cluster 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd Traceback (most recent call last): 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 262, in check_for_setup_error 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd with RADOSClient(self): 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 234, in __init__ 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd self.cluster, self.ioctx = driver._connect_to_rados(pool) 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 282, in _connect_to_rados 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd client.connect() 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd File "/usr/lib/python2.7/dist-packages/rados.py", line 367, in connect 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd raise make_ex(ret, "error calling connect") 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd ObjectNotFound: error calling connect 2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd 2013-12-18 09:06:50.319 12380 ERROR cinder.volume.manager [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Error encountered during initialization of driver: RBDDriver 2013-12-18 09:06:50.319 12380 ERROR cinder.volume.manager [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Bad or unexpected response from the storage volume backend API: error connecting to ceph cluster 2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager Traceback (most recent call last): 2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager File "/usr/lib/python2.7/dist-packages/cinder/volume/manager.py", line 190, in init_host 2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager self.driver.check_for_setup_error() 2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 267, in check_for_setup_error 2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager raise exception.VolumeBackendAPIException(data=""> 2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager VolumeBackendAPIException: Bad or unexpected response from the storage volume backend API: error connecting to ceph cluster 2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager This is my cinder.conf file: cat /etc/cinder/cinder.conf [DEFAULT] rootwrap_config=/etc/cinder/rootwrap.conf sql_connection = mysql://cinderUser:cinderPass@10.193.0.120/cinder api_paste_config = /etc/cinder/api-paste.ini iscsi_helper=ietadm volume_name_template = volume-%s volume_group = cinder-volumes verbose = True auth_strategy = keystone iscsi_ip_address=10.193.0.120 volume_driver=cinder.volume.drivers.rbd.RBDDriver rbd_pool=volumes glance_api_version=2 rdb_user=volumes rdb_secret_uuid=19365acb-10b4-44c9-9a28-f948e8128e91 and ceph.conf file [global] fsid = 633accd0-dd09-4d97-ab40-2aca79f44d1c mon_initial_members = ceph-1 mon_host = 10.193.0.111 auth_supported = cephx osd_journal_size = 1024 filestore_xattr_use_omap = true Thanks for the help Song |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com