Can you send your ceph.conf too?
Is /etc/ceph/ceph.conf present? Is the key of user volume present too? –––– Sébastien Han Cloud Engineer "Always give 100%. Unless you're giving blood."
Phone : +33 (0)1 49 70 99 72 – Mobile : +33 (0)6 52 84 44 70 Address : 10, rue de la Victoire – 75009 Paris
On Jul 23, 2013, at 5:39 AM, johnu < johnugeorge109@xxxxxxxxx> wrote: Hi, I have a three node ceph cluster. ceph -w says health ok . I have openstack in the same cluster and trying to map cinder and glance onto rbd.
I have followed steps as given in http://ceph.com/docs/next/rbd/rbd-openstack/
New Settings that is added in cinder.conf for three files
volume_driver=cinder.volume.drivers.rbd.RBDDriver rbd_pool=volumes glance_api_version=2 rbd_user=volumes rbd_secret_uuid=62d0b384-50ad-2e17-15ed-66bfeda40252 ( different for each node)
LOGS seen when I run ./rejoin.sh
2013-07-22 20:35:01.900 INFO cinder.service [-] Starting 1 workers 2013-07-22 20:35:01.909 INFO cinder.service [-] Started child 2290 2013-07-22 20:35:01.965 AUDIT cinder.service [-] Starting cinder-volume node (version 2013.2) 2013-07-22 20:35:02.129 ERROR cinder.volume.drivers.rbd [req-d3bc2e86-e9db-40e8-bcdb-08c609ce44c3 None None] error connecting to ceph cluster 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd Traceback (most recent call last): 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd File "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 243, in check_for_setup_error 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd with RADOSClient(self): 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd File "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 215, in __init__ 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd self.cluster, self.ioctx = driver._connect_to_rados(pool) 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd File "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 263, in _connect_to_rados 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd client.connect() 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd File "/usr/lib/python2.7/dist-packages/rados.py", line 192, in connect 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd raise make_ex(ret, "error calling connect") 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd ObjectNotFound: error calling connect 2013-07-22 20:35:02.129 TRACE cinder.volume.drivers.rbd 2013-07-22 20:35:02.149 ERROR cinder.service [req-d3bc2e86-e9db-40e8-bcdb-08c609ce44c3 None None] Unhandled exception 2013-07-22 20:35:02.149 TRACE cinder.service Traceback (most recent call last): 2013-07-22 20:35:02.149 TRACE cinder.service File "/opt/stack/cinder/cinder/service.py", line 228, in _start_child 2013-07-22 20:35:02.149 TRACE cinder.service self._child_process(wrap.server) 2013-07-22 20:35:02.149 TRACE cinder.service File "/opt/stack/cinder/cinder/service.py", line 205, in _child_process 2013-07-22 20:35:02.149 TRACE cinder.service launcher.run_server(server) 2013-07-22 20:35:02.149 TRACE cinder.service File "/opt/stack/cinder/cinder/service.py", line 96, in run_server 2013-07-22 20:35:02.149 TRACE cinder.service server.start() 2013-07-22 20:35:02.149 TRACE cinder.service File "/opt/stack/cinder/cinder/service.py", line 359, in start 2013-07-22 20:35:02.149 TRACE cinder.service self.manager.init_host() 2013-07-22 20:35:02.149 TRACE cinder.service File "/opt/stack/cinder/cinder/volume/manager.py", line 139, in init_host 2013-07-22 20:35:02.149 TRACE cinder.service self.driver.check_for_setup_error() 2013-07-22 20:35:02.149 TRACE cinder.service File "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 248, in check_for_setup_error 2013-07-22 20:35:02.149 TRACE cinder.service raise exception.VolumeBackendAPIException(data="">2013-07-22 20:35:02.149 TRACE cinder.service VolumeBackendAPIException: Bad or unexpected response from the storage volume backend API: error connecting to ceph cluster 2013-07-22 20:35:02.149 TRACE cinder.service 2013-07-22 20:35:02.191 INFO cinder.service [-] Child 2290 exited with status 2 2013-07-22 20:35:02.192 INFO cinder.service [-] _wait_child 1 2013-07-22 20:35:02.193 INFO cinder.service [-] wait wrap.failed True
Can someone help me with some debug points and solve it ? _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
|