Error connecting to ceph cluster in openstack cinder

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



hi,
I am trying to get ceph working with cinder. I have an existing openstack setup with one nova-controller and five compute-nodes. I have setup another three separate servers as a ceph cluster. Following the instruction at http://ceph.com/docs/master/rbd/rbd-openstack/, I am getting this error when starting cinder-volume:


2013-12-18 09:06:49.756 12380 AUDIT cinder.service [-] Starting cinder-volume node (version 2013.2)
2013-12-18 09:06:50.286 12380 INFO cinder.openstack.common.rpc.common [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Connected to AMQP server on localhost:5672
2013-12-18 09:06:50.297 12380 INFO cinder.volume.manager [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Starting volume driver RBDDriver (1.1.0)
2013-12-18 09:06:50.316 12380 ERROR cinder.volume.drivers.rbd [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] error connecting to ceph cluster
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd Traceback (most recent call last):
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd   File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 262, in check_for_setup_error
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd     with RADOSClient(self):
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd   File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 234, in __init__
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd     self.cluster, self.ioctx = driver._connect_to_rados(pool)
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd   File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 282, in _connect_to_rados
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd     client.connect()
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd   File "/usr/lib/python2.7/dist-packages/rados.py", line 367, in connect
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd     raise make_ex(ret, "error calling connect")
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd ObjectNotFound: error calling connect
2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd
2013-12-18 09:06:50.319 12380 ERROR cinder.volume.manager [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Error encountered during initialization of driver: RBDDriver
2013-12-18 09:06:50.319 12380 ERROR cinder.volume.manager [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Bad or unexpected response from the storage volume backend API: error connecting to ceph cluster
2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager Traceback (most recent call last):
2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager   File "/usr/lib/python2.7/dist-packages/cinder/volume/manager.py", line 190, in init_host
2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager     self.driver.check_for_setup_error()
2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager   File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 267, in check_for_setup_error
2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager     raise exception.VolumeBackendAPIException(data="">
2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager VolumeBackendAPIException: Bad or unexpected response from the storage volume backend API: error connecting to ceph cluster
2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager

This is my cinder.conf file:
cat /etc/cinder/cinder.conf
[DEFAULT]
rootwrap_config=/etc/cinder/rootwrap.conf
sql_connection = mysql://cinderUser:cinderPass@10.193.0.120/cinder
api_paste_config = /etc/cinder/api-paste.ini
iscsi_helper=ietadm
volume_name_template = volume-%s
volume_group = cinder-volumes
verbose = True
auth_strategy = keystone
iscsi_ip_address=10.193.0.120
volume_driver=cinder.volume.drivers.rbd.RBDDriver
rbd_pool=volumes
glance_api_version=2
rdb_user=volumes
rdb_secret_uuid=19365acb-10b4-44c9-9a28-f948e8128e91

and ceph.conf file
[global]
fsid = 633accd0-dd09-4d97-ab40-2aca79f44d1c
mon_initial_members = ceph-1
mon_host = 10.193.0.111
auth_supported = cephx
osd_journal_size = 1024
filestore_xattr_use_omap = true

Thanks for the help

Song

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux