Re: Error connecting to ceph cluster in openstack cinder

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Check that cinder has access to read your ceph.conf file. I’ve had to 644 mine.

 

From: ceph-users-bounces@xxxxxxxxxxxxxx [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of bigbird Lim
Sent: Wednesday, December 18, 2013 10:19 AM
To: ceph-users@xxxxxxxxxxxxxx
Subject: Error connecting to ceph cluster in openstack cinder

 

hi,

I am trying to get ceph working with cinder. I have an existing openstack setup with one nova-controller and five compute-nodes. I have setup another three separate servers as a ceph cluster. Following the instruction at http://ceph.com/docs/master/rbd/rbd-openstack/, I am getting this error when starting cinder-volume:

 

 

2013-12-18 09:06:49.756 12380 AUDIT cinder.service [-] Starting cinder-volume node (version 2013.2)

2013-12-18 09:06:50.286 12380 INFO cinder.openstack.common.rpc.common [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Connected to AMQP server on localhost:5672

2013-12-18 09:06:50.297 12380 INFO cinder.volume.manager [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Starting volume driver RBDDriver (1.1.0)

2013-12-18 09:06:50.316 12380 ERROR cinder.volume.drivers.rbd [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] error connecting to ceph cluster

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd Traceback (most recent call last):

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd   File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 262, in check_for_setup_error

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd     with RADOSClient(self):

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd   File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 234, in __init__

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd     self.cluster, self.ioctx = driver._connect_to_rados(pool)

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd   File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 282, in _connect_to_rados

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd     client.connect()

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd   File "/usr/lib/python2.7/dist-packages/rados.py", line 367, in connect

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd     raise make_ex(ret, "error calling connect")

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd ObjectNotFound: error calling connect

2013-12-18 09:06:50.316 12380 TRACE cinder.volume.drivers.rbd

2013-12-18 09:06:50.319 12380 ERROR cinder.volume.manager [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Error encountered during initialization of driver: RBDDriver

2013-12-18 09:06:50.319 12380 ERROR cinder.volume.manager [req-925fa7e8-1ccf-474d-a3a8-646e0f9ec93e None None] Bad or unexpected response from the storage volume backend API: error connecting to ceph cluster

2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager Traceback (most recent call last):

2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager   File "/usr/lib/python2.7/dist-packages/cinder/volume/manager.py", line 190, in init_host

2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager     self.driver.check_for_setup_error()

2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager   File "/usr/lib/python2.7/dist-packages/cinder/volume/drivers/rbd.py", line 267, in check_for_setup_error

2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager     raise exception.VolumeBackendAPIException(data="">

2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager VolumeBackendAPIException: Bad or unexpected response from the storage volume backend API: error connecting to ceph cluster

2013-12-18 09:06:50.319 12380 TRACE cinder.volume.manager

 

This is my cinder.conf file:

cat /etc/cinder/cinder.conf

[DEFAULT]

rootwrap_config=/etc/cinder/rootwrap.conf

api_paste_config = /etc/cinder/api-paste.ini

iscsi_helper=ietadm

volume_name_template = volume-%s

volume_group = cinder-volumes

verbose = True

auth_strategy = keystone

iscsi_ip_address=10.193.0.120

volume_driver=cinder.volume.drivers.rbd.RBDDriver

rbd_pool=volumes

glance_api_version=2

rdb_user=volumes

rdb_secret_uuid=19365acb-10b4-44c9-9a28-f948e8128e91

 

and ceph.conf file

[global]

fsid = 633accd0-dd09-4d97-ab40-2aca79f44d1c

mon_initial_members = ceph-1

mon_host = 10.193.0.111

auth_supported = cephx

osd_journal_size = 1024

filestore_xattr_use_omap = true

 

Thanks for the help

 

Song

 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux