Re: Block Devices and OpenStack

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

Can I see your ceph.conf?
I suspect that [client.cinder] and [client.glance] sections are missing.

Cheers.
–––– 
Sébastien Han 
Cloud Engineer 

"Always give 100%. Unless you're giving blood.” 

Phone: +33 (0)1 49 70 99 72 
Mail: sebastien.han@xxxxxxxxxxxx 
Address : 10, rue de la Victoire - 75009 Paris 
Web : www.enovance.com - Twitter : @enovance 

On 16 Feb 2014, at 06:55, Ashish Chandra <mail.ashishchandra@xxxxxxxxx> wrote:

> Hi Jean,
> 
> Here is the output for ceph auth list for client.cinder
> 
> client.cinder
>         key: AQCKaP9ScNgiMBAAwWjFnyL69rBfMzQRSHOfoQ==
>         caps: [mon] allow r
>         caps: [osd] allow class-read object_prefix rbd_children, allow rwx pool=volumes, allow rx pool=images
> 
> 
> Here is the output of ceph -s:
> 
> ashish@ceph-client:~$ ceph -s
>     cluster afa13fcd-f662-4778-8389-85047645d034
>      health HEALTH_OK
>      monmap e1: 1 mons at {ceph-node1=10.0.1.11:6789/0}, election epoch 1, quorum 0 ceph-node1
>      osdmap e37: 3 osds: 3 up, 3 in
>       pgmap v84: 576 pgs, 6 pools, 0 bytes data, 0 objects
>             106 MB used, 9076 MB / 9182 MB avail
>                  576 active+clean
> 
> I created all the keyrings and copied as suggested by the guide.
> 
> 
> 
> 
> 
> 
> On Sun, Feb 16, 2014 at 3:08 AM, Jean-Charles LOPEZ <jc.lopez@xxxxxxxxxxx> wrote:
> Hi,
> 
> what do you get when you run a 'ceph auth list' command for the user name (client.cinder) you created for cinder? Are the caps and the key for this user correct? No typo in the hostname in the cinder.conf file (host=) ? Did you copy the keyring to the cinder running cinder (can’t really say from your output and there is no ceph-s command to check the monitor names)?
> 
> It could just be a typo in the ceph auth get-or-create command that’s causing it.
> 
> Rgds
> JC
> 
> 
> 
> On Feb 15, 2014, at 10:35, Ashish Chandra <mail.ashishchandra@xxxxxxxxx> wrote:
> 
>> Hi Cephers,
>> 
>> I am trying to configure ceph rbd as backend for cinder and glance by following the steps mentioned in:
>> 
>> http://ceph.com/docs/master/rbd/rbd-openstack/
>> 
>> Before I start all openstack services are running normally and ceph cluster health shows "HEALTH_OK"
>> 
>> But once I am done with all steps and restart openstack services, cinder-volume fails to start and throws an error.
>> 
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd Traceback (most recent call last):
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd   File "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 262, in check_for_setup_error
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd     with RADOSClient(self):
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd   File "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 234, in __init__
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd     self.cluster, self.ioctx = driver._connect_to_rados(pool)
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd   File "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 282, in _connect_to_rados
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd     client.connect()
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd   File "/usr/lib/python2.7/dist-packages/rados.py", line 185, in connect
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd     raise make_ex(ret, "error calling connect")
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd Error: error calling connect: error code 95
>> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd
>> 2014-02-16 00:01:42.591 ERROR cinder.volume.manager [req-8134a4d7-53f8-4ada-b4b5-4d96d7cad4bc None None] Error encountered during initialization of driver: RBDDriver
>> 2014-02-16 00:01:42.592 ERROR cinder.volume.manager [req-8134a4d7-53f8-4ada-b4b5-4d96d7cad4bc None None] Bad or unexpected response from the storage volume backend API: error connecting to ceph cluster
>> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager Traceback (most recent call last):
>> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager   File "/opt/stack/cinder/cinder/volume/manager.py", line 190, in init_host
>> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager     self.driver.check_for_setup_error()
>> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager   File "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 267, in check_for_setup_error
>> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager     raise exception.VolumeBackendAPIException(data=msg)
>> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager VolumeBackendAPIException: Bad or unexpected response from the storage volume backend API: error connecting to ceph cluster
>> 
>> 
>> Here is the content of my /etc/ceph in openstack node: 
>> 
>> ashish@ubuntu:/etc/ceph$ ls -lrt
>> total 16
>> -rw-r--r-- 1 cinder cinder 229 Feb 15 23:45 ceph.conf
>> -rw-r--r-- 1 glance glance  65 Feb 15 23:46 ceph.client.glance.keyring
>> -rw-r--r-- 1 cinder cinder  65 Feb 15 23:47 ceph.client.cinder.keyring
>> -rw-r--r-- 1 cinder cinder  72 Feb 15 23:47 ceph.client.cinder-backup.keyring
>> 
>> I am really stuck and tried a lot. What Could possibly I be doing wrong.
>> 
>> 
>> HELP.
>> 
>> 
>> Thanks and Regards
>> Ashish Chandra
>> 
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux