Re: how to Testing cinder and glance with CEPH

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks Jens / Sebastien

It worked for me now ,  Thanks a lot for your suggestions , they were worth.

Many Thanks
Karan Singh



From: "Jens-Christian Fischer" <jens-christian.fischer@xxxxxxxxx>
To: "Karan Singh" <ksingh@xxxxxx>
Cc: "Sebastien Han" <sebastien.han@xxxxxxxxxxxx>, ceph-users@xxxxxxxxxxxxxx
Sent: Wednesday, 27 November, 2013 11:57:53 AM
Subject: Re: how to Testing cinder and glance with CEPH

Hi Karan

your cinder.conf looks sensible to me, I have posted mine here:

--- cut ---

[DEFAULT]
rootwrap_config = /etc/cinder/rootwrap.conf
api_paste_confg = /etc/cinder/api-paste.ini
iscsi_helper = tgtadm
volume_name_template = volume-%s
volume_group = cinder-volumes
verbose = True
auth_strategy = keystone
state_path = /var/lib/cinder
lock_path = /var/lock/cinder
volumes_dir = /var/lib/cinder/volumes

volume_driver=cinder.volume.drivers.rbd.RBDDriver
rbd_pool=volumes
glance_api_version=2

rbd_user=volumes
rbd_secret_uuid=e1915277-e3a5-4547-bc9e-xxxxxxx

rpc_backend = cinder.openstack.common.rpc.impl_kombu
rabbit_host = xxx.yyy.cc
rabbit_port = 5672

quota_volumes=20
quota_snapshots=20

debug = False
use_syslog = True
syslog_log_facility = LOG_LOCAL0



[database]
connection = mysql://cinder:xxx@xxxxxxxxxx/cinder


[keystone_authtoken]
# keystone public API
auth_protocol = https
auth_host = xxx.yyy.cc
auth_port = 5000
admin_tenant_name = service
admin_user = cinder
admin_password =xxx

--- cut ---

what are the different cinder*.log files telling you?

Is /etc/ceph/ceph.conf readable for other processes? (chmod 644 /etc/ceph/ceph.conf)
Are the key rings available and readable?

good luck
jc


-- 
SWITCH
Jens-Christian Fischer, Peta Solutions
Werdstrasse 2, P.O. Box, 8021 Zurich, Switzerland
phone +41 44 268 15 15, direct +41 44 268 15 71
jens-christian.fischer@xxxxxxxxx

On 27.11.2013, at 08:51, Karan Singh <ksingh@xxxxxx> wrote:

Hello Sebastien / Community


I tried the commands mentioned in below email.


[root@rdo ~]#
[root@rdo ~]# cinder create 1
+---------------------+--------------------------------------+
|       Property      |                Value                 |
+---------------------+--------------------------------------+
|     attachments     |                  []                  |
|  availability_zone  |                 nova                 |
|       bootable      |                false                 |
|      created_at     |      2013-11-27T07:40:54.161478      |
| display_description |                 None                 |
|     display_name    |                 None                 |
|          id         | ae8cd686-5f1d-4c05-8c42-cb7622122a3e |
|       metadata      |                  {}                  |
|         size        |                  1                   |
|     snapshot_id     |                 None                 |
|     source_volid    |                 None                 |
|        status       |               creating               |
|     volume_type     |                 None                 |
+---------------------+--------------------------------------+
[root@rdo ~]#
[root@rdo ~]# cinder list
+--------------------------------------+--------+--------------+------+-------------+----------+-------------+
|                  ID                  | Status | Display Name | Size | Volume Type | Bootable | Attached to |
+--------------------------------------+--------+--------------+------+-------------+----------+-------------+
| ae8cd686-5f1d-4c05-8c42-cb7622122a3e | error  |     None     |  1   |     None    |  false   |             |
+--------------------------------------+--------+--------------+------+-------------+----------+-------------+
[root@rdo ~]#
[root@rdo ~]#
[root@rdo ~]#
[root@rdo ~]# rbd -p ceph-volumes ls
rbd: pool ceph-volumes doesn't contain rbd images
[root@rdo ~]#
[root@rdo ~]#
[root@rdo ~]# rados lspools
data
metadata
rbd
ceph-images
ceph-volumes
[root@rdo ~]# rbd -p rbd ls
[root@rdo ~]# rbd -p data ls
foo
foo1
[root@rdo ~]#




I checked in cinder.log and got the below errors.


2013-11-27 09:44:14.830 3273 INFO cinder.volume.manager [-] Updating volume status
2013-11-27 09:44:14.830 3273 WARNING cinder.volume.manager [-] Unable to update stats, driver is uninitialized
2013-11-27 09:44:42.407 12007 INFO cinder.volume.manager [-] Updating volume status
2013-11-27 09:44:42.408 12007 WARNING cinder.volume.manager [-] Unable to update stats, driver is uninitialized
2013-11-27 09:44:51.799 4943 INFO cinder.volume.manager [-] Updating volume status
2013-11-27 09:44:51.799 4943 WARNING cinder.volume.manager [-] Unable to update stats, driver is uninitialized
2013-11-27 09:45:14.834 3273 INFO cinder.volume.manager [-] Updating volume status
2013-11-27 09:45:14.834 3273 WARNING cinder.volume.manager [-] Unable to update stats, driver is uninitialized
[root@rdo cinder]#




Output from my cinder.conf file



# Options defined in cinder.volume.utils
#

# The default block size used when copying/clearing volumes
# (string value)
#volume_dd_blocksize=1M


# Total option count: 382
volume_driver=cinder.volume.drivers.rbd.RBDDriver
rbd_pool=ceph-volumes
glance_api_version=2
rbd_user=volumes
rbd_secret_uuid=801a42ec-aec1-3ea8-d869-823c2de56b83

rootwrap_config=/etc/cinder/rootwrap.conf
sql_connection = mysql://cinder:root@localhost/cinder
api_paste_config = /etc/cinder/api-paste.ini

iscsi_helper=tgtadm
volume_name_template = volume-%s
volume_group = cinder-volumes
verbose = True
auth_strategy = keystone
#osapi_volume_listen_port=5900
[keystone_authtoken]
admin_tenant_name = admin
admin_user = admin
admin_password = root@123
auth_host = 192.168.1.49
auth_port = 35357
auth_protocol = http
signing_dirname = /tmp/keystone-signing-cinder
[root@rdo cinder]#



Can you suggest what is wrong with this configuration , i followed everything mentioned on http://ceph.com/docs/master/rbd/rbd-openstack/   and there is not much information available on internet WRT this error.

Please help.



Many Thanks
Karan Singh


----- Original Message -----
From: "Sebastien Han" <sebastien.han@xxxxxxxxxxxx>
To: "Karan Singh" <ksingh@xxxxxx>
Cc: ceph-users@xxxxxxxxxxxxxx
Sent: Wednesday, 27 November, 2013 1:08:35 AM
Subject: Re: how to Testing cinder and glance with CEPH

Hi,

Well after restarting the services run:

$ cinder create 1

Then you can check both status in Cinder and Ceph:

For Cinder run:
$ cinder list

For Ceph run:
$ rbd -p <cinder-pool> ls

If the image is there, you’re good.

Cheers.

––––
Sébastien Han
Cloud Engineer

"Always give 100%. Unless you're giving blood.”

Phone: +33 (0)1 49 70 99 72
Mail: sebastien.han@xxxxxxxxxxxx
Address : 10, rue de la Victoire - 75009 Paris
Web : www.enovance.com - Twitter : @enovance

On 27 Nov 2013, at 00:04, Karan Singh <ksingh@xxxxxx> wrote:

Hello Cephers

I was following http://ceph.com/docs/master/rbd/rbd-openstack/  for ceph and openstack Integration , using this document ih ave done all the changes required for this integration.

I am not sure how should i test my configuration , how should i make sure integration is successful. Can you suggest some test that i can perform to check my ceph and openstack integration .

FYI , in the document http://ceph.com/docs/master/rbd/rbd-openstack/  , Nothing is mentioned after " Restart Openstack Services "  heading , but there should be steps to test this inttegration , please suggest me here , i am new to openstack great if you can give me some commanes used for testing.



Karan Singh
CSC - IT Center for Science Ltd.
P.O. Box 405, FI-02101 Espoo, FINLAND
http://www.csc.fi/ | +358 (0) 503 812758

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux