Re: Ceph Openstack deployment

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Dear Cephers,

   I still can attach volume to my cloud machines, ceph version is  0.94.5 (9764da52395923e0b32908d83a9f7304401fee43) and Openstack Juno

   Nova+cinder are able to create volumes on Ceph
cephvolume:~ # rados ls --pool volumes
rbd_header.1f7784a9e1c2e
rbd_id.volume-5e2ab5c2-4710-4c28-9755-b5bc4ff6a52a
rbd_directory
rbd_id.volume-7da08f12-fb0f-4269-931a-d528c1507fee
rbd_header.23d5e33b4c15c
rbd_id.volume-4d26bb31-91e8-4646-8010-82127b775c8e
rbd_header.20407190ce77f

cloud:~ # cinder list
+--------------------------------------+--------+--------------+------+-------------+----------+------------------------------------------------------------------------------------------+
|                  ID                                               | Status  | Display Name | Size | Volume Type | Bootable |                                    Attached to              |
+--------------------------------------+--------+--------------+------+-------------+----------+--------------------------------|---------------------------------------------------------+
| 4d26bb31-91e8-4646-8010-82127b775c8e | in-use |     None     |         2   |             rbd     |  false       | 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb |
+--------------------------------------+--------+--------------+------+-------------+----------+------------------------------------------------------------------------------------------+


   nova:~ # nova volume-attach 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb 4d26bb31-91e8-4646-8010-82127b775c8e auto
+----------+------------------------------------------------------------+
| Property |                              Value                                |
+----------+------------------------------------------------------------+
| device      | /dev/xvdd                                                     |
| id             | 4d26bb31-91e8-4646-8010-82127b775c8e |
| serverId   | 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb  |
| volumeId | 4d26bb31-91e8-4646-8010-82127b775c8e |
+----------+--------------------------------------+

From nova-compute (Ubuntu 14.04 LTS \n \l) node I see the attaching/detaching:
cloud01:~ # dpkg -l | grep ceph
ii  ceph-common                         0.94.5-1trusty                      amd64        common utilities to mount and interact with a ceph storage cluster
ii  libcephfs1                               0.94.5-1trusty                      amd64        Ceph distributed file system client library
ii  python-cephfs                         0.94.5-1trusty                      amd64        Python libraries for the Ceph libcephfs library
ii  librbd1                                    0.94.5-1trusty                      amd64        RADOS block device client library
ii  python-rbd                              0.94.5-1trusty                      amd64        Python libraries for the Ceph librbd library

at cinder.conf

 rbd_user = cinder
rbd_secret_uuid = 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxxx

[rbd-cephvolume]
volume_backend_name=rbd
volume_driver = cinder.volume.drivers.rbd.RBDDriver
rbd_pool = volumes
rbd_ceph_conf = /etc/ceph/ceph.conf
rbd_flatten_volume_from_snapshot = false
rbd_max_clone_depth = 5
rbd_store_chunk_size = 4
rados_connect_timeout = -1
glance_api_version = 2

in nova.conf
rbd_user=cinder

# The libvirt UUID of the secret for the rbd_uservolumes
# (string value)
rbd_secret_uuid=67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxxx

images_rbd_pool=volumes

# Path to the ceph configuration file to use (string value)
images_rbd_ceph_conf=/etc/ceph/ceph.conf

ls -la /etc/libvirt/secrets
total 16
drwx------ 2 root root 4096 Nov  4 10:28 .
drwxr-xr-x 7 root root 4096 Oct 22 13:15 ..
-rw------- 1 root root   40 Nov  4 10:28 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxx.base64
-rw------- 1 root root  170 Nov  4 10:25 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxx.xml



2015-11-04 10:39:42.573 11653 INFO nova.compute.manager [req-8b2a9793-4b39-4cb0-b291-e492c350387e b7aadbb4a85745feb498b74e437129cc ce2dd2951bd24c1ea3b43c3b3716f604 - - -] [instance: 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb] Detach volume 4d26bb31-91e8-4646-8010-82127b775c8e from mountpoint /dev/xvdd
2015-11-04 10:40:43.266 11653 INFO nova.compute.manager [req-35218de0-3f26-496b-aad9-5c839143da17 b7aadbb4a85745feb498b74e437129cc ce2dd2951bd24c1ea3b43c3b3716f604 - - -] [instance: 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb] Attaching volume 4d26bb31-91e8-4646-8010-82127b775c8e to /dev/xvdd

but one on cloud machine (SL6) the volume y never showed (xvdd).
[root@cloud5 ~]# cat /proc/partitions 
major minor  #blocks  name

 202        0   20971520 xvda
 202       16  209715200 xvdb
 202       32   10485760 xvdc

Thanks in advance, I

2015-11-03 11:18 GMT+01:00 Iban Cabrillo <cabrillo@xxxxxxxxxxxxxx>:
Hi all,
    During last week I been trying to deploy the pre-existing ceph cluster with out openstack intance.
    The ceph-cinder integration was easy (or at least I think so!!)
    There is only one volume to attach block storage to out cloud machines.

    The client.cinder has permission on this volume (following the guides)
    ...............
    client.cinder
key: AQAonXXXXXXXRAAPIAj9iErv001a0k+vyFdUg==
caps: [mon] allow r
caps: [osd] allow class-read object_prefix rbd_children, allow rwx pool=volumes

   ceph.conf file seems to be OK:
   
[global]
fsid = 6f5a65a7-316c-4825-afcb-428608941dd1
mon_initial_members = cephadm, cephmon02, cephmon03
mon_host = 10.10.3.1,10.10.3.2,10.10.3.3
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
filestore_xattr_use_omap = true
osd_pool_default_size = 2
public_network = 10.10.0.0/16
cluster_network = 192.168.254.0/27

[osd]
osd_journal_size = 20000

[client.cinder]
keyring = /etc/ceph/ceph.client.cinder.keyring

[client]
rbd cache = true
rbd cache writethrough until flush = true
admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok


The trouble seems that blocks are created using the client.admin instead of client.cinder

From cinder machine:

cinder:~ # rados ls --pool volumes
rbd_id.volume-5e2ab5c2-4710-4c28-9755-b5bc4ff6a52a
rbd_directory
rbd_id.volume-7da08f12-fb0f-4269-931a-d528c1507fee
rbd_header.23d5e33b4c15c
rbd_header.20407190ce77f

But if I try to look for using cinder client:


  cinder:~ #rados ls --pool volumes --secret client.cinder
  "empty answer"

cinder:~ # ls -la /etc/ceph
total 24
drwxr-xr-x   2 root   root   4096 nov  3 10:17 .
drwxr-xr-x 108 root   root   4096 oct 29 09:52 ..
-rw-------   1 root   root     63 nov  3 10:17 ceph.client.admin.keyring
-rw-r--r--   1 cinder cinder   67 oct 28 13:44 ceph.client.cinder.keyring
-rw-r--r--   1 root   root    454 oct  1 13:56 ceph.conf
-rw-r--r--   1 root   root     73 sep 27 09:36 ceph.mon.keyring


from a client (I have supposed that this machine only need the cinder key...)

cloud28:~ # ls -la /etc/ceph/
total 28
drwx------   2 root root  4096 nov  3 11:01 .
drwxr-xr-x 116 root root 12288 oct 30 14:37 ..
-rw-r--r--   1 nova nova    67 oct 28 11:43 ceph.client.cinder.keyring
-rw-r--r--   1 root root   588 nov  3 10:59 ceph.conf
-rw-r--r--   1 root root    92 oct 26 16:59 rbdmap

cloud28:~ # rbd -p volumes ls
2015-11-03 11:01:58.782795 7fc6c714b840 -1 monclient(hunting): ERROR: missing keyring, cannot use cephx for authentication
2015-11-03 11:01:58.782800 7fc6c714b840  0 librados: client.admin initialization error (2) No such file or directory
rbd: couldn't connect to the cluster!

Any help will be welcome.




--
############################################################################
Iban Cabrillo Bartolome
Instituto de Fisica de Cantabria (IFCA)
Santander, Spain
Tel: +34942200969
PGP PUBLIC KEY: http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC
############################################################################
Bertrand Russell:
"El problema con el mundo es que los estúpidos están seguros de todo y los inteligentes están llenos de dudas"
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux