IIRC, secret-define will give you a uuid to use, but you can also just tell it to use a predefined one. You need to do so. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Fri, Jul 26, 2013 at 10:32 AM, johnu <johnugeorge109@xxxxxxxxx> wrote: > Greg, > :) I am not getting where was the mistake in the configuration. > virsh secret-define gave different secrets > > sudo virsh secret-define --file secret.xml > <uuid of secret is output here> > sudo virsh secret-set-value --secret {uuid of secret} --base64 $(cat > client.volumes.key) > > > > On Fri, Jul 26, 2013 at 10:16 AM, Gregory Farnum <greg@xxxxxxxxxxx> wrote: >> >> On Fri, Jul 26, 2013 at 10:11 AM, johnu <johnugeorge109@xxxxxxxxx> wrote: >> > Greg, >> > Yes, the outputs match >> >> Nope, they don't. :) You need the secret_uuid to be the same on each >> node, because OpenStack is generating configuration snippets on one >> node (which contain these secrets) and then shipping them to another >> node where they're actually used. >> >> Your secrets are also different despite having the same rbd user >> specified, so that's broken too; not quite sure how you got there... >> -Greg >> Software Engineer #42 @ http://inktank.com | http://ceph.com >> >> > >> > master node: >> > >> > ceph auth get-key client.volumes >> > AQC/ze1R2EOWNBAAmLUE4U7zO1KafZ/CzVVTqQ== >> > >> > virsh secret-get-value bdf77f5d-bf0b-1053-5f56-cd76b32520dc >> > AQC/ze1R2EOWNBAAmLUE4U7zO1KafZ/CzVVTqQ== >> > >> > /etc/cinder/cinder.conf >> > >> > volume_driver=cinder.volume.drivers.rbd.RBDDriver >> > rbd_pool=volumes >> > glance_api_version=2 >> > rbd_user=volumes >> > rbd_secret_uuid=bdf77f5d-bf0b-1053-5f56-cd76b32520dc >> > >> > >> > slave1 >> > >> > /etc/cinder/cinder.conf >> > >> > volume_driver=cinder.volume.drivers.rbd.RBDDriver >> > rbd_pool=volumes >> > glance_api_version=2 >> > rbd_user=volumes >> > rbd_secret_uuid=62d0b384-50ad-2e17-15ed-66bfeda40252 >> > >> > >> > virsh secret-get-value 62d0b384-50ad-2e17-15ed-66bfeda40252 >> > AQC/ze1R2EOWNBAAmLUE4U7zO1KafZ/CzVVTqQ== >> > >> > slave2 >> > >> > /etc/cinder/cinder.conf >> > >> > volume_driver=cinder.volume.drivers.rbd.RBDDriver >> > rbd_pool=volumes >> > glance_api_version=2 >> > rbd_user=volumes >> > rbd_secret_uuid=33651ba9-5145-1fda-3e61-df6a5e6051f5 >> > >> > virsh secret-get-value 33651ba9-5145-1fda-3e61-df6a5e6051f5 >> > AQC/ze1R2EOWNBAAmLUE4U7zO1KafZ/CzVVTqQ== >> > >> > >> > Yes, Openstack horizon is showing same host for all volumes. Somehow, if >> > volume is attached to an instance lying on the same host, it works >> > otherwise, it doesn't. Might be a coincidence. And I am surprised that >> > no >> > one else has seen or reported this issue. Any idea? >> > >> > On Fri, Jul 26, 2013 at 9:45 AM, Gregory Farnum <greg@xxxxxxxxxxx> >> > wrote: >> >> >> >> On Fri, Jul 26, 2013 at 9:35 AM, johnu <johnugeorge109@xxxxxxxxx> >> >> wrote: >> >> > Greg, >> >> > I verified in all cluster nodes that rbd_secret_uuid is same >> >> > as >> >> > virsh secret-list. And If I do virsh secret-get-value of this uuid, i >> >> > getting back the auth key for client.volumes. What did you mean by >> >> > same >> >> > configuration?. Did you mean same secret for all compute nodes? >> >> >> >> If you run "virsh secret-get-value" with that rbd_secret_uuid on each >> >> compute node, does it return the right secret for client.volumes? >> >> >> >> > when we login as admin, There is a column in admin panel >> >> > which >> >> > gives >> >> > the 'host' where the volumes lie. I know that volumes are striped >> >> > across >> >> > the >> >> > cluster but it gives same host for all volumes. That is why ,I got >> >> > little >> >> > confused. >> >> >> >> That's not something you can get out of the RBD stack itself; is this >> >> something that OpenStack is showing you? I suspect it's just making up >> >> information to fit some API expectations, but somebody more familiar >> >> with the OpenStack guts can probably chime in. >> >> -Greg >> >> Software Engineer #42 @ http://inktank.com | http://ceph.com >> > >> > > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com