Failure to run "rbd" commands

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



G'day all

Today's testing was on having a client (i.e. Ubuntu 12.04 server running KVM to be virtualisation host, named "server") connect to the 3x node v0.42 Ceph Cluster (names "ceph1, ceph2 & ceph3").

A very simple command "rbd ls" is failing per below:

root@server:/# rbd ls
terminate called after throwing an instance of 'ceph::buffer::error'
  what():  buffer::exception
*** Caught signal (Aborted) **
 in thread 7fe239f7d700
 ceph version 0.41 (commit:c1345f7136a0af55d88280ffe4b58339aaf28c9d)
 1: rbd() [0x459316]
 2: (()+0xfcb0) [0x7fe23c055cb0]
 3: (gsignal()+0x35) [0x7fe23a7b5475]
 4: (abort()+0x17b) [0x7fe23a7b8bdb]
 5: (__gnu_cxx::__verbose_terminate_handler()+0x11d) [0x7fe23b10169d]
 6: (()+0xb5846) [0x7fe23b0ff846]
 7: (()+0xb5873) [0x7fe23b0ff873]
 8: (()+0xb596e) [0x7fe23b0ff96e]
 9: (pg_pool_t::decode(ceph::buffer::list::iterator&)+0x859) [0x7fe23c3d0e99]
 10: (void decode<long, pg_pool_t>(std::map<long, pg_pool_t, std::less<long>, std::allocator<std::pair<long const, pg_pool_t> > >&, ceph::buffer::list::iterator&)+0x91) [0x7fe23c3c8851]
 11: (OSDMap::decode(ceph::buffer::list&)+0x17d) [0x7fe23c3befcd]
 12: (Objecter::handle_osd_map(MOSDMap*)+0x1e9f) [0x7fe23c3033ff]
 13: (librados::RadosClient::_dispatch(Message*)+0x54) [0x7fe23c2e3e44]
 14: (librados::RadosClient::ms_dispatch(Message*)+0x33) [0x7fe23c2e3ef3]
 15: (SimpleMessenger::dispatch_entry()+0x76b) [0x47ea1b]
 16: (SimpleMessenger::DispatchThread::entry()+0x1c) [0x46995c]
 17: (()+0x7e9a) [0x7fe23c04de9a]
 18: (clone()+0x6d) [0x7fe23a86f77d]
Aborted (core dumped)

However this command works on the same box:
root@server:/# echo "10.112.9.11,10.112.9.12,10.112.9.13 name=admin,secret=AQBBg0NPeFt9MRAAg8AWG108I8sje82hzdjkiw== rbd demo.rbd" > /sys/bus/rbd/add
root@server:/# ls -alh /dev/rbd*
brw-rw---- 1 root disk 251, 0 Feb 22 14:12 /dev/rbd0

/dev/rbd:
total 0
drwxr-xr-x  3 root root   60 Feb 22 14:12 .
drwxr-xr-x 16 root root 4.2K Feb 22 14:12 ..
drwxr-xr-x  2 root root   60 Feb 22 14:12 rbd


Confirming the below works as expected when run from any of the 3x servers within the Cluster:
root@ceph1:/etc/ceph# rbd ls
demo.rbd
root@ceph3:~# rbd -p rbd ls
demo.rbd

On the "client" KVM box, I have installed:
apt-get install ceph-common
modprobe rbd

Then created these two files:
;-------------------------------------------------------------------
/etc/ceph/ceph.conf
;-------------------------------------------------------------------
[global]
    ; enable secure authentication
    auth supported = cephx
    keyring = /etc/ceph/keyring.bin

[mon]
[mon.alpha]
        host = ceph1
        mon addr = 10.112.9.11:6789

[mon.bravo]
        host = ceph2
        mon addr = 10.112.9.12:6789

[mon.charlie]
        host = ceph3
        mon addr = 10.112.9.13:6789


;-------------------------------------------------------------------
/etc/ceph/keyring.bin
;-------------------------------------------------------------------
[client.admin]
        key = AQBBg0NPeFt9MRAAg8AWG108I8sje82hzdjkiw==





--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux