Python exception on ceph-deploy admin

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello
I followed everything in the set up documentation setting up a test cluster on an XCP install and got this:

Invoked (1.3.5): /usr/bin/ceph--
deploy admin domUs1 domUs2 domUs3 domUca
Pushing admin keys and conf to domUs1
connected to host: domUs1
detect platform information from remote host
detect machine type
get remote short hostname
write cluster configuration to /etc/ceph/{cluster}.conf
Pushing admin keys and conf to domUs2
connected to host: domUs2
detect platform information from remote host
detect machine type
get remote short hostname
write cluster configuration to /etc/ceph/{cluster}.conf
Pushing admin keys and conf to domUs3
connected to host: domUs3
detect platform information from remote host
detect machine type
get remote short hostname
write cluster configuration to /etc/ceph/{cluster}.conf
Pushing admin keys and conf to domUca
connected to host: domUca
detect platform information from remote host
detect machine type
get remote short hostname
write cluster configuration to /etc/ceph/{cluster}.conf

Unhandled exception in thread started by <function run_and_release at 0xc7bb90>
Error in sys.excepthook:
Traceback (most recent call last):
  File "/usr/lib64/python2.6/site-packages/abrt_exception_handler.py", line 204,,
 in <lambda>
    sys.excepthook = lambda etype, value, tb: handleMyException((etype, value, tt
b))
TypeError: 'NoneType' object is not callable

Original exception was:
Traceback (most recent call last):
  File "/usr/lib/python2.6/site-packages/ceph_deploy/lib/remoto/lib/execnet/gatee
way_base.py", line 245, in run_and_release
    with self._running_lock:
  File "/usr/lib64/python2.6/threading.py", line 117, in acquire
    me = _get_ident()
TypeError: 'NoneType' object is not callable

I must've missed the advice to install ceph on the admin node cos I hadn't done that.  When I did, thinking this maybe a spurious error, I get this:

2014-03-12 01:10:08.094837 7fe8c8626700  0 -- :/1011655 >> 192.168.10.25:6789/0 pipe(0x7fe8c4024440 sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7fe8c40246a0).fault
2014-03-12 01:10:11.091931 7fe8c8525700  0 -- :/1011655 >> 192.168.10.25:6789/0 pipe(0x7fe8b8000c00 sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7fe8b8000e60).fault
...

I specified 2 OSDs on 2 virtualdisks plugged into each of domUs1-3.  The setup is running on a new HP DL360p h/w RAID across 4 x 1 Tb disks.

Anyone seen this before?
Thanks
Mark
ps. note I captured the above output with a typescript so there maybe duplicate chars in certain places.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux