Re: Python exception on ceph-deploy admin

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Mar 11, 2014 at 9:12 PM, Mark s2c <mark@xxxxxxxxxxxxxxx> wrote:
>
> Hello
> I followed everything in the set up documentation setting up a test cluster on an XCP install and got this:
>
> Invoked (1.3.5): /usr/bin/ceph--
> deploy admin domUs1 domUs2 domUs3 domUca
> Pushing admin keys and conf to domUs1
> connected to host: domUs1
> detect platform information from remote host
> detect machine type
> get remote short hostname
> write cluster configuration to /etc/ceph/{cluster}.conf
> Pushing admin keys and conf to domUs2
> connected to host: domUs2
> detect platform information from remote host
> detect machine type
> get remote short hostname
> write cluster configuration to /etc/ceph/{cluster}.conf
> Pushing admin keys and conf to domUs3
> connected to host: domUs3
> detect platform information from remote host
> detect machine type
> get remote short hostname
> write cluster configuration to /etc/ceph/{cluster}.conf
> Pushing admin keys and conf to domUca
> connected to host: domUca
> detect platform information from remote host
> detect machine type
> get remote short hostname
> write cluster configuration to /etc/ceph/{cluster}.conf
>
> Unhandled exception in thread started by <function run_and_release at 0xc7bb90>
> Error in sys.excepthook:
> Traceback (most recent call last):
>   File "/usr/lib64/python2.6/site-packages/abrt_exception_handler.py", line 204,,
>  in <lambda>
>     sys.excepthook = lambda etype, value, tb: handleMyException((etype, value, tt
> b))
> TypeError: 'NoneType' object is not callable
>
> Original exception was:
> Traceback (most recent call last):
>   File "/usr/lib/python2.6/site-packages/ceph_deploy/lib/remoto/lib/execnet/gatee
> way_base.py", line 245, in run_and_release
>     with self._running_lock:
>   File "/usr/lib64/python2.6/threading.py", line 117, in acquire
>     me = _get_ident()
> TypeError: 'NoneType' object is not callable
>

That exception you see is fixed in the current master and should be
part of a ceph-deploy release soon. It also
has nothing to do with errors with ceph-deploy actually working, it is
just garbage from disconnecting from the remote
server.

> I must've missed the advice to install ceph on the admin node cos I hadn't done that.  When I did, thinking this maybe a spurious error, I get this:
>
> 2014-03-12 01:10:08.094837 7fe8c8626700  0 -- :/1011655 >> 192.168.10.25:6789/0 pipe(0x7fe8c4024440 sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7fe8c40246a0).fault
> 2014-03-12 01:10:11.091931 7fe8c8525700  0 -- :/1011655 >> 192.168.10.25:6789/0 pipe(0x7fe8b8000c00 sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7fe8b8000e60).fault
> ...
>
> I specified 2 OSDs on 2 virtualdisks plugged into each of domUs1-3.  The setup is running on a new HP DL360p h/w RAID across 4 x 1 Tb disks.
>
> Anyone seen this before?
> Thanks
> Mark
> ps. note I captured the above output with a typescript so there maybe duplicate chars in certain places.
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux