On Tue, Mar 11, 2014 at 9:12 PM, Mark s2c <mark@xxxxxxxxxxxxxxx> wrote: > > Hello > I followed everything in the set up documentation setting up a test cluster on an XCP install and got this: > > Invoked (1.3.5): /usr/bin/ceph-- > deploy admin domUs1 domUs2 domUs3 domUca > Pushing admin keys and conf to domUs1 > connected to host: domUs1 > detect platform information from remote host > detect machine type > get remote short hostname > write cluster configuration to /etc/ceph/{cluster}.conf > Pushing admin keys and conf to domUs2 > connected to host: domUs2 > detect platform information from remote host > detect machine type > get remote short hostname > write cluster configuration to /etc/ceph/{cluster}.conf > Pushing admin keys and conf to domUs3 > connected to host: domUs3 > detect platform information from remote host > detect machine type > get remote short hostname > write cluster configuration to /etc/ceph/{cluster}.conf > Pushing admin keys and conf to domUca > connected to host: domUca > detect platform information from remote host > detect machine type > get remote short hostname > write cluster configuration to /etc/ceph/{cluster}.conf > > Unhandled exception in thread started by <function run_and_release at 0xc7bb90> > Error in sys.excepthook: > Traceback (most recent call last): > File "/usr/lib64/python2.6/site-packages/abrt_exception_handler.py", line 204,, > in <lambda> > sys.excepthook = lambda etype, value, tb: handleMyException((etype, value, tt > b)) > TypeError: 'NoneType' object is not callable > > Original exception was: > Traceback (most recent call last): > File "/usr/lib/python2.6/site-packages/ceph_deploy/lib/remoto/lib/execnet/gatee > way_base.py", line 245, in run_and_release > with self._running_lock: > File "/usr/lib64/python2.6/threading.py", line 117, in acquire > me = _get_ident() > TypeError: 'NoneType' object is not callable > That exception you see is fixed in the current master and should be part of a ceph-deploy release soon. It also has nothing to do with errors with ceph-deploy actually working, it is just garbage from disconnecting from the remote server. > I must've missed the advice to install ceph on the admin node cos I hadn't done that. When I did, thinking this maybe a spurious error, I get this: > > 2014-03-12 01:10:08.094837 7fe8c8626700 0 -- :/1011655 >> 192.168.10.25:6789/0 pipe(0x7fe8c4024440 sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7fe8c40246a0).fault > 2014-03-12 01:10:11.091931 7fe8c8525700 0 -- :/1011655 >> 192.168.10.25:6789/0 pipe(0x7fe8b8000c00 sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x7fe8b8000e60).fault > ... > > I specified 2 OSDs on 2 virtualdisks plugged into each of domUs1-3. The setup is running on a new HP DL360p h/w RAID across 4 x 1 Tb disks. > > Anyone seen this before? > Thanks > Mark > ps. note I captured the above output with a typescript so there maybe duplicate chars in certain places. > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com