Unable to start ceph monitor in V0.59

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi List,

 

         I cannot start my monitor when I update my cluster to v0.59, pls note that I am not trying to upgrade,but by reinstall the ceph software stack and rerunning mkcephfs.  I have seen that the monitor change a lot after 0.58, is the mkcephfs still have bugs ?

 

Below is the log:

 

 

2013-03-21 08:17:41.127576 7f71c3610780  0 ceph version 0.59 (cbae6a435c62899f857775f66659de052fb0e759), process ceph-mon, pid 1550

2013-03-21 08:17:41.131271 7f71c3610780  1 unable to open monitor store at /data/mon.ceph1

2013-03-21 08:17:41.131281 7f71c3610780  1 check for old monitor store format

2013-03-21 08:17:41.131409 7f71c3610780  1 store(/data/mon.ceph1) mount

2013-03-21 08:17:41.131430 7f71c3610780  1 store(/data/mon.ceph1) mount

2013-03-21 08:17:41.131659 7f71c3610780  1 found old GV monitor store format -- should convert!

2013-03-21 08:17:41.136476 7f71c3610780  1 store(/data/mon.ceph1) mount

2013-03-21 08:17:46.098118 7f71c3610780  1 _convert_paxos first gv 2 last gv 475156

2013-03-21 08:17:47.131667 7f71c3610780  0 convert finished conversion

2013-03-21 08:17:47.185261 7f71c3610780  1 mon.ceph1@-1(probing) e1 preinit fsid 6d4e68d7-8959-4e8e-90c9-7e43f508f16a

2013-03-21 08:17:47.220874 7f71c3610780  0 mon.ceph1@-1(probing) e1  my rank is now 0 (was -1)

2013-03-21 08:17:47.220905 7f71c3610780  1 mon.ceph1@0(probing) e1 win_standalone_election

2013-03-21 08:17:47.221808 7f71c3610780  0 log [INF] : mon.ceph1@0 won leader election with quorum 0

2013-03-21 08:17:47.238542 7f71c3610780  0 log [INF] : pgmap v217425: 10368 pgs: 140 active+clean, 3 stale+active+recovering, 2 stal

e, 67 stale+active, 2 active+recovery_wait, 99 stale+active+clean, 819 peering, 5 stale+active+degraded+wait_backfill, 3 stale+activ

e+recovery_wait, 3760 down+peering, 12 stale+active+recovering+degraded, 305 stale+peering, 3977 stale+down+peering, 1135 stale+acti

ve+degraded, 2 stale+active+degraded+backfilling, 4 stale+active+degraded+remapped+wait_backfill, 6 incomplete, 1 stale+remapped+pee

ring, 17 stale+incomplete, 1 stale+active+degraded+remapped, 8 active+recovering; 2717 GB data, 2015 GB used, 15441 GB / 17457 GB av

ail; 90413/1391836 degraded (6.496%); 250/695918 unfound (0.036%)

2013-03-21 08:17:47.239560 7f71c3610780  0 log [INF] : mdsmap e1: 0/0/1 up

2013-03-21 08:17:47.240448 7f71c3610780  0 log [INF] : osdmap e5056: 80 osds: 25 up, 25 in

 

 

2013-03-21 08:17:47.241019 7f71c3610780  0 log [INF] : monmap e1: 1 mons at {ceph1=192.168.10.11:6789/0}

2013-03-21 08:17:47.441000 7f71bc1d1700 -1 auth/none/AuthNoneServiceHandler.h: In function 'virtual int AuthNoneServiceHandler::hand

le_request(ceph::buffer::list::iterator&, ceph::bufferlist&, uint64_t&, AuthCapsInfo&, uint64_t*)' thread 7f71bc1d1700 time 2013-03-

21 08:17:47.440030

auth/none/AuthNoneServiceHandler.h: 35: FAILED assert(0)

 

ceph version 0.59 (cbae6a435c62899f857775f66659de052fb0e759)

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux