Re: mds cluster degraded

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hmm, last time we saw this it meant that the MDS log had gotten
corrupted somehow and was a little short (in that case due to the OSDs
filling up). What do you mean by "rebuilt the OSDs"?
-Greg

On Mon, Nov 17, 2014 at 12:52 PM, JIten Shah <jshah2005@xxxxxx> wrote:
> After i rebuilt the OSD’s, the MDS went into the degraded mode and will not
> recover.
>
>
> [jshah@Lab-cephmon001 ~]$ sudo tail -100f
> /var/log/ceph/ceph-mds.Lab-cephmon001.log
> 2014-11-17 17:55:27.855861 7fffef5d3700  0 -- X.X.16.111:6800/3046050 >>
> X.X.16.114:0/838757053 pipe(0x1e18000 sd=22 :6800 s=0 pgs=0 cs=0 l=0
> c=0x1e02c00).accept peer addr is really X.X.16.114:0/838757053 (socket is
> X.X.16.114:34672/0)
> 2014-11-17 17:57:27.855519 7fffef5d3700  0 -- X.X.16.111:6800/3046050 >>
> X.X.16.114:0/838757053 pipe(0x1e18000 sd=22 :6800 s=2 pgs=2 cs=1 l=0
> c=0x1e02c00).fault with nothing to send, going to standby
> 2014-11-17 17:58:47.883799 7fffef3d1700  0 -- X.X.16.111:6800/3046050 >>
> X.X.16.114:0/26738200 pipe(0x1e1be80 sd=23 :6800 s=0 pgs=0 cs=0 l=0
> c=0x1e04ba0).accept peer addr is really X.X.16.114:0/26738200 (socket is
> X.X.16.114:34699/0)
> 2014-11-17 18:00:47.882484 7fffef3d1700  0 -- X.X.16.111:6800/3046050 >>
> X.X.16.114:0/26738200 pipe(0x1e1be80 sd=23 :6800 s=2 pgs=2 cs=1 l=0
> c=0x1e04ba0).fault with nothing to send, going to standby
> 2014-11-17 18:01:47.886662 7fffef1cf700  0 -- X.X.16.111:6800/3046050 >>
> X.X.16.114:0/3673954317 pipe(0x1e1c380 sd=24 :6800 s=0 pgs=0 cs=0 l=0
> c=0x1e05540).accept peer addr is really X.X.16.114:0/3673954317 (socket is
> X.X.16.114:34718/0)
> 2014-11-17 18:03:47.885488 7fffef1cf700  0 -- X.X.16.111:6800/3046050 >>
> X.X.16.114:0/3673954317 pipe(0x1e1c380 sd=24 :6800 s=2 pgs=2 cs=1 l=0
> c=0x1e05540).fault with nothing to send, going to standby
> 2014-11-17 18:04:47.888983 7fffeefcd700  0 -- X.X.16.111:6800/3046050 >>
> X.X.16.114:0/3403131574 pipe(0x1e18a00 sd=25 :6800 s=0 pgs=0 cs=0 l=0
> c=0x1e05280).accept peer addr is really X.X.16.114:0/3403131574 (socket is
> X.X.16.114:34744/0)
> 2014-11-17 18:06:47.888427 7fffeefcd700  0 -- X.X.16.111:6800/3046050 >>
> X.X.16.114:0/3403131574 pipe(0x1e18a00 sd=25 :6800 s=2 pgs=2 cs=1 l=0
> c=0x1e05280).fault with nothing to send, going to standby
> 2014-11-17 20:02:03.558250 7ffff07de700 -1 mds.0.1 *** got signal Terminated
> ***
> 2014-11-17 20:02:03.558297 7ffff07de700  1 mds.0.1 suicide.  wanted
> down:dne, now up:active
> 2014-11-17 20:02:56.053339 7ffff7fe77a0  0 ceph version 0.80.5
> (38b73c67d375a2552d8ed67843c8a65c2c0feba6), process ceph-mds, pid 3424727
> 2014-11-17 20:02:56.121367 7ffff30e4700  1 mds.-1.0 handle_mds_map standby
> 2014-11-17 20:02:56.124343 7ffff30e4700  1 mds.0.2 handle_mds_map i am now
> mds.0.2
> 2014-11-17 20:02:56.124345 7ffff30e4700  1 mds.0.2 handle_mds_map state
> change up:standby --> up:replay
> 2014-11-17 20:02:56.124348 7ffff30e4700  1 mds.0.2 replay_start
> 2014-11-17 20:02:56.124359 7ffff30e4700  1 mds.0.2  recovery set is
> 2014-11-17 20:02:56.124362 7ffff30e4700  1 mds.0.2  need osdmap epoch 93,
> have 92
> 2014-11-17 20:02:56.124363 7ffff30e4700  1 mds.0.2  waiting for osdmap 93
> (which blacklists prior instance)
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux