Re: mds crash

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
I've found out, what I make wrong: stop the cluster and forget a client, which as mounting the cephfs. I simply forget the client. 
With a 
ceph mds newfs 0 1 --yes-i-really-mean-it
(I dont really what the parameters are), but the mds is restarting with
an empty fs. 
I tried the patch version and I'am able to reproduce the problem
(the log file is much shorter).

Regards,
  Steffen

and On Fri, Feb 22, 2013 at 03:25:15PM -0800, Gregory Farnum wrote:
> Hurray! Thanks to the log you sent me (off-list) I was able to
> identify the issue and create a patch. Can you test it and make sure
> things work out? It should be available shortly at
> http://gitbuilder.ceph.com/ceph-deb--x86_64-basic/ref/testing-4248-on-57
> See http://ceph.com/docs/master/install/debian/#development-testing-packages
> for more instructions if necessary. :)
> Thanks!
> -Greg
> 
> On Wed, Feb 20, 2013 at 11:12 PM, Steffen Thorhauer
> <thorhaue@xxxxxxxxxxxxxxxxxxxxxxx> wrote:
> > Hello,
> > I have a test ceph cluster on ubuntu 12.04 and made yesterday an upgrade to 0.57 .
> > But after the upgrade the mds dies.
> > ceph -s says
> >   health HEALTH_WARN mds 0 is laggy
> >    monmap e2: 5 mons at {0=10.37.124.161:6789/0,1=10.37.124.162:6789/0,2=10.37.124.163:6789/0,3=10.37.124.164:6789/0,4=10.37.124.167:6789/0}, election epoch 46, quorum 0,1,2,3,4 0,1,2,3,4
> >    osdmap e901: 6 osds: 6 up, 6 in
> >     pgmap v89987: 1280 pgs: 1280 active+clean; 123 GB data, 373 GB used, 225 GB / 599 GB avail
> >    mdsmap e2296: 1/1/1 up {0=0=up:replay(laggy or crashed)}
> > i
> > I get only a crash report in
> > /var/log/ceph/ceph-mds.0.log
> >    0> 2013-02-21 08:04:02.892956 7f8d50955700 -1 *** Caught signal (Aborted) **
> >  in thread 7f8d50955700
> >
> >  ceph version 0.57 (9a7a9d06c0623ccc116a1d3b71c765c20a17e98e)
> >  1: /usr/bin/ceph-mds() [0x81b03a]
> >  2: (()+0xfcb0) [0x7f8d58a59cb0]
> >  3: (gsignal()+0x35) [0x7f8d57835425]
> >  4: (abort()+0x17b) [0x7f8d57838b8b]
> >  5: (__gnu_cxx::__verbose_terminate_handler()+0x11d) [0x7f8d5818769d]
> >  6: (()+0xb5846) [0x7f8d58185846]
> >  7: (()+0xb5873) [0x7f8d58185873]
> >  8: (()+0xb596e) [0x7f8d5818596e]
> >  9: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x1df) [0x7833cf]
> >  10: (EMetaBlob::replay(MDS*, LogSegment*, MDSlaveUpdate*)+0x1e34) [0x4e0d94]
> >  11: (EUpdate::replay(MDS*)+0x3a) [0x4e8e1a]
> >  12: (MDLog::_replay_thread()+0x438) [0x6ad448]
> >  13: (MDLog::ReplayThread::entry()+0xd) [0x4cc0ad]
> >  14: (()+0x7e9a) [0x7f8d58a51e9a]
> >  15: (clone()+0x6d) [0x7f8d578f2cbd]
> >  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.
> >
> > --- logging levels ---
> >    0/ 5 none
> >    0/ 1 lockdep
> >    0/ 1 context
> >    1/ 1 crush
> >    1/ 5 mds
> >    1/ 5 mds_balancer
> >    1/ 5 mds_locker
> >    1/ 5 mds_log
> >    1/ 5 mds_log_expire
> >    1/ 5 mds_migrator
> >    0/ 1 buffer
> >    0/ 1 timer
> >    0/ 1 filer
> >    0/ 1 striper
> >    0/ 1 objecter
> >    0/ 5 rados
> >    0/ 5 rbd
> >    0/ 5 journaler
> >    0/ 5 objectcacher
> >    0/ 5 client
> >    0/ 5 osd
> >    0/ 5 optracker
> >    0/ 5 objclass
> >    1/ 3 filestore
> >    1/ 3 journal
> >    0/ 5 ms
> >    1/ 5 mon
> >    0/10 monc
> >    0/ 5 paxos
> >    0/ 5 tp
> >    1/ 5 auth
> >    1/ 5 crypto
> >    1/ 1 finisher
> >    1/ 5 heartbeatmap
> >    1/ 5 perfcounter
> >    1/ 5 rgw
> >    1/ 5 hadoop
> >    1/ 5 javaclient
> >    1/ 5 asok
> >    1/ 1 throttle
> >   -2/-2 (syslog threshold)
> >   -1/-1 (stderr threshold)
> >   max_recent    100000
> >   max_new         1000
> >   log_file /var/log/ceph/ceph-mds.0.log
> > --- end dump of recent events ---
> >
> > Regards,
> >   Steffen Thorhauer
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@xxxxxxxxxxxxxx
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

-- 
______________________________________________________________________
Steffen Thorhauer

email: sth@xxxxxxxxxxxxxxxxxxxxxxx
url: http://wwwiti.cs.uni-magdeburg.de/~thorhaue                     
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux