On Mon, 25 Feb 2013, Steffen Thorhauer wrote: > Hi, > I've found out, what I make wrong: stop the cluster and forget a client, which as mounting the cephfs. I simply forget the client. > With a > ceph mds newfs 0 1 --yes-i-really-mean-it Oh... the 'newfs' resets the MDSMap in the monitor, but doesn't erase any data objects. It's meant to be run with newly created or empty pools, not ones with existing data objects. sage > (I dont really what the parameters are), but the mds is restarting with > an empty fs. > I tried the patch version and I'am able to reproduce the problem > (the log file is much shorter). > > Regards, > Steffen > > and On Fri, Feb 22, 2013 at 03:25:15PM -0800, Gregory Farnum wrote: > > Hurray! Thanks to the log you sent me (off-list) I was able to > > identify the issue and create a patch. Can you test it and make sure > > things work out? It should be available shortly at > > http://gitbuilder.ceph.com/ceph-deb--x86_64-basic/ref/testing-4248-on-57 > > See http://ceph.com/docs/master/install/debian/#development-testing-packages > > for more instructions if necessary. :) > > Thanks! > > -Greg > > > > On Wed, Feb 20, 2013 at 11:12 PM, Steffen Thorhauer > > <thorhaue@xxxxxxxxxxxxxxxxxxxxxxx> wrote: > > > Hello, > > > I have a test ceph cluster on ubuntu 12.04 and made yesterday an upgrade to 0.57 . > > > But after the upgrade the mds dies. > > > ceph -s says > > > health HEALTH_WARN mds 0 is laggy > > > monmap e2: 5 mons at {0=10.37.124.161:6789/0,1=10.37.124.162:6789/0,2=10.37.124.163:6789/0,3=10.37.124.164:6789/0,4=10.37.124.167:6789/0}, election epoch 46, quorum 0,1,2,3,4 0,1,2,3,4 > > > osdmap e901: 6 osds: 6 up, 6 in > > > pgmap v89987: 1280 pgs: 1280 active+clean; 123 GB data, 373 GB used, 225 GB / 599 GB avail > > > mdsmap e2296: 1/1/1 up {0=0=up:replay(laggy or crashed)} > > > i > > > I get only a crash report in > > > /var/log/ceph/ceph-mds.0.log > > > 0> 2013-02-21 08:04:02.892956 7f8d50955700 -1 *** Caught signal (Aborted) ** > > > in thread 7f8d50955700 > > > > > > ceph version 0.57 (9a7a9d06c0623ccc116a1d3b71c765c20a17e98e) > > > 1: /usr/bin/ceph-mds() [0x81b03a] > > > 2: (()+0xfcb0) [0x7f8d58a59cb0] > > > 3: (gsignal()+0x35) [0x7f8d57835425] > > > 4: (abort()+0x17b) [0x7f8d57838b8b] > > > 5: (__gnu_cxx::__verbose_terminate_handler()+0x11d) [0x7f8d5818769d] > > > 6: (()+0xb5846) [0x7f8d58185846] > > > 7: (()+0xb5873) [0x7f8d58185873] > > > 8: (()+0xb596e) [0x7f8d5818596e] > > > 9: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x1df) [0x7833cf] > > > 10: (EMetaBlob::replay(MDS*, LogSegment*, MDSlaveUpdate*)+0x1e34) [0x4e0d94] > > > 11: (EUpdate::replay(MDS*)+0x3a) [0x4e8e1a] > > > 12: (MDLog::_replay_thread()+0x438) [0x6ad448] > > > 13: (MDLog::ReplayThread::entry()+0xd) [0x4cc0ad] > > > 14: (()+0x7e9a) [0x7f8d58a51e9a] > > > 15: (clone()+0x6d) [0x7f8d578f2cbd] > > > NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. > > > > > > --- logging levels --- > > > 0/ 5 none > > > 0/ 1 lockdep > > > 0/ 1 context > > > 1/ 1 crush > > > 1/ 5 mds > > > 1/ 5 mds_balancer > > > 1/ 5 mds_locker > > > 1/ 5 mds_log > > > 1/ 5 mds_log_expire > > > 1/ 5 mds_migrator > > > 0/ 1 buffer > > > 0/ 1 timer > > > 0/ 1 filer > > > 0/ 1 striper > > > 0/ 1 objecter > > > 0/ 5 rados > > > 0/ 5 rbd > > > 0/ 5 journaler > > > 0/ 5 objectcacher > > > 0/ 5 client > > > 0/ 5 osd > > > 0/ 5 optracker > > > 0/ 5 objclass > > > 1/ 3 filestore > > > 1/ 3 journal > > > 0/ 5 ms > > > 1/ 5 mon > > > 0/10 monc > > > 0/ 5 paxos > > > 0/ 5 tp > > > 1/ 5 auth > > > 1/ 5 crypto > > > 1/ 1 finisher > > > 1/ 5 heartbeatmap > > > 1/ 5 perfcounter > > > 1/ 5 rgw > > > 1/ 5 hadoop > > > 1/ 5 javaclient > > > 1/ 5 asok > > > 1/ 1 throttle > > > -2/-2 (syslog threshold) > > > -1/-1 (stderr threshold) > > > max_recent 100000 > > > max_new 1000 > > > log_file /var/log/ceph/ceph-mds.0.log > > > --- end dump of recent events --- > > > > > > Regards, > > > Steffen Thorhauer > > > > > > _______________________________________________ > > > ceph-users mailing list > > > ceph-users@xxxxxxxxxxxxxx > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -- > ______________________________________________________________________ > Steffen Thorhauer > > email: sth@xxxxxxxxxxxxxxxxxxxxxxx > url: http://wwwiti.cs.uni-magdeburg.de/~thorhaue > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com