You can find the upstream fix here https://github.com/ceph/ceph/pull/46833 Thanks, Kotresh HR On Mon, Sep 26, 2022 at 3:17 PM Dhairya Parmar <dparmar@xxxxxxxxxx> wrote: > Patch for this has already been merged and backported to quincy as well. It > will be there in the next Quincy release. > > On Thu, Sep 22, 2022 at 5:12 PM E Taka <0etaka0@xxxxxxxxx> wrote: > > > Ceph 17.2.3 (dockerized in Ubuntu 20.04) > > > > The subject says it. The MDS process always crashes after evicting. ceph > -w > > shows: > > > > 2022-09-22T13:26:23.305527+0200 mds.ksz-cephfs2.ceph00.kqjdwe [INF] > > Evicting (and blocklisting) client session 5181680 ( > > 10.149.12.21:0/3369570791) > > 2022-09-22T13:26:35.729317+0200 mon.ceph00 [INF] daemon > > mds.ksz-cephfs2.ceph03.vsyrbk restarted > > 2022-09-22T13:26:36.039678+0200 mon.ceph00 [INF] daemon > > mds.ksz-cephfs2.ceph01.xybiqv restarted > > 2022-09-22T13:29:21.000392+0200 mds.ksz-cephfs2.ceph04.ekmqio [INF] > > Evicting (and blocklisting) client session 5249349 ( > > 10.149.12.22:0/2459302619) > > 2022-09-22T13:29:32.069656+0200 mon.ceph00 [INF] daemon > > mds.ksz-cephfs2.ceph01.xybiqv restarted > > 2022-09-22T13:30:00.000101+0200 mon.ceph00 [INF] overall HEALTH_OK > > 2022-09-22T13:30:20.710271+0200 mon.ceph00 [WRN] Health check failed: 1 > > daemons have recently crashed (RECENT_CRASH) > > > > The crash info of the crashed MDS is: > > # ceph crash info > > 2022-09-22T11:26:24.013274Z_b005f3fc-7704-4cfc-96c5-f2a9c993f166 > > { > > "assert_condition": "!mds->is_any_replay()", > > "assert_file": > > > > > "/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/17.2.3/rpm/el8/BUILD/ceph-17.2.3/src/mds/MDLog.cc", > > > > "assert_func": "void MDLog::_submit_entry(LogEvent*, > > MDSLogContextBase*)", > > "assert_line": 283, > > "assert_msg": > > > > > "/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/17.2.3/rpm/el8/BUILD/ceph-17.2.3/src/mds/MDLog.cc: > > In function 'void MDLog::_submit_entry(LogEvent*, MDSLogContextBase*)' > > thread 7f76fa8f6700 time > > > > > 2022-09-22T11:26:23.992050+0000\n/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/17.2.3/rpm/el8/BUILD/ceph-17.2.3/src/mds/MDLog.cc: > > 283: FAILED ceph_assert(!mds->is_any_replay())\n", > > "assert_thread_name": "ms_dispatch", > > "backtrace": [ > > "/lib64/libpthread.so.0(+0x12ce0) [0x7f770231bce0]", > > "gsignal()", > > "abort()", > > "(ceph::__ceph_assert_fail(char const*, char const*, int, char > > const*)+0x1b0) [0x7f770333bcd2]", > > "/usr/lib64/ceph/libceph-common.so.2(+0x283e95) [0x7f770333be95]", > > "(MDLog::_submit_entry(LogEvent*, MDSLogContextBase*)+0x3f) > > [0x55991905efdf]", > > "(Server::journal_close_session(Session*, int, Context*)+0x78c) > > [0x559918d7d63c]", > > "(Server::kill_session(Session*, Context*)+0x212) > [0x559918d7dd92]", > > "(Server::apply_blocklist()+0x10d) [0x559918d7e04d]", > > "(MDSRank::apply_blocklist(std::set<entity_addr_t, > > std::less<entity_addr_t>, std::allocator<entity_addr_t> > const&, > unsigned > > int)+0x34) [0x559918d39d74]", > > "(MDSRankDispatcher::handle_osd_map()+0xf6) [0x559918d3a0b6]", > > "(MDSDaemon::handle_core_message(boost::intrusive_ptr<Message > const> > > const&)+0x39b) [0x559918d2330b]", > > "(MDSDaemon::ms_dispatch2(boost::intrusive_ptr<Message> > > const&)+0xc3) [0x559918d23cc3]", > > "(DispatchQueue::entry()+0x14fa) [0x7f77035c240a]", > > "(DispatchQueue::DispatchThread::entry()+0x11) [0x7f7703679481]", > > "/lib64/libpthread.so.0(+0x81ca) [0x7f77023111ca]", > > "clone()" > > ], > > "ceph_version": "17.2.3", > > "crash_id": > > "2022-09-22T11:26:24.013274Z_b005f3fc-7704-4cfc-96c5-f2a9c993f166", > > "entity_name": "mds.ksz-cephfs2.ceph03.vsyrbk", > > "os_id": "centos", > > "os_name": "CentOS Stream", > > "os_version": "8", > > "os_version_id": "8", > > "process_name": "ceph-mds", > > "stack_sig": > > "b75e46941b5f6b7c05a037f9af5d42bb19d82ab7fc6a3c168533fc31a42b4de8", > > "timestamp": "2022-09-22T11:26:24.013274Z", > > "utsname_hostname": "ceph03", > > "utsname_machine": "x86_64", > > "utsname_release": "5.4.0-125-generic", > > "utsname_sysname": "Linux", > > "utsname_version": "#141-Ubuntu SMP Wed Aug 10 13:42:03 UTC 2022" > > } > > > > (Don't be confused by the time information, "ceph -w" is UTC+2, "crash > > info" is UTC) > > > > Should I report this a bug or did I miss something which caused the > error? > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > > > > -- > *Dhairya Parmar* > > He/Him/His > > Associate Software Engineer, CephFS > > Red Hat Inc. <https://www.redhat.com/> > > dparmar@xxxxxxxxxx > <https://www.redhat.com/> > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx