Marco, Could you please share what was done to make your cluster stable again? On Fri, May 1, 2020 at 4:47 PM Marco Pizzolo <marcopizzolo@xxxxxxxxx> wrote: > > Thanks Everyone, > > I was able to address the issue at least temporarily. The filesystem and > MDSes are for the time staying online and the pgs are being remapped. > > What i'm not sure about is the best tuning for MDS given our use case, nor > am i sure of exactly what caused the OSD to flap as they did, so I don't > yet know how to avoid a recurrence. > > I do very much like Ceph though.... > > Best wishes, > > Marco > > On Fri, May 1, 2020 at 3:49 PM Marco Pizzolo <marcopizzolo@xxxxxxxxx> wrote: > > > Understood Paul, thanks. > > > > In case this helps to shed any further light...Digging through logs I'm > > also seeing this: > > > > 2020-05-01 10:06:55.984 7eff10cc3700 1 mds.prdceph01 Updating MDS map to > > version 1487236 from mon.2 > > 2020-05-01 10:06:56.398 7eff0e4be700 0 log_channel(cluster) log [WRN] : > > 17 slow requests, 1 included below; oldest blocked for > 254.203584 secs > > 2020-05-01 10:06:56.398 7eff0e4be700 0 log_channel(cluster) log [WRN] : > > slow request 60.552277 seconds old, received at 2020-05-01 10:05:55.846466: > > client_request(client.2525280:277916371 mkdir #0x10014e76974/1f 2020-05-01 > > 10:05:55.844490 caller_uid=1010, caller_gid=1015{}) currently submit entry: > > journal_and_reply > > 2020-05-01 10:06:57.400 7eff0e4be700 0 log_channel(cluster) log [WRN] : > > 17 slow requests, 2 included below; oldest blocked for > 255.205489 secs > > 2020-05-01 10:06:57.400 7eff0e4be700 0 log_channel(cluster) log [WRN] : > > slow request 60.564545 seconds old, received at 2020-05-01 10:05:56.836104: > > client_request(client.2525280:277921203 create > > #0x10014f12b86/9254b3f0-1d5a-4e88-8d41-d36f244bcb12.zip 2020-05-01 > > 10:05:56.834494 caller_uid=1010, caller_gid=1015{}) currently submit entry: > > journal_and_reply > > 2020-05-01 10:06:57.400 7eff0e4be700 0 log_channel(cluster) log [WRN] : > > slow request 60.550874 seconds old, received at 2020-05-01 10:05:56.849775: > > client_request(client.2525280:277921267 mkdir #0x10014e78bec/e0 2020-05-01 > > 10:05:56.848494 caller_uid=1010, caller_gid=1015{}) currently submit entry: > > journal_and_reply > > 2020-05-01 10:06:58.400 7eff0e4be700 0 log_channel(cluster) log [WRN] : > > 17 slow requests, 0 included below; oldest blocked for > 256.205519 secs > > 2020-05-01 10:07:15.250 7eff0dcbd700 1 heartbeat_map is_healthy 'MDSRank' > > had timed out after 15 > > 2020-05-01 10:07:15.250 7eff0dcbd700 0 mds.beacon.prdceph01 Skipping > > beacon heartbeat to monitors (last acked 3.99999s ago); MDS internal > > heartbeat is not healthy! > > 2020-05-01 10:07:15.750 7eff0dcbd700 1 heartbeat_map is_healthy 'MDSRank' > > had timed out after 15 > > 2020-05-01 10:07:15.750 7eff0dcbd700 0 mds.beacon.prdceph01 Skipping > > beacon heartbeat to monitors (last acked 4.49999s ago); MDS internal > > heartbeat is not healthy! > > 2020-05-01 10:07:16.250 7eff0dcbd700 1 heartbeat_map is_healthy 'MDSRank' > > had timed out after 15 > > 2020-05-01 10:07:16.250 7eff0dcbd700 0 mds.beacon.prdceph01 Skipping > > beacon heartbeat to monitors (last acked 4.99998s ago); MDS internal > > heartbeat is not healthy! > > 2020-05-01 10:07:16.750 7eff0dcbd700 1 heartbeat_map is_healthy 'MDSRank' > > had timed out after 15 > > 2020-05-01 10:07:16.750 7eff0dcbd700 0 mds.beacon.prdceph01 Skipping > > beacon heartbeat to monitors (last acked 5.49998s ago); MDS internal > > heartbeat is not healthy! > > 2020-05-01 10:07:17.250 7eff0dcbd700 1 heartbeat_map is_healthy 'MDSRank' > > had timed out after 15 > > 2020-05-01 10:07:17.250 7eff0dcbd700 0 mds.beacon.prdceph01 Skipping > > beacon heartbeat to monitors (last acked 5.99998s ago); MDS internal > > heartbeat is not healthy! > > > > > > THEN about 5 minutes later....... > > > > > > > > 2020-05-01 10:07:35.559 7eff10cc3700 1 mds.prdceph01 9: 'ceph' > > 2020-05-01 10:07:35.559 7eff10cc3700 1 mds.prdceph01 respawning with exe > > /usr/bin/ceph-mds > > 2020-05-01 10:07:35.559 7eff10cc3700 1 mds.prdceph01 exe_path > > /proc/self/exe > > 2020-05-01 10:07:50.785 7fbff66291c0 0 ceph version 14.2.9 > > (581f22da52345dba46ee232b73b990f06029a2a0) nautilus (stable), process > > ceph-mds, pid 9710 > > 2020-05-01 10:07:50.787 7fbff66291c0 0 pidfile_write: ignore empty > > --pid-file > > 2020-05-01 10:07:50.817 7fbfe4408700 1 mds.prdceph01 Updating MDS map to > > version 1487238 from mon.2 > > 2020-05-01 10:07:55.820 7fbfe4408700 1 mds.prdceph01 Updating MDS map to > > version 1487239 from mon.2 > > 2020-05-01 10:07:55.820 7fbfe4408700 1 mds.prdceph01 Map has assigned me > > to become a standby > > 2020-05-01 10:11:07.369 7fbfe4408700 1 mds.prdceph01 Updating MDS map to > > version 1487282 from mon.2 > > 2020-05-01 10:11:07.373 7fbfe4408700 1 mds.0.1487282 handle_mds_map i am > > now mds.0.1487282 > > 2020-05-01 10:11:07.373 7fbfe4408700 1 mds.0.1487282 handle_mds_map state > > change up:boot --> up:replay > > 2020-05-01 10:11:07.373 7fbfe4408700 1 mds.0.1487282 replay_start > > 2020-05-01 10:11:07.374 7fbfe4408700 1 mds.0.1487282 recovery set is > > 2020-05-01 10:11:07.374 7fbfe4408700 1 mds.0.1487282 waiting for osdmap > > 2198096 (which blacklists prior instance) > > 2020-05-01 10:11:09.517 7fbfdd3fa700 0 mds.0.cache creating system inode > > with ino:0x100 > > 2020-05-01 10:11:09.518 7fbfdd3fa700 0 mds.0.cache creating system inode > > with ino:0x1 > > 2020-05-01 10:11:09.753 7fbfddbfb700 -1 > > /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/gigantic/release/14.2.9/rpm/el7/BUILD/ceph-14.2.9/src/mds/OpenFileTable.cc: > > In function 'void OpenFileTable::_load_finish(int, int, int, unsigned int, > > bool, bool, ceph::bufferlist&, std::map<std::basic_string<char>, > > ceph::buffer::v14_2_0::list>&)' thread 7fbfddbfb700 time 2020-05-01 > > 10:11:09.752945 > > /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/gigantic/release/14.2.9/rpm/el7/BUILD/ceph-14.2.9/src/mds/OpenFileTable.cc: > > 777: FAILED ceph_assert(omap_num_objs == num_objs) > > > > ceph version 14.2.9 (581f22da52345dba46ee232b73b990f06029a2a0) nautilus > > (stable) > > 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char > > const*)+0x14a) [0x7fbfed691875] > > 2: (()+0x253a3d) [0x7fbfed691a3d] > > 3: (OpenFileTable::_load_finish(int, int, int, unsigned int, bool, bool, > > ceph::buffer::v14_2_0::list&, std::map<std::string, > > ceph::buffer::v14_2_0::list, std::less<std::string>, > > std::allocator<std::pair<std::string const, ceph::buffer::v14_2_0::list> > > > >&)+0x1bb0) [0x563c26b1e8b0] > > 4: (C_IO_OFT_Load::finish(int)+0x3b) [0x563c26b2467b] > > 5: (MDSContext::complete(int)+0x74) [0x563c26afbce4] > > 6: (MDSIOContextBase::complete(int)+0x177) [0x563c26afbf47] > > 7: (Finisher::finisher_thread_entry()+0x16f) [0x7fbfed71a44f] > > 8: (()+0x7e65) [0x7fbfeb551e65] > > 9: (clone()+0x6d) [0x7fbfea1ff88d] > > > > 2020-05-01 10:11:09.755 7fbfddbfb700 -1 *** Caught signal (Aborted) ** > > in thread 7fbfddbfb700 thread_name:fn_anonymous > > > > ceph version 14.2.9 (581f22da52345dba46ee232b73b990f06029a2a0) nautilus > > (stable) > > 1: (()+0xf5f0) [0x7fbfeb5595f0] > > 2: (gsignal()+0x37) [0x7fbfea137337] > > 3: (abort()+0x148) [0x7fbfea138a28] > > 4: (ceph::__ceph_assert_fail(char const*, char const*, int, char > > const*)+0x199) [0x7fbfed6918c4] > > 5: (()+0x253a3d) [0x7fbfed691a3d] > > 6: (OpenFileTable::_load_finish(int, int, int, unsigned int, bool, bool, > > ceph::buffer::v14_2_0::list&, std::map<std::string, > > ceph::buffer::v14_2_0::list, std::less<std::string>, > > std::allocator<std::pair<std::string const, ceph::buffer::v14_2_0::list> > > > >&)+0x1bb0) [0x563c26b1e8b0] > > 7: (C_IO_OFT_Load::finish(int)+0x3b) [0x563c26b2467b] > > 8: (MDSContext::complete(int)+0x74) [0x563c26afbce4] > > 9: (MDSIOContextBase::complete(int)+0x177) [0x563c26afbf47] > > 10: (Finisher::finisher_thread_entry()+0x16f) [0x7fbfed71a44f] > > 11: (()+0x7e65) [0x7fbfeb551e65] > > 12: (clone()+0x6d) [0x7fbfea1ff88d] > > NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed > > to interpret this. > > > > > > I've taken the filesystem down to see if allowing autonomous healing of > > the undersized and degraded PGs (15 at this time) will help, but would > > greatly appreciate any assistance/guidance anyone can offer. > > > > Thanks in advance, > > > > Marco > > > > On Fri, May 1, 2020 at 3:36 PM Paul Emmerich <paul.emmerich@xxxxxxxx> > > wrote: > > > >> > >> On Fri, May 1, 2020 at 9:27 PM Paul Emmerich <paul.emmerich@xxxxxxxx> > >> wrote: > >> > >>> The OpenFileTable objects are safe to delete while the MDS is offline > >>> anyways, the RADOS object names are mds*_openfiles* > >>> > >> > >> I should clarify this a little bit: you shouldn't touch the CephFS > >> internal state or data structures unless you know *exactly* what you are > >> doing. > >> > >> However, it is pretty safe to delete these files in general, running a > >> scrub afterwards is a good idea anyways. > >> > >> But only do this after reading up on lots of details or consulting an > >> expert. > >> My assessment here is purely an educated guess based on the error and can > >> be wrong or counter-productive. All of my mailing list advice is just > >> things that I know off the top of my head with no further research. > >> Take with a grain of salt. Don't touch stuff that you don't understand if > >> you have important data in there. > >> > >> > >> Paul > >> > >> > >> -- > >> Paul Emmerich > >> > >> Looking for help with your Ceph cluster? Contact us at https://croit.io > >> > >> croit GmbH > >> Freseniusstr. 31h > >> 81247 München > >> www.croit.io > >> Tel: +49 89 1896585 90 > >> > >> > >>> > >>> > >>> > >>> Paul > >>> > >>> -- > >>> Paul Emmerich > >>> > >>> Looking for help with your Ceph cluster? Contact us at https://croit.io > >>> > >>> croit GmbH > >>> Freseniusstr. 31h > >>> 81247 München > >>> www.croit.io > >>> Tel: +49 89 1896585 90 > >>> > >>> > >>> On Fri, May 1, 2020 at 9:04 PM Marco Pizzolo <marcopizzolo@xxxxxxxxx> > >>> wrote: > >>> > >>>> Also seeing errors such as this: > >>>> > >>>> > >>>> [2020-05-01 13:15:20,970][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:20,970][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 11 > >>>> [2020-05-01 13:15:20,974][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.13 with osd_fsid > >>>> dd49cd80-418e-4a8c-8ebf-a33d339663ff > >>>> [2020-05-01 13:15:20,989][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:20,989][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 11 > >>>> [2020-05-01 13:15:20,998][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.5 with osd_fsid > >>>> 4eaf2baa-60f2-4045-8964-6152608c742a > >>>> [2020-05-01 13:15:21,014][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:21,014][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 11 > >>>> [2020-05-01 13:15:21,019][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.9 with osd_fsid > >>>> 32f4a716-f26e-4579-a074-5d6452c22e34 > >>>> [2020-05-01 13:15:21,035][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:21,035][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 11 > >>>> [2020-05-01 13:15:25,972][ceph_volume.process][INFO ] Running command: > >>>> /usr/sbin/ceph-volume lvm trigger 1-0f0e6dd7-9dd8-4b48-beaa-084f55f73b32 > >>>> [2020-05-01 13:15:25,994][ceph_volume.process][INFO ] Running command: > >>>> /usr/sbin/ceph-volume lvm trigger > >>>> 13-dd49cd80-418e-4a8c-8ebf-a33d339663ff > >>>> [2020-05-01 13:15:26,020][ceph_volume.process][INFO ] Running command: > >>>> /usr/sbin/ceph-volume lvm trigger 5-4eaf2baa-60f2-4045-8964-6152608c742a > >>>> [2020-05-01 13:15:26,040][ceph_volume.process][INFO ] Running command: > >>>> /usr/sbin/ceph-volume lvm trigger 9-32f4a716-f26e-4579-a074-5d6452c22e34 > >>>> [2020-05-01 13:15:26,388][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.1 with osd_fsid > >>>> 0f0e6dd7-9dd8-4b48-beaa-084f55f73b32 > >>>> [2020-05-01 13:15:26,389][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.13 with osd_fsid > >>>> dd49cd80-418e-4a8c-8ebf-a33d339663ff > >>>> [2020-05-01 13:15:26,391][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.5 with osd_fsid > >>>> 4eaf2baa-60f2-4045-8964-6152608c742a > >>>> [2020-05-01 13:15:26,402][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:26,403][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 10 > >>>> [2020-05-01 13:15:26,403][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:26,404][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 10 > >>>> [2020-05-01 13:15:26,404][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:26,405][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 10 > >>>> [2020-05-01 13:15:26,411][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.9 with osd_fsid > >>>> 32f4a716-f26e-4579-a074-5d6452c22e34 > >>>> [2020-05-01 13:15:26,424][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:26,424][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 10 > >>>> [2020-05-01 13:15:31,408][ceph_volume.process][INFO ] Running command: > >>>> /usr/sbin/ceph-volume lvm trigger 1-0f0e6dd7-9dd8-4b48-beaa-084f55f73b32 > >>>> [2020-05-01 13:15:31,408][ceph_volume.process][INFO ] Running command: > >>>> /usr/sbin/ceph-volume lvm trigger 5-4eaf2baa-60f2-4045-8964-6152608c742a > >>>> [2020-05-01 13:15:31,409][ceph_volume.process][INFO ] Running command: > >>>> /usr/sbin/ceph-volume lvm trigger > >>>> 13-dd49cd80-418e-4a8c-8ebf-a33d339663ff > >>>> [2020-05-01 13:15:31,429][ceph_volume.process][INFO ] Running command: > >>>> /usr/sbin/ceph-volume lvm trigger 9-32f4a716-f26e-4579-a074-5d6452c22e34 > >>>> [2020-05-01 13:15:31,743][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.5 with osd_fsid > >>>> 4eaf2baa-60f2-4045-8964-6152608c742a > >>>> [2020-05-01 13:15:31,750][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.13 with osd_fsid > >>>> dd49cd80-418e-4a8c-8ebf-a33d339663ff > >>>> [2020-05-01 13:15:31,752][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:31,752][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 9 > >>>> [2020-05-01 13:15:31,754][ceph_volume.process][INFO ] stderr --> > >>>> RuntimeError: could not find osd.1 with osd_fsid > >>>> 0f0e6dd7-9dd8-4b48-beaa-084f55f73b32 > >>>> [2020-05-01 13:15:31,761][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:31,762][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 9 > >>>> [2020-05-01 13:15:31,764][systemd][WARNING] command returned non-zero > >>>> exit > >>>> status: 1 > >>>> [2020-05-01 13:15:31,765][systemd][WARNING] failed activating OSD, > >>>> retries > >>>> left: 9 > >>>> > >>>> On Fri, May 1, 2020 at 2:23 PM Marco Pizzolo <marcopizzolo@xxxxxxxxx> > >>>> wrote: > >>>> > >>>> > Hi Ashley, > >>>> > > >>>> > Thanks for your response. Nothing that I can think of would have > >>>> > happened. We are using max_mds =1. We do have 4 so used to have 3 > >>>> > standby. Within minutes they all crash. > >>>> > > >>>> > On Fri, May 1, 2020 at 2:21 PM Ashley Merrick < > >>>> singapore@xxxxxxxxxxxxxx> > >>>> > wrote: > >>>> > > >>>> >> Quickly checking the code that calls that assert > >>>> >> > >>>> >> > >>>> >> > >>>> >> > >>>> >> if (version > omap_version) { > >>>> >> > >>>> >> omap_version = version; > >>>> >> > >>>> >> omap_num_objs = num_objs; > >>>> >> > >>>> >> omap_num_items.resize(omap_num_objs); > >>>> >> > >>>> >> journal_state = jstate; > >>>> >> > >>>> >> } else if (version == omap_version) { > >>>> >> > >>>> >> ceph_assert(omap_num_objs == num_objs); > >>>> >> > >>>> >> if (jstate > journal_state) > >>>> >> > >>>> >> journal_state = jstate; > >>>> >> > >>>> >> } > >>>> >> > >>>> >> } > >>>> >> > >>>> >> > >>>> >> Im not a dev, but not sure if this will help, seems could mean that > >>>> MDS > >>>> >> thinks its behind on omaps/too far ahead. > >>>> >> > >>>> >> > >>>> >> Anything happened recently? Just running a single MDS? > >>>> >> > >>>> >> > >>>> >> Hopefully someone else may see this and shine some light on what > >>>> could be > >>>> >> causing it. > >>>> >> > >>>> >> > >>>> >> > >>>> >> ---- On Sat, 02 May 2020 02:10:58 +0800 marcopizzolo@xxxxxxxxx > >>>> wrote ---- > >>>> >> > >>>> >> > >>>> >> Hello, > >>>> >> > >>>> >> Hoping you can help me. > >>>> >> > >>>> >> Ceph had been largely problem free for us for the better part of a > >>>> year. > >>>> >> We have a high file count in a single CephFS filesystem, and are > >>>> seeing > >>>> >> this error in the logs: > >>>> >> > >>>> >> > >>>> >> > >>>> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/gigantic/release/14.2.9/rpm/el7/BUILD/ceph-14.2.9/src/mds/OpenFileTable.cc: > >>>> >> 777: FAILED ceph_assert(omap_num_objs == num_objs) > >>>> >> > >>>> >> The issued seemed to occur this morning, and restarting the MDS as > >>>> well as > >>>> >> rebooting the servers doesn't correct the problem. > >>>> >> > >>>> >> Not really sure where to look next as the MDS daemons crash. > >>>> >> > >>>> >> Appreciate any help you can provide > >>>> >> > >>>> >> Marco > >>>> >> _______________________________________________ > >>>> >> ceph-users mailing list -- ceph-users@xxxxxxx > >>>> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > >>>> >> _______________________________________________ > >>>> >> ceph-users mailing list -- ceph-users@xxxxxxx > >>>> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > >>>> >> > >>>> > > >>>> _______________________________________________ > >>>> ceph-users mailing list -- ceph-users@xxxxxxx > >>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx > >>>> > >>> > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx