Eugen, thank you very much. This (ceph1n021) is indeed the one using leveldb (in kv_backend) Other mons have kv_backend "rocksdb" but unfortunately after reinstalling ceph-mon@ceph1n021 we get no ceph status anymore and our mon logs get filed with: 2022-09-28T13:10:54.822+0200 7fbc6a863700 0 log_channel(cluster) log [INF] : mon.ceph1n011 calling monitor election 2022-09-28T13:10:54.824+0200 7fbc6a863700 1 paxos.0).electionLogic(789785) init, last seen epoch 789785, mid-election, bumping 2022-09-28T13:10:54.827+0200 7fbc6a863700 1 mon.ceph1n011@0(electing) e46 collect_metadata sda: no unique device id for sda: fallback method has model 'Virtual disk ' but no serial' 2022-09-28T07:11:37.380-0400 7efc2d6ec700 1 mon.ceph1n020@2(electing) e46 collect_metadata md125: no unique device id for md125: fallback method has no model nor serial 2022-09-28T07:11:37.384-0400 7efc2d6ec700 1 mon.ceph1n020@2(electing) e46 collect_metadata md125: no unique device id for md125: fallback method has no model nor serial 2022-09-28T07:11:37.389-0400 7efc2d6ec700 1 mon.ceph1n020@2(electing) e46 collect_metadata md125: no unique device id for md125: fallback method has no model nor serial 2022-09-28T07:11:37.393-0400 7efc2d6ec700 1 mon.ceph1n020@2(electing) e46 collect_metadata md125: no unique device id for md125: fallback method has no model nor serial [root@ceph1n020 ~]# It seems there is a problem with the /var device or some ... Now we have an urgent state of our produktion cluster. :-( Do you have some hints for us? Best regards, Christoph Am Mi., 28. Sept. 2022 um 12:21 Uhr schrieb Eugen Block <eblock@xxxxxx>: > Hi, > > there was a thread about deprecating leveldb [1], but I didn't get the > impression that it already has been deprecated. But the thread > mentions that it's not tested anymore, so that might explain it. To > confirm that you use leveldb you can run: > > cat /var/lib/ceph/mon/ceph-<MON>/kv_backend > > So you already have successfully upgraded other MONs, what kv_backend > do they use? If this is the last one with leveldb you can probably > move the old store content and recreate an empty MON. > > [1] > > https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/K4OSAA4AJS2V7FQI6GNCKCK3IRQDBQRS/ > > Zitat von "Ackermann, Christoph" <c.ackermann@xxxxxxxxxxxx>: > > > Hello List, > > > > i'm on the way to upgrade our "non cephadm" from Octopus to Quiny. It > > fails/stuck on third ceph-mon ceph1n021 with an strange error: > > > > 2022-09-28T11:04:27.691+0200 7f8681543880 -1 _open error initializing > > leveldb db back storage in /var/lib/ceph/mon/ceph-ceph1n021/store.db > > > > This monitor contains a lot of ldb files, wondering if we do not use > > leveldb anymore.. > > > > [root@ceph1n021 ~]# ls /var/lib/ceph/mon/ceph-ceph1n021/store.db/ > > 3251327.ldb 5720254.ldb 6568574.ldb 6652800.ldb 6726397.ldb > > 6726468.ldb 6726623.ldb 6726631.ldb 6726638.ldb 6726646.ldb > > 6726653.ldb IDENTITY > > 3251520.ldb 6497196.ldb 6575398.ldb 6654280.ldb 6726398.ldb > > 6726469.ldb 6726624.ldb 6726632.ldb 6726639.ldb 6726647.ldb > > 6726654.ldb LOCK > > 3251566.ldb 6517010.ldb 6595757.ldb 6680000.ldb 6726399.ldb > > 6726588.ldb 6726627.ldb 6726634.ldb 6726642.ldb 6726648.ldb > > 6726655.ldb MANIFEST-5682438 > > 3251572.ldb 6523701.ldb 6601653.ldb 6699521.ldb 6726400.ldb > > 6726608.ldb 6726628.ldb 6726635.ldb 6726643.ldb 6726649.ldb > > 6726656.ldb OPTIONS-000005 > > 3251583.ldb 6543819.ldb 6624261.ldb 6706116.ldb 6726401.ldb > > 6726618.log 6726629.ldb 6726636.ldb 6726644.ldb 6726650.ldb > > 6726657.ldb > > 3251584.ldb 6549696.ldb 6627961.ldb 6725307.ldb 6726467.ldb > > 6726622.ldb 6726630.ldb 6726637.ldb 6726645.ldb 6726651.ldb CURRENT > > > > All other ceph-mon "store.db" folder consists only expected files like: > > > > [root@ceph1n020 ~]# ls -l /var/lib/ceph/mon/ceph-ceph1n020/store.db/ > > total 153252 > > -rw-------. 1 ceph ceph 11230512 Sep 28 05:13 1040392.log > > -rw-------. 1 ceph ceph 67281589 Sep 28 05:11 1040394.sst > > -rw-------. 1 ceph ceph 40121324 Sep 28 05:11 1040395.sst > > -rw-------. 1 ceph ceph 16 Aug 19 06:29 CURRENT > > -rw-------. 1 ceph ceph 37 Feb 21 2022 IDENTITY > > -rw-r--r--. 1 ceph ceph 0 Feb 21 2022 LOCK > > -rw-------. 1 ceph ceph 8465618 Sep 28 05:11 MANIFEST-898389 > > -rw-------. 1 ceph ceph 4946 Aug 19 04:51 OPTIONS-898078 > > -rw-------. 1 ceph ceph 4946 Aug 19 06:29 OPTIONS-898392 > > > > > > "mon": { > > "ceph version 15.2.17 (8a82819d84cf884bd39c17e3236e0632ac146dc4) > > octopus (stable)": 3, > > "ceph version 17.2.3 (dff484dfc9e19a9819f375586300b3b79d80034d) > > quincy (stable)": (ceph-mon@ceph1n011 and ceph-mon@ceph1n012) > > > > Is it safe to go forward restarting the rest of these monitors (ceph1n019 > > and ceph1n020) and what can we do to fix errors on ceph-mon@ceph1n021 ? > > > > Best regards, > > Christoph > > > > > > > > Christoph Ackermann | System Engineer > > INFOSERVE GmbH | Am Felsbrunnen 15 | D-66119 Saarbrücken > > Fon +49 (0)681 88008-59 | Fax +49 (0)681 88008-33 | > c.ackermann@xxxxxxxxxxxx > > | www.infoserve.de > > INFOSERVE Datenschutzhinweise: www.infoserve.de/datenschutz > > Handelsregister: Amtsgericht Saarbrücken, HRB 11001 | Erfüllungsort: > > Saarbrücken > > Geschäftsführer: Dr. Stefan Leinenbach | Ust-IdNr.: DE168970599 > > > > <https://facebook.com/infoserve.de> > > <https://www.xing.com/companies/infoservegmbh> > > <https://www.youtube.com/channel/UCUj8C3TGGhQZPVvxu4woXmQ> > > <https://www.linkedin.com/company-beta/10095540> > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx