You need to run a full deep scrub before continuing the upgrade, the reason for this is that the deep scrub migrates the format of some snapshot-related on-disk data structure. Looks like you only tried a normal scrub, not a deep-scrub Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io Tel: +49 89 1896585 90 On Thu, Feb 7, 2019 at 4:34 PM Eugen Block <eblock@xxxxxx> wrote: > > Hi, > > could it be a missing 'ceph osd require-osd-release luminous' on your cluster? > > When I check a luminous cluster I get this: > > host1:~ # ceph osd dump | grep recovery > flags sortbitwise,recovery_deletes,purged_snapdirs > > The flags in the code you quote seem related to that. > Can you check that output on your cluster? > > Found this in a thread from last year [1]. > > > Regards, > Eugen > > [1] https://www.spinics.net/lists/ceph-devel/msg40191.html > > Zitat von Andrew Bruce <dbmail1771@xxxxxxxxx>: > > > Hello All! Yesterday started upgrade from luminous to mimic with one > > of my 3 MONs. > > > > After applying mimic yum repo and updating - a restart reports the > > following error from the MON log file: > > > > ==> /var/log/ceph/ceph-mon.lvtncephx121.log <== > > 2019-02-07 10:02:40.110 7fc8283ed700 -1 mon.lvtncephx121@0(probing) > > e4 handle_probe_reply existing cluster has not completed a full > > luminous scrub to purge legacy snapdir objects; please scrub before > > upgrading beyond luminous. > > > > My question is simply: What exactly does this require? > > > > Yesterday afternoon I did a manual: > > > > ceph osd scrub all > > > > But that has zero effect. I still get the same message on restarting the MON > > > > I have no errors in the cluster except for the single MON > > (lvtncephx121) that I'm working to migrate to mimic first: > > > > [root@lvtncephx110 ~]# ceph status > > cluster: > > id: 5fabf1b2-cfd0-44a8-a6b5-fb3fd0545517 > > health: HEALTH_WARN > > 1/3 mons down, quorum lvtncephx122,lvtncephx123 > > > > services: > > mon: 3 daemons, quorum lvtncephx122,lvtncephx123, out of quorum: > > lvtncephx121 > > mgr: lvtncephx122(active), standbys: lvtncephx123, lvtncephx121 > > mds: cephfs-1/1/1 up {0=lvtncephx151=up:active}, 1 up:standby > > osd: 18 osds: 18 up, 18 in > > rgw: 2 daemons active > > > > data: > > pools: 23 pools, 2016 pgs > > objects: 2608k objects, 10336 GB > > usage: 20689 GB used, 39558 GB / 60247 GB avail > > pgs: 2016 active+clean > > > > io: > > client: 5612 B/s rd, 3756 kB/s wr, 1350 op/s rd, 412 op/s wr > > > > FWIW: The source code has the following: > > > > // Monitor.cc > > if (!osdmon()->osdmap.test_flag(CEPH_OSDMAP_PURGED_SNAPDIRS) || > > !osdmon()->osdmap.test_flag(CEPH_OSDMAP_RECOVERY_DELETES)) { > > derr << __func__ << " existing cluster has not completed a > > full luminous" > > << " scrub to purge legacy snapdir objects; please scrub before" > > << " upgrading beyond luminous." << dendl; > > exit(0); > > } > > } > > > > So two question: > > How to show the current flags in the OSD map checked by the monitor? > > How to get these flags set so the MON will actually start. > > > > Thanks, > > Andy > > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com