Re: Upgrade from Jewel to Luminous. REQUIRE_JEWEL OSDMap

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, 28 Nov 2017, Cary wrote:
> I get this error when I try to start the OSD that has been downgraded
> to 10.2.3-r2.
> 
> 2017-11-28 03:42:35.989754 7fa5e6429940  1
> filestore(/var/lib/ceph/osd/ceph-3) upgrade
> 2017-11-28 03:42:35.989788 7fa5e6429940  2 osd.3 0 boot
> 2017-11-28 03:42:35.990132 7fa5e6429940 -1 osd.3 0 The disk uses
> features unsupported by the executable.
> 2017-11-28 03:42:35.990142 7fa5e6429940 -1 osd.3 0  ondisk features
> compat={},rocompat={},incompat={1=initial feature set(~v.18),2=pginfo
> object,3=object
> locator,4=last_epoch_clean,5=categories,6=hobjectpool,7=biginfo,8=leveldbinfo,9=leveldblog,10=snapmapper,12=transaction
> hints,13=pg meta object,14=explicit missing set,15=fastinfo pg
> attr,16=deletes in missing set}
> 2017-11-28 03:42:35.990150 7fa5e6429940 -1 osd.3 0  daemon features
> compat={},rocompat={},incompat={1=initial feature set(~v.18),2=pginfo
> object,3=object
> locator,4=last_epoch_clean,5=categories,6=hobjectpool,7=biginfo,8=leveldbinfo,9=leveldblog,10=snapmapper,11=sharded
> objects,12=transaction hints,13=pg meta object}
> 2017-11-28 03:42:35.990160 7fa5e6429940 -1 osd.3 0 Cannot write to
> disk! Missing features: compat={},rocompat={},incompat={14=explicit
> missing set,15=fastinfo pg attr,16=deletes in missing set}
> 2017-11-28 03:42:35.990775 7fa5e6429940  1 journal close
> /dev/disk/by-partlabel/ceph-3
> 2017-11-28 03:42:35.992960 7fa5e6429940 -1  ** ERROR: osd init failed:
> (95) Operation not supported

Oh, right.  In that case, install the 'luminous' branch[1] on the monitors 
(or just the primary monitor if you're being conservative), restrart it, 
and you'll be able to do

 ceph osd set require_jewel_osds --yes-i-really-mean-it

sage


[1] ceph-deploy install --dev luminous HOST




> Cary
> 
> On Tue, Nov 28, 2017 at 3:09 AM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
> > On Tue, 28 Nov 2017, Cary wrote:
> >> Hello,
> >>
> >>  Could someone please help me complete my botched upgrade from Jewel
> >> 10.2.3-r1 to Luminous 12.2.1. I have 9 Gentoo servers, 4 of which have
> >> 2 OSDs each.
> >>
> >>  My OSD servers were accidentally rebooted before the monitor servers
> >> causing them to be running Luminous before the monitors. All services
> >> have been restarted and running ceph versions gives the following:
> >>
> >> # ceph versions
> >> 2017-11-27 21:27:24.356940 7fed67efe700 -1 WARNING: the following
> >> dangerous and experimental features are enabled: btrfs
> >> 2017-11-27 21:27:24.368469 7fed67efe700 -1 WARNING: the following
> >> dangerous and experimental features are enabled: btrfs
> >>
> >>     "mon": {
> >>         "ceph version 12.2.1
> >> (3e7492b9ada8bdc9a5cd0feafd42fbca27f9c38e) luminous (stable)": 4
> >>     },
> >>     "mgr": {
> >>         "ceph version 12.2.1
> >> (3e7492b9ada8bdc9a5cd0feafd42fbca27f9c38e) luminous (stable)": 3
> >>     },
> >>     "osd": {},
> >>     "mds": {
> >>         "ceph version 12.2.1
> >> (3e7492b9ada8bdc9a5cd0feafd42fbca27f9c38e) luminous (stable)": 1
> >>     },
> >>     "overall": {
> >>         "ceph version 12.2.1
> >> (3e7492b9ada8bdc9a5cd0feafd42fbca27f9c38e) luminous (stable)": 8
> >>
> >>
> >>
> >> For some reason the OSDs do not show what version they are running,
> >> and a ceph osd tree shows all of the OSD as being down.
> >>
> >>  # ceph osd tree
> >> 2017-11-27 21:32:51.969335 7f483d9c2700 -1 WARNING: the following
> >> dangerous and experimental features are enabled: btrfs
> >> 2017-11-27 21:32:51.980976 7f483d9c2700 -1 WARNING: the following
> >> dangerous and experimental features are enabled: btrfs
> >> ID CLASS WEIGHT   TYPE NAME              STATUS REWEIGHT PRI-AFF
> >> -1       27.77998 root default
> >> -3       27.77998     datacenter DC1
> >> -6       27.77998         rack 1B06
> >> -5        6.48000             host ceph3
> >>  1        1.84000                 osd.1    down        0 1.00000
> >>  3        4.64000                 osd.3    down        0 1.00000
> >> -2        5.53999             host ceph4
> >>  5        4.64000                 osd.5    down        0 1.00000
> >>  8        0.89999                 osd.8    down        0 1.00000
> >> -4        9.28000             host ceph6
> >>  0        4.64000                 osd.0    down        0 1.00000
> >>  2        4.64000                 osd.2    down        0 1.00000
> >> -7        6.48000             host ceph7
> >>  6        4.64000                 osd.6    down        0 1.00000
> >>  7        1.84000                 osd.7    down        0 1.00000
> >>
> >> The OSD logs all have this message:
> >>
> >> 20235 osdmap REQUIRE_JEWEL OSDMap flag is NOT set; please set it.
> >
> > THis is an annoying corner condition.  12.2.2 (out soon!)  will have a
> > --force option to set the flag even tho no osds are up.  Until then, the
> > workaround is to downgrade one host to jewel, start one jewel osd, then
> > set the flag.  Then upgrade to luminous again and restart all osds.
> >
> > sage
> >
> >
> >>
> >> When I try to set it with "ceph osd set require_jewel_osds" I get this error:
> >>
> >> Error EPERM: not all up OSDs have CEPH_FEATURE_SERVER_JEWEL feature
> >>
> >>
> >>
> >> A "ceph features" returns:
> >>
> >>     "mon": {
> >>         "group": {
> >>             "features": "0x1ffddff8eea4fffb",
> >>             "release": "luminous",
> >>             "num": 4
> >>         }
> >>     },
> >>     "mds": {
> >>         "group": {
> >>             "features": "0x1ffddff8eea4fffb",
> >>             "release": "luminous",
> >>             "num": 1
> >>         }
> >>     },
> >>     "osd": {
> >>         "group": {
> >>             "features": "0x1ffddff8eea4fffb",
> >>             "release": "luminous",
> >>             "num": 8
> >>         }
> >>     },
> >>     "client": {
> >>         "group": {
> >>             "features": "0x1ffddff8eea4fffb",
> >>             "release": "luminous",
> >>             "num": 3
> >>
> >>  # ceph tell osd.* versions
> >> 2017-11-28 02:29:28.565943 7f99c6aee700 -1 WARNING: the following
> >> dangerous and experimental features are enabled: btrfs
> >> 2017-11-28 02:29:28.578956 7f99c6aee700 -1 WARNING: the following
> >> dangerous and experimental features are enabled: btrfs
> >> Error ENXIO: problem getting command descriptions from osd.0
> >> osd.0: problem getting command descriptions from osd.0
> >> Error ENXIO: problem getting command descriptions from osd.1
> >> osd.1: problem getting command descriptions from osd.1
> >> Error ENXIO: problem getting command descriptions from osd.2
> >> osd.2: problem getting command descriptions from osd.2
> >> Error ENXIO: problem getting command descriptions from osd.3
> >> osd.3: problem getting command descriptions from osd.3
> >> Error ENXIO: problem getting command descriptions from osd.5
> >> osd.5: problem getting command descriptions from osd.5
> >> Error ENXIO: problem getting command descriptions from osd.6
> >> osd.6: problem getting command descriptions from osd.6
> >> Error ENXIO: problem getting command descriptions from osd.7
> >> osd.7: problem getting command descriptions from osd.7
> >> Error ENXIO: problem getting command descriptions from osd.8
> >> osd.8: problem getting command descriptions from osd.8
> >>
> >>  # ceph daemon osd.1 status
> >>
> >>     "cluster_fsid": "CENSORED",
> >>     "osd_fsid": "CENSORED",
> >>     "whoami": 1,
> >>     "state": "preboot",
> >>     "oldest_map": 19482,
> >>     "newest_map": 20235,
> >>     "num_pgs": 141
> >>
> >>  # ceph -s
> >> 2017-11-27 22:04:10.372471 7f89a3935700 -1 WARNING: the following
> >> dangerous and experimental features are enabled: btrfs
> >> 2017-11-27 22:04:10.375709 7f89a3935700 -1 WARNING: the following
> >> dangerous and experimental features are enabled: btrfs
> >>   cluster:
> >>     id:     CENSORED
> >>     health: HEALTH_ERR
> >>             513 pgs are stuck inactive for more than 60 seconds
> >>             126 pgs backfill_wait
> >>             52 pgs backfilling
> >>             435 pgs degraded
> >>             513 pgs stale
> >>             435 pgs stuck degraded
> >>             513 pgs stuck stale
> >>             435 pgs stuck unclean
> >>             435 pgs stuck undersized
> >>             435 pgs undersized
> >>             recovery 854719/3688140 objects degraded (23.175%)
> >>             recovery 838607/3688140 objects misplaced (22.738%)
> >>             mds cluster is degraded
> >>             crush map has straw_calc_version=0
> >>
> >>   services:
> >>     mon: 4 daemons, quorum 0,1,3,2
> >>     mgr: 0(active), standbys: 1, 5
> >>     mds: cephfs-1/1/1 up  {0=a=up:replay}, 1 up:standby
> >>     osd: 8 osds: 0 up, 0 in
> >>
> >>   data:
> >>     pools:   7 pools, 513 pgs
> >>     objects: 1199k objects, 4510 GB
> >>     usage:   13669 GB used, 15150 GB / 28876 GB avail
> >>     pgs:     854719/3688140 objects degraded (23.175%)
> >>              838607/3688140 objects misplaced (22.738%)
> >>              257 stale+active+undersized+degraded
> >>              126 stale+active+undersized+degraded+remapped+backfill_wait
> >>              78  stale+active+clean
> >>              52  stale+active+undersized+degraded+remapped+backfilling
> >>
> >>
> >> I ran "ceph auth list", and client.admin has the following permissions.
> >> auid: 0
> >> caps: [mds] allow
> >> caps: [mgr] allow *
> >> caps: [mon] allow *
> >> caps: [osd] allow *
> >>
> >> Thank you for your time.
> >>
> >> Is there any way I can get these OSDs to join the cluster now, or
> >> recover my data?
> >>
> >> Cary
> >> -Dynamic
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> the body of a message to majordomo@xxxxxxxxxxxxxxx
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>
> >>
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux