Could you post your crushmap? PGs mapping to no OSDs is a symptom of something wrong there.
You can stop the osds from changing position at startup with 'osd crush update on start = false':
http://docs.ceph.com/docs/master/rados/operations/crush-map/#crush-location
Josh
Sent: Sep 11, 2017 23:50
To: ceph-users@xxxxxxxxxxxxxx
Subject: [ceph-users] Oeps: lost cluster with: ceph osd require-osd-release luminous
We have a kraken cluster, at the time newly build, with bluestore enabled.
it is 8 systems, with each 10 disks 10TB , and each computer has 1 NVME
2TB disk
3 monitor etc
About 700 TB and 300TB used. Mainly S3 objectstore
Of course there is more to the story: We have one strange thing in our
cluster.
We tried to create two pools of storage, default and ssd, and created a
new crush rule.
Worked without problems for months
But when we restart a computer / nvme-osd, it would "forget" that the
nvme should be connected the SSD pool ( for that particular computer).
Since we don't restart systems, we didn't notice that.
The nvme would appear back a default pool.
When we re-apply the same crush rule again it would go back to the SSD
pool.
All while data kept working on the nvme disks
Clearly something is not ideal there. And luminous has a different
approach to separating SSD from HDD.
So we thought first go to luminous 12.2.0 and later see how we fix this.
We did an upgrade to luminous and that went well. That requires a reboot
/ restart off osd's, so all nvme devices where a default.
Reapplying the crush rule brought them back to the ssd pool.
Also while doing the upgrade we switched off in ceph.conf the rule:
# enable experimental unrecoverable data corrupting features =
bluestore, sine in luminous that was no problem
Everything was working fine.
In Ceph -s we had this health warning
all OSDs are running luminous or later but
require_osd_release < luminous
So i thought i would set the minimum OSD version to luminous with;
ceph osd require-osd-release luminous
To us that seemed nothing more than a minimum software version that was
required to connect tot the cluster
the system answered back
recovery_deletes is set
and that was it, the same second, ceph-s went to "0"
ceph -s
cluster:
id: 5bafad08-31b2-4716-be77-07ad2e2647eb
health: HEALTH_WARN
noout flag(s) set
Reduced data availability: 3248 pgs inactive
Degraded data redundancy: 3248 pgs unclean
services:
mon: 3 daemons, quorum Ceph-Mon1,Ceph-Mon2,Ceph-Mon3
mgr: Ceph-Mon2(active), standbys: Ceph-Mon3, Ceph-Mon1
osd: 88 osds: 88 up, 88 in; 297 remapped pgs
flags noout
data:
pools: 26 pools, 3248 pgs
objects: 0 objects, 0 bytes
usage: 0 kB used, 0 kB / 0 kB avail
pgs: 100.000% pgs unknown
3248 unknown
And it was something like this. The errors (apart from the scrub error)
you see would where from the upgrade / restarting, and I would expect
them to go away very fast.
ceph -s
cluster:
id: 5bafad08-31b2-4716-be77-07ad2e2647eb
health: HEALTH_ERR
385 pgs backfill_wait
5 pgs backfilling
135 pgs degraded
1 pgs inconsistent
1 pgs peering
4 pgs recovering
131 pgs recovery_wait
98 pgs stuck degraded
525 pgs stuck unclean
recovery 119/612465488 objects degraded (0.000%)
recovery 24/612465488 objects misplaced (0.000%)
1 scrub errors
noout flag(s) set
all OSDs are running luminous or later but
require_osd_release < luminous
services:
mon: 3 daemons, quorum Ceph-Mon1,Ceph-Mon2,Ceph-Mon3
mgr: Ceph-Mon2(active), standbys: Ceph-Mon1, Ceph-Mon3
osd: 88 osds: 88 up, 88 in; 387 remapped pgs
flags noout
data:
pools: 26 pools, 3248 pgs
objects: 87862k objects, 288 TB
usage: 442 TB used, 300 TB / 742 TB avail
pgs: 0.031% pgs not active
119/612465488 objects degraded (0.000%)
24/612465488 objects misplaced (0.000%)
2720 active+clean
385 active+remapped+backfill_wait
131 active+recovery_wait+degraded
5 active+remapped+backfilling
4 active+recovering+degraded
1 active+clean+inconsistent
1 peering
1 active+clean+scrubbing+deep
io:
client: 34264 B/s rd, 2091 kB/s wr, 38 op/s rd, 48 op/s wr
recovery: 4235 kB/s, 6 objects/s
current ceph health detail
HEALTH_WARN noout flag(s) set; Reduced data availability: 3248 pgs
inactive; Degraded data redundancy: 3248 pgs unclean
OSDMAP_FLAGS noout flag(s) set
PG_AVAILABILITY Reduced data availability: 3248 pgs inactive
pg 15.7cd is stuck inactive for 24780.157341, current state
unknown, last acting []
pg 15.7ce is stuck inactive for 24780.157341, current state
unknown, last acting []
pg 15.7cf is stuck inactive for 24780.157341, current state
unknown, last acting []
..
pg 15.7ff is stuck inactive for 24728.059692, current state
unknown, last acting []
PG_DEGRADED Degraded data redundancy: 3248 pgs unclean
pg 15.7cd is stuck unclean for 24728.059692, current state unknown,
last acting []
pg 15.7ce is stuck unclean for 24728.059692, current state unknown,
last acting []
....
pg 15.7fc is stuck unclean for 21892.783340, current state unknown,
last acting []
pg 15.7fd is stuck unclean for 21892.783340, current state unknown,
last acting []
pg 15.7fe is stuck unclean for 21892.783340, current state unknown,
last acting []
pg 15.7ff is stuck unclean for 21892.783340, current state unknown,
last acting []
ceph pg dump_stuck unclean | more
15.46b unknown [] -1 [] -1
15.46a unknown [] -1 [] -1
15.469 unknown [] -1 [] -1
15.468 unknown [] -1 [] -1
15.467 unknown [] -1 [] -1
15.466 unknown [] -1 [] -1
15.465 unknown [] -1 [] -1
15.464 unknown [] -1 [] -1
15.463 unknown [] -1 [] -1
....
Any idea's
Greetings JW
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com