Oeps: lost cluster with: ceph osd require-osd-release luminous

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We have a kraken cluster,  at the time newly build, with bluestore enabled.
it is 8 systems, with each 10 disks 10TB , and each computer has 1 NVME 2TB disk
3 monitor etc
About 700 TB and 300TB used. Mainly S3 objectstore

Of course there is more to the story: We have one strange thing in our cluster. We tried to create two pools of storage, default and ssd, and created a new crush rule.
Worked without problems for months
But when we restart a computer / nvme-osd, it would "forget" that the nvme should be connected the SSD pool ( for that particular computer).
Since we don't restart systems, we didn't notice that.
The nvme would appear back a default  pool.
When we re-apply the same crush rule again it would go back to the SSD pool.
All while data kept working on the nvme disks

Clearly something is not ideal there. And luminous has a different approach to separating SSD from HDD.
So we thought first go to luminous 12.2.0 and later see how we fix this.

We did an upgrade to luminous and that went well. That requires a reboot / restart off osd's, so all nvme devices where a default.
Reapplying the crush rule  brought them back to the ssd pool.
Also while doing the upgrade we switched off in ceph.conf the rule:
# enable experimental unrecoverable data corrupting features = bluestore, sine in luminous that was no problem

Everything was working fine.
In Ceph -s we had this health warning

all OSDs are running luminous or later but require_osd_release < luminous

So i thought i would set the minimum  OSD version to luminous with;

ceph osd require-osd-release luminous

To us that seemed nothing more than a minimum software version that was required to connect tot the cluster
the system answered back

recovery_deletes is set

and that was it, the same second, ceph-s went to "0"

 ceph -s
  cluster:
    id:     5bafad08-31b2-4716-be77-07ad2e2647eb
    health: HEALTH_WARN
            noout flag(s) set
            Reduced data availability: 3248 pgs inactive
            Degraded data redundancy: 3248 pgs unclean

  services:
    mon: 3 daemons, quorum Ceph-Mon1,Ceph-Mon2,Ceph-Mon3
    mgr: Ceph-Mon2(active), standbys: Ceph-Mon3, Ceph-Mon1
    osd: 88 osds: 88 up, 88 in; 297 remapped pgs
         flags noout

  data:
    pools:   26 pools, 3248 pgs
    objects: 0 objects, 0 bytes
    usage:   0 kB used, 0 kB / 0 kB avail
    pgs:     100.000% pgs unknown
             3248 unknown

And it was something like this. The errors (apart from the scrub error) you see would where from the upgrade / restarting, and I would expect them to go away very fast.

ceph -s
  cluster:
    id:     5bafad08-31b2-4716-be77-07ad2e2647eb
    health: HEALTH_ERR
            385 pgs backfill_wait
            5 pgs backfilling
            135 pgs degraded
            1 pgs inconsistent
            1 pgs peering
            4 pgs recovering
            131 pgs recovery_wait
            98 pgs stuck degraded
            525 pgs stuck unclean
            recovery 119/612465488 objects degraded (0.000%)
            recovery 24/612465488 objects misplaced (0.000%)
            1 scrub errors
            noout flag(s) set
all OSDs are running luminous or later but require_osd_release < luminous

  services:
    mon: 3 daemons, quorum Ceph-Mon1,Ceph-Mon2,Ceph-Mon3
    mgr: Ceph-Mon2(active), standbys: Ceph-Mon1, Ceph-Mon3
    osd: 88 osds: 88 up, 88 in; 387 remapped pgs
         flags noout

  data:
    pools:   26 pools, 3248 pgs
    objects: 87862k objects, 288 TB
    usage:   442 TB used, 300 TB / 742 TB avail
    pgs:     0.031% pgs not active
             119/612465488 objects degraded (0.000%)
             24/612465488 objects misplaced (0.000%)
             2720 active+clean
             385  active+remapped+backfill_wait
             131  active+recovery_wait+degraded
             5    active+remapped+backfilling
             4    active+recovering+degraded
             1    active+clean+inconsistent
             1    peering
             1    active+clean+scrubbing+deep

  io:
    client:   34264 B/s rd, 2091 kB/s wr, 38 op/s rd, 48 op/s wr
    recovery: 4235 kB/s, 6 objects/s

current ceph health detail

HEALTH_WARN noout flag(s) set; Reduced data availability: 3248 pgs inactive; Degraded data redundancy: 3248 pgs unclean
OSDMAP_FLAGS noout flag(s) set
PG_AVAILABILITY Reduced data availability: 3248 pgs inactive
pg 15.7cd is stuck inactive for 24780.157341, current state unknown, last acting [] pg 15.7ce is stuck inactive for 24780.157341, current state unknown, last acting [] pg 15.7cf is stuck inactive for 24780.157341, current state unknown, last acting []
..
pg 15.7ff is stuck inactive for 24728.059692, current state unknown, last acting []
PG_DEGRADED Degraded data redundancy: 3248 pgs unclean
pg 15.7cd is stuck unclean for 24728.059692, current state unknown, last acting [] pg 15.7ce is stuck unclean for 24728.059692, current state unknown, last acting []
....
pg 15.7fc is stuck unclean for 21892.783340, current state unknown, last acting [] pg 15.7fd is stuck unclean for 21892.783340, current state unknown, last acting [] pg 15.7fe is stuck unclean for 21892.783340, current state unknown, last acting [] pg 15.7ff is stuck unclean for 21892.783340, current state unknown, last acting []

 ceph pg dump_stuck unclean | more
15.46b  unknown []         -1     []             -1
15.46a  unknown []         -1     []             -1
15.469  unknown []         -1     []             -1
15.468  unknown []         -1     []             -1
15.467  unknown []         -1     []             -1
15.466  unknown []         -1     []             -1
15.465  unknown []         -1     []             -1
15.464  unknown []         -1     []             -1
15.463  unknown []         -1     []             -1
....


Any idea's

Greetings JW

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux