Weird pg degradation behavior

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I lost a node due to cpu failure so until hardware fix, I let it down as Ceph marked osds out, however I keep it in the tree because in 1-2 days will be online again.

This is the current state:
    health: HEALTH_WARN                        Degraded data redundancy: 1726/23917349718 objects degraded(0.000%), 1 pg degraded, 1 pg undersized

Out of the 9x osd nodes 2x of them have 4x nvme osd (temporarily) and 7x has only 2x (for index pool on 3 replicas) where this pg is located.
(The down osd node is a 2x nvme osd one)

This is some part of the query of that specific degraded pg:
...
 "up": [
     233,
     202
 ],
 "acting": [
     233,
     202
 ],
 "avail_no_missing": [
     "233",
     "202"
 ],
...

The above mentioned osds are on the 2x osd nodes with 4x nvme.

What is not clear, I don't see any probing osd or any indicator for the 3rd missing part of that specific pg.

This is the pgs_brief:
PG_STAT  STATE                        UP                         UP_PRIMARY  ACTING                     ACTING_PRIMARY
10.f6     active+undersized+degraded                  [233,202]         233                  [233,202]             233

Shouldn't have any indicator what it is looking for?
Which osd or some probing osd? (It's quincy 17.2.7).

If I purge the down osds on the down server,  I guess it would kick off recovery, I'm just curious why don't show anywhere what it is looking for?

Ty
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux