Hello!
Today, I started the morning with a WARNING STATUS on our Ceph cluster.
# ceph health detail
HEALTH_WARN Too many repaired reads on 1 OSDs
[WRN] OSD_TOO_MANY_REPAIRS: Too many repaired reads on 1 OSDs
osd.67 had 399911 reads repaired
I made "ceph osd out 67" and PGs where migrated to another OSDs.
I stopped the osd.67 daemon, inspected the logs, etc...
Then I started the daemon and made "# ceph osd in 67".
OSD started backfilling with some PGs and no other error appeared in the
rest of the day, but Warning status still remains.
Can I clear it? Shoud I remove the osd and start with a new one?
Thanks in advance for your time!
Javier.-
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx