Hello,
Since we upgraded ceph
cluster we are facing a lot of problems. Most of them due to osd
crashing. What can cause this?
This morning I woke up
with thi message:
root@red-compute:~#
ceph -w
cluster 9028f4da-0d77-462b-be9b-dbdf7fa57771
health HEALTH_ERR
1 pgs are stuck inactive for more than 300 seconds
7 pgs inconsistent
1 pgs stale
1 pgs stuck stale
recovery 20266198323167232/287940 objects degraded
(7038340738753.641%)
37154696925806626 scrub errors
too many PGs per OSD (305 > max 300)
monmap e12: 2 mons at
{blue-compute=172.16.0.119:6789/0,red-compute=172.16.0.100:6789/0}
election epoch 4986, quorum 0,1
red-compute,blue-compute
fsmap e913: 1/1/1 up {0=blue-compute=up:active}
osdmap e8096: 5 osds: 5 up, 5 in
flags require_jewel_osds
pgmap v68755349: 764 pgs, 6 pools, 558 GB data, 140 kobjects
1119 GB used, 3060 GB / 4179 GB avail
20266198323167232/287940 objects degraded
(7038340738753.641%)
756 active+clean
7 active+clean+inconsistent
1 stale+active+clean
client io 1630 B/s rd, 552 kB/s wr, 0 op/s rd, 64 op/s wr
2017-10-22 18:10:13.000812 mon.0 [INF] pgmap v68755348: 764 pgs: 7
active+clean+inconsistent, 756 active+clean, 1 stale+active+clean;
558 GB data, 1119 GB used, 3060 GB / 4179 GB avail; 1641 B/s rd,
229 kB/s wr, 39 op/s; 20266198323167232/287940 objects
degraded (7038340738753.641%)
|
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com