Good to know. I tried reset-failed and restart several times, it didn't work on any of them. I also rebooted one of the hosts, didn't help. Thankfully it seems they failed far enough apart that our nearly-empty cluster rebuilt in time. But it's rather worrying. On Fri, Jul 19, 2019 at 10:09 PM Nigel Williams <nigel.williams@xxxxxxxxxxx> wrote: > > > On Sat, 20 Jul 2019 at 04:28, Nathan Fish <lordcirth@xxxxxxxxx> wrote: >> >> On further investigation, it seems to be this bug: >> http://tracker.ceph.com/issues/38724 > > > We just upgraded to 14.2.2, and had a dozen OSDs at 14.2.2 go down this bug, recovered with: > > systemctl reset-failed ceph-osd@160 > systemctl start ceph-osd@160 > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com