Hello community.
We run Ceph on quite old hardware with quite low traffic. Yesterday we
had to reboot one of the OSDs and after reboot it did not came up. The
error message is:
[2019-11-02 15:05:07,317][ceph_volume.process][INFO ] Running command:
/usr/sbin/ceph-volume lvm trigger 22-1c0b3fd7-7d80-4de9-9594-17ac5b2bf92f
[2019-11-02 15:05:07,473][ceph_volume.process][INFO ] stderr -->
RuntimeError: could not find osd.22 with fsid
1c0b3fd7-7d80-4de9-9594-17ac5b2bf92f
This OSD has 2 disks, which are put into one logical volume (basically
raid0) and then used for osd storage.
We are quite a beginners with Ceph and this error stuck us. What should
we do? Change fsid (where?)? Right now cluster is in repair-state.. As
the last resort we would drop osd and rebuild it, but it would be very
important for us to understand - what and why happenned. Is it faulty
config or did something bad happen with the disks?
Regards,
Anton.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com