I am really hoping you can help. THANKS in advance. I have inherited a Docker swarm running CEPH but I know very little about it. Current I have an unhealthy ceph environment that will not mount my data drive. Its a cluster of 4 vm servers. docker01,docker02, docker03, docker-cloud CL has the /data that is on a separate drive, currently failing to mount. how can I recover this without loosing the data? on server docker-cloud, mount /data returns: mount error 113 = No route to host docker ps is healthy on all nodes. bc81d14dde92 ceph/daemon:latest-mimic "/opt/ceph-container…" 2 years ago Up 34 minutes ceph-mds d4fecec5e0e8 ceph/daemon:latest-mimic "/opt/ceph-container…" 2 years ago Up 34 minutes ceph-osd 482ba41803af ceph/daemon:latest-mimic "/opt/ceph-container…" 2 years ago Up 34 minutes ceph-mgr d6a5c44179c7 ceph/daemon:latest-mimic "/opt/ceph-container…" 2 years ago Up 32 minutes ceph-mon ceph -s: cluster: id: 7a5b2243-8e92-4e03-aee7-aa64cea666ec health: HEALTH_ERR 1 filesystem is degraded 1 filesystem is offline 1 mds daemon damaged noout,noscrub,nodeep-scrub flag(s) set clock skew detected on mon.docker02, mon.docker03, mon.docker-cloud mons docker-cloud,docker01,docker02,docker03 are low on available space services: mon: 4 daemons, quorum docker01,docker02,docker03,docker-cloud mgr: docker01(active), standbys: docker02, docker03, docker-cloud mds: cephfs-0/1/1 up , 4 up:standby, 1 damaged osd: 4 osds: 4 up, 4 in flags noout,noscrub,nodeep-scrub data: pools: 2 pools, 256 pgs objects: 194.2 k objects, 241 GiB usage: 499 GiB used, 1.5 TiB / 2.0 TiB avail pgs: 256 active+clean _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx