Hello Cephers I am struggling with my ceph cluster health , PGS are not getting clean , i waited for recovery process to get end was hoping after recovery PG will become clean , but it didn’t. Can you please share your suggestions. cluster 0ff473d9-0670-42a3-89ff-81bbfb2e676a health HEALTH_WARN 119 pgs down; 303 pgs peering; 303 pgs stuck inactive; 303 pgs stuck unclean; mds cluster is degraded; crush map has no n-optimal tunables monmap e3: 3 mons at {ceph-mon1=192.168.1.38:6789/0,ceph-mon2=192.168.1.33:6789/0,ceph-mon3=192.168.1.31:6789/0}, election epoch 4226, quo rum 0,1,2 ceph-mon1,ceph-mon2,ceph-mon3 mdsmap e8465: 1/1/1 up {0=ceph-mon1=up:replay} osdmap e250466: 10 osds: 10 up, 10 in pgmap v585809: 576 pgs, 6 pools, 101933 MB data, 25453 objects 343 GB used, 5423 GB / 5767 GB avail 273 active+clean 108 peering 119 down+peering 76 remapped+peering # id weight type name up/down reweight -1 5.65 root default -2 0 host ceph-node1 -3 1.72 host ceph-node2 4 0.43 osd.4 up 1 5 0.43 osd.5 up 1 6 0.43 osd.6 up 1 7 0.43 osd.7 up 1 -4 1.31 host ceph-node4 8 0.88 osd.8 up 1 1 0.43 osd.1 up 1 -5 1.31 host ceph-node5 9 0.88 osd.9 up 1 2 0.43 osd.2 up 1 -6 0.88 host ceph-node6 10 0.88 osd.10 up 1 -7 0.43 host ceph-node3 0 0.43 osd.0 up 1 Regards karan _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com