Hi Cephers, We've been testing drive failures and we're just trying to see if the behaviour of our cluster is normal, or if we've setup something wrong. In summary; the OSD is down and out, but the PGs are showing as degraded and don't seem to want to remap. We'd have assumed once the OSD was marked out, that a re-map should have happened and we'd see misplaced rather than degraded PGs. cluster bfb7e824-f37d-45c0-a4fc-a98182fed985 health HEALTH_WARN 43 pgs degraded 43 pgs stuck degraded 44 pgs stuck unclean 43 pgs stuck undersized 43 pgs undersized recovery 36899/6822836 objects degraded (0.541%) recovery 813/6822836 objects misplaced (0.012%) monmap e3: 3 mons at {ceph-admin1=10.66.8.1:6789/0,ceph-store1=10.66.8.2:6789/0,ceph-store2=10. 66.8.3:6789/0} election epoch 950, quorum 0,1,2 ceph-admin1,ceph-store1,ceph-store2 osdmap e6342: 36 osds: 35 up, 35 in; 1 remapped pgs pgmap v11805515: 1700 pgs, 3 pools, 13165 GB data, 3331 kobjects 25941 GB used, 30044 GB / 55986 GB avail 36899/6822836 objects degraded (0.541%) 813/6822836 objects misplaced (0.012%) 1656 active+clean 43 active+undersized+degraded 1 active+remapped client io 491 kB/s rd, 3998 kB/s wr, 480 op/s # id weight type name up/down reweight -6 43.56 root hdd -2 21.78 host ceph-store1-hdd 0 3.63 osd.0 up 1 2 3.63 osd.2 up 1 4 3.63 osd.4 up 1 6 3.63 osd.6 up 1 8 3.63 osd.8 up 1 10 3.63 osd.10 up 1 -3 21.78 host ceph-store2-hdd 1 3.63 osd.1 up 1 3 3.63 osd.3 up 1 5 3.63 osd.5 up 1 7 3.63 osd.7 up 1 9 3.63 osd.9 up 1 11 3.63 osd.11 up 1 -1 11.48 root ssd -4 5.74 host ceph-store1-ssd 12 0.43 osd.12 up 1 13 0.43 osd.13 up 1 14 0.43 osd.14 up 1 16 0.43 osd.16 up 1 18 0.43 osd.18 down 0 19 0.43 osd.19 up 1 20 0.43 osd.20 up 1 21 0.43 osd.21 up 1 32 0.72 osd.32 up 1 33 0.72 osd.33 up 1 17 0.43 osd.17 up 1 15 0.43 osd.15 up 1 -5 5.74 host ceph-store2-ssd 22 0.43 osd.22 up 1 23 0.43 osd.23 up 1 24 0.43 osd.24 up 1 25 0.43 osd.25 up 1 26 0.43 osd.26 up 1 27 0.43 osd.27 up 1 28 0.43 osd.28 up 1 29 0.43 osd.29 up 1 30 0.43 osd.30 up 1 31 0.43 osd.31 up 1 34 0.72 osd.34 up 1 35 0.72 osd.35 up 1 Are we misunderstanding the default behaviour? Any help you can provide will be very much appreciated. Regards, Daniel W: www.3ca.com.au _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com