Hi, ceph -s reports most of my PGs as "active+clean", but a small number will stay at just "active": # ceph -s 2012-01-06 11:16:44.832625 pg v278953: 396 pgs: 8 active, 388 active+clean; 70764 MB data, 157 GB used, 4994 GB / 5257 GB avail; 170/38388 degraded (0.443%) 2012-01-06 11:16:44.842568 mds e9: 1/1/1 up {0=alpha=up:active} 2012-01-06 11:16:44.842648 osd e243: 6 osds: 6 up, 6 in 2012-01-06 11:16:44.842799 log 2012-01-06 06:12:00.829148 osd.4 10.3.1.35:6800/1490 343 : [INF] 0.4e scrub ok 2012-01-06 11:16:44.844179 mon e5: 3 mons at {ceph1=10.3.1.33:6789/0,ceph2=10.3.1.34:6789/0,ceph3=10.3.1.35:6789/0} It's been like that for several days now. IIRC, the last thing I did with the cluster was to add some more OSDs. The last time this happened, the problem went away after restarting some of the OSDs. What does this mean? Is that a bug? Regards, Guido -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html