I recently started with Ceph, and overall had very few issues. However during a process of Cluster creation I must have done something wrong which created orphan PG groups. I suspect it was broken when I removed OSD right after initial creation, but I am guessing. Currently here is the Ceph output: #ceph -s cluster 30266c5f-5e10-4027-936c-e4409667b409 health HEALTH_WARN 65 pgs stale 22 pgs stuck inactive 65 pgs stuck stale 22 pgs stuck unclean too many PGs per OSD (456 > max 300) monmap e7: 7 mons at <servers> election epoch 156, quorum 0,1,2,3,4,5,6 <servers> mdsmap e5: 1/1/1 up {0=<server>=up:active} osdmap e579: 14 osds: 14 up, 14 in pgmap v1229872: 2152 pgs, 10 pools, 193 GB data, 54315 objects 661 GB used, 7299 GB / 8319 GB avail 2065 active+clean 65 stale+active+clean 22 creating client io 287 kB/s wr, 108 op/s One of the Stuck Inactive is 0.21 and here is the output of ceph pg map #ceph pg map 0.21 osdmap e579 pg 0.21 (0.21) -> up [] acting [] #ceph pg dump_stuck stale ok pg_stat state up up_primary acting acting_primary 0.22 stale+active+clean [5,1,6] 5 [5,1,6] 5 0.1f stale+active+clean [2,0,4] 2 [2,0,4] 2 <reducted for ease of reading> # ceph osd stat osdmap e579: 14 osds: 14 up, 14 in If I do #ceph pg 0.21 query The command freezes and never returns any output. I suspect that the problem is that these PGs were created but the OSD that they were initially created under disappeared. So I believe that I should just remove these PGs, but honestly I don’t see how. Does anybody have any ideas as to what to do next? _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com