On Mon, Jun 29, 2015 at 1:44 PM, Burkhard Linke <Burkhard.Linke@xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx> wrote: > Hi, > > I've noticed that a number of placement groups in our setup contain objects, > but no actual data > (ceph pg dump | grep remapped during a hard disk replace operation): > > 7.616 2636 0 0 5272 0 4194304 3003 3003 > active+remapped+wait_backfill 2015-06-29 13:43:28.716687 159913'33987 > 160091:526298 [30,6,36] 30 > [30,36,3] 30 153699'33892 2015-06-29 07:30:16.030470 > 149573'32565 2015-06-23 07:00:21.948563 > 7.60a 2696 0 0 5392 0 0 3046 3046 > active+remapped+wait_backfill 2015-06-29 13:43:09.847541 159919'34627 > 160091:388532 [2,36,3] 2 > [2,36,31] 2 153669'34496 2015-06-28 20:09:51.850005 > 153669'34496 2015-06-28 20:09:51.850005 > 7.60d 2694 0 2 5388 0 0 3026 3026 > active+remapped+wait_backfill 2015-06-29 13:43:27.202928 159939'33708 > 160091:392535 [31,6,38] 31 > [31,38,3] 31 152584'33610 2015-06-29 07:11:37.484500 > 152584'33610 2015-06-29 07:11:37.484500 > .... > > Pool 7 was used a data pool in cephfs, but almost all files stored in that > pool have been removed: > ~# rados df > pool name KB objects clones degraded unfound > rd rd KB wr wr KB > cephfs_test_data 940066 5537838 0 202 0 > 2022238 1434381904 21823705 3064326550 > > Is it possible to remove these "zombie" objects, since they influence > maintenance operations like backfilling or recovery? That's odd; the actual objects should have been deleted (not just truncated). Have you used this pool for anything else (CephFS metadata storage, RGW bucket indexes, etc)? What version of Ceph are you running and what workload did you do to induce this issue? -Greg _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com