We have a cluster (10.2.9 based) with a cephfs filesytem that has 4800+ snapshots. We want to delete most of the very old ones to get it to a more manageable number (such as 0). However, deleting even 1 snapshot right now takes up to a full 24 hours due to their age and size. It would literally take 13 years to delete all of them at the current pace. Here is one snapshot directory statistics: # file: cephfs/.snap/snapshot.2017-02-24_22_17_01-1487992621 ceph.dir.entries="3" ceph.dir.files="0" ceph.dir.rbytes="30500769204664" ceph.dir.rctime="1504695439.09966088000" ceph.dir.rentries="7802785" ceph.dir.rfiles="7758691" ceph.dir.rsubdirs="44094" ceph.dir.subdirs="3" There is a bug filed with details here: http://tracker.ceph.com/issues/21412 Im wondering if there is a faster, undocumented, "backdoor" way to clean up our snapshot mess without destroying the entire filesystem and recreating it. -Wyllys Ingersoll Keeper Technology, LLC -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html