Forgot some important info: I'm running Mimic 13.2.5. On Mon, Sep 23, 2019 at 8:49 AM Josh Haft <paccrap@xxxxxxxxx> wrote: > > Hi, > > I've been migrating data from one EC pool to another EC pool: two > directories are mounted with ceph.dir.layout.pool file attribute set > appropriately, then rsync from old to new and finally, delete the old > files. I'm using the kernel client to do this. While the removed files > are no longer present on the filesystem, they still appear to be > accounted for via "ceph df". > > When I tally up the sizes reported by "ls -lh" on all subdirectories > under the root CephFS using a FUSE client mount (except for those on > the new EC pool), it totals just under 2PiB. However, "ceph df" shows > the original EC pool as 2.5PiB used. I've copied + deleted > approximately 545TiB so far, so it seems like the unlinked files > aren't being fully released/purged. > > I've only observed the num_strays counter from "ceph daemon mds.$name > perf dump" for a few days now since I first suspected an issue, but > I've never seen it drop below roughly 310k. From other ML postings > I've gathered that stat has something to do with files pending > deletion, but I'm not positive. > > So far all I've done is restart the mds and mon daemons, which hasn't > helped. What are the next steps for troubleshooting? I can turn up mds > debug logging, but am not sure what to look for. > > Thanks for your help! > Josh _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx