On Wed, Nov 11, 2015 at 5:39 PM, Eric Eastman <eric.eastman@xxxxxxxxxxxxxx> wrote: > I am trying to figure out why my Ceph file system is not freeing > space. Using Ceph 9.1.0 I created a file system with snapshots > enabled, filled up the file system over days while taking snapshots > hourly. I then deleted all files and all snapshots, but Ceph is not > returning the space. I left the cluster sit for two days to see if the > cleanup process was being done in the background and it still has not > freed the space. I tried rebooting the cluster and clients and the > space is still not returned. Preface: snapshots are disabled by default for a reason -- we don't have the test coverage for this stuff yet. Things to try: * Looking at MDS statistics (ceph daemon mds.foo perf dump) with "stray" in the name to see if your inodes are stuck in stray state * Dumping MDS cache to see what it thinks about it, if you can see references to the files that should have been deleted John > > The file system was created with the command: > # ceph fs new cephfs cephfs_metadata cephfs_data > > Info on the Ceph file system: > > # getfattr -d -m ceph.dir.* /cephfs/ > getfattr: Removing leading '/' from absolute path names > # file: cephfs/ > ceph.dir.entries="0" > ceph.dir.files="0" > ceph.dir.rbytes="0" > ceph.dir.rctime="1447033469.0920991041" > ceph.dir.rentries="4" > ceph.dir.rfiles="1" > ceph.dir.rsubdirs="3" > ceph.dir.subdirs="0" > > ls -l /cephfs/ > total 0 > > # ls -l /cephfs/.snap > total 0 > > # grep ceph /proc/mounts > ceph-fuse /cephfs fuse.ceph-fuse > rw,noatime,user_id=0,group_id=0,default_permissions,allow_other 0 0 > > # df /cephfs/ > Filesystem 1K-blocks Used Available Use% Mounted on > ceph-fuse 276090880 194162688 81928192 71% /cephfs > > # df -i /cephfs/ > Filesystem Inodes IUsed IFree IUse% Mounted on > ceph-fuse 2501946 - - - /cephfs > > # ceph df detail > GLOBAL: > SIZE AVAIL RAW USED %RAW USED OBJECTS > 263G 80009M 181G 68.78 2443k > POOLS: > NAME ID CATEGORY USED %USED MAX > AVAIL OBJECTS DIRTY READ WRITE > rbd 0 - 0 0 > 27826M 0 0 0 0 > cephfs_data 1 - 76846M 28.50 > 27826M 2501672 2443k 345k 32797k > cephfs_metadata 2 - 34868k 0.01 > 27826M 259 259 480k 23327k > kSAFEbackup 3 - 108M 0.04 > 27826M 15 15 0 49 > > The Ceph cluster and client systems are running on Trusty with a 4.3.0 > kernel and Ceph version 9.1.0 > # ceph -v > ceph version 9.1.0 (3be81ae6cf17fcf689cd6f187c4615249fea4f61) > # uname -a > Linux ede-c2-adm01 4.3.0-040300-generic #201511020949 SMP Mon Nov 2 > 14:50:44 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux > > Any ideas on why the space is not being freed? > > Thanks, > Eric > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com