have you ever used fancy layout? see http://tracker.ceph.com/issues/15050 On Wed, Apr 20, 2016 at 3:17 AM, Simion Rad <Simion.Rad@xxxxxxxxx> wrote: > Mounting and unmount doesn't change anyting. > The used space reported by df command is nearly the same as the values returned by ceph -s command. > > Example 1, df output: > ceph-fuse 334T 134T 200T 41% /cephfs > > Example 2, ceph -s output: > health HEALTH_WARN > mds0: Many clients (22) failing to respond to cache pressure > noscrub,nodeep-scrub,sortbitwise flag(s) set > monmap e1: 5 mons at {r730-12=10.103.213.12:6789/0,r730-4=10.103.213.4:6789/0,r730-5= > 10.103.213.5:6789/0,r730-8=10.103.213.8:6789/0,r730-9=10.103.213.9:6789/0} > election epoch 132, quorum 0,1,2,3,4 r730-4,r730-5,r730-8,r730-9,r730-12 > mdsmap e14637: 1/1/1 up {0=ceph2-mds-2=up:active} > osdmap e6549: 68 osds: 68 up, 68 in > flags noscrub,nodeep-scrub,sortbitwise > pgmap v4394151: 896 pgs, 3 pools, 54569 GB data, 56582 kobjects > 133 TB used, 199 TB / 333 TB avail > 896 active+clean > client io 47395 B/s rd, 1979 kB/s wr, 388 op/s > > > ________________________________________ > From: John Spray <jspray@xxxxxxxxxx> > Sent: Tuesday, April 19, 2016 22:04 > To: Simion Rad > Cc: ceph-users@xxxxxxxxxxxxxx > Subject: Re: cephfs does not seem to properly free up space > > On Tue, Apr 19, 2016 at 2:40 PM, Simion Rad <Simion.Rad@xxxxxxxxx> wrote: >> Hello, >> >> >> At my workplace we have a production cephfs cluster (334 TB on 60 OSDs) >> which was recently upgraded from Infernalis 9.2.0 to Infernalis 9.2.1 on >> Ubuntu 14.04.3 (linux 3.19.0-33). >> >> It seems that cephfs still doesn't free up space at all or at least that's >> what df command tells us. > > Hmm, historically there were bugs with the purging code, but I thought > we fixed them before Infernalis. > > Does the space get freed after you unmount the client? Some issues > have involved clients holding onto references to unlinked inodes. > > John > >> >> Is there a better way of getting a df-like output with other command for >> cephfs ? >> >> >> Thank you, >> >> Marius Rad >> >> SysAdmin >> >> www.propertyshark.com >> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com