I'm asking that you get the new state of the file system tree after recovering from the data pool. Florian wrote that before I asked you to do this... How long did it take to run the cephfs-data-scan commands? On Tue, Jan 14, 2020 at 11:58 AM Oskar Malnowicz <oskar.malnowicz@xxxxxxxxxxxxxx> wrote: > > as florian already wrote, `du -hc` shows a total usage of 31G, but `ceph > df` show us an usage of 2.1 > > </ mds># du -hs > 31G > > # ceph df > cephfs_data 6 2.1 TiB 2.48M 2.1 TiB 25.00 3.1 TiB > > Am 14.01.20 um 20:44 schrieb Patrick Donnelly: > > On Tue, Jan 14, 2020 at 11:40 AM Oskar Malnowicz > > <oskar.malnowicz@xxxxxxxxxxxxxx> wrote: > >> i run this commands, but still the same problems > > Which problems? > > > >> $ cephfs-data-scan scan_extents cephfs_data > >> > >> $ cephfs-data-scan scan_inodes cephfs_data > >> > >> $ cephfs-data-scan scan_links > >> 2020-01-14 20:36:45.110 7ff24200ef80 -1 mds.0.snap updating last_snap 1 > >> -> 27 > >> > >> $ cephfs-data-scan cleanup cephfs_data > >> > >> do you have other ideas ? > > After you complete this, you should see the deleted files in your file > > system tree (if this is indeed the issue). What's the output of `du > > -hc`? > > > > -- Patrick Donnelly, Ph.D. He / Him / His Senior Software Engineer Red Hat Sunnyvale, CA GPG: 19F28A586F808C2402351B93C3301A3E258DD79D _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx