Hey All, It might be easier to check using cephfs dir stats using getfattr, ex. getfattr -n ceph.dir.rentries /path/to/dir Regards, Bailey > -----Original Message----- > From: Igor Fedotov <igor.fedotov@xxxxxxxx> > Sent: March 14, 2024 1:37 PM > To: Thorne Lawler <thorne@xxxxxxxxxxx>; ceph-users@xxxxxxx; > etienne.menguy@xxxxxxxxxxx; vbogdan@xxxxxxxxx > Subject: Re: CephFS space usage > > Thorn, > > you might want to assess amount of files on the mounted fs by runnning "du > -h | wc". Does it differ drastically from amount of objects in the pool = ~3.8 > M? > > And just in case - please run "rados lssnap -p cephfs.shared.data". > > > Thanks, > > Igor > > On 3/14/2024 1:42 AM, Thorne Lawler wrote: > > > > Igor, Etienne, Bogdan, > > > > The system is a four node cluster. Each node has 12 3.8TB SSDs, and > > each SSD is an OSD. > > > > I have not defined any separate DB / WAL devices - this cluster is > > mostly at cephadm defaults. > > > > Everything is currently configured to have x3 replicas. > > > > The system also does various RBD workloads from other pools. > > > > There are no subvolumes and no snapshots on the CephFS volume in > question. > > > > The CephFS volume I am concerned about is called 'shared'. For the > > purposes of this question I am omitting information about the other pools. > > > > [root@san1 ~]# rados df > > POOL_NAME USED OBJECTS CLONES COPIES > > MISSING_ON_PRIMARY UNFOUND DEGRADED RD_OPS RD > WR_OPS WR > > USED COMPR UNDER COMPR cephfs.shared.data 41 TiB 3834689 > > 0 > > 11504067 0 0 0 3219785418 175 TiB > > 9330001764 229 TiB 7.0 MiB 12 MiB cephfs.shared.meta > > 757 MiB 85 0 > > 255 0 0 0 5306018840 26 TiB > > 9170232158 24 TiB 0 B 0 B > > > > total_objects 13169948 > > total_used 132 TiB > > total_avail 33 TiB > > total_space 166 TiB > > > > [root@san1 ~]# ceph df detail > > --- RAW STORAGE --- > > CLASS SIZE AVAIL USED RAW USED %RAW USED ssd 166 TiB > > 33 TiB 132 TiB 132 TiB 79.82 TOTAL 166 TiB 33 TiB 132 TiB > > 132 TiB 79.82 > > > > --- POOLS --- > > POOL ID PGS STORED (DATA) (OMAP) OBJECTS > > USED (DATA) (OMAP) %USED MAX AVAIL QUOTA OBJECTS QUOTA > BYTES > > DIRTY USED COMPR UNDER COMPR cephfs.shared.meta 3 32 251 > > MiB 208 MiB 42 MiB > > 84 752 MiB 625 MiB 127 MiB 0 3.4 TiB N/A N/A > > N/A 0 B 0 B cephfs.shared.data 4 512 14 > > TiB 14 TiB 0 B 3.83M 41 TiB 41 TiB 0 B 79.90 3.4 > > TiB N/A N/A N/A 7.0 MiB 12 MiB > > > > [root@san1 ~]# ceph osd pool get cephfs.shared.data size > > size: 3 > > > > ...however running 'du' in the root directory of the 'shared' volume says: > > > > # du -sh . > > 5.5T . > > > > So yeah - 14TB is replicated to 41TB, that's fine, but 14TB is a lot > > more than 5.5TB, so... where is that space going? > > > > On 14/03/2024 2:09 am, Igor Fedotov wrote: > >> Hi Thorn, > >> > >> could you please share the output of "ceph df detail" command > >> representing the problem? > >> > >> > >> And please give an overview of your OSD layout - amount of OSDs, > >> shared or dedicated DB/WAL, main and DB volume sizes. > >> > >> > >> Thanks, > >> > >> Igor > >> > >> > >> On 3/13/2024 5:58 AM, Thorne Lawler wrote: > >>> Hi everyone! > >>> > >>> My Ceph cluster (17.2.6) has a CephFS volume which is showing 41TB > >>> usage for the data pool, but there are only 5.5TB of files in it. > >>> There are fewer than 100 files on the filesystem in total, so where > >>> is all that space going? > >>> > >>> How can I analyze my cephfs to understand what is using that space, > >>> and if possible, how can I reclaim that space? > >>> > >>> Thank you. > >>> > > -- > > > > Regards, > > > > Thorne Lawler - Senior System Administrator > > *DDNS* | ABN 76 088 607 265 > > First registrar certified ISO 27001-2013 Data Security Standard > > ITGOV40172 P +61 499 449 170 > > > > _DDNS > > > > /_*Please note:* The information contained in this email message and > > any attached files may be confidential information, and may also be > > the subject of legal professional privilege. _If you are not the > > intended recipient any use, disclosure or copying of this email is > > unauthorised. _If you received this email in error, please notify > > Discount Domain Name Services Pty Ltd on 03 9815 6868 to report this > > matter and delete all copies of this transmission together with any > > attachments. / > > > -- > Igor Fedotov > Ceph Lead Developer > > Looking for help with your Ceph cluster? Contact us athttps://croit.io > > croit GmbH, Freseniusstr. 31h, 81247 Munich > CEO: Martin Verges - VAT-ID: DE310638492 Com. register: Amtsgericht > Munich HRB 231263 Web:https://croit.io | YouTube:https://goo.gl/PGE1Bx > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email > to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx