Did you try options from cache-sizing <https://docs.ceph.com/en/quincy/dev/cache-pool/#cache-sizing> or other-tunables <https://docs.ceph.com/en/quincy/dev/cache-pool/#other-tunables>? - Dhairya On Fri, Mar 3, 2023 at 5:39 AM <xadhoom76@xxxxxxxxx> wrote: > Hi, we have a cluster with this ceph df > > --- RAW STORAGE --- > CLASS SIZE AVAIL USED RAW USED %RAW USED > hdd 240 GiB 205 GiB 29 GiB 35 GiB 14.43 > hddvm 1.6 TiB 1.2 TiB 277 GiB 332 GiB 20.73 > TOTAL 1.8 TiB 1.4 TiB 305 GiB 366 GiB 19.91 > > --- POOLS --- > POOL ID PGS STORED (DATA) (OMAP) OBJECTS USED > (DATA) (OMAP) %USED MAX AVAIL QUOTA OBJECTS QUOTA BYTES DIRTY > USED COMPR UNDER COMPR > device_health_metrics 1 1 0 B 0 B 0 B 0 0 > B 0 B 0 B 0 308 GiB N/A N/A 0 > 0 B 0 B > rbd-pool 2 32 539 B 19 B 520 B 9 539 > B 19 B 520 B 0 462 GiB N/A N/A 9 > 0 B 0 B > cephfs.sharedfs.meta 3 32 299 MiB 190 MiB 109 MiB 87.10k 299 > MiB 190 MiB 109 MiB 0.03 308 GiB N/A N/A > 87.10k 0 B 0 B > cephfs.sharedfs.data 4 32 2.2 GiB 2.2 GiB 0 B 121.56k 2.2 > GiB 2.2 GiB 0 B 0.23 308 GiB N/A N/A > 121.56k 0 B 0 B > rbd-pool-proddeb02 5 32 712 MiB 712 MiB 568 B 201 712 > MiB 712 MiB 568 B 0.08 308 GiB N/A N/A > 201 0 B 0 B > > > So as you can see we have 332GB RAW but data really are 299+2.2G+712M > > POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL > device_health_metrics 1 1 0 B 0 0 B 0 308 GiB > rbd-pool 2 32 539 B 9 539 B 0 462 GiB > cephfs.sharedfs.meta 3 32 299 MiB 87.10k 299 MiB 0.03 308 GiB > cephfs.sharedfs.data 4 32 2.2 GiB 121.56k 2.2 GiB 0.23 308 GiB > rbd-pool-proddeb02 5 32 712 MiB 201 712 MiB 0.08 308 GiB > > How to clean Dirty ? How is that possible ? any cache issue or not > committed flush from client ? > Best regards > Alessandro > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx