Hello, I have a Nautilus cluster with a cephfs volume, on grafana, it shows that cephfs_data pool is almost full[1] but if I give a look to the pool usage, it looks like I have plenty of space. Which metrics are used by grafana? 1. https://framapic.org/5r7J86s55x6k/jGSIsjEUPYMU.png pool usage: > artemis@icitsrv5:~$ ceph df detail > RAW STORAGE: > CLASS SIZE AVAIL USED RAW USED %RAW USED > hdd 662 TiB 296 TiB 366 TiB 366 TiB 55.32 > TOTAL 662 TiB 296 TiB 366 TiB 366 TiB 55.32 > > POOLS: > POOL ID STORED OBJECTS USED %USED MAX AVAIL QUOTA OBJECTS QUOTA BYTES DIRTY USED COMPR UNDER COMPR > .rgw.root 3 8.1 KiB 15 2.8 MiB 0 63 TiB N/A N/A 15 0 B 0 B > default.rgw.control 4 0 B 8 0 B 0 63 TiB N/A N/A 8 0 B 0 B > default.rgw.meta 5 26 KiB 85 16 MiB 0 63 TiB N/A N/A 85 0 B 0 B > default.rgw.log 6 0 B 207 0 B 0 63 TiB N/A N/A 207 0 B 0 B > cephfs_data 7 113 TiB 139.34M 186 TiB 49.47 138 TiB N/A N/A 139.34M 0 B 0 B > cephfs_metadata 8 54 GiB 10.21M 57 GiB 0.03 63 TiB N/A N/A 10.21M 0 B 0 B > default.rgw.buckets.data 9 122 TiB 54.57M 173 TiB 47.70 138 TiB N/A N/A 54.57M 0 B 0 B > default.rgw.buckets.index 10 2.6 GiB 19.97k 2.6 GiB 0 63 TiB N/A N/A 19.97k 0 B 0 B > default.rgw.buckets.non-ec 11 67 MiB 186 102 MiB 0 63 TiB N/A N/A 186 0 B 0 B > device_health_metrics 12 1.2 MiB 145 1.2 MiB 0 63 TiB N/A N/A 145 0 B 0 B Best, -- Yoann Moulin EPFL IC-IT _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx