Re: Increase of osd space usage on cephfs heavy load

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

On 4/6/21 2:20 PM, Olivier AUDRY wrote:
hello

now backup is running since 3hours and cephfs metadata goes from 20G to
479Go...
POOL
   ID  STORED   OBJECTS  USED     %USED  MAX AVAIL
cephfs-metadata            12  479 GiB   642.26k    1.4 TiB
18.79       2.0 TiB
cephfs-data0                    13  2.9 TiB     9.23M        9.4 TiB
10.67       26 TiB

is that a normal behaviour ?


The MDS maintains a list of open files in the metadata pool. If your backup is scanning a lot of files, and caps are not reclaimed by the MDS, this list will become large.


The corresponding objects are called 'mds<rank>_openfiles.<chunk>', e.g. mds0_openfiles.0, mds0_openfiles.1 etc. You can check the size of these objects with the rados command.


If this is the reason for the large pool, I would recommend to restrict the number of caps per client, otherwise you might run into out of memory problems if the MDS is restarted during the backup.


Regards,

Burkhard

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux