Re: CephFS metadata pool size

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

thank you for your response.

I don't think this thread covers my problem, because the OSDs for the
metadata pool fill up at different rates. So I would think this is no
direct problem with the journal.
Because we had earlier problems with the journal I changed some
settings(see below). I already restarted all MDS multiple times but no
change here.

The health warnings regarding cache pressure resolve normally after a
short period of time, when the heavy load on the client ends. Sometimes it
stays a bit longer because an rsync is running and copying data on the
cluster(rsync is not good at releasing the caps).

Could it be a problem if scrubs run most of the time in the background? Can
this block any other tasks or generate new data itself?

Best regards,
Lars


global                      basic     mds_cache_memory_limit
    17179869184
global                      advanced  mds_max_caps_per_client
   16384
global                      advanced  mds_recall_global_max_decay_threshold
   262144
global                      advanced  mds_recall_max_decay_rate
   1.000000
global                      advanced  mds_recall_max_decay_threshold
    262144
mds                         advanced  mds_cache_trim_threshold
    131072
mds                         advanced  mds_heartbeat_grace
   120.000000
mds                         advanced  mds_heartbeat_reset_grace
   7400
mds                         advanced  mds_tick_interval
   3.000000


[image: ariadne.ai Logo] Lars Köppel
Developer
Email: lars.koeppel@xxxxxxxxxx
Phone: +49 6221 5993580 <+4962215993580>
ariadne.ai (Germany) GmbH
Häusserstraße 3, 69115 Heidelberg
Amtsgericht Mannheim, HRB 744040
Geschäftsführer: Dr. Fabian Svara
https://ariadne.ai


On Tue, Jun 11, 2024 at 2:05 PM Eugen Block <eblock@xxxxxx> wrote:

> Hi,
>
> can you check if this thread [1] applies to your situation? You don't
> have multi-active MDS enabled, but maybe it's still some journal
> trimming, or maybe misbehaving clients? In your first post there were
> health warnings regarding cache pressure and cache size. Are those
> resolved?
>
> [1]
>
> https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/7U27L27FHHPDYGA6VNNVWGLTXCGP7X23/#VOOV235D4TP5TEOJUWHF4AVXIOTHYQQE
>
> Zitat von Lars Köppel <lars.koeppel@xxxxxxxxxx>:
>
> > Hello everyone,
> >
> > short update to this problem.
> > The zapped OSD is rebuilt and it has now 1.9 TiB (the expected size
> ~50%).
> > The other 2 OSDs are now at 2.8 respectively 3.2 TiB. They jumped up and
> > down a lot but the higher one has now also reached 'nearfull' status. How
> > is this possible? What is going on?
> >
> > Does anyone have a solution how to fix this without zapping the OSD?
> >
> > Best regards,
> > Lars
> >
> >
> > [image: ariadne.ai Logo] Lars Köppel
> > Developer
> > Email: lars.koeppel@xxxxxxxxxx
> > Phone: +49 6221 5993580 <+4962215993580>
> > ariadne.ai (Germany) GmbH
> > Häusserstraße 3, 69115 Heidelberg
> > Amtsgericht Mannheim, HRB 744040
> > Geschäftsführer: Dr. Fabian Svara
> > https://ariadne.ai
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux