Re: CephFS metadata pool size

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Since my last update the size of the largest OSD increased by 0.4 TiB while
the smallest one only increased by 0.1 TiB. How is this possible?

Because the metadata pool reported to have only 900MB space left, I stopped
the hot-standby MDS. This gave me 8GB back but these filled up in the last
2h.
I think I have to zap the next OSD because the filesystem is getting read
only...

How is it possible that an OSD has over 1 TiB less data on it after a
rebuild? And how is it possible to have so different sizes of OSDs?


[image: ariadne.ai Logo] Lars Köppel
Developer
Email: lars.koeppel@xxxxxxxxxx
Phone: +49 6221 5993580 <+4962215993580>
ariadne.ai (Germany) GmbH
Häusserstraße 3, 69115 Heidelberg
Amtsgericht Mannheim, HRB 744040
Geschäftsführer: Dr. Fabian Svara
https://ariadne.ai


On Tue, Jun 11, 2024 at 3:47 PM Lars Köppel <lars.koeppel@xxxxxxxxxx> wrote:

> Only in warning mode. And there were no PG splits or merges in the last 2
> month.
>
>
> [image: ariadne.ai Logo] Lars Köppel
> Developer
> Email: lars.koeppel@xxxxxxxxxx
> Phone: +49 6221 5993580 <+4962215993580>
> ariadne.ai (Germany) GmbH
> Häusserstraße 3, 69115 Heidelberg
> Amtsgericht Mannheim, HRB 744040
> Geschäftsführer: Dr. Fabian Svara
> https://ariadne.ai
>
>
> On Tue, Jun 11, 2024 at 3:32 PM Eugen Block <eblock@xxxxxx> wrote:
>
>> I don't think scrubs can cause this. Do you have autoscaler enabled?
>>
>> Zitat von Lars Köppel <lars.koeppel@xxxxxxxxxx>:
>>
>> > Hi,
>> >
>> > thank you for your response.
>> >
>> > I don't think this thread covers my problem, because the OSDs for the
>> > metadata pool fill up at different rates. So I would think this is no
>> > direct problem with the journal.
>> > Because we had earlier problems with the journal I changed some
>> > settings(see below). I already restarted all MDS multiple times but no
>> > change here.
>> >
>> > The health warnings regarding cache pressure resolve normally after a
>> > short period of time, when the heavy load on the client ends. Sometimes
>> it
>> > stays a bit longer because an rsync is running and copying data on the
>> > cluster(rsync is not good at releasing the caps).
>> >
>> > Could it be a problem if scrubs run most of the time in the background?
>> Can
>> > this block any other tasks or generate new data itself?
>> >
>> > Best regards,
>> > Lars
>> >
>> >
>> > global                      basic     mds_cache_memory_limit
>> >     17179869184
>> > global                      advanced  mds_max_caps_per_client
>> >    16384
>> > global                      advanced
>> mds_recall_global_max_decay_threshold
>> >    262144
>> > global                      advanced  mds_recall_max_decay_rate
>> >    1.000000
>> > global                      advanced  mds_recall_max_decay_threshold
>> >     262144
>> > mds                         advanced  mds_cache_trim_threshold
>> >     131072
>> > mds                         advanced  mds_heartbeat_grace
>> >    120.000000
>> > mds                         advanced  mds_heartbeat_reset_grace
>> >    7400
>> > mds                         advanced  mds_tick_interval
>> >    3.000000
>> >
>> >
>> > [image: ariadne.ai Logo] Lars Köppel
>> > Developer
>> > Email: lars.koeppel@xxxxxxxxxx
>> > Phone: +49 6221 5993580 <+4962215993580>
>> > ariadne.ai (Germany) GmbH
>> > Häusserstraße 3, 69115 Heidelberg
>> > Amtsgericht Mannheim, HRB 744040
>> > Geschäftsführer: Dr. Fabian Svara
>> > https://ariadne.ai
>> >
>> >
>> > On Tue, Jun 11, 2024 at 2:05 PM Eugen Block <eblock@xxxxxx> wrote:
>> >
>> >> Hi,
>> >>
>> >> can you check if this thread [1] applies to your situation? You don't
>> >> have multi-active MDS enabled, but maybe it's still some journal
>> >> trimming, or maybe misbehaving clients? In your first post there were
>> >> health warnings regarding cache pressure and cache size. Are those
>> >> resolved?
>> >>
>> >> [1]
>> >>
>> >>
>> https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/7U27L27FHHPDYGA6VNNVWGLTXCGP7X23/#VOOV235D4TP5TEOJUWHF4AVXIOTHYQQE
>> >>
>> >> Zitat von Lars Köppel <lars.koeppel@xxxxxxxxxx>:
>> >>
>> >> > Hello everyone,
>> >> >
>> >> > short update to this problem.
>> >> > The zapped OSD is rebuilt and it has now 1.9 TiB (the expected size
>> >> ~50%).
>> >> > The other 2 OSDs are now at 2.8 respectively 3.2 TiB. They jumped up
>> and
>> >> > down a lot but the higher one has now also reached 'nearfull'
>> status. How
>> >> > is this possible? What is going on?
>> >> >
>> >> > Does anyone have a solution how to fix this without zapping the OSD?
>> >> >
>> >> > Best regards,
>> >> > Lars
>> >> >
>> >> >
>> >> > [image: ariadne.ai Logo] Lars Köppel
>> >> > Developer
>> >> > Email: lars.koeppel@xxxxxxxxxx
>> >> > Phone: +49 6221 5993580 <+4962215993580>
>> >> > ariadne.ai (Germany) GmbH
>> >> > Häusserstraße 3, 69115 Heidelberg
>> >> > Amtsgericht Mannheim, HRB 744040
>> >> > Geschäftsführer: Dr. Fabian Svara
>> >> > https://ariadne.ai
>> >> > _______________________________________________
>> >> > ceph-users mailing list -- ceph-users@xxxxxxx
>> >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >>
>> >>
>> >> _______________________________________________
>> >> ceph-users mailing list -- ceph-users@xxxxxxx
>> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >>
>>
>>
>>
>>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux