Re: High read throughput on BlueFS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I found that bluefs_max_prefetch is set to 1048576 which equals to 1MiB! So
why it's reading about 1GiB/s?

On Thu, Dec 3, 2020 at 8:03 PM Seena Fallah <seenafallah@xxxxxxxxx> wrote:

> My first question is about this metric: ceph_bluefs_read_prefetch_bytes
> and I want to know what operation is related to this metric?
>
> On Thu, Dec 3, 2020 at 7:49 PM Seena Fallah <seenafallah@xxxxxxxxx> wrote:
>
>> Hi all,
>>
>> When my cluster gets into a recovery state (adding new node) I see a huge
>> read throughput on its disks and it affects the latency! Disks are SSD and
>> they don't have a separate WAL/DB.
>> I'm using nautilus 14.2.14 and bluefs_buffered_io is false by default.
>> When this throughput came on my disk it will get too much high latency.
>> After I turned on bluefs_buffered_io another huge throughput around 1.2GB/s
>> came in and it against affect my latency but much less than the previous
>> one! (Graphs are attached and bluefs_buffered_io was turned on with ceph
>> tell injectargs at 13:41 also I have restarted the OSD at 13:16 because it
>> doesn't get better at the moment)
>>
>> I have four questions:
>> 1. What are they? I see the recovery speed is 20MB/s and client io on
>> that OSD is 10MB/s so what is this high throughput for?!
>> 2. How can I control this throughput? Because my disks don't support this
>> much throughput!
>> 3. I see a common issue here https://tracker.ceph.com/issues/36482 that
>> I think it's similar to my case. It was discussed about read_ahead, well
>> should I change the read_ahead_kb config of my disk to support this type of
>> request? I'm using the default value in ubuntu (128)
>> 4. Is there any tuning required that can help to turn off the
>> bluefs_buffered_io again?
>>
>> Configs I used for recovery:
>> osd max backfills = 1
>> osd recovery max active = 1
>> osd recovery op priority = 1
>> osd recovery priority = 1
>> osd recovery sleep ssd = 0.2
>>
>> My OSD memory target is around 6GB.
>>
>> Thanks.
>>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux