I found that bluefs_max_prefetch is set to 1048576 which equals to 1MiB! So why it's reading about 1GiB/s? On Thu, Dec 3, 2020 at 8:03 PM Seena Fallah <seenafallah@xxxxxxxxx> wrote: > My first question is about this metric: ceph_bluefs_read_prefetch_bytes > and I want to know what operation is related to this metric? > > On Thu, Dec 3, 2020 at 7:49 PM Seena Fallah <seenafallah@xxxxxxxxx> wrote: > >> Hi all, >> >> When my cluster gets into a recovery state (adding new node) I see a huge >> read throughput on its disks and it affects the latency! Disks are SSD and >> they don't have a separate WAL/DB. >> I'm using nautilus 14.2.14 and bluefs_buffered_io is false by default. >> When this throughput came on my disk it will get too much high latency. >> After I turned on bluefs_buffered_io another huge throughput around 1.2GB/s >> came in and it against affect my latency but much less than the previous >> one! (Graphs are attached and bluefs_buffered_io was turned on with ceph >> tell injectargs at 13:41 also I have restarted the OSD at 13:16 because it >> doesn't get better at the moment) >> >> I have four questions: >> 1. What are they? I see the recovery speed is 20MB/s and client io on >> that OSD is 10MB/s so what is this high throughput for?! >> 2. How can I control this throughput? Because my disks don't support this >> much throughput! >> 3. I see a common issue here https://tracker.ceph.com/issues/36482 that >> I think it's similar to my case. It was discussed about read_ahead, well >> should I change the read_ahead_kb config of my disk to support this type of >> request? I'm using the default value in ubuntu (128) >> 4. Is there any tuning required that can help to turn off the >> bluefs_buffered_io again? >> >> Configs I used for recovery: >> osd max backfills = 1 >> osd recovery max active = 1 >> osd recovery op priority = 1 >> osd recovery priority = 1 >> osd recovery sleep ssd = 0.2 >> >> My OSD memory target is around 6GB. >> >> Thanks. >> > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx