Hi Rok, On Sun, 22 Dec 2024 at 20:19, Rok Jaklič <rjaklic@xxxxxxxxx> wrote: > First I tried with osd reweight, waited a few hours then osd crush > reweight, then with pg-umpap from Laimis. Seems to crush reweight was most > effective, but not for "all" osds I tried. > > Uh, probably I've set ceph config set osd osd_max_backfills to high number > in the past, probably better to reduce it to 1 in steps, since now much > backfilling is already going on? > Every time a backfill finishes, a new one will be placed in the queue. The number of backfills won't reduce as long as you don't lower it. You can adjust it and see if it improves the backfill process or not (wait an hour or two). > > Output of commands in attachment. > There seems to be a low amount of PGs for the rgw data pool, compared to the amount of OSDs. Though it depends on the EC profile and size of a shard (`ceph pg <id> query`) if this is really an issue. But in general the amount of PGs is important, because too few of them will make them grow larger. Hence backfilling a PG will take a longer time and easier tilts the usage of OSDs, as the algorithm works by pseudo-randomly placing PGs and not taking its size into account. I'd wait with the PG adjustment after the backfilling to the HDDs has finished, should you need to adjust the number of PGs. As this will create more data movement. Cheers, Alwin croit GmbH, https://croit.io/ _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx