Hi Stefan, Thank you, that definitely helped. I bumped it to 20% for now and that's giving me around 124 PGs backfilling at 187 MiB/s, 47 Objects/s. I'll see how that runs and then increase it a bit more if the cluster handles it ok. Do you think it's worth enabling scrubbing while backfilling? Since this is going to take a while. I do have 1 inconsistent PG that has now become 10 as it splits. ceph health detail HEALTH_ERR 21 scrub errors; Possible data damage: 10 pgs inconsistent; 2 pgs not deep-scrubbed in time [ERR] OSD_SCRUB_ERRORS: 21 scrub errors [ERR] PG_DAMAGED: Possible data damage: 10 pgs inconsistent pg 12.12 is active+clean+inconsistent, acting [28,1,37,0] pg 12.32 is active+clean+inconsistent, acting [37,3,14,22] pg 12.52 is active+clean+inconsistent, acting [4,33,7,23] pg 12.72 is active+remapped+inconsistent+backfilling, acting [37,3,14,22] pg 12.92 is active+remapped+inconsistent+backfilling, acting [28,1,37,0] pg 12.b2 is active+remapped+inconsistent+backfilling, acting [37,3,14,22] pg 12.d2 is active+clean+inconsistent, acting [4,33,7,23] pg 12.f2 is active+remapped+inconsistent+backfilling, acting [37,3,14,22] pg 12.112 is active+clean+inconsistent, acting [28,1,37,0] pg 12.132 is active+clean+inconsistent, acting [37,3,14,22] [WRN] PG_NOT_DEEP_SCRUBBED: 2 pgs not deep-scrubbed in time pg 4.13 not deep-scrubbed since 2022-06-16T03:15:16.758943+0000 pg 7.1 not deep-scrubbed since 2022-06-16T20:51:12.211259+0000 Thanks, Curt On Wed, Jun 29, 2022 at 5:53 PM Stefan Kooman <stefan@xxxxxx> wrote: > On 6/29/22 15:14, Curt wrote: > > > > > > Hi Stefan, > > > > Good to know. I see the default if .05 for misplaced_ratio. What do > > you recommend would be a safe number to increase it to? > > It depends. It might be safe to put it to 1. But I would slowly increase > it, have the manager increase pgp_num and see how the cluster copes with > the increased load. If you have hardly any client workload you might > bump this ratio quite a bit. At some point you would need to increase > osd max backfill to avoid having PGs waiting on backfill. > > Gr. Stefan > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx