Re: Ceph recovery network speed

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Stefan,

Thank you, that definitely helped. I bumped it to 20% for now and that's
giving me around 124 PGs backfilling at 187 MiB/s, 47 Objects/s.  I'll see
how that runs and then increase it a bit more if the cluster handles it ok.

Do you think it's worth enabling scrubbing while backfilling?  Since this
is going to take a while. I do have 1 inconsistent PG that has now become
10 as it splits.

ceph health detail
HEALTH_ERR 21 scrub errors; Possible data damage: 10 pgs inconsistent; 2
pgs not deep-scrubbed in time
[ERR] OSD_SCRUB_ERRORS: 21 scrub errors
[ERR] PG_DAMAGED: Possible data damage: 10 pgs inconsistent
    pg 12.12 is active+clean+inconsistent, acting [28,1,37,0]
    pg 12.32 is active+clean+inconsistent, acting [37,3,14,22]
    pg 12.52 is active+clean+inconsistent, acting [4,33,7,23]
    pg 12.72 is active+remapped+inconsistent+backfilling, acting
[37,3,14,22]
    pg 12.92 is active+remapped+inconsistent+backfilling, acting [28,1,37,0]
    pg 12.b2 is active+remapped+inconsistent+backfilling, acting
[37,3,14,22]
    pg 12.d2 is active+clean+inconsistent, acting [4,33,7,23]
    pg 12.f2 is active+remapped+inconsistent+backfilling, acting
[37,3,14,22]
    pg 12.112 is active+clean+inconsistent, acting [28,1,37,0]
    pg 12.132 is active+clean+inconsistent, acting [37,3,14,22]
[WRN] PG_NOT_DEEP_SCRUBBED: 2 pgs not deep-scrubbed in time
    pg 4.13 not deep-scrubbed since 2022-06-16T03:15:16.758943+0000
    pg 7.1 not deep-scrubbed since 2022-06-16T20:51:12.211259+0000

Thanks,
Curt

On Wed, Jun 29, 2022 at 5:53 PM Stefan Kooman <stefan@xxxxxx> wrote:

> On 6/29/22 15:14, Curt wrote:
>
>
> >
> > Hi Stefan,
> >
> > Good to know.  I see the default if .05 for misplaced_ratio.  What do
> > you recommend would be a safe number to increase it to?
>
> It depends. It might be safe to put it to 1. But I would slowly increase
> it, have the manager increase pgp_num and see how the cluster copes with
> the increased load. If you have hardly any client workload you might
> bump this ratio quite a bit. At some point you would need to increase
> osd max backfill to avoid having PGs waiting on backfill.
>
> Gr. Stefan
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux