Re: About scrub and deep-scrub

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Eugen

Adding PG to the cluster only helps to reduce PG size and reduce scrub time
is right? >150 PG and <200 in osd i think it's good number

My cluster I/O 2-3GB/s real time in 24h IOPS 2K. If choosing to increase PG
is it a good choice?

Vào Th 2, 7 thg 10, 2024 vào lúc 15:49 Eugen Block <eblock@xxxxxx> đã
viết:

> So your PGs for pool 52 have a size of around 320 GB, that is quite a
> lot and not surprising that deep-scrubs take a long time. At the same
> time, your PGs per OSD are already > 150. We had a similar situation
> on a customer cluster this year as well, also with 12 TB drives. We
> decided to increase the pg_num anyway to reduce the pg sizes. They
> currently have around 380 PGs per large OSD (they have lots of smaller
> OSDs as well) which still works fine. But they're using it as an
> archive, so the IO is not very high. If you would decide to split PGs,
> keep in mind to increase mon_max_pg_per_osd and
> osd_max_pg_per_osd_hard_ratio as well. I can't explicitly recommend to
> double your PGs per OSD as I'm not familiar with your cluster, the
> load etc. It's just something to think about.
> Doubling the PG count would reduce the PG size to around 160 GB, which
> is still a lot, but I probably wouldn't go further than that.
> The OSD utilization is only around 40%, in this case a cluster with
> more (smaller) OSDs would probably have made more sense.
>
> Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:
>
> > Hi Eugen
> >
> > Can you see and give me some advice, number of PG and PG size..
> >
> > ID   CLASS  WEIGHT    REWEIGHT  SIZE     RAW USE  DATA     OMAP     META
> >   AVAIL    %USE   VAR   PGS  STATUS
> >   2    hdd  10.98349   1.00000   11 TiB  4.5 TiB  4.2 TiB  5.0 MiB   12
> GiB
> >  6.5 TiB  40.77  1.00  182      up
> >  17    hdd  10.98349   1.00000   11 TiB  4.8 TiB  4.6 TiB   22 MiB   14
> GiB
> >  6.1 TiB  44.16  1.08  196      up
> >  32    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   37 MiB   12
> GiB
> >  6.7 TiB  38.80  0.95  173      up
> >  47    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB  655 KiB   11
> GiB
> >  6.8 TiB  38.16  0.93  184      up
> >  60    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   19 MiB   12
> GiB
> >  6.6 TiB  39.47  0.96  176      up
> >  74    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB   28 MiB   12
> GiB
> >  6.8 TiB  38.10  0.93  187      up
> >  83    hdd  10.98349   1.00000   11 TiB  4.8 TiB  4.5 TiB  1.9 MiB   14
> GiB
> >  6.2 TiB  43.47  1.06  180      up
> >  96    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   38 MiB   12
> GiB
> >  6.7 TiB  38.80  0.95  181      up
> > 110    hdd  10.98349   1.00000   11 TiB  4.5 TiB  4.2 TiB  4.3 MiB   13
> GiB
> >  6.5 TiB  40.79  1.00  174      up
> > 123    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB  1.9 MiB   13
> GiB
> >  6.8 TiB  38.11  0.93  173      up
> > 136    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   43 MiB   12
> GiB
> >  6.6 TiB  39.46  0.96  179      up
> > .....
> >
> > PG      OBJECTS  DEGRADED  MISPLACED  UNFOUND  BYTES         OMAP_BYTES*
> >  OMAP_KEYS*  LOG   LOG_DUPS  STATE                        SINCE
> > 52.0      80121         0          0        0  323105248            0
> >         0  1747      3000                 active+clean    94m
> > 52.1      79751         0          0        0  321711556766            0
> >         0  1727      3000                 active+clean    21h
> > 52.2      80243         0          0        0  323711878626            0
> >         0  1618      3000                 active+clean    30h
> > 52.3      79892         0          0        0  322166010020            0
> >         0  1627      3000                 active+clean     9h
> > 52.4      80267         0          0        0  323708219486            0
> >         0  1658      3000                 active+clean     5h
> > 52.5      79996         0          0        0  322331504454            0
> >         0  1722      3000                 active+clean    18h
> > 52.6      80190         0          0        0  323460394402            0
> >         0  1759      3000                 active+clean    15h
> > 52.7      79998         0          0        0  322769143546            0
> >         0  1720      3000                 active+clean    26h
> > 52.8      80292         0          0        0  323932173152            0
> >         0  1691      3000                 active+clean    21h
> > 52.9      79808         0          0        0  321910742702            0
> >         0  1675      3000                 active+clean     7h
> > 52.a      79751         0          0        0  321578061334            0
> >         0  1822      3000                 active+clean    26h
> > 52.b      80287         0          0        0  323905164642            0
> >         0  1793      3000                 active+clean     6h
> > ....
> > Thanks Eugen
> >
> > Vào Th 2, 7 thg 10, 2024 vào lúc 14:45 Eugen Block <eblock@xxxxxx> đã
> > viết:
> >
> >> Hi,
> >>
> >> disabling scrubbing in general is bad idea, because you won't notice
> >> any data corruption except when it might be too late.
> >> But you can fine tune scrubbing, for example increase the interval to
> >> allow fewer scrubs to finish in a longer interval. Or if the client
> >> load is mainly during business hours, adjust osd_scrub_begin_hour and
> >> osd_scrub_end_hour to your needs.
> >> And it also depends on the size of your PGs. The larger the PGs are,
> >> the longer a deep-scrub would take. So splitting PGs can have a quite
> >> positive effect in general. Inspect 'ceph osd df' output as well as
> >> 'ceph pg ls' (BYTES column), you can also share it here if you need
> >> assistance interpreting those values.
> >>
> >> Regards,
> >> Eugen
> >>
> >> Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:
> >>
> >> > Hi ceph users!
> >> >
> >> > How about the disable scrub and deep-scrub, i want to disable it
> because
> >> of
> >> > its effect on many I/O of my cluster.
> >> > If I disable scrub how will it affect my cluster?
> >> > If enabled, scrubbing is not complete and takes a long time.
> >> >
> >> >
> >> > Thank
> >> > Skype: tranphong079
> >> > _______________________________________________
> >> > ceph-users mailing list -- ceph-users@xxxxxxx
> >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>
> >>
> >> _______________________________________________
> >> ceph-users mailing list -- ceph-users@xxxxxxx
> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>
> >
> >
> > --
> > Trân trọng,
> >
> ----------------------------------------------------------------------------
> >
> > *Tran Thanh Phong*
> >
> > Email: tranphong079@xxxxxxxxx
> > Skype: tranphong079
>
>
>
>

-- 
Trân trọng,
----------------------------------------------------------------------------

*Tran Thanh Phong*

Email: tranphong079@xxxxxxxxx
Skype: tranphong079
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux