Re: About scrub and deep-scrub

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Well, if you have more PGs, the deep-scrubs per PG finish faster, but you have more PGs to scrub. But that way you can stretch the deep-scrub interval. As I said, I can not recommend to increase them in general. Overloading OSDs with too many PGs can have a negative effect as well. If you can, I'd probably rather add more (smaller) OSDs to spread the PGs.

Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:

Hi Eugen

Adding PG to the cluster only helps to reduce PG size and reduce scrub time
is right? >150 PG and <200 in osd i think it's good number

My cluster I/O 2-3GB/s real time in 24h IOPS 2K. If choosing to increase PG
is it a good choice?

Vào Th 2, 7 thg 10, 2024 vào lúc 15:49 Eugen Block <eblock@xxxxxx> đã
viết:

So your PGs for pool 52 have a size of around 320 GB, that is quite a
lot and not surprising that deep-scrubs take a long time. At the same
time, your PGs per OSD are already > 150. We had a similar situation
on a customer cluster this year as well, also with 12 TB drives. We
decided to increase the pg_num anyway to reduce the pg sizes. They
currently have around 380 PGs per large OSD (they have lots of smaller
OSDs as well) which still works fine. But they're using it as an
archive, so the IO is not very high. If you would decide to split PGs,
keep in mind to increase mon_max_pg_per_osd and
osd_max_pg_per_osd_hard_ratio as well. I can't explicitly recommend to
double your PGs per OSD as I'm not familiar with your cluster, the
load etc. It's just something to think about.
Doubling the PG count would reduce the PG size to around 160 GB, which
is still a lot, but I probably wouldn't go further than that.
The OSD utilization is only around 40%, in this case a cluster with
more (smaller) OSDs would probably have made more sense.

Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:

> Hi Eugen
>
> Can you see and give me some advice, number of PG and PG size..
>
> ID   CLASS  WEIGHT    REWEIGHT  SIZE     RAW USE  DATA     OMAP     META
>   AVAIL    %USE   VAR   PGS  STATUS
>   2    hdd  10.98349   1.00000   11 TiB  4.5 TiB  4.2 TiB  5.0 MiB   12
GiB
>  6.5 TiB  40.77  1.00  182      up
>  17    hdd  10.98349   1.00000   11 TiB  4.8 TiB  4.6 TiB   22 MiB   14
GiB
>  6.1 TiB  44.16  1.08  196      up
>  32    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   37 MiB   12
GiB
>  6.7 TiB  38.80  0.95  173      up
>  47    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB  655 KiB   11
GiB
>  6.8 TiB  38.16  0.93  184      up
>  60    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   19 MiB   12
GiB
>  6.6 TiB  39.47  0.96  176      up
>  74    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB   28 MiB   12
GiB
>  6.8 TiB  38.10  0.93  187      up
>  83    hdd  10.98349   1.00000   11 TiB  4.8 TiB  4.5 TiB  1.9 MiB   14
GiB
>  6.2 TiB  43.47  1.06  180      up
>  96    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   38 MiB   12
GiB
>  6.7 TiB  38.80  0.95  181      up
> 110    hdd  10.98349   1.00000   11 TiB  4.5 TiB  4.2 TiB  4.3 MiB   13
GiB
>  6.5 TiB  40.79  1.00  174      up
> 123    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB  1.9 MiB   13
GiB
>  6.8 TiB  38.11  0.93  173      up
> 136    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   43 MiB   12
GiB
>  6.6 TiB  39.46  0.96  179      up
> .....
>
> PG      OBJECTS  DEGRADED  MISPLACED  UNFOUND  BYTES         OMAP_BYTES*
>  OMAP_KEYS*  LOG   LOG_DUPS  STATE                        SINCE
> 52.0      80121         0          0        0  323105248            0
>         0  1747      3000                 active+clean    94m
> 52.1      79751         0          0        0  321711556766            0
>         0  1727      3000                 active+clean    21h
> 52.2      80243         0          0        0  323711878626            0
>         0  1618      3000                 active+clean    30h
> 52.3      79892         0          0        0  322166010020            0
>         0  1627      3000                 active+clean     9h
> 52.4      80267         0          0        0  323708219486            0
>         0  1658      3000                 active+clean     5h
> 52.5      79996         0          0        0  322331504454            0
>         0  1722      3000                 active+clean    18h
> 52.6      80190         0          0        0  323460394402            0
>         0  1759      3000                 active+clean    15h
> 52.7      79998         0          0        0  322769143546            0
>         0  1720      3000                 active+clean    26h
> 52.8      80292         0          0        0  323932173152            0
>         0  1691      3000                 active+clean    21h
> 52.9      79808         0          0        0  321910742702            0
>         0  1675      3000                 active+clean     7h
> 52.a      79751         0          0        0  321578061334            0
>         0  1822      3000                 active+clean    26h
> 52.b      80287         0          0        0  323905164642            0
>         0  1793      3000                 active+clean     6h
> ....
> Thanks Eugen
>
> Vào Th 2, 7 thg 10, 2024 vào lúc 14:45 Eugen Block <eblock@xxxxxx> đã
> viết:
>
>> Hi,
>>
>> disabling scrubbing in general is bad idea, because you won't notice
>> any data corruption except when it might be too late.
>> But you can fine tune scrubbing, for example increase the interval to
>> allow fewer scrubs to finish in a longer interval. Or if the client
>> load is mainly during business hours, adjust osd_scrub_begin_hour and
>> osd_scrub_end_hour to your needs.
>> And it also depends on the size of your PGs. The larger the PGs are,
>> the longer a deep-scrub would take. So splitting PGs can have a quite
>> positive effect in general. Inspect 'ceph osd df' output as well as
>> 'ceph pg ls' (BYTES column), you can also share it here if you need
>> assistance interpreting those values.
>>
>> Regards,
>> Eugen
>>
>> Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:
>>
>> > Hi ceph users!
>> >
>> > How about the disable scrub and deep-scrub, i want to disable it
because
>> of
>> > its effect on many I/O of my cluster.
>> > If I disable scrub how will it affect my cluster?
>> > If enabled, scrubbing is not complete and takes a long time.
>> >
>> >
>> > Thank
>> > Skype: tranphong079
>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users@xxxxxxx
>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>
>
> --
> Trân trọng,
>
----------------------------------------------------------------------------
>
> *Tran Thanh Phong*
>
> Email: tranphong079@xxxxxxxxx
> Skype: tranphong079





--
Trân trọng,
----------------------------------------------------------------------------

*Tran Thanh Phong*

Email: tranphong079@xxxxxxxxx
Skype: tranphong079


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux