Re: About scrub and deep-scrub

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



So your PGs for pool 52 have a size of around 320 GB, that is quite a lot and not surprising that deep-scrubs take a long time. At the same time, your PGs per OSD are already > 150. We had a similar situation on a customer cluster this year as well, also with 12 TB drives. We decided to increase the pg_num anyway to reduce the pg sizes. They currently have around 380 PGs per large OSD (they have lots of smaller OSDs as well) which still works fine. But they're using it as an archive, so the IO is not very high. If you would decide to split PGs, keep in mind to increase mon_max_pg_per_osd and osd_max_pg_per_osd_hard_ratio as well. I can't explicitly recommend to double your PGs per OSD as I'm not familiar with your cluster, the load etc. It's just something to think about. Doubling the PG count would reduce the PG size to around 160 GB, which is still a lot, but I probably wouldn't go further than that. The OSD utilization is only around 40%, in this case a cluster with more (smaller) OSDs would probably have made more sense.

Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:

Hi Eugen

Can you see and give me some advice, number of PG and PG size..

ID   CLASS  WEIGHT    REWEIGHT  SIZE     RAW USE  DATA     OMAP     META
  AVAIL    %USE   VAR   PGS  STATUS
  2    hdd  10.98349   1.00000   11 TiB  4.5 TiB  4.2 TiB  5.0 MiB   12 GiB
 6.5 TiB  40.77  1.00  182      up
 17    hdd  10.98349   1.00000   11 TiB  4.8 TiB  4.6 TiB   22 MiB   14 GiB
 6.1 TiB  44.16  1.08  196      up
 32    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   37 MiB   12 GiB
 6.7 TiB  38.80  0.95  173      up
 47    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB  655 KiB   11 GiB
 6.8 TiB  38.16  0.93  184      up
 60    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   19 MiB   12 GiB
 6.6 TiB  39.47  0.96  176      up
 74    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB   28 MiB   12 GiB
 6.8 TiB  38.10  0.93  187      up
 83    hdd  10.98349   1.00000   11 TiB  4.8 TiB  4.5 TiB  1.9 MiB   14 GiB
 6.2 TiB  43.47  1.06  180      up
 96    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   38 MiB   12 GiB
 6.7 TiB  38.80  0.95  181      up
110    hdd  10.98349   1.00000   11 TiB  4.5 TiB  4.2 TiB  4.3 MiB   13 GiB
 6.5 TiB  40.79  1.00  174      up
123    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB  1.9 MiB   13 GiB
 6.8 TiB  38.11  0.93  173      up
136    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   43 MiB   12 GiB
 6.6 TiB  39.46  0.96  179      up
.....

PG      OBJECTS  DEGRADED  MISPLACED  UNFOUND  BYTES         OMAP_BYTES*
 OMAP_KEYS*  LOG   LOG_DUPS  STATE                        SINCE
52.0      80121         0          0        0  323105248            0
        0  1747      3000                 active+clean    94m
52.1      79751         0          0        0  321711556766            0
        0  1727      3000                 active+clean    21h
52.2      80243         0          0        0  323711878626            0
        0  1618      3000                 active+clean    30h
52.3      79892         0          0        0  322166010020            0
        0  1627      3000                 active+clean     9h
52.4      80267         0          0        0  323708219486            0
        0  1658      3000                 active+clean     5h
52.5      79996         0          0        0  322331504454            0
        0  1722      3000                 active+clean    18h
52.6      80190         0          0        0  323460394402            0
        0  1759      3000                 active+clean    15h
52.7      79998         0          0        0  322769143546            0
        0  1720      3000                 active+clean    26h
52.8      80292         0          0        0  323932173152            0
        0  1691      3000                 active+clean    21h
52.9      79808         0          0        0  321910742702            0
        0  1675      3000                 active+clean     7h
52.a      79751         0          0        0  321578061334            0
        0  1822      3000                 active+clean    26h
52.b      80287         0          0        0  323905164642            0
        0  1793      3000                 active+clean     6h
....
Thanks Eugen

Vào Th 2, 7 thg 10, 2024 vào lúc 14:45 Eugen Block <eblock@xxxxxx> đã
viết:

Hi,

disabling scrubbing in general is bad idea, because you won't notice
any data corruption except when it might be too late.
But you can fine tune scrubbing, for example increase the interval to
allow fewer scrubs to finish in a longer interval. Or if the client
load is mainly during business hours, adjust osd_scrub_begin_hour and
osd_scrub_end_hour to your needs.
And it also depends on the size of your PGs. The larger the PGs are,
the longer a deep-scrub would take. So splitting PGs can have a quite
positive effect in general. Inspect 'ceph osd df' output as well as
'ceph pg ls' (BYTES column), you can also share it here if you need
assistance interpreting those values.

Regards,
Eugen

Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:

> Hi ceph users!
>
> How about the disable scrub and deep-scrub, i want to disable it because
of
> its effect on many I/O of my cluster.
> If I disable scrub how will it affect my cluster?
> If enabled, scrubbing is not complete and takes a long time.
>
>
> Thank
> Skype: tranphong079
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



--
Trân trọng,
----------------------------------------------------------------------------

*Tran Thanh Phong*

Email: tranphong079@xxxxxxxxx
Skype: tranphong079


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux