Re: About scrub and deep-scrub

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Eugen

Can you see and give me some advice, number of PG and PG size..

ID   CLASS  WEIGHT    REWEIGHT  SIZE     RAW USE  DATA     OMAP     META
  AVAIL    %USE   VAR   PGS  STATUS
  2    hdd  10.98349   1.00000   11 TiB  4.5 TiB  4.2 TiB  5.0 MiB   12 GiB
 6.5 TiB  40.77  1.00  182      up
 17    hdd  10.98349   1.00000   11 TiB  4.8 TiB  4.6 TiB   22 MiB   14 GiB
 6.1 TiB  44.16  1.08  196      up
 32    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   37 MiB   12 GiB
 6.7 TiB  38.80  0.95  173      up
 47    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB  655 KiB   11 GiB
 6.8 TiB  38.16  0.93  184      up
 60    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   19 MiB   12 GiB
 6.6 TiB  39.47  0.96  176      up
 74    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB   28 MiB   12 GiB
 6.8 TiB  38.10  0.93  187      up
 83    hdd  10.98349   1.00000   11 TiB  4.8 TiB  4.5 TiB  1.9 MiB   14 GiB
 6.2 TiB  43.47  1.06  180      up
 96    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   38 MiB   12 GiB
 6.7 TiB  38.80  0.95  181      up
110    hdd  10.98349   1.00000   11 TiB  4.5 TiB  4.2 TiB  4.3 MiB   13 GiB
 6.5 TiB  40.79  1.00  174      up
123    hdd  10.98349   1.00000   11 TiB  4.2 TiB  3.9 TiB  1.9 MiB   13 GiB
 6.8 TiB  38.11  0.93  173      up
136    hdd  10.98349   1.00000   11 TiB  4.3 TiB  4.0 TiB   43 MiB   12 GiB
 6.6 TiB  39.46  0.96  179      up
.....

PG      OBJECTS  DEGRADED  MISPLACED  UNFOUND  BYTES         OMAP_BYTES*
 OMAP_KEYS*  LOG   LOG_DUPS  STATE                        SINCE
52.0      80121         0          0        0  323105721248            0
        0  1747      3000                 active+clean    94m
52.1      79751         0          0        0  321711556766            0
        0  1727      3000                 active+clean    21h
52.2      80243         0          0        0  323711878626            0
        0  1618      3000                 active+clean    30h
52.3      79892         0          0        0  322166010020            0
        0  1627      3000                 active+clean     9h
52.4      80267         0          0        0  323708219486            0
        0  1658      3000                 active+clean     5h
52.5      79996         0          0        0  322331504454            0
        0  1722      3000                 active+clean    18h
52.6      80190         0          0        0  323460394402            0
        0  1759      3000                 active+clean    15h
52.7      79998         0          0        0  322769143546            0
        0  1720      3000                 active+clean    26h
52.8      80292         0          0        0  323932173152            0
        0  1691      3000                 active+clean    21h
52.9      79808         0          0        0  321910742702            0
        0  1675      3000                 active+clean     7h
52.a      79751         0          0        0  321578061334            0
        0  1822      3000                 active+clean    26h
52.b      80287         0          0        0  323905164642            0
        0  1793      3000                 active+clean     6h
....
Thanks Eugen

Vào Th 2, 7 thg 10, 2024 vào lúc 14:45 Eugen Block <eblock@xxxxxx> đã
viết:

> Hi,
>
> disabling scrubbing in general is bad idea, because you won't notice
> any data corruption except when it might be too late.
> But you can fine tune scrubbing, for example increase the interval to
> allow fewer scrubs to finish in a longer interval. Or if the client
> load is mainly during business hours, adjust osd_scrub_begin_hour and
> osd_scrub_end_hour to your needs.
> And it also depends on the size of your PGs. The larger the PGs are,
> the longer a deep-scrub would take. So splitting PGs can have a quite
> positive effect in general. Inspect 'ceph osd df' output as well as
> 'ceph pg ls' (BYTES column), you can also share it here if you need
> assistance interpreting those values.
>
> Regards,
> Eugen
>
> Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:
>
> > Hi ceph users!
> >
> > How about the disable scrub and deep-scrub, i want to disable it because
> of
> > its effect on many I/O of my cluster.
> > If I disable scrub how will it affect my cluster?
> > If enabled, scrubbing is not complete and takes a long time.
> >
> >
> > Thank
> > Skype: tranphong079
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>


-- 
Trân trọng,
----------------------------------------------------------------------------

*Tran Thanh Phong*

Email: tranphong079@xxxxxxxxx
Skype: tranphong079
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux