So your PGs for pool 52 have a size of around 320 GB, that is quite a
lot and not surprising that deep-scrubs take a long time. At the same
time, your PGs per OSD are already > 150. We had a similar situation
on a customer cluster this year as well, also with 12 TB drives. We
decided to increase the pg_num anyway to reduce the pg sizes. They
currently have around 380 PGs per large OSD (they have lots of smaller
OSDs as well) which still works fine. But they're using it as an
archive, so the IO is not very high. If you would decide to split PGs,
keep in mind to increase mon_max_pg_per_osd and
osd_max_pg_per_osd_hard_ratio as well. I can't explicitly recommend to
double your PGs per OSD as I'm not familiar with your cluster, the
load etc. It's just something to think about.
Doubling the PG count would reduce the PG size to around 160 GB, which
is still a lot, but I probably wouldn't go further than that.
The OSD utilization is only around 40%, in this case a cluster with
more (smaller) OSDs would probably have made more sense.
Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:
Hi Eugen
Can you see and give me some advice, number of PG and PG size..
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META
AVAIL %USE VAR PGS STATUS
2 hdd 10.98349 1.00000 11 TiB 4.5 TiB 4.2 TiB 5.0 MiB 12 GiB
6.5 TiB 40.77 1.00 182 up
17 hdd 10.98349 1.00000 11 TiB 4.8 TiB 4.6 TiB 22 MiB 14 GiB
6.1 TiB 44.16 1.08 196 up
32 hdd 10.98349 1.00000 11 TiB 4.3 TiB 4.0 TiB 37 MiB 12 GiB
6.7 TiB 38.80 0.95 173 up
47 hdd 10.98349 1.00000 11 TiB 4.2 TiB 3.9 TiB 655 KiB 11 GiB
6.8 TiB 38.16 0.93 184 up
60 hdd 10.98349 1.00000 11 TiB 4.3 TiB 4.0 TiB 19 MiB 12 GiB
6.6 TiB 39.47 0.96 176 up
74 hdd 10.98349 1.00000 11 TiB 4.2 TiB 3.9 TiB 28 MiB 12 GiB
6.8 TiB 38.10 0.93 187 up
83 hdd 10.98349 1.00000 11 TiB 4.8 TiB 4.5 TiB 1.9 MiB 14 GiB
6.2 TiB 43.47 1.06 180 up
96 hdd 10.98349 1.00000 11 TiB 4.3 TiB 4.0 TiB 38 MiB 12 GiB
6.7 TiB 38.80 0.95 181 up
110 hdd 10.98349 1.00000 11 TiB 4.5 TiB 4.2 TiB 4.3 MiB 13 GiB
6.5 TiB 40.79 1.00 174 up
123 hdd 10.98349 1.00000 11 TiB 4.2 TiB 3.9 TiB 1.9 MiB 13 GiB
6.8 TiB 38.11 0.93 173 up
136 hdd 10.98349 1.00000 11 TiB 4.3 TiB 4.0 TiB 43 MiB 12 GiB
6.6 TiB 39.46 0.96 179 up
.....
PG OBJECTS DEGRADED MISPLACED UNFOUND BYTES OMAP_BYTES*
OMAP_KEYS* LOG LOG_DUPS STATE SINCE
52.0 80121 0 0 0 323105248 0
0 1747 3000 active+clean 94m
52.1 79751 0 0 0 321711556766 0
0 1727 3000 active+clean 21h
52.2 80243 0 0 0 323711878626 0
0 1618 3000 active+clean 30h
52.3 79892 0 0 0 322166010020 0
0 1627 3000 active+clean 9h
52.4 80267 0 0 0 323708219486 0
0 1658 3000 active+clean 5h
52.5 79996 0 0 0 322331504454 0
0 1722 3000 active+clean 18h
52.6 80190 0 0 0 323460394402 0
0 1759 3000 active+clean 15h
52.7 79998 0 0 0 322769143546 0
0 1720 3000 active+clean 26h
52.8 80292 0 0 0 323932173152 0
0 1691 3000 active+clean 21h
52.9 79808 0 0 0 321910742702 0
0 1675 3000 active+clean 7h
52.a 79751 0 0 0 321578061334 0
0 1822 3000 active+clean 26h
52.b 80287 0 0 0 323905164642 0
0 1793 3000 active+clean 6h
....
Thanks Eugen
Vào Th 2, 7 thg 10, 2024 vào lúc 14:45 Eugen Block <eblock@xxxxxx> đã
viết:
Hi,
disabling scrubbing in general is bad idea, because you won't notice
any data corruption except when it might be too late.
But you can fine tune scrubbing, for example increase the interval to
allow fewer scrubs to finish in a longer interval. Or if the client
load is mainly during business hours, adjust osd_scrub_begin_hour and
osd_scrub_end_hour to your needs.
And it also depends on the size of your PGs. The larger the PGs are,
the longer a deep-scrub would take. So splitting PGs can have a quite
positive effect in general. Inspect 'ceph osd df' output as well as
'ceph pg ls' (BYTES column), you can also share it here if you need
assistance interpreting those values.
Regards,
Eugen
Zitat von Phong Tran Thanh <tranphong079@xxxxxxxxx>:
> Hi ceph users!
>
> How about the disable scrub and deep-scrub, i want to disable it because
of
> its effect on many I/O of my cluster.
> If I disable scrub how will it affect my cluster?
> If enabled, scrubbing is not complete and takes a long time.
>
>
> Thank
> Skype: tranphong079
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
--
Trân trọng,
----------------------------------------------------------------------------
*Tran Thanh Phong*
Email: tranphong079@xxxxxxxxx
Skype: tranphong079
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx