Oh ok. Because we have two types of ssds. 1.8TB and 3.6TB. The 1.8TB got around 90-100pgs and the 3.6TB around 150-190pgs Here is the output: RAW STORAGE: CLASS SIZE AVAIL USED RAW USED %RAW USED ssd 956 TiB 360 TiB 595 TiB 596 TiB 62.35 TOTAL 956 TiB 360 TiB 595 TiB 596 TiB 62.35 POOLS: POOL ID STORED OBJECTS USED %USED MAX AVAIL pool 1 1 580 GiB 149.31k 1.7 TiB 0.58 96 TiB pool 3 3 208 TiB 61.66M 593 TiB 67.21 96 TiB pool 4 4 7.9 MiB 2.01k 68 MiB 0 96 TiB pool 5 5 19 B 2 36 KiB 0 96 TiB Am Di., 16. Juni 2020 um 14:13 Uhr schrieb Dan van der Ster <dan@xxxxxxxxxxxxxx>: > > On Tue, Jun 16, 2020 at 2:00 PM Boris Behrens <bb@xxxxxxxxx> wrote: > > > > See inline comments > > Dan van der Ster <dan@xxxxxxxxxxxxxx> 于2020年6月16日周二 下午7:07写道: > > > > > > Could you share the output of > > > > > > ceph osd pool ls detail > > > > pool 1 'pool 1' replicated size 3 min_size 2 crush_rule 0 object_hash > > rjenkins pg_num 256 pgp_num 256 autoscale_mode warn last_change > > 2318859 flags hashpspool min_write_recency_for_promote 1 stripe_width > > 0 application rbd > > pool 3 'pool 3' replicated size 3 min_size 2 crush_rule 0 object_hash > > rjenkins pg_num 16384 pgp_num 16384 autoscale_mode warn last_change > > 2544040 lfor 0/0/1952329 flags hashpspool,selfmanaged_snaps > > min_write_recency_for_promote 1 stripe_width 0 application rbd > > pool 4 'pool 4' replicated size 3 min_size 2 crush_rule 0 object_hash > > rjenkins pg_num 8 pgp_num 8 autoscale_mode warn last_change 2318859 > > flags hashpspool min_write_recency_for_promote 1 stripe_width 0 > > application rbd > > pool 5 'pool 5' replicated size 3 min_size 2 crush_rule 0 object_hash > > rjenkins pg_num 256 pgp_num 256 autoscale_mode warn last_change > > 2318859 flags hashpspool,selfmanaged_snaps > > min_write_recency_for_promote 1 stripe_width 0 application rbd > > > > OK now maybe share the output of `ceph df` so we can see how much data > is in each pool? > > Assuming that the majority of your data is in 'pool 3' with 16384 PGs, > then your current PG values are just fine. (You should have around 110 > PGs per OSD). > The pg_autoscaler aims for 100 per OSD and doesn't make changes unless > a pool has 4x too few or too many PGs. > > Unless you are planning to put a large proportion of data into the > other pools, I'd leave pg_autoscaler disabled and move on to the next > task. > > -- Dan > > > the mgr module is not enabled yet. > > > > > > > > ? > > > > > > This way we can see how the pools are configured and help recommend if > > > pg_autoscaler is worth enabling. > > > > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx