Hi,
We are currently upgrading our cluster from Nautilus to Octupus.
After upgrade of the mons and mgrs, we get warnings about the number of PGS.
Which parameter did change during upgrade to explain those new warnings.
Nothing else was changed.
Is it risky to change the pgs/pool as proposed in the warnings ? In
particular, to reduce from 4096 to 64 !!!
Thanks in advance,
Patrick
root@server4 ~]# ceph -s
cluster:
id: ba00c030-382f-4d75-b150-5b17f77e57fe
health: HEALTH_WARN
clients are using insecure global_id reclaim
6 pools have too few placement groups
9 pools have too many placement groups
services:
mon: 3 daemons, quorum server2,server5,server6 (age 66m)
mgr: server8(active, since 67m), standbys: server4, server1
osd: 244 osds: 244 up (since 12m), 244 in (since 2w)
rgw: 2 daemons active (server1, server4)
task status:
data:
pools: 16 pools, 11441 pgs
objects: 2.02M objects, 5.9 TiB
usage: 18 TiB used, 982 TiB / 1000 TiB avail
pgs: 11441 active+clean
io:
client: 862 KiB/s rd, 1.4 MiB/s wr, 61 op/s rd, 100 op/s wr
root@server4 ~]# ceph health detail
...
[WRN] POOL_TOO_MANY_PGS: 9 pools have too many placement groups
Pool default.rgw.buckets.index has 128 placement groups, should have 32
Pool default.rgw.buckets.data has 4096 placement groups, should have 64
Pool os_glance has 1024 placement groups, should have 32
...
[root@server4 ~]# ceph config get mon mon_max_pg_per_osd
250
In ceph.conf, we set also:
osd_max_pg_per_osd_hard_ratio = 3
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx