Hello,
still do not really understand why this error message comes up.
The error message contains two significant numbers. The first one which
is easy to understand is the maximal value of pgs for each osd a
precompiled config variable (mon_max_pg_per_osd). The value on my
cluster is 250. This value is multiplied by the total number of osds
(88): The result is the value of maximal 22000 pgs for the whole cluster
(based on mon_max_pg_per_osd).
Well as said the whole cluster does only have 2592 pgs at the moment and
I wanted to add another 512+128 pgs.
Then there is the second value "22148". The meaning can only be pgs
since this value is beeing compared to the total number of pgs for the
whole cluster (based on "mon_max_pg_per_osd").
One solution to that problem could be that mon_max_pg_per_osd does not
as the named suggests mean: max pgs/osd, but instead max shards/osd. If
this is true, then multiplying my pgs by the number of shards (5+3 for
me) would be a value higher than 22000 if I add the shards for the new
pool I wanted to create.
If this is true, then the chosen naming of mon_max_pg_per_osd would
simply be misleading.
Thanks
Rainer
Am 01.12.22 um 19:00 schrieb Eugen Block:
"got unexpected control message: TASK ERROR: error with 'osd pool
create': mon_command failed - pg_num 512 size 8 would mean 22148
total pgs, which exceeds max 22000 (mon_max_pg_per_osd 250 *
num_in_osds 88)"
I also tried the direct way to create a new pool using:
ceph osd pool create <pool> 512 128 erasure <profile> but the error
message below remains.
What I do not understand now are the calculations behind the scenes
for the calculated total pg number of 22148. How is this total number
"22148" calculated?
I already reduced the number of pgs for the metadata pool of each
ec-pool and so I was able to create 4 pools in this way. But just for
fun I now tried to create ec-pool number 5 and I see the message from
above again.
Here are the pools created by now (scraped from ceph osd pool
autoscale-status):
Pool: Size: Bias: PG_NUM:
rbd 4599 1.0 32
px-a-data 528.2G 1.0 512
px-a-metadata 838.1k 1.0 128
px-b-data 0 1.0 512
px-b-metadata 19 1.0 128
px-c-data 0 1.0 512
px-c-metadata 19 1.0 128
px-d-data 0 1.0 512
px-d-metadata 0 1.0 128
So the total number of pgs for all pools is currently 2592 which is
far from 22148 pgs?
Any ideas?
Thanks Rainer
--
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Web: http://www.uni-koblenz.de/~krienke, Tel: +49261287
1312
PGP: http://www.uni-koblenz.de/~krienke/mypgp.html, ; Fax: +49261287
1001312
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
--
Rainer Krienke, Uni Koblenz, Rechenzentrum, A22, Universitaetsstrasse 1
56070 Koblenz, Tel: +49261287 1312 Fax +49261287 100 1312
Web: http://userpages.uni-koblenz.de/~krienke
PGP: http://userpages.uni-koblenz.de/~krienke/mypgp.html
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx