I got it, the pg_num suggested is the total, I need to divide it by the number of replications.
Thanks Oliver, your answer is very thorough and helpful!
On 23 March 2016 at 02:19, Oliver Dzombic <info@xxxxxxxxxxxxxxxxx> wrote:
Hi Zhang,
yeah i saw your answer already.
At very first, you should make sure that there is no clock skew.
This can cause some sideeffects.
----
According to
http://docs.ceph.com/docs/master/rados/operations/placement-groups/
you have to:
(OSDs * 100)
Total PGs = ------------
pool size
Means:
20 OSD's of you * 100 = 2000
Poolsize is:
Where pool size is either the number of replicas for replicated pools or
the K+M sum for erasure coded pools (as returned by ceph osd
erasure-code-profile get).
--
So lets say, you have 2 replications, you should have 1000 PG's.
If you have 3 replications, you should have 2000 / 3 = 666 PG's.
But you configured 4096 PGs. Thats simply far too much.
Reduce it. Or, if you can not, get more OSD's into this.
I dont know any other way.
Good luck !
--
Mit freundlichen Gruessen / Best regards
Oliver Dzombic
IP-Interactive
mailto:info@xxxxxxxxxxxxxxxxx
Anschrift:
IP Interactive UG ( haftungsbeschraenkt )
Zum Sonnenberg 1-3
63571 Gelnhausen
HRB 93402 beim Amtsgericht Hanau
Geschäftsführung: Oliver Dzombic
Steuer Nr.: 35 236 3622 1
UST ID: DE274086107
Am 22.03.2016 um 19:02 schrieb Zhang Qiang:
> Hi Oliver,
>
> Thanks for your reply to my question on Ceph mailing list. I somehow
> wasn't able to receive your reply in my mailbox, but I saw your reply in
> the archive, so I have to mail you personally.
>
> I have pasted the whole ceph health output on gist:
> https://gist.github.com/dotSlashLu/22623b4cefa06a46e0d4
>
> Hope this will help. Thank you!
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com