Hi Zhang, yeah i saw your answer already. At very first, you should make sure that there is no clock skew. This can cause some sideeffects. ---- According to http://docs.ceph.com/docs/master/rados/operations/placement-groups/ you have to: (OSDs * 100) Total PGs = ------------ pool size Means: 20 OSD's of you * 100 = 2000 Poolsize is: Where pool size is either the number of replicas for replicated pools or the K+M sum for erasure coded pools (as returned by ceph osd erasure-code-profile get). -- So lets say, you have 2 replications, you should have 1000 PG's. If you have 3 replications, you should have 2000 / 3 = 666 PG's. But you configured 4096 PGs. Thats simply far too much. Reduce it. Or, if you can not, get more OSD's into this. I dont know any other way. Good luck ! -- Mit freundlichen Gruessen / Best regards Oliver Dzombic IP-Interactive mailto:info@xxxxxxxxxxxxxxxxx Anschrift: IP Interactive UG ( haftungsbeschraenkt ) Zum Sonnenberg 1-3 63571 Gelnhausen HRB 93402 beim Amtsgericht Hanau Geschäftsführung: Oliver Dzombic Steuer Nr.: 35 236 3622 1 UST ID: DE274086107 Am 22.03.2016 um 19:02 schrieb Zhang Qiang: > Hi Oliver, > > Thanks for your reply to my question on Ceph mailing list. I somehow > wasn't able to receive your reply in my mailbox, but I saw your reply in > the archive, so I have to mail you personally. > > I have pasted the whole ceph health output on gist: > https://gist.github.com/dotSlashLu/22623b4cefa06a46e0d4 > > Hope this will help. Thank you! _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com