On Mon, Sep 8, 2014 at 10:08 AM, JIten Shah <jshah2005 at me.com> wrote: > While checking the health of the cluster, I ran to the following error: > > warning: health HEALTH_WARN too few pgs per osd (1< min 20) > > When I checked the pg and php numbers, I saw the value was the default value > of 64 > > ceph osd pool get data pg_num > pg_num: 64 > ceph osd pool get data pgp_num > pgp_num: 64 > > Checking the ceph documents, I updated the numbers to 2000 using the > following commands: > > ceph osd pool set data pg_num 2000 > ceph osd pool set data pgp_num 2000 > > It started resizing the data and saw health warnings again: > > health HEALTH_WARN 1 requests are blocked > 32 sec; pool data pg_num 2000 > > pgp_num 64 > > and then: > > ceph health detail > HEALTH_WARN 6 requests are blocked > 32 sec; 3 osds have slow requests > 5 ops are blocked > 65.536 sec > 1 ops are blocked > 32.768 sec > 1 ops are blocked > 32.768 sec on osd.16 > 1 ops are blocked > 65.536 sec on osd.77 > 4 ops are blocked > 65.536 sec on osd.98 > 3 osds have slow requests > > This error also went away after a day. > > ceph health detail > HEALTH_OK > > > Now, the question I have is, will this pg number remain effective on the > cluster, even if we restart MON or OSD?s on the individual disks? I haven?t > changed the values in /etc/ceph/ceph.conf. Do I need to make a change to the > ceph.conf and push that change to all the MON, MSD and OSD?s ? It's durable once the commands are successful on the monitors. You're all done. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com