So, if it doesn?t refer to the entry in ceph.conf. Where does it actually store the new value? ?Jiten On Sep 8, 2014, at 10:31 AM, Gregory Farnum <greg at inktank.com> wrote: > On Mon, Sep 8, 2014 at 10:08 AM, JIten Shah <jshah2005 at me.com> wrote: >> While checking the health of the cluster, I ran to the following error: >> >> warning: health HEALTH_WARN too few pgs per osd (1< min 20) >> >> When I checked the pg and php numbers, I saw the value was the default value >> of 64 >> >> ceph osd pool get data pg_num >> pg_num: 64 >> ceph osd pool get data pgp_num >> pgp_num: 64 >> >> Checking the ceph documents, I updated the numbers to 2000 using the >> following commands: >> >> ceph osd pool set data pg_num 2000 >> ceph osd pool set data pgp_num 2000 >> >> It started resizing the data and saw health warnings again: >> >> health HEALTH_WARN 1 requests are blocked > 32 sec; pool data pg_num 2000 > >> pgp_num 64 >> >> and then: >> >> ceph health detail >> HEALTH_WARN 6 requests are blocked > 32 sec; 3 osds have slow requests >> 5 ops are blocked > 65.536 sec >> 1 ops are blocked > 32.768 sec >> 1 ops are blocked > 32.768 sec on osd.16 >> 1 ops are blocked > 65.536 sec on osd.77 >> 4 ops are blocked > 65.536 sec on osd.98 >> 3 osds have slow requests >> >> This error also went away after a day. >> >> ceph health detail >> HEALTH_OK >> >> >> Now, the question I have is, will this pg number remain effective on the >> cluster, even if we restart MON or OSD?s on the individual disks? I haven?t >> changed the values in /etc/ceph/ceph.conf. Do I need to make a change to the >> ceph.conf and push that change to all the MON, MSD and OSD?s ? > > It's durable once the commands are successful on the monitors. You're all done. > -Greg > Software Engineer #42 @ http://inktank.com | http://ceph.com > _______________________________________________ > ceph-users mailing list > ceph-users at lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com