omg I'm soooo stupid, we just said that monitors handle the pool creation and I knew that the [mon] has to be edited so I really don't know what happened into my mind to restart my OSDs... Anyway I just checked the mon socket and it works, so no bug here. Thanks again :) On Fri, Oct 26, 2012 at 6:18 PM, Sage Weil <sage@xxxxxxxxxxx> wrote: > On Fri, 26 Oct 2012, S?bastien Han wrote: >> > It's the monitor who is preparing the new pool before creating it. I had a >> > discussion regarding this with Greg some time ago, can't re-call if this was >> > on IRC or the ml though. >> >> Well actually I think I was in the conversation and it was on IRC, but >> I had completely forget about it. -_- >> >> Thanks the clarification. It may sound weird but if the monitors >> handles the pool creation, it makes sense. >> >> Now I'm surprised to see that the admin socket has a different value. > > The admin socket on the monitor? Keep in mind that the actual creation is > coordinated by the lead monitor... 'ceph quorum_status' will tell you > who that is (the first one listed who is in the quorum). *that's* the one > who's config value for 'osd pool default size' matters. > >> Moreover after changing the configuration and set the default size to >> 2 and restarted every OSD. It looks like the cluster continue to >> create a pool size of 3... > > OSDs are not involved; no need to touch their config or restart them. > >> Could be a bug then, any idea? > > If you can confirm that the monitor who did the creation created a pool of > a size different that its osd default pool size, then yes... but let's > confirm that! > > In any case, it's no big deal to change it after the pool is created: > > ceph osd pool set <poolname> size <num replicas> > > sage > > >> >> On Fri, Oct 26, 2012 at 3:41 PM, Wido den Hollander <wido@xxxxxxxxx> wrote: >> > On 10/26/2012 10:17 AM, S?bastien Han wrote: >> >> >> >> Hi Cephers! >> >> >> >> Some question about this parameter: >> >> >> >> - Why does this parameter need to be in the [mon] section to work? >> > >> > >> > It's the monitor who is preparing the new pool before creating it. I had a >> > discussion regarding this with Greg some time ago, can't re-call if this was >> > on IRC or the ml though. >> > >> > It however contradicts what's in the docs: >> > http://ceph.com/docs/master/config-cluster/osd-config-ref/ >> > >> > The docs seem to be wrong at this point: >> > >> > * src/mon/OSDMonitor.cc * >> > >> > int OSDMonitor::prepare_new_pool(string& name, uint64_t auid, int >> > crush_rule, >> > unsigned pg_num, unsigned pgp_num) >> > { >> > .. >> > .. >> > pending_inc.new_pools[pool].size = g_conf->osd_pool_default_size; >> > .. >> > .. >> > } >> > >> > This is done by the monitor. >> > >> > There seems to be some reference for this in src/osd/OSDMap.cc, but that >> > seems to be for initializing the cluster since it creates the data, metadata >> > and rbd pool there. >> > >> > This was fixed in this commit: >> > https://github.com/ceph/ceph/commit/1292436b8f88ce203bdba97278ce368b1dfa025f >> > >> > Seems to be because of this message on the ml last year: >> > http://comments.gmane.org/gmane.comp.file-systems.ceph.devel/2983 >> > >> > So yes, it is odd that something prefixed with "osd" should go into the mon >> > section. >> > >> > Wido >> > >> >> - In my ceph.conf I set a default size of 3: >> >> >> >> # rados mkpool lol >> >> successfully created pool lol >> >> >> >> # ceph osd dump | grep lol >> >> pool 31 'lol' rep size 3 crush_ruleset 0 object_hash rjenkins pg_num 8 >> >> pgp_num 8 last_change 430 owner 18446744073709551615 >> >> >> >> Now if I retrieve the admin daemon I get, >> >> >> >> # ceph --admin-daemon /var/run/ceph/ceph-osd.0.asok config show | grep >> >> 'osd_pool_default_size' >> >> osd_pool_default_size = 2 >> >> >> >> Why? Did I something wrong? >> >> >> >> Thanks :) >> >> >> >> Cheers! >> >> -- >> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> >> More majordomo info at http://vger.kernel.org/majordomo-info.html >> >> >> > >> -- >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> More majordomo info at http://vger.kernel.org/majordomo-info.html >> >> -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html