Re: osd_pool_default_size

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, 26 Oct 2012, S?bastien Han wrote:
> > It's the monitor who is preparing the new pool before creating it. I had a
> > discussion regarding this with Greg some time ago, can't re-call if this was
> > on IRC or the ml though.
> 
> Well actually I think I was in the conversation and it was on IRC, but
> I had completely forget about it. -_-
> 
> Thanks the clarification. It may sound weird but if the monitors
> handles the pool creation, it makes sense.
> 
> Now I'm surprised to see that the admin socket has a different value.

The admin socket on the monitor?  Keep in mind that the actual creation is 
coordinated by the lead monitor... 'ceph quorum_status' will tell you 
who that is (the first one listed who is in the quorum).  *that's* the one 
who's config value for 'osd pool default size' matters.

> Moreover after changing the configuration and set the default size to
> 2 and restarted every OSD. It looks like the cluster continue to
> create a pool size of 3...

OSDs are not involved; no need to touch their config or restart them.

> Could be a bug then, any idea?

If you can confirm that the monitor who did the creation created a pool of 
a size different that its osd default pool size, then yes... but let's 
confirm that!

In any case, it's no big deal to change it after the pool is created:

	ceph osd pool set <poolname> size <num replicas>

sage


> 
> On Fri, Oct 26, 2012 at 3:41 PM, Wido den Hollander <wido@xxxxxxxxx> wrote:
> > On 10/26/2012 10:17 AM, S?bastien Han wrote:
> >>
> >> Hi Cephers!
> >>
> >> Some question about this parameter:
> >>
> >> - Why does this parameter need to be in the [mon] section to work?
> >
> >
> > It's the monitor who is preparing the new pool before creating it. I had a
> > discussion regarding this with Greg some time ago, can't re-call if this was
> > on IRC or the ml though.
> >
> > It however contradicts what's in the docs:
> > http://ceph.com/docs/master/config-cluster/osd-config-ref/
> >
> > The docs seem to be wrong at this point:
> >
> > * src/mon/OSDMonitor.cc *
> >
> > int OSDMonitor::prepare_new_pool(string& name, uint64_t auid, int
> > crush_rule,
> >                                  unsigned pg_num, unsigned pgp_num)
> > {
> > ..
> > ..
> >   pending_inc.new_pools[pool].size = g_conf->osd_pool_default_size;
> > ..
> > ..
> > }
> >
> > This is done by the monitor.
> >
> > There seems to be some reference for this in src/osd/OSDMap.cc, but that
> > seems to be for initializing the cluster since it creates the data, metadata
> > and rbd pool there.
> >
> > This was fixed in this commit:
> > https://github.com/ceph/ceph/commit/1292436b8f88ce203bdba97278ce368b1dfa025f
> >
> > Seems to be because of this message on the ml last year:
> > http://comments.gmane.org/gmane.comp.file-systems.ceph.devel/2983
> >
> > So yes, it is odd that something prefixed with "osd" should go into the mon
> > section.
> >
> > Wido
> >
> >> - In my ceph.conf I set a default size of 3:
> >>
> >> # rados mkpool lol
> >> successfully created pool lol
> >>
> >> # ceph osd dump | grep lol
> >> pool 31 'lol' rep size 3 crush_ruleset 0 object_hash rjenkins pg_num 8
> >> pgp_num 8 last_change 430 owner 18446744073709551615
> >>
> >> Now if I retrieve the admin daemon I get,
> >>
> >> # ceph --admin-daemon /var/run/ceph/ceph-osd.0.asok config show | grep
> >> 'osd_pool_default_size'
> >> osd_pool_default_size = 2
> >>
> >> Why? Did I something wrong?
> >>
> >> Thanks :)
> >>
> >> Cheers!
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> the body of a message to majordomo@xxxxxxxxxxxxxxx
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>
> >
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux