On Mon, Apr 22, 2013 at 9:42 AM, John Nielsen <lists@xxxxxxxxxxxx> wrote: > On Mar 15, 2013, at 8:15 PM, Yehuda Sadeh <yehuda@xxxxxxxxxxx> wrote: > >> On Fri, Mar 15, 2013 at 5:06 PM, Mandell Degerness >> <mandell@xxxxxxxxxxxxxxx> wrote: >>> How are the pools used by rgw defined? >>> >>> Specifically, if I want to ensure that all of the data stored by rgw >>> uses pools which are replicated 3 times and have a pgnum and a pgpnum >>> greater than 8, what do I need to set? >> >> There are a bunch of pools that are create automatically. Currently >> the best way to avoid them being created with a very low pg number is >> to pre-create them before starting the gateways. Also, there's the >> actual data pool (the pool that holds the user data and bucket >> indexes), which is the .rgw.buckets pool, and which you can modify by >> using the 'radosgw-admin pool pool add/rm. The following is >> (currently) the default pools that are being used. The ability to set >> and modify these will be part of the disaster-recovery/georeplication >> feature. > > Is there an ETA for this work? I suppose seeing it in Cuttlefish isn't likely, is it? > > I'd like the ability to run two (or more) completely distinct gateways on the same rados cluster. Each gateway should have its own (ceph/authx) user and its own set of pools. It sounds like that is not currently possible since the pool names are read-only and apparently global. Is having separate pools per gateway user/instance on the roadmap at all? > >> Current version only allows to view this list. >> >> { "domain_root": ".rgw", >> "control_pool": ".rgw.control", >> "gc_pool": ".rgw.gc", >> "log_pool": ".log", >> "intent_log_pool": ".intent-log", >> "usage_log_pool": ".usage", >> "user_keys_pool": ".users", >> "user_email_pool": ".users.email", >> "user_swift_pool": ".users.swift", >> "user_uid_pool ": ".users.uid"} The ability to configure the pools that are being used will make it to Cuttlefish. The 'radosgw-admin zone info' will replace 'radosgw-admin cluster info', and there will be a new 'radosgw-admin zone set' command that will allow setting that info. In practice you'll also be able to have more than one setup on a single cluster, leveraging the 'rgw root zone pool' configurable. Basically you'd need to do something like this: # radosgw-admin --rgw-root-zone-pool=.zone2.root zone info > zone.info <now edit zone.info, replace the default pool names with new pools> #radosgw-admin --rgw-root-zone-pool=.zone2.root zone set < zone.info and for the second instance run radosgw with --rgw-root-zone-pool=.zone2.root (or put it in your ceph.conf as 'rgw root zone pool = .zone2.root). > > I'm also trying to fine-tune the pools used by our gateway. Obviously pools that store lots of objects and/or data (like .rgw.buckets) should have more than the low default number of placement groups. Is the same true for the rest of the pools? Or in other words, if a pool is only going to contain a few small objects, does it make sense to inflate the number of placement groups in the cluster as a whole? > > If such a distinction makes sense, which pools are typically 'large'? > Well, pools other than the data pool can have smaller number of pgs. Though you'd definitely want to have more than the default 8. Yehuda _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com