Hello! Fri, Aug 28, 2020 at 04:05:55PM +0700, mrxlazuardin wrote: > Hi Konstantin, > > I hope you or anybody still follows this old thread. > > Can this EC data pool be configured per pool, not per client? If we follow > https://docs.ceph.com/docs/master/rbd/rbd-openstack/ we may see that cinder > client will access vms and volumes pools, both with read and write > permission. How can we handle this? > > If we config with different clients for nova (vms) and cinder (volumes), I > think there will be a problem if there is cross pool access, especially on > write. Let's say that client nova will create volume on instance creation > for booting from that volume. Any thoughts? As of this docs, nova will use pools as client.cinder user. When using replicated + erasure pools with cinder I have created two different users for them and two different backends in in cinder.conf for the same cluster with different credentials as rbd_default_data_pool may be set only per-user in ceph.conf. So it was 2 different rdb uids installed in libvirt and 2 different volume types in cinder. As I understand, you need something like my setup. > > Best regards, > > > > Date: Wed, 11 Jul 2018 11:16:27 +0700 > > From: Konstantin Shalygin <k0ste@xxxxxxxx> > > To: ceph-users@xxxxxxxxxxxxxx > > Subject: Re: Erasure coding RBD pool for OpenStack > > Glance, Nova and Cinder > > Message-ID: <069ac368-22b0-3d18-937b-70ce39287cb1@xxxxxxxx> > > Content-Type: text/plain; charset=utf-8; format=flowed > > > > > So if you want, two more questions to you : > > > > > > - How do you handle your ceph.conf configuration (default data pool by > > > user) / distribution ? Manually, config management, openstack-ansible... > > > ? > > > - Did you made comparisons, benchmarks between replicated pools and EC > > > pools, on the same hardware / drives ? I read that small writes are not > > > very performant with EC. > > > > > > ceph.conf with default data pool is only need for Cinder at image > > creation time, after this luminous+ rbd client will be found feature > > "data-pool" and will perform data-io to this pool. > > > > > > > # rbd info > > > erasure_rbd_meta/volume-09ed44bf-7d16-453a-b712-a636a0d3d812 <----- > > > meta pool ! > > > rbd image 'volume-09ed44bf-7d16-453a-b712-a636a0d3d812': > > > ??????? size 1500 GB in 384000 objects > > > ??????? order 22 (4096 kB objects) > > > ??????? data_pool: erasure_rbd_data??????? <----- our data pool > > > ??????? block_name_prefix: rbd_data.6.a2720a1ec432bf > > > ??????? format: 2 > > > ??????? features: layering, exclusive-lock, object-map, fast-diff, > > > deep-flatten, data-pool????????? <----- "data-pool" feature > > > ??????? flags: > > > ??????? create_timestamp: Sat Jan 27 20:24:04 2018 > > > > > > > > k > > > > > > > > ------------------------------ > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx