Hi David, put that way, the docs make complete sense, thank you! i.e. to allow writing to a 5+2 EC cluster with one node down: default is: # ceph osd pool get ecpool min_size min_size: 7 to tolerate one node failure, set: # ceph osd pool set ecpool min_size 6 set pool 1 min_size to 6 to tolerate two nodes failing, set: # ceph osd pool set ecpool min_size 5 set pool 1 min_size to 5 thanks again! Jake On 12/07/17 14:36, David Turner wrote: > As long as you have the 7 copies online if you're using 7+2 then you can > still work and read to the EC pool. For EC pool, size is equivalently 9 > and min_size is 7. > > I have a 3 node cluster with 2+1 and I can restart 1 node at a time with > host failure domain. > > > On Wed, Jul 12, 2017, 6:34 AM Jake Grimmett <jog@xxxxxxxxxxxxxxxxx > <mailto:jog@xxxxxxxxxxxxxxxxx>> wrote: > > Dear All, > > Quick question; is it possible to write to a degraded EC pool? > > i.e. is there an equivalent to this setting for a replicated pool.. > > osd pool default size = 3 > osd pool default min size = 2 > > My reason for asking, is that it would be nice if we could build a EC > 7+2 cluster, and actively use the cluster while a node was off-line by > setting osd noout. > > BTW, Currently testing the Luminous RC, it's looking really nice! > > thanks, > > Jake > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com