Inline data is officially an experimental feature. I know of a production cluster that's running with inline data enabled, no problems so far (but it was only enabled two months ago or so). You can reduce the bluestore min alloc size; it's only 16kb for SSDs by default. But the main overhead will then be the metadata required for every object. Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io Tel: +49 89 1896585 90 On Mon, Feb 18, 2019 at 7:06 AM <jesper@xxxxxxxx> wrote: > > Hi Paul. > > Thanks for you comments. > > > For your examples: > > > > 16 MB file -> 4x 4 MB objects -> 4x 4x 1 MB data chunks, 4x 2x 1 MB > > coding chunks > > > > 512 kB file -> 1x 512 kB object -> 4x 128 kB data chunks, 2x 128 kb > > coding chunks > > > > > > You'll run into different problems once the erasure coded chunks end > > up being smaller than 64kb each due to bluestore min allocation sizes > > and general metadata overhead making erasure coding a bad fit for very > > small files. > > Thanks for the clairification, which makes this a "very bad fit" for CephFS: > > # find . -type f -print0 | xargs -0 stat | grep Size | perl -ane '/Size: > (\d+)/; print $1 . "\n";' | ministat -n > x <stdin> > N Min Max Median Avg Stddev > x 12651568 0 1.0840049e+11 9036 2217611.6 > 32397960 > > Gives me 6,3M files < 9036 bytes in size, that'll be stored as 6 x 64KB at > the bluestore > level if I understand it correctly. > > We come from a xfs world where default blocksize is 4K so above situation > worked quite nicely. Guess I probably would be way better off with a > RBD with xfs on top to solve this case using Ceph. > > Is it fair to summarize your input as: > > In a EC4+2 configuration, minimal used space is 256KB+128KB(coding) > regardless of file-size > In a EC8+3 configuraiton, minimal used space is 512KB+192KB(coding) > regardless of file-size > > And for the access side: > All access to files in EC pool requires as a minimum IO-requests to > k-shards for the first > bytes to be returned, with fast_read it becomes k+n, but returns when k > has responded. > > Any experience with inlining data on the MDS - that would obviously help > here I guess. > > Thanks. > > -- > Jesper > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com