There are no problems with mixed bluestore_min_alloc_size; that's an abstraction layer lower than the concept of multiple OSDs. (Also, you always have that when mixing SSDs and HDDs) I'm not sure about the real-world impacts of a lower min alloc size or the rationale behind the default values for HDDs (64) and SSDs (16kb). Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io Tel: +49 89 1896585 90 On Mon, Apr 1, 2019 at 10:36 AM Clausen, Jörn <jclausen@xxxxxxxxx> wrote: > > Hi Paul! > > Thanks for your answer. Yep, bluestore_min_alloc_size and your > calculation sounds very reasonable to me :) > > Am 29.03.2019 um 23:56 schrieb Paul Emmerich: > > Are you running on HDDs? The minimum allocation size is 64kb by > > default here. You can control that via the parameter > > bluestore_min_alloc_size during OSD creation. > > 64 kb times 8 million files is 512 GB which is the amount of usable > > space you reported before running the test, so that seems to add up. > > My test cluster is virtualized on vSphere, but the OSDs are reported as > HDDs. And our production cluster also uses HDDs only. All OSDs use the > default value for bluestore_min_alloc_size. > > If we should really consider tinkering with bluestore_min_alloc_size: As > this is probably not tunable afterwards, we would need to replace all > OSDs in a rolling update. Should we expect any problems while we have > OSDs with mixed min_alloc_sizes? > > > There's also some metadata overhead etc. You might want to consider > > enabling inline data in cephfs to handle small files in a > > store-efficient way (note that this feature is officially marked as > > experimental, though). > > http://docs.ceph.com/docs/master/cephfs/experimental-features/#inline-data > > I'll give it a try on my test cluster. > > -- > Jörn Clausen > Daten- und Rechenzentrum > GEOMAR Helmholtz-Zentrum für Ozeanforschung Kiel > Düsternbrookerweg 20 > 24105 Kiel > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com