Hello community,
I am building first cluster for project, that hosts millions of small
(from 20kb) and big (up to 10mb) files. Right now we are moving from
local 16tb raid storage to cluster of 12 small machines. We are
planning to have 11 OSD nodes, use erasure coding pool (10+1) and one
host for MDS.
On my local tests I see, that available space decrease unproportionally
to the amount of data copied into cluster. With clean cluster I have,
for example 100gb available space, but after copying 40gb in - size
decreases for about 5-10%. Is that normal?
Is there any term, that would specify cluster's minimal object size?
I also have question if having so many small files (current number is
about 50'000'000 files at least) - could have negative impact and where
would be our bottleneck? As we don't have money for SSD, we will have
WAL/DB on separate simple HDD.
Also - would that help to put Metadata pool on separate disks, away from
Data pool drives for CephFS?
Regards,
Anton.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com