glusterFS, many small files and data replication

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello list members,

at the moment we are searching for a storage cluster solution that
should fulfill the following specifications:
* 3 cluster nodes, each node should have 24x SATA disks (750GB) in a RAID 6
* data replication (each file stored on at least 2 of 3 nodes)
* easy desaster recovery (in case a node crashes fully)
* high speed read access to many small files at the same time
(20.000.000-200.000.000 files of sizes 5kB to 60kB)
* files are delivered by a web server
* average writing speed
* system should even work if one(two) storage node is completely down

At the moment we think most of this is already possible with
glusterFS. Only the recovery part is not yet possible (should be done
with 1.4).
But how does glusterFS perform with that many small files? I could
imagine that glusterFS is not optimized for that usage because it does
not have a meta data server that helps the clients to find the right
server to ask for a file.

Is glusterFS the right system for us?


Best wishes

Bernhard J. M. Gruen




[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux