Hello Metin, > Problem is, some files can be demanded much more than other files. And > we cant replicate every file on other nodes, the cost would increase > exponentialy. If you really need selective replication take a look at MooseFS. It is open source, posix compliant, runs entirely in user space (both client and servers) and it is production ready. The architecture is similar to Ceph or GoogleFS - metadata server and bunch of storage nodes. You can use just about any unix as a host OS. Clients need FUSE. > So i need a cloud based system, which will push the files into > replicated nodes automatically when demanded to those files are high, > and when the demand is low, they will delete from other nodes and it > will stay in only 1 node. The replication ratio in MooseFS is set per file or per directory via command line tools (at the client). Files and subdirectories inherit their replication ratio from parent directory unless you set it otherwise. Thus your frontend nodes might set the desired replication level per file and then increase or decrease it as needed. Or you you can create several directories, each one set for certain replication ratio and then instruct your frontend servers to move the files around depending on their popularity. Big files are sliced and slices stored (striped) on different nodes to improve aggregate bandwidth. As to the performance, I had only a small test cluster. Single file read performance over gigabit ethernet was very much the same as Samba. When reading several files I was able to saturate gigabit connection at the client. But my cluster was very small (1MDS, 3 OSD), so I cannot really tell how scalable it is. Regards Ales Blaha -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html