Re: CephFS vs Lustre performance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



So despite the performance overhead of replication (or EC + cache tiering) I think CephFS is still a really good solution going forward. We still have a lot of testing/tuning to do, but as you said there are definitely advantages.

I haven't looked closely at either Lustre or Gluster for several years, so I'd prefer not to comment on the state of either these days. :)

Hope that helps!

Mark

On 08/04/2015 05:38 AM, jupiter wrote:
Hi Mark,

Thanks for the comments, that was the same arguments people  concern
CephFS performance here. But one thing I like the Ceph is it is
capable to run everything including replications directly to XFS on
commodity hardware disks, I am not clear if the Lustre can do it as
well, or did you allude that the Lustre has to run on top of the RAID
for replications and fault tolerance?

We are also looking for CephFS and Gluster, apart from the main
difference that Gluster is based on block storage and CephFS is based
on object storage, Ceph is cetainly has much better scalibility, any
insight comments of pros, cons and performance between CephFS and
Gluster?

Thank you and appreciate it.

- jupiter

On 8/3/15, Mark Nelson <mnelson@xxxxxxxxxx> wrote:
On 08/03/2015 06:31 AM, jupiter wrote:
Hi,

I'd like to deploy Cephfs in a cluster, but I need to have a performance
report compared with Lustre and Gluster. Could anyone point me documents
/ links for performance between CephFS, Gluster and Lustre?

Thank you.

Kind regards,

- j

Hi,

I don't know that anything like this really exists yet to be honest.  We
wrote a paper with ORNL several years ago looking at Ceph performance on
a DDN SFA10K and basically saw that we could hit about 6GB/s with CephFS
while Lustre could do closer to 11GB/s.  Primarily that was due to the
journal on the write side (using local SSDs for journal would have
improved things dramatically as the limitation was the IB connections
between the SFA10K and the OSD nodes rather than the disks).  On the
read side we ended up running out of time to figure it out.  We could do
about 8GB/s with RADOS but CephFS was again limited to about 6GB/s.
This was several years ago now so things may have changed.

In general you should expect that Lustre will probably be faster for
large sequential writes (especially if you use Ceph replication vs RAID6
for Lustre) and may be faster for large sequential reads.  For small IO
I suspect that Ceph may do better, and for metadata I would expect the
situation will be mixed with Ceph faster at some things but possibly
slower at others since afaik we haven't done a lot of tuning of the MDS
yet.

Mark





_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux