Jan, I would like to echo Sage's response here. It seems you only want a subset of what Ceph offers, whereas RADOS is designed to offer a whole lot more, which requires a lot more intelligence at the lower levels. I must say I have found your attitude to both Sage and the Ceph project as a whole over the last few emails quite disrespectful. I spend a lot of my time trying to sell the benefits of open source, which centre on the openness of the idea/code and not around the fact that you can get it for free. One of the things that I like about open source is the constructive, albeit sometimes abrupt, constructive criticism that results in a better product. Simply shouting Ceph is slow and it's because dev's don't understand filesystems is not constructive. I've just come back from an expo at ExCel London where many providers are passionately talking about Ceph. There seems to be a lot of big money sloshing about for something that is inherently "wrong" Sage and the core Ceph team seem like very clever people to me and I trust that over the years of development, that if they have decided that standard FS's are not the ideal backing store for Ceph, that this is probably correct decision. However I am also aware that the human condition "Can't see the wood for the trees" is everywhere and I'm sure if you have any clever insights into filesystem behaviour, the Ceph Dev team would be more than open to suggestions. Personally I wish I could contribute more to the project as I feel that I (any my company) get more from Ceph than we put in, but it strikes a nerve when there is such negative criticism for what effectively is a free product. Yes, I also suffer from the problem of slow sync writes, but the benefit of being able to shift 1U servers around a Rack/DC compared to a SAS tethered 4U jbod somewhat outweighs that as well as several other advanatages. A new cluster that we are deploying has several hardware choices which go a long way to improve this performance as well. Coupled with the coming Bluestore, the future looks bright. > -----Original Message----- > From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of > Sage Weil > Sent: 12 April 2016 21:48 > To: Jan Schermer <jan@xxxxxxxxxxx> > Cc: ceph-devel <ceph-devel@xxxxxxxxxxxxxxx>; ceph-users <ceph- > users@xxxxxxxx>; ceph-maintainers@xxxxxxxx > Subject: Re: Deprecating ext4 support > > On Tue, 12 Apr 2016, Jan Schermer wrote: > > Still the answer to most of your points from me is "but who needs that?" > > Who needs to have exactly the same data in two separate objects > > (replicas)? Ceph needs it because "consistency"?, but the app (VM > > filesystem) is fine with whatever version because the flush didn't > > happen (if it did the contents would be the same). > > If you want replicated VM store that isn't picky about consistency, try > Sheepdog. Or your mdraid over iSCSI proposal. > > We care about these things because VMs are just one of many users of > rados, and because even if we could get away with being sloppy in some (or > even most) cases with VMs, we need the strong consistency to build other > features people want, like RBD journaling for multi-site async replication. > > Then there's the CephFS MDS, RGW, and a pile of out-of-tree users that > chose rados for a reason. > > And we want to make sense of an inconsistency when we find one on scrub. > (Does it mean the disk is returning bad data, or we just crashed during a write > a while back?) > > ... > > Cheers- > sage > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com