--- On Wed, 5/5/10, Alex Elsayed <eternaleye@xxxxxxxxx> wrote: Sorry about the accidental off-list reply, thanks for replying on list... > I would recommend benchmarking to have empirical results > rather than going with my presumptions, Or my presumptions, :) agreed. > but in Ceph the metadata servers cache the metadata and > the OSDs journal writes, so any writes which fit in the > journal will be quite fast. Yes, but the kernel will do that also (albeit differently) for local file systems even if the block device is remote. > Also, RBD has no way of knowing what reads/writes are > 'small' in the RBD block device, because it works by splitting > the disk image into 4MB chunks and deals with those. Good point. I assume that is tunable, at least by editing the rbd driver, no? > This is an advantage in the container virtualization case > because you can (say) mount the entire Ceph FS on the > host and treat the containers simply run the containers > from a very basic LXC or other container config, > treating the Ceph filesystem as just another directory > tree from the point of view of the container. > This simplifies your container config, and gives the > advantages I named earlier (online resize, etc). Hmm, while some of those are good advantages (and some may not be depending on your mindset), I am missing the main point as the why this is different than with "real VMs" except for maybe your claim of "not being usual" ... > Ceph has POSIX (or as close as possible) semantics, > matching local filesystems, and provides more features > than any local FS except BtrFS, which is similarly > under heavy development. Perhaps it has many of the features that you want, but there are many things that many other FSs can do (good and bad depending again on your mindset), that ceph cannot and will likely never be able to do. For example, can it be case insensitive such as a DOS FS? > RBD is actually a rather recent addition - the first > mailing list message about it was on March 7th, 2010, > whereas Ceph has been in development since 2007. True, I guess I meant that using the OSDs will likely always be simpler and more stable than using them through ceph. Thanks, -Martin -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html