On Tue, May 29, 2012 at 12:25 AM, Quenten Grasso <QGrasso@xxxxxxxxxx> wrote: > So if we have 10 nodes vs. 3 nodes with the same mount of disks we should see better write and read performance as you would have less "overlap". First of all, a typical way to run Ceph is with say 8-12 disks per node, and an OSD per disk. That means your 3-10 node clusters actually have 24-120 OSDs on them. The number of physical machines is not really a factor, number of OSDs is what matters. Secondly, 10-node or 3-node clusters are fairly uninteresting for Ceph. The real challenge is at the hundreds, thousands and above range. > Now we take BTRFS into the picture as I understand journals are not necessary due to the nature of the way it writes/snapshots and reads data this alone would be a major performance increase on a BTRFS Raid level (like ZFS RAIDZ). A journal is still needed on btrfs, snapshots just enable us to write to the journal in parallel to the real write, instead of needing to journal first. -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html