Re: distributed cluster

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, May 28, 2012 at 4:28 AM, Jerker Nyberg <jerker@xxxxxxxxxxxx> wrote:
> This may not really be a subject ceph-devel mailinglist but rather a
> potential ceph-users? I hope it is ok to write here.

It's absolutely ok to talk on this mailing list about using Ceph. We
may create a separate ceph-users later on, but right now this list is
where the conversation should go.

> Let us assume we have a couple of sites distributed over a metro network
> with at least gigabit interconnect. The demands for storage capacity and
> speed at our sites are increasing together with the demands for reasonably
> stable storage. May Ceph be a port of a solution?

Ceph was designed to work within a single data center. If parts of the
cluster reside in remote locations, you essentially suffer the worst
combination of their latency and bandwidth limits. A write that gets
replicated to three different data centers is not complete until the
data has been transferred to all three, and an acknowledgement has
been received.

For example: with data replicated over data centers A, B, C, connected
at 1Gb/s, the fastest all of A will ever handle writes is 0.5Gb/s --
it'll need to replicate everything to B and C, over that single pipe.

I am aware of a few people building multi-dc Ceph clusters. Some have
shared their network latency, bandwidth and availability numbers with
me (confidentially), and at first glance their wide-area network
performs better than many single-dc networks. They are far above a 1
gigabit interconnect.

I would really recommend you embark on a project like this only if you
are able to understand the Ceph replication model, and do the math for
yourself and figure out what your expected service levels for Ceph
operations would be. (Naturally, Inktank Professional Services will
help you in your endeavors, though their first response should be
"that's not a recommended setup".)

> One idea is to set up Ceph distributed over this metro network. A public
> service network is announced at all sites, anycasted from the storage
> SMB/NFS/RGW(?)-to-Ceph gateway. (for stateless connections). Statefull
> connections (iSCSI?) has to contact the individual storage gateways and
> redundancy is handled at the application level (dual path). Ceph kernel
> clients contact the storage servers directly.

The Ceph Distributed File System is not considered production ready yet.
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux