Largest Production Ceph Cluster

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Our (DreamHost's) largest cluster is roughly the same size as yours,
~3PB on just shy of 1100 OSDs currently.  The architecture's quite
similar too, except we have "separate" 10G front-end and back-end
networks with a partial spine-leaf architecture using 40G
interconnects.  I say "separate" because the networks only exist in
the logical space; they aren't separated among different bits of
network gear.  Another thing of note is that Ceph will prevent the
marking down of entire racks of OSDs without crazy tweaks of your
CRUSH map (via the mon_osd_down_out_subtree_limit config option).
That actually saved us recently when we suffered a couple of switch
crashes one weekend.


On Tue, Apr 1, 2014 at 7:18 AM, Dan Van Der Ster
<daniel.vanderster at cern.ch> wrote:
> Hi,
>
> On 1 Apr 2014 at 15:59:07, Andrey Korolyov (andrey at xdel.ru) wrote:
>
> On 04/01/2014 03:51 PM, Robert Sander wrote:
>> On 01.04.2014 13:38, Karol Kozubal wrote:
>>
>>> I am curious to know what is the largest known ceph production
>>> deployment?
>>
>> I would assume it is the CERN installation.
>>
>> Have a look at the slides from Frankfurt Ceph Day:
>>
>> http://www.slideshare.net/Inktank_Ceph/scaling-ceph-at-cern
>>
>> Regards
>>
>
> Just curious, how CERN guys built the network topology to prevent
> possible cluster splits, because split in the middle will cause huge
> downtime even for a relatively short split time enough to mark half of
> those 1k OSDs as down by remaining MON majority.
>
>
> The mons are distributed around the data centre, across N switches.
> The OSDs are across a few switches -- actually, we could use CRUSH rules to
> replicate across switches but didn't do so because of an (unconfirmed) fear
> that the uplinks would become a bottleneck.
> So a switch or routing outage scenario is clearly a point of failure where
> some PGs could become stale, but we've been lucky enough not to suffer from
> that yet.
>
> BTW, this 3PB cluster was built to test the scalability of Ceph's
> implementation, not because we have 3PB of data to store in Ceph today (most
> of the results of those tests are discussed in that presentation.). And we
> are currently partitioning this cluster down into a smaller production
> instance for Cinder and other instances for ongoing tests.
>
> BTW#2, I don't think the CERN cluster is the largest. Isn't DreamHost's
> bigger?
>
> Cheers, Dan
>
> -- Dan van der Ster || Data & Storage Services || CERN IT Department --
>
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux