Re: Ceph block storage cluster limitations

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> Hello,
> 
> I wanted to know if there are any max limitations on
> 
> - Max number of Ceph data nodes
> - Max number of OSDs per data node
> - Global max on number of OSDs
> - Any limitations on the size of each drive managed by OSD?
> - Any limitation on number of client nodes?
> - Any limitation on maximum number of RBD volumes that can be created?

I don’t think there any *architectural* limits, but there can be *practical* limits.  There are a lot of variables and everyone has a unique situation, but some thoughts:

> Max number of Ceph data nodes

May be limited at some extreme by networking.  Don’t cheap out on your switches.

> - Max number of OSDs per data node

People have run at least 72.  Consider RAM required for a given set of drives, and that a single host/chassis isn’t a big percentage of your cluster.  Ie., don’t have a huge fault domain that will bite you later.  For a production cluster at scale I would suggest at least 12 OSD nodes, but this depends on lots of variables.  Conventional wisdom is 1GB RAM per 1TB of OSD; in practice for a large cluster I would favor somewhat more.  A cluster with, say, 3 nodes of 72 OSDs each is going to be in bad way when one fails.

> - Global max on number of OSDs

A cluster with at lest 10800 has existed.

https://indico.cern.ch/event/542464/contributions/2202295/attachments/1289543/1921810/cephday-dan.pdf
https://indico.cern.ch/event/649159/contributions/2761965/attachments/1544385/2423339/hroussea-storage-at-CERN.pdf

The larger a cluster becomes, the more careful attention must be paid to topology and tuning.

> Also, any advise on using NVMes for OSD drives?

They rock.  Evaluate your servers carefully:
* Some may route PCI through a multi-mode SAS/SATA HBA
* Watch for PCI bridges or multiplexing
* Pinning, minimize data over QPI links
* Faster vs more cores can squeeze out more performance 

AMD Epyc single-socket systems may be very interesting for NVMe OSD nodes.

> What is the known maximum cluster size that Ceph RBD has been deployed to?

See above.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux