Re: 6 Node cluster with 24 SSD per node: Hardwareplanning/ agreement

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,


some thoughts about network and disks inline

On 10/04/2016 03:43 PM, Denny Fuchs wrote:
Hello,

*snipsnap*


* Storage NIC: 1 x Infiniband MCX314A-BCCT
** I red, that ConnectX-3 Pro is better supported, than the X-4 and a bit cheaper
** Switch: 2 x Mellanox SX6012 (56Gb/s)
** Active FC cables
** Maybe VPI is nice to have, but unsure.
The Infiniband support is Ceph is experimental and not recommended for production uses. You'll have to fall back to IPoIB for the moment. The ConnectX-3 has configurable ports and also supports 40GbE, so ethernet switches might be an alternative for your setup. (some of the Mellanox switch support both infiniband and ethernet).

* Production NIC: 1 x Intel 520 dual port SFP+
** Connected each to one of a HP 2920 10Gb/s ports via 802.3ad

All nodes are connected over cross to every switch, so if one switch goes down, a second path is available.


* Disk:
** Storage: 24 x Crucial MX300 250GB (maybe for production 12xSSD / 12x big Sata disks)
** OSD journal: 1 x Intel SSD DC P3700 PCIe

Not sure about the OSD SSDs, but keep in mind that consumer SSDs are not intended for running under load 24/7. Have a closer look at their write endurance and ensure that the SSDs are monitored properly. If you take ll of them into production at the same time and got a good data distribution over all SSDs, they might also be failing within a very short time span...

The journal SSD is OK (we have the same model), but according to tests it is only capable of writing about 1 GB/s as journal SSD (obligatory blog link: http://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-ssd-is-suitable-as-a-journal-device/). With 2x 10GbE public network links, the SSD might become the bottleneck in large scale write operations.

For read operations with 24 SSDs (assuming 400MB/s per SSD -> 9,6 GB/s) the network will definitely become the bottleneck. You also might want to check whether the I/O subsystem is able to drive 24 SSDs (SAS-3 has 12 GBit/s, expander are usually connected with 4 channels -> 6 GB/s).

Regards,
Burkhard

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux