Re: high density machines

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



My take is that you really only want to do these kinds of systems if you have massive deployments. At least 10 of them, but probably more like 20-30+. You do get massive density with them, but I think if you are considering 5 of these, you'd be better off with 10 of the 36 drive units. An even better solution might be ~30-40 of these:

http://www.supermicro.com/products/system/1U/6017/SYS-6017R-73THDP_.cfm

An extremely compelling solution would be if they took this system:

http://www.supermicro.com/products/system/1U/5018/SSG-5018A-AR12L.cfm?parts=SHOW

and replaced the C2750 with a Xeon-D 1540 (but keep the same number of SATA ports).

Potentially you could have:

- 8x 2.0GHz Xeon Broadwell-DE Cores, 45W TDP
- Up to 128GB RAM (32GB probably the sweet spot)
- 2x 10GbE
- 12x 3.5" spinning disks
- single PCIe slot for PCIe SSD/NVMe

The density would be higher than the 36 drive units but lower than the 72 drive units (though with shorter rack depth afaik). Probably more CPU per OSD and far better distribution of OSDs across servers. Given that the 10GbE and processor are embedded on the motherboard, there's a decent chance these systems could be priced reasonably and wouldn't have excessive power/cooling requirements.

Mark

On 09/03/2015 09:13 AM, Jan Schermer wrote:
It's not exactly a single system

SSG-F618H-OSD288P*
4U-FatTwin, 4x 1U 72TB per node, Ceph-OSD-Storage Node

This could actually be pretty good, it even has decent CPU power.

I'm not a big fan of blades and blade-like systems - sooner or later a backplane will die and you'll need to power off everything, which is a huge PITA.
But assuming you get 3 of these it could be pretty cool!
It would be interesting to have a price comparison to a SC216 chassis or similiar, I'm afraid it won't be much cheaper.

Jan

On 03 Sep 2015, at 16:09, Kris Gillespie <kgillespie@xxxxxxx> wrote:

It's funny cause in my mind, such dense servers seems like a bad idea to
me for exactly the reason you mention, what if it fails. Losing 400+TB
of storage is going to have quite some impact, 40G interfaces or not and
no matter what options you tweak.
Sure it'll be cost effective per TB, but that isn't the only aspect to
look at (for production use anyways).

But I'd also be curious about real world feedback.

Cheers

Kris

The 09/03/2015 16:01, Gurvinder Singh wrote:
Hi,

I am wondering if anybody in the community is running ceph cluster with
high density machines e.g. Supermicro SYS-F618H-OSD288P (288 TB),
Supermicro SSG-6048R-OSD432 (432 TB) or some other high density
machines. I am assuming that the installation will be of petabyte scale
as you would want to have at least 3 of these boxes.

It would be good to hear their experiences in terms of reliability,
performance (specially during node failures). As these machines have
40Gbit network connection it can be ok, but experience from real users
would be  great to hear. As these are mentioned in the reference
architecture published by red hat and supermicro.

Thanks for your time.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

De informatie verzonden met dit e-mailbericht is uitsluitend bestemd voor de geadresseerde. Gebruik van deze informatie door anderen dan de geadresseerde is uitdrukkelijk verboden. Indien u dit bericht per vergissing heeft ontvangen, verzoeken wij u ons onmiddelijk hiervan op de hoogte te stellen en het bericht te vernietigen. Openbaarmaking, vermenigvuldiging, verspreiding en/of verstrekking van deze informatie aan derden is niet toegestaan. Bol.com b.v. staat niet in voor de juiste en volledige overbrenging van de inhoud van een verzonden e-mail, noch voor tijdige ontvangst daarvan en aanvaardt geen aansprakelijkheid in dezen.
The information contained in this communication is confidential and may be legally privileged. It is intended solely for the use of the individual or entity to whom it is addressed and others authorised to receive it. If you are not the intended recipient please notify the sender and destroy this message. Any disclosure, copying, distribution or taking any action in reliance on the contents of this information is strictly prohibited and may be unlawful. Bol.com b.v. is neither liable for the proper and complete transmission of the information contained in this communication nor for delay in its receipt.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux