Re: high density machines

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On 09/03/2015 02:49 PM, Gurvinder Singh wrote:
Thanks everybody for the feedback.
On 09/03/2015 05:09 PM, Mark Nelson wrote:
My take is that you really only want to do these kinds of systems if you
have massive deployments.  At least 10 of them, but probably more like
20-30+.  You do get massive density with them, but I think if you are
considering 5 of these, you'd be better off with 10 of the 36 drive
units.  An even better solution might be ~30-40 of these:

http://www.supermicro.com/products/system/1U/6017/SYS-6017R-73THDP_.cfm

This one does look interesting.
An extremely compelling solution would be if they took this system:

http://www.supermicro.com/products/system/1U/5018/SSG-5018A-AR12L.cfm?parts=SHOW

This one can be really good solution for archiving purpose with replaced
CPU to get more juice into it.

and replaced the C2750 with a Xeon-D 1540 (but keep the same number of
SATA ports).

Potentially you could have:

- 8x 2.0GHz Xeon Broadwell-DE Cores, 45W TDP
- Up to 128GB RAM (32GB probably the sweet spot)
- 2x 10GbE
- 12x 3.5" spinning disks
- single PCIe slot for PCIe SSD/NVMe
I am wondering does single PCIe SSD/NVMe device can support 12 OSDs
journals and still perform the same as 4 OSD per SSD ?

Basically the limiting factor is how fast the device can do O_DSYNC writes. We've seen that some PCIe SSD and NVME devices can do 1-2GB/s depending on the capacity which is enough to reasonably support 12-24 OSDs. Whether or not it's good to have a single PCIe card to be a point of failure is a worthwhile topic (Probably only high write endurance cards should be considered). There are plenty of other things that can bring the node down too though (motherboard, ram, cpu, etc) though. A single node failure will also have less impact if there are lots of small nodes vs a couple big ones.


The density would be higher than the 36 drive units but lower than the
72 drive units (though with shorter rack depth afaik).
You mean the 1U solution with 12 disk is longer in length than 72 disk
4U version ?

Sorry, the other way around I believe.


- Gurvinder
   Probably more
CPU per OSD and far better distribution of OSDs across servers.  Given
that the 10GbE and processor are embedded on the motherboard, there's a
decent chance these systems could be priced reasonably and wouldn't have
excessive power/cooling requirements.

Mark

On 09/03/2015 09:13 AM, Jan Schermer wrote:
It's not exactly a single system

SSG-F618H-OSD288P*
4U-FatTwin, 4x 1U 72TB per node, Ceph-OSD-Storage Node

This could actually be pretty good, it even has decent CPU power.

I'm not a big fan of blades and blade-like systems - sooner or later a
backplane will die and you'll need to power off everything, which is a
huge PITA.
But assuming you get 3 of these it could be pretty cool!
It would be interesting to have a price comparison to a SC216 chassis
or similiar, I'm afraid it won't be much cheaper.

Jan

On 03 Sep 2015, at 16:09, Kris Gillespie <kgillespie@xxxxxxx> wrote:

It's funny cause in my mind, such dense servers seems like a bad idea to
me for exactly the reason you mention, what if it fails. Losing 400+TB
of storage is going to have quite some impact, 40G interfaces or not and
no matter what options you tweak.
Sure it'll be cost effective per TB, but that isn't the only aspect to
look at (for production use anyways).

But I'd also be curious about real world feedback.

Cheers

Kris

The 09/03/2015 16:01, Gurvinder Singh wrote:
Hi,

I am wondering if anybody in the community is running ceph cluster with
high density machines e.g. Supermicro SYS-F618H-OSD288P (288 TB),
Supermicro SSG-6048R-OSD432 (432 TB) or some other high density
machines. I am assuming that the installation will be of petabyte scale
as you would want to have at least 3 of these boxes.

It would be good to hear their experiences in terms of reliability,
performance (specially during node failures). As these machines have
40Gbit network connection it can be ok, but experience from real users
would be  great to hear. As these are mentioned in the reference
architecture published by red hat and supermicro.

Thanks for your time.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

De informatie verzonden met dit e-mailbericht is uitsluitend bestemd
voor de geadresseerde. Gebruik van deze informatie door anderen dan
de geadresseerde is uitdrukkelijk verboden. Indien u dit bericht per
vergissing heeft ontvangen, verzoeken wij u ons onmiddelijk hiervan
op de hoogte te stellen en het bericht te vernietigen.
Openbaarmaking, vermenigvuldiging, verspreiding en/of verstrekking
van deze informatie aan derden is niet toegestaan. Bol.com b.v. staat
niet in voor de juiste en volledige overbrenging van de inhoud van
een verzonden e-mail, noch voor tijdige ontvangst daarvan en
aanvaardt geen aansprakelijkheid in dezen.
The information contained in this communication is confidential and
may be legally privileged. It is intended solely for the use of the
individual or entity to whom it is addressed and others authorised to
receive it. If you are not the intended recipient please notify the
sender and destroy this message. Any disclosure, copying,
distribution or taking any action in reliance on the contents of this
information is strictly prohibited and may be unlawful. Bol.com b.v.
is neither liable for the proper and complete transmission of the
information contained in this communication nor for delay in its
receipt.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux