On Fri, 10 Apr 2015 09:00:40 +0200 Francois Lafont wrote: > Hi Christian, > > Christian Balzer wrote: > > >> Sorry in advance for this thread not directly linked to Ceph. ;) > >> We are thinking about buying servers to build a ceph cluster and we > >> would like to have, if possible, a *approximative* power usage > >> estimation of these servers (this parameter could be important in > >> your choice): > >> > > In short, way, way, way too many variables. > > Which CPUs, HDDs/SSDs, PSUs. > > And a lightly loaded cluster/node will consume like 1/3rd of the power > > CPU wise than a very busy one does. > > Yes indeed. It's just to have a very approximative idea. > > >> 1. the 12xbays supermicro OSD node > >> (here https://www.supermicro.com/solutions/datasheet_Ceph.pdf, > >> page 2, model SSG-6027R-OSD040H in the table) > >> > > I'd really wish SM would revise that pamphlet, for nearly all the > > roles in there they have better suited models. > > And models that fill requirements not really covered in that sheet. > > Ah, err... could you be more precise? Which models have you in the head? > Do you have links? > Depends on your use case really, density, cost, HDDs or SSDs for OSDs, etc. See below for an example. > > If you're willing to take the 1:5 SSD journal to OSD ratio risk, as > > proposed by that configuration, why not go all out to a chassis that > > has 2 hotswap bays in the back and 1:6. Much better density and you'll > > have journals and HDDs on different SATA buses. > > I'm not sure to well understand: the model that I indicated in the link > above (page 2, model SSG-6027R-OSD040H in the table) already have hotswap > bays in the back, for OS drives. > Yes, but that model is pre-configured: 2x 2.5" 400GB SSDs, 10x 3.5" 4TB SATA3 HDDs Rear 2.5" Hot-swap OS drives (mirrored 80GB SSD) What model SSDs and HDDs are those anyway? Instead you could use the basically same thing: http://www.supermicro.com.tw/products/system/2U/6028/SSG-6028R-E1CR12L.cfm And put 12 HDDs (of your choice) in the front and 2 fast and durable SSDs for journals (and OS) in the back. > >> 2. SC216-based chassis 2U, 24xbays 2.5" (like this one for instance > >> http://www.supermicro.com/products/chassis/2U/216/SC216BA-R1K28LP.cfm) > >> > > > > At this level of density, you'd need about 24GHz combined CPU power to > > fully utilize the IOPS potentioal of a pure HDD based node. > > Ok, can I consider that the general rule is ~1Ghz per OSD in HDD (no > separate journal)? > Yes, as per the numerous hardware configuration guides. No separate SSD journal, journal on the same HDD. > > The moment you add SSD journals to this picture, that number at _least_ > > doubles, making it a potentially very power hungry unit. > > So, if I understand well, I should estimate ~2Ghz per OSD with journal in > separate SSD. Is that correct? > At least. With a fio like this inside a VM: --- fio --size=4G --ioengine=libaio --invalidate=1 --direct=1 --numjobs=1 --rw=randwrite --name=fiojob --blocksize=4K --iodepth=32 -- I can make all 8 3.1GHz core on an 8 OSD server with 4 journal SSDs reach 100% utilization. CPU to IOPS ratio is likely to improve as Ceph improves, but this is with Firefly and I doubt even the just released Hammer would change that very much. > > You'll also need a HBA/RAID card to connect up those 6 mini-SAS ports > > on the backplane. > > Is a HBA/RAID systematically necessary? > With the model you cited, yes. All those ports need to be connected up (4 SATA links per port). > Are there some hardware configurations where it could possible to > consider disks directly connected to the controller of the motherboard? > Depends on the motherboard, see the concurrent "Motherboard" thread on this ML. > > If you're concerned about power, look at their X10 offerings with > > Titanium level PSUs and pick CPUs that are energy efficient while > > still having enough capacity to satisfy your IOPS needs. > > Ok. > > >> If someone here has a server as above, we would be curious to have > >> a appromative power usage estimation (for instance in volt-ampere). > >> > > A SM server (not running Ceph, but as a mailbox server being somewhat > > comparable) here with Platinum (94% efficiency supposedly) PSUs > > consumes while being basically idle 105W on the input side (100V in > > Japan) and 95W on the output side. > > This triples basically during peak utilization times. > > Ok, thank you for your help Christian. :) > > PS: it's curious your message doesn't appear in the archive: > http://www.mail-archive.com/ceph-users@xxxxxxxxxxxxxx/msg18699.html > Given how unreliable/delayed the CephML in general is and that the ceph/inktank sites are currently unreachable from here to boot, I'm not particular suprised (or worried). Christian -- Christian Balzer Network/Systems Engineer chibi@xxxxxxx Global OnLine Japan/Fusion Communications http://www.gol.com/ _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com