There are a lot of advantages to going bare metal if you can make use of
all of the cores. It's sort of ironic that it's one of the things Ceph
is fairly good at. If you need more parallelism you can throw more OSDs
at the problem. Failure domain and general simplicity have always been
the wildcards that keep me on the side of smaller dense nodes with fewer
cores, straightforward topography, and surprisingly sometimes lower
cost. As far as ARM processors go, it's still pretty wild west imho.
We might be able to get away with it for functional testing, but we
can't necessarily buy ampre and expect that it gives us a good picture
of how we would behave on Graviton2, Grace, M1, or other setups
(especially if we were to eventually target things like GPU erasure
coding offload).
Mark
On 10/11/21 7:56 PM, Dan Mick wrote:
We have some experience testing Ceph on x86 VMs; we used to do that a
lot, but have move to mostly physical hosts. I could be wrong, but I
think our experience is that the cross-loading from one swamped VM to
another on the same physical host can skew the load/failure recovery
testing enough that it's attractive for our normal test strategy/load
to have separate physical hosts.
On 10/11/2021 12:00 AM, Martin Verges wrote:
Hello Dan,
why not using a bit bigger machines and use VMs for tests? We have
quite good experience with that and it works like a charm. If you
plan them as hypervisors, you can run a lot of tests simultaneous.
Use the 80 core ARM, put 512GB or more in them and use some good NVMe
like P55XX or so. In addition put 2*25GbE/40GbE in the servers and
you need only a few of them to simulate a lot. This would save costs,
makes it easier to maintain, and you are much more flexible. For
example running tests on different OS, injecting latency, simulating
errors and more.
--
Martin Verges
Managing director
Mobile: +49 174 9335695 | Chat: https://t.me/MartinVerges
<https://t.me/MartinVerges>
croit GmbH, Freseniusstr. 31h, 81247 Munich
CEO: Martin Verges - VAT-ID: DE310638492
Com. register: Amtsgericht Munich HRB 231263
Web: https://croit.io <https://croit.io> | YouTube:
https://goo.gl/PGE1Bx <https://goo.gl/PGE1Bx>
On Sat, 9 Oct 2021 at 01:25, Dan Mick <dmick@xxxxxxxxxx
<mailto:dmick@xxxxxxxxxx>> wrote:
Ceph has been completely ported to build and run on ARM hardware
(architecture arm64/aarch64), but we're unable to test it due to
lack of
hardware. We propose to purchase a significant number of ARM
servers
(50+?) to install in our upstream Sepia test lab to use for upstream
testing of Ceph, alongside the x86 hardware we already own.
This message is to start a discussion of what the nature of that
hardware should be, and an investigation as to what's available
and how
much it might cost. The general idea is to build something
arm64-based
that is similar to the smithi/gibba nodes:
https://wiki.sepia.ceph.com/doku.php?id=hardware:gibba
<https://wiki.sepia.ceph.com/doku.php?id=hardware:gibba>
Some suggested features:
* base hardware/peripheral support for current releases of RHEL,
CentOS,
Ubuntu
* 1 fast and largish (400GB+) NVME drive for OSDs (it will be
partitioned into 4-5 subdrives for tests)
* 1 large (1TB+) SSD/HDD for boot/system and logs (faster is
better but
not as crucial as for cluster storage)
* Remote/headless management (IPMI?)
* At least 1 10G network interface per host
* Order of 64GB main memory per host
Density is valuable to the lab; we have space but not an unlimited
amount.
Any suggestions on vendors or specific server configurations?
Thanks!
_______________________________________________
Dev mailing list -- dev@xxxxxxx <mailto:dev@xxxxxxx>
To unsubscribe send an email to dev-leave@xxxxxxx
<mailto:dev-leave@xxxxxxx>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx