2GB ram is gonna be really tight, probably. However, I do something similar at home with a bunch of rock64 4gb boards, and it works well. There are sometimes issues with the released ARM packages (frequently crc32 doesn;'t work, which isn't great), so you may have to build your own on the board you're targeting or on something like scaleway, YMMV.
On Fri, Sep 6, 2019 at 6:16 PM Cranage, Steve <scranage@xxxxxxxxxxxxxxxxxxxx> wrote:
_______________________________________________I use those HC2 nodes for my home Ceph cluster, but my setup only has to support the librados API, my software does HSM between regular XFS file systems and the RADOS api so I don’t need the other MDS and the rest so I can’t tell you if you’ll be happy in your configuration.
Steve Cranage
Principal Architect, Co-Founder
DeepSpace Storage
719-930-6960
From: ceph-users <ceph-users-bounces@xxxxxxxxxxxxxx> on behalf of William Ferrell <willfe@xxxxxxxxx>
Sent: Friday, September 6, 2019 3:16:30 PM
To: ceph-users@xxxxxxxxxxxxxx <ceph-users@xxxxxxxxxxxxxx>
Subject: Ceph for "home lab" / hobbyist use?Hello everyone!
After years of running several ZFS pools on a home server and several
disk failures along the way, I've decided that my current home storage
setup stinks. So far there hasn't been any data loss, but
recovering/"resilvering" a ZFS pool after a disk failure is a
nail-biting experience. I also think the way things are set up now
isn't making the best use of all the disks attached to the server;
they were acquired over time instead of all at once, so I've got 4
4-disk raidz1 pools, each in their own enclosures. If any enclosure
dies, all that pool's data is lost. Despite having a total of 16 disks
in use for storage, the entire system can only "safely" lose one disk
before there's a risk of a second failure taking a bunch of data with
it.
I'd like to ask the list's opinions on running a Ceph cluster in a
home environment as a filer using cheap, low-power systems. I don't
have any expectations for high performance (this will be built on a
gigabit network, and just used for backups and streaming videos,
music, etc. for two people); the main concern is resiliency if one or
two disks fail, and the secondary concern is having a decent usable
storage capacity. Being able to slowly add capacity to the cluster one
disk at a time is a very appealing bonus.
I'm interested in using these things as OSDs (and hopefully monitors
and metadata servers):
https://www.hardkernel.com/shop/odroid-hc2-home-cloud-two/
They're about $50 each, can boot from MicroSD or eMMC flash (basically
an SSD with a custom connector), and have one SATA port. They have
8-core 32-bit CPUs, 2GB of RAM and a gigabit ethernet port. Four of
them (including disks) can run off a single 12V/8A power adapter
(basically 100 watts per set of 4). The obvious appeal is price, plus
they're stackable so they'd be easy to hide away in a closet.
Is it feasible for these to work as OSDs at all? The Ceph hardware
recommendations page suggests OSDs need 1GB per TB of space, so does
this mean these wouldn't be suitable with, say, a 4TB or 8TB disk? Or
would they work, but just more slowly?
Pushing my luck further (assuming the HC2 can handle OSD duties at
all), is that enough muscle to run the monitor and/or metadata
servers? Should monitors and MDS's be run separately, or can/should
they piggyback on hosts running OSDs?
I'd be perfectly happy with a setup like this even if it could only
achieve speeds in the 20-30MB/sec range.
Is this a dumb idea, or could it actually work? Are there any other
recommendations among Ceph users for low-end hardware to cobble
together a working cluster?
Any feedback is sincerely appreciated.
Thanks!
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com