It’s usually no problem to shut down a cluster. Set at least the noout
flag, the other flags like norebalance, nobackfill etc won’t hurt
either. Then shut down the servers. I do that all the time with test
clusters (they do have data, just not important at all), and I’ve
never had data loss after powering them back on. When all OSDs are up,
unset the flags and let it scrub. Usually, the (deep-)scrubbing will
start almost immediately.
Zitat von "adam.ther" <adam.ther@xxxxxxx>:
Thanks guys,
I think ill just risk it cause it's just for backup, then write
something up later as a follow up on what happens in-case others
want to do similar. I agree it not typical, im a bit of an odd-duck
datahorder.
Regards,
Adam
On 5/21/24 14:21, Matt Vandermeulen wrote:
I would normally vouch for ZFS for this sort of thing, but the mix
of drive sizes will be... and inconvenience, at best. You could get
creative with the hierarchy (making zraid{2,3} of mirrors of
same-sized drives, or something), but it would be far from ideal. I
use ZFS for my own home machines however, all the drives are
identical.
I'm curious about this application of Ceph though, in home-lab use.
Performance likely isn't a top concern, just a durable persistent
storage target, so this is an interesting use case.
On 2024-05-21 17:02, adam.ther wrote:
Hello,
It's all non-corperate data, I'm just trying to cut back on
wattage (removes around 450W of the 2.4KW) by powering down backup
servers that house 208TB while not being backed up or restoring.
ZFS sounds interesting however does it play nice with a mix of
drive sizes? That's primarily why I use Ceph, it's okay (if not
ideal) with 4x 22TB, 8x 10TB, 10x 4TB.
So that said, would Ceph have any known issues with long power
downs aside from it nagging about the scrubbing schedule? Mark i
see you said it wouldn't matter but does Ceph not use a date based
scheduler?
Thanks,
Adam
On 5/21/24 13:29, Marc wrote:
I think it is his lab so maybe it is a test setup for production.
Home production?
A home setup to test on, before he applies changes to his production
Saluti 🍷 ;)
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx