On Tue, May 21, 2024 at 08:54:26PM +0000, Eugen Block wrote: > It’s usually no problem to shut down a cluster. Set at least the noout flag, > the other flags like norebalance, nobackfill etc won’t hurt either. Then > shut down the servers. I do that all the time with test clusters (they do > have data, just not important at all), and I’ve never had data loss after > powering them back on. When all OSDs are up, unset the flags and let it > scrub. Usually, the (deep-)scrubbing will start almost immediately. One surprise I have with the Ubuntu test cluster (non-containerized, Ubuntu packages) that I regularly shut down is that the signals from log rotation (I assume) to Ceph daemons interfere with Ceph startup. When rebooted after some days, it is the same on all nodes: no Ceph daemon is running. Workaround: another reboot Matthias > > Zitat von "adam.ther" <adam.ther@xxxxxxx>: > > > Thanks guys, > > > > I think ill just risk it cause it's just for backup, then write > > something up later as a follow up on what happens in-case others want to > > do similar. I agree it not typical, im a bit of an odd-duck datahorder. > > > > Regards, > > > > Adam > > > > On 5/21/24 14:21, Matt Vandermeulen wrote: > > > I would normally vouch for ZFS for this sort of thing, but the mix > > > of drive sizes will be... and inconvenience, at best. You could get > > > creative with the hierarchy (making zraid{2,3} of mirrors of > > > same-sized drives, or something), but it would be far from ideal. I > > > use ZFS for my own home machines however, all the drives are > > > identical. > > > > > > I'm curious about this application of Ceph though, in home-lab use. > > > Performance likely isn't a top concern, just a durable persistent > > > storage target, so this is an interesting use case. > > > > > > > > > On 2024-05-21 17:02, adam.ther wrote: > > > > Hello, > > > > > > > > It's all non-corperate data, I'm just trying to cut back on > > > > wattage (removes around 450W of the 2.4KW) by powering down > > > > backup servers that house 208TB while not being backed up or > > > > restoring. > > > > > > > > ZFS sounds interesting however does it play nice with a mix of > > > > drive sizes? That's primarily why I use Ceph, it's okay (if not > > > > ideal) with 4x 22TB, 8x 10TB, 10x 4TB. > > > > > > > > So that said, would Ceph have any known issues with long power > > > > downs aside from it nagging about the scrubbing schedule? Mark i > > > > see you said it wouldn't matter but does Ceph not use a date > > > > based scheduler? > > > > > > > > Thanks, > > > > > > > > Adam > > > > > > > > On 5/21/24 13:29, Marc wrote: > > > > > > > I think it is his lab so maybe it is a test setup for production. > > > > > > Home production? > > > > > A home setup to test on, before he applies changes to his production > > > > > > > > > > Saluti 🍷 ;) > > > > > > > > > _______________________________________________ > > > > ceph-users mailing list -- ceph-users@xxxxxxx > > > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > > > ceph-users mailing list -- ceph-users@xxxxxxx > > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx