On 11/08/2020 05:42, George Rapp wrote:
Hello Linux RAID community - I've been running an assortment of software RAID arrays for a while now (my oldest Creation Time according to 'mdadm --detail' is April 2011) but have been meaning to consolidate my five active arrays into something easier to manage. This weekend, I finally found enough cheap 2TB disks to get started. I'm planning on creating a RAID 6 array due to the age and consumer-grade quality of my 16 2TB disks.
SCT/ERC ???
Use case is long-term storage of many small files and a few large ones (family photos and videos, backups of other systems, working copies of photo, audio, and video edits, etc.)? Current usable space is about 10TB but my end state vision is probably upwards of 20TB. I'll probably consign the slowest working disks in the server to an archive filesystem, either RAID 1 or RAID 5, for stuff I care less about and backups; the archive part can be ignored for the purposes of this exercise.
If you haven't got ERC, I'd be more inclined to raid-10 than raid 6. Your 16 disks would give you 10TB if you used a 3-way mirror, or 15TB if it's a 2-way.
My question is: what filesystem type would be best practice for my use case and size requirements on the big array? (I have reviewed https://raid.wiki.kernel.org/index.php/RAID_and_filesystems, but am looking for practitioners' recommendations.) I've run ext4 exclusively on my arrays to date, but have been reading up on xfs; is there another filesystem type I should consider? Finally, are there any pitfalls I should know about in my high-level design?
Think about dm-integrity and LVM. Take a look at https://raid.wiki.kernel.org/index.php/System2020
I'm working my way through with that system, so that page is an incomplete mess of musings at the moment but it might give you ideas.
File systems? Ext4 is a good choice. Remember that filesystems like btrfs and zfs are trying to replace raid, lvm etc and subsume it all into the filesystem. Do you want a layered KISS setup, or an all-things-to-all-men filesystem.
And look at slowly upgrading your disks to decent raid-happy 4TB drives or so ... Ironwolves aren't that expensive ...
Details: # uname -a Linux backend5 5.7.11-200.fc32.x86_64 #1 SMP Wed Jul 29 17:15:52 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux # mdadm --version mdadm - v4.1 - 2018-10-01 Finally, and this is inadequate given the value I've received from the efforts of this group, but thanks for many years of supporting mdadm and helping with software RAID issues, including the recovery procedures you have written up and guided me through. This group's efforts have saved my data, my bacon, and my sanity on more than one occasion.
Thanks very much :-) Cheers, Wol