Re: Sign, more experimentation and broken features...

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> It may be worth mounting the file system read-only and copying all of
> your data off before you do anything else....

Okay, I managed to borrow enough drives (and SATA ports) to do that.
(The while issue started after I consolidated several separate
drives onto one RAID, then added the source drives to the RAID and
tried to grow the file system.  So the FS corruption was discovered
just a bit too late to go back to the originals.)

> Also, it looks like there may be some problems with the metadata_csum
> option when resizing, either alone or in combination with bigalloc.
> Please note that I have ___not___ really done a lot of exhaustive
> testing with metadata_csum, since it's not in a released final state
> in e2fsprogs, and I've had lots of other things I've been busy trying
> to make sure is stablized.  For example, we are still working on
> fixing various test failures with bigalloc.  It's probably good enough
> for fairly simple workloads (mostly using fallocate and direct I/O),
> but there are corner cases which we are still working on fixing.

I think the big issue is with resizing bigalloc file systems.
Which, as the ext4 wiki page says, currently Doesn't Work.

(One random question I'm curious about is how a larger cluster size
differs from just having a larger block size and why you bothered
creating a new superblock field.  There's probably some discussion
on a mailing list if I search hard enough.)

Anyway, I *think* I have all the drives I intend to add to my big
RAID for a while.  (This is yet another "personal media server" box.)
I now have (borrowed) disjoint source data drives, so I can just create
the RAID and file system in its "final" size.

But it might grow in future.  What I'm trying to decide is if it's worth
risking using bigalloc and relying on resizing becoming reliable in 6
months or so.


With the 2 TB drives that are the best GB/$ these days, it's possible to
break the the 2^32 block limit I ran into last time I complained about
a file system corrupted by resizing.  Even one or two extra bits of
addressing pushes that limit comfortably far away.  (An 8-data-drive
array is practical, a 16-drive array is not really, 32x2 TB would just
be perverse.)

So a large cluster size lets me avoid a 64-bit file system (*another*
new and not-so-well-tested feature).

This RAID probably won't grow beyond 16 TB (it's at 10 TB right now)
before it's time to switch to larger disks, which will probably involve
rebuilding the FS.  But a slightly larger cluster size would make sure
I wouldn't need to start with a 64-bit FS.  (Bigger in-kernel data
structures, and it's *another* not-so-thoroughly-tested feature.)


Any guidance on this question (or general mke2fs parameter suggestions)
is greatly appreciated!
--
To unsubscribe from this list: send the line "unsubscribe linux-ext4" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Reiser Filesystem Development]     [Ceph FS]     [Kernel Newbies]     [Security]     [Netfilter]     [Bugtraq]     [Linux FS]     [Yosemite National Park]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Samba]     [Device Mapper]     [Linux Media]

  Powered by Linux